Hi all,I build two node master/master esos cluster, all working good.I have problem with quorum on clvm. I add two_node: 1 to corosync.conf and also no-quorum-policy=ignore in pacemaker, but clvm still waiting for quorum. Clvm is not managed by pacemaker, because clvm RA is missing in esos, and I don't know, if RA solve this problem.
So, with one node down, storage is unavailable. Solution should be quorum disk, but mkqdisk command is missing in esos.
Another question, what is advantage in master/master vs master/slave config? SCST daemon still must run in master/slave, so ALL initiators read/write to only one node (no performance benefit) or there is some override to spread initiators across storage nodes?
If there is no benefit then will be better configure master/slave cluster without use of clvm.
Thanks for replies.Best regards Feldsam.
--
You received this message because you are subscribed to the Google Groups "esos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to esos-users+...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
On Fri, May 1, 2015 at 1:10 PM, Kristián Feldsam <fel...@gmail.com> wrote:Hi all,I build two node master/master esos cluster, all working good.I have problem with quorum on clvm. I add two_node: 1 to corosync.conf and also no-quorum-policy=ignore in pacemaker, but clvm still waiting for quorum. Clvm is not managed by pacemaker, because clvm RA is missing in esos, and I don't know, if RA solve this problem.Its been quite a while since I've done anything with clvm but if I remember correctly, it doesn't have anything to do with Corosync/Pacemaker... its simply a front-end to locking so if you have LVM shared between nodes (LVM cluster) when doing LVM management/provisioning tasks like PV/VG/LV create/destroy/etc. you can do those tasks safely (from one node). The activation of volume groups is handled by the "LVM" RA. And if I recall correctly, clvm relies on DLM (distributed lock manager) so rc.dlm needs to be enabled/running as well, otherwise I think clvm will just hanging waiting.
So, with one node down, storage is unavailable. Solution should be quorum disk, but mkqdisk command is missing in esos.If something relevant is missing from ESOS, I'd like to add it. What package does 'mkqdisk' come from?
Another question, what is advantage in master/master vs master/slave config? SCST daemon still must run in master/slave, so ALL initiators read/write to only one node (no performance benefit) or there is some override to spread initiators across storage nodes?There should never be master/master for SCST across more than 1 node... the blocks of data can be replicated between between the two nodes, but other SCSI functions like SCSI reservations and persistent reservations are held at the SCST layer, and those are NOT replicated between nodes. So if an initiator issues a SCSI reservation and it goes to one host, and then expects to release it on the other host (another path) its not gonna work... very bad.In order to bring more performance, you can still use multipath (like round-robin) across target interfaces that belong to the same host, not round-robin I/O across targets on different hosts. So say I have two nodes in the cluster, and each has (2) 4 Gb Fibre Channel target interfaces. Configure ALUA with the two targets in the "local" group and the two targets for the other host in the "remote" group and then your initiators can use a round-robin policy and they should obey and only send data across the two paths that belong to one node... the two paths marked as "active" and the two paths on the other host are "nonoptimized" (or whatever state you choose). So you now effectively get 8 Gb of bandwidth since its going across 2 target interfaces. I've tested this personally with vSphere 5.5 and can changing the pathing policy when doing IO tests in a VM from single path to round-robin across multiple, and its very clear to see the performance gain (the bandwidth nearly doubles when doing a large I/O sequential test). Test the multipath setup with other types of initiators, but I will say it works correctly from my experience in vSphere.
If there is no benefit then will be better configure master/slave cluster without use of clvm.Are you wanting to use clvm just for LVM "cluster"? That's the only thing I know it as, but maybe there is another use? Hope this helps.
I know I know, I read docs, but I am asking for performace (or any other) benefit of two master/master cluster nodes. I use multipath, I have two dual port cards in both nodes, so 4 + 4 targets. I follow your article from 2013, where you setup drbd master/master, clone lvm with clvm and scst master/slave. So WHY go for master/master drbd and clvm when initiators can use only one node? Would not be better to use master/slave drbd + just simple lvm with locking_type = 1 without dlm and clvm, and master/slave scst? I don't see any advantage of master/master setup when scst must be master/slave.