Once you have configured the Solaris cluster , you have to add the quorum device for additional voting process. Without Quorum devices, cluster will be in installed mode. You can verify the status using “cluster show -t global | grep installmode” command.Each node in a configured cluster has one ( 1 ) quorum vote and cluster require minimum two vote to run the cluster.If any one node goes down, cluster won’t get 2 votes and it will panic second node also to avoid the data corruption on shared storage. To avoid this situation, we can make one small SAN disk as quorum device which can provide one vote.So that, if one node fails, system can still get two votes all the time on two node cluster.
Once you have configured the two node Solaris cluster, you can start configure the quorum device.
1.Check the cluster node status.
UASOL1:#clnode status === Cluster Nodes === --- Node Status --- Node Name Status --------- ------ UASOL2 Online UASOL1 Online UASOL1:#
2.You can see that ,currently cluster is in install mode.
# cluster show -t global | grep installmode installmode: enabled
3.Current cluster quorum status
UASOL1:#clq status === Cluster Quorum === --- Quorum Votes Summary from (latest node reconfiguration) --- Needed Present Possible ------ ------- -------- 1 1 1 --- Quorum Votes by Node (current status) --- Node Name Present Possible Status --------- ------- -------- ------ UASOL2 1 1 Online UASOL1 0 0 Online UASOL1:#
4.Make sure you have small size LUN is assigned to both the cluster node from SAN.
UASOL1:#echo |format Searching for disks...done AVAILABLE DISK SELECTIONS: 0. c1t0d0 VMware,-VMware Virtual -1.0 cyl 1824 alt 2 hd 255 sec 63 /pci@0,0/pci15ad,1976@10/sd@0,0 1. c1t1d0 VMware,-VMware Virtual -1.0 cyl 508 alt 2 hd 64 sec 32 /pci@0,0/pci15ad,1976@10/sd@1,0 Specify disk (enter its number): Specify disk (enter its number): UASOL1:#
5.Let me label the disk and naming the disk.
UASOL1:#format c1t1d0 selecting c1t1d0: quorum [disk formatted] FORMAT MENU: disk - select a disk type - select (define) a disk type partition - select (define) a partition table current - describe the current disk format - format and analyze the disk fdisk - run the fdisk program repair - repair a defective sector label - write label to the disk analyze - surface analysis defect - defect list management backup - search for backup labels verify - read and display labels save - save new disk/partition definitions inquiry - show vendor, product and revision volname - set 8-character volume name ! - execute , then return quit format> fdisk The default partition for the disk is: a 100% "SOLARIS System" partition Type "y" to accept the default partition, otherwise type "n" to edit the partition table. y format> volname quorum format> quit UASOL1:#
6.You can see the same LUN on UASOL2 node as well.
UASOL2:#echo |format Searching for disks...done AVAILABLE DISK SELECTIONS: 0. c1t0d0 VMware,-VMware Virtual -1.0 cyl 1824 alt 2 hd 255 sec 63 /pci@0,0/pci15ad,1976@10/sd@0,0 1. c1t1d0 VMware,-VMware Virtual -1.0 cyl 508 alt 2 hd 64 sec 32 quorum /pci@0,0/pci15ad,1976@10/sd@1,0 Specify disk (enter its number): Specify disk (enter its number): UASOL2:#
7. Populate the disks in solaris cluster.
UASOL2:#cldev populate Configuring DID devices did instance 4 created. did subpath UASOL2:/dev/rdsk/c1t1d0 created for instance 4. Configuring the /dev/global directory (global devices) obtaining access to all attached disks UASOL2:# UASOL1:#cldev populate Configuring DID devices did instance 4 created. did subpath UASOL1:/dev/rdsk/c1t1d0 created for instance 4. Configuring the /dev/global directory (global devices) obtaining access to all attached disks UASOL1:#
8.Check the devices status.
UASOL1:#cldevice list -v DID Device Full Device Path ---------- ---------------- d1 UASOL2:/dev/rdsk/c1t0d0 d1 UASOL1:/dev/rdsk/c1t0d0 d4 UASOL2:/dev/rdsk/c1t1d0 d4 UASOL1:/dev/rdsk/c1t1d0 UASOL1:#cldev show d4 === DID Device Instances === DID Device Name: /dev/did/rdsk/d4 Full Device Path: UASOL1:/dev/rdsk/c1t1d0 Full Device Path: UASOL2:/dev/rdsk/c1t1d0 Replication: none default_fencing: global UASOL1:#
9.Add the d4 as quorum device in cluster.
UASOL1:#clquorum add d4 UASOL1:#
10.Check the Quorum status
UASOL1:#clq status === Cluster Quorum === --- Quorum Votes Summary from (latest node reconfiguration) --- Needed Present Possible ------ ------- -------- 2 3 3 --- Quorum Votes by Node (current status) --- Node Name Present Possible Status --------- ------- -------- ------ UASOL2 1 1 Online UASOL1 1 1 Online --- Quorum Votes by Device (current status) --- Device Name Present Possible Status ----------- ------- -------- ------ d4 1 1 Online UASOL1:#
We have successfully configured the quorum on two node Solaris cluster 3.3 u2.
How can we test quorum device is working or not ?
Just reboot any one of the node and you can see the voting status .
UASOL2:#reboot updating /platform/i86pc/boot_archive Connection to UASOL2 closed by remote host. Connection to UASOL2 closed. UASOL1:# UASOL1:#clq status === Cluster Quorum === --- Quorum Votes Summary from (latest node reconfiguration) --- Needed Present Possible ------ ------- -------- 2 2 3 --- Quorum Votes by Node (current status) --- Node Name Present Possible Status --------- ------- -------- ------ UASOL2 0 1 Offline UASOL1 1 1 Online --- Quorum Votes by Device (current status) --- Device Name Present Possible Status ----------- ------- -------- ------ d4 1 1 Online UASOL1:#
We can see that UASOL1 is not panic by cluster. So quorum device worked well.
If you don’t have real SAN storage for shared LUN, you can use openfiler.
What’s Next ? We will configure resource group for failover local zone and perform the test.
Share it ! Comment it !! Be Sociable !!