>On 2013/8/27 wrote
>Are you using san with emc powerpath? Yes >drbd service start on boot? >start drbd by hand and give the command ls -l dev/drbd/by-res/r0/1 and ls -l It works fine if we do things manually. The only issue comes when we try to run things under rgmanager. >>On 2013/8/27 Goutam Baul wrote >>Dear List, >>We have configured DRBD 8.4.3 (community version) in the sense that we are able to manually start DRBD at both the nodes, make nodes primary or secondary and >>data is getting replicated between the primary and the secondary node. We are doing this for our postfix mail server. The drbd configuration file is >>resource r0 { >>startup { >> wfc-timeout 1; >> degr-wfc-timeout 2; >> } >> volume 0 { >> device /dev/drbd1; >> disk /dev/emcpowera2; >> meta-disk /dev/emcpowera5; >> } >> volume 1 { >> device /dev/drbd2; >> disk /dev/emcpowera3; >> meta-disk /dev/emcpowera6; >> } >> on drmail1.cesc.co.in { >> address 10.50.4.14:7789; >> } >> on drbdtest2.cesc.co.in { >> address 10.50.81.253:7789; >> } >>} >>We want to run postfix mailing system on a RHCS cluster suite for high availability. The mail-store and the mail-queue of postfix will be on the DRBD setup so that >>they are replicated from our data center to our disaster recovery site. The cluster.conf is as follows: >><?xml version="1.0"?> >><cluster config_version="27" name="dr-mail"> >> <fence_daemon clean_start="1" post_join_delay="30"/> >> <clusternodes> >> <clusternode name="drmail1.cesc.co.in" nodeid="1"> >> <fence> >> <method name="Method1_drac"> >> <device name="fence_drac1"/> >> </method> >> </fence> >> </clusternode> >> <clusternode name="drmail2.cesc.co.in" nodeid="2"> >> <fence> >> <method name="Method1_drac"> >> <device name="fence_drac2"/> >> </method> >> </fence> >> </clusternode> >> </clusternodes> >> <cman expected_votes="1" two_node="1"/> >> <fencedevices> >> <fencedevice agent="fence_drac5" cmd_prompt="admin1->" ipaddr="10.50.4.22" login="root" module_name="fence_drac1" name="fence_drac1" passwd="calvin" secure="on"/> >> <fencedevice agent="fence_drac5" cmd_prompt="admin1->" ipaddr="10.50.4.23" login="root" module_name="fence_drac2" name="fence_drac2" passwd="calvin" secure="on"/> >> </fencedevices> >> <rm> >> <failoverdomains> >> <failoverdomain name="dr_mail" nofailback="0" ordered="1" restricted="1"> >> <failoverdomainnode name="drmail1.cesc.co.in" priority="1"/> >> <failoverdomainnode name="drmail2.cesc.co.in" priority="2"/> </failoverdomain> >> </failoverdomains> >> <resources> >> <ip address="10.50.4.20" monitor_link="on" sleeptime="10"/> >> <drbd name="drbd-mail" resource="r0"> >> <fs device="/dev/drbd/by-res/r0/0" force_unmount="1" fsid="61850" mountpoint="/home" fstype="ext3" name="drbd-mailstore" options="noatime"/> >> <fs device="/dev/drbd/by-res/r0/1" force_unmount="1" fsid="61855" mountpoint="/var/spool/postfix" fstype="ext3" name="drbd-spool" options="noatime"/> >> </drbd> >> <script file="/etc/init.d/postfix" name="postfix" /> >> </resources> >> <service domain="dr_mail" name="mail" recovery="relocate"> >> <ip ref="10.50.4.20"/> >> <drbd ref="drbd-mail" /> <script ref="postfix" /> >> </service> >> </rm> >> </cluster> >> If we run this RHCS cluster, we get the error in the log as >> Aug 27 17:38:19 drmail1 rgmanager[4695]: drbd not found in /proc/modules. Do you need to modprobe? >> If we run the RHCS cluster after issuing the modprobe drbd command we get the following message >> 1: Failure: (127) Device minor not allocated >> additional info from kernel: >> unknown minor >> Command 'drbdsetup role 1' terminated with exit code 10 >> <debug> DRBD resource r0 already configured >> DRBD resource r0 already configured >> Start of mail complete >> But the DRBD service is not starting up and therefore the corresponding devices are not being created. Thus the log file returns the error >> Aug 27 17:38:33 drmail1 rgmanager[5147]: stop: Could not match /dev/drbd/by-res/r0/1 with a real device >> Aug 27 17:38:33 drmail1 rgmanager[5166]: stop: Could not match /dev/drbd/by-res/r0/0 with a real device >> We have built the DRBD rpm files from the source using rpmbuild and copied the drbd.sh.rhcs and drbd.metadata.rhcs files from the script folder of the source to the >> /usr/share/cluster folder as drbd.sh and drbd.metadata respectively. We are using RHEL 6.1 (64 bit). |
-- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster