Re: RHCS + DRBD - DRBD file systems not mounting

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Are you using san with emc powerpath?
drbd service start on boot?

start drbd by hand and give the command ls -l dev/drbd/by-res/r0/1 and ls -l
dev/drbd/by-res/r0/0


2013/8/27 Goutam Baul <goutam.baul@xxxxxxxx>

Dear List,

 

We have configured DRBD 8.4.3 (community version) in the sense that we are able to manually  start DRBD at both the nodes, make nodes primary or secondary and data is getting replicated between the primary and the secondary node. We are doing this for our postfix mail server. The drbd configuration file is

 

resource r0 {

startup {

    wfc-timeout 1;

    degr-wfc-timeout 2;

  }

  volume 0 {

    device    /dev/drbd1;

    disk      /dev/emcpowera2;

    meta-disk /dev/emcpowera5;

  }

  volume 1 {

    device    /dev/drbd2;

    disk      /dev/emcpowera3;

    meta-disk /dev/emcpowera6;

  }

  on drmail1.cesc.co.in {

    address   10.50.4.14:7789;

  }

  on drbdtest2.cesc.co.in {

    address   10.50.81.253:7789;

  }

}

 

We want to run postfix mailing system on a RHCS cluster suite for high availability. The mail-store and the mail-queue of postfix will be on the DRBD setup so that they are replicated from our data center to our disaster recovery site. The cluster.conf is as follows:

 

<?xml version="1.0"?>

<cluster config_version="27" name="dr-mail">

        <fence_daemon clean_start="1" post_join_delay="30"/>

        <clusternodes>

                <clusternode name="drmail1.cesc.co.in" nodeid="1">

                        <fence>

                                <method name="Method1_drac">

                                        <device name="fence_drac1"/>

                                </method>

 

                        </fence>

                </clusternode>

                <clusternode name="drmail2.cesc.co.in" nodeid="2">

                        <fence>

                                <method name="Method1_drac">

                                        <device name="fence_drac2"/>

                                </method>

 

                        </fence>

                </clusternode>

        </clusternodes>

        <cman expected_votes="1" two_node="1"/>

        <fencedevices>

                <fencedevice agent="fence_drac5" cmd_prompt="admin1-&gt;" ipaddr="10.50.4.22" login="root" module_name="fence_drac1" name="fence_drac1" passwd="calvin" secure="on"/>

                <fencedevice agent="fence_drac5" cmd_prompt="admin1-&gt;" ipaddr="10.50.4.23" login="root" module_name="fence_drac2" name="fence_drac2" passwd="calvin" secure="on"/>

 

        </fencedevices>

        <rm>

                <failoverdomains>

                        <failoverdomain name="dr_mail" nofailback="0" ordered="1" restricted="1">

                                <failoverdomainnode name="drmail1.cesc.co.in" priority="1"/>

                                <failoverdomainnode name="drmail2.cesc.co.in" priority="2"/>

                        </failoverdomain>

                </failoverdomains>

                <resources>

                        <ip address="10.50.4.20" monitor_link="on" sleeptime="10"/>

                        <drbd name="drbd-mail" resource="r0">

                              <fs device="/dev/drbd/by-res/r0/0" force_unmount="1" fsid="61850" mountpoint="/home" fstype="ext3" name="drbd-mailstore" options="noatime"/>

                              <fs device="/dev/drbd/by-res/r0/1" force_unmount="1" fsid="61855" mountpoint="/var/spool/postfix" fstype="ext3" name="drbd-spool" options="noatime"/>

                        </drbd>

                        <script file="/etc/init.d/postfix" name="postfix" />

                </resources>

                <service domain="dr_mail" name="mail" recovery="relocate">

                        <ip ref="10.50.4.20"/>

                        <drbd ref="drbd-mail" />

                        <script ref="postfix" />

                </service>

        </rm>

</cluster>

 

If we run this RHCS cluster, we get the error in the log as

 

Aug 27 17:38:19 drmail1 rgmanager[4695]: drbd not found in /proc/modules. Do you need to modprobe?

 

If we run the RHCS cluster after issuing the modprobe drbd command we get the following message

 

1: Failure: (127) Device minor not allocated

additional info from kernel:

unknown minor

Command 'drbdsetup role 1' terminated with exit code 10

<debug>  DRBD resource r0 already configured

DRBD resource r0 already configured

Start of mail complete

 

But the DRBD service is not starting up and therefore the corresponding devices are not being created. Thus the log file returns the error

 

Aug 27 17:38:33 drmail1 rgmanager[5147]: stop: Could not match /dev/drbd/by-res/r0/1 with a real device

Aug 27 17:38:33 drmail1 rgmanager[5166]: stop: Could not match /dev/drbd/by-res/r0/0 with a real device

 

We have built the DRBD rpm files from the source using rpmbuild and copied the drbd.sh.rhcs and drbd.metadata.rhcs files from the script folder of the source to the /usr/share/cluster folder as drbd.sh and drbd.metadata respectively. We are using RHEL 6.1 (64 bit).

 

Any help please.

 

With regards,

 

Goutam

 

 

 


--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster



--
esta es mi vida e me la vivo hasta que dios quiera
-- 
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux