Re: CLVM error 6 nodes

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Alasdair,


We did the modifications but still the same error:

Since the lock servers does not have access to the LUNs that will be in use by the GFS I used the option:

<clusternode name="lock-node3" votes="1" clvmd="0">

Only the rac-nodes have access to the shared LUNs...


[root@rac-node1 ~]# lvcreate -L4G -n lv01 vg01

  Found duplicate PV o2Xf8uUmskTL5fiVQAgY0nJ1ZJSMA9U3: using /dev/sds1 not /dev/emcpowerb1

  Found duplicate PV 3yduFLdX3FWPWWMb9lIbtBf3JIPjYnHF: using /dev/emcpowerc1 not /dev/sdr1

  Found duplicate PV Kyha3qI6nVE4odg77UXf7igS3FenGJNn: using /dev/sdd1 not /dev/emcpowerq1

  Found duplicate PV WS1LyhqQ8HaE2fIuSnXNd5sgTRtNzNAJ: using /dev/sdt1 not /dev/emcpowera1

  Found duplicate PV DuBJ7dZsS3PIO7n5U6hINxPkWorZDzvx: using /dev/emcpowerd1 not /dev/sdq1

  Found duplicate PV ZECZzAtbA0e9pFbl9oL0lZg4q7fkS5x4: using /dev/sdk1 not /dev/emcpowerj1

  Found duplicate PV bnVVmL6WhS2mesnOFUkT4fEfR0cFhybD: using /dev/emcpowerk1 not /dev/sdj1

  Found duplicate PV XyXrg2zdxxMS5jo03f9I4QYtGM3ILLGV: using /dev/sdl1 not /dev/emcpoweri1

  Found duplicate PV SLE5v2eTD7cJlpRUDGG35xfXkRbW86i1: using /dev/emcpowerl1 not /dev/sdi1

  Found duplicate PV acGyUd2wX7FnOF94Cbt0ombp10iUWMSf: using /dev/sdm1 not /dev/emcpowerh1

  Found duplicate PV ll8eNZ0JRh9katV0eui4BcxSc6HBggSI: using /dev/emcpowerm1 not /dev/sdh1

  Found duplicate PV ptGubq8R16LxywZ458P7ebmdG3Fq2aJo: using /dev/sdn1 not /dev/emcpowerg1

  Found duplicate PV PLQ3uON7pYe7nY16gRmAP94WBaEydRwf: using /dev/emcpowern1 not /dev/sdg1

  Found duplicate PV PsVYTeKNy6EcqWYbJwQ4KEbPp2Q8HjWv: using /dev/sdo1 not /dev/emcpowerf1

  Found duplicate PV hvekbzDAltJ3t23QveOMz1axfhj9Mp2j: using /dev/emcpowero1 not /dev/sdf1

  Found duplicate PV 5OhUbKbZLW5bTc3tpJeU4YlH0dTttJHF: using /dev/sdp1 not /dev/emcpowere1

  Found duplicate PV dFtPhq6pkwFdl41NTrAguAEFB3601CTb: using /dev/emcpowerp1 not /dev/sde1

  Error locking on node lock-node1: Internal lvm error, check syslog

  Error locking on node lock-node2: Internal lvm error, check syslog

  Error locking on node lock-node3: Internal lvm error, check syslog

  Failed to activate new LV.

File:
/var/log/messages:

Oct 13 21:20:46 lock-node1 lvm[5478]: Volume group for uuid not found: UzvBBmBj7m53APMbye1XXztWjdIavfgX8L5rGTOB3i3KGYPazw1AVaGCmWsXZpqR


Here is the cluster.conf

[root@rac-node1 ~]# cat /etc/cluster/cluster.conf
<?xml version="1.0 "?>
<cluster alias="cluster" config_version="6" name="cluster">
<fence_daemon post_fail_delay="0" post_join_delay="120"/>
<clusternodes>
<clusternode name="lock-node1" votes="1" clvmd="0">
<fence>
<method name="1">
<device name="lock-node1-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="
lock-node2" votes="1" clvmd="0">
<fence>
<method name="1">
<device name="lock-node2-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="lock-node3" votes="1" clvmd="0">
<fence>
<method name="1">
<device name="lock-node3-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="rac-node1" votes="1">
<fence>
<method name="1">
<device name="rac-node1-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="rac-node2" votes="1">
<fence>
<method name="1">
<device name="rac-node2-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="rac-node3" votes="1">
<fence>
<method name="1">
<device name="rac-node3-fence"/>
</method>
</fence>
</clusternode>
</clusternodes>
<gulm>
<lockserver name="lock-node1"/>
<lockserver name="lock-node2"/>
<lockserver name="lock-node3"/>
</gulm>
<fencedevices>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.4" login="xxxx" name="lock-node1-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.5" login="xxxx" name="lock-node2-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.6" login="xxxx" name="lock-node3-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.1" login="xxxx" name="rac-node1-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.2" login="xxxx" name="rac-node2-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.3" login="xxxx" name="rac-node3-fence" passwd="xxxx"/>
</fencedevices>
<rm>
<failoverdomains/>
<resources/>
</rm>

</cluster>

/etc/lvm/lvm.conf

# By default we accept every block device:

    # filter = [ "a/.*/" ]

    filter = [ "r|/dev/sda|", "a/.*/" ]

 

Regards,
Filipe Miranda

On 10/13/06, Filipe Miranda <filipe.miranda@xxxxxxxxx> wrote:
Alasdair,

Do I need to reboot the machine to test this configurations changes? Or is there a way do testing it without rebooting the machine?

I will try to swap the order of the filters,

Thanks for the hint


On 10/13/06, Alasdair G Kergon < agk@xxxxxxxxxx> wrote:
On Fri, Oct 13, 2006 at 06:50:24PM -0300, Filipe Miranda wrote:
> filter = [ "a/.*/", "r|/dev/sda|" ] (forgot to use the close pipe)

Still won't work - swap the order of the two items; first matches
everything so second isn't looked at.
(man lvm.conf)

Alasdair
--
agk@xxxxxxxxxx

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster



--
---
Filipe T Miranda
Red Hat Certified Engineer



--
---
Filipe T Miranda
Red Hat Certified Engineer
--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux