Alasdair,
We did the modifications but still the same error:
Since the lock servers does not have access to the LUNs that will be in use by the GFS I used the option:
<clusternode name="lock-node3" votes="1" clvmd="0">
Only the rac-nodes have access to the shared LUNs...
[root@rac-node1 ~]# lvcreate -L4G -n lv01 vg01
Found duplicate PV o2Xf8uUmskTL5fiVQAgY0nJ1ZJSMA9U3: using /dev/sds1 not /dev/emcpowerb1
Found duplicate PV 3yduFLdX3FWPWWMb9lIbtBf3JIPjYnHF: using /dev/emcpowerc1 not /dev/sdr1
Found duplicate PV Kyha3qI6nVE4odg77UXf7igS3FenGJNn: using /dev/sdd1 not /dev/emcpowerq1
Found duplicate PV WS1LyhqQ8HaE2fIuSnXNd5sgTRtNzNAJ: using /dev/sdt1 not /dev/emcpowera1
Found duplicate PV DuBJ7dZsS3PIO7n5U6hINxPkWorZDzvx: using /dev/emcpowerd1 not /dev/sdq1
Found duplicate PV ZECZzAtbA0e9pFbl9oL0lZg4q7fkS5x4: using /dev/sdk1 not /dev/emcpowerj1
Found duplicate PV bnVVmL6WhS2mesnOFUkT4fEfR0cFhybD: using /dev/emcpowerk1 not /dev/sdj1
Found duplicate PV XyXrg2zdxxMS5jo03f9I4QYtGM3ILLGV: using /dev/sdl1 not /dev/emcpoweri1
Found duplicate PV SLE5v2eTD7cJlpRUDGG35xfXkRbW86i1: using /dev/emcpowerl1 not /dev/sdi1
Found duplicate PV acGyUd2wX7FnOF94Cbt0ombp10iUWMSf: using /dev/sdm1 not /dev/emcpowerh1
Found duplicate PV ll8eNZ0JRh9katV0eui4BcxSc6HBggSI: using /dev/emcpowerm1 not /dev/sdh1
Found duplicate PV ptGubq8R16LxywZ458P7ebmdG3Fq2aJo: using /dev/sdn1 not /dev/emcpowerg1
Found duplicate PV PLQ3uON7pYe7nY16gRmAP94WBaEydRwf: using /dev/emcpowern1 not /dev/sdg1
Found duplicate PV PsVYTeKNy6EcqWYbJwQ4KEbPp2Q8HjWv: using /dev/sdo1 not /dev/emcpowerf1
Found duplicate PV hvekbzDAltJ3t23QveOMz1axfhj9Mp2j: using /dev/emcpowero1 not /dev/sdf1
Found duplicate PV 5OhUbKbZLW5bTc3tpJeU4YlH0dTttJHF: using /dev/sdp1 not /dev/emcpowere1
Found duplicate PV dFtPhq6pkwFdl41NTrAguAEFB3601CTb: using /dev/emcpowerp1 not /dev/sde1
Error locking on node lock-node1: Internal lvm error, check syslog
Error locking on node lock-node2: Internal lvm error, check syslog
Error locking on node lock-node3: Internal lvm error, check syslog
Failed to activate new LV.File:
/var/log/messages:
Oct 13 21:20:46 lock-node1 lvm[5478]: Volume group for uuid not found: UzvBBmBj7m53APMbye1XXztWjdIavfgX8L5rGTOB3i3KGYPazw1AVaGCmWsXZpqR
Here is the cluster.conf
[root@rac-node1 ~]# cat /etc/cluster/cluster.conf
<?xml version="1.0 "?>
<cluster alias="cluster" config_version="6" name="cluster">
<fence_daemon post_fail_delay="0" post_join_delay="120"/>
<clusternodes>
<clusternode name="lock-node1" votes="1" clvmd="0">
<fence>
<method name="1">
<device name="lock-node1-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="lock-node2" votes="1" clvmd="0">
<fence>
<method name="1">
<device name="lock-node2-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="lock-node3" votes="1" clvmd="0">
<fence>
<method name="1">
<device name="lock-node3-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="rac-node1" votes="1">
<fence>
<method name="1">
<device name="rac-node1-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="rac-node2" votes="1">
<fence>
<method name="1">
<device name="rac-node2-fence"/>
</method>
</fence>
</clusternode>
<clusternode name="rac-node3" votes="1">
<fence>
<method name="1">
<device name="rac-node3-fence"/>
</method>
</fence>
</clusternode>
</clusternodes>
<gulm>
<lockserver name="lock-node1"/>
<lockserver name="lock-node2"/>
<lockserver name="lock-node3"/>
</gulm>
<fencedevices>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.4" login="xxxx" name="lock-node1-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.5" login="xxxx" name="lock-node2-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.6" login="xxxx" name="lock-node3-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.1" login="xxxx" name="rac-node1-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.2" login="xxxx" name="rac-node2-fence" passwd="xxxx"/>
<fencedevice agent="fence_ipmilan" auth="none"
ipaddr="20.20.20.3" login="xxxx" name="rac-node3-fence" passwd="xxxx"/>
</fencedevices>
<rm>
<failoverdomains/>
<resources/>
</rm>
</cluster>
/etc/lvm/lvm.conf
# By default we accept every block device:
# filter = [ "a/.*/" ]
filter = [ "r|/dev/sda|", "a/.*/" ]
Regards,
Filipe Miranda
Alasdair,
Do I need to reboot the machine to test this configurations changes? Or is there a way do testing it without rebooting the machine?
I will try to swap the order of the filters,
Thanks for the hintOn 10/13/06, Alasdair G Kergon < agk@xxxxxxxxxx> wrote:On Fri, Oct 13, 2006 at 06:50:24PM -0300, Filipe Miranda wrote:
> filter = [ "a/.*/", "r|/dev/sda|" ] (forgot to use the close pipe)
Still won't work - swap the order of the two items; first matches
everything so second isn't looked at.
(man lvm.conf)
Alasdair
--
agk@xxxxxxxxxx
--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster--
---
Filipe T Miranda
Red Hat Certified Engineer
--
---
Filipe T Miranda
Red Hat Certified Engineer
-- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster