I am working on configuring cluster environment for NFS share using pacemaker. Below are the resources I have configured.
PCS Status
I followedthe redhat link to configure.
Once configured, I could mount the directory from nfs client with no issues. However, wehn I try entering the node to standby, the resources not starting up in secondary node.
After entering active node to standby,
I am seeing this error,
This resolves when I create the lvm manually on secondary but I expect the resources to do the job. Am I missing something in this configuration?
--
Quote:
Group: nfsgroup Resource: my_lvm (class=ocf provider=heartbeat type=LVM) Attributes: volgrpname=my_vg exclusive=true Operations: start interval=0s timeout=30 (my_lvm-start-interval-0s) stop interval=0s timeout=30 (my_lvm-stop-interval-0s) monitor interval=10 timeout=30 (my_lvm-monitor-interval-10) Resource: nfsshare (class=ocf provider=heartbeat type=Filesystem) Attributes: device=/dev/my_vg/my_lv directory=/nfsshare fstype=ext4 Operations: start interval=0s timeout=60 (nfsshare-start-interval-0s) stop interval=0s timeout=60 (nfsshare-stop-interval-0s) monitor interval=20 timeout=40 (nfsshare-monitor-interval-20) Resource: nfs-daemon (class=ocf provider=heartbeat type=nfsserver) Attributes: nfs_shared_infodir=/nfsshare/nfsinfo nfs_no_notify=true Operations: start interval=0s timeout=40 (nfs-daemon-start-interval-0s) stop interval=0s timeout=20s (nfs-daemon-stop-interval-0s) monitor interval=10 timeout=20s (nfs-daemon-monitor-interval-10) Resource: nfs-root (class=ocf provider=heartbeat type=exportfs) Attributes: clientspec=10.199.1.0/255.255.255.0 options=rw,sync,no_root_squash directory=/nfsshare/exports fsid=0 Operations: start interval=0s timeout=40 (nfs-root-start-interval-0s) stop interval=0s timeout=120 (nfs-root-stop-interval-0s) monitor interval=10 timeout=20 (nfs-root-monitor-interval-10) Resource: nfs-export1 (class=ocf provider=heartbeat type=exportfs) Attributes: clientspec=10.199.1.0/255.255.255.0 options=rw,sync,no_root_squash directory=/nfsshare/exports/export1 fsid=1 Operations: start interval=0s timeout=40 (nfs-export1-start-interval-0s) stop interval=0s timeout=120 (nfs-export1-stop-interval-0s) monitor interval=10 timeout=20 (nfs-export1-monitor-interval-10) Resource: nfs-export2 (class=ocf provider=heartbeat type=exportfs) Attributes: clientspec=10.199.1.0/255.255.255.0 options=rw,sync,no_root_squash directory=/nfsshare/exports/export2 fsid=2 Operations: start interval=0s timeout=40 (nfs-export2-start-interval-0s) stop interval=0s timeout=120 (nfs-export2-stop-interval-0s) monitor interval=10 timeout=20 (nfs-export2-monitor-interval-10) Resource: nfs_ip (class=ocf provider=heartbeat type=IPaddr2) Attributes: ip=10.199.1.86 cidr_netmask=24 Operations: start interval=0s timeout=20s (nfs_ip-start-interval-0s) stop interval=0s timeout=20s (nfs_ip-stop-interval-0s) monitor interval=10s timeout=20s (nfs_ip-monitor-interval-10s) Resource: nfs-notify (class=ocf provider=heartbeat type=nfsnotify) Attributes: source_host=10.199.1.86 Operations: start interval=0s timeout=90 (nfs-notify-start-interval-0s) stop interval=0s timeout=90 (nfs-notify-stop-interval-0s) monitor interval=30 timeout=90 (nfs-notify-monitor-interval-30) |
PCS Status
Quote:
Cluster name: my_cluster Stack: corosync Current DC: node3.cluster.com (version 1.1.15-11.el7_3.5-e174ec8) - partition with quorum Last updated: Wed Jul 5 13:12:48 2017 Last change: Wed Jul 5 13:11:52 2017 by root via crm_attribute on node3.cluster.com 2 nodes and 10 resources configured Online: [ node3.cluster.com node4.cluster.com ] Full list of resources: fence-3 (stonith:fence_vmware_soap): Started node4.cluster.com fence-4 (stonith:fence_vmware_soap): Started node3.cluster.com Resource Group: nfsgroup my_lvm (ocf::heartbeat:LVM): Started node3.cluster.com nfsshare (ocf::heartbeat:Filesystem): Started node3.cluster.com nfs-daemon (ocf::heartbeat:nfsserver): Started node3.cluster.com nfs-root (ocf::heartbeat:exportfs): Started node3.cluster.com nfs-export1 (ocf::heartbeat:exportfs): Started node3.cluster.com nfs-export2 (ocf::heartbeat:exportfs): Started node3.cluster.com nfs_ip (ocf::heartbeat:IPaddr2): Started node3.cluster.com nfs-notify (ocf::heartbeat:nfsnotify): Started node3.cluster.com |
Once configured, I could mount the directory from nfs client with no issues. However, wehn I try entering the node to standby, the resources not starting up in secondary node.
After entering active node to standby,
Quote:
[root@node3 ~]# pcs status Cluster name: my_cluster Stack: corosync Current DC: node3.cluster.com (version 1.1.15-11.el7_3.5-e174ec8) - partition with quorum Last updated: Wed Jul 5 13:16:05 2017 Last change: Wed Jul 5 13:15:38 2017 by root via crm_attribute on node3.cluster.com 2 nodes and 10 resources configured Node node3.cluster.com: standby Online: [ node4.cluster.com ] Full list of resources: fence-3 (stonith:fence_vmware_soap): Started node4.cluster.com fence-4 (stonith:fence_vmware_soap): Started node4.cluster.com Resource Group: nfsgroup my_lvm (ocf::heartbeat:LVM): Stopped nfsshare (ocf::heartbeat:Filesystem): Stopped nfs-daemon (ocf::heartbeat:nfsserver): Stopped nfs-root (ocf::heartbeat:exportfs): Stopped nfs-export1 (ocf::heartbeat:exportfs): Stopped nfs-export2 (ocf::heartbeat:exportfs): Stopped nfs_ip (ocf::heartbeat:IPaddr2): Stopped nfs-notify (ocf::heartbeat:nfsnotify): Stopped Failed Actions: * fence-3_monitor_60000 on node4.cluster.com 'unknown error' (1): call=50, status=Timed Out, exitreason='none', last-rc-change='Wed Jul 5 13:11:54 2017', queued=0ms, exec=20012ms * fence-4_monitor_60000 on node4.cluster.com 'unknown error' (1): call=47, status=Timed Out, exitreason='none', last-rc-change='Wed Jul 5 13:05:32 2017', queued=0ms, exec=20028ms * my_lvm_start_0 on node4.cluster.com 'unknown error' (1): call=49, status=complete, exitreason='Volume group [my_vg] does not exist or contains error! Volume group "my_vg" not found', last-rc-change='Wed Jul 5 13:05:39 2017', queued=0ms, exec=1447ms Daemon Status: corosync: active/enabled pacemaker: active/enabled pcsd: active/enabled |
I am seeing this error,
Quote:
ERROR: Volume group [my_vg] does not exist or contains error! Volume group "my_vg" not found#012 Cannot process volume group my_vg |
--
Regards,
Pradeep Anandh-- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster