Cluster IP alias doesn't work

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I setuped and configured a clustered NFS.
I have created 2 quorum partitions /dev/sdd2 and /dev/sdd3  (100MB each).

I created another huge partition /dev/sdd4 (over 600GB) and formatted it in ext3 file system.

I installed the cluster suite on the 1st node (RAC1) and 2nd node RAC2 and I started the rawdevices on the two nodes RAC1 and RAC2 (it's OK).

This the hosts file /etc/host on the node1 (RAC1) and node2 RAC2

Do not remove the following line, or various programs
# that require network functionality will fail.
#127.0.0.1		rac1 localhost.localdomain localhost
127.0.0.1              localhost.localdomain localhost
#
# Private hostnames
#
192.168.253.3           rac1.project.net     rac1
192.168.253.4           rac2.project.net     rac2
192.168.253.10          raclu.project.net	raclu
192.168.253.20		raclu_nfs.project.net	raclu_nfs
#
# Hostnames used for Interconnect
#
1.1.1.1                 rac1i.project.net    rac1i
1.1.1.2                 rac2i.project.net    rac2i
#
192.168.253.5           infra.project.net       infra
192.168.253.7		ractest.project.net     ractest
#
I generated a /etc/cluster.conf on the 1st node RAC1 as following:

# This file is automatically generated.  Do not manually edit!

[cluhbd]
 logLevel = 4

[clupowerd]
 logLevel = 4

[cluquorumd]
 logLevel = 4

[cluster]
 alias_ip = raclu
 name = project
 timestamp = 1121957827

[clusvcmgrd]
 logLevel = 4

[database]
 version = 2.0

[members]
start member0
start chan0
 name = rac1
 type = net
end chan0
 id = 0
 name = rac1
 powerSwitchIPaddr = rac1
 powerSwitchPortName = unused
 quorumPartitionPrimary = /dev/raw/raw1
 quorumPartitionShadow = /dev/raw/raw2
end member0
start member1
start chan0
 name = rac2
 type = net
end chan0
 id = 1
 name = rac2
 powerSwitchIPaddr = rac2
 powerSwitchPortName = unused
 quorumPartitionPrimary = /dev/raw/raw1
 quorumPartitionShadow = /dev/raw/raw2
end member1

[powercontrollers]
start powercontroller0
 IPaddr = rac1
 login = unused
 passwd = unused
 type = null
end powercontroller0
start powercontroller1
 IPaddr = rac2
 login = unused
 passwd = unused
 type = null
end powercontroller1

[services]
start service0
 checkInterval = 0
start device0
start mount
start NFSexports
start directory0
start client0
 name = *
 options = rw
end client0
 name = /u04
end directory0
end NFSexports
 forceUnmount = yes
 fstype = ext3
 name = /u04
 options = rw,nosuid,sync
end mount
 name = /dev/sdd4
 sharename = None
end device0
 name = nfs_project
start network0
 ipAddress = 192.168.253.20
end network0
 preferredNode = rac2
 relocateOnPreferredNodeBoot = yes
end service0


I generated a /etc/cluster.conf on the 2nd node RAC2 as following:


# This file is automatically generated.  Do not manually edit!

[cluhbd]
 logLevel = 4

[clupowerd]
 logLevel = 4

[cluquorumd]
 logLevel = 4

[cluster]
 alias_ip = raclu
 name = project
 timestamp = 1121957827

[clusvcmgrd]
 logLevel = 4

[database]
 version = 2.0

[members]
start member0
start chan0
 name = rac1
 type = net
end chan0
 id = 0
 name = rac1
 powerSwitchIPaddr = rac1
 powerSwitchPortName = unused
 quorumPartitionPrimary = /dev/raw/raw1
 quorumPartitionShadow = /dev/raw/raw2
end member0
start member1
start chan0
 name = rac2
 type = net
end chan0
 id = 1
 name = rac2
 powerSwitchIPaddr = rac2
 powerSwitchPortName = unused
 quorumPartitionPrimary = /dev/raw/raw1
 quorumPartitionShadow = /dev/raw/raw2
end member1

[powercontrollers]
start powercontroller0
 IPaddr = rac1
 login = unused
 passwd = unused
 type = null
end powercontroller0
start powercontroller1
 IPaddr = rac2
 login = unused
 passwd = unused
 type = null
end powercontroller1

[services]
start service0
 checkInterval = 0
start device0
start mount
start NFSexports
start directory0
start client0
 name = *
 options = rw
end client0
 name = /u04
end directory0
end NFSexports
 forceUnmount = yes
 fstype = ext3
 name = /u04
 options = rw,nosuid,sync
end mount
 name = /dev/sdd4
 sharename = None
end device0
 name = nfs_project
start network0
 ipAddress = 192.168.253.20
end network0
 preferredNode = rac2
 relocateOnPreferredNodeBoot = yes
end service0

[session]
 lock = rac1-cluadmin-16970-root 1121957668

I created a NFS share on /u04 (mount on /dev/sdd4) using the following command cluadmin on RAC1
I launched on 2 nodes Rac1 and RAC2 the following command:
Service cluster start
I checked the result on the 2 nodes:
clustat



Cluster Status Monitor (project)   14:21:38

Cluster alias: raclu

==========  M e m b e r   S t a t u s  =============

 Member         Status     Node Id    Power Switch
 -------------- ---------- ---------- ------------
 rac1        Up         0          Good
 rac2        Up         1          Good

===========  H e a r t b e a t   S t a t u s  ======

 Name                           Type       Status
 ------------------------------ ---------- ------------
 rac1      <--> rac2      network    ONLINE

========== S e r v i c e   S t a t u s  =======

                                        Last             Monitor  Restart
 Service        Status   Owner          Transition       Interval Count
-------------- -------- -------------- ---------------- -------- nfs_project started rac2 13:01:36 Jul 21 0 0


And I launched on RAC1 and other servers the following command:
mount ?t nfs 192.168.253.20:/u04 /u04

And all are OK, I can list the /u04 content from any server (if I mount it).

But my only problem is:

When I want to try a test, I stop the network service on RAC2:
ifconfig eth0 down
Then when I try from another server as RAC1 to list the /u04 content, it doesn?t work and doesn?t respond, and when I ping the IP alias 192.168.253.20 it doesn?t respond also.

Have you any idea to fix this problem?

Thanks for your replies and help


Haydar

From: "haydar Ali" <haydar2906@xxxxxxxxxxx>
Reply-To: linux clustering <linux-cluster@xxxxxxxxxx>
To: linux-cluster@xxxxxxxxxx
Subject:  Need help for Clustered NFS
Date: Wed, 20 Jul 2005 10:10:35 -0400

Hi,

I want to setup and configure clustered NFS.
I have created 2 quorum partitions /dev/sdd2 and /dev/sdd3 (100MB each) and formatted them

mkfs -t ext2 -b 4096 /dev/sdd2
mkfs -t ext2 -b 4096 /dev/sdd3

I created another huge partition /dev/sdd4 (over 600GB) and formatted it in ext3 filesystem.

I installed the cluster suite on the 1st node (RAC1) and I started the rawdevices on the two nodes RAC1 and RAC2 (it's OK).

This the hosts file /etc/host on the node1 (RAC1)

# Do not remove the following line, or various programs
# that require network functionality will fail.
127.0.0.1              localhost.localdomain localhost
#
# Private hostnames
#
192.168.253.3           rac1.domain.net     rac1
192.168.253.4           rac2.domain.net     rac2
192.168.253.10          rac1
#
# Hostnames used for Interconnect
#
1.1.1.1                 rac1i.domain.net    rac1i
1.1.1.2                 rac2i.domain.net    rac2i
#
-----------------------


I launched the command cluconfig and it generated /etc/cluster.conf, you can list its content:

-------------------------------
This file is automatically generated.  Do not manually edit!

[cluhbd]
 logLevel = 4

[clupowerd]
 logLevel = 4

[cluquorumd]
 logLevel = 4

[cluster]
 alias_ip = 192.168.253.10
 name = project
 timestamp = 1121804245

[clusvcmgrd]
 logLevel = 4

[database]
 version = 2.0

[members]
start member0
start chan0
 name = rac1
 type = net
end chan0
 id = 0
 name = rac1
 powerSwitchIPaddr = rac1
 powerSwitchPortName = unused
 quorumPartitionPrimary = /dev/raw/raw1
 quorumPartitionShadow = /dev/raw/raw2
end member0
start member1
start chan0
 name = rac2
 type = net
end chan0
 id = 1
 name = rac2
 powerSwitchIPaddr = rac2
 powerSwitchPortName = unused
 quorumPartitionPrimary = /dev/raw/raw1
 quorumPartitionShadow = /dev/raw/raw2
end member1

[powercontrollers]
start powercontroller0
 IPaddr = rac1
 login = unused
 passwd = unused
 type = null
end powercontroller0
start powercontroller1
 IPaddr = rac2
 login = unused
 passwd = unused
 type = null
end powercontroller1

[services]
start service0
 checkInterval = 30
start device0
start mount
start NFSexports
start directory0
start client0
 name = rac1
 options = rw
end client0
 name = /u04
end directory0
end NFSexports
 forceUnmount = yes
 fstype = ext3
 name = /u04
 options = rw,nosuid,sync
end mount
 name = /dev/sdd4
 sharename = None
end device0
 name = nfs_project
 preferredNode = rac2
 relocateOnPreferredNodeBoot = yes
end service0
------------------------------------

I created a NFS share on /u04 using the following command cluadmin

[root@rac1 root]# cluadmin
Wed Jul 20 10:02:20 EDT 2005

You can obtain help by entering help and one of the following commands:

cluster     service        clear
help        apropos        exit
version         quit
cluadmin> service show
 1) state
 2) config
 3) services
service show what? 2
 0) nfs_project
 c) cancel

Choose service: 0
name: nfs_project
preferred node: rac2
relocate: yes
monitor interval: 30
device 0: /dev/sdd4
 mount point, device 0: /u04
 mount fstype, device 0: ext3
 mount options, device 0: rw,nosuid,sync
 force unmount, device 0: yes
 samba share, device 0: None
NFS export 0: /u04
 Client 0: rac1, rw
cluadmin> service show state
========================= S e r v i c e S t a t u s ========================

                                        Last             Monitor  Restart
 Service        Status   Owner          Transition       Interval Count
 -------------- -------- -------------- ---------------- -------- -------
 nfs_project       started  rac1        16:21:23 Jul 19  30       1
cluadmin>


And when I launched clustat, I expected this error message:

clustat
Cluster Status Monitor (Fileserver Test Cluster)
07:46:05
Cluster alias: rac1

===================== M e m b e r   S t a t u s ================
 Member         Status     Node Id    Power Switch
 -------------- ---------- ---------- ------------
 rac1           Up         0          Good
 rac2           Down    1          Unknown

=================== H e a r t b e a t   S t a t u s ===============
 Name                           Type       Status
 ------------------------------ ---------- ------------
 rac1         <--> rac2         network    OFFLINE

=================== S e r v i c e   S t a t u s ==================
Last Monitor Restart
Service         Status   Owner          Transition    Interval  Count

 ------------- -------- ------------- ---------------- ------------
nfs_project started rac1 16:07:42 Jul 19 30 0



And when I launched this command on RAC2:
mount ?t nfs rac1:/u04 /u04
It list the following error message :
Mount: rac1:/u04 failed, reason given by server: Permission denied

Can someone help me to fix this problem in this configuration?

Thanks

Cheers!

Haydar


--

Linux-cluster@xxxxxxxxxx
http://www.redhat.com/mailman/listinfo/linux-cluster


--

Linux-cluster@xxxxxxxxxx
http://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux