Hi,
I setuped and configured a clustered NFS.
I have created 2 quorum partitions /dev/sdd2 and /dev/sdd3 (100MB each).
I created another huge partition /dev/sdd4 (over 600GB) and formatted it in
ext3 file system.
I installed the cluster suite on the 1st node (RAC1) and 2nd node RAC2 and I
started the rawdevices on the two nodes RAC1 and RAC2 (it's OK).
This the hosts file /etc/host on the node1 (RAC1) and node2 RAC2
Do not remove the following line, or various programs
# that require network functionality will fail.
#127.0.0.1 rac1 localhost.localdomain localhost
127.0.0.1 localhost.localdomain localhost
#
# Private hostnames
#
192.168.253.3 rac1.project.net rac1
192.168.253.4 rac2.project.net rac2
192.168.253.10 raclu.project.net raclu
192.168.253.20 raclu_nfs.project.net raclu_nfs
#
# Hostnames used for Interconnect
#
1.1.1.1 rac1i.project.net rac1i
1.1.1.2 rac2i.project.net rac2i
#
192.168.253.5 infra.project.net infra
192.168.253.7 ractest.project.net ractest
#
I generated a /etc/cluster.conf on the 1st node RAC1 as following:
# This file is automatically generated. Do not manually edit!
[cluhbd]
logLevel = 4
[clupowerd]
logLevel = 4
[cluquorumd]
logLevel = 4
[cluster]
alias_ip = raclu
name = project
timestamp = 1121957827
[clusvcmgrd]
logLevel = 4
[database]
version = 2.0
[members]
start member0
start chan0
name = rac1
type = net
end chan0
id = 0
name = rac1
powerSwitchIPaddr = rac1
powerSwitchPortName = unused
quorumPartitionPrimary = /dev/raw/raw1
quorumPartitionShadow = /dev/raw/raw2
end member0
start member1
start chan0
name = rac2
type = net
end chan0
id = 1
name = rac2
powerSwitchIPaddr = rac2
powerSwitchPortName = unused
quorumPartitionPrimary = /dev/raw/raw1
quorumPartitionShadow = /dev/raw/raw2
end member1
[powercontrollers]
start powercontroller0
IPaddr = rac1
login = unused
passwd = unused
type = null
end powercontroller0
start powercontroller1
IPaddr = rac2
login = unused
passwd = unused
type = null
end powercontroller1
[services]
start service0
checkInterval = 0
start device0
start mount
start NFSexports
start directory0
start client0
name = *
options = rw
end client0
name = /u04
end directory0
end NFSexports
forceUnmount = yes
fstype = ext3
name = /u04
options = rw,nosuid,sync
end mount
name = /dev/sdd4
sharename = None
end device0
name = nfs_project
start network0
ipAddress = 192.168.253.20
end network0
preferredNode = rac2
relocateOnPreferredNodeBoot = yes
end service0
I generated a /etc/cluster.conf on the 2nd node RAC2 as following:
# This file is automatically generated. Do not manually edit!
[cluhbd]
logLevel = 4
[clupowerd]
logLevel = 4
[cluquorumd]
logLevel = 4
[cluster]
alias_ip = raclu
name = project
timestamp = 1121957827
[clusvcmgrd]
logLevel = 4
[database]
version = 2.0
[members]
start member0
start chan0
name = rac1
type = net
end chan0
id = 0
name = rac1
powerSwitchIPaddr = rac1
powerSwitchPortName = unused
quorumPartitionPrimary = /dev/raw/raw1
quorumPartitionShadow = /dev/raw/raw2
end member0
start member1
start chan0
name = rac2
type = net
end chan0
id = 1
name = rac2
powerSwitchIPaddr = rac2
powerSwitchPortName = unused
quorumPartitionPrimary = /dev/raw/raw1
quorumPartitionShadow = /dev/raw/raw2
end member1
[powercontrollers]
start powercontroller0
IPaddr = rac1
login = unused
passwd = unused
type = null
end powercontroller0
start powercontroller1
IPaddr = rac2
login = unused
passwd = unused
type = null
end powercontroller1
[services]
start service0
checkInterval = 0
start device0
start mount
start NFSexports
start directory0
start client0
name = *
options = rw
end client0
name = /u04
end directory0
end NFSexports
forceUnmount = yes
fstype = ext3
name = /u04
options = rw,nosuid,sync
end mount
name = /dev/sdd4
sharename = None
end device0
name = nfs_project
start network0
ipAddress = 192.168.253.20
end network0
preferredNode = rac2
relocateOnPreferredNodeBoot = yes
end service0
[session]
lock = rac1-cluadmin-16970-root 1121957668
I created a NFS share on /u04 (mount on /dev/sdd4) using the following
command cluadmin on RAC1
I launched on 2 nodes Rac1 and RAC2 the following command:
Service cluster start
I checked the result on the 2 nodes:
clustat
Cluster Status Monitor (project) 14:21:38
Cluster alias: raclu
========== M e m b e r S t a t u s =============
Member Status Node Id Power Switch
-------------- ---------- ---------- ------------
rac1 Up 0 Good
rac2 Up 1 Good
=========== H e a r t b e a t S t a t u s ======
Name Type Status
------------------------------ ---------- ------------
rac1 <--> rac2 network ONLINE
========== S e r v i c e S t a t u s =======
Last Monitor Restart
Service Status Owner Transition Interval Count
-------------- -------- -------------- ---------------- --------
nfs_project started rac2 13:01:36 Jul 21 0 0
And I launched on RAC1 and other servers the following command:
mount ?t nfs 192.168.253.20:/u04 /u04
And all are OK, I can list the /u04 content from any server (if I mount it).
But my only problem is:
When I want to try a test, I stop the network service on RAC2:
ifconfig eth0 down
Then when I try from another server as RAC1 to list the /u04 content, it
doesn?t work and doesn?t respond, and when I ping the IP alias
192.168.253.20 it doesn?t respond also.
Have you any idea to fix this problem?
Thanks for your replies and help
Haydar
From: "haydar Ali" <haydar2906@xxxxxxxxxxx>
Reply-To: linux clustering <linux-cluster@xxxxxxxxxx>
To: linux-cluster@xxxxxxxxxx
Subject: Need help for Clustered NFS
Date: Wed, 20 Jul 2005 10:10:35 -0400
Hi,
I want to setup and configure clustered NFS.
I have created 2 quorum partitions /dev/sdd2 and /dev/sdd3 (100MB each)
and formatted them
mkfs -t ext2 -b 4096 /dev/sdd2
mkfs -t ext2 -b 4096 /dev/sdd3
I created another huge partition /dev/sdd4 (over 600GB) and formatted it in
ext3 filesystem.
I installed the cluster suite on the 1st node (RAC1) and I started the
rawdevices on the two nodes RAC1 and RAC2 (it's OK).
This the hosts file /etc/host on the node1 (RAC1)
# Do not remove the following line, or various programs
# that require network functionality will fail.
127.0.0.1 localhost.localdomain localhost
#
# Private hostnames
#
192.168.253.3 rac1.domain.net rac1
192.168.253.4 rac2.domain.net rac2
192.168.253.10 rac1
#
# Hostnames used for Interconnect
#
1.1.1.1 rac1i.domain.net rac1i
1.1.1.2 rac2i.domain.net rac2i
#
-----------------------
I launched the command cluconfig and it generated /etc/cluster.conf, you
can list its content:
-------------------------------
This file is automatically generated. Do not manually edit!
[cluhbd]
logLevel = 4
[clupowerd]
logLevel = 4
[cluquorumd]
logLevel = 4
[cluster]
alias_ip = 192.168.253.10
name = project
timestamp = 1121804245
[clusvcmgrd]
logLevel = 4
[database]
version = 2.0
[members]
start member0
start chan0
name = rac1
type = net
end chan0
id = 0
name = rac1
powerSwitchIPaddr = rac1
powerSwitchPortName = unused
quorumPartitionPrimary = /dev/raw/raw1
quorumPartitionShadow = /dev/raw/raw2
end member0
start member1
start chan0
name = rac2
type = net
end chan0
id = 1
name = rac2
powerSwitchIPaddr = rac2
powerSwitchPortName = unused
quorumPartitionPrimary = /dev/raw/raw1
quorumPartitionShadow = /dev/raw/raw2
end member1
[powercontrollers]
start powercontroller0
IPaddr = rac1
login = unused
passwd = unused
type = null
end powercontroller0
start powercontroller1
IPaddr = rac2
login = unused
passwd = unused
type = null
end powercontroller1
[services]
start service0
checkInterval = 30
start device0
start mount
start NFSexports
start directory0
start client0
name = rac1
options = rw
end client0
name = /u04
end directory0
end NFSexports
forceUnmount = yes
fstype = ext3
name = /u04
options = rw,nosuid,sync
end mount
name = /dev/sdd4
sharename = None
end device0
name = nfs_project
preferredNode = rac2
relocateOnPreferredNodeBoot = yes
end service0
------------------------------------
I created a NFS share on /u04 using the following command cluadmin
[root@rac1 root]# cluadmin
Wed Jul 20 10:02:20 EDT 2005
You can obtain help by entering help and one of the following commands:
cluster service clear
help apropos exit
version quit
cluadmin> service show
1) state
2) config
3) services
service show what? 2
0) nfs_project
c) cancel
Choose service: 0
name: nfs_project
preferred node: rac2
relocate: yes
monitor interval: 30
device 0: /dev/sdd4
mount point, device 0: /u04
mount fstype, device 0: ext3
mount options, device 0: rw,nosuid,sync
force unmount, device 0: yes
samba share, device 0: None
NFS export 0: /u04
Client 0: rac1, rw
cluadmin> service show state
========================= S e r v i c e S t a t u s
========================
Last Monitor Restart
Service Status Owner Transition Interval Count
-------------- -------- -------------- ---------------- -------- -------
nfs_project started rac1 16:21:23 Jul 19 30 1
cluadmin>
And when I launched clustat, I expected this error message:
clustat
Cluster Status Monitor (Fileserver Test Cluster)
07:46:05
Cluster alias: rac1
===================== M e m b e r S t a t u s ================
Member Status Node Id Power Switch
-------------- ---------- ---------- ------------
rac1 Up 0 Good
rac2 Down 1 Unknown
=================== H e a r t b e a t S t a t u s ===============
Name Type Status
------------------------------ ---------- ------------
rac1 <--> rac2 network OFFLINE
=================== S e r v i c e S t a t u s ==================
Last Monitor
Restart
Service Status Owner Transition Interval Count
------------- -------- ------------- ---------------- ------------
nfs_project started rac1 16:07:42 Jul 19 30
0
And when I launched this command on RAC2:
mount ?t nfs rac1:/u04 /u04
It list the following error message :
Mount: rac1:/u04 failed, reason given by server: Permission denied
Can someone help me to fix this problem in this configuration?
Thanks
Cheers!
Haydar
--
Linux-cluster@xxxxxxxxxx
http://www.redhat.com/mailman/listinfo/linux-cluster
--
Linux-cluster@xxxxxxxxxx
http://www.redhat.com/mailman/listinfo/linux-cluster