Hi,
I configured a cluster using gfs1 on rhel-4 kernel version 2.6.9-55.16.EL.
Using iscsi-target and initiator.
gfs1 mount is exported via nfs service.
I can manually stop all services in following sequence:
nfs, portmap, rgmanager, gfs, clvmd, fenced, cman, ccsd.
to stop my iscsi service first I give 'vgchange -aln' then I stop iscsi service, otherwise i get an error of module in use, as I have an clusterd lvm over iscsi device (/dev/sda1)
Everything works fine, but when i am trying to simulate a possible problem, f.e. iscsi service is stopped I get following error.
Test1:
When cluster is working I stop iscsi service with
/etc/init.d/iscsi stop
Searching for iscsi-based multipath maps
Found 0 maps
Stopping iscsid: [ OK ]
Removing iscsi driver: ERROR: Module iscsi_sfnet is in use
[FAILED]
To stop my iscsi service without a failure, I stop all cluster services as follows.
/etc/init.d/nfs stop
/etc/init.d/portmap stop
/etc/init.d/rgmanager stop
/etc/init.d/gfs stop
/etc/init.d/clvmd stop
/etc/init.d/fenced stop
/etc/init.d/cman stop
/etc/init.d/ccsd stop
Every service stops with a ok message. now again when i stop my iscsi service I get same error
/etc/init.d/iscsi stop
Removing iscsi driver: ERROR: Module iscsi_sfnet is in use [FAILED]
On my iscsi device (which is /dev/sd1), i have a LVM with gfs1 file-system,
as all the cluster services are stopped, I try to deactivate the lvm with:
vgchange -aln
/dev/dm-0: read failed after 0 of 4096 at 0: Input/output error
No volume groups found
At the moment if I start my iscsi service, my /dev/sda becomes /dev/sdb as well as iscsi service gives me following error:
[root@pr0031 new]# /sbin/service iscsi start
Checking iscsi config: [ OK ]
Loading iscsi driver: [ OK ]
mknod: `/dev/iscsictl': File exists
Starting iscsid: [ OK ]
Sep 5 16:42:37 pr0031 iscsi: iscsi config check succeeded
Sep 5 16:42:37 pr0031 iscsi: Loading iscsi driver: succeeded
Sep 5 16:42:42 pr0031 iscsid[20732]: version 4:0.1.11-7 variant (14-Apr-2008)
Sep 5 16:42:42 pr0031 iscsi: iscsid startup succeeded
Sep 5 16:42:42 pr0031 iscsid[20736]: Connected to Discovery Address 192.168.10.199
Sep 5 16:42:42 pr0031 kernel: iscsi-sfnet:host16: Session established
Sep 5 16:42:42 pr0031 kernel: scsi16 : SFNet iSCSI driver
Sep 5 16:42:42 pr0031 kernel: Vendor: IET Model: VIRTUAL-DISK Rev: 0
Sep 5 16:42:42 pr0031 kernel: Type: Direct-Access ANSI SCSI revision: 04
Sep 5 16:42:42 pr0031 kernel: SCSI device sdb: 1975932 512-byte hdwr sectors (1012 MB)
Sep 5 16:42:42 pr0031 kernel: SCSI device sdb: drive cache: write through
Sep 5 16:42:42 pr0031 kernel: SCSI device sdb: 1975932 512-byte hdwr sectors (1012 MB)
Sep 5 16:42:42 pr0031 kernel: SCSI device sdb: drive cache: write through
Sep 5 16:42:42 pr0031 kernel: sdb: sdb1
Sep 5 16:42:42 pr0031 kernel: Attached scsi disk sdb at scsi16, channel 0, id 0, lun 0
Sep 5 16:42:43 pr0031 scsi.agent[20764]: disk at /devices/platform/host16/target16:0:0/16:0:0:0
As my /dev/sda1 became /dev/sdb1, if i start cluster services, I have no gfs mount.
clurgmgrd[21062]: <notice> Starting stopped service flx
Sep 5 16:47:16 pr0031 kernel: scsi15 (0:0): rejecting I/O to dead device
Sep 5 16:47:16 pr0031 clurgmgrd: [21062]: <err> 'mount -t gfs /dev/mapper/VG01-LV01 /u01' failed, error=32
Sep 5 16:47:16 pr0031 clurgmgrd[21062]: <notice> start on clusterfs:gfsmount_u01 returned 2 (invalid argument(s))
Sep 5 16:47:16 pr0031 clurgmgrd[21062]: <warning> #68: Failed to start flx; return value: 1
Sep 5 16:47:16 pr0031 clurgmgrd[21062]: <notice> Stopping service flx
After the above situation I need to restart the nodes, which I don't want to, I created a script to handle all this, in which if i restart all the services first, first I get the same /dev/sdb ( which should be /dev/sda so that my cluster can have a gfs mount). When I restart all the services second time, I get no error (this time iscsi disk is attached with /dev/sda device name and I don't see any /dev/iscsctl exist error at the iscsi startup time) and cluster starts working.
my script : http://www.grex.org/~anuj/cluster.txt
So, how to get my cluster working if my /dev/sda becomes /dev/sdb?
Thanks and Regards
Anuj Singh
I configured a cluster using gfs1 on rhel-4 kernel version 2.6.9-55.16.EL.
Using iscsi-target and initiator.
gfs1 mount is exported via nfs service.
I can manually stop all services in following sequence:
nfs, portmap, rgmanager, gfs, clvmd, fenced, cman, ccsd.
to stop my iscsi service first I give 'vgchange -aln' then I stop iscsi service, otherwise i get an error of module in use, as I have an clusterd lvm over iscsi device (/dev/sda1)
Everything works fine, but when i am trying to simulate a possible problem, f.e. iscsi service is stopped I get following error.
Test1:
When cluster is working I stop iscsi service with
/etc/init.d/iscsi stop
Searching for iscsi-based multipath maps
Found 0 maps
Stopping iscsid: [ OK ]
Removing iscsi driver: ERROR: Module iscsi_sfnet is in use
[FAILED]
To stop my iscsi service without a failure, I stop all cluster services as follows.
/etc/init.d/nfs stop
/etc/init.d/portmap stop
/etc/init.d/rgmanager stop
/etc/init.d/gfs stop
/etc/init.d/clvmd stop
/etc/init.d/fenced stop
/etc/init.d/cman stop
/etc/init.d/ccsd stop
Every service stops with a ok message. now again when i stop my iscsi service I get same error
/etc/init.d/iscsi stop
Removing iscsi driver: ERROR: Module iscsi_sfnet is in use [FAILED]
On my iscsi device (which is /dev/sd1), i have a LVM with gfs1 file-system,
as all the cluster services are stopped, I try to deactivate the lvm with:
vgchange -aln
/dev/dm-0: read failed after 0 of 4096 at 0: Input/output error
No volume groups found
At the moment if I start my iscsi service, my /dev/sda becomes /dev/sdb as well as iscsi service gives me following error:
[root@pr0031 new]# /sbin/service iscsi start
Checking iscsi config: [ OK ]
Loading iscsi driver: [ OK ]
mknod: `/dev/iscsictl': File exists
Starting iscsid: [ OK ]
Sep 5 16:42:37 pr0031 iscsi: iscsi config check succeeded
Sep 5 16:42:37 pr0031 iscsi: Loading iscsi driver: succeeded
Sep 5 16:42:42 pr0031 iscsid[20732]: version 4:0.1.11-7 variant (14-Apr-2008)
Sep 5 16:42:42 pr0031 iscsi: iscsid startup succeeded
Sep 5 16:42:42 pr0031 iscsid[20736]: Connected to Discovery Address 192.168.10.199
Sep 5 16:42:42 pr0031 kernel: iscsi-sfnet:host16: Session established
Sep 5 16:42:42 pr0031 kernel: scsi16 : SFNet iSCSI driver
Sep 5 16:42:42 pr0031 kernel: Vendor: IET Model: VIRTUAL-DISK Rev: 0
Sep 5 16:42:42 pr0031 kernel: Type: Direct-Access ANSI SCSI revision: 04
Sep 5 16:42:42 pr0031 kernel: SCSI device sdb: 1975932 512-byte hdwr sectors (1012 MB)
Sep 5 16:42:42 pr0031 kernel: SCSI device sdb: drive cache: write through
Sep 5 16:42:42 pr0031 kernel: SCSI device sdb: 1975932 512-byte hdwr sectors (1012 MB)
Sep 5 16:42:42 pr0031 kernel: SCSI device sdb: drive cache: write through
Sep 5 16:42:42 pr0031 kernel: sdb: sdb1
Sep 5 16:42:42 pr0031 kernel: Attached scsi disk sdb at scsi16, channel 0, id 0, lun 0
Sep 5 16:42:43 pr0031 scsi.agent[20764]: disk at /devices/platform/host16/target16:0:0/16:0:0:0
As my /dev/sda1 became /dev/sdb1, if i start cluster services, I have no gfs mount.
clurgmgrd[21062]: <notice> Starting stopped service flx
Sep 5 16:47:16 pr0031 kernel: scsi15 (0:0): rejecting I/O to dead device
Sep 5 16:47:16 pr0031 clurgmgrd: [21062]: <err> 'mount -t gfs /dev/mapper/VG01-LV01 /u01' failed, error=32
Sep 5 16:47:16 pr0031 clurgmgrd[21062]: <notice> start on clusterfs:gfsmount_u01 returned 2 (invalid argument(s))
Sep 5 16:47:16 pr0031 clurgmgrd[21062]: <warning> #68: Failed to start flx; return value: 1
Sep 5 16:47:16 pr0031 clurgmgrd[21062]: <notice> Stopping service flx
After the above situation I need to restart the nodes, which I don't want to, I created a script to handle all this, in which if i restart all the services first, first I get the same /dev/sdb ( which should be /dev/sda so that my cluster can have a gfs mount). When I restart all the services second time, I get no error (this time iscsi disk is attached with /dev/sda device name and I don't see any /dev/iscsctl exist error at the iscsi startup time) and cluster starts working.
my script : http://www.grex.org/~anuj/cluster.txt
So, how to get my cluster working if my /dev/sda becomes /dev/sdb?
Thanks and Regards
Anuj Singh
-- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster