Hi,
Now removed the disk which has hosted one of the brick of the volume.
Storage.health-check-interval is set to 30 seconds.
I could see the disk is unavailable using zpool command of zfs on linux but the gluster volume status still displays the brick process running which should have been shutdown by this time.
Is this a bug in 3.6 since it is mentioned as feature "https://github.com/gluster/glusterfs/blob/release-3.6/doc/features/brick-failure-detection.md" or am I doing any mistakes here?
[root@fractal-c92e gluster-3.6]# gluster volume status
Status of volume: repvol
Gluster process Port Online Pid
------------------------------------------------------------------------------
Brick 192.168.1.246:/zp1/brick1 49154 Y 17671
Brick 192.168.1.246:/zp2/brick2 49155 Y 17682
NFS Server on localhost 2049 Y 17696
Self-heal Daemon on localhost N/A Y 17701
Task Status of Volume repvol
------------------------------------------------------------------------------
There are no active volume tasks
[root@fractal-c92e gluster-3.6]# gluster volume info
Volume Name: repvol
Type: Replicate
Volume ID: d4f992b1-1393-43b8-9fda-2e2b6e3b5039
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 192.168.1.246:/zp1/brick1
Brick2: 192.168.1.246:/zp2/brick2
Options Reconfigured:
storage.health-check-interval: 30
[root@fractal-c92e gluster-3.6]# zpool status zp2
pool: zp2
state: UNAVAIL
status: One or more devices are faulted in response to IO failures.
action: Make sure the affected devices are connected, then run 'zpool clear'.
scan: none requested
config:
NAME STATE READ WRITE CKSUM
zp2 UNAVAIL 0 0 0 insufficient replicas
sdb UNAVAIL 0 0 0
errors: 2 data errors, use '-v' for a list
Thanks,
Kiran.
_______________________________________________ Gluster-devel mailing list Gluster-devel@xxxxxxxxxxx http://supercolony.gluster.org/mailman/listinfo/gluster-devel