Could you share the logs?
I'd like to look at the glustershd logs and etc-glusterfs-glusterd.vol.log files.
-Krutika
From: "Lindsay Mathieson" <lindsay.mathieson@xxxxxxxxx>
To: "gluster-users" <Gluster-users@xxxxxxxxxxx>
Sent: Saturday, January 23, 2016 9:20:52 AM
Subject: More Peculiar heal behaviour after removing brickMaybe I'm doing some wrong here but I'm not sure what, or maybe this is normal behaviour?
All of the following is performed from my vna node which has the highest numbered uuid. Indenting applied by me for readability.
Spoiler, because it happens at the end: removing the vng brick followed by a full heal gives this error:
"Commit failed on vng.proxmox.softlog. Please check log file for details."
Sets to recreate:
1. Create a test volume:vna$ gluster volume create test3 rep 3 transport tcp vnb.proxmox.softlog:/vmdata/test3 vng.proxmox.softlog:/vmdata/test3 vna.proxmox.softlog:/vmdata/test32. Immediately remove the vng brick:
vna$ gluster volume set test3 group softlog
vna$ gluster volume info test3Volume Name: test3vna$ gluster volume start test3
Type: Replicate
Volume ID: 0be89d63-775c-4eb5-9d98-0a4a87f30fbf
Status: Created
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: vnb.proxmox.softlog:/vmdata/test3
Brick2: vng.proxmox.softlog:/vmdata/test3
Brick3: vna.proxmox.softlog:/vmdata/test3
Options Reconfigured:
cluster.data-self-heal-algorithm: full
network.remote-dio: enable
cluster.eager-lock: enable
performance.io-cache: off
performance.read-ahead: off
performance.quick-read: off
performance.stat-prefetch: off
performance.strict-write-ordering: on
performance.write-behind: off
nfs.enable-ino32: off
nfs.addr-namelookup: off
nfs.disable: on
performance.cache-refresh-timeout: 4
performance.io-thread-count: 32
performance.low-prio-threads: 32
cluster.server-quorum-type: server
cluster.quorum-type: auto
client.event-threads: 4
server.event-threads: 4
cluster.self-heal-window-size: 256
features.shard-block-size: 512MB
features.shard: on
performance.readdir-ahead: off
vna$ gluster volume remove-brick test3 replica 2 vng.proxmox.softlog:/vmdata/test3 force
vna$ gluster volume info test3Volume Name: test3
Type: Replicate
Volume ID: 36421a23-68c4-455d-8d4c-e21d9428e1da
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: vnb.proxmox.softlog:/vmdata/test3
Brick2: vna.proxmox.softlog:/vmdata/test3
Options Reconfigured:
cluster.data-self-heal-algorithm: full
network.remote-dio: enable
cluster.eager-lock: enable
performance.io-cache: off
performance.read-ahead: off
performance.quick-read: off
performance.stat-prefetch: off
performance.strict-write-ordering: on
performance.write-behind: off
nfs.enable-ino32: off
nfs.addr-namelookup: off
nfs.disable: on
performance.cache-refresh-timeout: 4
performance.io-thread-count: 32
performance.low-prio-threads: 32
cluster.server-quorum-type: server
cluster.quorum-type: auto
client.event-threads: 4
server.event-threads: 4
cluster.self-heal-window-size: 256
features.shard-block-size: 512MB
features.shard: on
performance.readdir-ahead: off
3. Then run a full heal:
vna$ gluster volume heal test3 full
Commit failed on vng.proxmox.softlog. Please check log file for details.
Weird, because of cause the vng brick has been removed. This happens every time.
I have preserved the glustershd logs from vna & vng if needed. There were no heal logs.
-- Lindsay Mathieson
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users