Continual heals happening on cluster

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



gluster 3.7.10
Proxmox (debian jessie)

I'm finding the following more than a little concerning. I've created a datastore with the following settings:

Volume Name: datastore4
Type: Replicate
Volume ID: 0ba131ef-311d-4bb1-be46-596e83b2f6ce
Status: Started
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: vnb.proxmox.softlog:/tank/vmdata/datastore4
Brick2: vng.proxmox.softlog:/tank/vmdata/datastore4
Brick3: vna.proxmox.softlog:/tank/vmdata/datastore4
Options Reconfigured:
features.shard-block-size: 64MB
network.remote-dio: enable
cluster.eager-lock: enable
performance.io-cache: off
performance.read-ahead: off
performance.quick-read: off
performance.stat-prefetch: on
performance.strict-write-ordering: off
nfs.enable-ino32: off
nfs.addr-namelookup: off
nfs.disable: on
cluster.server-quorum-type: server
cluster.quorum-type: auto
features.shard: on
cluster.data-self-heal: on
cluster.self-heal-window-size: 1024
transport.address-family: inet
performance.readdir-ahead: on


I've transferred 12 Windows VM's to it (gfapi) and am running them all, spread across three nodes.

"gluster volume heal datastore3 statistics heal-count" shows zero heals on all nodes.

but "gluster volume heal datastore4 info" shows heals occurring on mutliple shards on all nodes, different shards each time its called.

gluster volume heal datastore4 info
Brick vnb.proxmox.softlog:/tank/vmdata/datastore4
/.shard/d297f8d6-e263-4af3-9384-6492614dc115.221
/.shard/744c5059-303d-4e82-b5be-0a5f53b1aeff.1362
/.shard/bbdff876-290a-4e5e-93ef-a95276d57220.942
/.shard/eaeb41ec-9c0d-4fed-984f-cf832d8d33e0.1032
/.shard/f8ce4b49-14d0-46ef-9a95-456884f34fd4.623
/.shard/e9a39d2e-a1b7-4ea0-9d8c-b55370048d03.483
/.shard/f8ce4b49-14d0-46ef-9a95-456884f34fd4.47
/.shard/eaeb41ec-9c0d-4fed-984f-cf832d8d33e0.160
Status: Connected
Number of entries: 8

Brick vng.proxmox.softlog:/tank/vmdata/datastore4
/.shard/bd493985-2ee6-43f1-b8d5-5f0d5d3aa6f5.33
/.shard/d297f8d6-e263-4af3-9384-6492614dc115.48
/.shard/744c5059-303d-4e82-b5be-0a5f53b1aeff.1304
/.shard/d297f8d6-e263-4af3-9384-6492614dc115.47
/.shard/719041d0-d755-4bc6-a5fc-6b59071fac17.142
Status: Connected
Number of entries: 5

Brick vna.proxmox.softlog:/tank/vmdata/datastore4
/.shard/d297f8d6-e263-4af3-9384-6492614dc115.357
/.shard/bbdff876-290a-4e5e-93ef-a95276d57220.996
/.shard/d297f8d6-e263-4af3-9384-6492614dc115.679
/.shard/d297f8d6-e263-4af3-9384-6492614dc115.496
/.shard/eaeb41ec-9c0d-4fed-984f-cf832d8d33e0.160
/.shard/719041d0-d755-4bc6-a5fc-6b59071fac17.954
/.shard/d297f8d6-e263-4af3-9384-6492614dc115.678
/.shard/719041d0-d755-4bc6-a5fc-6b59071fac17.852
/.shard/bbdff876-290a-4e5e-93ef-a95276d57220.1544
Status: Connected
Number of entries: 9



-- 
Lindsay Mathieson
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux