On 10/23/2018 03:00 PM, hsafe wrote:
Hello all,
Can somebody please respond to this? as of now if I run "gluster
volume heal gv1 info"
there is infinite number of lines of gfid which never ends...usually
and in stable scenario this ended with some numbers and status but
currently it never finishes...is it a bad sign ? is it a loop? are
there any actions required to do beside gluster?
Are the files list of files constant or are they varying?
Are any of the files in split-brain?
Do you see messages related to self-heal completion in
/var/log/glusterfs/glustershd.log on your nodes?
If you see different set of files each time, or the same set of files
after the got healed (timestamps and gfids can be seen in
glustershd.log), then you might be experiencing network issues between
your client and the bricks resulting in constant heals. Check your
client (fuse?) logs for disconnects to the bricks.
-Ravi
Appreciate any help...
On 10/21/18 8:05 AM, hsafe wrote:
Hello all gluster community,
I am in a scenario unmatched for the past year of using glusterfs in
a 2 replica set on glusterfs 3.10.12 servers where they are the
storage back of my application which saves small images into them.
Now the problem I face and unique for the time is that whenever we
were asynced or one server went down; bringing the other one will
start the self heal and eventually we could see the clustered volume
in sync, but now if I run the volume heal info the list of the gfid
does not even finish after couple of hours. if I look at the heal log
I can see that the process is ongoing but it a very small scale and
speed!
My question is how can I expect it finished and how can I speed it up
there?
Here is a bit of info:
Status of volume: gv1
Gluster process TCP Port RDMA Port
Online Pid
------------------------------------------------------------------------------
Brick IMG-01:/images/storage/brick1 49152 0 Y 4176
Brick IMG-02:/images/storage/brick1 49152 0 Y 4095
Self-heal Daemon on localhost N/A N/A Y 4067
Self-heal Daemon on IMG-01 N/A N/A Y 4146
Task Status of Volume gv1
------------------------------------------------------------------------------
There are no active volume tasks
Status of volume: gv2
Gluster process TCP Port RDMA Port
Online Pid
------------------------------------------------------------------------------
Brick IMG-01:/data/brick2 49153 0 Y 4185
Brick IMG-02:/data/brick2 49153 0 Y 4104
NFS Server on localhost N/A N/A N N/A
Self-heal Daemon on localhost N/A N/A Y 4067
NFS Server on IMG-01 N/A N/A N N/A
Self-heal Daemon on IMG-01 N/A N/A Y 4146
Task Status of Volume gv2
------------------------------------------------------------------------------
There are no active volume tasks
gluster> peer status
Number of Peers: 1
Hostname: IMG-01
Uuid: 5faf60fc-7f5c-4c6e-aa3f-802482391c1b
State: Peer in Cluster (Connected)
Hostname: IMG-01
Uuid: 5faf60fc-7f5c-4c6e-aa3f-802482391c1b
State: Peer in Cluster (Connected)
gluster> exit
root@NAS02:/var/log/glusterfs# gluster volume gv1 info
unrecognized word: gv1 (position 1)
root@NAS02:/var/log/glusterfs# gluster volume info
Volume Name: gv1
Type: Replicate
Volume ID: f1c955a1-7a92-4b1b-acb5-8b72b41aaace
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: IMG-01:/images/storage/brick1
Brick2: IMG-02:/images/storage/brick1
Options Reconfigured:
server.event-threads: 4
performance.cache-invalidation: on
performance.stat-prefetch: on
features.cache-invalidation-timeout: 600
features.cache-invalidation: on
cluster.lookup-optimize: on
cluster.shd-max-threads: 4
cluster.readdir-optimize: on
performance.md-cache-timeout: 30
cluster.background-self-heal-count: 32
server.statedump-path: /tmp
performance.readdir-ahead: on
nfs.disable: true
network.inode-lru-limit: 50000
features.bitrot: off
features.scrub: Inactive
performance.cache-max-file-size: 16MB
client.event-threads: 8
cluster.eager-lock: on
cluster.self-heal-daemon: enable
Please do help me out...Thanks
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users