These symptoms appear to be the same as I've recorded in this post:
http://lists.gluster.org/pipermail/gluster-users/2017-September/032435.html
http://lists.gluster.org/pipermail/gluster-users/2017-September/032435.html
On Wed, Sep 13, 2017 at 7:01 AM, Atin Mukherjee <atin.mukherjee83@xxxxxxxxx> wrote:
Additionally the brick log file of the same brick would be required. Please look for if brick process went down or crashed. Doing a volume start force should resolve the issue.On Wed, 13 Sep 2017 at 16:28, Gaurav Yadav <gyadav@xxxxxxxxxx> wrote:Please send me the logs as well i.e glusterd.logs and cmd_history.log.On Wed, Sep 13, 2017 at 1:45 PM, lejeczek <peljasz@xxxxxxxxxxx> wrote:
On 13/09/17 06:21, Gaurav Yadav wrote:
Please provide the output of gluster volume info, gluster volume status and gluster peer status.
Apart from above info, please provide glusterd logs, cmd_history.log.
Thanks
Gaurav
On Tue, Sep 12, 2017 at 2:22 PM, lejeczek <peljasz@xxxxxxxxxxx <mailto:peljasz@xxxxxxxxxxx>> wrote:
hi everyone
I have 3-peer cluster with all vols in replica mode, 9
vols.
What I see, unfortunately, is one brick fails in one
vol, when it happens it's always the same vol on the
same brick.
Command: gluster vol status $vol - would show brick
not online.
Restarting glusterd with systemclt does not help, only
system reboot seem to help, until it happens, next time.
How to troubleshoot this weird misbehaviour?
many thanks, L.
.
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx<mailto:Gluster-users@gluster.org >
http://lists.gluster.org/mailman/listinfo/gluster-users
<http://lists.gluster.org/mailman/listinfo/gluster-users >
hi, here:
$ gluster vol info C-DATA
Volume Name: C-DATA
Type: Replicate
Volume ID: 18ffba73-532e-4a4d-84da-fceea52f8c2e
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: 10.5.6.49:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-C-DATA
Brick2: 10.5.6.100:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-C-DATA
Brick3: 10.5.6.32:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-C-DATA
Options Reconfigured:
performance.md-cache-timeout: 600
performance.cache-invalidation: on
performance.stat-prefetch: on
features.cache-invalidation-timeout: 600
features.cache-invalidation: on
performance.io-thread-count: 64
performance.cache-size: 128MB
cluster.self-heal-daemon: enable
features.quota-deem-statfs: on
changelog.changelog: on
geo-replication.ignore-pid-check: on
geo-replication.indexing: on
features.inode-quota: on
features.quota: on
performance.readdir-ahead: on
nfs.disable: on
transport.address-family: inet
performance.cache-samba-metadata: on
$ gluster vol status C-DATA
Status of volume: C-DATA
Gluster processTCP Port RDMA Port Online Pid
------------------------------------------------------------ ------------------
Brick 10.5.6.49:/__.aLocalStorages/0/0-GLUS
TERs/0GLUSTER-C-DATAN/A N/A N N/A
Brick 10.5.6.100:/__.aLocalStorages/0/0-GLU
STERs/0GLUSTER-C-DATA49152 0 Y 9376
Brick 10.5.6.32:/__.aLocalStorages/0/0-GLUS
TERs/0GLUSTER-C-DATA49152 0 Y 8638
Self-heal Daemon on localhost N/A N/A Y 387879
Quota Daemon on localhost N/A N/A Y 387891
Self-heal Daemon on rider.private.ccnr.ceb.
private.cam.ac.ukN/A N/A Y 16439
Quota Daemon on rider.private.ccnr.ceb.priv
ate.cam.ac.ukN/A N/A Y 16451
Self-heal Daemon on 10.5.6.32 N/A N/A Y 7708
Quota Daemon on 10.5.6.32 N/A N/A Y 8623
Self-heal Daemon on 10.5.6.17 N/A N/A Y 20549
Quota Daemon on 10.5.6.17 N/A N/A Y 9337
Task Status of Volume C-DATA
------------------------------------------------------------ ------------------
There are no active volume tasks______________________________
.
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users _________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users ----Atin
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://lists.gluster.org/mailman/listinfo/gluster-users