Adding OSD's results in slow ops, inactive PG's

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi

We have a cluster with which currently looks like so:

    services:
      mon: 5 daemons, quorum lazy,jolly,happy,dopey,sleepy (age 13d)
      mgr: jolly.tpgixt(active, since 25h), standbys: dopey.lxajvk, lazy.xuhetq
      mds: 1/1 daemons up, 2 standby
      osd: 449 osds: 425 up (since 15m), 425 in (since 5m); 5104 remapped pgs
data:
      volumes: 1/1 healthy
      pools:   13 pools, 11153 pgs
      objects: 304.11M objects, 988 TiB
      usage:   1.6 PiB used, 1.4 PiB / 2.9 PiB avail
      pgs:     6/1617270006 objects degraded (0.000%)
               366696947/1617270006 objects misplaced (22.674%)
               6043 active+clean
               5041 active+remapped+backfill_wait
               66   active+remapped+backfilling
               2    active+recovery_wait+degraded+remapped
               1    active+recovering+degraded


It's currently rebalancing after adding a node, but this rebalance has been rather slow -- right now it's running 66 backfills, but it seems to stabilize around 8 backfills eventually. We figured that perhaps adding another node might speed things up.

Immediately upon adding the node, we get slow ops and inactive PG's. Removing the new node gets us back in working order.

It turns out that even adding 1 OSD breaks the cluster, and immediately sends it here:

    [WRN] PG_DEGRADED: Degraded data redundancy: 6/1617265712 objects degraded (0.000%), 3 pgs degraded
        pg 37.c8 is active+recovery_wait+degraded+remapped, acting [410,163,236,209,7,283,155,143,78]
        pg 37.1a1 is active+recovering+degraded, acting [234,424,163,74,22,128,177,153,181]
        pg 37.1da is active+recovery_wait+degraded+remapped, acting [163,408,230,190,93,284,50,78,44]
    [WRN] SLOW_OPS: 22 slow ops, oldest one blocked for 54 sec, daemons [osd.11,osd.110,osd.112,osd.117,osd.120,osd.123,osd.13,osd.136,osd.144,osd.157]... have slow ops.

The OSD added had number 431, so it does not appear to be the immediate cause of the slow ops, however, removing 431 immediately clears the problem.

We thought we might be experiencing 'Crush giving up too soon' symptoms [1], as we have seen similar behaviour on another pool, but it does not appear to be the case here. We went through the motions described on the page and everything looked OK.

At least one pool which stops working is a 4+2 EC pool, placed on spinning rust, some 200-ish disks distributed across 13 nodes. I'm not sure if other pools break, but that particular 4+2 EC pool is rather important so I'm a little wary of experimenting blindly.

Any thoughts on where to look next?

Thanks,
Ruben Vestergaard

[1] https://docs.ceph.com/en/reef/rados/troubleshooting/troubleshooting-pg/#crush-gives-up-too-soon
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux