14.2.16 Low space hindering backfill after reboot

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello Everyone,

We seem to be having a problem on one of our ceph clusters post the OS
patch and reboot of one of the nodes.  The three other nodes are showing
OSD fill rates of 77%-81%, but the 60 OSDs contained in the host that was
just rebooted are varying between 64% and 90% since the reboot occurred.
The three other nodes have not yet been patched or rebooted.

The result is:

    health: HEALTH_WARN
            15 nearfull osd(s)
            7 pool(s) nearfull
            Low space hindering backfill (add storage if this doesn't
resolve itself): 15 pgs backfill_toofull
            Degraded data redundancy: 170940/1437684990 objects degraded
(0.012%), 4 pgs degraded, 4 pgs undersized

  services:
    mon: 3 daemons, quorum prdceph01,prdceph02,prdceph03 (age 6h)
    mgr: prdceph01(active, since 5w), standbys: prdceph02, prdceph03,
prdceph04
    mds: ArchiveRepository:1 {0=prdceph01=up:active} 3 up:standby
    osd: 240 osds: 240 up (since 6h), 240 in (since 27h); 16 remapped pgs

  task status:
    scrub status:
        mds.prdceph01: idle

  data:
    pools:   7 pools, 8384 pgs
    objects: 479.23M objects, 557 TiB
    usage:   1.7 PiB used, 454 TiB / 2.1 PiB avail
    pgs:     170940/1437684990 objects degraded (0.012%)
             4155186/1437684990 objects misplaced (0.289%)
             8332 active+clean
             36   active+clean+scrubbing+deep
             11   active+remapped+backfill_toofull
             2    active+undersized+degraded+remapped+backfill_toofull
             2
 active+forced_recovery+undersized+degraded+remapped+forced_backfill+backfill_toofull
             1    active+remapped+backfilling

  io:
    client:   9.6 MiB/s rd, 820 KiB/s wr, 1.02k op/s rd, 189 op/s wr
    recovery: 0 B/s, 25 keys/s, 10 objects/s

Any suggestions would be greatly appreciated, as currently it is not able
to complete the repair, nor will it backfill, even when attempting to force.

Many thanks in advance.

Marco
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux