Re: 14.2.16 Low space hindering backfill after reboot

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



What are your full ratios? The defaults are:

    "mon_osd_backfillfull_ratio": "0.900000",
    "mon_osd_full_ratio": "0.950000",

You could temporarily increase the mon_osd_backfillfull_ratio a bit and see if it resolves. But it's not recommended to get an OSD really full, so be careful with that. Do you have the option to add more disks?


Zitat von Marco Pizzolo <marcopizzolo@xxxxxxxxx>:

Hello Everyone,

We seem to be having a problem on one of our ceph clusters post the OS
patch and reboot of one of the nodes.  The three other nodes are showing
OSD fill rates of 77%-81%, but the 60 OSDs contained in the host that was
just rebooted are varying between 64% and 90% since the reboot occurred.
The three other nodes have not yet been patched or rebooted.

The result is:

    health: HEALTH_WARN
            15 nearfull osd(s)
            7 pool(s) nearfull
            Low space hindering backfill (add storage if this doesn't
resolve itself): 15 pgs backfill_toofull
            Degraded data redundancy: 170940/1437684990 objects degraded
(0.012%), 4 pgs degraded, 4 pgs undersized

  services:
    mon: 3 daemons, quorum prdceph01,prdceph02,prdceph03 (age 6h)
    mgr: prdceph01(active, since 5w), standbys: prdceph02, prdceph03,
prdceph04
    mds: ArchiveRepository:1 {0=prdceph01=up:active} 3 up:standby
    osd: 240 osds: 240 up (since 6h), 240 in (since 27h); 16 remapped pgs

  task status:
    scrub status:
        mds.prdceph01: idle

  data:
    pools:   7 pools, 8384 pgs
    objects: 479.23M objects, 557 TiB
    usage:   1.7 PiB used, 454 TiB / 2.1 PiB avail
    pgs:     170940/1437684990 objects degraded (0.012%)
             4155186/1437684990 objects misplaced (0.289%)
             8332 active+clean
             36   active+clean+scrubbing+deep
             11   active+remapped+backfill_toofull
             2    active+undersized+degraded+remapped+backfill_toofull
             2
active+forced_recovery+undersized+degraded+remapped+forced_backfill+backfill_toofull
             1    active+remapped+backfilling

  io:
    client:   9.6 MiB/s rd, 820 KiB/s wr, 1.02k op/s rd, 189 op/s wr
    recovery: 0 B/s, 25 keys/s, 10 objects/s

Any suggestions would be greatly appreciated, as currently it is not able
to complete the repair, nor will it backfill, even when attempting to force.

Many thanks in advance.

Marco
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux