Re: PG_BACKFILL_FULL

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thank for your response and advice.

On 16/01/23 15:17, Boris Behrens wrote:
Hmm.. I ran into some similar issue.

IMHO there are two ways to work around the problem until the new disk in place: 1. change the backfill full threshold (I use these commands: https://www.suse.com/support/kb/doc/?id=000019724 <https://www.suse.com/support/kb/doc/?id=000019724>)

If I understand correctly the "backfillfull_ratio" is a ratio above which a warning is triggered and the cluster will deny backfilling to the OSD in question. But my OSD (87.53% ) is not above the ratio (90%). Granted, it is possible that "after" the 3 PGs are moved to that OSD the ratio will be crossed, but right now we are bellow.

My final goal is to empty the "damaged" OSD/disk, replace it and fill it up again. But I can't do that because it has still 3 PGs on it.

2. reweight the backfill full OSDs just a little bit, so they move data to disks that are free enough (i.e. `ceph osd reweight osd.60 0.9`) - if you have enough capacity in the cluster (577+ OSDs should be able to take that :) )

You mean first reweight to (something like) .9, let the OSD free up space and then put it back to 1 so the PGs can be backfilled?

Idally I would like just to manually set the new "location" of the PGs away from the nearfull OSD.60. I see there are some commands called "ceph osd pg-upmap" and "ceph osd pg-upmap-items" which could be the right tool for what I want to achieve. But I didn't found a lot of information about it, sombody knows something more, are those tools "safe" to run in my case?

Thanks a lot
Iztok



Cheers
  Boris


Am Mo., 16. Jan. 2023 um 15:01 Uhr schrieb Iztok Gregori <iztok.gregori@xxxxxxxxxx <mailto:iztok.gregori@xxxxxxxxxx>>:

    Hi to all!

    We are in a situation where we have 3 PG in
    "active+remapped+backfill_toofull". It happened when we executed a
    "gentle-reweight" to zero of one OSD (osd.77) to swap it with a new one
    (the current one registered some read errors and it's to be replaced
    just-in-case).

     > # ceph health detail:
     > [WRN] PG_BACKFILL_FULL: Low space hindering backfill (add storage
    if this doesn't resolve itself): 3 pgs backfill_toofull
     >     pg 10.46c is active+remapped+backfill_toofull, acting [77,607,96]
     >     pg 10.8ad is active+remapped+backfill_toofull, acting
    [577,152,77]
     >     pg 10.b15 is active+remapped+backfill_toofull, acting
    [483,348,77]

    Our cluster is a little unbalanced and we have 7 OSD nearfull (I think
    it's because we have 4 nodes with 6 TB disks and the other 19 have 10
    TB, but should be unrelated, why is the cluster unbalanced I mean, to
    the backfill_toofull ) not by too much (less then 88%). I'm not too
    much
    worried about it, we will add new storage this month (if the servers
    will arrive) and we will get rid of the old 6 TB servers.

    If I dump the PGs I see, if I'm not mistaken, that the osd.77 will be
    "replaced" by the osd.60, which is one of the nearfull ones (the top
    one
    with 87.53% used).


     > # ceph pg dump:
     >
> 10.b15     37236                   0         0      37236 0  155249620992            0           0   5265      5265 active+remapped+backfill_toofull  2023-01-16T14:45:46.155801+0100   305742'144106      305742:901513   [483,348,60]         483  [483,348,77]             483      305211'144056 2023-01-11T10:20:56.600135+0100      305211'144056 2023-01-11T10:20:56.600135+0100              0 > 10.8ad     37518                   0         0      37518 0  156345024512            0           0   5517      5517 active+remapped+backfill_toofull  2023-01-16T14:45:38.510038+0100   305213'142117      305742:937228   [577,60,152]         577  [577,152,77]             577      303828'142043 2023-01-06T17:52:02.523104+0100      303334'141645 2022-12-20T17:39:22.668083+0100              0 > 10.46c     36710                   0         0      36710 0  153023443456            0           0   8172      8172 active+remapped+backfill_toofull  2023-01-16T14:45:29.284223+0100   305298'141437      305741:877331    [60,607,96]          60 [77,607,96]              77      304802'141358 2023-01-08T21:39:23.469198+0100      304363'141349 2023-01-01T18:13:45.645494+0100              0

     > # ceph osd df:
     >  60    hdd   5.45999   1.00000  5.5 TiB  4.8 TiB   697 GiB  128
    MiB       0 B   697 GiB  87.53  1.29   37      up

    In this situation was the correct way to address the problem?
    reweight-by-utilization the osd.60 to free up space (the OSD is a 6 TB
    disk, and other OSD on the same host are nearfull)? There is other way
    to manually map a PG to a different OSD?

    Thank you for your attention

    Iztok Gregori
    _______________________________________________
    ceph-users mailing list -- ceph-users@xxxxxxx
    <mailto:ceph-users@xxxxxxx>
    To unsubscribe send an email to ceph-users-leave@xxxxxxx
    <mailto:ceph-users-leave@xxxxxxx>



--
Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im groüen Saal.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux