Im not sure that setup to 128 max backfills is a good idea, I shared our config for recovery and backfilling: osd recovery threads = 4 osd recovery op priority = 1 osd recovery max active = 2 osd recovery max single start = 1 osd max backfills = 4 osd backfill scan max = 16 osd backfill scan min = 4 osd client op priority = 63 Check the fullest osd my be one is full and this prevent of use the other 131TB raw or distribution is not even across osd Check mon osd full ratio and mon osd nearfull ratio , maybe put just a 2% more in full ratio makes different for you. Regards De: ceph-users <ceph-users-bounces@xxxxxxxxxxxxxx> En nombre de Kári Bertilsson Hello I am running cephfs with 8/2 erasure coding. I had about 40tb usable free(110tb raw), one small disk crashed and i added 2x10tb disks. Now it's backfilling & recovering with 0B free and i can't read a single file from the file system... This happend with max-backfilling 4, but i have increased max backfills to 128, to hopefully get this over a little faster since system has been unusable for 12 hours anyway. Not sure yet if that was a good idea. 131TB of raw space was somehow not enough to keep things running. Any tips to avoid this kind of scenario in the future ? GLOBAL: data: Is there anything i can do to restore reading ? I can understand writing not working, but why is it blocking reading also ? Any tips ? |
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com