Re: 17.2.7: Backfilling deadlock / stall / stuck / standstill

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I faced a similar issue. The PG just would never finish recovery. Changing
all OSDs in the PG to "osd_op_queue wpq" and then restarting them serially
ultimately allowed the PG to recover. Seemed to be some issue with mclock.

Respectfully,

*Wes Dillingham*
wes@xxxxxxxxxxxxxxxxx
LinkedIn <http://www.linkedin.com/in/wesleydillingham>


On Fri, Jan 26, 2024 at 7:57 AM Kai Stian Olstad <ceph+list@xxxxxxxxxx>
wrote:

> Hi,
>
> This is a cluster running 17.2.7 upgraded from 16.2.6 on the 15 January
> 2024.
>
> On Monday 22 January we had 4 HDD all on different server with I/O-error
> because of some damage sectors, the OSD is hybrid so the DB is on SSD, 5
> HDD share 1 SSD.
> I set the OSD out, ceph osd out 223 269 290 318 and all hell broke
> loose.
>
> I took only minutes before the users complained about Ceph not working.
> Ceph status reportet slow OPS on the OSDs that was set to out, and “ceph
> tell osd.<id> dump_ops_in_flight” against the out OSDs it just hang,
> after 30 minutes I stopped the dump command.
> Long story short I ended up running “ceph osd set nobackfill” to slow
> ops was gone and then unset it when the slow ops message disappeared.
> I needed to run that all the time so the cluster didn’t come to a holt
> so this oneliner loop was used
> “while true; do ceph -s | grep -qE "oldest one blocked for [0-9]{2,}" &&
> (date; ceph osd set nobackfill; sleep 15; ceph osd unset nobackfill);
> sleep 10; done”
>
>
> But now 4 days later the backfilling has stopped progressing completely
> and the number of misplaced object is increasing.
> Some PG has 0 misplaced object but sill have backfilling state, and been
> in this state for over 24 hours now.
>
> I have a hunch that it’s because of PG 404.6e7 is in state
> “active+recovering+degraded+remapped” it’s been in this state for over
> 48 hours.
> It’s has possible 2 missing object, but since they are not unfound I
> can’t delete them with “ceph pg 404.6e7 mark_unfound_lost delete”
>
> Could someone please help to solve this?
> Down below is some output of ceph commands, I’ll also attache them.
>
>
> ceph status (only removed information about no running scrub and
> deep_scrub)
> ---
>    cluster:
>      id:     b321e76e-da3a-11eb-b75c-4f948441dcd0
>      health: HEALTH_WARN
>              Degraded data redundancy: 2/6294904971 objects degraded
> (0.000%), 1 pg degraded
>
>    services:
>      mon: 3 daemons, quorum ceph-mon-1,ceph-mon-2,ceph-mon-3 (age 11d)
>      mgr: ceph-mon-1.ptrsea(active, since 11d), standbys:
> ceph-mon-2.mfdanx
>      mds: 1/1 daemons up, 1 standby
>      osd: 355 osds: 355 up (since 22h), 351 in (since 4d); 18 remapped
> pgs
>      rgw: 7 daemons active (7 hosts, 1 zones)
>
>    data:
>      volumes: 1/1 healthy
>      pools:   14 pools, 3945 pgs
>      objects: 1.14G objects, 1.1 PiB
>      usage:   1.8 PiB used, 1.2 PiB / 3.0 PiB avail
>      pgs:     2/6294904971 objects degraded (0.000%)
>               2980455/6294904971 objects misplaced (0.047%)
>               3901 active+clean
>               22   active+clean+scrubbing+deep
>               17   active+remapped+backfilling
>               4    active+clean+scrubbing
>               1    active+recovering+degraded+remapped
>
>    io:
>      client:   167 MiB/s rd, 13 MiB/s wr, 6.02k op/s rd, 2.35k op/s wr
>
>
> ceph health detail (only removed information about no running scrub and
> deep_scrub)
> ---
> HEALTH_WARN Degraded data redundancy: 2/6294902067 objects degraded
> (0.000%), 1 pg degraded
> [WRN] PG_DEGRADED: Degraded data redundancy: 2/6294902067 objects
> degraded (0.000%), 1 pg degraded
>      pg 404.6e7 is active+recovering+degraded+remapped, acting
> [223,274,243,290,286,283]
>
>
> ceph pg 202.6e7 list_unfound
> ---
> {
>      "num_missing": 2,
>      "num_unfound": 0,
>      "objects": [],
>      "state": "Active",
>      "available_might_have_unfound": true,
>      "might_have_unfound": [],
>      "more": false
> }
>
> ceph pg 404.6e7 query | jq .recovery_state
> ---
> [
>    {
>      "name": "Started/Primary/Active",
>      "enter_time": "2024-01-26T09:08:41.918637+0000",
>      "might_have_unfound": [
>        {
>          "osd": "243(2)",
>          "status": "already probed"
>        },
>        {
>          "osd": "274(1)",
>          "status": "already probed"
>        },
>        {
>          "osd": "275(0)",
>          "status": "already probed"
>        },
>        {
>          "osd": "283(5)",
>          "status": "already probed"
>        },
>        {
>          "osd": "286(4)",
>          "status": "already probed"
>        },
>        {
>          "osd": "290(3)",
>          "status": "already probed"
>        },
>        {
>          "osd": "335(3)",
>          "status": "already probed"
>        }
>      ],
>      "recovery_progress": {
>        "backfill_targets": [
>          "275(0)",
>          "335(3)"
>        ],
>        "waiting_on_backfill": [],
>        "last_backfill_started":
>
> "404:e76011a9:::1f244892-a2e7-406b-aa62-1b13511333a2.625411.18_56463c71-286c-4399-8d5d-0c278b7c97fd:head",
>        "backfill_info": {
>          "begin": "MIN",
>          "end": "MIN",
>          "objects": []
>        },
>        "peer_backfill_info": [],
>        "backfills_in_flight": [],
>        "recovering": [],
>        "pg_backend": {
>          "recovery_ops": [],
>          "read_ops": []
>        }
>      }
>    },
>    {
>      "name": "Started",
>      "enter_time": "2024-01-26T09:08:40.909151+0000"
>    }
> ]
>
>
> ceph pg ls recovering backfilling
> ---
> PG       OBJECTS  DEGRADED  MISPLACED  UNFOUND  BYTES
> OMAP_BYTES*  OMAP_KEYS*  LOG    LOG_DUPS  STATE
>        SINCE  VERSION          REPORTED          UP
>        ACTING
> 404.bc    287986         0          0        0  512046716673
> 0           0  10091         0           active+recovering+remapped
> 2h   217988'1385478   217988:10897565  [193,297,279,276,136,197]p193
> [223,297,269,276,136,197]p223
> 404.c4    288236         0     288236        0  511669837559
> 0           0  10063         0          active+remapped+backfilling
> 24h   217988'1378228   217988:11719855  [156,186,178,345,339,177]p156
> [223,186,178,345,339,177]p223
> 404.12a   287544         0          0        0  512246100354
> 0           0  10009         0          active+remapped+backfilling
> 24h   217988'1392371   217988:13739524  [248,178,250,145,304,272]p248
> [223,178,250,145,304,272]p223
> 404.1c1   287739         0     286969        0  511800674008
> 0           0  10047         0          active+remapped+backfilling
> 2d   217988'1402889   217988:10975174  [332,246,183,169,280,255]p332
> [318,246,183,169,280,255]p318
> 404.258   287737         0     277111        0  510099501390
> 0           0  10077         0          active+remapped+backfilling
> 24h   217988'1451778   217988:12780104  [308,199,134,342,188,221]p308
> [318,199,134,342,188,221]p318
> 404.269   287990         0          0        0  512343190608
> 0           0  10043         0          active+remapped+backfilling
> 24h   217988'1358446   217988:14020217  [275,205,283,247,211,292]p275
> [223,205,283,247,211,292]p223
> 404.34e   287624         0     277899        0  510447074297
> 0           0  10002         0          active+remapped+backfilling
> 24h   217988'1392933   217988:12636557  [322,141,338,168,251,218]p322
> [318,141,338,168,251,218]p318
> 404.39c   287844         0     286692        0  512947685682
> 0           0  10017         0          active+remapped+backfilling
> 2d   217988'1414697   217988:11004944  [288,188,131,299,295,181]p288
> [318,188,131,299,295,181]p318
> 404.511   287589         0          0        0  512014863711
> 0           0  10057         0          active+remapped+backfilling
> 24h   217988'1368741   217988:11544729  [166,151,327,333,186,150]p166
> [223,151,327,333,186,150]p223
> 404.5f1   288126         0     286621        0  510850256945
> 0           0  10071         0          active+remapped+backfilling
> 24h   217988'1365831   217988:10348125  [214,332,289,184,255,160]p214
> [223,332,289,184,255,160]p223
> 404.62a   288035         0          0        0  511318662269
> 0           0  10014         0          active+remapped+backfilling
> 3h   217988'1358010   217988:12528704  [322,260,259,319,149,152]p322
> [318,260,259,319,149,152]p318
> 404.63d   287372         0     286559        0  508783837699
> 0           0  10074         0          active+remapped+backfilling
> 24h   217988'1402174   217988:11685744  [303,307,186,350,161,267]p303
> [318,307,186,350,161,267]p318
> 404.6e3   288110         0          0        0  509047569016
> 0           0  10049         0          active+remapped+backfilling
> 24h   217988'1368547   217988:12202278  [166,317,233,144,337,240]p166
> [223,317,233,144,337,240]p223
> 404.6e7   287856         2          2        0  510383394904
> 0           0  10047         0  active+recovering+degraded+remapped
> 3h   217988'1356501   217988:13157749  [275,274,243,335,286,283]p275
> [223,274,243,290,286,283]p223
> 404.7d2   287619         0     286026        0  510708533087
> 0           0  10093         0          active+remapped+backfilling
> 3d   217988'1397393   217988:12146656  [185,139,299,222,155,149]p185
> [223,139,299,222,155,149]p223
> 412.119   711468         0          0        0  207473602580
> 0           0  10099         0          active+remapped+backfilling
> 24h  217988'21613330   217988:87589096  [352,207,292,314,230,262]p352
> [318,207,292,314,230,262]p318
> 412.12f   711529         0     701279        0  208498170310
> 0           0  10033         0          active+remapped+backfilling
> 24h  217988'14873593   217988:86198113  [303,305,183,215,130,244]p303
> [318,305,183,215,130,244]p318
> 412.1fb   713044         0       3166        0  207787641403
> 0           0  10097         0          active+remapped+backfilling
> 2d  217988'14893270  217988:105346132  [156,137,228,241,262,353]p156
> [223,137,228,241,262,353]p223
>
>
> ceph osd tree out
> ---
> ID   CLASS  WEIGHT      TYPE NAME             STATUS  REWEIGHT  PRI-AFF
>   -1         3112.43481  root default
> -67          192.35847      host ceph-hd-001
> 269    hdd    12.82390          osd.269           up         0  1.00000
> -49          192.35847      host ceph-hd-003
> 223    hdd    12.82390          osd.223           up         0  1.00000
> -73          192.35847      host ceph-hd-011
> 290    hdd    12.82390          osd.290           up         0  1.00000
> -79          192.35847      host ceph-hd-014
> 318    hdd    12.82390          osd.318           up         0
> 1.00000_______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux