Subject: OSDs added, remapped pgs and objects misplaced cycling up and down

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

ceph-16.2.9

I've added some new osds - some added to existing hosts and some on newly-commissioned hosts. The new osds were added to the data side of an existing EC 8.3 pool.

I've been waiting for the system to finish remapping / backfilling for some time. Originally the number of remapped pgs and objects misplaced was steadily decreasing. However in the last few days both remapped pgs and objects misplaced have increased several times. Both then steadily decrease before jumping up again. See below
for some details on this.

I'm guessing these are associated stats, i.e. the number of objects misplaced reflects the objects in the remapped pgs so it's not surprising they're moving in concert.

I thought at first that this cycling might be due to the balancer (which was enabled), but the cycling is continuing after the balancer was disabled, e.g.:

$ ts; ceph balancer status
2023-02-09 12:39:14
{
    "active": true,
    "last_optimize_duration": "0:00:00.000313",
    "last_optimize_started": "Thu Feb  9 01:38:55 2023",
    "mode": "upmap",
    "optimize_result": "Too many objects (0.078537 > 0.050000) are misplaced; try again later",
    "plans": []
}
$ ts; ceph balancer off
2023-02-09 13:03:27 ceph balancer off

4 days later, the balancer is still inactive, and "last_optimize_started" indicates it hasn't run since before it was turned off:

$ ts; ceph balancer status
2023-02-13 10:18:18
{
    "active": false,
    "last_optimize_duration": "0:00:00.000420",
    "last_optimize_started": "Thu Feb  9 02:02:56 2023",
    "mode": "upmap",
    "optimize_result": "Too many objects (0.078101 > 0.050000) are misplaced; try again later",
    "plans": []
}

Since the balancer was turned off the cycling continues: "misplaced" gets down to 5% and then rapidly increases to 9%, in conjunction with "starting backfill" messages, before starting to slowly decrease once again. See below for a log extract.

Is this "sawtooth" pattern of remapped pgs and misplaced objects a normal consequence of adding OSDs?

Will it eventually settle done itself or do I need to "do something"?

Is there some way of telling how much work remains until it all settles down?

Cheers,

Chris

------------------------------------------------------------------------------
#
# General health
#   scrubs disabled until added OSDs complete backfilling etc.
#
$ ceph -s
  cluster:
    id:     c6618970-0ce0-4cb2-bc9a-dd5f29b62e24
    health: HEALTH_WARN
            noout,noscrub,nodeep-scrub flag(s) set
            5945 pgs not deep-scrubbed in time
            5945 pgs not scrubbed in time
            5 slow ops, oldest one blocked for 464 sec, mon.k2 has slow ops
services:
    mon: 5 daemons, quorum k2,b2,b4,k1,b5 (age 5d)
    mgr: b4(active, since 3M), standbys: b5, b2
    osd: 82 osds: 82 up (since 2w), 82 in (since 2w); 12 remapped pgs
         flags noout,noscrub,nodeep-scrub
data:
    pools:   16 pools, 5945 pgs
    objects: 70.35M objects, 88 TiB
    usage:   186 TiB used, 242 TiB / 428 TiB avail
    pgs:     50050192/700705500 objects misplaced (7.143%)
             5933 active+clean
             10   active+remapped+backfill_wait
             2    active+remapped+backfilling
io:
    client:   2.2 MiB/s rd, 1.8 MiB/s wr, 395 op/s rd, 122 op/s wr
    recovery: 24 MiB/s, 24 objects/s

#
# remapped pgs steadily decreasing then started periodically increasing
#   (extract from periodic "ceph -s")
#
2023-02-06 00:00:00    osd: 82 osds: 82 up (since 7d), 82 in (since 6d); 26 remapped pgs
2023-02-06 04:00:00    osd: 82 osds: 82 up (since 7d), 82 in (since 6d); 25 remapped pgs
2023-02-06 06:00:00    osd: 82 osds: 82 up (since 7d), 82 in (since 6d); 24 remapped pgs
2023-02-06 09:00:00    osd: 82 osds: 82 up (since 7d), 82 in (since 6d); 23 remapped pgs
2023-02-06 12:30:00    osd: 82 osds: 82 up (since 7d), 82 in (since 7d); 22 remapped pgs
2023-02-06 14:00:00    osd: 82 osds: 82 up (since 7d), 82 in (since 7d); 21 remapped pgs
2023-02-06 14:30:00    osd: 82 osds: 82 up (since 7d), 82 in (since 7d); 20 remapped pgs
2023-02-06 20:30:00    osd: 82 osds: 82 up (since 7d), 82 in (since 7d); 19 remapped pgs
2023-02-07 00:00:00    osd: 82 osds: 82 up (since 8d), 82 in (since 7d); 18 remapped pgs
2023-02-07 01:00:00    osd: 82 osds: 82 up (since 8d), 82 in (since 7d); 17 remapped pgs
2023-02-07 11:00:00    osd: 82 osds: 82 up (since 8d), 82 in (since 8d); 15 remapped pgs
2023-02-07 16:30:00    osd: 82 osds: 82 up (since 8d), 82 in (since 8d); 13 remapped pgs
2023-02-07 22:30:00    osd: 82 osds: 82 up (since 9d), 82 in (since 8d); 11 remapped pgs
2023-02-08 15:30:00    osd: 82 osds: 82 up (since 9d), 82 in (since 9d); 9 remapped pgs
2023-02-08 21:30:00    osd: 82 osds: 82 up (since 10d), 82 in (since 9d); 15 remapped pgs	<<< increase
2023-02-09 07:30:00    osd: 82 osds: 82 up (since 10d), 82 in (since 9d); 13 remapped pgs
2023-02-09 13:03:27 ceph balancer off
2023-02-09 22:30:00    osd: 82 osds: 82 up (since 11d), 82 in (since 10d); 11 remapped pgs
2023-02-10 14:00:00    osd: 82 osds: 82 up (since 11d), 82 in (since 11d); 9 remapped pgs
2023-02-10 19:00:00    osd: 82 osds: 82 up (since 11d), 82 in (since 11d); 15 remapped pgs	<<< increase
2023-02-11 04:30:00    osd: 82 osds: 82 up (since 12d), 82 in (since 12d); 13 remapped pgs
2023-02-11 19:00:00    osd: 82 osds: 82 up (since 12d), 82 in (since 12d); 11 remapped pgs
2023-02-12 12:00:00    osd: 82 osds: 82 up (since 13d), 82 in (since 13d); 9 remapped pgs
2023-02-12 18:00:00    osd: 82 osds: 82 up (since 13d), 82 in (since 13d); 15 remapped pgs	<<< increase
2023-02-12 20:30:00    osd: 82 osds: 82 up (since 13d), 82 in (since 13d); 14 remapped pgs
2023-02-13 04:00:00    osd: 82 osds: 82 up (since 2w), 82 in (since 13d); 12 remapped pgs

#
# objects misplaced increases when the remapped pgs increases
#
2023-02-06 00:00:00    pgs:     62898818/685798373 objects misplaced (9.172%)
2023-02-06 04:00:00    pgs:     61995946/686058837 objects misplaced (9.037%)
2023-02-06 06:00:00    pgs:     61090940/686122303 objects misplaced (8.904%)
2023-02-06 09:00:00    pgs:     59285766/686306445 objects misplaced (8.638%)
2023-02-06 12:30:00    pgs:     57626520/686387322 objects misplaced (8.396%)
2023-02-06 14:00:00    pgs:     56813753/686388870 objects misplaced (8.277%)
2023-02-06 14:30:00    pgs:     56578343/686389185 objects misplaced (8.243%)
2023-02-06 20:30:00    pgs:     54255700/686969448 objects misplaced (7.898%)
2023-02-07 00:00:00    pgs:     53089526/687266865 objects misplaced (7.725%)
2023-02-07 01:00:00    pgs:     52810518/687391707 objects misplaced (7.683%)
2023-02-07 11:00:00    pgs:     50169973/688319189 objects misplaced (7.289%)
2023-02-07 16:30:00    pgs:     49312642/689410730 objects misplaced (7.153%)
2023-02-07 22:30:00    pgs:     48244798/689771838 objects misplaced (6.994%)
2023-02-08 15:30:00    pgs:     38058106/691166947 objects misplaced (5.506%)
2023-02-08 21:30:00    pgs:     62851681/691618387 objects misplaced (9.088%)	<<< increase
2023-02-09 07:30:00    pgs:     57693559/693214621 objects misplaced (8.323%)
2023-02-09 13:03:27 ceph balancer off
2023-02-09 22:30:00    pgs:     48248474/694618590 objects misplaced (6.946%)
2023-02-10 14:00:00    pgs:     37872501/696136659 objects misplaced (5.440%)
2023-02-10 19:00:00    pgs:     64062339/696645992 objects misplaced (9.196%)	<<< increase
2023-02-11 04:30:00    pgs:     52183052/699438360 objects misplaced (7.461%)
2023-02-11 19:00:00    pgs:     49167191/699631520 objects misplaced (7.028%)
2023-02-12 12:00:00    pgs:     39485735/700248360 objects misplaced (5.639%)
2023-02-12 18:00:00    pgs:     63540276/700256808 objects misplaced (9.074%)	<<< increase
2023-02-12 20:30:00    pgs:     61766426/700291865 objects misplaced (8.820%)
2023-02-13 04:00:00    pgs:     57622457/700365805 objects misplaced (8.227%)


--- annotated log extract ---
##
## "misplaced" decreasing...
##
Feb 12 17:50:20 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:18.782191+0000 mgr.b4 (mgr.17297882) 4129637 : cluster [DBG] pgmap v4094135: 5945 pgs: 3 active+remapped+backfilling, 6 active+remapped+backfill_wait, 5936 active+clean; 88 TiB data, 185 TiB used, 243 TiB / 428 TiB avail; 1.5 MiB/s rd, 4.1 MiB/s wr, 178 op/s; 35014018/700256808 objects misplaced (5.000%); 34 MiB/s, 34 objects/s recovering
Feb 12 17:50:20 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:18.782191+0000 mgr.b4 (mgr.17297882) 4129637 : cluster [DBG] pgmap v4094135: 5945 pgs: 3 active+remapped+backfilling, 6 active+remapped+backfill_wait, 5936 active+clean; 88 TiB data, 185 TiB used, 243 TiB / 428 TiB avail; 1.5 MiB/s rd, 4.1 MiB/s wr, 178 op/s; 35014018/700256808 objects misplaced (5.000%); 34 MiB/s, 34 objects/s recovering
Feb 12 17:50:27 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:26.833542+0000 mgr.b4 (mgr.17297882) 4129641 : cluster [DBG] pgmap v4094141: 5945 pgs: 3 active+remapped+backfilling, 6 active+remapped+backfill_wait, 5936 active+clean; 88 TiB data, 185 TiB used, 243 TiB / 428 TiB avail; 2.7 MiB/s rd, 4.9 MiB/s wr, 226 op/s; 35012325/700256808 objects misplaced (5.000%); 37 MiB/s, 36 objects/s recovering
Feb 12 17:50:30 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:28.848020+0000 mgr.b4 (mgr.17297882) 4129642 : cluster [DBG] pgmap v4094143: 5945 pgs: 3 active+remapped+backfilling, 6 active+remapped+backfill_wait, 5936 active+clean; 88 TiB data, 185 TiB used, 243 TiB / 428 TiB avail; 2.0 MiB/s rd, 4.9 MiB/s wr, 436 op/s; 35011830/700256808 objects misplaced (5.000%); 36 MiB/s, 35 objects/s recovering
##
## shortly after, "misplaced" has increased, followed by the "starting backfill" lines
## I'm not sure if the line ordering here is a quirk of journalctl, maybe the "starting backfill" came first which would make some sense
##
Feb 12 17:50:32 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:30.860404+0000 mgr.b4 (mgr.17297882) 4129643 : cluster [DBG] pgmap v4094145: 5945 pgs: 3 active+remapped+backfilling, 8 active+remapped+backfill_wait, 5934 active+clean; 88 TiB data, 185 TiB used, 243 TiB / 428 TiB avail; 419 KiB/s rd, 1.2 MiB/s wr, 388 op/s; 44579289/700256808 objects misplaced (6.366%); 31 MiB/s, 30 objects/s recovering
Feb 12 17:50:32 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:31.031822+0000 osd.64 (osd.64) 362 : cluster [DBG] 33.2cs0 starting backfill to osd.25(4) from (0'0,0'0] MAX to 639282'24881196
Feb 12 17:50:32 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:31.031951+0000 osd.64 (osd.64) 363 : cluster [DBG] 33.6cs0 starting backfill to osd.25(4) from (0'0,0'0] MAX to 639282'24593871
Feb 12 17:50:32 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:31.044285+0000 osd.64 (osd.64) 364 : cluster [DBG] 33.6cs0 starting backfill to osd.51(5) from (0'0,0'0] MAX to 639282'24593871
Feb 12 17:50:32 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:31.044655+0000 osd.64 (osd.64) 365 : cluster [DBG] 33.2cs0 starting backfill to osd.51(5) from (0'0,0'0] MAX to 639282'24881196
Feb 12 17:50:32 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:31.057888+0000 osd.64 (osd.64) 366 : cluster [DBG] 33.6cs0 starting backfill to osd.63(7) from (0'0,0'0] MAX to 639282'24593871
Feb 12 17:50:32 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:31.058039+0000 osd.64 (osd.64) 367 : cluster [DBG] 33.2cs0 starting backfill to osd.63(7) from (0'0,0'0] MAX to 639282'24881196
Feb 12 17:50:32 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:31.077517+0000 osd.64 (osd.64) 368 : cluster [DBG] 33.6cs0 starting backfill to osd.68(2) from (0'0,0'0] MAX to 639282'24593871
  ## (more "starting backfill" lines elided here)
##
## "misplaced" continues to increase rapidly to 9%...
##
Feb 12 17:50:34 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:32.871242+0000 mgr.b4 (mgr.17297882) 4129644 : cluster [DBG] pgmap v4094148: 5945 pgs: 2 activating+remapped, 3 active+remapped+backfilling, 8 active+remapped+backfill_wait, 5932 active+clean; 88 TiB data, 185 TiB used, 243 TiB / 428 TiB avail; 424 KiB/s rd, 1.2 MiB/s wr, 440 op/s; 54133989/700256808 objects misplaced (7.731%); 31 MiB/s, 31 objects/s recovering
Feb 12 17:50:36 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:34.886503+0000 mgr.b4 (mgr.17297882) 4129645 : cluster [DBG] pgmap v4094149: 5945 pgs: 2 activating+remapped, 3 active+remapped+backfilling, 8 active+remapped+backfill_wait, 5932 active+clean; 88 TiB data, 185 TiB used, 243 TiB / 428 TiB avail; 442 KiB/s rd, 1.1 MiB/s wr, 483 op/s; 54133670/700256808 objects misplaced (7.731%); 41 MiB/s, 41 objects/s recovering
Feb 12 17:50:37 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:36.901531+0000 mgr.b4 (mgr.17297882) 4129646 : cluster [DBG] pgmap v4094150: 5945 pgs: 4 activating+remapped, 3 active+remapped+backfilling, 8 active+remapped+backfill_wait, 5930 active+clean; 88 TiB data, 185 TiB used, 243 TiB / 428 TiB avail; 281 KiB/s rd, 902 KiB/s wr, 324 op/s; 63666090/700256808 objects misplaced (9.092%); 17 MiB/s, 17 objects/s recovering
##
## then starts to fall again...
##
Feb 12 17:50:42 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:40.930869+0000 mgr.b4 (mgr.17297882) 4129648 : cluster [DBG] pgmap v4094152: 5945 pgs: 4 activating+remapped, 3 active+remapped+backfilling, 8 active+remapped+backfill_wait, 5930 active+clean; 88 TiB data, 185 TiB used, 243 TiB / 428 TiB avail; 407 KiB/s rd, 1.4 MiB/s wr, 479 op/s; 63665652/700256808 objects misplaced (9.092%); 28 MiB/s, 28 objects/s recovering
Feb 12 17:50:43 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:42.948421+0000 mgr.b4 (mgr.17297882) 4129649 : cluster [DBG] pgmap v4094153: 5945 pgs: 3 active+remapped+backfilling, 12 active+remapped+backfill_wait, 5930 active+clean; 88 TiB data, 185 TiB used, 243 TiB / 428 TiB avail; 455 KiB/s rd, 1.5 MiB/s wr, 542 op/s; 63665652/700256808 objects misplaced (9.092%); 27 MiB/s, 26 objects/s recovering
Feb 12 17:50:46 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:44.961474+0000 mgr.b4 (mgr.17297882) 4129650 : cluster [DBG] pgmap v4094154: 5945 pgs: 3 active+remapped+backfilling, 12 active+remapped+backfill_wait, 5930 active+clean; 88 TiB data, 185 TiB used, 243 TiB / 428 TiB avail; 436 KiB/s rd, 1.4 MiB/s wr, 518 op/s; 63664579/700256808 objects misplaced (9.092%); 37 MiB/s, 37 objects/s recovering
Feb 12 17:50:50 b2 ceph-c6618970-0ce0-4cb2-bc9a-dd5f29b62e24-mon-b2[1613051]: cluster 2023-02-12T06:50:48.991308+0000 mgr.b4 (mgr.17297882) 4129652 : cluster [DBG] pgmap v4094156: 5945 pgs: 3 active+remapped+backfilling, 12 active+remapped+backfill_wait, 5930 active+clean; 88 TiB data, 185 TiB used, 243 TiB / 428 TiB avail; 513 KiB/s rd, 1.6 MiB/s wr, 534 op/s; 63664502/700256808 objects misplaced (9.092%); 33 MiB/s, 33 objects/s recovering
------------------------------------------------------------------------------
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux