Re: _delete_some new onodes has appeared since PG removal started

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Yes, with the fixes in 14.2.19 PG removal is really much much much
better than before.

But on some clusters (in particular with rocksdb on the hdd) there is
still a rare osd flap at the end of the PG removal -- indicated by the
logs I shared earlier.
Our workaround to prevent that new flap is to increase
osd_heartbeat_grace (e.g. to 45).

With 3.5M objects in a PG, I suggest that you try moving one PG with
upmap and watch how it goes (especially at the end).

-- dan

On Wed, Apr 21, 2021 at 4:01 PM Konstantin Shalygin <k0ste@xxxxxxxx> wrote:
>
> Dan, you about this issue [1] ?
>
> I was start to backfill new OSD's on 14.2.19: pool have 2048PG with 7.14G objects...
> avg number of PG is 3486328 objects
>
>
>
> [1] https://tracker.ceph.com/issues/47044
>
> k
>
>
> On 21 Apr 2021, at 16:37, Dan van der Ster <dan@xxxxxxxxxxxxxx> wrote:
>
> Do we have a tracker for this?
>
> We should ideally be able to remove that final collection_list from
> the optimized pg removal.
> It can take a really long time and lead to osd flapping:
>
> 2021-04-21 15:23:37.003 7f51c273c700  1 heartbeat_map is_healthy
> 'OSD::osd_op_tp thread 0x7f51a3e81700' had timed out after 15
> 2021-04-21 15:23:41.595 7f51a3e81700  0
> bluestore(/var/lib/ceph/osd/ceph-941) log_latency_fn slow operation
> observed for _collection_list, latency = 67.7234s, lat = 67s cid
> =10.14aes4_head start GHMAX end GHMAX max 30
> 2021-04-21 15:23:41.595 7f51a3e81700  0 osd.941 pg_epoch: 43004
> pg[10.14aes4( v 42754'296580 (40999'293500,42754'296580] lb MIN
> (bitwise) local-lis/les=41331/41332 n=159058 ec=4951/4937 lis/c
> 41331/41331 les/c/f 41332/42758/0 41330/42759/33461)
> [171,903,106,27,395,773]p171(0) r=-1 lpr=42759 DELETING
> pi=[41331,42759)/1 crt=42754'296580 unknown NOTIFY mbc={}]
> _delete_some additional unexpected onode list (new onodes has appeared
> since PG removal started[4#10:75280000::::head#]
> 2021-04-21 15:23:50.061 7f51a3e81700  0
> bluestore(/var/lib/ceph/osd/ceph-941) log_latency slow operation
> observed for submit_transact, latency = 8.46584s
> 2021-04-21 15:23:50.062 7f51a3e81700  1 heartbeat_map reset_timeout
> 'OSD::osd_op_tp thread 0x7f51a3e81700' had timed out after 15
> 2021-04-21 15:23:50.115 7f51b6ca1700  0
> bluestore(/var/lib/ceph/osd/ceph-941) log_latency_fn slow operation
> observed for _txc_committed_kv, latency = 8.51916s, txc =
> 0x5573928a7340
> 2021-04-21 15:23:50.473 7f51b2498700  0 log_channel(cluster) log [WRN]
> : Monitor daemon marked osd.941 down, but it is still running
>
> -- dan
>
> On Thu, Apr 15, 2021 at 10:32 AM Dan van der Ster <dan@xxxxxxxxxxxxxx> wrote:
>
>
> Thanks Igor and Neha for the quick responses.
>
> I posted an osd log with debug_osd 10 and debug_bluestore 20:
> ceph-post-file: 09094430-abdb-4248-812c-47b7babae06c
>
> Hope that helps,
>
> Dan
>
> On Thu, Apr 15, 2021 at 1:27 AM Neha Ojha <nojha@xxxxxxxxxx> wrote:
>
>
> We saw this warning once in testing
> (https://tracker.ceph.com/issues/49900#note-1), but there, the problem
> was different, which also led to a crash. That issue has been fixed
> but if you can provide osd logs with verbose logging, we might be able
> to investigate further.
>
> Neha
>
> On Wed, Apr 14, 2021 at 4:14 PM Igor Fedotov <ifedotov@xxxxxxx> wrote:
>
>
> Hi Dan,
>
> Seen that once before and haven't thoroughly investigated yet but I
> think the new PG removal stuff just revealed this "issue". In fact it
> had been in the code before the patch.
>
> The warning means that new object(s) (given the object names these are
> apparently system objects, don't remember what's this exactly)  has been
> written to a PG after it was staged for removal.
>
> New PG removal properly handles that case - that was just a paranoid
> check for an unexpected situation which has actually triggered. Hence
> IMO no need to worry at this point but developers might want to validate
> why this is happening....
>
>
> Thanks,
>
> Igor
>
> On 4/14/2021 10:26 PM, Dan van der Ster wrote:
>
> Hi Igor,
>
> After updating to 14.2.19 and then moving some PGs around we have a
> few warnings related to the new efficient PG removal code, e.g. [1].
> Is that something to worry about?
>
> Best Regards,
>
> Dan
>
> [1]
>
> /var/log/ceph/ceph-osd.792.log:2021-04-14 20:34:34.353 7fb2439d4700  0
> osd.792 pg_epoch: 40906 pg[10.14b2s0( v 40734'290069
> (33782'287000,40734'290069] lb MIN (bitwise) local-lis/les=33990/33991
> n=36272 ec=4951/4937 lis/c 33990/33716 les/c/f 33991/33747/0
> 40813/40813/37166) [933,626,260,804,503,491]p933(0) r=-1 lpr=40813
> DELETING pi=[33716,40813)/4 crt=40734'290069 unknown NOTIFY mbc={}]
> _delete_some additional unexpected onode list (new onodes has appeared
> since PG removal started[0#10:4d280000::::head#]
>
> /var/log/ceph/ceph-osd.851.log:2021-04-14 18:40:13.312 7fd87bded700  0
> osd.851 pg_epoch: 40671 pg[10.133fs5( v 40662'288967
> (33782'285900,40662'288967] lb MIN (bitwise) local-lis/les=33786/33787
> n=13 ec=4947/4937 lis/c 40498/33714 les/c/f 40499/33747/0
> 40670/40670/33432) [859,199,913,329,439,79]p859(0) r=-1 lpr=40670
> DELETING pi=[33714,40670)/4 crt=40662'288967 unknown NOTIFY mbc={}]
> _delete_some additional unexpected onode list (new onodes has appeared
> since PG removal started[5#10:fcc80000::::head#]
>
> /var/log/ceph/ceph-osd.851.log:2021-04-14 20:58:14.393 7fd87adeb700  0
> osd.851 pg_epoch: 40906 pg[10.2e8s3( v 40610'288991
> (33782'285900,40610'288991] lb MIN (bitwise) local-lis/les=33786/33787
> n=161220 ec=4937/4937 lis/c 39826/33716 les/c/f 39827/33747/0
> 40617/40617/39225) [717,933,727,792,607,129]p717(0) r=-1 lpr=40617
> DELETING pi=[33716,40617)/3 crt=40610'288991 unknown NOTIFY mbc={}]
> _delete_some additional unexpected onode list (new onodes has appeared
> since PG removal started[3#10:17400000::::head#]
>
> /var/log/ceph/ceph-osd.883.log:2021-04-14 18:55:16.822 7f78c485d700  0
> osd.883 pg_epoch: 40857 pg[7.d4( v 40804'9911289
> (35835'9908201,40804'9911289] lb MIN (bitwise)
> local-lis/les=40782/40783 n=195 ec=2063/1989 lis/c 40782/40782 les/c/f
> 40783/40844/0 40781/40845/40845) [877,870,894] r=-1 lpr=40845 DELETING
> pi=[40782,40845)/1 crt=40804'9911289 lcod 40804'9911288 unknown NOTIFY
> mbc={}] _delete_some additional unexpected onode list (new onodes has
> appeared since PG removal started[#7:2b000000::::head#]
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux