On 2020/9/30 11:30, Marc Smith wrote: > On Tue, Sep 29, 2020 at 9:37 PM Coly Li <colyli@xxxxxxx> wrote: >> >> On 2020/9/30 03:18, Marc Smith wrote: >>> Hi Coly, >>> >>> So nearly one year later, and I haven't hit this issue since running >>> 5.4.x until now. It seems to happen on a system for just ONE of the >>> backing devices (4 backing devices per cache device). Running Linux >>> 5.4.45 (vanilla kernel.org): >>> >>> --snip-- >>> [ 1597.707235] bcache: bch_journal_replay() journal replay done, 0 >>> keys in 1 entries, seq 1412486 >>> [ 1597.712131] bcache: bch_cached_dev_attach() Caching md123 as >>> bcache3 on set 521f7664-690d-4680-9b0d-1299fcee4321 >>> [ 1597.715109] bcache: bch_cached_dev_attach() Caching md124 as >>> bcache2 on set 521f7664-690d-4680-9b0d-1299fcee4321 >>> [ 1597.718674] bcache: bch_cached_dev_attach() Caching md125 as >>> bcache1 on set 521f7664-690d-4680-9b0d-1299fcee4321 >>> [ 1597.723378] ------------[ cut here ]------------ >>> [ 1597.723381] kernel BUG at drivers/md/bcache/writeback.c:562! >>> [ 1597.723389] invalid opcode: 0000 [#1] SMP NOPTI >>> [ 1597.723514] CPU: 0 PID: 2356 Comm: bcache_writebac Kdump: loaded >>> Tainted: G OE 5.4.45-esos.prod #1 >>> [ 1597.723746] Hardware name: Red Hat KVM, BIOS 0.5.1 01/01/2011 >>> [ 1597.723877] RIP: 0010:dirty_pred+0x17/0x21 [bcache] >>> [ 1597.723991] Code: 09 89 43 68 5b 5d 41 5c e9 e4 fd ff ff f0 48 0f >>> b3 3e c3 48 8b 06 8b 8f 38 f4 ff ff 48 89 c2 81 e2 ff ff 0f 00 48 39 >>> d1 74 02 <0f> 0b 48 c1 e8 24 83 e0 01 c3 41 56 41 55 41 54 55 48 89 fd >>> 48 83 >>> [ 1597.724426] RSP: 0018:ffffc900007f7bf0 EFLAGS: 00010297 >>> [ 1597.724543] RAX: 9000000001000002 RBX: ffff88855f800c50 RCX: 0000000000000001 >>> [ 1597.724699] RDX: 0000000000000002 RSI: ffff8885125a4f08 RDI: ffff88855f800c50 >>> [ 1597.724857] RBP: ffff8885125a4f08 R08: 0000000000000001 R09: 0000000000000001 >>> [ 1597.725015] R10: 9000000001000002 R11: 0000000000000001 R12: ffffc900007f7dd0 >>> [ 1597.725175] R13: ffff8885a8797400 R14: ffff8885a87974c8 R15: 0000000000000000 >>> [ 1597.725309] FS: 0000000000000000(0000) GS:ffff888627a00000(0000) >>> knlGS:0000000000000000 >>> [ 1597.725428] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 >>> [ 1597.725517] CR2: 0000000000ce9d00 CR3: 000000059f304000 CR4: 0000000000340ef0 >>> [ 1597.725623] Call Trace: >>> [ 1597.725672] refill_keybuf_fn+0x66/0x1a3 [bcache] >>> [ 1597.725743] ? mca_find+0x50/0x50 [bcache] >>> [ 1597.726310] bch_btree_map_keys_recurse+0x79/0x14c [bcache] >>> [ 1597.726433] bcache: bch_cached_dev_attach() Caching md126 as >>> bcache0 on set 521f7664-690d-4680-9b0d-1299fcee4321 >>> [ 1597.726880] ? __switch_to_asm+0x34/0x70 >>> [ 1597.728612] bcache: register_cache() registered cache device dm-0 >>> [ 1597.729128] ? bch_btree_node_get+0xce/0x1bd [bcache] >>> [ 1597.730553] ? mca_find+0x50/0x50 [bcache] >>> [ 1597.731102] bch_btree_map_keys_recurse+0xc1/0x14c [bcache] >>> [ 1597.731672] ? __switch_to_asm+0x40/0x70 >>> [ 1597.732219] ? __switch_to_asm+0x34/0x70 >>> [ 1597.732770] ? __schedule+0x492/0x4b5 >>> [ 1597.733319] ? rwsem_optimistic_spin+0x186/0x1ae >>> [ 1597.733873] ? mca_find+0x50/0x50 [bcache] >>> [ 1597.734431] bch_btree_map_keys+0x87/0xd5 [bcache] >>> [ 1597.734987] ? clear_bit+0x6/0x6 [bcache] >>> [ 1597.735537] bch_refill_keybuf+0x81/0x1ae [bcache] >>> [ 1597.736091] ? remove_wait_queue+0x41/0x41 >>> [ 1597.736645] ? clear_bit+0x6/0x6 [bcache] >>> [ 1597.737189] bch_writeback_thread+0x35e/0x507 [bcache] >>> [ 1597.737755] ? read_dirty+0x448/0x448 [bcache] >>> [ 1597.738311] kthread+0xe4/0xe9 >>> [ 1597.738843] ? kthread_flush_worker+0x70/0x70 >>> [ 1597.739405] ret_from_fork+0x22/0x40 >>> [ 1597.739951] Modules linked in: nvmet_fc(O) nvmet_rdma(O) nvmet(O) >>> bcache qla2xxx_scst(OE) nvme_fc(O) nvme_fabrics(O) bonding >>> cls_switchtec(O) qede qed mlx5_core(O) mlxfw(O) bna rdmavt(O) >>> ib_umad(O) rdma_ucm(O) ib_uverbs(O) ib_srp(O) rdma_cm(O) ib_cm(O) >>> iw_cm(O) iw_cxgb4(O) iw_cxgb3(O) ib_qib(O) mlx4_ib(O) mlx4_core(O) >>> ib_core(O) ib_mthca(O) nvme(O) nvme_core(O) mlx_compat(O) >>> [ 1597.743363] ---[ end trace 095defe1dc682fed ]--- >>> [ 1597.743983] RIP: 0010:dirty_pred+0x17/0x21 [bcache] >>> [ 1597.746015] Code: 09 89 43 68 5b 5d 41 5c e9 e4 fd ff ff f0 48 0f >>> b3 3e c3 48 8b 06 8b 8f 38 f4 ff ff 48 89 c2 81 e2 ff ff 0f 00 48 39 >>> d1 74 02 <0f> 0b 48 c1 e8 24 83 e0 01 c3 41 56 41 55 41 54 55 48 89 fd >>> 48 83 >>> [ 1597.747806] RSP: 0018:ffffc900007f7bf0 EFLAGS: 00010297 >>> [ 1597.748449] RAX: 9000000001000002 RBX: ffff88855f800c50 RCX: 0000000000000001 >>> [ 1597.749097] RDX: 0000000000000002 RSI: ffff8885125a4f08 RDI: ffff88855f800c50 >>> [ 1597.749769] RBP: ffff8885125a4f08 R08: 0000000000000001 R09: 0000000000000001 >>> [ 1597.750427] R10: 9000000001000002 R11: 0000000000000001 R12: ffffc900007f7dd0 >>> [ 1597.751073] R13: ffff8885a8797400 R14: ffff8885a87974c8 R15: 0000000000000000 >>> [ 1597.751745] FS: 0000000000000000(0000) GS:ffff888627a00000(0000) >>> knlGS:0000000000000000 >>> [ 1597.752900] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 >>> [ 1597.753544] CR2: 0000000000ce9d00 CR3: 000000059f304000 CR4: 0000000000340ef0 >>> --snip-- >>> >>> This line at the beginning of that snippet above catches my eye (when >>> registering the cache device): >>> [ 1597.707235] bcache: bch_journal_replay() journal replay done, 0 >>> keys in 1 entries, seq 1412486 >>> >>> This machine was in a crashed state, and then upon resetting it, I hit >>> the issue above when registering the backing/cache devices. Is the >>> cache/journal perhaps corrupted? Any way to manually >>> intervene/resolve? >>> >>> Thanks for your time. >> >> Hi Marc, >> >> Thanks for the reporting. Maybe this is a hidden issue in btree code >> (just a quick guess without any evidence). >> >> For the journal information you mentioned, >> "[ 1597.707235] bcache: bch_journal_replay() journal replay done, 0 >> keys in 1 entries, seq 1412486" >> Because a journal_meta() operation may only flush a journal set block >> with cache set meta data without any extra btree key, the message itself >> is legal. I am not able to get suspicious clue from it. > > Okay, thank you for the explanation. I took a peek at > bch_journal_replay() and understand now. > > >> >> Maybe the new kernel just improves to make less mistake for the hidden >> bug, but the root cause is not identified yet. >> >> BTW, there are 2 things in my brain, >> 1) Do you have this patch in your kernel, >> commit be23e837333a ("bcache: fix potential deadlock problem in >> btree_gc_coalesce") > > No, looks like this fix came with 5.4.49 (I'm using .45 currently). > Will update ASAP. > > >> 2) The journal and btree flush issue was not totally fixed in v5.4. If I >> remember correctly the (currently) final fixes went into mainline kernel >> in v5.6. Could you please to try v5.6 stable tree see whether the >> situation might be better. > > I can't move to or really test all that well 5.6.x at this moment. > However, if you know of a specific patch or set of patches you could > point me at, I could look at back porting these to 5.4.x for my use. > Here are the following btree flush fixes for your reference, commit 34cf78bf34d4 ("bcache: fix a lost wake-up problem caused by mca_cannibalize_lock") commit 84c529aea182 ("bcache: fix deadlock in bcache_allocator") commit 9fcc34b1a6dd ("bcache: at least try to shrink 1 node in bch_mca_scan()") commit 2aa8c529387c ("bcache: avoid unnecessary btree nodes flushing in btree_flush_write()") commit 125d98edd114 ("bcache: remove member accessed from struct btree") commit d5c9c470b011 ("bcache: reap c->btree_cache_freeable from the tail in bch_mca_scan()") commit 5bebf7486d4f ("bcache: fix memory corruption in bch_cache_accounting_clear()") Thanks. Coly Li