Thank you for your reply. I am now running kernel 3.15.4-200.fc20.x86_64 and the problem still exists. If the bad record could be found could it be repaired manually? The stuck backing devices total 17TB and I don't have that much spare capacity to relocate the data so I am motivated to try to repair in-place rather than wipe and rebuild. If you have any tips for where to look in the source code for the on-disk record format for the cache store and what to look for to identify the bad record(s) I would greatly appreciate it. Memory on this box is only lightly used. Of the 16GB, there is never less than 10GB free (free + buffers + cache). Here's a stack dump of the bcache_writebac thread: task: ffff880408ce4f00 ti: ffff8803f4eb4000 task.ti: ffff8803f4eb4000 RIP: 0010:[<ffffffffa0005c11>] [<ffffffffa0005c11>] refill_keybuf_fn+0x61/0x1d0 [bcache] RSP: 0018:ffff8803f4eb7b78 EFLAGS: 00000246 RAX: 0000000000000001 RBX: ffff8803fdc13000 RCX: ffff8803f4eb7e40 RDX: 0000000023ad5e98 RSI: ffff8802820a41b8 RDI: ffff8803ff830bc8 RBP: ffff8803f4eb7b78 R08: ffff880402520000 R09: 0000000000000001 R10: ffff88040fbdc000 R11: 000007ffffffffff R12: ffff8803f4eb7d70 R13: 0000000000000000 R14: ffff8803f4eb7d00 R15: ffff8803fdc13000 FS: 00007fecc4d59700(0000) GS:ffff880427c00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 00007fa1adad3000 CR3: 00000003f17dc000 CR4: 00000000000007f0 Stack: ffff8803f4eb7c30 ffffffffa0008292 00000001d06df860 ffffffffa0005bb0 ffff8803fdc130c8 ffff880131682990 ffff88040cc0e8c8 0000000000000004 0000000000000001 ffff8802820a41d0 ffff8802820cc118 ffff880402520000 Call Trace: [<ffffffffa0008292>] bch_btree_map_keys_recurse+0x72/0x1d0 [bcache] [<ffffffffa0005bb0>] ? btree_insert_key+0xe0/0xe0 [bcache] [<ffffffffa0007ed5>] ? bch_btree_node_get+0xd5/0x290 [bcache] [<ffffffffa0008327>] bch_btree_map_keys_recurse+0x107/0x1d0 [bcache] [<ffffffffa0005bb0>] ? btree_insert_key+0xe0/0xe0 [bcache] [<ffffffffa000b3f7>] bch_btree_map_keys+0x127/0x150 [bcache] [<ffffffffa0005bb0>] ? btree_insert_key+0xe0/0xe0 [bcache] [<ffffffffa00209b0>] ? bch_crc64+0x50/0x50 [bcache] [<ffffffffa000b4d4>] bch_refill_keybuf+0xb4/0x220 [bcache] [<ffffffff810d0350>] ? abort_exclusive_wait+0xb0/0xb0 [<ffffffffa00209b0>] ? bch_crc64+0x50/0x50 [bcache] [<ffffffffa0021204>] bch_writeback_thread+0x154/0x840 [bcache] [<ffffffffa00210b0>] ? write_dirty_finish+0x260/0x260 [bcache] [<ffffffff810ac538>] kthread+0xd8/0xf0 [<ffffffff810ac460>] ? insert_kthread_work+0x40/0x40 [<ffffffff816ff23c>] ret_from_fork+0x7c/0xb0 [<ffffffff810ac460>] ? insert_kthread_work+0x40/0x40 When this snapshot was taken both bcache_writebac processes had similar stack traces but one was writing data and the other was not. The one that was writing was only writing at 3-4 4K blocks per second. I tried setting writeback_rate to a very large value and writeback_percent to zero but neither changed the rate. The backing device that is now writing does not write consistently. It will be idle for many hours then spontaneously start writing at ~4 4K blocks per second for several hours then stop again. --Larkin On 7/16/2014 1:27 AM, Samuel Huo wrote: > It seems your original 3.14 kernel hit the "btree split failed" bug > when bcache wrote back some dirty data into backing device and > inserted keys to invalidate those dirty data in btree index. But I > think that shouldn't prevent future writeback. > The "btree split failed" problem gets fixed in 3.15, as Slava > suggested in another post. > > You said writeback takes near 100% CPU, can you try to find out where > the writeback thread gets stuck at? and how is your memory usage? > > -Jianjian > > On Mon, Jul 14, 2014 at 5:47 AM, Larkin Lowrey > <llowrey@xxxxxxxxxxxxxxxxx> wrote: >> I am still unable to detach either backing device. There is still dirty >> data and it's not being written out. The bcache_writebac processes are >> at max CPU and the backing devices are idle. It appears to me that >> something is corrupt in the cache device which prevents the full dirty >> data from being written. >> >> Is there anything useful I can do to debug? >> >> Several GB of dirty data had accumulated since the data reported below. >> I set writeback_percent to zero in order to flush the dirty data and >> most did but not all. In the case of md1 it had grown to 6GB but only >> dropped to 2GB (it had been as low as 1.1GB a few days ago). The md3 >> device had 1.5GB but only dropped to 696KB (the same as the minimum from >> a few days ago). I have switch to writethrough mode in hopes that I >> don't get into further trouble. >> >> Attempting to detach (via /sys/block/md?/bcache/detach) has had no effect. >> >> Any help would be much appreciated! >> >> --Larkin >> >> On 7/8/2014 12:18 PM, Larkin Lowrey wrote: >>> %CPU %MEM TIME+ COMMAND >>> 98.8 0.0 645:05.50 bcache_writebac >>> 97.2 0.0 600:50.18 bcache_writebac >>> >>> The machine in question is a backup server. Backups usually take >>> 40-60min but this morning took 8.5-9 hours. I did make several changes >>> yesterday. >>> >>> There is a single (md raid 1) cache device and I had 2 md raid6 arrays >>> attached. Yesterday I attached a third raid6. >>> >>> Since noticing the high CPU usage (and poor IO performance) I attempted >>> to detach each of the three backing devices with the idea that I would >>> rebuild the cache set. One of the backing devices did detach but two >>> have not. One of the two remaining devices has 1GB of dirty data and the >>> other has 696KB. >>> >>> I had been running kernel 3.14.9-200.fc20.x86_64 but once this high load >>> situation happened I switched to 3.15.3-200.fc20.x86_64. >>> >>> When I reboot the two bcache_writebac processes start out immediately at >>> high load. >>> >>> /sys/block/md[13]/bcache/writeback_percent start out at 10%. >>> >>> There is no IO other than to the root fs which is on a separate raid1. >>> >>> ==> /sys/block/md1/bcache/writeback_rate_debug <== >>> rate: 512/sec >>> dirty: 1.1G >>> target: 15.5G >>> proportional: -13.7M >>> derivative: 0 >>> change: -13.7M/sec >>> next io: -62ms >>> >>> ==> /sys/block/md3/bcache/writeback_rate_debug <== >>> rate: 512/sec >>> dirty: 696k >>> target: 4.4G >>> proportional: -4.2M >>> derivative: 0 >>> change: -4.2M/sec >>> next io: 0ms >>> >>> After switching writeback_percent to 0%, there is still no IO and... >>> >>> ==> /sys/block/md1/bcache/writeback_rate_debug <== >>> rate: 512/sec >>> dirty: 1.1G >>> target: 15.5G >>> proportional: -13.7M >>> derivative: 0 >>> change: -13.7M/sec >>> next io: -50ms >>> >>> ==> /sys/block/md3/bcache/writeback_rate_debug <== >>> rate: 512/sec >>> dirty: 696k >>> target: 4.4G >>> proportional: -4.2M >>> derivative: 0 >>> change: -4.2M/sec >>> next io: 0ms >>> >>> ... and switching back to 10%, still no IO and... >>> >>> ==> /sys/block/md1/bcache/writeback_rate_debug <== >>> rate: 512/sec >>> dirty: 1.1G >>> target: 15.5G >>> proportional: -13.7M >>> derivative: 0 >>> change: -13.7M/sec >>> next io: -67ms >>> >>> ==> /sys/block/md3/bcache/writeback_rate_debug <== >>> rate: 512/sec >>> dirty: 696k >>> target: 4.4G >>> proportional: -4.2M >>> derivative: 0 >>> change: -4.2M/sec >>> next io: 0ms >>> >>> The only log messages for bcache are: >>> >>> [ 23.728302] bcache: register_bdev() registered backing device md2 >>> [ 23.750124] bcache: register_bdev() registered backing device md1 >>> [ 23.776249] bcache: register_bdev() registered backing device md3 >>> [ 25.414604] bcache: bch_journal_replay() journal replay done, 66 keys >>> in 10 entries, seq 26249932 >>> [ 25.534038] bcache: bch_cached_dev_attach() Caching md3 as bcache2 on >>> set 66c1a39b-5898-4aae-abe4-6ab609c18155 >>> [ 25.682785] bcache: bch_cached_dev_attach() Caching md1 as bcache1 on >>> set 66c1a39b-5898-4aae-abe4-6ab609c18155 >>> [ 25.695961] bcache: register_cache() registered cache device dm-2 >>> >>> Status (shortly after reboot): >>> >>> # bcache-status -s >>> --- bcache --- >>> UUID 66c1a39b-5898-4aae-abe4-6ab609c18155 >>> Block Size 4.00 KiB >>> Bucket Size 2.0 MiB >>> Congested? False >>> Read Congestion 2.0ms >>> Write Congestion 20.0ms >>> Total Cache Size 200 GiB >>> Total Cache Used 148 GiB (74%) >>> Total Cache Unused 52 GiB (26%) >>> Evictable Cache 200 GiB (100%) >>> Replacement Policy [lru] fifo random >>> Cache Mode writethrough [writeback] writearound none >>> Total Hits 6174 (94%) >>> Total Misses 380 >>> Total Bypass Hits 0 >>> Total Bypass Misses 0 >>> Total Bypassed 0 B >>> --- Backing Device --- >>> Device File /dev/md1 (9:1) >>> bcache Device File /dev/bcache1 (252:1) >>> Size 13 TiB >>> Cache Mode writethrough [writeback] writearound none >>> Readahead 0 >>> Sequential Cutoff 16.0 MiB >>> Merge sequential? False >>> State dirty >>> Writeback? True >>> Dirty Data 1 GiB >>> Total Hits 6108 (99%) >>> Total Misses 12 >>> Total Bypass Hits 0 >>> Total Bypass Misses 0 >>> Total Bypassed 0 B >>> --- Backing Device --- >>> Device File /dev/md3 (9:3) >>> bcache Device File /dev/bcache2 (252:2) >>> Size 4 TiB >>> Cache Mode writethrough [writeback] writearound none >>> Readahead 0 >>> Sequential Cutoff 16.0 MiB >>> Merge sequential? False >>> State dirty >>> Writeback? True >>> Dirty Data 696.00 KiB >>> Total Hits 66 (15%) >>> Total Misses 368 >>> Total Bypass Hits 0 >>> Total Bypass Misses 0 >>> Total Bypassed 0 B >>> --- Cache Device --- >>> Device File /dev/dm-2 (253:2) >>> Size 200 GiB >>> Block Size 4.00 KiB >>> Bucket Size 2.0 MiB >>> Replacement Policy [lru] fifo random >>> Discard? False >>> I/O Errors 0 >>> Metadata Written 2.0 MiB >>> Data Written 1.4 MiB >>> Buckets 102400 >>> Cache Used 148 GiB (74%) >>> Cache Unused 52 GiB (26%) >>> >>> I did find the following in the logs from a day prior to when I did the >>> work. The cacti graphs show high load at that time but a raid-check >>> started shortly before this which usually causes this kind of load. So, >>> the problem may have been occurring since the 6th without being noticing. >>> >>> Jul 6 00:26:32 mcp kernel: WARNING: CPU: 1 PID: 15568 at >>> drivers/md/bcache/btree.c:1979 btree_split+0x5bb/0x630 [bcache]() >>> Jul 6 00:26:32 mcp kernel: bcache: btree split failed >>> Jul 6 00:26:32 mcp kernel: Modules linked in: >>> Jul 6 00:26:32 mcp kernel: [30957.447979] bcache: btree split failed >>> Jul 6 00:26:32 mcp kernel: [30957.451654] Modules linked in: >>> binfmt_misc cfg80211 rfkill it87 hwmon_vid nf_conntrack_ipv4 ip6t_REJECT >>> nf_defrag_ipv4 nf_conntrack_ipv6 nf_defrag_ip >>> v6 xt_conntrack ip6table_filter nf_conntrack ip6_tables raid456 >>> async_raid6_recov async_memcpy async_pq async_xor async_tx kvm microcode >>> edac_core serio_raw sp5100_tco edac_mce_amd >>> k10temp snd_hda_codec_realtek i2c_piix4 snd_hda_codec_generic >>> snd_hda_codec_hdmi snd_hda_intel snd_hda_codec snd_hwdep snd_seq >>> snd_seq_device snd_pcm snd_timer snd r8169 soundcore >>> mii shpchp wmi acpi_cpufreq nfsd auth_rpcgss nfs_acl lockd sunrpc btrfs >>> xor raid6_pq raid1 radeon i2c_algo_bit firewire_ohci drm_kms_helper >>> firewire_core ttm crc_itu_t mvsas drm l >>> ibsas scsi_transport_sas i2c_core cpufreq_stats bcache >>> Jul 6 00:26:32 mcp kernel: [30957.528183] CPU: 1 PID: 15568 Comm: >>> kworker/1:2 Not tainted 3.14.8-200.fc20.x86_64 #1 >>> Jul 6 00:26:32 mcp kernel: [30957.537550] Hardware name: Gigabyte >>> Technology Co., Ltd. GA-890GPA-UD3H/GA-890GPA-UD3H, BIOS F9 09/09/2011 >>> Jul 6 00:26:32 mcp kernel: binfmt_misc cfg80211 rfkill it87 hwmon_vid >>> nf_conntrack_ipv4 ip6t_REJECT nf_defrag_ipv4 nf_conntrack_ipv6 >>> nf_defrag_ipv6 xt_conntrack ip6table_filter nf >>> _conntrack ip6_tables raid456 async_raid6_recov async_memcpy async_pq >>> async_xor async_tx kvm microcode edac_core serio_raw sp5100_tco >>> edac_mce_amd k10temp snd_hda_codec_realtek i2c >>> _piix4 snd_hda_codec_generic snd_hda_codec_hdmi snd_hda_intel >>> snd_hda_codec snd_hwdep snd_seq snd_seq_device snd_pcm snd_timer snd >>> r8169 soundcore mii shpchp wmi acpi_cpufreq nfsd >>> auth_rpcgss nfs_acl lockd sunrpc btrfs xor raid6_pq raid1 radeon >>> i2c_algo_bit firewire_ohci drm_kms_helper firewire_core ttm crc_itu_t >>> mvsas drm libsas scsi_transport_sas i2c_core >>> cpufreq_stats bcache >>> Jul 6 00:26:32 mcp kernel: [30957.548771] Workqueue: events >>> write_dirty_finish [bcache] >>> Jul 6 00:26:32 mcp kernel: CPU: 1 PID: 15568 Comm: kworker/1:2 Not >>> tainted 3.14.8-200.fc20.x86_64 #1 >>> Jul 6 00:26:32 mcp kernel: [30957.555773] 0000000000000000 >>> 00000000dea1a02f ffff880007ffd8b0 ffffffff816f0502 >>> Jul 6 00:26:32 mcp kernel: Hardware name: Gigabyte Technology Co., Ltd. >>> GA-890GPA-UD3H/GA-890GPA-UD3H, BIOS F9 09/09/2011 >>> Jul 6 00:26:32 mcp kernel: [30957.564855] ffff880007ffd8f8 >>> ffff880007ffd8e8 ffffffff8108a1cd ffff88040e6b7800 >>> Jul 6 00:26:32 mcp kernel: Workqueue: events write_dirty_finish [bcache] >>> Jul 6 00:26:32 mcp kernel: 0000000000000000 00000000dea1a02f >>> ffff880007ffd8b0 ffffffff816f0502 >>> Jul 6 00:26:32 mcp kernel: ffff880007ffd8f8 ffff880007ffd8e8 >>> ffffffff8108a1cd ffff88040e6b7800 >>> Jul 6 00:26:32 mcp kernel: [30957.573943] ffffffffffffffe4 >>> ffff880007ffdd58 ffff880007ffd980 0000000000000000 >>> Jul 6 00:26:32 mcp kernel: [30957.583045] Call Trace: >>> Jul 6 00:26:32 mcp kernel: [30957.587031] [<ffffffff816f0502>] >>> dump_stack+0x45/0x56 >>> Jul 6 00:26:32 mcp kernel: ffffffffffffffe4 ffff880007ffdd58 >>> ffff880007ffd980 0000000000000000 >>> Jul 6 00:26:32 mcp kernel: Call Trace: >>> Jul 6 00:26:32 mcp kernel: [<ffffffff816f0502>] dump_stack+0x45/0x56 >>> Jul 6 00:26:32 mcp kernel: [30957.593729] [<ffffffff8108a1cd>] >>> warn_slowpath_common+0x7d/0xa0 >>> Jul 6 00:26:32 mcp kernel: [<ffffffff8108a1cd>] >>> warn_slowpath_common+0x7d/0xa0 >>> Jul 6 00:26:32 mcp kernel: [<ffffffff8108a24c>] warn_slowpath_fmt+0x5c/0x80 >>> Jul 6 00:26:32 mcp kernel: [30957.601287] [<ffffffff8108a24c>] >>> warn_slowpath_fmt+0x5c/0x80 >>> Jul 6 00:26:32 mcp kernel: [30957.608568] [<ffffffffa000965b>] >>> btree_split+0x5bb/0x630 [bcache] >>> Jul 6 00:26:32 mcp kernel: [<ffffffffa000965b>] btree_split+0x5bb/0x630 >>> [bcache] >>> Jul 6 00:26:32 mcp kernel: [30957.616248] [<ffffffff816f3d22>] ? >>> __schedule+0x2e2/0x740 >>> Jul 6 00:26:32 mcp kernel: [<ffffffff816f3d22>] ? __schedule+0x2e2/0x740 >>> Jul 6 00:26:32 mcp kernel: [30957.623240] [<ffffffff8136decd>] ? >>> list_del+0xd/0x30 >>> Jul 6 00:26:32 mcp kernel: [<ffffffff8136decd>] ? list_del+0xd/0x30 >>> Jul 6 00:26:32 mcp kernel: [30957.629792] [<ffffffffa0009771>] >>> bch_btree_insert_node+0xa1/0x2c0 [bcache] >>> Jul 6 00:26:32 mcp kernel: [30957.638293] [<ffffffffa000a724>] >>> btree_insert_fn+0x24/0x50 [bcache] >>> Jul 6 00:26:32 mcp kernel: [<ffffffffa0009771>] >>> bch_btree_insert_node+0xa1/0x2c0 [bcache] >>> Jul 6 00:26:32 mcp kernel: [<ffffffffa000a724>] >>> btree_insert_fn+0x24/0x50 [bcache] >>> Jul 6 00:26:32 mcp kernel: [<ffffffffa0007861>] >>> bch_btree_map_nodes_recurse+0x51/0x180 [bcache] >>> Jul 6 00:26:32 mcp kernel: [30957.646175] [<ffffffffa0007861>] >>> bch_btree_map_nodes_recurse+0x51/0x180 [bcache] >>> Jul 6 00:26:32 mcp kernel: [30957.655171] [<ffffffffa000a700>] ? >>> bch_btree_insert_check_key+0x1b0/0x1b0 [bcache] >>> Jul 6 00:26:32 mcp kernel: [<ffffffffa000a700>] ? >>> bch_btree_insert_check_key+0x1b0/0x1b0 [bcache] >>> Jul 6 00:26:32 mcp kernel: [<ffffffffa000da62>] ? >>> bch_btree_ptr_invalid+0x12/0x20 [bcache] >>> Jul 6 00:26:32 mcp kernel: [30957.664290] [<ffffffffa000da62>] ? >>> bch_btree_ptr_invalid+0x12/0x20 [bcache] >>> Jul 6 00:26:32 mcp kernel: [30957.672814] [<ffffffffa000c48d>] ? >>> bch_btree_ptr_bad+0x4d/0x110 [bcache] >>> Jul 6 00:26:32 mcp kernel: [<ffffffffa000c48d>] ? >>> bch_btree_ptr_bad+0x4d/0x110 [bcache] >>> Jul 6 00:26:32 mcp kernel: [30957.681045] [<ffffffff816f6592>] ? >>> down_write+0x12/0x30 >>> Jul 6 00:26:32 mcp kernel: [<ffffffff816f6592>] ? down_write+0x12/0x30 >>> Jul 6 00:26:32 mcp kernel: [30957.731889] [<ffffffff810b2d66>] ? >>> srcu_notifier_call_chain+0x16/0x20 >>> Jul 6 00:26:32 mcp kernel: [<ffffffff810b2d66>] ? >>> srcu_notifier_call_chain+0x16/0x20 >>> Jul 6 00:26:32 mcp kernel: [<ffffffffa000b1e1>] >>> bch_btree_insert+0xf1/0x170 [bcache] >>> Jul 6 00:26:32 mcp kernel: [30957.739905] [<ffffffffa000b1e1>] >>> bch_btree_insert+0xf1/0x170 [bcache] >>> Jul 6 00:26:32 mcp kernel: [30957.747908] [<ffffffff810d2180>] ? >>> abort_exclusive_wait+0xb0/0xb0 >>> Jul 6 00:26:32 mcp kernel: [<ffffffff810d2180>] ? >>> abort_exclusive_wait+0xb0/0xb0 >>> Jul 6 00:26:33 mcp kernel: [<ffffffffa0021d66>] >>> write_dirty_finish+0x1f6/0x260 [bcache] >>> Jul 6 00:26:33 mcp kernel: [30957.755602] [<ffffffffa0021d66>] >>> write_dirty_finish+0x1f6/0x260 [bcache] >>> Jul 6 00:26:33 mcp kernel: [30957.763908] [<ffffffff810c3e76>] ? >>> __dequeue_entity+0x26/0x40 >>> Jul 6 00:26:33 mcp kernel: [<ffffffff810c3e76>] ? >>> __dequeue_entity+0x26/0x40 >>> Jul 6 00:26:33 mcp kernel: [30957.771228] [<ffffffff810135d6>] ? >>> __switch_to+0x126/0x4c0 >>> Jul 6 00:26:33 mcp kernel: [<ffffffff810135d6>] ? __switch_to+0x126/0x4c0 >>> Jul 6 00:26:33 mcp kernel: [30957.778283] [<ffffffff810a68e6>] >>> process_one_work+0x176/0x430 >>> Jul 6 00:26:33 mcp kernel: [<ffffffff810a68e6>] >>> process_one_work+0x176/0x430 >>> Jul 6 00:26:33 mcp kernel: [30957.785598] [<ffffffff810a757b>] >>> worker_thread+0x11b/0x3a0 >>> Jul 6 00:26:33 mcp kernel: [<ffffffff810a757b>] worker_thread+0x11b/0x3a0 >>> Jul 6 00:26:33 mcp kernel: [30957.792656] [<ffffffff810a7460>] ? >>> rescuer_thread+0x3b0/0x3b0 >>> Jul 6 00:26:33 mcp kernel: [<ffffffff810a7460>] ? >>> rescuer_thread+0x3b0/0x3b0 >>> Jul 6 00:26:33 mcp kernel: [30957.799982] [<ffffffff810ae2d1>] >>> kthread+0xe1/0x100 >>> Jul 6 00:26:33 mcp kernel: [<ffffffff810ae2d1>] kthread+0xe1/0x100 >>> Jul 6 00:26:33 mcp kernel: [30957.806427] [<ffffffff810ae1f0>] ? >>> insert_kthread_work+0x40/0x40 >>> Jul 6 00:26:33 mcp kernel: [<ffffffff810ae1f0>] ? >>> insert_kthread_work+0x40/0x40 >>> Jul 6 00:26:33 mcp kernel: [30957.813995] [<ffffffff8170083c>] >>> ret_from_fork+0x7c/0xb0 >>> Jul 6 00:26:33 mcp kernel: [<ffffffff8170083c>] ret_from_fork+0x7c/0xb0 >>> Jul 6 00:26:33 mcp kernel: [30957.820844] [<ffffffff810ae1f0>] ? >>> insert_kthread_work+0x40/0x40 >>> Jul 6 00:26:33 mcp kernel: [<ffffffff810ae1f0>] ? >>> insert_kthread_work+0x40/0x40 >>> Jul 6 00:26:33 mcp kernel: [30957.828357] ---[ end trace >>> 874ec8b4276a8f33 ]--- >>> Jul 6 00:26:33 mcp kernel: ---[ end trace 874ec8b4276a8f33 ]--- >>> Jul 6 00:26:33 mcp kernel: [30957.834379] bcache: bch_btree_insert() >>> error -12 >>> Jul 6 00:26:33 mcp kernel: bcache: bch_btree_insert() error -12 >>> >>> --Larkin >> -- >> To unsubscribe from this list: send the line "unsubscribe linux-bcache" in >> the body of a message to majordomo@xxxxxxxxxxxxxxx >> More majordomo info at http://vger.kernel.org/majordomo-info.html > -- > To unsubscribe from this list: send the line "unsubscribe linux-bcache" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe linux-bcache" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html