Re: bcache_writebac at nearly 100% CPU

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thank you for this information. I was finally able to detach all backing
devices by (carefully) over writing the dirty blocks in writethrough
mode. Once I did that the remaining backing devices became clean and I
was able to detach them.

I did not see any "split failed" messages while running a 3.15 kernel.

--Larkin

On 7/16/2014 10:09 PM, Jianjian Huo wrote:
> So you still see "btree split failed" errors in your new 3.15 kernel?
>
> All the data items stored in cache are in B+ tree structure. You can
> use the bcache debugfs file to find out where your dirty data items
> are, just
>
> cat /sys/kernel/debug/bcache/<your_cache_set_uuid> | grep dirty
>
> On the left side of each item, it shows the data size and starting
> location of backing device, the right side shows the stored location
> in cache device.
>
> -Jianjian
>
> On Wed, Jul 16, 2014 at 8:04 AM, Larkin Lowrey
> <llowrey@xxxxxxxxxxxxxxxxx> wrote:
>> Thank you for your reply.
>>
>> I am now running kernel 3.15.4-200.fc20.x86_64 and the problem still
>> exists. If the bad record could be found could it be repaired manually?
>>
>> The stuck backing devices total 17TB and I don't have that much spare
>> capacity to relocate the data so I am motivated to try to repair
>> in-place rather than wipe and rebuild. If you have any tips for where to
>> look in the source code for the on-disk record format for the cache
>> store and what to look for to identify the bad record(s) I would greatly
>> appreciate it.
>>
>> Memory on this box is only lightly used. Of the 16GB, there is never
>> less than 10GB free (free + buffers + cache).
>>
>> Here's a stack dump of the bcache_writebac thread:
>>
>> task: ffff880408ce4f00 ti: ffff8803f4eb4000 task.ti: ffff8803f4eb4000
>> RIP: 0010:[<ffffffffa0005c11>]  [<ffffffffa0005c11>]
>> refill_keybuf_fn+0x61/0x1d0 [bcache]
>> RSP: 0018:ffff8803f4eb7b78  EFLAGS: 00000246
>> RAX: 0000000000000001 RBX: ffff8803fdc13000 RCX: ffff8803f4eb7e40
>> RDX: 0000000023ad5e98 RSI: ffff8802820a41b8 RDI: ffff8803ff830bc8
>> RBP: ffff8803f4eb7b78 R08: ffff880402520000 R09: 0000000000000001
>> R10: ffff88040fbdc000 R11: 000007ffffffffff R12: ffff8803f4eb7d70
>> R13: 0000000000000000 R14: ffff8803f4eb7d00 R15: ffff8803fdc13000
>> FS:  00007fecc4d59700(0000) GS:ffff880427c00000(0000) knlGS:0000000000000000
>> CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
>> CR2: 00007fa1adad3000 CR3: 00000003f17dc000 CR4: 00000000000007f0
>> Stack:
>>  ffff8803f4eb7c30 ffffffffa0008292 00000001d06df860 ffffffffa0005bb0
>>  ffff8803fdc130c8 ffff880131682990 ffff88040cc0e8c8 0000000000000004
>>  0000000000000001 ffff8802820a41d0 ffff8802820cc118 ffff880402520000
>> Call Trace:
>>  [<ffffffffa0008292>] bch_btree_map_keys_recurse+0x72/0x1d0 [bcache]
>>  [<ffffffffa0005bb0>] ? btree_insert_key+0xe0/0xe0 [bcache]
>>  [<ffffffffa0007ed5>] ? bch_btree_node_get+0xd5/0x290 [bcache]
>>  [<ffffffffa0008327>] bch_btree_map_keys_recurse+0x107/0x1d0 [bcache]
>>  [<ffffffffa0005bb0>] ? btree_insert_key+0xe0/0xe0 [bcache]
>>  [<ffffffffa000b3f7>] bch_btree_map_keys+0x127/0x150 [bcache]
>>  [<ffffffffa0005bb0>] ? btree_insert_key+0xe0/0xe0 [bcache]
>>  [<ffffffffa00209b0>] ? bch_crc64+0x50/0x50 [bcache]
>>  [<ffffffffa000b4d4>] bch_refill_keybuf+0xb4/0x220 [bcache]
>>  [<ffffffff810d0350>] ? abort_exclusive_wait+0xb0/0xb0
>>  [<ffffffffa00209b0>] ? bch_crc64+0x50/0x50 [bcache]
>>  [<ffffffffa0021204>] bch_writeback_thread+0x154/0x840 [bcache]
>>  [<ffffffffa00210b0>] ? write_dirty_finish+0x260/0x260 [bcache]
>>  [<ffffffff810ac538>] kthread+0xd8/0xf0
>>  [<ffffffff810ac460>] ? insert_kthread_work+0x40/0x40
>>  [<ffffffff816ff23c>] ret_from_fork+0x7c/0xb0
>>  [<ffffffff810ac460>] ? insert_kthread_work+0x40/0x40
>>
>> When this snapshot was taken both bcache_writebac processes had similar
>> stack traces but one was writing data and the other was not. The one
>> that was writing was only writing at 3-4 4K blocks per second. I tried
>> setting writeback_rate to a very large value and writeback_percent to
>> zero but neither changed the rate.
>>
>> The backing device that is now writing does not write consistently. It
>> will be idle for many hours then spontaneously start writing at ~4 4K
>> blocks per second for several hours then stop again.
>>
>> --Larkin
>>
>> On 7/16/2014 1:27 AM, Samuel Huo wrote:
>>> It seems your original 3.14 kernel hit the "btree split failed" bug
>>> when bcache wrote back some dirty data into backing device and
>>> inserted keys to invalidate those dirty data in btree index. But I
>>> think that shouldn't prevent future writeback.
>>> The "btree split failed" problem gets fixed in 3.15, as Slava
>>> suggested in another post.
>>>
>>> You said writeback takes near 100% CPU, can you try to find out where
>>> the writeback thread gets stuck at? and how is your memory usage?
>>>
>>> -Jianjian
>>>
>>> On Mon, Jul 14, 2014 at 5:47 AM, Larkin Lowrey
>>> <llowrey@xxxxxxxxxxxxxxxxx> wrote:
>>>> I am still unable to detach either backing device. There is still dirty
>>>> data and it's not being written out. The bcache_writebac processes are
>>>> at max CPU and the backing devices are idle. It appears to me that
>>>> something is corrupt in the cache device which prevents the full dirty
>>>> data from being written.
>>>>
>>>> Is there anything useful I can do to debug?
>>>>
>>>> Several GB of dirty data had accumulated since the data reported below.
>>>> I set writeback_percent to zero in order to flush the dirty data and
>>>> most did but not all. In the case of md1 it had grown to 6GB but only
>>>> dropped to 2GB (it had been as low as 1.1GB a few days ago). The md3
>>>> device had 1.5GB but only dropped to 696KB (the same as the minimum from
>>>> a few days ago). I have switch to writethrough mode in hopes that I
>>>> don't get into further trouble.
>>>>
>>>> Attempting to detach (via /sys/block/md?/bcache/detach) has had no effect.
>>>>
>>>> Any help would be much appreciated!
>>>>
>>>> --Larkin
>>>>
>>>> On 7/8/2014 12:18 PM, Larkin Lowrey wrote:
>>>>> %CPU %MEM     TIME+ COMMAND
>>>>> 98.8  0.0 645:05.50 bcache_writebac
>>>>> 97.2  0.0 600:50.18 bcache_writebac
>>>>>
>>>>> The machine in question is a backup server. Backups usually take
>>>>> 40-60min but this morning took 8.5-9 hours. I did make several changes
>>>>> yesterday.
>>>>>
>>>>> There is a single (md raid 1) cache device and I had 2 md raid6 arrays
>>>>> attached. Yesterday I attached a third raid6.
>>>>>
>>>>> Since noticing the high CPU usage (and poor IO performance) I attempted
>>>>> to detach each of the three backing devices with the idea that I would
>>>>> rebuild the cache set. One of the backing devices did detach but two
>>>>> have not. One of the two remaining devices has 1GB of dirty data and the
>>>>> other has 696KB.
>>>>>
>>>>> I had been running kernel 3.14.9-200.fc20.x86_64 but once this high load
>>>>> situation happened I switched to 3.15.3-200.fc20.x86_64.
>>>>>
>>>>> When I reboot the two bcache_writebac processes start out immediately at
>>>>> high load.
>>>>>
>>>>> /sys/block/md[13]/bcache/writeback_percent start out at 10%.
>>>>>
>>>>> There is no IO other than to the root fs which is on a separate raid1.
>>>>>
>>>>> ==> /sys/block/md1/bcache/writeback_rate_debug <==
>>>>> rate:           512/sec
>>>>> dirty:          1.1G
>>>>> target:         15.5G
>>>>> proportional:   -13.7M
>>>>> derivative:     0
>>>>> change:         -13.7M/sec
>>>>> next io:        -62ms
>>>>>
>>>>> ==> /sys/block/md3/bcache/writeback_rate_debug <==
>>>>> rate:           512/sec
>>>>> dirty:          696k
>>>>> target:         4.4G
>>>>> proportional:   -4.2M
>>>>> derivative:     0
>>>>> change:         -4.2M/sec
>>>>> next io:        0ms
>>>>>
>>>>> After switching writeback_percent to 0%, there is still no IO and...
>>>>>
>>>>> ==> /sys/block/md1/bcache/writeback_rate_debug <==
>>>>> rate:           512/sec
>>>>> dirty:          1.1G
>>>>> target:         15.5G
>>>>> proportional:   -13.7M
>>>>> derivative:     0
>>>>> change:         -13.7M/sec
>>>>> next io:        -50ms
>>>>>
>>>>> ==> /sys/block/md3/bcache/writeback_rate_debug <==
>>>>> rate:           512/sec
>>>>> dirty:          696k
>>>>> target:         4.4G
>>>>> proportional:   -4.2M
>>>>> derivative:     0
>>>>> change:         -4.2M/sec
>>>>> next io:        0ms
>>>>>
>>>>> ... and switching back to 10%, still no IO and...
>>>>>
>>>>> ==> /sys/block/md1/bcache/writeback_rate_debug <==
>>>>> rate:           512/sec
>>>>> dirty:          1.1G
>>>>> target:         15.5G
>>>>> proportional:   -13.7M
>>>>> derivative:     0
>>>>> change:         -13.7M/sec
>>>>> next io:        -67ms
>>>>>
>>>>> ==> /sys/block/md3/bcache/writeback_rate_debug <==
>>>>> rate:           512/sec
>>>>> dirty:          696k
>>>>> target:         4.4G
>>>>> proportional:   -4.2M
>>>>> derivative:     0
>>>>> change:         -4.2M/sec
>>>>> next io:        0ms
>>>>>
>>>>> The only log messages for bcache are:
>>>>>
>>>>> [   23.728302] bcache: register_bdev() registered backing device md2
>>>>> [   23.750124] bcache: register_bdev() registered backing device md1
>>>>> [   23.776249] bcache: register_bdev() registered backing device md3
>>>>> [   25.414604] bcache: bch_journal_replay() journal replay done, 66 keys
>>>>> in 10 entries, seq 26249932
>>>>> [   25.534038] bcache: bch_cached_dev_attach() Caching md3 as bcache2 on
>>>>> set 66c1a39b-5898-4aae-abe4-6ab609c18155
>>>>> [   25.682785] bcache: bch_cached_dev_attach() Caching md1 as bcache1 on
>>>>> set 66c1a39b-5898-4aae-abe4-6ab609c18155
>>>>> [   25.695961] bcache: register_cache() registered cache device dm-2
>>>>>
>>>>> Status (shortly after reboot):
>>>>>
>>>>> # bcache-status -s
>>>>> --- bcache ---
>>>>> UUID                        66c1a39b-5898-4aae-abe4-6ab609c18155
>>>>> Block Size                  4.00 KiB
>>>>> Bucket Size                 2.0 MiB
>>>>> Congested?                  False
>>>>> Read Congestion             2.0ms
>>>>> Write Congestion            20.0ms
>>>>> Total Cache Size            200 GiB
>>>>> Total Cache Used            148 GiB     (74%)
>>>>> Total Cache Unused          52 GiB      (26%)
>>>>> Evictable Cache             200 GiB     (100%)
>>>>> Replacement Policy          [lru] fifo random
>>>>> Cache Mode                  writethrough [writeback] writearound none
>>>>> Total Hits                  6174        (94%)
>>>>> Total Misses                380
>>>>> Total Bypass Hits           0
>>>>> Total Bypass Misses         0
>>>>> Total Bypassed              0 B
>>>>> --- Backing Device ---
>>>>>   Device File               /dev/md1 (9:1)
>>>>>   bcache Device File        /dev/bcache1 (252:1)
>>>>>   Size                      13 TiB
>>>>>   Cache Mode                writethrough [writeback] writearound none
>>>>>   Readahead                 0
>>>>>   Sequential Cutoff         16.0 MiB
>>>>>   Merge sequential?         False
>>>>>   State                     dirty
>>>>>   Writeback?                True
>>>>>   Dirty Data                1 GiB
>>>>>   Total Hits                6108        (99%)
>>>>>   Total Misses              12
>>>>>   Total Bypass Hits         0
>>>>>   Total Bypass Misses       0
>>>>>   Total Bypassed            0 B
>>>>> --- Backing Device ---
>>>>>   Device File               /dev/md3 (9:3)
>>>>>   bcache Device File        /dev/bcache2 (252:2)
>>>>>   Size                      4 TiB
>>>>>   Cache Mode                writethrough [writeback] writearound none
>>>>>   Readahead                 0
>>>>>   Sequential Cutoff         16.0 MiB
>>>>>   Merge sequential?         False
>>>>>   State                     dirty
>>>>>   Writeback?                True
>>>>>   Dirty Data                696.00 KiB
>>>>>   Total Hits                66  (15%)
>>>>>   Total Misses              368
>>>>>   Total Bypass Hits         0
>>>>>   Total Bypass Misses       0
>>>>>   Total Bypassed            0 B
>>>>> --- Cache Device ---
>>>>>   Device File               /dev/dm-2 (253:2)
>>>>>   Size                      200 GiB
>>>>>   Block Size                4.00 KiB
>>>>>   Bucket Size               2.0 MiB
>>>>>   Replacement Policy        [lru] fifo random
>>>>>   Discard?                  False
>>>>>   I/O Errors                0
>>>>>   Metadata Written          2.0 MiB
>>>>>   Data Written              1.4 MiB
>>>>>   Buckets                   102400
>>>>>   Cache Used                148 GiB     (74%)
>>>>>   Cache Unused              52 GiB      (26%)
>>>>>
>>>>> I did find the following in the logs from a day prior to when I did the
>>>>> work. The cacti graphs show high load at that time but a raid-check
>>>>> started shortly before this which usually causes this kind of load. So,
>>>>> the problem may have been occurring since the 6th without being noticing.
>>>>>
>>>>> Jul  6 00:26:32 mcp kernel: WARNING: CPU: 1 PID: 15568 at
>>>>> drivers/md/bcache/btree.c:1979 btree_split+0x5bb/0x630 [bcache]()
>>>>> Jul  6 00:26:32 mcp kernel: bcache: btree split failed
>>>>> Jul  6 00:26:32 mcp kernel: Modules linked in:
>>>>> Jul  6 00:26:32 mcp kernel: [30957.447979] bcache: btree split failed
>>>>> Jul  6 00:26:32 mcp kernel: [30957.451654] Modules linked in:
>>>>> binfmt_misc cfg80211 rfkill it87 hwmon_vid nf_conntrack_ipv4 ip6t_REJECT
>>>>> nf_defrag_ipv4 nf_conntrack_ipv6 nf_defrag_ip
>>>>> v6 xt_conntrack ip6table_filter nf_conntrack ip6_tables raid456
>>>>> async_raid6_recov async_memcpy async_pq async_xor async_tx kvm microcode
>>>>> edac_core serio_raw sp5100_tco edac_mce_amd
>>>>>  k10temp snd_hda_codec_realtek i2c_piix4 snd_hda_codec_generic
>>>>> snd_hda_codec_hdmi snd_hda_intel snd_hda_codec snd_hwdep snd_seq
>>>>> snd_seq_device snd_pcm snd_timer snd r8169 soundcore
>>>>>  mii shpchp wmi acpi_cpufreq nfsd auth_rpcgss nfs_acl lockd sunrpc btrfs
>>>>> xor raid6_pq raid1 radeon i2c_algo_bit firewire_ohci drm_kms_helper
>>>>> firewire_core ttm crc_itu_t mvsas drm l
>>>>> ibsas scsi_transport_sas i2c_core cpufreq_stats bcache
>>>>> Jul  6 00:26:32 mcp kernel: [30957.528183] CPU: 1 PID: 15568 Comm:
>>>>> kworker/1:2 Not tainted 3.14.8-200.fc20.x86_64 #1
>>>>> Jul  6 00:26:32 mcp kernel: [30957.537550] Hardware name: Gigabyte
>>>>> Technology Co., Ltd. GA-890GPA-UD3H/GA-890GPA-UD3H, BIOS F9 09/09/2011
>>>>> Jul  6 00:26:32 mcp kernel: binfmt_misc cfg80211 rfkill it87 hwmon_vid
>>>>> nf_conntrack_ipv4 ip6t_REJECT nf_defrag_ipv4 nf_conntrack_ipv6
>>>>> nf_defrag_ipv6 xt_conntrack ip6table_filter nf
>>>>> _conntrack ip6_tables raid456 async_raid6_recov async_memcpy async_pq
>>>>> async_xor async_tx kvm microcode edac_core serio_raw sp5100_tco
>>>>> edac_mce_amd k10temp snd_hda_codec_realtek i2c
>>>>> _piix4 snd_hda_codec_generic snd_hda_codec_hdmi snd_hda_intel
>>>>> snd_hda_codec snd_hwdep snd_seq snd_seq_device snd_pcm snd_timer snd
>>>>> r8169 soundcore mii shpchp wmi acpi_cpufreq nfsd
>>>>> auth_rpcgss nfs_acl lockd sunrpc btrfs xor raid6_pq raid1 radeon
>>>>> i2c_algo_bit firewire_ohci drm_kms_helper firewire_core ttm crc_itu_t
>>>>> mvsas drm libsas scsi_transport_sas i2c_core
>>>>> cpufreq_stats bcache
>>>>> Jul  6 00:26:32 mcp kernel: [30957.548771] Workqueue: events
>>>>> write_dirty_finish [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: CPU: 1 PID: 15568 Comm: kworker/1:2 Not
>>>>> tainted 3.14.8-200.fc20.x86_64 #1
>>>>> Jul  6 00:26:32 mcp kernel: [30957.555773]  0000000000000000
>>>>> 00000000dea1a02f ffff880007ffd8b0 ffffffff816f0502
>>>>> Jul  6 00:26:32 mcp kernel: Hardware name: Gigabyte Technology Co., Ltd.
>>>>> GA-890GPA-UD3H/GA-890GPA-UD3H, BIOS F9 09/09/2011
>>>>> Jul  6 00:26:32 mcp kernel: [30957.564855]  ffff880007ffd8f8
>>>>> ffff880007ffd8e8 ffffffff8108a1cd ffff88040e6b7800
>>>>> Jul  6 00:26:32 mcp kernel: Workqueue: events write_dirty_finish [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: 0000000000000000 00000000dea1a02f
>>>>> ffff880007ffd8b0 ffffffff816f0502
>>>>> Jul  6 00:26:32 mcp kernel: ffff880007ffd8f8 ffff880007ffd8e8
>>>>> ffffffff8108a1cd ffff88040e6b7800
>>>>> Jul  6 00:26:32 mcp kernel: [30957.573943]  ffffffffffffffe4
>>>>> ffff880007ffdd58 ffff880007ffd980 0000000000000000
>>>>> Jul  6 00:26:32 mcp kernel: [30957.583045] Call Trace:
>>>>> Jul  6 00:26:32 mcp kernel: [30957.587031]  [<ffffffff816f0502>]
>>>>> dump_stack+0x45/0x56
>>>>> Jul  6 00:26:32 mcp kernel: ffffffffffffffe4 ffff880007ffdd58
>>>>> ffff880007ffd980 0000000000000000
>>>>> Jul  6 00:26:32 mcp kernel: Call Trace:
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff816f0502>] dump_stack+0x45/0x56
>>>>> Jul  6 00:26:32 mcp kernel: [30957.593729]  [<ffffffff8108a1cd>]
>>>>> warn_slowpath_common+0x7d/0xa0
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff8108a1cd>]
>>>>> warn_slowpath_common+0x7d/0xa0
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff8108a24c>] warn_slowpath_fmt+0x5c/0x80
>>>>> Jul  6 00:26:32 mcp kernel: [30957.601287]  [<ffffffff8108a24c>]
>>>>> warn_slowpath_fmt+0x5c/0x80
>>>>> Jul  6 00:26:32 mcp kernel: [30957.608568]  [<ffffffffa000965b>]
>>>>> btree_split+0x5bb/0x630 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa000965b>] btree_split+0x5bb/0x630
>>>>> [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [30957.616248]  [<ffffffff816f3d22>] ?
>>>>> __schedule+0x2e2/0x740
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff816f3d22>] ? __schedule+0x2e2/0x740
>>>>> Jul  6 00:26:32 mcp kernel: [30957.623240]  [<ffffffff8136decd>] ?
>>>>> list_del+0xd/0x30
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff8136decd>] ? list_del+0xd/0x30
>>>>> Jul  6 00:26:32 mcp kernel: [30957.629792]  [<ffffffffa0009771>]
>>>>> bch_btree_insert_node+0xa1/0x2c0 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [30957.638293]  [<ffffffffa000a724>]
>>>>> btree_insert_fn+0x24/0x50 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa0009771>]
>>>>> bch_btree_insert_node+0xa1/0x2c0 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa000a724>]
>>>>> btree_insert_fn+0x24/0x50 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa0007861>]
>>>>> bch_btree_map_nodes_recurse+0x51/0x180 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [30957.646175]  [<ffffffffa0007861>]
>>>>> bch_btree_map_nodes_recurse+0x51/0x180 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [30957.655171]  [<ffffffffa000a700>] ?
>>>>> bch_btree_insert_check_key+0x1b0/0x1b0 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa000a700>] ?
>>>>> bch_btree_insert_check_key+0x1b0/0x1b0 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa000da62>] ?
>>>>> bch_btree_ptr_invalid+0x12/0x20 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [30957.664290]  [<ffffffffa000da62>] ?
>>>>> bch_btree_ptr_invalid+0x12/0x20 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [30957.672814]  [<ffffffffa000c48d>] ?
>>>>> bch_btree_ptr_bad+0x4d/0x110 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa000c48d>] ?
>>>>> bch_btree_ptr_bad+0x4d/0x110 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [30957.681045]  [<ffffffff816f6592>] ?
>>>>> down_write+0x12/0x30
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff816f6592>] ? down_write+0x12/0x30
>>>>> Jul  6 00:26:32 mcp kernel: [30957.731889]  [<ffffffff810b2d66>] ?
>>>>> srcu_notifier_call_chain+0x16/0x20
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff810b2d66>] ?
>>>>> srcu_notifier_call_chain+0x16/0x20
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa000b1e1>]
>>>>> bch_btree_insert+0xf1/0x170 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [30957.739905]  [<ffffffffa000b1e1>]
>>>>> bch_btree_insert+0xf1/0x170 [bcache]
>>>>> Jul  6 00:26:32 mcp kernel: [30957.747908]  [<ffffffff810d2180>] ?
>>>>> abort_exclusive_wait+0xb0/0xb0
>>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff810d2180>] ?
>>>>> abort_exclusive_wait+0xb0/0xb0
>>>>> Jul  6 00:26:33 mcp kernel: [<ffffffffa0021d66>]
>>>>> write_dirty_finish+0x1f6/0x260 [bcache]
>>>>> Jul  6 00:26:33 mcp kernel: [30957.755602]  [<ffffffffa0021d66>]
>>>>> write_dirty_finish+0x1f6/0x260 [bcache]
>>>>> Jul  6 00:26:33 mcp kernel: [30957.763908]  [<ffffffff810c3e76>] ?
>>>>> __dequeue_entity+0x26/0x40
>>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810c3e76>] ?
>>>>> __dequeue_entity+0x26/0x40
>>>>> Jul  6 00:26:33 mcp kernel: [30957.771228]  [<ffffffff810135d6>] ?
>>>>> __switch_to+0x126/0x4c0
>>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810135d6>] ? __switch_to+0x126/0x4c0
>>>>> Jul  6 00:26:33 mcp kernel: [30957.778283]  [<ffffffff810a68e6>]
>>>>> process_one_work+0x176/0x430
>>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810a68e6>]
>>>>> process_one_work+0x176/0x430
>>>>> Jul  6 00:26:33 mcp kernel: [30957.785598]  [<ffffffff810a757b>]
>>>>> worker_thread+0x11b/0x3a0
>>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810a757b>] worker_thread+0x11b/0x3a0
>>>>> Jul  6 00:26:33 mcp kernel: [30957.792656]  [<ffffffff810a7460>] ?
>>>>> rescuer_thread+0x3b0/0x3b0
>>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810a7460>] ?
>>>>> rescuer_thread+0x3b0/0x3b0
>>>>> Jul  6 00:26:33 mcp kernel: [30957.799982]  [<ffffffff810ae2d1>]
>>>>> kthread+0xe1/0x100
>>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810ae2d1>] kthread+0xe1/0x100
>>>>> Jul  6 00:26:33 mcp kernel: [30957.806427]  [<ffffffff810ae1f0>] ?
>>>>> insert_kthread_work+0x40/0x40
>>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810ae1f0>] ?
>>>>> insert_kthread_work+0x40/0x40
>>>>> Jul  6 00:26:33 mcp kernel: [30957.813995]  [<ffffffff8170083c>]
>>>>> ret_from_fork+0x7c/0xb0
>>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff8170083c>] ret_from_fork+0x7c/0xb0
>>>>> Jul  6 00:26:33 mcp kernel: [30957.820844]  [<ffffffff810ae1f0>] ?
>>>>> insert_kthread_work+0x40/0x40
>>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810ae1f0>] ?
>>>>> insert_kthread_work+0x40/0x40
>>>>> Jul  6 00:26:33 mcp kernel: [30957.828357] ---[ end trace
>>>>> 874ec8b4276a8f33 ]---
>>>>> Jul  6 00:26:33 mcp kernel: ---[ end trace 874ec8b4276a8f33 ]---
>>>>> Jul  6 00:26:33 mcp kernel: [30957.834379] bcache: bch_btree_insert()
>>>>> error -12
>>>>> Jul  6 00:26:33 mcp kernel: bcache: bch_btree_insert() error -12
>>>>>
>>>>> --Larkin
>>>> --
>>>> To unsubscribe from this list: send the line "unsubscribe linux-bcache" in
>>>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe linux-bcache" in
>>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe linux-bcache" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe linux-bcache" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[Index of Archives]     [Linux ARM Kernel]     [Linux Filesystem Development]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [ECOS]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux