Re: bcache_writebac at nearly 100% CPU

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



So you still see "btree split failed" errors in your new 3.15 kernel?

All the data items stored in cache are in B+ tree structure. You can
use the bcache debugfs file to find out where your dirty data items
are, just

cat /sys/kernel/debug/bcache/<your_cache_set_uuid> | grep dirty

On the left side of each item, it shows the data size and starting
location of backing device, the right side shows the stored location
in cache device.

-Jianjian

On Wed, Jul 16, 2014 at 8:04 AM, Larkin Lowrey
<llowrey@xxxxxxxxxxxxxxxxx> wrote:
> Thank you for your reply.
>
> I am now running kernel 3.15.4-200.fc20.x86_64 and the problem still
> exists. If the bad record could be found could it be repaired manually?
>
> The stuck backing devices total 17TB and I don't have that much spare
> capacity to relocate the data so I am motivated to try to repair
> in-place rather than wipe and rebuild. If you have any tips for where to
> look in the source code for the on-disk record format for the cache
> store and what to look for to identify the bad record(s) I would greatly
> appreciate it.
>
> Memory on this box is only lightly used. Of the 16GB, there is never
> less than 10GB free (free + buffers + cache).
>
> Here's a stack dump of the bcache_writebac thread:
>
> task: ffff880408ce4f00 ti: ffff8803f4eb4000 task.ti: ffff8803f4eb4000
> RIP: 0010:[<ffffffffa0005c11>]  [<ffffffffa0005c11>]
> refill_keybuf_fn+0x61/0x1d0 [bcache]
> RSP: 0018:ffff8803f4eb7b78  EFLAGS: 00000246
> RAX: 0000000000000001 RBX: ffff8803fdc13000 RCX: ffff8803f4eb7e40
> RDX: 0000000023ad5e98 RSI: ffff8802820a41b8 RDI: ffff8803ff830bc8
> RBP: ffff8803f4eb7b78 R08: ffff880402520000 R09: 0000000000000001
> R10: ffff88040fbdc000 R11: 000007ffffffffff R12: ffff8803f4eb7d70
> R13: 0000000000000000 R14: ffff8803f4eb7d00 R15: ffff8803fdc13000
> FS:  00007fecc4d59700(0000) GS:ffff880427c00000(0000) knlGS:0000000000000000
> CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
> CR2: 00007fa1adad3000 CR3: 00000003f17dc000 CR4: 00000000000007f0
> Stack:
>  ffff8803f4eb7c30 ffffffffa0008292 00000001d06df860 ffffffffa0005bb0
>  ffff8803fdc130c8 ffff880131682990 ffff88040cc0e8c8 0000000000000004
>  0000000000000001 ffff8802820a41d0 ffff8802820cc118 ffff880402520000
> Call Trace:
>  [<ffffffffa0008292>] bch_btree_map_keys_recurse+0x72/0x1d0 [bcache]
>  [<ffffffffa0005bb0>] ? btree_insert_key+0xe0/0xe0 [bcache]
>  [<ffffffffa0007ed5>] ? bch_btree_node_get+0xd5/0x290 [bcache]
>  [<ffffffffa0008327>] bch_btree_map_keys_recurse+0x107/0x1d0 [bcache]
>  [<ffffffffa0005bb0>] ? btree_insert_key+0xe0/0xe0 [bcache]
>  [<ffffffffa000b3f7>] bch_btree_map_keys+0x127/0x150 [bcache]
>  [<ffffffffa0005bb0>] ? btree_insert_key+0xe0/0xe0 [bcache]
>  [<ffffffffa00209b0>] ? bch_crc64+0x50/0x50 [bcache]
>  [<ffffffffa000b4d4>] bch_refill_keybuf+0xb4/0x220 [bcache]
>  [<ffffffff810d0350>] ? abort_exclusive_wait+0xb0/0xb0
>  [<ffffffffa00209b0>] ? bch_crc64+0x50/0x50 [bcache]
>  [<ffffffffa0021204>] bch_writeback_thread+0x154/0x840 [bcache]
>  [<ffffffffa00210b0>] ? write_dirty_finish+0x260/0x260 [bcache]
>  [<ffffffff810ac538>] kthread+0xd8/0xf0
>  [<ffffffff810ac460>] ? insert_kthread_work+0x40/0x40
>  [<ffffffff816ff23c>] ret_from_fork+0x7c/0xb0
>  [<ffffffff810ac460>] ? insert_kthread_work+0x40/0x40
>
> When this snapshot was taken both bcache_writebac processes had similar
> stack traces but one was writing data and the other was not. The one
> that was writing was only writing at 3-4 4K blocks per second. I tried
> setting writeback_rate to a very large value and writeback_percent to
> zero but neither changed the rate.
>
> The backing device that is now writing does not write consistently. It
> will be idle for many hours then spontaneously start writing at ~4 4K
> blocks per second for several hours then stop again.
>
> --Larkin
>
> On 7/16/2014 1:27 AM, Samuel Huo wrote:
>> It seems your original 3.14 kernel hit the "btree split failed" bug
>> when bcache wrote back some dirty data into backing device and
>> inserted keys to invalidate those dirty data in btree index. But I
>> think that shouldn't prevent future writeback.
>> The "btree split failed" problem gets fixed in 3.15, as Slava
>> suggested in another post.
>>
>> You said writeback takes near 100% CPU, can you try to find out where
>> the writeback thread gets stuck at? and how is your memory usage?
>>
>> -Jianjian
>>
>> On Mon, Jul 14, 2014 at 5:47 AM, Larkin Lowrey
>> <llowrey@xxxxxxxxxxxxxxxxx> wrote:
>>> I am still unable to detach either backing device. There is still dirty
>>> data and it's not being written out. The bcache_writebac processes are
>>> at max CPU and the backing devices are idle. It appears to me that
>>> something is corrupt in the cache device which prevents the full dirty
>>> data from being written.
>>>
>>> Is there anything useful I can do to debug?
>>>
>>> Several GB of dirty data had accumulated since the data reported below.
>>> I set writeback_percent to zero in order to flush the dirty data and
>>> most did but not all. In the case of md1 it had grown to 6GB but only
>>> dropped to 2GB (it had been as low as 1.1GB a few days ago). The md3
>>> device had 1.5GB but only dropped to 696KB (the same as the minimum from
>>> a few days ago). I have switch to writethrough mode in hopes that I
>>> don't get into further trouble.
>>>
>>> Attempting to detach (via /sys/block/md?/bcache/detach) has had no effect.
>>>
>>> Any help would be much appreciated!
>>>
>>> --Larkin
>>>
>>> On 7/8/2014 12:18 PM, Larkin Lowrey wrote:
>>>> %CPU %MEM     TIME+ COMMAND
>>>> 98.8  0.0 645:05.50 bcache_writebac
>>>> 97.2  0.0 600:50.18 bcache_writebac
>>>>
>>>> The machine in question is a backup server. Backups usually take
>>>> 40-60min but this morning took 8.5-9 hours. I did make several changes
>>>> yesterday.
>>>>
>>>> There is a single (md raid 1) cache device and I had 2 md raid6 arrays
>>>> attached. Yesterday I attached a third raid6.
>>>>
>>>> Since noticing the high CPU usage (and poor IO performance) I attempted
>>>> to detach each of the three backing devices with the idea that I would
>>>> rebuild the cache set. One of the backing devices did detach but two
>>>> have not. One of the two remaining devices has 1GB of dirty data and the
>>>> other has 696KB.
>>>>
>>>> I had been running kernel 3.14.9-200.fc20.x86_64 but once this high load
>>>> situation happened I switched to 3.15.3-200.fc20.x86_64.
>>>>
>>>> When I reboot the two bcache_writebac processes start out immediately at
>>>> high load.
>>>>
>>>> /sys/block/md[13]/bcache/writeback_percent start out at 10%.
>>>>
>>>> There is no IO other than to the root fs which is on a separate raid1.
>>>>
>>>> ==> /sys/block/md1/bcache/writeback_rate_debug <==
>>>> rate:           512/sec
>>>> dirty:          1.1G
>>>> target:         15.5G
>>>> proportional:   -13.7M
>>>> derivative:     0
>>>> change:         -13.7M/sec
>>>> next io:        -62ms
>>>>
>>>> ==> /sys/block/md3/bcache/writeback_rate_debug <==
>>>> rate:           512/sec
>>>> dirty:          696k
>>>> target:         4.4G
>>>> proportional:   -4.2M
>>>> derivative:     0
>>>> change:         -4.2M/sec
>>>> next io:        0ms
>>>>
>>>> After switching writeback_percent to 0%, there is still no IO and...
>>>>
>>>> ==> /sys/block/md1/bcache/writeback_rate_debug <==
>>>> rate:           512/sec
>>>> dirty:          1.1G
>>>> target:         15.5G
>>>> proportional:   -13.7M
>>>> derivative:     0
>>>> change:         -13.7M/sec
>>>> next io:        -50ms
>>>>
>>>> ==> /sys/block/md3/bcache/writeback_rate_debug <==
>>>> rate:           512/sec
>>>> dirty:          696k
>>>> target:         4.4G
>>>> proportional:   -4.2M
>>>> derivative:     0
>>>> change:         -4.2M/sec
>>>> next io:        0ms
>>>>
>>>> ... and switching back to 10%, still no IO and...
>>>>
>>>> ==> /sys/block/md1/bcache/writeback_rate_debug <==
>>>> rate:           512/sec
>>>> dirty:          1.1G
>>>> target:         15.5G
>>>> proportional:   -13.7M
>>>> derivative:     0
>>>> change:         -13.7M/sec
>>>> next io:        -67ms
>>>>
>>>> ==> /sys/block/md3/bcache/writeback_rate_debug <==
>>>> rate:           512/sec
>>>> dirty:          696k
>>>> target:         4.4G
>>>> proportional:   -4.2M
>>>> derivative:     0
>>>> change:         -4.2M/sec
>>>> next io:        0ms
>>>>
>>>> The only log messages for bcache are:
>>>>
>>>> [   23.728302] bcache: register_bdev() registered backing device md2
>>>> [   23.750124] bcache: register_bdev() registered backing device md1
>>>> [   23.776249] bcache: register_bdev() registered backing device md3
>>>> [   25.414604] bcache: bch_journal_replay() journal replay done, 66 keys
>>>> in 10 entries, seq 26249932
>>>> [   25.534038] bcache: bch_cached_dev_attach() Caching md3 as bcache2 on
>>>> set 66c1a39b-5898-4aae-abe4-6ab609c18155
>>>> [   25.682785] bcache: bch_cached_dev_attach() Caching md1 as bcache1 on
>>>> set 66c1a39b-5898-4aae-abe4-6ab609c18155
>>>> [   25.695961] bcache: register_cache() registered cache device dm-2
>>>>
>>>> Status (shortly after reboot):
>>>>
>>>> # bcache-status -s
>>>> --- bcache ---
>>>> UUID                        66c1a39b-5898-4aae-abe4-6ab609c18155
>>>> Block Size                  4.00 KiB
>>>> Bucket Size                 2.0 MiB
>>>> Congested?                  False
>>>> Read Congestion             2.0ms
>>>> Write Congestion            20.0ms
>>>> Total Cache Size            200 GiB
>>>> Total Cache Used            148 GiB     (74%)
>>>> Total Cache Unused          52 GiB      (26%)
>>>> Evictable Cache             200 GiB     (100%)
>>>> Replacement Policy          [lru] fifo random
>>>> Cache Mode                  writethrough [writeback] writearound none
>>>> Total Hits                  6174        (94%)
>>>> Total Misses                380
>>>> Total Bypass Hits           0
>>>> Total Bypass Misses         0
>>>> Total Bypassed              0 B
>>>> --- Backing Device ---
>>>>   Device File               /dev/md1 (9:1)
>>>>   bcache Device File        /dev/bcache1 (252:1)
>>>>   Size                      13 TiB
>>>>   Cache Mode                writethrough [writeback] writearound none
>>>>   Readahead                 0
>>>>   Sequential Cutoff         16.0 MiB
>>>>   Merge sequential?         False
>>>>   State                     dirty
>>>>   Writeback?                True
>>>>   Dirty Data                1 GiB
>>>>   Total Hits                6108        (99%)
>>>>   Total Misses              12
>>>>   Total Bypass Hits         0
>>>>   Total Bypass Misses       0
>>>>   Total Bypassed            0 B
>>>> --- Backing Device ---
>>>>   Device File               /dev/md3 (9:3)
>>>>   bcache Device File        /dev/bcache2 (252:2)
>>>>   Size                      4 TiB
>>>>   Cache Mode                writethrough [writeback] writearound none
>>>>   Readahead                 0
>>>>   Sequential Cutoff         16.0 MiB
>>>>   Merge sequential?         False
>>>>   State                     dirty
>>>>   Writeback?                True
>>>>   Dirty Data                696.00 KiB
>>>>   Total Hits                66  (15%)
>>>>   Total Misses              368
>>>>   Total Bypass Hits         0
>>>>   Total Bypass Misses       0
>>>>   Total Bypassed            0 B
>>>> --- Cache Device ---
>>>>   Device File               /dev/dm-2 (253:2)
>>>>   Size                      200 GiB
>>>>   Block Size                4.00 KiB
>>>>   Bucket Size               2.0 MiB
>>>>   Replacement Policy        [lru] fifo random
>>>>   Discard?                  False
>>>>   I/O Errors                0
>>>>   Metadata Written          2.0 MiB
>>>>   Data Written              1.4 MiB
>>>>   Buckets                   102400
>>>>   Cache Used                148 GiB     (74%)
>>>>   Cache Unused              52 GiB      (26%)
>>>>
>>>> I did find the following in the logs from a day prior to when I did the
>>>> work. The cacti graphs show high load at that time but a raid-check
>>>> started shortly before this which usually causes this kind of load. So,
>>>> the problem may have been occurring since the 6th without being noticing.
>>>>
>>>> Jul  6 00:26:32 mcp kernel: WARNING: CPU: 1 PID: 15568 at
>>>> drivers/md/bcache/btree.c:1979 btree_split+0x5bb/0x630 [bcache]()
>>>> Jul  6 00:26:32 mcp kernel: bcache: btree split failed
>>>> Jul  6 00:26:32 mcp kernel: Modules linked in:
>>>> Jul  6 00:26:32 mcp kernel: [30957.447979] bcache: btree split failed
>>>> Jul  6 00:26:32 mcp kernel: [30957.451654] Modules linked in:
>>>> binfmt_misc cfg80211 rfkill it87 hwmon_vid nf_conntrack_ipv4 ip6t_REJECT
>>>> nf_defrag_ipv4 nf_conntrack_ipv6 nf_defrag_ip
>>>> v6 xt_conntrack ip6table_filter nf_conntrack ip6_tables raid456
>>>> async_raid6_recov async_memcpy async_pq async_xor async_tx kvm microcode
>>>> edac_core serio_raw sp5100_tco edac_mce_amd
>>>>  k10temp snd_hda_codec_realtek i2c_piix4 snd_hda_codec_generic
>>>> snd_hda_codec_hdmi snd_hda_intel snd_hda_codec snd_hwdep snd_seq
>>>> snd_seq_device snd_pcm snd_timer snd r8169 soundcore
>>>>  mii shpchp wmi acpi_cpufreq nfsd auth_rpcgss nfs_acl lockd sunrpc btrfs
>>>> xor raid6_pq raid1 radeon i2c_algo_bit firewire_ohci drm_kms_helper
>>>> firewire_core ttm crc_itu_t mvsas drm l
>>>> ibsas scsi_transport_sas i2c_core cpufreq_stats bcache
>>>> Jul  6 00:26:32 mcp kernel: [30957.528183] CPU: 1 PID: 15568 Comm:
>>>> kworker/1:2 Not tainted 3.14.8-200.fc20.x86_64 #1
>>>> Jul  6 00:26:32 mcp kernel: [30957.537550] Hardware name: Gigabyte
>>>> Technology Co., Ltd. GA-890GPA-UD3H/GA-890GPA-UD3H, BIOS F9 09/09/2011
>>>> Jul  6 00:26:32 mcp kernel: binfmt_misc cfg80211 rfkill it87 hwmon_vid
>>>> nf_conntrack_ipv4 ip6t_REJECT nf_defrag_ipv4 nf_conntrack_ipv6
>>>> nf_defrag_ipv6 xt_conntrack ip6table_filter nf
>>>> _conntrack ip6_tables raid456 async_raid6_recov async_memcpy async_pq
>>>> async_xor async_tx kvm microcode edac_core serio_raw sp5100_tco
>>>> edac_mce_amd k10temp snd_hda_codec_realtek i2c
>>>> _piix4 snd_hda_codec_generic snd_hda_codec_hdmi snd_hda_intel
>>>> snd_hda_codec snd_hwdep snd_seq snd_seq_device snd_pcm snd_timer snd
>>>> r8169 soundcore mii shpchp wmi acpi_cpufreq nfsd
>>>> auth_rpcgss nfs_acl lockd sunrpc btrfs xor raid6_pq raid1 radeon
>>>> i2c_algo_bit firewire_ohci drm_kms_helper firewire_core ttm crc_itu_t
>>>> mvsas drm libsas scsi_transport_sas i2c_core
>>>> cpufreq_stats bcache
>>>> Jul  6 00:26:32 mcp kernel: [30957.548771] Workqueue: events
>>>> write_dirty_finish [bcache]
>>>> Jul  6 00:26:32 mcp kernel: CPU: 1 PID: 15568 Comm: kworker/1:2 Not
>>>> tainted 3.14.8-200.fc20.x86_64 #1
>>>> Jul  6 00:26:32 mcp kernel: [30957.555773]  0000000000000000
>>>> 00000000dea1a02f ffff880007ffd8b0 ffffffff816f0502
>>>> Jul  6 00:26:32 mcp kernel: Hardware name: Gigabyte Technology Co., Ltd.
>>>> GA-890GPA-UD3H/GA-890GPA-UD3H, BIOS F9 09/09/2011
>>>> Jul  6 00:26:32 mcp kernel: [30957.564855]  ffff880007ffd8f8
>>>> ffff880007ffd8e8 ffffffff8108a1cd ffff88040e6b7800
>>>> Jul  6 00:26:32 mcp kernel: Workqueue: events write_dirty_finish [bcache]
>>>> Jul  6 00:26:32 mcp kernel: 0000000000000000 00000000dea1a02f
>>>> ffff880007ffd8b0 ffffffff816f0502
>>>> Jul  6 00:26:32 mcp kernel: ffff880007ffd8f8 ffff880007ffd8e8
>>>> ffffffff8108a1cd ffff88040e6b7800
>>>> Jul  6 00:26:32 mcp kernel: [30957.573943]  ffffffffffffffe4
>>>> ffff880007ffdd58 ffff880007ffd980 0000000000000000
>>>> Jul  6 00:26:32 mcp kernel: [30957.583045] Call Trace:
>>>> Jul  6 00:26:32 mcp kernel: [30957.587031]  [<ffffffff816f0502>]
>>>> dump_stack+0x45/0x56
>>>> Jul  6 00:26:32 mcp kernel: ffffffffffffffe4 ffff880007ffdd58
>>>> ffff880007ffd980 0000000000000000
>>>> Jul  6 00:26:32 mcp kernel: Call Trace:
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff816f0502>] dump_stack+0x45/0x56
>>>> Jul  6 00:26:32 mcp kernel: [30957.593729]  [<ffffffff8108a1cd>]
>>>> warn_slowpath_common+0x7d/0xa0
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff8108a1cd>]
>>>> warn_slowpath_common+0x7d/0xa0
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff8108a24c>] warn_slowpath_fmt+0x5c/0x80
>>>> Jul  6 00:26:32 mcp kernel: [30957.601287]  [<ffffffff8108a24c>]
>>>> warn_slowpath_fmt+0x5c/0x80
>>>> Jul  6 00:26:32 mcp kernel: [30957.608568]  [<ffffffffa000965b>]
>>>> btree_split+0x5bb/0x630 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa000965b>] btree_split+0x5bb/0x630
>>>> [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [30957.616248]  [<ffffffff816f3d22>] ?
>>>> __schedule+0x2e2/0x740
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff816f3d22>] ? __schedule+0x2e2/0x740
>>>> Jul  6 00:26:32 mcp kernel: [30957.623240]  [<ffffffff8136decd>] ?
>>>> list_del+0xd/0x30
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff8136decd>] ? list_del+0xd/0x30
>>>> Jul  6 00:26:32 mcp kernel: [30957.629792]  [<ffffffffa0009771>]
>>>> bch_btree_insert_node+0xa1/0x2c0 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [30957.638293]  [<ffffffffa000a724>]
>>>> btree_insert_fn+0x24/0x50 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa0009771>]
>>>> bch_btree_insert_node+0xa1/0x2c0 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa000a724>]
>>>> btree_insert_fn+0x24/0x50 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa0007861>]
>>>> bch_btree_map_nodes_recurse+0x51/0x180 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [30957.646175]  [<ffffffffa0007861>]
>>>> bch_btree_map_nodes_recurse+0x51/0x180 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [30957.655171]  [<ffffffffa000a700>] ?
>>>> bch_btree_insert_check_key+0x1b0/0x1b0 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa000a700>] ?
>>>> bch_btree_insert_check_key+0x1b0/0x1b0 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa000da62>] ?
>>>> bch_btree_ptr_invalid+0x12/0x20 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [30957.664290]  [<ffffffffa000da62>] ?
>>>> bch_btree_ptr_invalid+0x12/0x20 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [30957.672814]  [<ffffffffa000c48d>] ?
>>>> bch_btree_ptr_bad+0x4d/0x110 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa000c48d>] ?
>>>> bch_btree_ptr_bad+0x4d/0x110 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [30957.681045]  [<ffffffff816f6592>] ?
>>>> down_write+0x12/0x30
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff816f6592>] ? down_write+0x12/0x30
>>>> Jul  6 00:26:32 mcp kernel: [30957.731889]  [<ffffffff810b2d66>] ?
>>>> srcu_notifier_call_chain+0x16/0x20
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff810b2d66>] ?
>>>> srcu_notifier_call_chain+0x16/0x20
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffffa000b1e1>]
>>>> bch_btree_insert+0xf1/0x170 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [30957.739905]  [<ffffffffa000b1e1>]
>>>> bch_btree_insert+0xf1/0x170 [bcache]
>>>> Jul  6 00:26:32 mcp kernel: [30957.747908]  [<ffffffff810d2180>] ?
>>>> abort_exclusive_wait+0xb0/0xb0
>>>> Jul  6 00:26:32 mcp kernel: [<ffffffff810d2180>] ?
>>>> abort_exclusive_wait+0xb0/0xb0
>>>> Jul  6 00:26:33 mcp kernel: [<ffffffffa0021d66>]
>>>> write_dirty_finish+0x1f6/0x260 [bcache]
>>>> Jul  6 00:26:33 mcp kernel: [30957.755602]  [<ffffffffa0021d66>]
>>>> write_dirty_finish+0x1f6/0x260 [bcache]
>>>> Jul  6 00:26:33 mcp kernel: [30957.763908]  [<ffffffff810c3e76>] ?
>>>> __dequeue_entity+0x26/0x40
>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810c3e76>] ?
>>>> __dequeue_entity+0x26/0x40
>>>> Jul  6 00:26:33 mcp kernel: [30957.771228]  [<ffffffff810135d6>] ?
>>>> __switch_to+0x126/0x4c0
>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810135d6>] ? __switch_to+0x126/0x4c0
>>>> Jul  6 00:26:33 mcp kernel: [30957.778283]  [<ffffffff810a68e6>]
>>>> process_one_work+0x176/0x430
>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810a68e6>]
>>>> process_one_work+0x176/0x430
>>>> Jul  6 00:26:33 mcp kernel: [30957.785598]  [<ffffffff810a757b>]
>>>> worker_thread+0x11b/0x3a0
>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810a757b>] worker_thread+0x11b/0x3a0
>>>> Jul  6 00:26:33 mcp kernel: [30957.792656]  [<ffffffff810a7460>] ?
>>>> rescuer_thread+0x3b0/0x3b0
>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810a7460>] ?
>>>> rescuer_thread+0x3b0/0x3b0
>>>> Jul  6 00:26:33 mcp kernel: [30957.799982]  [<ffffffff810ae2d1>]
>>>> kthread+0xe1/0x100
>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810ae2d1>] kthread+0xe1/0x100
>>>> Jul  6 00:26:33 mcp kernel: [30957.806427]  [<ffffffff810ae1f0>] ?
>>>> insert_kthread_work+0x40/0x40
>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810ae1f0>] ?
>>>> insert_kthread_work+0x40/0x40
>>>> Jul  6 00:26:33 mcp kernel: [30957.813995]  [<ffffffff8170083c>]
>>>> ret_from_fork+0x7c/0xb0
>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff8170083c>] ret_from_fork+0x7c/0xb0
>>>> Jul  6 00:26:33 mcp kernel: [30957.820844]  [<ffffffff810ae1f0>] ?
>>>> insert_kthread_work+0x40/0x40
>>>> Jul  6 00:26:33 mcp kernel: [<ffffffff810ae1f0>] ?
>>>> insert_kthread_work+0x40/0x40
>>>> Jul  6 00:26:33 mcp kernel: [30957.828357] ---[ end trace
>>>> 874ec8b4276a8f33 ]---
>>>> Jul  6 00:26:33 mcp kernel: ---[ end trace 874ec8b4276a8f33 ]---
>>>> Jul  6 00:26:33 mcp kernel: [30957.834379] bcache: bch_btree_insert()
>>>> error -12
>>>> Jul  6 00:26:33 mcp kernel: bcache: bch_btree_insert() error -12
>>>>
>>>> --Larkin
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe linux-bcache" in
>>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> --
>> To unsubscribe from this list: send the line "unsubscribe linux-bcache" in
>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>
--
To unsubscribe from this list: send the line "unsubscribe linux-bcache" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[Index of Archives]     [Linux ARM Kernel]     [Linux Filesystem Development]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [ECOS]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux