Kernel BUG using RBD module

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello folks,

Ran into this one today on a machine using kRBD.  Completely locked the machine:

Apr  1 17:35:10 nfs1 kernel: [492418.251665] Call Trace:
Apr  1 17:35:10 nfs1 kernel: [492418.251675]  [<ffffffffa02086b4>]
reset_changed_osds+0x74/0xa0 [libceph]
Apr  1 17:35:10 nfs1 kernel: [492418.251682]  [<ffffffffa020b0c2>]
ceph_osdc_handle_map+0x212/0x3e0 [libceph]
Apr  1 17:35:10 nfs1 kernel: [492418.251689]  [<ffffffffa0207547>]
dispatch+0xa7/0x120 [libceph]
Apr  1 17:35:10 nfs1 kernel: [492418.251694]  [<ffffffffa0201045>]
process_message+0xa5/0xc0 [libceph]
Apr  1 17:35:10 nfs1 kernel: [492418.251700]  [<ffffffffa0204fa1>]
try_read+0x2e1/0x440 [libceph]
Apr  1 17:35:10 nfs1 kernel: [492418.251705]  [<ffffffffa0205100>] ?
try_read+0x440/0x440 [libceph]
Apr  1 17:35:10 nfs1 kernel: [492418.251711]  [<ffffffffa0205192>]
con_work+0x92/0x1c0 [libceph]
Apr  1 17:35:10 nfs1 kernel: [492418.251716]  [<ffffffff81071baa>]
process_one_work+0x11a/0x480
Apr  1 17:35:10 nfs1 kernel: [492418.251720]  [<ffffffff81072bc5>]
worker_thread+0x165/0x370
Apr  1 17:35:10 nfs1 kernel: [492418.251723]  [<ffffffff81072a60>] ?
manage_workers.isra.29+0x130/0x130
Apr  1 17:35:10 nfs1 kernel: [492418.251727]  [<ffffffff81077b63>]
kthread+0x93/0xa0
Apr  1 17:35:10 nfs1 kernel: [492418.251733]  [<ffffffff816a3ee4>]
kernel_thread_helper+0x4/0x10
Apr  1 17:35:10 nfs1 kernel: [492418.251736]  [<ffffffff81077ad0>] ?
flush_kthread_worker+0xb0/0xb0
Apr  1 17:35:10 nfs1 kernel: [492418.251739]  [<ffffffff816a3ee0>] ?
gs_change+0x13/0x13


Anybody run into this before?

I'm running kernel 3.5.7.2

I came across a few patches that touch relevant functions:
https://patchwork.kernel.org/patch/1913871/

>From what i can tell, this can pulled into 3.4.26 and 3.7.3.  I can't
tell if it went into 3.5.x at all.

Is this a known bug that's already been fixed?

This happened while I was doing a rolling upgrade of the Ceph cluster
-- doing a "service ceph restart osd.$x" for 60 OSDs.  30 second sleep
in between each to allow things to settle.

 - Travis
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux