Re: linux 4.7.0 rbd client kernel panic when OSD process was killed by OOM

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, Aug 8, 2016 at 9:57 PM, Victor Payno <vpayno@xxxxxxxxxx> wrote:
> We have another problem where an RBD client was killed when an OSD was
> killed by the OOM on a server. The servers have 4.4.16 kernels.
>
> ams2 login: [789881.620147] ------------[ cut here ]------------
> [789881.625094] kernel BUG at drivers/block/rbd.c:4638!
> [789881.630311] invalid opcode: 0000 [#1] SMP
> [789881.634650] Modules linked in: rbd libceph sg rpcsec_gss_krb5
> xt_nat xt_UDPLB(O) xt_multiport xt_addrtype iptable_mangle iptable_raw
> iptable_nat nf_nat_ipv4 nf_nat ext4 jbd2 mbcache x86_pkg_temp_thermal
> gkuart(O) usbserial ie31200_edac edac_core tpm_tis raid1 crc32c_intel
> [789881.661718] CPU: 4 PID: 4111 Comm: kworker/u16:0 Tainted: G
>    O    4.7.0-vanilla-ams-3 #1
> [789881.671091] Hardware name: Quanta T6BC-S1N/T6BC, BIOS T6BC2A01 03/26/2014
> [789881.678212] Workqueue: ceph-watch-notify do_watch_notify [libceph]
> [789881.684814] task: ffff88032069ea00 ti: ffff8803f0c90000 task.ti:
> ffff8803f0c90000
> [789881.692802] RIP: 0010:[<ffffffffa016d1c9>]  [<ffffffffa016d1c9>]
> rbd_dev_header_info+0x5a9/0x940 [rbd]
> [789881.702702] RSP: 0018:ffff8803f0c93d30  EFLAGS: 00010286
> [789881.708344] RAX: 0000000000000077 RBX: ffff8802a6a63800 RCX:
> 0000000000000000
> [789881.715985] RDX: 0000000000000077 RSI: ffff88041fd0dd08 RDI:
> ffff88041fd0dd08
> [789881.723625] RBP: ffff8803f0c93d98 R08: 0000000000000030 R09:
> 0000000000000000
> [789881.731261] R10: 0000000000000000 R11: 0000000000004479 R12:
> ffff8800d6eaf000
> [789881.738899] R13: ffff8802a6a639b0 R14: 0000000000000000 R15:
> ffff880327e6e780
> [789881.746533] FS:  0000000000000000(0000) GS:ffff88041fd00000(0000)
> knlGS:0000000000000000
> [789881.755120] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> [789881.761197] CR2: 00007fbb18242838 CR3: 0000000001e07000 CR4:
> 00000000001406e0
> [789881.768846] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
> 0000000000000000
> [789881.776482] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7:
> 0000000000000400
> [789881.784118] Stack:
> [789881.786457]  ffffffff8113a91a ffff88032069ea00 ffff88041fd17ef0
> ffff88041fd17ef0
> [789881.794713]  ffff88041fd17ef0 0000000000030289 ffff8803f0c93dd8
> ffffffff8113d968
> [789881.802965]  ffff8802a6a63800 ffff8800d6eaf000 ffff8802a6a639b0
> 0000000000000000
> [789881.811207] Call Trace:
> [789881.813988]  [<ffffffff8113a91a>] ? update_curr+0x8a/0x110
> [789881.819810]  [<ffffffff8113d968>] ? dequeue_task_fair+0x618/0x1150
> [789881.826321]  [<ffffffffa016d591>] rbd_dev_refresh+0x31/0xf0 [rbd]
> [789881.832760]  [<ffffffffa016d719>] rbd_watch_cb+0x29/0xa0 [rbd]
> [789881.838930]  [<ffffffffa0138fdc>] do_watch_notify+0x4c/0x80 [libceph]
> [789881.845706]  [<ffffffff811258e9>] process_one_work+0x149/0x3c0
> [789881.856639]  [<ffffffff81125bae>] worker_thread+0x4e/0x490
> [789881.862453]  [<ffffffff81125b60>] ? process_one_work+0x3c0/0x3c0
> [789881.868823]  [<ffffffff8112b1e9>] kthread+0xc9/0xe0
> [789881.874033]  [<ffffffff8185e4ff>] ret_from_fork+0x1f/0x40
> [789881.879764]  [<ffffffff8112b120>] ? kthread_create_on_node+0x170/0x170
> [789881.886618] Code: 0b 44 8b 6d b8 e9 1d ff ff ff 48 c7 c1 f0 00 17
> a0 ba 1e 12 00 00 48 c7 c6 90 0e 17 a0 48 c7 c7 20 f8 16 a0 31 c0 e8
> 8a 5d 08 e1 <0f> 0b 75 14 49 8b 7f 68 41 bd 92 ff ff ff e8 d4 e0 fc ff
> e9 dc
> [789881.911744] RIP  [<ffffffffa016d1c9>] rbd_dev_header_info+0x5a9/0x940 [rbd]
> [789881.919116]  RSP <ffff8803f0c93d30>
> [789881.922989] ---[ end trace 12b8d1c2ed74d6c1 ]---
> [789881.927971] BUG: unable to handle kernel paging request at ffffffffffffffd8
> [789881.935435] IP: [<ffffffff8112b821>] kthread_data+0x11/0x20
> [789881.941427] PGD 1e0a067 PUD 1e0c067 PMD 0
> [789881.946117] Oops: 0000 [#2] SMP
> [789881.949591] Modules linked in: rbd libceph sg rpcsec_gss_krb5
> xt_nat xt_UDPLB(O) xt_multiport xt_addrtype iptable_mangle iptable_raw
> iptable_nat nf_nat_ipv4 nf_nat ext4 jbd2 mbcache x86_pkg_temp_thermal
> gkuart(O) usbserial ie31200_edac edac_core tpm_tis raid1 crc32c_intel
> [789881.976900] CPU: 4 PID: 4111 Comm: kworker/u16:0 Tainted: G      D
>    O    4.7.0-vanilla-ams-3 #1
> [789881.986280] Hardware name: Quanta T6BC-S1N/T6BC, BIOS T6BC2A01 03/26/2014
> [789881.993410] task: ffff88032069ea00 ti: ffff8803f0c90000 task.ti:
> ffff8803f0c90000
> [789882.001411] RIP: 0010:[<ffffffff8112b821>]  [<ffffffff8112b821>]
> kthread_data+0x11/0x20
> [789882.010024] RSP: 0018:ffff8803f0c93a28  EFLAGS: 00010002
> [789882.015682] RAX: 0000000000000000 RBX: ffff88041fd17e80 RCX:
> 0000000000000004
> [789882.023342] RDX: ffff88040f005000 RSI: ffff88032069ea00 RDI:
> ffff88032069ea00
> [789882.030996] RBP: ffff8803f0c93a30 R08: 0000000000000000 R09:
> 0000000000079800
> [789882.038645] R10: 0000000000000001 R11: 0000000000000001 R12:
> 0000000000017e80
> [789882.046288] R13: 0000000000000000 R14: ffff88032069eec0 R15:
> ffff88032069ea00
> [789882.053926] FS:  0000000000000000(0000) GS:ffff88041fd00000(0000)
> knlGS:0000000000000000
> [789882.062524] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> [789882.068603] CR2: 0000000000000028 CR3: 0000000001e07000 CR4:
> 00000000001406e0
> [789882.076261] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
> 0000000000000000
> [789882.083920] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7:
> 0000000000000400
> [789882.091577] Stack:
> [789882.093935]  ffffffff8112645e ffff8803f0c93a78 ffffffff8185ab3e
> ffff88032069ea00
> [789882.102194]  ffff8803f0c93a78 ffff8803f0c94000 ffff8803f0c93ad0
> ffff8803f0c936e8
> [789882.110438]  ffff88040d5c8000 0000000000000000 ffff8803f0c93a90
> ffffffff8185aef5
> [789882.118689] Call Trace:
> [789882.121484]  [<ffffffff8112645e>] ? wq_worker_sleeping+0xe/0x90
> [789882.127752]  [<ffffffff8185ab3e>] __schedule+0x36e/0x6f0
> [789882.133411]  [<ffffffff8185aef5>] schedule+0x35/0x80
> [789882.138712]  [<ffffffff81110ff9>] do_exit+0x739/0xb50
> [789882.144098]  [<ffffffff8108833c>] oops_end+0x9c/0xd0
> [789882.149400]  [<ffffffff810887ab>] die+0x4b/0x70
> [789882.154276]  [<ffffffff81085b26>] do_trap+0xb6/0x150
> [789882.159583]  [<ffffffff81085d87>] do_error_trap+0x77/0xe0
> [789882.165322]  [<ffffffffa016d1c9>] ? rbd_dev_header_info+0x5a9/0x940 [rbd]
> [789882.172446]  [<ffffffff811d7a3d>] ? irq_work_queue+0x6d/0x80
> [789882.178441]  [<ffffffff811575d4>] ? wake_up_klogd+0x34/0x40
> [789882.184363]  [<ffffffff81157aa6>] ? console_unlock+0x4c6/0x510
> [789882.190532]  [<ffffffff810863c0>] do_invalid_op+0x20/0x30
> [789882.196265]  [<ffffffff8185fb6e>] invalid_op+0x1e/0x30
> [789882.201740]  [<ffffffffa016d1c9>] ? rbd_dev_header_info+0x5a9/0x940 [rbd]
> [789882.208866]  [<ffffffff8113a91a>] ? update_curr+0x8a/0x110
> [789882.214694]  [<ffffffff8113d968>] ? dequeue_task_fair+0x618/0x1150
> [789882.221225]  [<ffffffffa016d591>] rbd_dev_refresh+0x31/0xf0 [rbd]
> [789882.227662]  [<ffffffffa016d719>] rbd_watch_cb+0x29/0xa0 [rbd]
> [789882.233855]  [<ffffffffa0138fdc>] do_watch_notify+0x4c/0x80 [libceph]
> [789882.240647]  [<ffffffff811258e9>] process_one_work+0x149/0x3c0
> [789882.246811]  [<ffffffff81125bae>] worker_thread+0x4e/0x490
> [789882.252629]  [<ffffffff81125b60>] ? process_one_work+0x3c0/0x3c0
> [789882.258969]  [<ffffffff8112b1e9>] kthread+0xc9/0xe0
> [789882.264182]  [<ffffffff8185e4ff>] ret_from_fork+0x1f/0x40
> [789882.269917]  [<ffffffff8112b120>] ? kthread_create_on_node+0x170/0x170
> [789882.276784] Code: 02 00 00 00 e8 a1 fd ff ff 5d c3 0f 1f 44 00 00
> 66 2e 0f 1f 84 00 00 00 00 00 0f 1f 44 00 00 48 8b 87 60 04 00 00 55
> 48 89 e5 5d <48> 8b 40 d8 c3 66 2e 0f 1f 84 00 00 00 00 00 0f 1f 44 00
> 00 55
> [789882.301941] RIP  [<ffffffff8112b821>] kthread_data+0x11/0x20
> [789882.308016]  RSP <ffff8803f0c93a28>
> [789882.311847] CR2: ffffffffffffffd8
> [789882.315505] ---[ end trace 12b8d1c2ed74d6c2 ]---
> [789882.320462] Fixing recursive fault but reboot is needed!

That's the same one you've reported in the "ceph osd kernel divide
error", right?  I've filed http://tracker.ceph.com/issues/16963 and
should get to it later this week.

What did you mean by "no networking stack" in that thread?

Thanks,

                Ilya
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux