Re: Possible deadlock condition

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, 20 Jun 2012, Mandell Degerness wrote:
> The prior thread seems to refer to something fixed in 3.0.X, we are
> running 3.2.18.  Also, in answer to the previous question, we see the
> error on systems running at 82% full and systems running at 5% full
> disks.
> 
> Anyone have any ideas about how to resolve the deadlock?  Do we have
> to configure mysql differently?

I'm not sure that this is related to the previous problems.  I would start 
from square one to diagnose:

 - When you see the hang, are there blocked osd I/O operations?

	cat /sys/kernel/debug/ceph/*/osdc

 - Maybe this is a memory deadlock?  I'm guessing not (if you can 
   log into the system), but it's worth checking.  I think you should see 
   another blocked task in that case.

 - This might be an XFS thing unrelated to the fact that we're running on 
   RBD.  If it's reproducible, can you try on ext4 instead?


sage

> 
> -Mandell
> On Mon, Jun 18, 2012 at 4:34 PM, Dan Mick <dan.mick@xxxxxxxxxxx> wrote:
> > I don't know enough to know if there's a connection, but I do note this
> > prior thread that sounds kinda similar:
> >
> > http://thread.gmane.org/gmane.comp.file-systems.ceph.devel/6574
> >
> >
> >
> > On 06/18/2012 04:08 PM, Mandell Degerness wrote:
> >>
> >> None of the OSDs seem to be more than 82% full. I didn't think we were
> >> running quite that close to the margin, but it is still far from
> >> actually full.
> >>
> >>
> >> On Mon, Jun 18, 2012 at 3:57 PM, Dan Mick<dan.mick@xxxxxxxxxxx>  wrote:
> >>>
> >>> Does the xfs on the OSD have plenty of free space left, or could this be
> >>> an
> >>> allocation deadlock?
> >>>
> >>>
> >>> On 06/18/2012 03:17 PM, Mandell Degerness wrote:
> >>>>
> >>>>
> >>>> Here is, perhaps, a more useful traceback from a different run of
> >>>> tests that we just ran into:
> >>>>
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.680815] INFO: task
> >>>> flush-254:0:29582 blocked for more than 120 seconds.
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.681040] "echo 0>
> >>>> /proc/sys/kernel/hung_task_timeout_secs" disables this message.
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.681458] flush-254:0
> >>>>   D ffff880bd9ca2fc0     0 29582      2 0x00000000
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.681740]
> >>>> ffff88006e51d160 0000000000000046 0000000000000002 ffff88061b362040
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.682173]
> >>>> ffff88006e51d160 00000000000120c0 00000000000120c0 00000000000120c0
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.682659]
> >>>> ffff88006e51dfd8 00000000000120c0 00000000000120c0 ffff88006e51dfd8
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.683088] Call Trace:
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.683302]
> >>>> [<ffffffff81520132>] schedule+0x5a/0x5c
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.683514]
> >>>> [<ffffffff815203e7>] schedule_timeout+0x36/0xe3
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.683784]
> >>>> [<ffffffff8101e0b2>] ? physflat_send_IPI_mask+0xe/0x10
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.683999]
> >>>> [<ffffffff8101a237>] ? native_smp_send_reschedule+0x46/0x48
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.684219]
> >>>> [<ffffffff811e0071>] ? list_move_tail+0x27/0x2c
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.684432]
> >>>> [<ffffffff81520d13>] __down_common+0x90/0xd4
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.684708]
> >>>> [<ffffffff811e1120>] ? _xfs_buf_find+0x17f/0x210
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.684925]
> >>>> [<ffffffff81520dca>] __down+0x1d/0x1f
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.685139]
> >>>> [<ffffffff8105db4e>] down+0x2d/0x3d
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.685350]
> >>>> [<ffffffff811e0f68>] xfs_buf_lock+0x76/0xaf
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.685565]
> >>>> [<ffffffff811e1120>] _xfs_buf_find+0x17f/0x210
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.685836]
> >>>> [<ffffffff811e13b6>] xfs_buf_get+0x2a/0x177
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.686052]
> >>>> [<ffffffff811e19f6>] xfs_buf_read+0x1f/0xca
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.686270]
> >>>> [<ffffffff8122a0b7>] xfs_trans_read_buf+0x205/0x308
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.686490]
> >>>> [<ffffffff81205e01>] xfs_btree_read_buf_block.clone.22+0x4f/0xa7
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.687015]
> >>>> [<ffffffff8122a3ee>] ? xfs_trans_log_buf+0xb2/0xc1
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.687232]
> >>>> [<ffffffff81205edd>] xfs_btree_lookup_get_block+0x84/0xac
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.687449]
> >>>> [<ffffffff81208e83>] xfs_btree_lookup+0x12b/0x3dc
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.687721]
> >>>> [<ffffffff811f6bb2>] ? xfs_alloc_vextent+0x447/0x469
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.687939]
> >>>> [<ffffffff811fd171>] xfs_bmbt_lookup_eq+0x1f/0x21
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.688156]
> >>>> [<ffffffff811ffa88>] xfs_bmap_add_extent_delay_real+0x5b5/0xfec
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.688378]
> >>>> [<ffffffff810f155b>] ? kmem_cache_alloc+0x87/0xf3
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.688650]
> >>>> [<ffffffff81204c40>] ? xfs_bmbt_init_cursor+0x3f/0x107
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.688867]
> >>>> [<ffffffff81201160>] xfs_bmapi_allocate+0x1f6/0x23a
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.689084]
> >>>> [<ffffffff812185bd>] ? xfs_iext_bno_to_irec+0x95/0xb9
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.689301]
> >>>> [<ffffffff81203414>] xfs_bmapi_write+0x32d/0x5a2
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.689519]
> >>>> [<ffffffff811e99e4>] xfs_iomap_write_allocate+0x1a5/0x29f
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.689797]
> >>>> [<ffffffff811df12a>] xfs_map_blocks+0x13e/0x1dd
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.690016]
> >>>> [<ffffffff811dfbff>] xfs_vm_writepage+0x24e/0x410
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.690233]
> >>>> [<ffffffff810bde1e>] __writepage+0x17/0x30
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.690446]
> >>>> [<ffffffff810be6ed>] write_cache_pages+0x276/0x3c8
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.690693]
> >>>> [<ffffffff810bde07>] ? set_page_dirty+0x60/0x60
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.690908]
> >>>> [<ffffffff810be884>] generic_writepages+0x45/0x5c
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.691123]
> >>>> [<ffffffff811defcb>] xfs_vm_writepages+0x4d/0x54
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.691337]
> >>>> [<ffffffff810bf832>] do_writepages+0x21/0x2a
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.691552]
> >>>> [<ffffffff811218f5>] writeback_single_inode+0x12a/0x2cc
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.691800]
> >>>> [<ffffffff81121d92>] writeback_sb_inodes+0x174/0x215
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.692016]
> >>>> [<ffffffff81122185>] __writeback_inodes_wb+0x78/0xb9
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.692231]
> >>>> [<ffffffff811224b5>] wb_writeback+0x136/0x22a
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.692444]
> >>>> [<ffffffff810becd1>] ? determine_dirtyable_memory+0x1d/0x26
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.692692]
> >>>> [<ffffffff81122d1e>] wb_do_writeback+0x19c/0x1b7
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.692907]
> >>>> [<ffffffff81122dc5>] bdi_writeback_thread+0x8c/0x20f
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.693122]
> >>>> [<ffffffff81122d39>] ? wb_do_writeback+0x1b7/0x1b7
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.693336]
> >>>> [<ffffffff81122d39>] ? wb_do_writeback+0x1b7/0x1b7
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.693553]
> >>>> [<ffffffff8105911d>] kthread+0x82/0x8a
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.693803]
> >>>> [<ffffffff81523c34>] kernel_thread_helper+0x4/0x10
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.694018]
> >>>> [<ffffffff8105909b>] ? kthread_worker_fn+0x13b/0x13b
> >>>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.694232]
> >>>> [<ffffffff81523c30>] ? gs_change+0xb/0xb
> >>>>
> >>>>
> >>>> On Mon, Jun 18, 2012 at 11:37 AM, Mandell Degerness
> >>>> <mandell@xxxxxxxxxxxxxxx>    wrote:
> >>>>>
> >>>>>
> >>>>> We've been seeing random issues of apparent deadlocks.  We are running
> >>>>> ceph 0.47 on kernel 3.2.18.  OSDs are running on XFS file system.
> >>>>> mysqld (which ran into the particular problems in the attached kernel
> >>>>> log) is running on an RBD with XFS (mounted on a system which includes
> >>>>> OSDs).  We have sync_fs, and gcc ver 4.5.3-r2.  The mysqld process in
> >>>>> both instances returned an error to the calling process.
> >>>>>
> >>>>> Regards,
> >>>>> Mandell Degerness
> >>>>
> >>>>
> >>>> --
> >>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> >>>> the body of a message to majordomo@xxxxxxxxxxxxxxx
> >>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 

[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux