Re: Possible deadlock condition

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



None of the OSDs seem to be more than 82% full. I didn't think we were
running quite that close to the margin, but it is still far from
actually full.


On Mon, Jun 18, 2012 at 3:57 PM, Dan Mick <dan.mick@xxxxxxxxxxx> wrote:
> Does the xfs on the OSD have plenty of free space left, or could this be an
> allocation deadlock?
>
>
> On 06/18/2012 03:17 PM, Mandell Degerness wrote:
>>
>> Here is, perhaps, a more useful traceback from a different run of
>> tests that we just ran into:
>>
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.680815] INFO: task
>> flush-254:0:29582 blocked for more than 120 seconds.
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.681040] "echo 0>
>> /proc/sys/kernel/hung_task_timeout_secs" disables this message.
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.681458] flush-254:0
>>   D ffff880bd9ca2fc0     0 29582      2 0x00000000
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.681740]
>> ffff88006e51d160 0000000000000046 0000000000000002 ffff88061b362040
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.682173]
>> ffff88006e51d160 00000000000120c0 00000000000120c0 00000000000120c0
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.682659]
>> ffff88006e51dfd8 00000000000120c0 00000000000120c0 ffff88006e51dfd8
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.683088] Call Trace:
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.683302]
>> [<ffffffff81520132>] schedule+0x5a/0x5c
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.683514]
>> [<ffffffff815203e7>] schedule_timeout+0x36/0xe3
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.683784]
>> [<ffffffff8101e0b2>] ? physflat_send_IPI_mask+0xe/0x10
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.683999]
>> [<ffffffff8101a237>] ? native_smp_send_reschedule+0x46/0x48
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.684219]
>> [<ffffffff811e0071>] ? list_move_tail+0x27/0x2c
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.684432]
>> [<ffffffff81520d13>] __down_common+0x90/0xd4
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.684708]
>> [<ffffffff811e1120>] ? _xfs_buf_find+0x17f/0x210
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.684925]
>> [<ffffffff81520dca>] __down+0x1d/0x1f
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.685139]
>> [<ffffffff8105db4e>] down+0x2d/0x3d
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.685350]
>> [<ffffffff811e0f68>] xfs_buf_lock+0x76/0xaf
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.685565]
>> [<ffffffff811e1120>] _xfs_buf_find+0x17f/0x210
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.685836]
>> [<ffffffff811e13b6>] xfs_buf_get+0x2a/0x177
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.686052]
>> [<ffffffff811e19f6>] xfs_buf_read+0x1f/0xca
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.686270]
>> [<ffffffff8122a0b7>] xfs_trans_read_buf+0x205/0x308
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.686490]
>> [<ffffffff81205e01>] xfs_btree_read_buf_block.clone.22+0x4f/0xa7
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.687015]
>> [<ffffffff8122a3ee>] ? xfs_trans_log_buf+0xb2/0xc1
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.687232]
>> [<ffffffff81205edd>] xfs_btree_lookup_get_block+0x84/0xac
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.687449]
>> [<ffffffff81208e83>] xfs_btree_lookup+0x12b/0x3dc
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.687721]
>> [<ffffffff811f6bb2>] ? xfs_alloc_vextent+0x447/0x469
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.687939]
>> [<ffffffff811fd171>] xfs_bmbt_lookup_eq+0x1f/0x21
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.688156]
>> [<ffffffff811ffa88>] xfs_bmap_add_extent_delay_real+0x5b5/0xfec
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.688378]
>> [<ffffffff810f155b>] ? kmem_cache_alloc+0x87/0xf3
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.688650]
>> [<ffffffff81204c40>] ? xfs_bmbt_init_cursor+0x3f/0x107
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.688867]
>> [<ffffffff81201160>] xfs_bmapi_allocate+0x1f6/0x23a
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.689084]
>> [<ffffffff812185bd>] ? xfs_iext_bno_to_irec+0x95/0xb9
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.689301]
>> [<ffffffff81203414>] xfs_bmapi_write+0x32d/0x5a2
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.689519]
>> [<ffffffff811e99e4>] xfs_iomap_write_allocate+0x1a5/0x29f
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.689797]
>> [<ffffffff811df12a>] xfs_map_blocks+0x13e/0x1dd
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.690016]
>> [<ffffffff811dfbff>] xfs_vm_writepage+0x24e/0x410
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.690233]
>> [<ffffffff810bde1e>] __writepage+0x17/0x30
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.690446]
>> [<ffffffff810be6ed>] write_cache_pages+0x276/0x3c8
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.690693]
>> [<ffffffff810bde07>] ? set_page_dirty+0x60/0x60
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.690908]
>> [<ffffffff810be884>] generic_writepages+0x45/0x5c
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.691123]
>> [<ffffffff811defcb>] xfs_vm_writepages+0x4d/0x54
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.691337]
>> [<ffffffff810bf832>] do_writepages+0x21/0x2a
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.691552]
>> [<ffffffff811218f5>] writeback_single_inode+0x12a/0x2cc
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.691800]
>> [<ffffffff81121d92>] writeback_sb_inodes+0x174/0x215
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.692016]
>> [<ffffffff81122185>] __writeback_inodes_wb+0x78/0xb9
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.692231]
>> [<ffffffff811224b5>] wb_writeback+0x136/0x22a
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.692444]
>> [<ffffffff810becd1>] ? determine_dirtyable_memory+0x1d/0x26
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.692692]
>> [<ffffffff81122d1e>] wb_do_writeback+0x19c/0x1b7
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.692907]
>> [<ffffffff81122dc5>] bdi_writeback_thread+0x8c/0x20f
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.693122]
>> [<ffffffff81122d39>] ? wb_do_writeback+0x1b7/0x1b7
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.693336]
>> [<ffffffff81122d39>] ? wb_do_writeback+0x1b7/0x1b7
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.693553]
>> [<ffffffff8105911d>] kthread+0x82/0x8a
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.693803]
>> [<ffffffff81523c34>] kernel_thread_helper+0x4/0x10
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.694018]
>> [<ffffffff8105909b>] ? kthread_worker_fn+0x13b/0x13b
>> Jun 18 17:58:51 node-172-29-0-15 kernel: [242522.694232]
>> [<ffffffff81523c30>] ? gs_change+0xb/0xb
>>
>>
>> On Mon, Jun 18, 2012 at 11:37 AM, Mandell Degerness
>> <mandell@xxxxxxxxxxxxxxx>  wrote:
>>>
>>> We've been seeing random issues of apparent deadlocks.  We are running
>>> ceph 0.47 on kernel 3.2.18.  OSDs are running on XFS file system.
>>> mysqld (which ran into the particular problems in the attached kernel
>>> log) is running on an RBD with XFS (mounted on a system which includes
>>> OSDs).  We have sync_fs, and gcc ver 4.5.3-r2.  The mysqld process in
>>> both instances returned an error to the calling process.
>>>
>>> Regards,
>>> Mandell Degerness
>>
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux