On Tue, Jan 23, 2018 at 06:22:07PM +0200, Avi Kivity wrote: > > > On 01/23/2018 06:11 PM, Brian Foster wrote: > > On Tue, Jan 23, 2018 at 05:45:39PM +0200, Avi Kivity wrote: > > > > > > On 01/23/2018 05:28 PM, Brian Foster wrote: > > > > On Tue, Jan 23, 2018 at 04:57:03PM +0200, Avi Kivity wrote: > > > > > I'm seeing the equivalent[*] of xfs_extent_busy_flush() sleeping in my > > > > > beautiful io_submit() calls. > > > > > > > > > > > > > > > Questions: > > > > > > > > > > - Is it correct that RWF_NOWAIT will not detect the condition that led to > > > > > the log being forced? > > > > > > > > > > - If so, can it be fixed? > > > > > > > > > > - Can I do something to reduce the odds of this occurring? larger logs, > > > > > more logs, flush more often, resurrect extinct species and sacrifice them to > > > > > the xfs gods? > > > > > > > > > > - Can an xfs developer do something? For example, make it RWF_NOWAIT > > > > > friendly (if the answer to the first question was "correct") > > > > > > > > > So RWF_NOWAIT eventually works its way to IOMAP_NOWAIT, which looks like > > > > it skips any write call that would require allocation in > > > > xfs_file_iomap_begin(). The busy flush should only happen in the block > > > > allocation path, so something is missing here. Do you have a backtrace > > > > for the log force you're seeing? > > > > > > > > > > > Here's a trace. It's from a kernel that lacks RWF_NOWAIT. > > > > > Oh, so the case below is roughly how I would have expected to hit the > > flush/wait without RWF_NOWAIT. The latter flag should prevent this, to > > answer your first question. > > Thanks, that's very encouraging. We are exploring recommending upstream-ish > kernels to users and customers, given their relative stability these days > and aio-related improvements (not to mention the shame of having to admit to > running an old kernel when reporting a problem to an upstream list). > > > > > For the follow up question, I think this should only occur when the fs > > is fairly low on free space. Is that the case here? > > No: > > /dev/md0 3.0T 1.2T 1.8T 40% /var/lib/scylla > > > > I'm not sure there's > > a specific metric, fwiw, but it's just a matter of attempting an (user > > data) allocation that only finds busy extents in the free space btrees > > and thus has to the force the log to satisfy the allocation. > > What does "busy" mean here? recently freed so we want to force the log to > make sure the extent isn't doubly-allocated? (wild guess) > Recently freed and the transaction that freed the blocks has not yet been persisted to the on-disk log. A subsequent attempt to allocate those blocks for user data waits for the transaction to commit to disk to ensure that the block is not written before the filesystem has persisted the fact that it has been freed. Otherwise, my understanding is that if the blocks are written to and the filesystem crashes before the previous free was persisted, we'd have allowed an overwrite of a still-used metadata block. > > I suppose > > running with more free space available would avoid this. I think running > > with less in-core log space could indirectly reduce extent busy time, > > but that may also have other performance ramifications and so is > > probably not a great idea. > > At 60%, I hope low free space is not a problem. > Yeah, that seems strange. I wouldn't expect busy extents to be a problem with that much free space. > btw, I'm also seeing 10ms+ periods of high CPU utilization: > > 0xffffffff816ab97a : _cond_resched+0x3a/0x50 [kernel] > 0xffffffff811e1495 : kmem_cache_alloc+0x35/0x1e0 [kernel] > 0xffffffffc00d8477 : kmem_zone_alloc+0x97/0x130 [xfs] > 0xffffffffc00deae2 : xfs_buf_item_init+0x42/0x190 [xfs] > 0xffffffffc00e89c3 : _xfs_trans_bjoin+0x23/0x60 [xfs] > 0xffffffffc00e8f17 : xfs_trans_read_buf_map+0x247/0x400 [xfs] > 0xffffffffc008f248 : xfs_btree_read_buf_block.constprop.29+0x78/0xc0 [xfs] > 0xffffffffc009221e : xfs_btree_increment+0x21e/0x350 [xfs] > 0xffffffffc00796a8 : xfs_alloc_ag_vextent_near+0x368/0xab0 [xfs] > 0xffffffffc0079efd : xfs_alloc_ag_vextent+0x10d/0x150 [xfs] > 0xffffffffc007abc6 : xfs_alloc_vextent+0x446/0x5f0 [xfs] > 0xffffffffc008b123 : xfs_bmap_btalloc+0x3f3/0x780 [xfs] > 0xffffffffc008b4be : xfs_bmap_alloc+0xe/0x10 [xfs] > > Is it normal for xfs to spend 10ms+ of CPU time to allocate an extent? > Should I be increasing my extent hint (currently at 32MB)? > I haven't done enough performance testing to have an intuition on the typical CPU time required to allocate blocks. Somebody else may be able to chime in on that. I suppose it could depend on the level of free space fragmentation, which can be observed via 'xfs_db -c "freesp -s" <dev>', whether I/Os or btree splits/joins were required, etc. FWIW, the above stack looks like it's stuck waiting on a memory allocation for a btree buffer xfs_buf_log_item, which is an internal data structure used to track metadata objects through the log subsystem. We have a kmem zone for such objects because they are allocated/freed frequently, but perhaps the zone had to grow..? We do pass KM_SLEEP there.. Brian > > > > Brian > > > > > 0xffffffff816ab231 : __schedule+0x531/0x9b0 [kernel] > > > 0xffffffff816ab6d9 : schedule+0x29/0x70 [kernel] > > > 0xffffffff816a90e9 : schedule_timeout+0x239/0x2c0 [kernel] > > > 0xffffffff816aba8d : wait_for_completion+0xfd/0x140 [kernel] > > > 0xffffffff810ab41d : flush_work+0xfd/0x190 [kernel] > > > 0xffffffffc00ddb3a : xlog_cil_force_lsn+0x8a/0x210 [xfs] > > > 0xffffffffc00dbbf5 : _xfs_log_force+0x85/0x2c0 [xfs] > > > 0xffffffffc00dbe5c : xfs_log_force+0x2c/0x70 [xfs] > > > 0xffffffffc0078f60 : xfs_alloc_ag_vextent_size+0x250/0x630 [xfs] > > > 0xffffffffc0079ed5 : xfs_alloc_ag_vextent+0xe5/0x150 [xfs] > > > 0xffffffffc007abc6 : xfs_alloc_vextent+0x446/0x5f0 [xfs] > > > 0xffffffffc008b123 : xfs_bmap_btalloc+0x3f3/0x780 [xfs] > > > 0xffffffffc008b4be : xfs_bmap_alloc+0xe/0x10 [xfs] > > > 0xffffffffc008bef9 : xfs_bmapi_write+0x499/0xab0 [xfs] > > > 0xffffffffc00c6ec8 : xfs_iomap_write_direct+0x1b8/0x390 [xfs] > > > > > -- > To unsubscribe from this list: send the line "unsubscribe linux-xfs" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe linux-xfs" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html