On Wed, May 28, 2014 at 03:41:11PM -0700, Linus Torvalds wrote: > On Wed, May 28, 2014 at 3:31 PM, Dave Chinner <david@xxxxxxxxxxxxx> wrote: > > > > Indeed, the call chain reported here is not caused by swap issuing > > IO. > > Well, that's one way of reading that callchain. > > I think it's the *wrong* way of reading it, though. Almost dishonestly > so. I guess you haven't met your insult quota for the day, Linus. :/ > Because very clearly, the swapout _is_ what causes the unplugging > of the IO queue, and does so because it is allocating the BIO for its > own IO. The fact that that then fails (because of other IO's in > flight), and causes *other* IO to be flushed, doesn't really change > anything fundamental. It's still very much swap that causes that > "let's start IO". It is not rocket science to see how plugging outside memory allocation context can lead to flushing that plug within direct reclaim without having dispatched any IO at all from direct reclaim.... You're focussing on the specific symptoms, not the bigger picture. i.e. you're ignoring all the other "let's start IO" triggers in direct reclaim. e.g there's two separate plug flush triggers in shrink_inactive_list(), one of which is: /* * Stall direct reclaim for IO completions if underlying BDIs or zone * is congested. Allow kswapd to continue until it starts encountering * unqueued dirty pages or cycling through the LRU too quickly. */ if (!sc->hibernation_mode && !current_is_kswapd()) wait_iff_congested(zone, BLK_RW_ASYNC, HZ/10); I'm not saying we shouldn't turn of swap from direct reclaim, just that all we'd be doing by turning off swap is playing whack-a-stack - the next report will simply be from one of the other direct reclaim IO schedule points. Regardless of whether it is swap or something external queues the bio on the plug, perhaps we should look at why it's done inline rather than by kblockd, where it was moved because it was blowing the stack from schedule(): commit f4af3c3d077a004762aaad052049c809fd8c6f0c Author: Jens Axboe <jaxboe@xxxxxxxxxxxx> Date: Tue Apr 12 14:58:51 2011 +0200 block: move queue run on unplug to kblockd There are worries that we are now consuming a lot more stack in some cases, since we potentially call into IO dispatch from schedule() or io_schedule(). We can reduce this problem by moving the running of the queue to kblockd, like the old plugging scheme did as well. This may or may not be a good idea from a performance perspective, depending on how many tasks have queue plugs running at the same time. For even the slightly contended case, doing just a single queue run from kblockd instead of multiple runs directly from the unpluggers will be faster. Signed-off-by: Jens Axboe <jaxboe@xxxxxxxxxxxx> commit a237c1c5bc5dc5c76a21be922dca4826f3eca8ca Author: Jens Axboe <jaxboe@xxxxxxxxxxxx> Date: Sat Apr 16 13:27:55 2011 +0200 block: let io_schedule() flush the plug inline Linus correctly observes that the most important dispatch cases are now done from kblockd, this isn't ideal for latency reasons. The original reason for switching dispatches out-of-line was to avoid too deep a stack, so by _only_ letting the "accidental" flush directly in schedule() be guarded by offload to kblockd, we should be able to get the best of both worlds. So add a blk_schedule_flush_plug() that offloads to kblockd, and only use that from the schedule() path. Signed-off-by: Jens Axboe <jaxboe@xxxxxxxxxxxx> And now we have too deep a stack due to unplugging from io_schedule()... > IOW, swap-out directly caused that extra 3kB of stack use in what was > a deep call chain (due to memory allocation). I really don't > understand why you are arguing anything else on a pure technicality. > > I thought you had some other argument for why swap was different, and > against removing that "page_is_file_cache()" special case in > shrink_page_list(). I've said in the past that swap is different to filesystem ->writepage implementations because it doesn't require significant stack to do block allocation and doesn't trigger IO deep in that allocation stack. Hence it has much lower stack overhead than the filesystem ->writepage implementations and so is much less likely to have stack issues. This stack overflow shows us that just the memory reclaim + IO layers are sufficient to cause a stack overflow, which is something I've never seen before. That implies no IO in direct reclaim context is safe - either from swap or io_schedule() unplugging. It also lends a lot of weight to my assertion that the majority of the stack growth over the past couple of years has been ocurring outside the filesystems.... Cheers, Dave. -- Dave Chinner david@xxxxxxxxxxxxx -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>