On Fri, 7 Jun 2019, Michal Hocko wrote: > > So my proposed change would be: > > - give the page allocator a consistent indicator that compaction failed > > because we are low on memory (make COMPACT_SKIPPED really mean this), > > - if we get this in the page allocator and we are allocating thp, fail, > > reclaim is unlikely to help here and is much more likely to be > > disruptive > > - we could retry compaction if we haven't scanned all memory and > > were contended, > > - if the hugepage allocation fails, have thp check watermarks for order-0 > > pages without any padding, > > - if watermarks succeed, fail the thp allocation: we can't allocate > > because of fragmentation and it's better to return node local memory, > > Doesn't this lead to the same THP low success rate we have seen with one > of the previous patches though? > >From my recollection, the only other patch that was tested involved __GFP_NORETRY and avoiding reclaim entirely for thp allocations when checking for high-order allocations. This in the page allocator: /* * Checks for costly allocations with __GFP_NORETRY, which * includes THP page fault allocations */ if (costly_order && (gfp_mask & __GFP_NORETRY)) { ... if (compact_result == COMPACT_DEFERRED) goto nopage; Yet there is no handling for COMPACT_SKIPPED (or what my plan above defines COMPACT_SKIPPED to be). I don't think anything was tried that tests why compaction failed, i.e. was it because the two scanners met, because hugepage-order memory was found available, because the zone lock was contended or we hit need_resched(), we're failing even order-0 watermarks, etc. I don't think the above plan has been tried, if someone has tried it, please let me know. I haven't seen any objection to disabling reclaim entirely when order-0 watermarks are failing in compaction. We simply can't guarantee that it is useful work with the current implementation of compaction. There are several reasons that I've enumerated why compaction can still fail even after successful reclaim. The point is that removing __GFP_THISNODE is not a fix for this if the remote memory is fragmented as well: it assumes that hugepages are available remotely when they aren't available locally otherwise we seem swap storms both locally and remotely. Relying on that is not in the best interest of any user of transparent hugepages. > Let me remind you of the previous semantic I was proposing > http://lkml.kernel.org/r/20181206091405.GD1286@xxxxxxxxxxxxxx and that > didn't get shot down. Linus had some follow up ideas on how exactly > the fallback order should look like and that is fine. We should just > measure differences between local node cheep base page vs. remote THP on > _real_ workloads. Any microbenchmark which just measures a latency is > inherently misleading. > I think both seek to add the possibility of allocating hugepages remotely in certain circumstances and that can be influenced by MADV_HUGEPAGE. I don't think we need to try hugepage specific mempolicies unless it is shown to be absolutely necessary although a usecase for this could be made separate to this discussion. There's a benefit to faulting remote hugepages over remote pages for everybody involved. My argument is that we can determine the need for that based on failed order-0 watermark checks in compaction: if the node would require reclaim to even fault a page, it is likely better over the long term to fault a remote hugepage. I think this can be made to work and is not even difficult to do. If anybody has any objection please let me know.