On Wed 02-10-19 16:37:57, Linus Torvalds wrote: > On Wed, Oct 2, 2019 at 4:03 PM David Rientjes <rientjes@xxxxxxxxxx> wrote: > > > > Since hugetlb allocations have explicitly preferred to loop and do reclaim > > and compaction, exempt them from this new behavior at least for the time > > being. It is not shown that hugetlb allocation success rate has been > > impacted by commit b39d0ee2632d but hugetlb allocations are admittedly > > beyond the scope of what the patch is intended to address (thp > > allocations). > > I'd like to see some numbers to show that this special case makes sense. http://lkml.kernel.org/r/20191001054343.GA15624@xxxxxxxxxxxxxx While the test is somehow artificial it is not too much different from real workloads which do preallocate a non trivial (50% in my case) of memory for hugetlb pages. Having a moderately utilized memory (by page cache in my case) is not really unexpected. > I understand the "this is what it used to do, and hugetlbfs wasn't the > intended recipient of the new semantics", and I don't think the patch > is wrong. This is not only about this used to work. It is an expected and documented semantic of __GFP_RETRY_MAYFAIL * %__GFP_RETRY_MAYFAIL: The VM implementation will retry memory reclaim * procedures that have previously failed if there is some indication * that progress has been made else where. It can wait for other * tasks to attempt high level approaches to freeing memory such as * compaction (which removes fragmentation) and page-out. * There is still a definite limit to the number of retries, but it is * a larger limit than with %__GFP_NORETRY. * Allocations with this flag may fail, but only when there is * genuinely little unused memory. While these allocations do not * directly trigger the OOM killer, their failure indicates that * the system is likely to need to use the OOM killer soon. The * caller must handle failure, but can reasonably do so by failing * a higher-level request, or completing it only in a much less * efficient manner. > But at the same time, we do know that swap storms happen for other > loads, and if we say "hugetlbfs is different" then there should at > least be some rationale for why it's different other than "history". > Some actual "yes, we _want_ the possibile swap storms, because load > XYZ". > > And I don't mean microbenchmark numbers for "look, behavior changed". > I mean "look, this is a real load, and now it runs X% slower because > it relied on this hugetlbfs behavior". It is not about running slower. It is about not getting the expected amount of hugetlb pages requested by admin who knows that that size is needed. -- Michal Hocko SUSE Labs