On Tue, Sep 24, 2019 at 7:23 AM Michal Hocko <mhocko@xxxxxxxxxx> wrote: > > On Wed 18-09-19 10:52:25, Alexander Duyck wrote: > [...] > > In order to try and keep the time needed to find a non-reported page to > > a minimum we maintain a "reported_boundary" pointer. This pointer is used > > by the get_unreported_pages iterator to determine at what point it should > > resume searching for non-reported pages. In order to guarantee pages do > > not get past the scan I have modified add_to_free_list_tail so that it > > will not insert pages behind the reported_boundary. > > > > If another process needs to perform a massive manipulation of the free > > list, such as compaction, it can either reset a given individual boundary > > which will push the boundary back to the list_head, or it can clear the > > bit indicating the zone is actively processing which will result in the > > reporting process resetting all of the boundaries for a given zone. > > Is this any different from the previous version? The last review > feedback (both from me and Mel) was that we are not happy to have an > externally imposed constrains on how the page allocator is supposed to > maintain its free lists. The main change for v10 versus v9 is that I allow the page reporting boundary to be overridden. Specifically there are two approaches that can be taken. The first is to simply reset the iterator for whatever list is updated. What this will do is reset the iterator back to list_head and then you can do whatever you want with that specific list. The other option is to simply clear the ZONE_PAGE_REPORTING_ACTIVE bit. That will essentially notify the page reporting code that any/all hints that were recorded have been discarded and that it needs to start over. All I am trying to do with this approach is reduce the work. Without doing this the code has to walk the entire free page list for the higher orders every iteration and that will not be cheap. Admittedly it is a bit more invasive than the cut/splice logic used in compaction which is taking the pages it has already processed and moving them to the other end of the list. However, I have reduced things so that we only really are limiting where add_to_free_list_tail can place pages, and we are having to check/push back the boundaries if a reported page is removed from a free_list. > If this is really the only way to go forward then I would like to hear > very convincing arguments about other approaches not being feasible. > There are none in this cover letter unfortunately. This will be really a > hard sell without them. So I had considered several different approaches. What I started out with was logic that was performing the hinting as a part of the architecture specific arch_free_page call. It worked but had performance issues as we were generating a hint per page freed which has fairly high overhead. The approach Nitesh has been using is to try and maintain a separate bitmap of "dirty" pages that have recently been freed. There are a few problems I saw with that approach. First is the fact that it becomes lossy in that pages could be reallocated out while we are waiting for the iterator to come through and process the page. This results in there being a greater amount of work as we have to hunt and peck for the pages, as such the zone lock has to be freed and reacquired often which slows this approach down further. Secondly there is the management of the bitmap itself and sparse memory which would likely necessitate doing something similar to pageblock_flags on order to support possible gaps in the zones. I had considered trying to maintain a separate list entirely and have the free pages placed there. However that was more invasive then this solution. In addition modifying the free_list/free_area in any way is problematic as it can result in the zone lock falling into the same cacheline as the highest order free_area. Ultimately what I settled on was the approach we have now where adding a page to the head of the free_list is unchanged, adding a page to the tail requires a check to see if the iterator is currently walking the list, and removing the page requires pushing back the iterator if the page is at the top of the reported list. I was trying to keep the amount of code that would have to be touched in the non-reported case to a minimum. With this we have to test for a bit in the zone flags if adding to tail, and we have to test for a bit in the page on a move/del from the freelist. So for the most common free/alloc cases we would only have the impact of the one additional page flag check.