Historically kswapd scanned from DMA->Movable in the opposite direction to the page allocator to avoid allocating behind kswapd direction of progress. The fair zone allocation policy altered this in a non-obvious manner. Traditionally, the page allocator prefers to use the highest eligible zone until the watermark is depleted, woke kswapd and moved onto the next zone. kswapd scans zones in the opposite direction so the scanning lists on 64-bit look like this; Page alloc Kswapd ---------- ------ Movable DMA Normal DMA32 DMA32 Normal DMA Movable If kswapd scanned in the same direction as the page allocator then it is possible that kswapd would proportionally reclaim the lower zones that were never used as the page allocator was always allocating behind the reclaim. This would work as follows pgalloc hits Normal low wmark kswapd reclaims Normal kswapd reclaims DMA32 pgalloc hits Normal low wmark kswapd reclaims Normal kswapd reclaims DMA32 The introduction of the fair zone allocation policy fundamentally altered this problem by interleaving between zones until the low watermark is reached. There are at least two issues with this o The page allocator can allocate behind kswapds progress (scans/reclaims lower zone and fair zone allocation policy then uses those pages) o When the low watermark of the high zone is reached there may recently allocated pages allocated from the lower zone but as kswapd scans dma->highmem to the highest zone needing balancing it'll reclaim the lower zone even if it was balanced. Let N = high_wmark(Normal) + high_wmark(DMA32). Of the last N allocations, some percentage will be allocated from Normal and some from DMA32. The percentage depends on the ratio of the zone sizes and when their watermarks were hit. If Normal is unbalanced, DMA32 will be shrunk by kswapd. If DMA32 is unbalanced only DMA32 will be shrunk. This leads to a difference of ages between DMA32 and Normal. Relatively young pages are then continually rotated and reclaimed from DMA32 due to the higher zone being unbalanced. Some of these pages may be recently read-ahead pages requiring that the page be re-read from disk and impacting overall performance. The problem is fundamental to the fact we have per-zone LRU and allocation policies and ideally we would only have per-node allocation and LRU lists. This would avoid the need for the fair zone allocation policy but the low-memory-starvation issue would have to be addressed again from scratch. This patch will only scan/reclaim from lower zones if they have not reached their watermark. This should not break the normal page aging as the proportional allocations due to the fair zone allocation policy should compensate. tiobench was used to evaluate this because it includes a simple sequential reader which is the most obvious regression. It also has threaded readers that produce reasonably steady figures. 3.16.0-rc2 3.16.0-rc2 3.0.0 vanilla checklow-v2r14 vanilla Min SeqRead-MB/sec-1 120.96 ( 0.00%) 140.63 ( 16.26%) 134.04 ( 10.81%) Min SeqRead-MB/sec-2 100.73 ( 0.00%) 117.95 ( 17.10%) 120.76 ( 19.88%) Min SeqRead-MB/sec-4 96.05 ( 0.00%) 109.54 ( 14.04%) 114.49 ( 19.20%) Min SeqRead-MB/sec-8 82.46 ( 0.00%) 88.22 ( 6.99%) 98.04 ( 18.89%) Min SeqRead-MB/sec-16 66.37 ( 0.00%) 69.14 ( 4.17%) 79.49 ( 19.77%) Mean RandWrite-MB/sec-16 1.34 ( 0.00%) 1.34 ( 0.00%) 1.34 ( 0.25%) It was also tested against xfs and there are similar gains. There are still regressions for higher number of threads but this is related to changes in the CFQ IO scheduler. Signed-off-by: Mel Gorman <mgorman@xxxxxxx> --- mm/vmscan.c | 16 ++++++++++------ 1 file changed, 10 insertions(+), 6 deletions(-) diff --git a/mm/vmscan.c b/mm/vmscan.c index 0f16ffe..40c3af8 100644 --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -3124,12 +3124,13 @@ static unsigned long balance_pgdat(pg_data_t *pgdat, int order, /* * Now scan the zone in the dma->highmem direction, stopping - * at the last zone which needs scanning. - * - * We do this because the page allocator works in the opposite - * direction. This prevents the page allocator from allocating - * pages behind kswapd's direction of progress, which would - * cause too much scanning of the lower zones. + * at the last zone which needs scanning. We do this because + * the page allocators prefers to work in the opposite + * direction and we want to avoid the page allocator reclaiming + * behind kswapd's direction of progress. Due to the fair zone + * allocation policy interleaving allocations between zones + * we no longer proportionally scan the lower zones if the + * watermarks are ok. */ for (i = 0; i <= end_zone; i++) { struct zone *zone = pgdat->node_zones + i; @@ -3152,6 +3153,9 @@ static unsigned long balance_pgdat(pg_data_t *pgdat, int order, &nr_soft_scanned); sc.nr_reclaimed += nr_soft_reclaimed; + if (zone_balanced(zone, order, 0, 0)) + continue; + /* * There should be no need to raise the scanning * priority if enough pages are already being scanned -- 1.8.4.5 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>