Shaohua Li reported commit aa45484031(mm: page allocator: calculate a better estimate of NR_FREE_PAGES when memory is low and kswapd is awake) made performance regression. | In a 4 socket 64 CPU system, zone_nr_free_pages() takes about 5% ~ 10% | cpu time | according to perf when memory pressure is high. The workload does | something like: | for i in `seq 1 $nr_cpu` | do | create_sparse_file $SPARSE_FILE-$i $((10 * mem / nr_cpu)) | $USEMEM -f $SPARSE_FILE-$i -j 4096 --readonly $((10 * mem / nr_cpu)) & | done | this simply reads a sparse file for each CPU. Apparently the | zone->percpu_drift_mark is too big, and guess zone_page_state_snapshot() | makes a lot of cache bounce for ->vm_stat_diff[]. below is the zoneinfo for | reference. | Is there any way to reduce the overhead? | | Node 3, zone Normal | pages free 2055926 | min 1441 | low 1801 | high 2161 | scanned 0 | spanned 2097152 | present 2068480 | vm stats threshold: 98 It mean zone_page_state_snapshot() is costly than we expected. This patch introduced very different approach. we are reserving max-drift pages at first instead runtime free page calculation. But, this technique can't be used on much cpus and few memory systems. On such system, we still need to use zone_page_state_snapshot(). Example1: typical desktop CPU: 2 MEM: 2GB old) zone->min = sqrt(2x1024x1024x16) = 5792 KB = 1448 pages new) max-drift = 2 x log2(2) x log2(2x1024/128) x 2 = 40 zone->min = 1448 + 40 = 1488 pages Example2: relatively large server CPU: 64 MEM: 8GBx4 (=32GB) old) zone->min = sqrt(32x1024x1024x16)/4 = 5792 KB = 1448 pages new) max-drift = 2 x log2(64) x log2(8x1024/128) x 64 = 6272 pages zone->min = 1448 + 6272 = 7720 pages Hmm, zone->min became almost 5x times. Is it acceptable? I think yes. Today, we can buy 8GB DRAM for $20. So, 6272 pages (=24.5MB) waste mean about 6 cent waste. It's good deal for getting good performance. Example3: ultimately big server CPU: 2048 MEM: 64GBx256 (=16TB) old) zone->min = sqrt(16x1024x1024x1024x16)/256 = 2048 KB = 512 pages (Wow!, it's smaller than desktop) new) max-drift = 125 x 2048 = 256000 pages = 1000MB (greater than 64GB/100) zone->min = 512 pages Reported-by: Shaohua Li <shaohua.li@xxxxxxxxx> Cc: Christoph Lameter <cl@xxxxxxxxx> Cc: Mel Gorman <mel@xxxxxxxxx> Signed-off-by: KOSAKI Motohiro <kosaki.motohiro@xxxxxxxxxxxxxx> --- mm/page_alloc.c | 9 +++++++++ 1 files changed, 9 insertions(+), 0 deletions(-) diff --git a/mm/page_alloc.c b/mm/page_alloc.c index 53627fa..194bdaa 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -4897,6 +4897,15 @@ static void setup_per_zone_wmarks(void) for_each_zone(zone) { u64 tmp; + /* + * If max drift are less than 1%, reserve max drift pages + * instead costly runtime calculation. + */ + if (zone->percpu_drift_mark < (zone->present_pages/100)) { + pages_min += zone->percpu_drift_mark; + zone->percpu_drift_mark = 0; + } + spin_lock_irqsave(&zone->lock, flags); tmp = (u64)pages_min * zone->present_pages; do_div(tmp, lowmem_pages); -- 1.6.5.2 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxxx For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>