The patch titled Subject: mm-vmscan-fix-the-page-state-calculation-in-too_many_isolated-fix has been removed from the -mm tree. Its filename was mm-vmscan-fix-the-page-state-calculation-in-too_many_isolated-fix.patch This patch was dropped because it was folded into mm-vmscan-fix-the-page-state-calculation-in-too_many_isolated.patch ------------------------------------------------------ From: Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> Subject: mm-vmscan-fix-the-page-state-calculation-in-too_many_isolated-fix Move the zone_page_state_snapshot() fallback logic into too_many_isolated(), so shrink_inactive_list() doesn't incorrectly call congestion_wait(). Cc: Johannes Weiner <hannes@xxxxxxxxxxx> Cc: Mel Gorman <mgorman@xxxxxxx> Cc: Michal Hocko <mhocko@xxxxxxx> Cc: Minchan Kim <minchan@xxxxxxxxxx> Cc: Vinayak Menon <vinmenon@xxxxxxxxxxxxxx> Cc: Vladimir Davydov <vdavydov@xxxxxxxxxxxxx> Signed-off-by: Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> --- mm/vmscan.c | 23 +++++++++++------------ 1 file changed, 11 insertions(+), 12 deletions(-) diff -puN mm/vmscan.c~mm-vmscan-fix-the-page-state-calculation-in-too_many_isolated-fix mm/vmscan.c --- a/mm/vmscan.c~mm-vmscan-fix-the-page-state-calculation-in-too_many_isolated-fix +++ a/mm/vmscan.c @@ -1402,7 +1402,7 @@ int isolate_lru_page(struct page *page) } static int __too_many_isolated(struct zone *zone, int file, - struct scan_control *sc, int safe) + struct scan_control *sc, int safe) { unsigned long inactive, isolated; @@ -1435,7 +1435,7 @@ static int __too_many_isolated(struct zo * unnecessary swapping, thrashing and OOM. */ static int too_many_isolated(struct zone *zone, int file, - struct scan_control *sc, int safe) + struct scan_control *sc) { if (current_is_kswapd()) return 0; @@ -1443,12 +1443,14 @@ static int too_many_isolated(struct zone if (!global_reclaim(sc)) return 0; - if (unlikely(__too_many_isolated(zone, file, sc, 0))) { - if (safe) - return __too_many_isolated(zone, file, sc, safe); - else - return 1; - } + /* + * __too_many_isolated(safe=0) is fast but inaccurate, because it + * doesn't account for the vm_stat_diff[] counters. So if it looks + * like too_many_isolated() is about to return true, fall back to the + * slower, more accurate zone_page_state_snapshot(). + */ + if (unlikely(__too_many_isolated(zone, file, sc, 0))) + return __too_many_isolated(zone, file, sc, 1); return 0; } @@ -1540,18 +1542,15 @@ shrink_inactive_list(unsigned long nr_to unsigned long nr_immediate = 0; isolate_mode_t isolate_mode = 0; int file = is_file_lru(lru); - int safe = 0; struct zone *zone = lruvec_zone(lruvec); struct zone_reclaim_stat *reclaim_stat = &lruvec->reclaim_stat; - while (unlikely(too_many_isolated(zone, file, sc, safe))) { + while (unlikely(too_many_isolated(zone, file, sc))) { congestion_wait(BLK_RW_ASYNC, HZ/10); /* We are about to die and free our memory. Return now. */ if (fatal_signal_pending(current)) return SWAP_CLUSTER_MAX; - - safe = 1; } lru_add_drain(); _ Patches currently in -mm which might be from akpm@xxxxxxxxxxxxxxxxxxxx are origin.patch mm-page_allocc-__alloc_pages_nodemask-dont-alter-arg-gfp_mask.patch vmscan-force-scan-offline-memory-cgroups-fix.patch mm-memcontrol-default-hierarchy-interface-for-memory.patch mm-memcontrol-fold-move_anon-and-move_file-fix.patch mm-compaction-enhance-tracepoint-output-for-compaction-begin-end-v4-fix.patch mm-vmscan-fix-the-page-state-calculation-in-too_many_isolated.patch mm-thp-allocate-transparent-hugepages-on-local-node-fix.patch docs-procs-describe-proc-pid-map_files-entry-fix.patch mm-page_ext-remove-unnecessary-stack_trace-field-fix.patch mm-fix-arithmetic-overflow-in-__vm_enough_memory-fix.patch x86-add-pmd_-for-thp-fix.patch sparc-add-pmd_-for-thp-fix.patch -- To unsubscribe from this list: send the line "unsubscribe mm-commits" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html