From: Shaohua Li <shli@xxxxxx> MADV_FREE clears pte dirty bit and then marks the page lazyfree (clear SwapBacked). There is no lock to prevent the page is added to swap cache between these two steps by page reclaim. If page reclaim finds such page, it will simply add the page to swap cache without pageout the page to swap because the page is marked as clean. Next time, page fault will read data from the swap slot which doesn't have the original data, so we have a data corruption. To fix issue, we mark the page dirty and pageout the page. However, we shouldn't dirty all pages which is clean and in swap cache. swapin page is swap cache and clean too. So we only dirty page which is added into swap cache in page reclaim, which shouldn't be swapin page. Normal anonymous pages should be dirty already. Reported-and-tested-y: Artem Savkov <asavkov@xxxxxxxxxx> Fix: 802a3a92ad7a(mm: reclaim MADV_FREE pages) Signed-off-by: Shaohua Li <shli@xxxxxx> Cc: Johannes Weiner <hannes@xxxxxxxxxxx> Cc: Michal Hocko <mhocko@xxxxxxxx> Cc: Hillf Danton <hillf.zj@xxxxxxxxxxxxxxx> Cc: Minchan Kim <minchan@xxxxxxxxxx> Cc: Hugh Dickins <hughd@xxxxxxxxxx> Cc: Rik van Riel <riel@xxxxxxxxxx> Cc: Mel Gorman <mgorman@xxxxxxxxxxxxxxxxxxx> Cc: Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> --- mm/vmscan.c | 12 ++++++++++++ 1 file changed, 12 insertions(+) diff --git a/mm/vmscan.c b/mm/vmscan.c index d811c81..820ee8d 100644 --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -980,6 +980,7 @@ static unsigned long shrink_page_list(struct list_head *page_list, int may_enter_fs; enum page_references references = PAGEREF_RECLAIM_CLEAN; bool dirty, writeback; + bool new_swap_page = false; cond_resched(); @@ -1165,6 +1166,7 @@ static unsigned long shrink_page_list(struct list_head *page_list, /* Adding to swap updated mapping */ mapping = page_mapping(page); + new_swap_page = true; } } else if (unlikely(PageTransHuge(page))) { /* Split file THP */ @@ -1185,6 +1187,16 @@ static unsigned long shrink_page_list(struct list_head *page_list, nr_unmap_fail++; goto activate_locked; } + + /* + * MADV_FREE clear pte dirty bit, but not yet clear + * SwapBacked for a page. We can't directly free the + * page because we already set swap entry in pte. The + * check guarantees this is such page and not a clean + * swapin page + */ + if (!PageDirty(page) && new_swap_page) + set_page_dirty(page); } if (PageDirty(page)) { -- 2.9.5 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>