On Tue, Mar 8, 2011 at 6:58 AM, Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> wrote: > On Sun, 6 Mar 2011 02:07:59 +0900 > Minchan Kim <minchan.kim@xxxxxxxxx> wrote: > >> On Sat, Mar 05, 2011 at 07:41:26PM +0300, Andrew Vagin wrote: >> > On 03/05/2011 06:53 PM, Minchan Kim wrote: >> > >On Sat, Mar 05, 2011 at 06:34:37PM +0300, Andrew Vagin wrote: >> > >>On 03/05/2011 06:20 PM, Minchan Kim wrote: >> > >>>On Sat, Mar 05, 2011 at 02:44:16PM +0300, Andrey Vagin wrote: >> > >>>>Check zone->all_unreclaimable in all_unreclaimable(), otherwise the >> > >>>>kernel may hang up, because shrink_zones() will do nothing, but >> > >>>>all_unreclaimable() will say, that zone has reclaimable pages. >> > >>>> >> > >>>>do_try_to_free_pages() >> > >>>> Â Â Â Âshrink_zones() >> > >>>> Â Â Â Â Â Â Â Â for_each_zone >> > >>>> Â Â Â Â Â Â Â Â Â Â Â Âif (zone->all_unreclaimable) >> > >>>> Â Â Â Â Â Â Â Â Â Â Â Â Â Â Â Âcontinue >> > >>>> Â Â Â Âif !all_unreclaimable(zonelist, sc) >> > >>>> Â Â Â Â Â Â Â Âreturn 1 >> > >>>> >> > >>>>__alloc_pages_slowpath() >> > >>>>retry: >> > >>>> Â Â Â Âdid_some_progress = do_try_to_free_pages(page) >> > >>>> Â Â Â Â... >> > >>>> Â Â Â Âif (!page&& Â did_some_progress) >> > >>>> Â Â Â Â Â Â Â Âretry; >> > >>>> >> > >>>>Signed-off-by: Andrey Vagin<avagin@xxxxxxxxxx> >> > >>>>--- >> > >>>> Âmm/vmscan.c | Â Â2 ++ >> > >>>> Â1 files changed, 2 insertions(+), 0 deletions(-) >> > >>>> >> > >>>>diff --git a/mm/vmscan.c b/mm/vmscan.c >> > >>>>index 6771ea7..1c056f7 100644 >> > >>>>--- a/mm/vmscan.c >> > >>>>+++ b/mm/vmscan.c >> > >>>>@@ -2002,6 +2002,8 @@ static bool all_unreclaimable(struct zonelist *zonelist, >> > >>>> >> > >>>> Â Â Â Âfor_each_zone_zonelist_nodemask(zone, z, zonelist, >> > >>>> Â Â Â Â Â Â Â Â Â Â Â Âgfp_zone(sc->gfp_mask), sc->nodemask) { >> > >>>>+ Â Â Â Â Â Â Â if (zone->all_unreclaimable) >> > >>>>+ Â Â Â Â Â Â Â Â Â Â Â continue; >> > >>>> Â Â Â Â Â Â Â Âif (!populated_zone(zone)) >> > >>>> Â Â Â Â Â Â Â Â Â Â Â Âcontinue; >> > >>>> Â Â Â Â Â Â Â Âif (!cpuset_zone_allowed_hardwall(zone, GFP_KERNEL)) >> > >>>zone_reclaimable checks it. Isn't it enough? >> > >>I sent one more patch [PATCH] mm: skip zombie in OOM-killer. >> > >>This two patches are enough. >> > >Sorry if I confused you. >> > >I mean zone->all_unreclaimable become true if !zone_reclaimable in balance_pgdat. >> > >zone_reclaimable compares recent pages_scanned with the number of zone lru pages. >> > >So too many page scanning in small lru pages makes the zone to unreclaimable zone. >> > > >> > >In all_unreclaimable, we calls zone_reclaimable to detect it. >> > >It's the same thing with your patch. >> > balance_pgdat set zone->all_unreclaimable, but the problem is that >> > it is cleaned late. >> >> Yes. It can be delayed by pcp so (zone->all_unreclaimable = true) is >> a false alram since zone have a free page and it can be returned >> to free list by drain_all_pages in next turn. >> >> > >> > The problem is that zone->all_unreclaimable = True, but >> > zone_reclaimable() returns True too. >> >> Why is it a problem? >> If zone->all_unreclaimable gives a false alram, we does need to check >> it again by zone_reclaimable call. >> >> If we believe a false alarm and give up the reclaim, maybe we have to make >> unnecessary oom kill. >> >> > >> > zone->all_unreclaimable will be cleaned in free_*_pages, but this >> > may be late. It is enough allocate one page from page cache, that >> > zone_reclaimable() returns True and zone->all_unreclaimable becomes >> > True. >> > >>>Does the hang up really happen or see it by code review? >> > >>Yes. You can reproduce it for help the attached python program. It's >> > >>not very clever:) >> > >>It make the following actions in loop: >> > >>1. fork >> > >>2. mmap >> > >>3. touch memory >> > >>4. read memory >> > >>5. munmmap >> > >It seems the test program makes fork bombs and memory hogging. >> > >If you applied this patch, the problem is gone? >> > Yes. >> >> Hmm.. Although it solves the problem, I think it's not a good idea that >> depends on false alram and give up the retry. > > Any alternative proposals? ÂWe should get the livelock fixed if possible.. > And we should avoid unnecessary OOM kill if possible. I think the problem is caused by (zone->pages_scanned < zone_reclaimable_pages(zone) * 6). I am not sure (* 6) is a best. It would be rather big on recent big DRAM machines. I think it is a trade-off between latency and OOM kill. If we decrease the magic value, maybe we should prevent the almost livelock but happens unnecessary OOM kill. And I think zone_reclaimable not fair. For example, too many scanning makes reclaimable state to unreclaimable state. Maybe it takes a very long time. But just some page free makes unreclaimable state to reclaimabe with very easy. So we need much painful reclaiming for changing reclaimable state with unreclaimabe state. it would affect latency very much. Maybe we need more smart zone_reclaimabe which is adaptive with memory pressure. -- Kind regards, Minchan Kim -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxxx For more info on Linux MM, see: http://www.linux-mm.org/ . Fight unfair telecom internet charges in Canada: sign http://stopthemeter.ca/ Don't email: <a href