Sorry about the word-wrap last email, here i resend it w/ hopefully better looking: On Mon, Apr 23, 2012 at 6:31 PM, Minchan Kim <minchan@xxxxxxxxxx> wrote: > Hi Ying, > > On 04/24/2012 08:18 AM, Ying Han wrote: > >> On Mon, Apr 23, 2012 at 3:20 PM, KOSAKI Motohiro >> <kosaki.motohiro@xxxxxxxxxxxxxx> wrote: >>> On Mon, Apr 23, 2012 at 4:56 PM, Ying Han <yinghan@xxxxxxxxxx> wrote: >>>> This is not a patch targeted to be merged at all, but trying to understand >>>> a logic in global direct reclaim. >>>> >>>> There is a logic in global direct reclaim where reclaim fails on priority 0 >>>> and zone->all_unreclaimable is not set, it will cause the direct to start over >>>> from DEF_PRIORITY. In some extreme cases, we've seen the system hang which is >>>> very likely caused by direct reclaim enters infinite loop. >>>> >>>> There have been serious patches trying to fix similar issue and the latest >>>> patch has good summary of all the efforts: >>>> >>>> commit 929bea7c714220fc76ce3f75bef9056477c28e74 >>>> Author: KOSAKI Motohiro <kosaki.motohiro@xxxxxxxxxxxxxx> >>>> Date: Thu Apr 14 15:22:12 2011 -0700 >>>> >>>> vmscan: all_unreclaimable() use zone->all_unreclaimable as a name >>>> >>>> Kosaki explained the problem triggered by async zone->all_unreclaimable and >>>> zone->pages_scanned where the later one was being checked by direct reclaim. >>>> However, after the patch, the problem remains where the setting of >>>> zone->all_unreclaimable is asynchronous with zone is actually reclaimable or not. >>>> >>>> The zone->all_unreclaimable flag is set by kswapd by checking zone->pages_scanned in >>>> zone_reclaimable(). Is that possible to have zone->all_unreclaimable == false while >>>> the zone is actually unreclaimable? >>>> >>>> 1. while kswapd in reclaim priority loop, someone frees a page on the zone. It >>>> will end up resetting the pages_scanned. >>>> >>>> 2. kswapd is frozen for whatever reason. I noticed Kosaki's covered the >>>> hibernation case by checking oom_killer_disabled, but not sure if that is >>>> everything we need to worry about. The key point here is that direct reclaim >>>> relies on a flag which is set by kswapd asynchronously, that doesn't sound safe. >>> >>> If kswapd was frozen except hibernation, why don't you add frozen >>> check instead of >>> hibernation check? And when and why is that happen? >> >> I haven't tried to reproduce the issue, so everything is based on >> eye-balling the code. The problem is that we have the potential >> infinite loop in direct reclaim where it keeps trying as long as >> !zone->all_unreclaimable. >> >> The flag is only set by kswapd and it will skip setting the flag if >> the following condition is true: >> >> zone->pages_scanned < zone_reclaimable_pages(zone) * 6; >> >> In a few-pages-on-lru condition, the zone->pages_scanned is easily >> remains 0 and also it is reset to 0 everytime a page being freed. >> Then, i will cause global direct reclaim entering infinite loop. >> > > > how does zone->pages_scanned become 0 easily in global reclaim? > Once VM has pages in LRU, it wouldn't be a zero. Look at isolate_lru_pages. > The problem is get_scan_count which could prevent scanning of LRU list but > it works well now. If the priority isn't zero and there are few pages in LRU, > it could be a zero scan but when the priority drop at zero, it could let VM scan > less pages under SWAP_CLUSTER_MAX. So pages_scanned would be increased. Yes, that is true. But the pages_scanned will be reset on freeing a page and that could happen asynchronously. For example I have only 2 pages on file_lru (w/o swap), and here is what is supposed to happen: A kswapd B direct reclaim priority DEP_PRIORITY to 0 zone->pages_scanned = 3 zone_reclaimable() == true zone->all_unreclaimable == 0 nr_reclaimed == 0 & !zone->all_unreclaimable retry priority DEP_PRIORITY to 0 zone->pages_scanned = 6 zone_reclaimable() == true zone->all_unreclaimable == 0 nr_reclaimed == 0 & !zone->all_unreclaimable retry repeat the above which eventually zone->pages_scanned will grow zone->pages_scanned to 12 zone_reclaimable() == false zone->all_unreclaimable == 1 nr_reclaimed == 0 & zone->all_unreclaimable oom However, what if B frees a pages everytime before pages_scanned reaches the point, then we won't set zone->all_unreclaimable at all. If so, we reaches a livelock here... --Ying > > I think the problem is live-lock as follows, > > > A kswapd B > > direct reclaim > reclaim a page > pages_scanned check <- skip > > steal a page reclaimed by A > use the page for user memory. > alloc failed > retry > > In this scenario, process A would be a live-locked. > Does it make sense for infinite loop case you mentioned? > > > -- > Kind regards, > Minchan Kim -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Fight unfair telecom internet charges in Canada: sign http://stopthemeter.ca/ Don't email: <a href