On Sat, Oct 08, 2011 at 01:48:21PM +0800, Shaohua Li wrote: > On Sat, 2011-10-08 at 12:32 +0800, Minchan Kim wrote: > > On Sat, Oct 08, 2011 at 11:09:51AM +0800, Shaohua Li wrote: > > > On Sat, 2011-10-01 at 14:59 +0800, Minchan Kim wrote: > > > > On Fri, Sep 30, 2011 at 10:12:23AM +0800, Shaohua Li wrote: > > > > > On Thu, 2011-09-29 at 17:18 +0800, Minchan Kim wrote: > > > > > > On Thu, Sep 29, 2011 at 09:14:51AM +0800, Shaohua Li wrote: > > > > > > > On Thu, 2011-09-29 at 01:57 +0800, Minchan Kim wrote: > > > > > > > > On Wed, Sep 28, 2011 at 03:08:31PM +0800, Shaohua Li wrote: > > > > > > > > > On Wed, 2011-09-28 at 14:57 +0800, Minchan Kim wrote: > > > > > > > > > > On Tue, Sep 27, 2011 at 03:23:04PM +0800, Shaohua Li wrote: > > > > > > > > > > > I saw DMA zone always has ->all_unreclaimable set. The reason is the high zones > > > > > > > > > > > are big, so zone_watermark_ok/_safe() will always return false with a high > > > > > > > > > > > classzone_idx for DMA zone, because DMA zone's lowmem_reserve is big for a high > > > > > > > > > > > classzone_idx. When kswapd runs into DMA zone, it doesn't scan/reclaim any > > > > > > > > > > > pages(no pages in lru), but mark the zone as all_unreclaimable. This can > > > > > > > > > > > happen in other low zones too. > > > > > > > > > > > > > > > > > > > > Good catch! > > > > > > > > > > > > > > > > > > > > > This is confusing and can potentially cause oom. Say a low zone has > > > > > > > > > > > all_unreclaimable when high zone hasn't enough memory. Then allocating > > > > > > > > > > > some pages in low zone(for example reading blkdev with highmem support), > > > > > > > > > > > then run into direct reclaim. Since the zone has all_unreclaimable set, > > > > > > > > > > > direct reclaim might reclaim nothing and an oom reported. If > > > > > > > > > > > all_unreclaimable is unset, the zone can actually reclaim some pages. > > > > > > > > > > > If all_unreclaimable is unset, in the inner loop of balance_pgdat we always have > > > > > > > > > > > all_zones_ok 0 when checking a low zone's watermark. If high zone watermark isn't > > > > > > > > > > > good, there is no problem. Otherwise, we might loop one more time in the outer > > > > > > > > > > > loop, but since high zone watermark is ok, the end_zone will be lower, then low > > > > > > > > > > > zone's watermark check will be ok and the outer loop will break. So looks this > > > > > > > > > > > doesn't bring any problem. > > > > > > > > > > > > > > > > > > > > I think it would be better to correct zone_reclaimable. > > > > > > > > > > My point is zone_reclaimable should consider zone->pages_scanned. > > > > > > > > > > The point of the function is how many pages scanned VS how many pages remained in LRU. > > > > > > > > > > If reclaimer doesn't scan the zone at all because of no lru pages, it shouldn't tell > > > > > > > > > > the zone is all_unreclaimable. > > > > > > > > > actually this is exact my first version of the patch. The problem is if > > > > > > > > > a zone is true unreclaimable (used by kenrel pages or whatever), we will > > > > > > > > > have zone->pages_scanned 0 too. I thought we should set > > > > > > > > > all_unreclaimable in this case. > > > > > > > > > > > > > > > > Let's think the problem again. > > > > > > > > Fundamental problem is that why the lower zone's lowmem_reserve for higher zone is huge big > > > > > > > > that might be bigger than the zone's size. > > > > > > > > I think we need the boundary for limiting lowmem_reseve. > > > > > > > > So how about this? > > > > > > > I didn't see a reason why high zone allocation should fallback to low > > > > > > > zone if high zone is big. Changing the lowmem_reserve can cause the > > > > > > > fallback. Has any rationale here? > > > > > > > > > > > > I try to think better solution than yours but I got failed. :( > > > > > > The why I try to avoid your patch is that kswapd is very complicated these days so > > > > > > I wanted to not add more logic for handling corner cases if we can solve it > > > > > > other ways. But as I said, but I got failed. > > > > > > > > > > > > It seems that it doesn't make sense that previous my patch that limit lowmem_reserve. > > > > > > Because we can have higher zone which is very big size so that lowmem_zone[higher_zone] of > > > > > > low zone could be bigger freely than the lowmem itself size. > > > > > > It implies the low zone should be not used for higher allocation. > > > > > > It has no reason to limit it. My brain was broken. :( > > > > > > > > > > > > But I have a question about your patch still. > > > > > > What happens if DMA zone sets zone->all_unreclaimable with 1? > > > > > > > > > > > > You said as follows, > > > > > > > > > > > > > This is confusing and can potentially cause oom. Say a low zone has > > > > > > > all_unreclaimable when high zone hasn't enough memory. Then allocating > > > > > > > some pages in low zone(for example reading blkdev with highmem support), > > > > > > > then run into direct reclaim. Since the zone has all_unreclaimable set, > > > > > > > > > > > > If low zone has enough pages for allocation, it cannot have entered reclaim. > > > > > > It means now low zone doesn't have enough free pages for the order allocation. > > > > > > So it's natural to enter reclaim path. > > > > > > > > > > > > > direct reclaim might reclaim nothing and an oom reported. If > > > > > > > > > > > > It's not correct "nothing". At least, it will do something in DEF_PRIORITY. > > > > > it does something, but might not reclaim any pages, for example, it > > > > > starts page write, but page isn't in disk yet in DEF_PRIORITY and it > > > > > skip further reclaiming in !DEF_PRIORITY. > > > > > > > > > > > > all_unreclaimable is unset, the zone can actually reclaim some pages. > > > > > > > > > > > > The reason of this problem is that the zone has no lru page, you said. > > > > > > Then how could we reclaim some pages in the zone even if the zone's all_unreclaimable is unset? > > > > > > You expect slab pages? > > > > > The zone could have lru pages. Let's take an example, allocation from > > > > > ZONE_HIGHMEM, then kswapd runs, ZONE_NORMAL gets all_unreclaimable set > > > > > even it has free pages. Then we do write blkdev device, which use > > > > > ZONE_NORMAL for page cache. Some pages in ZONE_NORMAL are in lru, then > > > > > we run into direct page reclaim for ZONE_NORMAL. Since all_unreclaimable > > > > > is set and pages in ZONE_NORMAL lru are dirty, direct reclaim could > > > > > fail. But I'd agree this is a corner case. > > > > > Besides when I saw ZONE_DMA has a lot of free pages and > > > > > all_unreclaimable is set, it's really confusing. > > > > > > > > Hi Shaohua, > > > > Sorry for late response and Thanks for your explanation. > > > > It's valuable to fix, I think. > > > > How about this? > > > > > > > > I hope other guys have a interest in the problem. > > > > Cced them. > > > Hi, > > > it's a long holiday here, so I'm late, sorry. > > > > No problem. I coundn't access internet freely, either. > > > > > > > > > From 070d5b1a69921bc71c6aaa5445fb1d29ecb38f74 Mon Sep 17 00:00:00 2001 > > > > From: Minchan Kim <minchan.kim@xxxxxxxxx> > > > > Date: Sat, 1 Oct 2011 15:26:08 +0900 > > > > Subject: [RFC] vmscan: set all_unreclaimable of zone carefully > > > > > > > > Shaohua Li reported all_unreclaimable of DMA zone is always set > > > > because the system has a big memory HIGH zone so that lowmem_reserve[HIGH] > > > > could be a big. > > > > > > > > It could be a problem as follows > > > > > > > > Assumption : > > > > 1. The system has a big high memory so that lowmem_reserve[HIGH] of DMA zone would be big. > > > > 2. HIGH/NORMAL zone are full but DMA zone has enough free pages. > > > > > > > > Scenario > > > > 1. A request to allocate a page in HIGH zone. > > > > 2. HIGH/NORMAL zone already consumes lots of pages so that it would be fall-backed to DMA zone. > > > > 3. In DMA zone, allocator got failed, too becuase lowmem_reserve[HIGH] is very big so that it wakes up kswapd > > > > 4. kswapd would call shrink_zone while it see DMA zone since DMA zone's lowmem_reserve[HIGHMEM] > > > > would be big so that it couldn't meet zone_watermark_ok_safe(high_wmark_pages(zone) + balance_gap, > > > > *end_zone*) > > > > 5. DMA zone doesn't meet stop condition(nr_slab != 0, !zone_reclaimable) because the zone has small lru pages > > > > and it doesn't have slab pages so that kswapd would set all_unreclaimable of the zone to *1* easily. > > > > 6. B request to allocate many pages in NORMAL zone but NORMAL zone has no free pages > > > > so that it would be fall-backed to DMA zone. > > > > 7. DMA zone would allocates many pages for NORMAL zone because lowmem_reserve[NORMAL] is small. > > > > These pages are used by application(ie, it menas LRU pages. Yes. Now DMA zone could have many reclaimable pages) > > > > 8. C request to allocate a page in NORMAL zone but he got failed because DMA zone doesn't have enough free pages. > > > > (Most of pages in DMA zone are consumed by B) > > > > 9. Kswapd try to reclaim lru pages in DMA zone but got failed because all_unreclaimable of the zone is 1. Otherwise, > > > > it could reclaim many pages which are used by B. > > > > > > > > Of coures, we can do something in DEF_PRIORITY but it couldn't do enough because it can't raise > > > > synchronus reclaim in direct reclaim path if the zone has many dirty pages > > > > so that the process is killed by OOM. > > > > > > > > The principal problem is caused by step 8. > > > > In step 8, we increased # of lru size very much but still the zone->all_unreclaimable is 1. > > > > If we increase lru size, it is valuable to try reclaiming again. > > > > The rationale is that we reset all_unreclaimable to 0 even if we free just a one page. > > > So this fixes the oom, but we still have DMA has all_unreclaimable set > > > always, because all_unreclaimable == zone_reclaimable_pages() + 1. Not a > > > problem? > > > > I think we can fix it if it is needeed. > > > > diff --git a/mm/vmstat.c b/mm/vmstat.c > > index 471b20b..ede852c 100644 > > --- a/mm/vmstat.c > > +++ b/mm/vmstat.c > > @@ -1019,7 +1019,7 @@ static void zoneinfo_show_print(struct seq_file *m, pg_data_t *pgdat, > > "\n all_unreclaimable: %u" > > "\n start_pfn: %lu" > > "\n inactive_ratio: %u", > > - zone->all_unreclaimable, > > + zone_unreclaimable(zone), > > zone->zone_start_pfn, > > zone->inactive_ratio); > > seq_putc(m, '\n'); > > > > I think it's not a big problem at the start. > > all_unreclamable doesn't mean "we have no free page in the zone" but "we have no reclaimable pages any more in the zone". > > It is possible in case of setting reserve memory very high for higher zones. > > If you think it's awkard, we could add description about that in Documentation. > this doesn't fix it, because all_unreclaimable == > zone_reclaimable_pages() + 1, zone_unreclaimable() will still be true. We set the zone into all_unreclaimabe when VM found much scanning in the zone without reclaiming a page or have no lru page/scanning. So if all_unreclaimable of the zone is set to 1, it is natural to contine remaining all_unreclaimable = 1 unless any page isn't freed or other LRU allocation happens. What should be fixed? > > > > What's wrong with my original patch? It appears reasonable if a zone has > > > a lot of free memory, don't set unreclaimable to it. > > > > As I said, all_unreclaimable doesn't mean "no free pages in the zone". > > So we shouldn't add new dependency between all_unreclaimable and # of free pages. > > > > And what's the purpose of high_wmark and some specific order check for it? > > Does it mean really "the zone has no free memory"? > > > > Your description try to explain about that and it seems to depend on outer/inner loop in balance_pgdat. > > (But it's very hard for dump me to parse :( ) > > I don't like that fragile code. If we might change it in future? > > Of course, we can add enough description about that but it means more complex. > > I don't want to add more complexity in kswapd unless we have a good reason. > > > > The most important thing is that the problem isn't related to # of free pages. > > As I state in my patch, the problem is caused by not considering LRU size changing. > > I would like to target a principal cause. > ok, if unreclaimable means no reclaimable pages. I'd consider the > original method (regards zone without lru pages as reclaimable), which Why do we have to consider zone which has no lru pages as reclaimable? It doesn't make sense to me. > is simpler. My original objection that method will regard zone which is > full and has no lru pages as reclaimable, but as you said reclaimable > isn't related to free pages. Yes. it's not related to free pages but lru pages. If we don't have no lru pages, we should mark the zone with all_unreclaimable. Of course, we should consider slab pages, too but it's not easy for VM to reclaim slabe pages compared to LRU pages so I decided ignoring it in this version but if anyone complains about that, I will consdier that, too. > I can't get a clear meaning for all_unreclaimable with your new patch. What I mean about all_unreclaimable is that the zone has no reclaimable pages which could be lru pages/slab pages. I ignore slab pages in this version > > Thanks, > Shaohua > -- Kinds regards, Minchan Kim -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Fight unfair telecom internet charges in Canada: sign http://stopthemeter.ca/ Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>