Re: another allocation livelock with zram

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, Nov 26, 2012 at 05:00:26PM -0800, Luigi Semenzato wrote:
> Still no measurable progress on this one, but some new information.
> To recapitulate:
> 
> --- vanilla 3.4 kernel + hacky min_filelist_kbytes patch + Minchan's
> patch below:
> 
> >> > --- a/mm/vmscan.c
> >> > +++ b/mm/vmscan.c
> >> > @@ -2101,7 +2101,7 @@ static bool all_unreclaimable(struct zonelist *zonelist,
> >> >                         continue;
> >> >                 if (!cpuset_zone_allowed_hardwall(zone, GFP_KERNEL))
> >> >                         continue;
> >> > -               if (!zone->all_unreclaimable)
> >> > +               if (zone->pages_scanned < zone_reclaimable_pages(zone) * 6)
> >> >                         return false;
> >> >         }
> 
> --- no longer running the Chrome browser; instead, running this
> synthetic load: several instances of a process that allocates 200MB,
> then touches some subset of its pages in an endless loop.  The
> process's data segment compresses well (10:1).
> 
> --- running the load on two similar systems: one ARM-based, the other
> x86-based.  Both systems run the same kernel and the same image
> (different but equivalent configurations).  Both have 2 GB RAM.
> 
> On the x86 system, the mm behaves as expected.  All 3 Gb of the zram
> device are consumed before OOM-kills happen.
> 
> On the ARM system, OOM kills start happening when there are still
> about 2.1 GB of swap available.  Because the compression ratio is so
> good, the zram disk is only using 100 to 150 MB.
> 
> The systems are pretty similar.  The x86 device has a rotating disk,
> vs. SSD on the ARM device.  This could affect the speed of paging in
> code, but the program is very small so I don't think that's a factor.
> 
> There are no messages from zram in the log.
> 
> It could be either an ARM bug, or maybe the bug is on both systems,
> and the performance behavior on ARM is different enough to expose it.

The scenario I can imagine is by kswapd.
Did you tried to move wakeup_all_kswapd to rebalance below?

The reason why I guess so is direct reclaim which caused by zram page
allocation is normally GFP_NOIO. It has big limits to reclaim pages,
for instance, it can't page out dirty pages and swap out of anon pages
for avoding deadlock so scanning rate of LRU would be high easily for 
finding easy-reclaimable pages(ie, clean page, likely code). Even, you
used min_filelist_kbytes, which prevent to reclaim clean pages which
are only reclaimable pages in case of GFP_NOIO.

High scanning speed without freeing the page would be short cut to OOM load.

> 
> I will continue trying to figure out why kswapd isn't more proactive on ARM.

Just for the hint. I'm not sure it's really problem.
Historically, kswapd had problems about balancing of zones,
(ex, size of zones for balancing) and it was not too long that ARM started
to use multiple zones with big RAM.

> 
> Thanks!
> Luigi
> 
> --
> To unsubscribe, send a message with 'unsubscribe linux-mm' in
> the body to majordomo@xxxxxxxxx.  For more info on Linux MM,
> see: http://www.linux-mm.org/ .
> Don't email: <a href=mailto:"dont@xxxxxxxxx";> email@xxxxxxxxx </a>

-- 
Kind regards,
Minchan Kim

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@xxxxxxxxx.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@xxxxxxxxx";> email@xxxxxxxxx </a>


[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [ECOS]     [Asterisk Internet PBX]     [Linux API]