On Thu, Jun 19, 2014 at 01:13:22PM -0700, Andrew Morton wrote: > On Thu, 19 Jun 2014 10:02:39 +0900 Minchan Kim <minchan@xxxxxxxxxx> wrote: > > > > > @@ -2057,8 +2057,7 @@ out: > > > > static void shrink_lruvec(struct lruvec *lruvec, struct scan_control > > > > *sc) > > > > { > > > > unsigned long nr[NR_LRU_LISTS]; > > > > - unsigned long targets[NR_LRU_LISTS]; > > > > - unsigned long nr_to_scan; > > > > + unsigned long file_target, anon_target; > > > > > > > > >From the above snippet, we can know that the "percent" locals come from > > > > targets[NR_LRU_LISTS]. So this fix does not increase the stack. > > > > > > OK. But I expect the stack use could be decreased by using more > > > complex expressions. > > > > I didn't look at this patch yet but want to say. > > > > The expression is not easy to follow since several people already > > confused/discuss/fixed a bit so I'd like to put more concern to clarity > > rather than stack footprint. > > That code is absolutely awful :( It's terribly difficult to work out > what the design is - what the code is actually setting out to achieve. > One is reduced to trying to reverse-engineer the intent from the > implementation and that becomes near impossible when the > implementation has bugs! > > Look at this miserable comment: > > /* > * For kswapd and memcg, reclaim at least the number of pages > * requested. Ensure that the anon and file LRUs are scanned > * proportionally what was requested by get_scan_count(). We > * stop reclaiming one LRU and reduce the amount scanning > * proportional to the original scan target. > */ > > > > For kswapd and memcg, reclaim at least the number of pages > > requested. > > *why*? > At the time of writing the intention was to reduce direct reclaim stall latency in the global case. Initially the following block was above it /* * For global direct reclaim, reclaim only the number of pages * requested. Less care is taken to scan proportionally as it * is more important to minimise direct reclaim stall latency * than it is to properly age the LRU lists. */ if (global_reclaim(sc) && !current_is_kswapd()) break; When that comment was removed then the remaining comment is less clear. > > Ensure that the anon and file LRUs are scanned > > proportionally what was requested by get_scan_count(). > > Ungramattical. Lacks specificity. Fails to explain *why*. > In the normal case, file/anon LRUs are scanned at a rate proportional to the value of vm.swappiness. get_scan_count() calculates the number of pages to scan from each LRU taking into account additional factors such as the availability of swap. When the requested number of pages have been reclaimed we adjust to scan targets to minimise the number of pages scanned while maintaining the ratio of file/anon pages that are scanned. -- Mel Gorman SUSE Labs -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>