On Mon, Dec 19, 2011 at 01:26:15PM +0000, Mel Gorman wrote: > On Sun, Dec 18, 2011 at 01:08:22AM +0900, Minchan Kim wrote: > > On Wed, Dec 14, 2011 at 03:41:33PM +0000, Mel Gorman wrote: > > > It was observed that scan rates from direct reclaim during tests > > > writing to both fast and slow storage were extraordinarily high. The > > > problem was that while pages were being marked for immediate reclaim > > > when writeback completed, the same pages were being encountered over > > > and over again during LRU scanning. > > > > > > This patch isolates file-backed pages that are to be reclaimed when > > > clean on their own LRU list. > > > > Please include your test result about reducing CPU usage. > > It makes this separate LRU list how vaule is. > > > > It's in the leader. The writebackCPDevicevfat tests should that System > CPU goes from 46.40 seconds to 4.44 seconds with this patch applied. Sorry I didn't read cover. Looks great. > > > > <SNIP> > > > > > > diff --git a/mm/swap.c b/mm/swap.c > > > index a91caf7..9973975 100644 > > > --- a/mm/swap.c > > > +++ b/mm/swap.c > > > @@ -39,6 +39,7 @@ int page_cluster; > > > > > > static DEFINE_PER_CPU(struct pagevec[NR_LRU_LISTS], lru_add_pvecs); > > > static DEFINE_PER_CPU(struct pagevec, lru_rotate_pvecs); > > > +static DEFINE_PER_CPU(struct pagevec, lru_putback_immediate_pvecs); > > > static DEFINE_PER_CPU(struct pagevec, lru_deactivate_pvecs); > > > > > > /* > > > @@ -255,24 +256,80 @@ static void pagevec_move_tail(struct pagevec *pvec) > > > } > > > > > > /* > > > + * Similar pair of functions to pagevec_move_tail except it is called when > > > + * moving a page from the LRU_IMMEDIATE to one of the [in]active_[file|anon] > > > + * lists > > > + */ > > > +static void pagevec_putback_immediate_fn(struct page *page, void *arg) > > > +{ > > > + struct zone *zone = page_zone(page); > > > + > > > + if (PageLRU(page)) { > > > + enum lru_list lru = page_lru(page); > > > + list_move(&page->lru, &zone->lru[lru].list); > > > + } > > > +} > > > + > > > +static void pagevec_putback_immediate(struct pagevec *pvec) > > > +{ > > > + pagevec_lru_move_fn(pvec, pagevec_putback_immediate_fn, NULL); > > > +} > > > + > > > +/* > > > * Writeback is about to end against a page which has been marked for immediate > > > * reclaim. If it still appears to be reclaimable, move it to the tail of the > > > * inactive list. > > > */ > > > void rotate_reclaimable_page(struct page *page) > > > { > > > + struct zone *zone = page_zone(page); > > > + struct list_head *page_list; > > > + struct pagevec *pvec; > > > + unsigned long flags; > > > + > > > + page_cache_get(page); > > > + local_irq_save(flags); > > > + __mod_zone_page_state(zone, NR_IMMEDIATE, -1); > > > + > > > > I am not sure underflow never happen. > > We do SetPageReclaim at several places but dont' increase NR_IMMEDIATE. > > > > In those cases, we do not move the page to the immedate list either. That's my concern. We didn't move the page to immediate list but set SetPageReclaim. It means we don't increate NR_IMMEDIATE. If end_page_writeback have called that page, rotate_reclimable_page would be called. Eventually, __mod_zone_page_state(zone, NR_IMMEDIATE, -1) is called. But I didn't look into the code yet for confirming it's possbile or not. > During one test I was recording /proc/vmstat every 10 seconds and never > saw an underflow. If it's very rare, it would be very hard to see it. > > > > if (!PageLocked(page) && !PageDirty(page) && !PageActive(page) && > > > !PageUnevictable(page) && PageLRU(page)) { > > > - struct pagevec *pvec; > > > - unsigned long flags; > > > > > > - page_cache_get(page); > > > - local_irq_save(flags); > > > pvec = &__get_cpu_var(lru_rotate_pvecs); > > > if (!pagevec_add(pvec, page)) > > > pagevec_move_tail(pvec); > > > - local_irq_restore(flags); > > > + } else { > > > + pvec = &__get_cpu_var(lru_putback_immediate_pvecs); > > > + if (!pagevec_add(pvec, page)) > > > + pagevec_putback_immediate(pvec); > > > > Nitpick about naming. > > Naming is important. > > > It doesn't say immediate is from or to. So I got confused > > which is source. I know comment of function already say it > > but good naming can reduce unnecessary comment. > > How about pagevec_putback_from_immediate_list? > > > > Sure. Done. > > > > + } > > > + > > > + /* > > > + * There is a potential race that if a page is set PageReclaim > > > + * and moved to the LRU_IMMEDIATE list after writeback completed, > > > + * it can be left on the LRU_IMMEDATE list with no way for > > > + * reclaim to find it. > > > + * > > > + * This race should be very rare but count how often it happens. > > > + * If it is a continual race, then it's very unsatisfactory as there > > > + * is no guarantee that rotate_reclaimable_page() will be called > > > + * to rescue these pages but finding them in page reclaim is also > > > + * problematic due to the problem of deciding when the right time > > > + * to scan this list is. > > > + */ > > > + page_list = &zone->lru[LRU_IMMEDIATE].list; > > > + if (!zone_page_state(zone, NR_IMMEDIATE) && !list_empty(page_list)) { > > > > How about this > > > > if (zone_page_state(zone, NR_IMMEDIATE)) { > > page_list = &zone->lru[LRU_IMMEDIATE].list; > > if (!list_empty(page_list)) > > ... > > ... > > } > > > > It can reduce a unnecessary reference. > > > > Ok, it mucks up the indentation a bit but with some renaming it looks > reasonable. > > > > + struct page *page; > > > + > > > + spin_lock(&zone->lru_lock); > > > + while (!list_empty(page_list)) { > > > + page = list_entry(page_list->prev, struct page, lru); > > > + list_move(&page->lru, &zone->lru[page_lru(page)].list); > > > + __count_vm_event(PGRESCUED); > > > + } > > > + spin_unlock(&zone->lru_lock); > > > } > > > + > > > + local_irq_restore(flags); > > > } > > > > > > static void update_page_reclaim_stat(struct zone *zone, struct page *page, > > > @@ -475,6 +532,13 @@ static void lru_deactivate_fn(struct page *page, void *arg) > > > * is _really_ small and it's non-critical problem. > > > */ > > > SetPageReclaim(page); > > > + > > > + /* > > > + * Move to the LRU_IMMEDIATE list to avoid being scanned > > > + * by page reclaim uselessly. > > > + */ > > > + list_move_tail(&page->lru, &zone->lru[LRU_IMMEDIATE].list); > > > + __mod_zone_page_state(zone, NR_IMMEDIATE, 1); > > > > It mekes below count of PGDEACTIVATE wrong in lru_deactivate_fn. > > Before this patch, all is from active to inacive so it was right. > > But with this patch, it can be from acdtive to immediate. > > > > I do not quite understand. PGDEACTIVATE is incremented if the page was > active and this is checked before the move to the immediate LRU. Whether > it moves to the immediate LRU or the end of the inactive list, it is > still a deactivation. What's wrong with incrementing the count if it Hmm, I have thought deactivation is only from active to deactive. I might be wrong but if we perhaps move page from active to unevictable list, is it deactivation, too? Maybe we need consistent count. > moves from active to immediate? > > ==== CUT HERE ==== > mm: Isolate pages for immediate reclaim on their own LRU fix > > Rename pagevec_putback_immediate_fn to pagevec_putback_from_immediate_fn > for clarity and alter flow of rotate_reclaimable_page() slightly to > avoid an unnecessary list reference. > > This is a fix to the patch > mm-isolate-pages-for-immediate-reclaim-on-their-own-lru.patch in mmotm. > > Signed-off-by: Mel Gorman <mgorman@xxxxxxx> Acked-by: Minchan Kim <minchan@xxxxxxxxxx> Thanks. -- Kind regards, Minchan Kim -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Fight unfair telecom internet charges in Canada: sign http://stopthemeter.ca/ Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>