On Mon, Sep 06, 2010 at 02:47:16PM +0900, KAMEZAWA Hiroyuki wrote: > > Now, sysfs interface of memory hotplug shows whether the section is > removable or not. But it checks only migrateype of pages and doesn't > check details of cluster of pages. > This was deliberate at the time. The intention was to avoid an expensive linear page scan where possible. > Next, memory hotplug's set_migratetype_isolate() has the same kind > of check, too. But the migrate-type is just a "hint" and the pageblock > can contain several types of pages if fragmentation is very heavy. > If fragmentation is very heavy on a system that requires memory hot-plug, I'd also be checking the value of min_free_kbytes. If it's low, I suggest an init script runs hugeadm --set-recommended-min_free_kbytes because it'll keep fragmentation-related events to a minimum. The mm_page_alloc_extfrag tracepoint can be used to measure fragmentation events if you want to see the effect of altering min_free_kbytes like this. > To get precise information, we need to check > - the pageblock only contains free pages or LRU pages. > > This patch adds the function __count_unmovable_pages() and makes > above 2 checks to use the same logic. This will improve user experience > of memory hotplug because sysfs interface tells accurate information. > > Note: > it may be better to check MIGRATE_UNMOVABLE for making failure case quick. > > Changelog: 2010/09/06 > - added comments. > - removed zone->lock. > - changed the name of the function to be is_pageblock_removable_async(). > because I removed the zone->lock. > > Reported-by: Michal Hocko <mhocko@xxxxxxx> > Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@xxxxxxxxxxxxxx> > --- > include/linux/memory_hotplug.h | 1 > mm/memory_hotplug.c | 15 ------- > mm/page_alloc.c | 77 ++++++++++++++++++++++++++++++----------- > 3 files changed, 60 insertions(+), 33 deletions(-) > > Index: kametest/mm/page_alloc.c > =================================================================== > --- kametest.orig/mm/page_alloc.c > +++ kametest/mm/page_alloc.c > @@ -5274,11 +5274,61 @@ void set_pageblock_flags_group(struct pa > * page allocater never alloc memory from ISOLATE block. > */ > > +static int __count_immobile_pages(struct zone *zone, struct page *page) > +{ This will also count RECLAIMABLE pages belonging to some slab objects. These are potentially hot-removable if slab is shrunk. Your function gives a more accurate count but not necessarily a better user-experience with respect to finding sections to hot-remove. You might like to detect PageSlab pages that belong to a RECLAIMABLE slab and not count these as immobile. > + unsigned long pfn, iter, found; > + /* > + * For avoiding noise data, lru_add_drain_all() should be called > + * If ZONE_MOVABLE, the zone never contains immobile pages > + */ > + if (zone_idx(zone) == ZONE_MOVABLE) > + return 0; > + > + pfn = page_to_pfn(page); > + for (found = 0, iter = 0; iter < pageblock_nr_pages; iter++) { > + unsigned long check = pfn + iter; > + > + if (!pfn_valid_within(check)) { > + iter++; > + continue; > + } > + page = pfn_to_page(check); > + if (!page_count(page)) { > + if (PageBuddy(page)) > + iter += (1 << page_order(page)) - 1; > + continue; > + } > + if (!PageLRU(page)) > + found++; Arguably, you do not care how many pages there are, you just care if there is one truely unmovable page. If you find one of them, then have this function return fail to avoid the rest of the scan. > + /* > + * If the page is not RAM, page_count()should be 0. > + * we don't need more check. This is an _used_ not-movable page. > + * > + * The problematic thing here is PG_reserved pages. PG_reserved > + * is set to both of a memory hole page and a _used_ kernel > + * page at boot. > + */ > + } > + return found; > +} > + > +bool is_pageblock_removable_async(struct page *page) > +{ > + struct zone *zone = page_zone(page); > + unsigned long flags; > + int num; > + /* Don't take zone->lock interntionally. */ intentionally? > + num = __count_immobile_pages(zone, page); > + > + if (num) > + return false; > + return true; > +} > + > int set_migratetype_isolate(struct page *page) > { > struct zone *zone; > - struct page *curr_page; > - unsigned long flags, pfn, iter; > + unsigned long flags, pfn; > unsigned long immobile = 0; > struct memory_isolate_notify arg; > int notifier_ret; > @@ -5289,11 +5339,6 @@ int set_migratetype_isolate(struct page > zone_idx = zone_idx(zone); > > spin_lock_irqsave(&zone->lock, flags); > - if (get_pageblock_migratetype(page) == MIGRATE_MOVABLE || > - zone_idx == ZONE_MOVABLE) { > - ret = 0; > - goto out; > - } > This will result in more scanning and a potentially more expensive memory hot-remove operation. I'm not massively concerned as such because memory hot-remove is not cheap but it's worth mentioning in the changelog that this is a consequence. > pfn = page_to_pfn(page); > arg.start_pfn = pfn; > @@ -5315,19 +5360,13 @@ int set_migratetype_isolate(struct page > notifier_ret = notifier_to_errno(notifier_ret); > if (notifier_ret) > goto out; > + immobile = __count_immobile_pages(zone ,page); > > - for (iter = pfn; iter < (pfn + pageblock_nr_pages); iter++) { > - if (!pfn_valid_within(pfn)) > - continue; > - > - curr_page = pfn_to_page(iter); > - if (!page_count(curr_page) || PageLRU(curr_page)) > - continue; > - > - immobile++; > - } > - > - if (arg.pages_found == immobile) > + /* > + * immobile means "not-on-lru" paes. If immobile is larger than > + * removable-by-driver pages reported by notifier, we'll fail. > + */ > + if (!immobile || arg.pages_found >= immobile) > ret = 0; > Here is where I'd suggest reimplementing __count_immobile_pages as pageblock_any_immobile() that returns true if it detects an immobile page in a given PFN range. > out: > Index: kametest/mm/memory_hotplug.c > =================================================================== > --- kametest.orig/mm/memory_hotplug.c > +++ kametest/mm/memory_hotplug.c > @@ -602,27 +602,14 @@ static struct page *next_active_pagebloc > /* Checks if this range of memory is likely to be hot-removable. */ > int is_mem_section_removable(unsigned long start_pfn, unsigned long nr_pages) > { > - int type; > struct page *page = pfn_to_page(start_pfn); > struct page *end_page = page + nr_pages; > > /* Check the starting page of each pageblock within the range */ > for (; page < end_page; page = next_active_pageblock(page)) { > - type = get_pageblock_migratetype(page); > - > - /* > - * A pageblock containing MOVABLE or free pages is considered > - * removable > - */ > - if (type != MIGRATE_MOVABLE && !pageblock_free(page)) > + if (!is_pageblock_removable_async(page)) > return 0; > > - /* > - * A pageblock starting with a PageReserved page is not > - * considered removable. > - */ > - if (PageReserved(page)) > - return 0; > } Bear in mind that a user or bad application constantly reading the sysfs file potentially causes a lot of cache trashing as a result of the linear scan instead of the pageblock type check. > > /* All pageblocks in the memory block are likely to be hot-removable */ > Index: kametest/include/linux/memory_hotplug.h > =================================================================== > --- kametest.orig/include/linux/memory_hotplug.h > +++ kametest/include/linux/memory_hotplug.h > @@ -69,6 +69,7 @@ extern void online_page(struct page *pag > /* VM interface that may be used by firmware interface */ > extern int online_pages(unsigned long, unsigned long); > extern void __offline_isolated_pages(unsigned long, unsigned long); > +extern bool is_pageblock_removable_async(struct page *page); > > /* reasonably generic interface to expand the physical pages in a zone */ > extern int __add_pages(int nid, struct zone *zone, unsigned long start_pfn, > -- Mel Gorman Part-time Phd Student Linux Technology Center University of Limerick IBM Dublin Software Lab -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxxx For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>