On Fri, Apr 08, 2022 at 03:12:20PM +0800, Huang Ying wrote: > To optimize page placement in a memory tiering system with NUMA > balancing, the hot pages in the slow memory node need to be > identified. Essentially, the original NUMA balancing implementation > selects the mostly recently accessed (MRU) pages as the hot pages. > But this isn't a very good algorithm to identify the hot pages. > > So, in this patch we implemented a better hot page selection > algorithm. Which is based on NUMA balancing page table scanning and > hint page fault as follows, > > - When the page tables of the processes are scanned to change PTE/PMD > to be PROT_NONE, the current time is recorded in struct page as scan > time. > > - When the page is accessed, hint page fault will occur. The scan > time is gotten from the struct page. And The hint page fault > latency is defined as > > hint page fault time - scan time > > The shorter the hint page fault latency of a page is, the higher the > probability of their access frequency to be higher. So the hint page > fault latency is a good estimation of the page hot/cold. > > But it's hard to find some extra space in struct page to hold the scan > time. Fortunately, we can reuse some bits used by the original NUMA > balancing. > > NUMA balancing uses some bits in struct page to store the page > accessing CPU and PID (referring to page_cpupid_xchg_last()). Which > is used by the multi-stage node selection algorithm to avoid to > migrate pages shared accessed by the NUMA nodes back and forth. But > for pages in the slow memory node, even if they are shared accessed by > multiple NUMA nodes, as long as the pages are hot, they need to be > promoted to the fast memory node. So the accessing CPU and PID > information are unnecessary for the slow memory pages. We can reuse > these bits in struct page to record the scan time for them. For the > fast memory pages, these bits are used as before. > > For the hot threshold, the default value is 1 second, which works well > in our performance test. All pages with hint page fault latency < the > threshold will be considered hot. A debugfs interface is also > provided to adjust the hot threshold. > > The downside of the above method is that the response time to the > workload hot spot changing may be much longer. For example, > > - A previous cold memory area becomes hot > > - The hint page fault will be triggered. But the hint page fault > latency isn't shorter than the hot threshold. So the pages will > not be promoted. > > - When the memory area is scanned again, maybe after a scan period, > the hint page fault latency measured will be shorter than the hot > threshold and the pages will be promoted. > > To mitigate this, > > - If there are enough free space in the fast memory node, the hot > threshold will not be used, all pages will be promoted upon the hint > page fault for fast response. > > - If fast response is more important for system performance, the > administrator can set a higher hot threshold. > > Thanks Zhong Jiang reported and tested the fix for a bug when > disabling memory tiering mode dynamically. > > Signed-off-by: "Huang, Ying" <ying.huang@xxxxxxxxx> > Cc: Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> > Cc: Michal Hocko <mhocko@xxxxxxxx> > Cc: Rik van Riel <riel@xxxxxxxxxxx> > Cc: Mel Gorman <mgorman@xxxxxxxxxxxxxxxxxxx> > Cc: Peter Zijlstra <peterz@xxxxxxxxxxxxx> > Cc: Dave Hansen <dave.hansen@xxxxxxxxxxxxxxx> > Cc: Yang Shi <shy828301@xxxxxxxxx> > Cc: Zi Yan <ziy@xxxxxxxxxx> > Cc: Wei Xu <weixugc@xxxxxxxxxx> > Cc: osalvador <osalvador@xxxxxxx> > Cc: Shakeel Butt <shakeelb@xxxxxxxxxx> > Cc: Zhong Jiang <zhongjiang-ali@xxxxxxxxxxxxxxxxx> > Cc: linux-kernel@xxxxxxxxxxxxxxx > Cc: linux-mm@xxxxxxxxx > --- > include/linux/mm.h | 30 ++++++++++++++++++ > kernel/sched/debug.c | 1 + > kernel/sched/fair.c | 74 ++++++++++++++++++++++++++++++++++++++++++++ > kernel/sched/sched.h | 1 + > mm/huge_memory.c | 13 ++++++-- > mm/memory.c | 11 ++++++- > mm/migrate.c | 12 +++++++ > mm/mprotect.c | 8 ++++- > 8 files changed, 145 insertions(+), 5 deletions(-) > > diff --git a/include/linux/mm.h b/include/linux/mm.h > index e34edb775334..455a3d0e699d 100644 > --- a/include/linux/mm.h > +++ b/include/linux/mm.h > @@ -1311,6 +1311,18 @@ static inline int folio_nid(const struct folio *folio) > } > > #ifdef CONFIG_NUMA_BALANCING > +/* page access time bits needs to hold at least 4 seconds */ > +#define PAGE_ACCESS_TIME_MIN_BITS 12 > +#if LAST_CPUPID_SHIFT < PAGE_ACCESS_TIME_MIN_BITS > +#define PAGE_ACCESS_TIME_BUCKETS \ > + (PAGE_ACCESS_TIME_MIN_BITS - LAST_CPUPID_SHIFT) > +#else > +#define PAGE_ACCESS_TIME_BUCKETS 0 > +#endif > + > +#define PAGE_ACCESS_TIME_MASK \ > + (LAST_CPUPID_MASK << PAGE_ACCESS_TIME_BUCKETS) > + > static inline int cpu_pid_to_cpupid(int cpu, int pid) > { > return ((cpu & LAST__CPU_MASK) << LAST__PID_SHIFT) | (pid & LAST__PID_MASK); > @@ -1346,6 +1358,11 @@ static inline bool __cpupid_match_pid(pid_t task_pid, int cpupid) > return (task_pid & LAST__PID_MASK) == cpupid_to_pid(cpupid); > } > > +static inline bool check_cpupid(int cpupid) > +{ > + return cpupid_to_cpu(cpupid) < nr_cpu_ids; > +} > + > #define cpupid_match_pid(task, cpupid) __cpupid_match_pid(task->pid, cpupid) > #ifdef LAST_CPUPID_NOT_IN_PAGE_FLAGS > static inline int page_cpupid_xchg_last(struct page *page, int cpupid) > @@ -1374,12 +1391,25 @@ static inline void page_cpupid_reset_last(struct page *page) > page->flags |= LAST_CPUPID_MASK << LAST_CPUPID_PGSHIFT; > } > #endif /* LAST_CPUPID_NOT_IN_PAGE_FLAGS */ > + > +static inline int xchg_page_access_time(struct page *page, int time) > +{ > + int last_time; > + > + last_time = page_cpupid_xchg_last(page, time >> PAGE_ACCESS_TIME_BUCKETS); > + return last_time << PAGE_ACCESS_TIME_BUCKETS; > +} > #else /* !CONFIG_NUMA_BALANCING */ > static inline int page_cpupid_xchg_last(struct page *page, int cpupid) > { > return page_to_nid(page); /* XXX */ > } > > +static inline int xchg_page_access_time(struct page *page, int time) > +{ > + return 0; > +} > + > static inline int page_cpupid_last(struct page *page) > { > return page_to_nid(page); /* XXX */ > diff --git a/kernel/sched/debug.c b/kernel/sched/debug.c > index bb3d63bdf4ae..ad63dbfc54f1 100644 > --- a/kernel/sched/debug.c > +++ b/kernel/sched/debug.c > @@ -333,6 +333,7 @@ static __init int sched_init_debug(void) > debugfs_create_u32("scan_period_min_ms", 0644, numa, &sysctl_numa_balancing_scan_period_min); > debugfs_create_u32("scan_period_max_ms", 0644, numa, &sysctl_numa_balancing_scan_period_max); > debugfs_create_u32("scan_size_mb", 0644, numa, &sysctl_numa_balancing_scan_size); > + debugfs_create_u32("hot_threshold_ms", 0644, numa, &sysctl_numa_balancing_hot_threshold); > #endif > > debugfs_create_file("debug", 0444, debugfs_sched, NULL, &sched_debug_fops); > diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c > index d4bd299d67ab..cb130ea46c71 100644 > --- a/kernel/sched/fair.c > +++ b/kernel/sched/fair.c > @@ -1058,6 +1058,9 @@ unsigned int sysctl_numa_balancing_scan_size = 256; > /* Scan @scan_size MB every @scan_period after an initial @scan_delay in ms */ > unsigned int sysctl_numa_balancing_scan_delay = 1000; > > +/* The page with hint page fault latency < threshold in ms is considered hot */ > +unsigned int sysctl_numa_balancing_hot_threshold = 1000; > + > struct numa_group { > refcount_t refcount; > > @@ -1400,6 +1403,37 @@ static inline unsigned long group_weight(struct task_struct *p, int nid, > return 1000 * faults / total_faults; > } > > +static bool pgdat_free_space_enough(struct pglist_data *pgdat) > +{ > + int z; > + unsigned long enough_mark; > + > + enough_mark = max(1UL * 1024 * 1024 * 1024 >> PAGE_SHIFT, > + pgdat->node_present_pages >> 4); > + for (z = pgdat->nr_zones - 1; z >= 0; z--) { > + struct zone *zone = pgdat->node_zones + z; > + > + if (!populated_zone(zone)) > + continue; > + > + if (zone_watermark_ok(zone, 0, > + high_wmark_pages(zone) + enough_mark, > + ZONE_MOVABLE, 0)) > + return true; > + } > + return false; > +} > + > +static int numa_hint_fault_latency(struct page *page) > +{ > + int last_time, time; > + > + time = jiffies_to_msecs(jiffies); > + last_time = xchg_page_access_time(page, time); > + > + return (time - last_time) & PAGE_ACCESS_TIME_MASK; This code can possibly consider cold page as hot, Assume, LAST_CPUPID_SHIFT = 12 PAGE_ACCESS_TIME_BUCKETS = 0 sysctl_numa_balancing_hot_threshold = 1000 Assume while changing pte, jiffies_to_msecs(jiffies) = 0xAABB0100 So value saved in page->flags will be lowest 12 bits of 0xAABB0100 which is 0x100. Assume when numa_hint_fault_latency() gets called, time = jiffies_to_msecs(jiffies) = 0xAACC0100 So, time = 0xAACC0100, and last_time = 0x100, time - last_time = 0xAACC0100 - 0x100 = 0xAACC0000 0xAACC0000 & PAGE_ACCESS_TIME_MASK = 0xAACC0000 & ((1 << 12) - 1) = 0 so the return value of this function is 0, the code will consider it as hot page but it is cold page because actual difference is 0xAACC0100 - 0xAABB0100 = 110000 ms There may be more such scenarios. What do you think? > +} > + > bool should_numa_migrate_memory(struct task_struct *p, struct page * page, > int src_nid, int dst_cpu) > { > @@ -1407,9 +1441,38 @@ bool should_numa_migrate_memory(struct task_struct *p, struct page * page, > int dst_nid = cpu_to_node(dst_cpu); > int last_cpupid, this_cpupid; > > + /* > + * The pages in slow memory node should be migrated according > + * to hot/cold instead of accessing CPU node. > + */ > + if (sysctl_numa_balancing_mode & NUMA_BALANCING_MEMORY_TIERING && > + !node_is_toptier(src_nid)) { > + struct pglist_data *pgdat; > + unsigned long latency, th; > + > + pgdat = NODE_DATA(dst_nid); > + if (pgdat_free_space_enough(pgdat)) > + return true; > + > + th = sysctl_numa_balancing_hot_threshold; > + latency = numa_hint_fault_latency(page); > + if (latency >= th) > + return false; > + > + return true; > + } > + > this_cpupid = cpu_pid_to_cpupid(dst_cpu, current->pid); > last_cpupid = page_cpupid_xchg_last(page, this_cpupid); > > + /* > + * The cpupid may be invalid when NUMA_BALANCING_MEMORY_TIERING > + * is disabled dynamically. > + */ > + if (!(sysctl_numa_balancing_mode & NUMA_BALANCING_MEMORY_TIERING) && > + !node_is_toptier(src_nid) && !check_cpupid(last_cpupid)) > + return false; > + > /* > * Allow first faults or private faults to migrate immediately early in > * the lifetime of a task. The magic number 4 is based on waiting for > @@ -2642,6 +2705,17 @@ void task_numa_fault(int last_cpupid, int mem_node, int pages, int flags) > if (!p->mm) > return; > > + /* > + * NUMA faults statistics are unnecessary for the slow memory node. > + * > + * And, the cpupid may be invalid when NUMA_BALANCING_MEMORY_TIERING > + * is disabled dynamically. > + */ > + if (!node_is_toptier(mem_node) && > + (sysctl_numa_balancing_mode & NUMA_BALANCING_MEMORY_TIERING || > + !check_cpupid(last_cpupid))) > + return; > + > /* Allocate buffer to track faults on a per-node basis */ > if (unlikely(!p->numa_faults)) { > int size = sizeof(*p->numa_faults) * > diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h > index 58263f90c559..86ce60d3d472 100644 > --- a/kernel/sched/sched.h > +++ b/kernel/sched/sched.h > @@ -2408,6 +2408,7 @@ extern unsigned int sysctl_numa_balancing_scan_delay; > extern unsigned int sysctl_numa_balancing_scan_period_min; > extern unsigned int sysctl_numa_balancing_scan_period_max; > extern unsigned int sysctl_numa_balancing_scan_size; > +extern unsigned int sysctl_numa_balancing_hot_threshold; > #endif > > #ifdef CONFIG_SCHED_HRTICK > diff --git a/mm/huge_memory.c b/mm/huge_memory.c > index 2fe38212e07c..3a715eeeebb5 100644 > --- a/mm/huge_memory.c > +++ b/mm/huge_memory.c > @@ -1401,7 +1401,7 @@ vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf) > struct page *page; > unsigned long haddr = vmf->address & HPAGE_PMD_MASK; > int page_nid = NUMA_NO_NODE; > - int target_nid, last_cpupid = -1; > + int target_nid, last_cpupid = (-1 & LAST_CPUPID_MASK); > bool migrated = false; > bool was_writable = pmd_savedwrite(oldpmd); > int flags = 0; > @@ -1422,7 +1422,8 @@ vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf) > flags |= TNF_NO_GROUP; > > page_nid = page_to_nid(page); > - last_cpupid = page_cpupid_last(page); > + if (node_is_toptier(page_nid)) > + last_cpupid = page_cpupid_last(page); > target_nid = numa_migrate_prep(page, vma, haddr, page_nid, > &flags); > > @@ -1740,6 +1741,7 @@ int change_huge_pmd(struct vm_area_struct *vma, pmd_t *pmd, > > if (prot_numa) { > struct page *page; > + bool toptier; > /* > * Avoid trapping faults against the zero page. The read-only > * data is likely to be read-cached on the local CPU and > @@ -1752,13 +1754,18 @@ int change_huge_pmd(struct vm_area_struct *vma, pmd_t *pmd, > goto unlock; > > page = pmd_page(*pmd); > + toptier = node_is_toptier(page_to_nid(page)); > /* > * Skip scanning top tier node if normal numa > * balancing is disabled > */ > if (!(sysctl_numa_balancing_mode & NUMA_BALANCING_NORMAL) && > - node_is_toptier(page_to_nid(page))) > + toptier) > goto unlock; > + > + if (sysctl_numa_balancing_mode & NUMA_BALANCING_MEMORY_TIERING && > + !toptier) > + xchg_page_access_time(page, jiffies_to_msecs(jiffies)); > } > /* > * In case prot_numa, we are under mmap_read_lock(mm). It's critical > diff --git a/mm/memory.c b/mm/memory.c > index b370ed118767..a8ac15ce7a75 100644 > --- a/mm/memory.c > +++ b/mm/memory.c > @@ -74,6 +74,7 @@ > #include <linux/perf_event.h> > #include <linux/ptrace.h> > #include <linux/vmalloc.h> > +#include <linux/sched/sysctl.h> > > #include <trace/events/kmem.h> > > @@ -4455,8 +4456,16 @@ static vm_fault_t do_numa_page(struct vm_fault *vmf) > if (page_mapcount(page) > 1 && (vma->vm_flags & VM_SHARED)) > flags |= TNF_SHARED; > > - last_cpupid = page_cpupid_last(page); > page_nid = page_to_nid(page); > + /* > + * In memory tiering mode, cpupid of slow memory page is used > + * to record page access time. So use default value. > + */ > + if ((sysctl_numa_balancing_mode & NUMA_BALANCING_MEMORY_TIERING) && > + !node_is_toptier(page_nid)) > + last_cpupid = (-1 & LAST_CPUPID_MASK); > + else > + last_cpupid = page_cpupid_last(page); > target_nid = numa_migrate_prep(page, vma, vmf->address, page_nid, > &flags); > if (target_nid == NUMA_NO_NODE) { > diff --git a/mm/migrate.c b/mm/migrate.c > index dc84edfae842..e73f26dfeb38 100644 > --- a/mm/migrate.c > +++ b/mm/migrate.c > @@ -534,6 +534,18 @@ void folio_migrate_flags(struct folio *newfolio, struct folio *folio) > * future migrations of this same page. > */ > cpupid = page_cpupid_xchg_last(&folio->page, -1); > + /* > + * If migrate between slow and fast memory node, reset cpupid, > + * because that is used to record page access time in slow > + * memory node > + */ > + if (sysctl_numa_balancing_mode & NUMA_BALANCING_MEMORY_TIERING) { > + bool f_toptier = node_is_toptier(page_to_nid(&folio->page)); > + bool t_toptier = node_is_toptier(page_to_nid(&newfolio->page)); > + > + if (f_toptier != t_toptier) > + cpupid = -1; > + } > page_cpupid_xchg_last(&newfolio->page, cpupid); > > folio_migrate_ksm(newfolio, folio); > diff --git a/mm/mprotect.c b/mm/mprotect.c > index b69ce7a7b2b7..e7cb90d84fac 100644 > --- a/mm/mprotect.c > +++ b/mm/mprotect.c > @@ -85,6 +85,7 @@ static unsigned long change_pte_range(struct vm_area_struct *vma, pmd_t *pmd, > if (prot_numa) { > struct page *page; > int nid; > + bool toptier; > > /* Avoid TLB flush if possible */ > if (pte_protnone(oldpte)) > @@ -114,14 +115,19 @@ static unsigned long change_pte_range(struct vm_area_struct *vma, pmd_t *pmd, > nid = page_to_nid(page); > if (target_node == nid) > continue; > + toptier = node_is_toptier(nid); > > /* > * Skip scanning top tier node if normal numa > * balancing is disabled > */ > if (!(sysctl_numa_balancing_mode & NUMA_BALANCING_NORMAL) && > - node_is_toptier(nid)) > + toptier) > continue; > + if (sysctl_numa_balancing_mode & NUMA_BALANCING_MEMORY_TIERING && > + !toptier) > + xchg_page_access_time(page, > + jiffies_to_msecs(jiffies)); > } > > oldpte = ptep_modify_prot_start(vma, addr, pte); > -- > 2.30.2 > >