Hi Bob, (2013/09/10 23:28), Bob Liu wrote: > Hi Yasuaki, > > On 09/10/2013 10:51 AM, Yasuaki Ishimatsu wrote: >> (2013/09/02 12:45), Bob Liu wrote: >>> Currently khugepaged will try to merge HPAGE_PMD_NR normal pages to a huge page >>> which is allocated from the node of the first normal page, this policy is very >>> rough and may affect userland applications. >>> Andrew Davidoff reported a related issue several days ago. >>> >>> Using "numactl --interleave=all ./test" to run the testcase, but the result >>> wasn't not as expected. >>> cat /proc/2814/numa_maps: >>> 7f50bd440000 interleave:0-3 anon=51403 dirty=51403 N0=435 N1=435 N2=435 >>> N3=50098 >>> The end results showed that most pages are from Node3 instead of interleave >>> among node0-3 which was unreasonable. >>> >> >>> This patch adds a more complicated policy. >>> When searching HPAGE_PMD_NR normal pages, record which node those pages come >>> from. Alway allocate hugepage from the node with the max record. If several >>> nodes have the same max record, try to interleave among them. >> >> I don't understand this policy. Why does ths patch allocate hugepage from the >> node with the max record? >> > > Thanks for your review. > > The reason is hupepaged always allocate huge pages from the node of the > first scanned normal page, which may break the original page-balancing > among all nodes. > Thinking about the case that the first scanned normal page is allocated > from node A, most of other scanned normal pages are allocated from node > B or C.. > But khugepaged will always allocate the huge page from node A which > will cause extra memory pressure on node A and is not the same as users > expected. > > The policy I used in this patch(allocate huge page from the node with > max record)is try to minimize the effect to original page balancing. > > The other thing is even normal pages are allocated from Node A,B and C > equally, after khugepaged started Node A will also suffer from memory > pressure because of huge pages. Thank you for your explanation. I understood it. > >>> >>> After this patch the result was as expected: >>> 7f78399c0000 interleave:0-3 anon=51403 dirty=51403 N0=12723 N1=12723 N2=13235 >>> N3=12722 >>> >>> The simple testcase is like this: >>> #include<stdio.h> >>> #include<stdlib.h> >>> >>> int main() { >>> char *p; >>> int i; >>> int j; >>> >>> for (i=0; i < 200; i++) { >>> p = (char *)malloc(1048576); >>> printf("malloc done\n"); >>> >>> if (p == 0) { >>> printf("Out of memory\n"); >>> return 1; >>> } >>> for (j=0; j < 1048576; j++) { >>> p[j] = 'A'; >>> } >>> printf("touched memory\n"); >>> >>> sleep(1); >>> } >>> printf("enter sleep\n"); >>> while(1) { >>> sleep(100); >>> } >>> } >>> >>> Reported-by: Andrew Davidoff <davidoff@xxxxxxxxx> >>> Signed-off-by: Bob Liu <bob.liu@xxxxxxxxxx> >>> --- >>> mm/huge_memory.c | 50 +++++++++++++++++++++++++++++++++++++++++--------- >>> 1 file changed, 41 insertions(+), 9 deletions(-) >>> >>> diff --git a/mm/huge_memory.c b/mm/huge_memory.c >>> index 7448cf9..86c7f0d 100644 >>> --- a/mm/huge_memory.c >>> +++ b/mm/huge_memory.c >>> @@ -2144,7 +2144,33 @@ static void khugepaged_alloc_sleep(void) >>> msecs_to_jiffies(khugepaged_alloc_sleep_millisecs)); >>> } >>> >>> +static int khugepaged_node_load[MAX_NUMNODES]; >>> #ifdef CONFIG_NUMA >>> +static int last_khugepaged_target_node = NUMA_NO_NODE; >>> +static int khugepaged_find_target_node(void) >>> +{ >> >>> + int i, target_node = 0, max_value = 1; >> >> i is used as node ids. So please use node or nid instead of i. >> > > Sure! > >>> + >> >>> + /* find first node with most normal pages hit */ >>> + for (i = 0; i < MAX_NUMNODES; i++) >>> + if (khugepaged_node_load[i] > max_value) { >>> + max_value = khugepaged_node_load[i]; >>> + target_node = i; >>> + } >> >> khugepaged_node_load[] is initialized as 0 and max_value is initialized >> as 1. So this loop does not work well until khugepage_node_load[] is set >> to 2 or more. How about initializing max_value to 0? >> > > Sure! > >> >>> + >>> + /* do some balance if several nodes have the same hit number */ >>> + if (target_node <= last_khugepaged_target_node) { >>> + for (i = last_khugepaged_target_node + 1; i < MAX_NUMNODES; i++) >>> + if (max_value == khugepaged_node_load[i]) { >>> + target_node = i; >>> + break; >>> + } >>> + } >>> + >>> + last_khugepaged_target_node = target_node; >>> + return target_node; >>> +} >>> + >>> static bool khugepaged_prealloc_page(struct page **hpage, bool *wait) >>> { >>> if (IS_ERR(*hpage)) { >>> @@ -2178,9 +2204,8 @@ static struct page >>> * mmap_sem in read mode is good idea also to allow greater >>> * scalability. >>> */ >> >>> - *hpage = alloc_hugepage_vma(khugepaged_defrag(), vma, address, >>> - node, __GFP_OTHER_NODE); >>> - >>> + *hpage = alloc_pages_exact_node(node, alloc_hugepage_gfpmask( >>> + khugepaged_defrag(), __GFP_OTHER_NODE), HPAGE_PMD_ORDER); >> >> Why do you use alloc_pages_exact_node()? >> > > alloc_hugepage_vma() will call alloc_pages_vma() which will use some > mempolicy. > But some time the mempolicy is not we want for khugepaged. > > In Andrew's example, he set his application's mempolicy to MPOL_INTERLEAVE. > But khugepaged doesn't know this, the mempolicy of khugepaged will be > used(MPOL_PREFERRED) when alloc_pages_vma() is called in khugepaged thread. > As a result, all huge pages are allocated from Node A which doesn't > match the requirement from user land. I understood it. Thanks, Yasuaki Ishimatsu > > Thanks, > -Bob > -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>