On 25.03.2014 [11:53:48 -0500], Christoph Lameter wrote: > On Tue, 25 Mar 2014, Nishanth Aravamudan wrote: > > > On 25.03.2014 [11:17:57 -0500], Christoph Lameter wrote: > > > On Mon, 24 Mar 2014, Nishanth Aravamudan wrote: > > > > > > > Anyone have any ideas here? > > > > > > Dont do that? Check on boot to not allow exhausting a node with huge > > > pages? > > > > Gigantic hugepages are allocated by the hypervisor (not the Linux VM), > > Ok so the kernel starts booting up and then suddenly the hypervisor takes > the 2 16G pages before even the slab allocator is working? There is nothing "sudden" about it. On power, very early, we find the 16G pages (gpages in the powerpc arch code) in the device-tree: early_setup -> early_init_mmu -> htab_initialize -> htab_init_page_sizes -> htab_dt_scan_hugepage_blocks -> memblock_reserve which marks the memory as reserved add_gpage which saves the address off so future calls for alloc_bootmem_huge_page() hugetlb_init -> hugetlb_init_hstates -> hugetlb_hstate_alloc_pages -> alloc_bootmem_huge_page > Not sure if I understand that correctly. Basically this is present memory that is "reserved" for the 16GB usage per the LPAR configuration. We honor that configuration in Linux based upon the contents of the device-tree. It just so happens in the configuration from my original e-mail that a consequence of this is that a NUMA node has memory (topologically), but none of that memory is free, nor will it ever be free. Perhaps, in this case, we could just remove that node from the N_MEMORY mask? Memory allocations will never succeed from the node, and we can never free these 16GB pages. It is really not any different than a memoryless node *except* when you are using the 16GB pages. Thanks, Nish -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>