On Thu, Jun 21, 2012 at 1:17 PM, Tejun Heo <tj@xxxxxxxxxx> wrote: > Hello, Yinghai. > > On Tue, Jun 19, 2012 at 07:57:45PM -0700, Yinghai Lu wrote: >> if it is that case, that change could fix other problem problem too. >> --- during the one free reserved.regions could double the array. > > Yeah, that sounds much more attractive to me too. Some comments on > the patch tho. > >> /** >> * memblock_double_array - double the size of the memblock regions array >> * @type: memblock type of the regions array being doubled >> @@ -216,7 +204,7 @@ static int __init_memblock memblock_doub >> >> /* Calculate new doubled size */ >> old_size = type->max * sizeof(struct memblock_region); >> - new_size = old_size << 1; >> + new_size = PAGE_ALIGN(old_size << 1); > > We definintely can use some comments explaining why we want page > alignment. It's kinda subtle. yes. > > This is a bit confusing here because old_size is the proper size > without padding while new_size is page aligned size with possible > padding. Maybe discerning {old|new}_alloc_size is clearer? Also, I > think adding @new_cnt variable which is calculated together would make > the code easier to follow. So, sth like, > > /* explain why page aligning is necessary */ > old_size = type->max * sizeof(struct memblock_region); > old_alloc_size = PAGE_ALIGN(old_size); > > new_max = type->max << 1; > new_size = new_max * sizeof(struct memblock_region); > new_alloc_size = PAGE_ALIGN(new_size); > > and use alloc_sizes for alloc/frees and sizes for everything else. ok, will add new_alloc_size, old_alloc_size. > >> unsigned long __init free_low_memory_core_early(int nodeid) >> { >> unsigned long count = 0; >> - phys_addr_t start, end; >> + phys_addr_t start, end, size; >> u64 i; >> >> - /* free reserved array temporarily so that it's treated as free area */ >> - memblock_free_reserved_regions(); >> + for_each_free_mem_range(i, MAX_NUMNODES, &start, &end, NULL) >> + count += __free_memory_core(start, end); >> >> - for_each_free_mem_range(i, MAX_NUMNODES, &start, &end, NULL) { >> - unsigned long start_pfn = PFN_UP(start); >> - unsigned long end_pfn = min_t(unsigned long, >> - PFN_DOWN(end), max_low_pfn); >> - if (start_pfn < end_pfn) { >> - __free_pages_memory(start_pfn, end_pfn); >> - count += end_pfn - start_pfn; >> - } >> - } >> + /* free range that is used for reserved array if we allocate it */ >> + size = get_allocated_memblock_reserved_regions_info(&start); >> + if (size) >> + count += __free_memory_core(start, start + size); > > I'm afraid this is too early. We don't want the region to be unmapped > yet. This should only happen after all memblock usages are finished > which I don't think is the case yet. No, it is not early. at that time memblock usage is done. Also I tested one system with huge memory, duplicated the problem on KVM that Sasha met. my patch fixes the problem. please check attached patch. Also I add another patch to double check if there is any reference with reserved.region. so far there is no reference found. Thanks Yinghai
Attachment:
fix_free_memblock_reserve_v4_5.patch
Description: Binary data
Attachment:
memblock_reserved_clear_check.patch
Description: Binary data