[Sorry if i missed or added anyone on cc, patchwork.kernel.org LKML is not working and I'm not subscribed to the list ] On Mon Jul 26 2010 about 12:47:37 EST, Christoph Lameter wrote: > On Tue, 27 Jul 2010, Minchan Kim wrote: > > > This patch registers address of mem_section to memmap itself's page struct's > > pg->private field. This means the page is used for memmap of the section. > > Otherwise, the page is used for other purpose and memmap has a hole. > > > +void mark_valid_memmap(unsigned long start, unsigned long end); > > + > > +#ifdef CONFIG_ARCH_HAS_HOLES_MEMORYMODEL > > +static inline int memmap_valid(unsigned long pfn) > > +{ > > + struct page *page = pfn_to_page(pfn); > > + struct page *__pg = virt_to_page(page); > > + return page_private(__pg) == (unsigned long)__pg; > > > What if page->private just happens to be the value of the page struct? > Even if that is not possible today, someday someone may add new > functionality to the kernel where page->pivage == page is used for some > reason. > > Checking for PG_reserved wont work? I had the same thought and suggest setting it to the memory section block, since that is a uniquie value (unlike PG_reserved), > > static inline int pfn_valid(unsigned long pfn) > > { > > + struct mem_section *ms; > > if (pfn_to_section_nr(pfn) >= NR_MEM_SECTIONS) > > return 0; > > - return valid_section(__nr_to_section(pfn_to_section_nr(pfn))); > > + ms = __nr_to_section(pfn_to_section_nr(pfn)); > > + return valid_section(ms) && memmap_valid(pfn); .. and we already have computed it when we use it so we could pass it as a parameter (to both _valid and mark_valid). milton -- To unsubscribe from this list: send the line "unsubscribe linux-mmc" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html