On Thu 31-08-17 09:55:25, Vlastimil Babka wrote: > On 08/23/2017 08:47 AM, Vlastimil Babka wrote: > > On 07/24/2017 02:38 PM, Michal Hocko wrote: > >> On Thu 20-07-17 15:40:26, Vlastimil Babka wrote: > >>> In init_pages_in_zone() we currently use the generic set_page_owner() function > >>> to initialize page_owner info for early allocated pages. This means we > >>> needlessly do lookup_page_ext() twice for each page, and more importantly > >>> save_stack(), which has to unwind the stack and find the corresponding stack > >>> depot handle. Because the stack is always the same for the initialization, > >>> unwind it once in init_pages_in_zone() and reuse the handle. Also avoid the > >>> repeated lookup_page_ext(). > >> > >> Yes this looks like an improvement but I have to admit that I do not > >> really get why we even do save_stack at all here. Those pages might > >> got allocated from anywhere so we could very well provide a statically > >> allocated "fake" stack trace, no? > > > > We could, but it's much simpler to do it this way than try to extend > > stack depot/stack saving to support creating such fakes. Would it be > > worth the effort? > > Ah, I've noticed we already do this for the dummy (prevent recursion) > stack and failure stack. So here you go. It will also make the fake > stack more obvious after "[PATCH 2/2] mm, page_owner: Skip unnecessary > stack_trace entries" is merged, which would otherwise remove > init_page_owner() from the stack. Yes this is what I've had in mind. > ----8<---- > >From 9804a5e62fc768e12b86fd4a3184e692c59ebfd1 Mon Sep 17 00:00:00 2001 > From: Vlastimil Babka <vbabka@xxxxxxx> > Date: Thu, 31 Aug 2017 09:46:46 +0200 > Subject: [PATCH] mm, page_owner: make init_pages_in_zone() faster-fix2 > > Create statically allocated fake stack trace for early allocated pages, per > Michal Hocko. Yes this looks good to me. I am just wondering why we need 3 different fake stacks. I do not see any code that would special case them when dumping traces. Maybe this can be done on top? > Signed-off-by: Vlastimil Babka <vbabka@xxxxxxx> Anyway Acked-by: Michal Hocko <mhocko@xxxxxxxx> > --- > mm/page_owner.c | 30 +++++++++++++++--------------- > 1 file changed, 15 insertions(+), 15 deletions(-) > > diff --git a/mm/page_owner.c b/mm/page_owner.c > index 54d49fc8035e..262503f3ea66 100644 > --- a/mm/page_owner.c > +++ b/mm/page_owner.c > @@ -30,6 +30,7 @@ DEFINE_STATIC_KEY_FALSE(page_owner_inited); > > static depot_stack_handle_t dummy_handle; > static depot_stack_handle_t failure_handle; > +static depot_stack_handle_t early_handle; > > static void init_early_allocated_pages(void); > > @@ -53,7 +54,7 @@ static bool need_page_owner(void) > return true; > } > > -static noinline void register_dummy_stack(void) > +static __always_inline depot_stack_handle_t create_dummy_stack(void) > { > unsigned long entries[4]; > struct stack_trace dummy; > @@ -64,21 +65,22 @@ static noinline void register_dummy_stack(void) > dummy.skip = 0; > > save_stack_trace(&dummy); > - dummy_handle = depot_save_stack(&dummy, GFP_KERNEL); > + return depot_save_stack(&dummy, GFP_KERNEL); > } > > -static noinline void register_failure_stack(void) > +static noinline void register_dummy_stack(void) > { > - unsigned long entries[4]; > - struct stack_trace failure; > + dummy_handle = create_dummy_stack(); > +} > > - failure.nr_entries = 0; > - failure.max_entries = ARRAY_SIZE(entries); > - failure.entries = &entries[0]; > - failure.skip = 0; > +static noinline void register_failure_stack(void) > +{ > + failure_handle = create_dummy_stack(); > +} > > - save_stack_trace(&failure); > - failure_handle = depot_save_stack(&failure, GFP_KERNEL); > +static noinline void register_early_stack(void) > +{ > + early_handle = create_dummy_stack(); > } > > static void init_page_owner(void) > @@ -88,6 +90,7 @@ static void init_page_owner(void) > > register_dummy_stack(); > register_failure_stack(); > + register_early_stack(); > static_branch_enable(&page_owner_inited); > init_early_allocated_pages(); > } > @@ -529,13 +532,10 @@ static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone) > unsigned long pfn = zone->zone_start_pfn, block_end_pfn; > unsigned long end_pfn = pfn + zone->spanned_pages; > unsigned long count = 0; > - depot_stack_handle_t init_handle; > > /* Scan block by block. First and last block may be incomplete */ > pfn = zone->zone_start_pfn; > > - init_handle = save_stack(0); > - > /* > * Walk the zone in pageblock_nr_pages steps. If a page block spans > * a zone boundary, it will be double counted between zones. This does > @@ -588,7 +588,7 @@ static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone) > continue; > > /* Found early allocated page */ > - __set_page_owner_handle(page_ext, init_handle, 0, 0); > + __set_page_owner_handle(page_ext, early_handle, 0, 0); > count++; > } > cond_resched(); > -- > 2.14.1 > -- Michal Hocko SUSE Labs -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>