Hello, David. First of all, thanks for review! On Thu, Sep 05, 2013 at 11:15:53AM +1000, David Gibson wrote: > On Fri, Aug 09, 2013 at 06:26:37PM +0900, Joonsoo Kim wrote: > > If parallel fault occur, we can fail to allocate a hugepage, > > because many threads dequeue a hugepage to handle a fault of same address. > > This makes reserved pool shortage just for a little while and this cause > > faulting thread who can get hugepages to get a SIGBUS signal. > > > > To solve this problem, we already have a nice solution, that is, > > a hugetlb_instantiation_mutex. This blocks other threads to dive into > > a fault handler. This solve the problem clearly, but it introduce > > performance degradation, because it serialize all fault handling. > > > > Now, I try to remove a hugetlb_instantiation_mutex to get rid of > > performance degradation. For achieving it, at first, we should ensure that > > no one get a SIGBUS if there are enough hugepages. > > > > For this purpose, if we fail to allocate a new hugepage when there is > > concurrent user, we return just 0, instead of VM_FAULT_SIGBUS. With this, > > these threads defer to get a SIGBUS signal until there is no > > concurrent user, and so, we can ensure that no one get a SIGBUS if there > > are enough hugepages. > > > > Signed-off-by: Joonsoo Kim <iamjoonsoo.kim@xxxxxxx> > > > > diff --git a/include/linux/hugetlb.h b/include/linux/hugetlb.h > > index e29e28f..981c539 100644 > > --- a/include/linux/hugetlb.h > > +++ b/include/linux/hugetlb.h > > @@ -242,6 +242,7 @@ struct hstate { > > int next_nid_to_free; > > unsigned int order; > > unsigned long mask; > > + unsigned long nr_dequeue_users; > > unsigned long max_huge_pages; > > unsigned long nr_huge_pages; > > unsigned long free_huge_pages; > > diff --git a/mm/hugetlb.c b/mm/hugetlb.c > > index 8743e5c..0501fe5 100644 > > --- a/mm/hugetlb.c > > +++ b/mm/hugetlb.c > > @@ -561,6 +561,7 @@ retry_cpuset: > > if (cpuset_zone_allowed_softwall(zone, htlb_alloc_mask)) { > > page = dequeue_huge_page_node(h, zone_to_nid(zone)); > > if (page) { > > + h->nr_dequeue_users++; > > So, nr_dequeue_users doesn't seem to be incremented in the > alloc_huge_page_node() path. I'm not sure exactly where that's used, > so I'm not sure if it's a problem. > Hmm.. I think that it isn't a problem. The point is that we want to avoid the race which kill the legitimate users of hugepages by out of resources. This allocation doesn't harm to the legitimate users. > > if (!use_reserve) > > break; > > > > @@ -577,6 +578,16 @@ retry_cpuset: > > return page; > > } > > > > +static void commit_dequeued_huge_page(struct hstate *h, bool do_dequeue) > > +{ > > + if (!do_dequeue) > > + return; > > Seems like it would be easier to do this test in the callers, but I > doubt it matters much. Yes, I will fix it. > > > + spin_lock(&hugetlb_lock); > > + h->nr_dequeue_users--; > > + spin_unlock(&hugetlb_lock); > > +} > > + > > static void update_and_free_page(struct hstate *h, struct page *page) > > { > > int i; > > @@ -1110,7 +1121,9 @@ static void vma_commit_reservation(struct hstate *h, > > } > > > > static struct page *alloc_huge_page(struct vm_area_struct *vma, > > - unsigned long addr, int use_reserve) > > + unsigned long addr, int use_reserve, > > + unsigned long *nr_dequeue_users, > > + bool *do_dequeue) > > { > > struct hugepage_subpool *spool = subpool_vma(vma); > > struct hstate *h = hstate_vma(vma); > > @@ -1138,8 +1151,11 @@ static struct page *alloc_huge_page(struct vm_area_struct *vma, > > return ERR_PTR(-ENOSPC); > > } > > spin_lock(&hugetlb_lock); > > + *do_dequeue = true; > > page = dequeue_huge_page_vma(h, vma, addr, use_reserve); > > if (!page) { > > + *nr_dequeue_users = h->nr_dequeue_users; > > So, the nr_dequeue_users parameter is only initialized if !page here. > It's not obvious to me that the callers only use it in hat case. Okay. I will fix it. > > > + *do_dequeue = false; > > spin_unlock(&hugetlb_lock); > > page = alloc_buddy_huge_page(h, NUMA_NO_NODE); > > if (!page) { > > I think the counter also needs to be incremented in the case where we > call alloc_buddy_huge_page() from alloc_huge_page(). Even though it's > new, it gets added to the hugepage pool at this point and could still > be a contended page for the last allocation, unless I'm missing > something. Your comment has reasonable point to me, but I have a different opinion. As I already mentioned, the point is that we want to avoid the race which kill the legitimate users of hugepages by out of resources. I increase 'h->nr_dequeue_users' when the hugepage allocated by administrator is dequeued. It is because what the hugepage I want to protect from the race is the one allocated by administrator via kernel param or /proc interface. Administrator may already know how many hugepages are needed for their application so that he may set nr_hugepage to reasonable value. I want to guarantee that these hugepages can be used for his application without any race, since he assume that the application would work fine with these hugepages. To protect hugepages returned from alloc_buddy_huge_page() from the race is different for me. Although it will be added to the hugepage pool, this doesn't guarantee certain application's success more. If certain application's success depends on the race of this new hugepage, it's death by the race doesn't matter, since nobody assume that it works fine. [snip..] > Otherwise I think it looks good. Really thanks! :) -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>