Re: "make -j" with memory.(memsw.)limit_in_bytes smaller than required -> livelock, even for unlimited processes

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, 22 Jun 2011 11:53:10 +0200
Lutz Vieweg <lvml@xxxxxx> wrote:

> On 06/22/2011 02:10 AM, KAMEZAWA Hiroyuki wrote:

> > Then, waiting for some page bit...I/O of libc mapped pages ?
> >
> > Hmm. it seems buggy behavior. Okay, I'll dig this.
> 
> Thanks a lot for investigating!
> 

This patch works for me. please see the thread
https://lkml.org/lkml/2011/6/22/163, too.
==

>From 7e9250da9ff529958d4c1ff511458dbdac8e4b81 Mon Sep 17 00:00:00 2001
From: KAMEZAWA Hiroyuki <kamezawa.hiroyu@xxxxxxxxxxxxxx>
Date: Thu, 23 Jun 2011 15:05:57 +0900
Subject: [PATCH] Currently we are keeping faulted page locked throughout whole __do_fault
 call (except for page_mkwrite code path). If we do early COW we allocate a
 new page which has to be charged for a memcg (mem_cgroup_newpage_charge).

This function, however, might block for unbounded amount of time if memcg
oom killer is disabled or fork-bomb is running because the only way out of
the OOM situation is either an external event or OOM-situation fix.

processes from faulting it in which is not good at all because we are
basically punishing potentially an unrelated process for OOM condition
in a different group (I have seen stuck system because of ld-2.11.1.so being
locked).

We can do test easily.

% cgcreate -g memory:A
% cgset -r memory.limit_in_bytes=64M A
% cgset -r memory.memsw.limit_in_bytes=64M A
% cd kernel_dir; cgexec -g memory:A make -j


Then, the whole system will live-locked until you kill 'make -j'
by hands (or push reboot...) This is because some important
page in a shared library are locked.

This patch delays "charge" until unlock_page() called. There is
no problem as far as we keep reference on a page.
(memcg doesn't require page_lock()).

Then, above livelock disappears.

Reported-by: Lutz Vieweg <lvml@xxxxxx>
Original-idea-by: Michal Hocko <mhocko@xxxxxxx>
Signed-off-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@xxxxxxxxxxxxxx>
---
 mm/memory.c |   28 +++++++++++++++++++---------
 1 files changed, 19 insertions(+), 9 deletions(-)

diff --git a/mm/memory.c b/mm/memory.c
index 87d9353..66442da 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -3129,7 +3129,7 @@ static int __do_fault(struct mm_struct *mm, struct vm_area_struct *vma,
 	struct page *page;
 	pte_t entry;
 	int anon = 0;
-	int charged = 0;
+	struct page *need_charge = NULL;
 	struct page *dirty_page = NULL;
 	struct vm_fault vmf;
 	int ret;
@@ -3177,12 +3177,7 @@ static int __do_fault(struct mm_struct *mm, struct vm_area_struct *vma,
 				ret = VM_FAULT_OOM;
 				goto out;
 			}
-			if (mem_cgroup_newpage_charge(page, mm, GFP_KERNEL)) {
-				ret = VM_FAULT_OOM;
-				page_cache_release(page);
-				goto out;
-			}
-			charged = 1;
+			need_charge = page;
 			copy_user_highpage(page, vmf.page, address, vma);
 			__SetPageUptodate(page);
 		} else {
@@ -3251,12 +3246,11 @@ static int __do_fault(struct mm_struct *mm, struct vm_area_struct *vma,
 		/* no need to invalidate: a not-present page won't be cached */
 		update_mmu_cache(vma, address, page_table);
 	} else {
-		if (charged)
-			mem_cgroup_uncharge_page(page);
 		if (anon)
 			page_cache_release(page);
 		else
 			anon = 1; /* no anon but release faulted_page */
+		need_charge = NULL;
 	}
 
 	pte_unmap_unlock(page_table, ptl);
@@ -3268,6 +3262,17 @@ out:
 		if (set_page_dirty(dirty_page))
 			page_mkwrite = 1;
 		unlock_page(dirty_page);
+		if (need_charge) {
+			/*
+			 * charge this page before we drop refcnt.
+			 * memory cgroup returns OOM condition when
+			 * this task is killed. So, it's not necesasry
+			 * to undo.
+			 */
+			if (mem_cgroup_newpage_charge(need_charge,
+					mm, GFP_KERNEL))
+				ret = VM_FAULT_OOM;
+		}
 		put_page(dirty_page);
 		if (page_mkwrite && mapping) {
 			/*
@@ -3282,6 +3287,11 @@ out:
 			file_update_time(vma->vm_file);
 	} else {
 		unlock_page(vmf.page);
+		if (need_charge) {
+			if (mem_cgroup_newpage_charge(need_charge,
+						mm, GFP_KERNEL))
+				ret = VM_FAULT_OOM;
+		}
 		if (anon)
 			page_cache_release(vmf.page);
 	}
-- 
1.7.4.1


--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@xxxxxxxxx.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Fight unfair telecom internet charges in Canada: sign http://stopthemeter.ca/
Don't email: <a href=mailto:"dont@xxxxxxxxx";> email@xxxxxxxxx </a>


[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [ECOS]     [Asterisk Internet PBX]     [Linux API]