Michal Hocko wrote: > On Sun 15-03-15 22:06:54, Tetsuo Handa wrote: > > Michal Hocko wrote: > [...] > > > this. I understand that the wording of the changelog might be confusing, > > > though. > > > > > > It says: "This implementation counts only those retries which involved > > > OOM killer because we do not want to be too eager to fail the request." > > > > > > Would it be more clear if I changed that to? > > > "This implemetnation counts only those retries when the system is > > > considered OOM because all previous reclaim attempts have resulted > > > in no progress because we do not want to be too eager to fail the > > > request." > > > > > > We definitely _want_ to fail GFP_NOFS allocations. > > > > I see. The updated changelog is much more clear. > > Patch with the updated changelog (no other changes) Now the changelog is clear that "Involved OOM killer" == "__GFP_FS allocation" and "Considered OOM" == "both __GFP_FS and !__GFP_FS allocation". One more thing I want to confirm about this patch's changelog. This patch will generate the same result shown below. Tetsuo Handa wrote: > I also tested on XFS. One is Linux 3.19 and the other is Linux 3.19 > with debug printk patch shown above. According to console logs, > oom_kill_process() is trivially called via pagefault_out_of_memory() > for the former kernel. Due to giving up !GFP_FS allocations immediately? > > (From http://I-love.SAKURA.ne.jp/tmp/serial-20150223-3.19-xfs-unpatched.txt.xz ) > ---------- xfs / Linux 3.19 ---------- > [ 793.283099] su invoked oom-killer: gfp_mask=0x0, order=0, oom_score_adj=0 > [ 793.283102] su cpuset=/ mems_allowed=0 > [ 793.283104] CPU: 3 PID: 9552 Comm: su Not tainted 3.19.0 #40 > [ 793.283159] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 07/31/2013 > [ 793.283161] 0000000000000000 ffff88007ac03bf8 ffffffff816ae9d4 000000000000bebe > [ 793.283162] ffff880078b0d740 ffff88007ac03c98 ffffffff816ac7ac 0000000000000206 > [ 793.283163] 0000000481f30298 ffff880073e55850 ffff88007ac03c88 ffff88007a20bef8 > [ 793.283164] Call Trace: > [ 793.283169] [<ffffffff816ae9d4>] dump_stack+0x45/0x57 > [ 793.283171] [<ffffffff816ac7ac>] dump_header+0x7f/0x1f1 > [ 793.283174] [<ffffffff8114b36b>] oom_kill_process+0x22b/0x390 > [ 793.283177] [<ffffffff810776d0>] ? has_capability_noaudit+0x20/0x30 > [ 793.283178] [<ffffffff8114bb72>] out_of_memory+0x4b2/0x500 > [ 793.283179] [<ffffffff8114bc37>] pagefault_out_of_memory+0x77/0x90 > [ 793.283180] [<ffffffff816aab2c>] mm_fault_error+0x67/0x140 > [ 793.283182] [<ffffffff8105a9f6>] __do_page_fault+0x3f6/0x580 > [ 793.283185] [<ffffffff810aed1d>] ? remove_wait_queue+0x4d/0x60 > [ 793.283186] [<ffffffff81070fcb>] ? do_wait+0x12b/0x240 > [ 793.283187] [<ffffffff8105abb1>] do_page_fault+0x31/0x70 > [ 793.283189] [<ffffffff816b83e8>] page_fault+0x28/0x30 > ---------- xfs / Linux 3.19 ---------- Are all memory allocations caused by page fault __GFP_FS allocation? If memory allocations caused by page fault are !__GFP_FS allocation (e.g. 0x2015a == __GFP_HARDWALL | __GFP_COLD | __GFP_IO | __GFP_WAIT | __GFP_HIGHMEM | __GFP_MOVABLE), this patch will start trivially involving OOM killer for !__GFP_FS allocation. I haven't tried how many processes can be killed by this path, but this path can potentially OOM-kill most of OOM-killable processes depending on how long the OOM condition lasts. It would be better to mention that a lot of processes might be OOM-killed by page faults due to this change. -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>