Michal Hocko wrote: > On Tue 17-10-17 22:04:59, Tetsuo Handa wrote: > > Below is updated patch. The motivation of this patch is to guarantee that > > the thread (it can be SCHED_IDLE priority) calling out_of_memory() can use > > enough CPU resource by saving CPU resource wasted by threads (they can be > > !SCHED_IDLE priority) waiting for out_of_memory(). Thus, replace > > mutex_trylock() with mutex_lock_killable(). > > So what exactly guanratees SCHED_IDLE running while other high priority > processes keep preempting it while it holds the oom lock? Not everybody > is inside the allocation path to get out of the way. I think that that is a too much worry. If you worry such possibility, current assumption /* * Acquire the oom lock. If that fails, somebody else is * making progress for us. */ is horribly broken. Also, high priority threads keep preempting will prevent low priority threads from reaching __alloc_pages_may_oom() because preemption will occur not only during a low priority thread is holding oom_lock but also while oom_lock is not held. We can try to reduce preemption while oom_lock is held by scattering around preempt_disable()/preempt_enable(). But you said you don't want to disable preemption during OOM kill operation when I proposed scattering patch, didn't you? So, I think that worrying about high priority threads preventing the low priority thread with oom_lock held is too much. Preventing high priority threads waiting for oom_lock from disturbing the low priority thread with oom_lock held by wasting CPU resource will be sufficient. If you don't like it, the only way will be to offload to a dedicated kernel thread (like the OOM reaper) so that allocating threads are no longer blocked by oom_lock. That's a big change. > > > > By replacing mutex_trylock() with mutex_lock_killable(), it might prevent > > the OOM reaper from start reaping immediately. Thus, remove mutex_lock() from > > the OOM reaper. > > oom_lock shouldn't be necessary in oom_reaper anymore and that is worth > a separate patch. I'll propose as a separate patch after we apply "mm, oom: task_will_free_mem(current) should ignore MMF_OOM_SKIP for once." or we call __alloc_pages_slowpath() with oom_lock held. > > > By removing mutex_lock() from the OOM reaper, the race window of needlessly > > selecting next OOM victim becomes wider, for the last second allocation > > attempt no longer waits for the OOM reaper. Thus, do the really last > > allocation attempt after selecting an OOM victim using the same watermark. > > > > Can we go with this direction? > > The patch is just too cluttered. You do not want to use > __alloc_pages_slowpath. get_page_from_freelist would be more > appropriate. Also doing alloc_pages_before_oomkill two times seems to be > excessive. This patch is intentionally calling __alloc_pages_slowpath() because it handles ALLOC_OOM by calling __gfp_pfmemalloc_flags(). If this patch calls only get_page_from_freelist(), we will fail to try ALLOC_OOM before calling out_of_memory() (when current thread is selected as OOM victim while waiting for oom_lock) and just before sending SIGKILL (when task_will_free_mem(current) in out_of_memory() returned false because MMF_OOM_SKIP was set before ALLOC_OOM allocation is attempted) unless we apply "mm, oom: task_will_free_mem(current) should ignore MMF_OOM_SKIP for once.". > > That being said, make sure you adrress all the concerns brought up by > Andrea and Johannes in the above email thread first. I don't think there are concerns if we wait for oom_lock. The only concern will be do not depend on __GFP_DIRECT_RECLAIM allocation while oom_lock is held. Andrea and Johannes, what are your concerns? -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>