Re: [RFC] [PATCH] Cgroup based OOM killer controller

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, 22 Jan 2009, Evgeniy Polyakov wrote:

> > In an exclusive cpuset, a task's memory is restricted to a set of mems 
> > that the administrator has designated.  If it is oom, the kernel must free 
> > memory on those nodes or the next allocation will again trigger an oom 
> > (leading to a needlessly killed task that was in a disjoint cpuset).
> > 
> > Really.
> 
> The whole point of oom-killer is to kill the most appropriate task to
> free the memory. And while task is selected system-wide and some
> tunables are added to tweak the behaviour local to some subsystems, this
> cpuset feature is hardcoded into the selection algorithm.

Of course, because the oom killer must be aware that tasks in disjoint 
cpusets are more likely than not to result in no memory freeing for 
current's subsequent allocations.

> And when some tunable starts doing own calculation, behaviour of this
> hardcoded feature changes.
> 

Yes, it is possible to elevate oom_adj scores to override the cpuset 
preference.  That's actually intended since it is now possible for the 
administrator to specify that, against the belief of the kernel, that 
killing a task will free memory in these cpuset-constrained ooms.  That's 
probably because it has either been moved to a different cpuset or its set 
of allowable nodes is dynamic.

> > Then the scope of this new cgroup is restricted to not being used with 
> > cpusets that could oom.
> 
> These are perpendicular tasks - cpusets limit one area of the oom
> handling, cgroup order - another. Some people needs cpusets, others want
> cgroups. cpusets are not something exceptional so that only they have to
> be taken into account when doing system-wide operation like OOM
> condition handling.
> 

A cpuset is a cgroup.  If I am using cpusets, this patch fails to 
adequately allow me to describe my oom preferences for both 
cpuset-constrained ooms and global unconstrained ooms, which is a major 
drawback.

I would encourage you to look at the per-cgroup oom notifier patch[*] that 
defers most of these decisions to userspace.  Given your interest in 
priority based oom preferences as exhibited by your oom_victim patch, I 
think you'll find it of interest since it allows you much greater 
flexibility than you could ever hope for from the kernel's heuristics.

 [*] http://marc.info/?l=linux-mm&m=122575082227252
_______________________________________________
Containers mailing list
Containers@xxxxxxxxxxxxxxxxxxxxxxxxxx
https://lists.linux-foundation.org/mailman/listinfo/containers

[Index of Archives]     [Cgroups]     [Netdev]     [Linux Wireless]     [Kernel Newbies]     [Security]     [Linux for Hams]     [Netfilter]     [Bugtraq]     [Yosemite Forum]     [MIPS Linux]     [ARM Linux]     [Linux RAID]     [Linux Admin]     [Samba]

  Powered by Linux