On Fri 28-06-13 22:55:17, Minchan Kim wrote: > Hi Michal, > > On Fri, Jun 28, 2013 at 02:24:12PM +0200, Michal Hocko wrote: > > On Fri 28-06-13 08:54:35, Minchan Kim wrote: > > > Hello Michal, > > > > > > On Thu, Jun 27, 2013 at 06:11:03PM +0200, Michal Hocko wrote: > > > > On Fri 28-06-13 00:35:28, Minchan Kim wrote: > > > > > Hi Michal, > > > > > > > > > > On Thu, Jun 27, 2013 at 11:37:21AM +0200, Michal Hocko wrote: > > > > > > On Thu 27-06-13 15:12:10, Hyunhee Kim wrote: > > > > > > > In vmpressure, the pressure level is calculated based on the ratio > > > > > > > of how many pages were scanned vs. reclaimed in a given time window. > > > > > > > However, there is a possibility that "scanned < reclaimed" in such a > > > > > > > case, when reclaiming ends by fatal signal in shrink_inactive_list. > > > > > > > So, with this patch, we just return "low" level when "scanned < reclaimed" > > > > > > > happens not to have userland miss reclaim activity. > > > > > > > > > > > > Hmm, fatal signal pending on kswapd doesn't make sense to me so it has > > > > > > to be a direct reclaim path. Does it really make sense to signal LOW > > > > > > when there is probably a big memory pressure and somebody is killing the > > > > > > current allocator? > > > > > > > > > > So, do you want to trigger critical instead of low? > > > > > > > > > > Now, current is going to die so we can expect shortly we can get a amount > > > > > of memory, normally. > > > > > > > > And also consider that this is per-memcg interface. And so it is even > > > > more complicated. If a task dies then there is _no_ guarantee that there > > > > will be an uncharge in that group (task could have been migrated to that > > > > group so the memory belongs to somebody else). > > > > > > Good point and that's one of the reason I hate memcg for just using > > > vmpressure. > > > > Well, the very same problem is present in the memcg OOM as well. oom > > score calculation is not memcg aware wrt charges. > > > > > Let's think over it. One of the very avaialbe scenario > > > which userland could do when notified from vmpressure is that manager > > > process sends signal for others to release own cached memory. > > > > Assuming those processes are in the same memcg, right? > > > > > If we use vmpressure without move_charge_at_immigrate in multiple memcg > > > group, it would be a disaster. But if we use move_charge_at_immigrate, > > > we will see long stall easily so it's not an option, either. > > > > I am not sure I am following you here. Could you be more specific what > > is the actual problem? > > From my POV, a manager can see a memory pressure, it notifies others in > > the same memcg and they will release their caches. With > > move_charge_at_immigrate == 0 some of those might release a memory in > > other group but somebody must be using memory from the currently > > signaled group, right? > > My concern is that manager process can send a signal to a process A > in same group but unfortunately, process A would release a memory > in other group so manager process can send a signal to a process B > in same group but unfortunately, process B would release a memory > in other group so manger process can ... > ... > ... > ... > in same group and at last, process Z would release a memory in same > group but we release all of cached from A-Y process. :( I would say, do not move tasks without move_at_immigrate if you are doing something like that. > > > So, IMO, it's not a good idea to use vmpressure with no-root memcg so > > > it could raise the question again "why vmpressure is part of memcg". > > > > Maybe I do not see the problem correctly, but making vmpressure memcg > > aware was a good idea. It is something like userspace pre-oom handling. > > I don't say that memcg-aware is bad. Surely it's good thing but > it's not good that we must enable memcg for just using memory notifier > globally. Even above problem would make memcg-vmpressure complicated and > memory reclaim behavior change compared to long history well-made global > page reclaim. > > I claim we should be able to use vmpressure without memcg as well as > memcg. This is not hard to do. Look at how we are handling lruvec for !memcg configurations. The similar thing could be done for vmpressure as well. You just need to find a way how to export eventfd which is not private to memcg AFAIK. -- Michal Hocko SUSE Labs -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>