On Thu 11-07-13 08:44:08, Tejun Heo wrote: > Hello, Michal. > > On Thu, Jul 11, 2013 at 11:33:00AM +0200, Michal Hocko wrote: > > +static inline > > +struct mem_cgroup *vmpressure_to_mem_cgroup(struct vmpressure *vmpr) > > +{ > > + return container_of(vmpr, struct mem_cgroup, vmpressure); > > +} > > + > > +void vmpressure_pin_memcg(struct vmpressure *vmpr) > > +{ > > + struct mem_cgroup *memcg = vmpressure_to_mem_cgroup(vmpr); > > + > > + css_get(&memcg->css); > > +} > > + > > +void vmpressure_unpin_memcg(struct vmpressure *vmpr) > > +{ > > + struct mem_cgroup *memcg = vmpressure_to_mem_cgroup(vmpr); > > + > > + css_put(&memcg->css); > > +} > > So, while this *should* work, can't we just cancel/flush the work item > from offline? I would rather not put vmpressure clean up code into memcg offlining. We have reference counting for exactly this purposes so it feels strange to overcome it like that. Besides that wouldn't be that racy? The work item could be already executing and preempted and we do not want vmpr to disappear from under our feet. I know this is _highly_ unlikely but why to play games? > There doesn't seem to be any possible deadlocks from my > shallow glance and those mutexes don't seem to be held for long (do > they actually need to be mutexes? what blocks inside them?). Dunno, to be honest. From a quick look they both can be turned to spin locks but events_lock might cause long preempt disabled periods when zillions of events are registered. > Also, while at it, can you please remove the work_pending() check? > They're almost always spurious or racy and should be avoided in > general. sure, this really looks bogus. I will cook patches for both issues and send them tomorrow. -- Michal Hocko SUSE Labs -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>