Re: [PATCH v8 11/12] writeback: make background writeback cgroup aware

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Jun 07, 2011 at 01:43:08PM -0700, Greg Thelen wrote:
> Vivek Goyal <vgoyal@xxxxxxxxxx> writes:
> 
> > On Fri, Jun 03, 2011 at 09:12:17AM -0700, Greg Thelen wrote:
> >> When the system is under background dirty memory threshold but a cgroup
> >> is over its background dirty memory threshold, then only writeback
> >> inodes associated with the over-limit cgroup(s).
> >> 
> >
> > [..]
> >> -static inline bool over_bground_thresh(void)
> >> +static inline bool over_bground_thresh(struct bdi_writeback *wb,
> >> +				       struct writeback_control *wbc)
> >>  {
> >>  	unsigned long background_thresh, dirty_thresh;
> >>  
> >>  	global_dirty_limits(&background_thresh, &dirty_thresh);
> >>  
> >> -	return (global_page_state(NR_FILE_DIRTY) +
> >> -		global_page_state(NR_UNSTABLE_NFS) > background_thresh);
> >> +	if (global_page_state(NR_FILE_DIRTY) +
> >> +	    global_page_state(NR_UNSTABLE_NFS) > background_thresh) {
> >> +		wbc->for_cgroup = 0;
> >> +		return true;
> >> +	}
> >> +
> >> +	wbc->for_cgroup = 1;
> >> +	wbc->shared_inodes = 1;
> >> +	return mem_cgroups_over_bground_dirty_thresh();
> >>  }
> >
> > Hi Greg,
> >
> > So all the logic of writeout from mem cgroup works only if system is
> > below background limit. The moment we cross background limit, looks
> > like we will fall back to existing way of writting inodes?
> 
> Correct.  If the system is over its background limit then the previous
> cgroup-unaware background writeback occurs.  I think of the system
> limits as those of the root cgroup.  If the system is over the global
> limit than all cgroups are eligible for writeback.  In this situation
> the current code does not distinguish between cgroups over or under
> their dirty background limit.
> 
> Vivek Goyal <vgoyal@xxxxxxxxxx> writes:
> > If yes, then from design point of view it is little odd that as long
> > as we are below background limit, we share the bdi between different
> > cgroups. The moment we are above background limit, we fall back to
> > algorithm of sharing the disk among individual inodes and forget
> > about memory cgroups. Kind of awkward.
> >
> > This kind of cgroup writeback I think will atleast not solve the problem
> > for CFQ IO controller, as we fall back to old ways of writting back inodes
> > the moment we cross dirty ratio.
> 
> It might make more sense to reverse the order of the checks in the
> proposed over_bground_thresh(): the new version would first check if any
> memcg are over limit; assuming none are over limit, then check global
> limits.  Assuming that the system is over its background limit and some
> cgroups are also over their limits, then the over limit cgroups would
> first be written possibly getting the system below its limit.  Does this
> address your concern?

Do you treat root group also as any other cgroup? If no, then above logic
can lead to issue of starvation of root group inode. Or unfair writeback.
So I guess it will be important to treat root group same as other groups.

> 
> Note: mem_cgroup_balance_dirty_pages() (patch 10/12) will perform
> foreground writeback when a memcg is above its dirty limit.  This would
> offer CFQ multiple tasks issuing IO.

I guess we can't rely on this as this will go away once IO less dirty
throttling is merged.

Thanks
Vivek
--
To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux Ext4 Filesystem]     [Union Filesystem]     [Filesystem Testing]     [Ceph Users]     [Ecryptfs]     [AutoFS]     [Kernel Newbies]     [Share Photos]     [Security]     [Netfilter]     [Bugtraq]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux Cachefs]     [Reiser Filesystem]     [Linux RAID]     [Samba]     [Device Mapper]     [CEPH Development]
  Powered by Linux