* Steven J. Magnani <steve@xxxxxxxxxxxxxxx> [2010-10-01 11:41:07]: > On Fri, 2010-10-01 at 16:07 +0100, David Howells wrote: > > Steve Magnani <steve@xxxxxxxxxxxxxxx> wrote: > > > > > If anything I think nommu is one of the better applications of memcg. Since > > > nommu typically embedded, being able to put potential memory pigs in a > > > sandbox so they can't destabilize the system is a Good Thing. That was my > > > motivation for doing this in the first place and it works quite well. > > > > I suspect it's not useful for a few reasons: > > > > (1) You don't normally run many applications on a NOMMU system. Typically, > > you'll run just one, probably threaded app, I think. > > Not always. > > > > > (2) In general, you won't be able to cull processes to make space. If the OOM > > killer runs your application has a bug in it. > > Not always. Every now and then applications have to deal with > user-supplied input of some sort. > > In our case it's a user-formatted disk drive that can have some > arbitrarily-sized FAT32 partition on which we are required to run > dosfsck. Now, dosfsck is the epitome of a memory pig; its memory > requirements scale with partition size, number of dentries, and any > damage encountered - none of which can be predicted. There is a set of > partitions we are able to check with no problem, but no guarantee the > user won't present us with one that would bring down the whole system, > were the OOM killer to get involved. Putting just dosfsck in its own > sandbox ensures this can't happen. See also my response to #4 below. > > > > > (3) memcg has a huge overhead. 20 bytes per page! On a 4K page 32-bit > > system, that's nearly 5% of your RAM, assuming I understand the > > CGROUP_MEM_RES_CTLR config help text correctly. > > When you use 16K pages, 20 bytes/page isn't so huge :) > > > > > (4) There's no swapping, no page faults, no migration and little shareable > > memory. Being able to allocate large blocks of contiguous memory is much > > more important and much more of a bottleneck than this. The 5% of RAM > > lost makes that just that little bit harder. > > > > If it's memory sandboxing you require, ulimit might be sufficient for NOMMU > > mode. > > dosfsck is written to handle memory allocation failures properly > (bailing out) but I have not been able to get this code to execute when > the system runs out of memory - the OOM killer gets invoked and that's > all she wrote. Will a ulimit violation return control back to the > process, or terminate it in some graceful manner? > > > > > However, I suppose there's little harm in letting the patch in. I would guess > > the additions all optimise away if memcg isn't enabled. > > > > A question for you: why does struct page_cgroup need a page pointer? If an > > array of page_cgroup structs is allocated per array of page structs, then you > > should be able to use the array index to map between them. > > Kame is probably better able to answer this. > To answer David's question: We have no notion of pfn in page_cgroup, how do we do the indexing? BTW, we are moving to cgroup ids that will take just 16 bits instead of 64 on a 64 bit system. -- Three Cheers, Balbir -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxxx For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>