Re: [PATCH 00 of 41] Transparent Hugepage Support #17

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 04/12/2010 09:09 AM, Nick Piggin wrote:
On Sun, Apr 11, 2010 at 02:08:00PM +0200, Ingo Molnar wrote:
* Avi Kivity<avi@xxxxxxxxxx>  wrote:

3) futility

I think Andrea and Mel and you demonstrated that while defrag is futile in
theory (we can always fill up all of RAM with dentries and there's no 2MB
allocation possible), it seems rather usable in practice.
One problem is that you need to keep a lot more memory free in order
for it to be reasonably effective.

It's the usual space-time tradeoff. You don't want to do it on a netbook, but it's worth it on a 16GB server, which is already not very high end.

Another thing is that the problem
of fragmentation breakdown is not just a one-shot event that fills
memory with pinned objects. It is a slow degredation.

Especially when you use something like SLUB as the memory allocator
which requires higher order allocations for objects which are pinned
in kernel memory.

Won't the usual antifrag tactics apply? Try to allocate those objects from the same block.

Just running a few minutes of testing with a kernel compile in the
background does not show the full picture. You really need a box that
has been up for days running a proper workload before you are likely
to see any breakdown.

I'm sure we'll be able to generate worst-case scenarios. I'm also reasonably sure we'll be able to deal with them. I hope we won't need to, but it's even possible to move dentries around.

I'm sure it's horrible for planning if the RDBMS or VM boxes gradually
get slower after X days of uptime. It's better to have consistent
performance really, for anything except pure benchmark setups.

If that were the case we'd disable caches everywhere. General purpose computing is a best effort thing, we try to be fast on the common case but we'll be slow on the uncommon case. Access to a bit of memory can take 3 ns if it's in cache, 100 ns if not, and 3 ms if it's on disk.

Here, the uncommon case will be really uncommon, most applications (that can benefit from large pages) I'm aware of don't switch from large anonymous working sets to a dcache load of many tiny files. They tend to keep doing the same thing over and over again.

I'm not saying we don't need to adapt to changing conditions (we do, especially for kvm, that's what khugepaged is for), but as long as we have a graceful fallback, we don't need to worry too much about failure in extreme conditions.

Defrag is not futile in theory, you just have to either have a reserve
of movable pages (and never allow pinned kernel pages in there), or
you need to allocate pinned kernel memory in units of the chunk size
goal (which just gives you different types of fragmentation problems)
or you need to do non-linear kernel mappings so you can defrag pinned
kernel memory (with *lots* of other problems of course). So you just
have a lot of downsides.

Non-linear kernel mapping moves the small page problem from userspace back to the kernel, a really unhappy solution.

Very large (object count, not object size) kernel caches can be addressed by compacting them, but I hope we won't need to do that.

--
I have a truly marvellous patch that fixes the bug which this
signature is too narrow to contain.

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@xxxxxxxxxx  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@xxxxxxxxx";> email@xxxxxxxxx </a>

[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [ECOS]     [Asterisk Internet PBX]     [Linux API]