On Tue, 2007-09-11 at 18:47 +0200, Andrea Arcangeli wrote: > Hi Mel, > Hi, > On Tue, Sep 11, 2007 at 04:36:07PM +0100, Mel Gorman wrote: > > that increasing the pagesize like what Andrea suggested would lead to > > internal fragmentation problems. Regrettably we didn't discuss Andrea's > > The config_page_shift guarantees the kernel stacks or whatever not > defragmentable allocation other allocation goes into the same 64k "not > defragmentable" page. Not like with SGI design that a 8k kernel stack > could be allocated in the first 64k page, and then another 8k stack > could be allocated in the next 64k page, effectively pinning all 64k > pages until Nick worst case scenario triggers. > In practice, it's pretty difficult to trigger. Buddy allocators always try and use the smallest possible sized buddy to split. Once a 64K is split for a 4K or 8K allocation, the remainder of that block will be used for other 4K, 8K, 16K, 32K allocations. The situation where multiple 64K blocks gets split does not occur. Now, the worst case scenario for your patch is that a hostile process allocates large amount of memory and mlocks() one 4K page per 64K chunk (this is unlikely in practice I know). The end result is you have many 64KB regions that are now unusable because 4K is pinned in each of them. Your approach is not immune from problems either. To me, only Nicks approach is bullet-proof in the long run. > What I said at the VM summit is that your reclaim-defrag patch in the > slub isn't necessarily entirely useless with config_page_shift, > because the larger the software page_size, the more partial pages we > could find in the slab, so to save some memory if there are tons of > pages very partially used, we could free some of them. > This is true. Slub targetted reclaim (Chrisophs work) is useful independent of this current problem. > But the whole point is that with the config_page_shift, Nick's worst > case scenario can't happen by design regardless of defrag or not > defrag. I agree with this. It's why I thought Nick's approach was where we were going to finish up ultimately. > While it can _definitely_ happen with SGI design (regardless > of any defrag thing). I have never stated that the SGI design is immune from this problem. > We can still try to save some memory by > defragging the slab a bit, but it's by far *not* required with > config_page_shift. No defrag at all is required infact. > You will need to take some sort of defragmentation to deal with internal fragmentation. It's a very similar problem to blasting away at slab pages and still not being able to free them because objects are in use. Replace "slab" with "large page" and "object" with "4k page" and the issues are similar. > Plus there's a cost in defragging and freeing cache... the more you > need defrag, the slower the kernel will be. > > > approach in depth. > > Well it wasn't my fault if we didn't discuss it in depth though. If it's my fault, sorry about that. It wasn't my intention. > I > tried to discuss it in all possible occasions where I was suggested to > talk about it and where it was somewhat on topic. Who said it was off-topic? Again, if this was me, sorry - you should have chucked something at my head to shut me up. > Given I wasn't even > invited at the KS, I felt it would not be appropriate for me to try to > monopolize the VM summit according to my agenda. So I happily listened > to what the top kernel developers are planning ;), while giving > some hints on what I think the right direction is instead. > Right, clearly we failed or at least had sub-optimal results dicussion this one at VM Summit. Good job we have mail to pick up the stick with. > > I *thought* that the end conclusion was that we would go with > > Frankly I don't care what the end conclusion was. > heh. Well we need to come to some sort of conclusion here or this will go around the merri-go-round till we're all bald. > > Christoph's approach pending two things being resolved; > > > > o mmap() support that we agreed on is good > > Let's see how good the mmap support for variable order page size will > work after the 2 weeks... > Ok, I'm ok with that. > > o A clear statement, with logging maybe for users that mounted a large > > block filesystem that it might blow up and they get to keep both parts > > when it does. Basically, for now it's only suitable in specialised > > environments. > > Yes, but perhaps you missed that such printk is needed exactly to > provide proof that SGI design is the wrong way and it needs to be > dumped. If that printk ever triggers it means you were totally wrong. > heh, I suggested printing the warning because I knew it had this problem. The purpose in my mind was to see how far the design could be brought before fs-block had to fill in the holes. > > I also thought there was an acknowledgement that long-term, fs-block was > > the way to go - possibly using contiguous pages optimistically instead > > of virtual mapping the pages. At that point, it would be a general > > solution and we could remove the warnings. > > fsblock should stack on top of config_page_shift simply. It should be able to stack on top of either approach and arguably setting slub_min_order=large_block_order with large block filesystems is 90% of your approach anyway. > Both are > needed. You don't want to use 64k pages on a laptop but you may want a > larger blocksize for the btrees etc... if you've a large harddisk and > not much ram. I am still failing to see what happens when there are pagetable pages, slab objects or mlocked 4k pages pinning the 64K pages and you need to allocate another 64K page for the filesystem. I *think* you deadlock in a similar fashion to Christoph's approach but the shape of the problem is different because we are dealing with internal instead of external fragmentation. Am I wrong? > > That's the absolute worst case but yes, in theory this can occur and > > it's safest to assume the situation will occur somewhere to someone. It > > Do you agree this worst case can't happen with config_page_shift? > Yes. I just think you have a different worst case that is just as bad. > > Where we expected to see the the use of this patchset was in specialised > > environments *only*. The SGI people can mitigate their mixed > > fragmentation problems somewhat by setting slub_min_order == > > large_block_order so that blocks get allocated and freed at the same > > size. This is partial way towards Andrea's solution of raising the size > > of an order-0 allocation. The point of printing out the warnings at > > Except you don't get all the full benefits of it... > > Even if I could end up mapping 4k kmalloced entries in userland for > the tail packing, that IMHO would still be a preferable solution than > to keep the base-page small and to make an hard effort to create large > pages out of small pages. The approach I advocate keeps the base page > big and the fast path fast, and it rather does some work to split the > base pages outside the buddy for the small files. > small files (you need something like Shaggy's page tail packing), pagetable pages, pte pages all have to be dealt with. These are the things I think will cause us internal fragmentaiton problems. > All your defrag work is still good to have, like I said at the VM > summit if you remember, to grow the hugetlbfs at runtime etc... I just > rather avoid to depend on it to avoid I/O failure in presence of > mlocked pagecache for example. > I'd rather avoid depending on it for the system to work 100% of the same. Hence I've been saying that we need fsblock ultimately for this to be a 100% supported feature. > > Independently of that, we would work on order-0 scalability, > > particularly readahead and batching operations on ranges of pages as > > much as possible. > > That's pretty much an unnecessary logic, if the order0 pages become > larger. > Quite possibly. -- Mel Gorman - To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html