On Friday 16 January 2009 21:55:30 Pekka Enberg wrote: > Hi Nick, > > On Fri, Jan 16, 2009 at 12:42 PM, Nick Piggin <nickpiggin@xxxxxxxxxxxx> wrote: > >> I don't have the exact config for the previous tests but it's was just > >> my laptop regular config whereas the new tests are x86-64 defconfig. > >> So I think I'm just hitting some of the other OLTP regressions here, > >> aren't I? There's some scheduler related options such as > >> CONFIG_GROUP_SCHED and CONFIG_FAIR_GROUP_SCHED enabled in defconfig > >> that I didn't have in the original tests. I can try without them if > >> you want but I'm not sure it's relevant for SLAB vs SLUB tests. > > > > Oh no that's fine. It just looked like you repeated the test but > > with lockdep disabled (and no other changes). > > Right. In any case, I am still unable to reproduce the OLTP issue and > I've seen SLUB beat SLAB on my machine in most of the benchmarks > you've posted. SLUB was distinctly slower on the tbench, netperf, and hackbench tests that I ran. These were faster with SLUB on your machine? What kind of system is it? > So I have very mixed feelings about SLQB. It's very > nice that it works for OLTP but we still don't have much insight (i.e. > numbers) on why it's better. According to estimates in this thread, I think Matthew said SLUB would be around 6% slower? SLQB is within measurement error of SLAB. Fair point about personally reproducing the OLTP problem yourself. But the fact is that we will get problem reports that cannot be reproduced. That does not make them less relevant. I can't reproduce the OLTP benchmark myself. And I'm fully expecting to get problem reports for SLQB against insanely sized SGI systems, which I will take very seriously and try to fix them. > I'm also bit worried if SLQB has gotten > enough attention from the NUMA and HPC folks that brought us SLUB. It hasn't, but that's the problem we're hoping to solve by getting it merged. People can give it more attention, and we can try to fix any problems. SLUB has been default for quite a while now and not able to solve all problems it has had reported against it. So I hope SLQB will be able to unblock this situation. > The good news is that SLQB can replace SLAB so either way, we're not > going to end up with four allocators. Whether it can replace SLUB > remains to be seen. Well I think being able to simply replace SLAB is not ideal. The plan I'm hoping is to have four allocators for a few releases, and then go back to having two. That is going to mean some groups might not have their ideal allocator merged... but I think it is crazy to settle with more than one main compile-time allocator for the long term. I don't know what the next redhat enterprise release is going to do, but if they go with SLAB, then I think that means no SGI systems would run in production with SLUB anyway, so what would be the purpose of having a special "HPC/huge system" allocator? Or... what other reasons should users select SLUB vs SLAB? (in terms of core allocator behaviour, versus extras that can be ported from one to the other) If we can't even make up our own minds, then will others be able to? -- To unsubscribe from this list: send the line "unsubscribe linux-scsi" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html