> On Thu, Dec 18, 2008 at 10:00 AM, Jan Kara <jack@xxxxxxx> wrote: > > Interesting although I'm not that surprised because those tests seem > > to do a lot of data changes (which are never journaled in fact) and tiny > > amount of metadata changes. If you run some benchmark doing lots of > > directory operations, I guess the numbers would be considerably > > different. > > Actually we have some compile bench numbers (coming to this list soon) > that also surprised us. > The stages of compile bench that I believe are dominated by directory > operations are also showing improvements without the journal. Yes, compilebench excercises directory operations quite heavily so that is the kind of benchmark I'd be interested in :). > > Maybe trying dbench (I know it's kind of stupid ;) or > > postmark will show the differences better. > > I admit also we see huge variance using dbench on subsequent runs. To > the point where I don't know how much I trust it's numbers. > Is this a tool that people on this list have a lot of faith in? Not really (at least as far as I'm concerned :). I had to deal with dbench recently tracking some reported performance regression and lost a lot of my faith in it ;). It measures something but the numbers vary greatly and also the load it puts on the filesystem is not much realistic (lots of rewrites of the same pages and file deletes just after creation). I think compilebench is a better choice, I just didn't remember its name when I was writing the email. Honza -- Jan Kara <jack@xxxxxxx> SuSE CR Labs -- To unsubscribe from this list: send the line "unsubscribe linux-ext4" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html