Le lundi 17 ao?t 2009 ? 19:49 +0400, Konstantin A. Lepikhov a ?crit : > Hi Julien! > > Monday 17, at 05:04:43 PM you wrote: > > > Le mardi 11 ao??t 2009 ?? 15:03 +0400, Konstantin A. Lepikhov a ??crit : > > > > > You can try to git clone kernel source and switch between different tags. > > > It's also very good test. > > > > Here are the final test results. The setup is : > > - 2 nodes, GbE, SATA drives, 2*4-cores Opteron 2.2Ghz, 16GB RAM > > - Ping between nodes is 0.120ms > > - GlusterFS 2.0.6 > > - Very simple setup : Replicate with readahead and writebehind. > > - Tests are done on only one node (no concurrent access) > Did you send this results to glusterfs-users list? Oops, sorry, I just hit 'reply'. Now it's done. > > The purpose of these tests is to compare GlusterFS versus local disk > > performances, on a two node cluster, as I want to host OpenVZ VEs on my > > servers. > Do you have disk load/network load statistics for this test? I haven't detailed stats, but for what I saw, there was no bottlenecks : - Load average never reached 1 - There was plenty of CPU power/RAM available during the tests - Network load was never above 30 percent of the bandwidth. It really looked as if the system was waiting for something, and my guess goes to the network. > > Untar a kernel archive : > > Local: 0:19 > > GlusterFS: 9:12 > > > > Kernel compilation : > > Local: 55:06 > > GlusterFS: 3:37:38 > > > > GIT clone kernel sources : > > Local: 5:31 > > GlusterFS: 2:49:09 > > > > So, clearly, GlusterFS solution is not viable here. I think this is > > because of network latency. As I don't think my hosting provider is > > likely to offer IB in the near future, this is a no-go. > > > > Maybe if I had dozens of servers, latency would be compensated by > > parallelism. I hope I'll be able to test it someday ;-) > Yes, latency is highly depends on configuration - I think DHT setup must be > much faster. > > > > > Anyway, thank you for your support and advice folks, I'll keep an eye on > > this project in the future. > IMHO in your setup pohmelfs/drbd8 are more acceptable. >