On Mon, Mar 16, 2009 at 9:54 PM, Greg Banks <gnb@xxxxxxx> wrote: > David, could you try your test case with this command running on the server? > > tethereal -i eth0 -f 'port 2049' -z rpc,rtt,100003,3 -w /dev/null > > and ^C it when you've done. You should see a table by RPC call with > minimum maximum and average "roundtrip times" (actually as they're > measured on the server, they should be server response times). This > should tell you which calls are slow and whether it's all those calls or > only a few outliers. OK, filtering out the empty lines, I ran the test (note that I modified it slightly, details later) with and without the server side large write running: Without: Procedure Calls Min RTT Max RTT Avg RTT GETATTR 3 0.00006 0.00133 0.00048 SETATTR 1 0.03878 0.03878 0.03878 ACCESS 9 0.00006 0.00232 0.00049 COMMIT 1 1.13381 1.13381 1.13381 With: Procedure Calls Min RTT Max RTT Avg RTT GETATTR 1 0.00016 0.00016 0.00016 SETATTR 1 30.14662 30.14662 30.14662 ACCESS 8 0.00005 0.00544 0.00154 COMMIT 1 0.34472 0.34472 0.34472 > Another experiment worth trying is a local filesystem load fairness test > on the server (no NFS client involved). This will tell you whether the > nfsd's IO priority is an issue, or if you're just seeing an IO > scheduling issue. > > 1. echo 3 > /proc/sys/vm/drop_caches > 2. time find /your/home/some/dir -ls > /dev/null (choose a directory > tree with a few hundred files) > 3. echo 3 > /proc/sys/vm/drop_caches > 4. start your large writes > 5. time find /your/home/some/dir -ls > /dev/null This test actually returned similar response times with or without the big server write running. Which led me to thinking that this isn't even a NFS problem - it actually appears to be a VM or filesystem(ext3) problem. Here's the test case again: 1. On the server, write out a large file: dd if=/dev/zero of=bigfile bs=1M count=5000 2. On the client, write out a small file: dd if=/dev/zero of=smallfile bs=4k count=1 Now, I just ran #2 both on a NFS client and on the local server - response time was the same, about 45 seconds. So it appears that my combination of server (older dual Xeon 3Ghz, 8GB RAM, SATA RAID1) lets too much dirty data accumulate creating a huge backlog of data to be written to the journal and leaving small, latency sensitive writes go very slowly. I seem to remember reading about this problem before - oh yeah: http://lwn.net/Articles/216853/ on top of this newer bug: http://bugzilla.kernel.org/show_bug.cgi?id=12309 http://lkml.org/lkml/2009/1/16/487 I'm off to go hack dirty_ratio and dirty_background_ratio for now until this bug gets fixed and ext4 appears to fully stabilize. :-) Thanks for your help, guys. -Dave -- To unsubscribe from this list: send the line "unsubscribe linux-nfs" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html