Some NFS performance numbers for 2.6.31

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I'm running some performance tests on NFSv3 on a slightly hacked 2.6.31 kernel.

Both:  64-bit linux.
  Dual-port 10G NIC, 82599 (I think, at any rate it's the new Intel 5GT/s pci-e gen2 chipset, ixgbe driver)
  MTU 1500

Server:  dual Intel E5530 2.4Ghz processors.
  Serving 2 25M files from tmpfs (RAM)

Client:  Core i7 3.2Ghz, quad-core.
  I'm running 10 mac-vlans on each physical interface, one NFS mount per interface
   (my patches are to allow multiple mounts per client OS).
  one reader for each mac-vlan and one on the physical, reading 16m chunks
  O_DIRECT is enabled for the readers.
  Mounts are using 'noatime', leaving everything else at defaults.

Total read bandwidth is about 12Gbps, but it varies quite a bit and I've seen short term (10 seconds or so)
 averages go up to 15Gbps.  These are on-the-wire stats reported by the NICs, not actual NFS throughput.

Some things of interest:
*  Rates bounce around several Gbps
*  I see tcp retransmits in netstat -s on the server
*  I see zero errors (pkt loss, etc) reported by the NICs.
*  Reading 100M files slows down the test.
*  2M and 16M reads are about equivalent (the normal bouncing of the rates makes it hard to tell)
*  Messing with rmem max, backlog, and other network tunings doesn't seem to matter.
*  Running 10 readers (5 on each physical NIC) ran at 16Gbps for a few seconds (higher than I'd
   seen with 10..but then it went back down to around 13Gbps.
*  Running 6 seems slower..about 11.5Gbps on average.
*  Using 9000 MTU yields a fairly steady 16.5Gbps throughput.  This may be about the max
    IO bandwidth for the server machine...but I know the client can do full 10Gbps tx + rx
    on both ports (using pktgen and 1514 byte pkts).

Here is snippet of top on the client.  bthelper is the thing doing the reads:

top - 17:26:25 up 13 min,  3 users,  load average: 25.45, 23.12, 13.98
Tasks: 227 total,   3 running, 224 sleeping,   0 stopped,   0 zombie
Cpu(s):  0.2%us,  1.4%sy,  0.0%ni, 74.4%id,  0.0%wa,  0.6%hi, 23.4%si,  0.0%st
Mem:  12326604k total,  1343680k used, 10982924k free,    30912k buffers
Swap: 14254072k total,        0k used, 14254072k free,   112084k cached

 1586 root      15  -5     0    0    0 R 39.2  0.0   1:47.91 rpciod/6
   22 root      15  -5     0    0    0 R 38.5  0.0   1:46.94 ksoftirqd/6
 3841 root       3 -17 49320  33m  980 D 20.3  0.3   0:23.05 bthelper
 3840 root       3 -17 49320  33m  980 D 18.9  0.3   0:23.43 bthelper
 3836 root       3 -17 49320  33m  980 D 12.0  0.3   0:26.39 bthelper
 3849 root       3 -17 49320  33m  980 D 11.3  0.3   0:22.74 bthelper


And, here is the server:

top - 17:28:02 up  2:03,  2 users,  load average: 0.08, 0.47, 0.68
Tasks: 291 total,   1 running, 290 sleeping,   0 stopped,   0 zombie
Cpu(s):  0.0%us,  2.5%sy,  0.0%ni, 95.5%id,  0.0%wa,  0.2%hi,  1.7%si,  0.0%st
Mem:  12325312k total,   617720k used, 11707592k free,     8240k buffers
Swap:        0k total,        0k used,        0k free,   101016k cached

  PID USER      PR  NI  VIRT  RES  SHR S %CPU %MEM    TIME+  COMMAND
 2171 root      15  -5     0    0    0 S  4.0  0.0   4:41.54 nfsd
 2163 root      15  -5     0    0    0 S  3.6  0.0   3:06.09 nfsd
 2166 root      15  -5     0    0    0 S  3.6  0.0   4:45.78 nfsd
 2176 root      15  -5     0    0    0 S  3.6  0.0   4:45.45 nfsd
 2164 root      15  -5     0    0    0 S  3.3  0.0   3:08.28 nfsd
 2165 root      15  -5     0    0    0 S  3.3  0.0   4:20.12 nfsd
 2167 root      15  -5     0    0    0 S  3.3  0.0   4:31.29 nfsd
 2170 root      15  -5     0    0    0 S  3.3  0.0   4:50.80 nfsd
 2174 root      15  -5     0    0    0 S  3.3  0.0   4:50.74 nfsd
 2168 root      15  -5     0    0    0 S  3.0  0.0   4:53.15 nfsd
 2169 root      15  -5     0    0    0 S  3.0  0.0   4:48.07 nfsd


I'm curious if anyone else has done similar testing...

Thanks,
Ben

--
Ben Greear <greearb@xxxxxxxxxxxxxxx>
Candela Technologies Inc  http://www.candelatech.com

--
To unsubscribe from this list: send the line "unsubscribe linux-nfs" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux Filesystem Development]     [Linux USB Development]     [Linux Media Development]     [Video for Linux]     [Linux NILFS]     [Linux Audio Users]     [Yosemite Info]     [Linux SCSI]

  Powered by Linux