Re: Poor Read Performance with Ubuntu 14.04 LTS 3.19.0-30 Kernel

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> Hi,
> 
> Very interesting!  Did you upgrade the kernel on both the OSDs and clients
or
> just some of them?  I remember there were some kernel performance
> regressions a little while back.  You might try running perf during your
tests
> and look for differences.  Also, iperf might be worth trying to see if
it's a
> network regression.
> 
> I also have a script that compares output from sysctl which might be worth
> trying to see if any defaults changes.
> 
> https://github.com/ceph/cbt/blob/master/tools/compare_sysctl.py
> 
> basically just save systctl -a with both kernels and pass them as
arguments to
> the python script.
> 
> Mark

Mark,

The testing was done with 3.19 on the client with 3.13 on the OSD nodes
using "rados bench -p bench 50 seq" with an initial "rados bench -p bench 50
write --no-cleanup". We suspected the network as well and tested with iperf
as one of our first steps and saw expected speeds (9.9Gb/s as we are using
bonded X540-T2 interfaces) on both kernels. As an added data point, we have
no problem with write performance to the same pool with the same kernel
configuration (~1GB/s). We also checked the values of read_ahead_kb of the
block devices but both were shown to be the default of 128 (we have since
changed these to 4096 in our configuration, but the results were seen with
the default of 128).

We are in the process of rebuilding the entire cluster to use 3.13 and a
completely fresh installation of Ceph to make sure nothing else is at play
here.

We did check a few things in iostat and collectl, but we didn't see any read
IO against the OSDs, so I am leaning towards something further up the stack.

Just a little more background on the cluster configuration:

Specific pool created just for benchmarking, using 512 pgs and pgps and 2
replicas. Using 3 OSD nodes (also handling MON duties) with 8 SATA 7.2K RPM
OSDs and 2 NVMe journals (4 OSD to 1 Journal ratio). 1 x Hex core CPUs with
32GB of RAM per OSD node.

Tom


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux