Jewel + kernel 4.4 Massive performance regression (-50%)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,


Just a quick update since I didn't have time for this yesterday.

I did a similar test as below with only the XFS node active and as expected
results are opposite:

3937 IOPS 3.16
3595 IOPS 4.9

As opposed to what I found out yesterday:
---
Thus I turned off the XFS node and ran the test again with just the EXT4
node active. And this time 4.9 came out (slightly) ahead:

3645 IOPS 3.16
3970 IOPS 4.9

---

Christian

On Mon, 20 Feb 2017 13:10:38 +0900 Christian Balzer wrote:

> Hello,
> 
> On Thu, 16 Feb 2017 17:51:18 +0200 Kostis Fardelas wrote:
> 
> > Hello,
> > we are on Debian Jessie and Hammer 0.94.9 and recently we decided to
> > upgrade our kernel from 3.16 to 4.9 (jessie-backports). We experience
> > the same regression but with some shiny points  
> 
> Same OS, kernels and Ceph version here, but I can't reproduce this for
> the most part, probably because of other differences.
> 
> 4 nodes, 
> 2 with 4 HDD based and SSD journal OSDs,
> 2 with 4 SSD based OSDs (cache-tier),
> replication 2.
> Half of the nodes/OSDs are using XFS, the other half EXT4.
> 
> > -- ceph tell osd average across the cluster --
> > 3.16.39-1: 204MB/s
> > 4.9.0-0    : 158MB/s
> >   
> The "ceph osd tell bench" is really way too imprecise and all over the
> place for me, but the average of the HDD based OSDs doesn't differ
> noticeably.
> 
> > -- 1 rados bench client 4K 2048 threads avg IOPS --
> > 3.16.39-1: 1604
> > 4.9.0-0    : 451
> >   
> I'd think 32-64 threads will do nicely.
> As discussed on the ML before, this test is also not particular realistic
> when it comes to actual client performance, but still, a data point is a
> data point. 
> 
> And incidentally this is the only test where I can clearly see something
> similar, with 64 threads and 4K:
> 
> 3400 IOPS 3.16
> 2600 IOPS 4.9
> 
> So where you are seeing a 70% reduction, I'm seeing "only" 25% less.
> 
> Which is of course a perfect match for my XFS vs. EXT4 OSD ratio.
> 
> Thus I turned off the XFS node and ran the test again with just the EXT4
> node active. And this time 4.9 came out (slightly) ahead:
> 
> 3645 IOPS 3.16
> 3970 IOPS 4.9
> 
> So this looks like a regression when it comes to CEPH interacting with XFS.
> Probably aggravated by how the "bench" tests work (lots of object
> creation), as opposed to normal usage with existing objects as tested
> below.
> 
> > -- 1 rados bench client 64K 512 threads avg BW MB/s--
> > 3.16.39-1: 78
> > 4.9.0-0    : 31
> >  
> With the default 4MB block size, no relevant difference here again.
> But then again, this creates only a few objects compared to 4KB.
>  
> I've run fio (4M write, 4K write, 4k randwrite) from within a VM against
> the cluster with both kernel versions, no noticeable difference there
> either.
> 
> Just to compare this to the rados bench tests above:
> ---
> root at tvm-01:~# fio --size=18G --ioengine=libaio --invalidate=1 --direct=1 --numjobs=1 --rw=write --name=fiojob --blocksize=4M --iodepth=64
> 
> fiojob: (g=0): rw=write, bs=4M-4M/4M-4M/4M-4M, ioengine=libaio, iodepth=64
> fio-2.1.11
>   write: io=18432MB, bw=359772KB/s, iops=87, runt= 52462msec
> ---
> OSD processes are at about 35% CPU usage (100% = 1 core), SSDs are at about
> 85% utilization. 
> 
> ---
> root at tvm-01:~# fio --size=4G --ioengine=libaio --invalidate=1 --direct=1 --numjobs=1 --rw=write --name=fiojob --blocksize=4K --iodepth=64
> 
> fiojob: (g=0): rw=write, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio, iodepth=64
> fio-2.1.11
>   write: io=4096.0MB, bw=241984KB/s, iops=60495, runt= 17333msec
> ---
> OSD processes are at about 20% CPU  usage, SSDs are at 50%
> utilization.
> 
> ---
> root at tvm-01:~# fio --size=2G --ioengine=libaio --invalidate=1 --direct=1 --numjobs=1 --rw=randwrite --name=fiojob --blocksize=4K --iodepth=64
> 
> fiojob: (g=0): rw=randwrite, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio, iodepth=64
> fio-2.1.11
>   write: io=2048.0MB, bw=36086KB/s, iops=9021, runt= 58115msec
> ---
> OSD processes are at 300% (and likely wanting more) CPU usage, SSDs at
> about 25% utilization
> 
> Christian
> 
> > The shiny points are on the following tests:
> > 1 rados bench client 4K 512 threads avg IOPS
> > 1 rados bench client 64K 2048 threads avg BW MB/s
> > 
> > where machines with kernel 4.9 seem to perform slightly better. The
> > overall impression though is that there is a serious regression or
> > something that should be tuned to get the same performance out of the
> > cluster.
> > 
> > Our demo cluster is 4 nodes X 12 OSDs, separate journal on SSD,
> > firefly tunables and everything else default considering our Ceph
> > installation and Debian OS. Each rados bench run 5 times to get an
> > average and caches were dropped before each test.
> > 
> > I wonder if anyone has discovered the culprit so far? Any hints from
> > others to focus our investigation on?
> > 
> > Best regards,
> > Kostis
> > 
> > On 19 December 2016 at 17:17, Yoann Moulin <yoann.moulin at epfl.ch> wrote:  
> > > Hello,
> > >
> > > Finally, I found time to do some new benchmarks with the latest jewel release (10.2.5) on 4 nodes. Each node has 10 OSDs.
> > >
> > > I ran 2 times "ceph tell osd.* bench" over 40 OSDs, here the average speed :
> > >
> > > 4.2.0-42-generic      97.45 MB/s
> > > 4.4.0-53-generic      55.73 MB/s
> > > 4.8.15-040815-generic 62.41 MB/s
> > > 4.9.0-040900-generic  60.88 MB/s
> > >
> > > I have the same behaviour with at least 35 to 40% performance drop between kernel 4.2 and kernel > 4.4
> > >
> > > I can do further benches if needed.
> > >
> > > Yoann
> > >
> > > Le 26/07/2016 ? 09:09, Lomayani S. Laizer a ?crit :    
> > >> Hello,
> > >> do you have journal on disk too ?
> > >>
> > >> Yes am having journal on same hard disk.
> > >>
> > >> ok and could you do bench with kernel 4.2 ? just to see if you have better
> > >> throughput. Thanks
> > >>
> > >> In ubuntu 14 I was running 4.2 kernel. the throughput was the same around 80-90MB/s per osd. I cant tell the difference because each test gives
> > >> the speeds on same range. I did not test kernel 4.4 in ubuntu 14
> > >>
> > >>
> > >> --
> > >> Lomayani
> > >>
> > >> On Tue, Jul 26, 2016 at 9:39 AM, Yoann Moulin <yoann.moulin at epfl.ch <mailto:yoann.moulin at epfl.ch>> wrote:
> > >>
> > >>     Hello,
> > >>    
> > >>     > Am running ubuntu 16 with kernel 4.4-0.31-generic and my speed are similar.    
> > >>
> > >>     do you have journal on disk too ?
> > >>    
> > >>     > I did tests on ubuntu 14 and Ubuntu 16 and the speed is similar. I have around
> > >>     > 80-90MB/s of OSD speeds in both operating systems    
> > >>
> > >>     ok and could you do bench with kernel 4.2 ? just to see if you have better
> > >>     throughput. Thanks
> > >>    
> > >>     > Only issue am observing now with ubuntu 16 is sometime osd fails on rebooting
> > >>     > until i start them manually or adding starting commands in rc.local.    
> > >>
> > >>     in my case, it's a test environment, so I don't have notice those behaviours
> > >>
> > >>     --
> > >>     Yoann
> > >>    
> > >>     > On Mon, Jul 25, 2016 at 6:45 PM, Yoann Moulin <yoann.moulin at epfl.ch <mailto:yoann.moulin at epfl.ch>
> > >>     > <mailto:yoann.moulin at epfl.ch <mailto:yoann.moulin at epfl.ch>>> wrote:
> > >>     >
> > >>     >     Hello,
> > >>     >
> > >>     >     (this is a repost, my previous message seems to be slipping under the radar)
> > >>     >
> > >>     >     Does anyone get a similar behaviour to the one described below ?
> > >>     >
> > >>     >     I found a big performance drop between kernel 3.13.0-88 (default kernel on
> > >>     >     Ubuntu Trusty 14.04) or kernel 4.2.0 and kernel 4.4.0.24.14 (default kernel on
> > >>     >     Ubuntu Xenial 16.04)
> > >>     >
> > >>     >     - ceph version is Jewel (10.2.2).
> > >>     >     - All tests have been done under Ubuntu 14.04 on
> > >>     >     - Each cluster has 5 nodes strictly identical.
> > >>     >     - Each node has 10 OSDs.
> > >>     >     - Journals are on the disk.
> > >>     >
> > >>     >     Kernel 4.4 has a drop of more than 50% compared to 4.2
> > >>     >     Kernel 4.4 has a drop of 40% compared to 3.13
> > >>     >
> > >>     >     details below :
> > >>     >
> > >>     >     With the 3 kernel I have the same performance on disks :
> > >>     >
> > >>     >     Raw benchmark:
> > >>     >     dd if=/dev/zero of=/dev/sdX bs=1M count=1024 oflag=direct    => average ~230MB/s
> > >>     >     dd if=/dev/zero of=/dev/sdX bs=1G count=1 oflag=direct       => average ~220MB/s
> > >>     >
> > >>     >     Filesystem mounted benchmark:
> > >>     >     dd if=/dev/zero of=/sdX1/test.img bs=1G count=1              => average ~205MB/s
> > >>     >     dd if=/dev/zero of=/sdX1/test.img bs=1G count=1 oflag=direct => average ~214MB/s
> > >>     >     dd if=/dev/zero of=/sdX1/test.img bs=1G count=1 oflag=sync   => average ~190MB/s
> > >>     >
> > >>     >     Ceph osd Benchmark:
> > >>     >     Kernel 3.13.0-88-generic : ceph tell osd.ID bench => average  ~81MB/s
> > >>     >     Kernel 4.2.0-38-generic  : ceph tell osd.ID bench => average ~109MB/s
> > >>     >     Kernel 4.4.0-24-generic  : ceph tell osd.ID bench => average  ~50MB/s
> > >>     >
> > >>     >     I did new benchmarks then on 3 new fresh clusters.
> > >>     >
> > >>     >     - Each cluster has 3 nodes strictly identical.
> > >>     >     - Each node has 10 OSDs.
> > >>     >     - Journals are on the disk.
> > >>     >
> > >>     >     bench5 : Ubuntu 14.04 / Ceph Infernalis
> > >>     >     bench6 : Ubuntu 14.04 / Ceph Jewel
> > >>     >     bench7 : Ubuntu 16.04 / Ceph jewel
> > >>     >
> > >>     >     this is the average of 2 runs of "ceph tell osd.* bench" on each cluster (2 x 30
> > >>     >     OSDs)
> > >>     >
> > >>     >     bench5 / 14.04 / Infernalis / kernel 3.13 :  54.35 MB/s
> > >>     >     bench6 / 14.04 / Jewel      / kernel 3.13 :  86.47 MB/s
> > >>     >
> > >>     >     bench5 / 14.04 / Infernalis / kernel 4.2  :  63.38 MB/s
> > >>     >     bench6 / 14.04 / Jewel      / kernel 4.2  : 107.75 MB/s
> > >>     >     bench7 / 16.04 / Jewel      / kernel 4.2  : 101.54 MB/s
> > >>     >
> > >>     >     bench5 / 14.04 / Infernalis / kernel 4.4  :  53.61 MB/s
> > >>     >     bench6 / 14.04 / Jewel      / kernel 4.4  :  65.82 MB/s
> > >>     >     bench7 / 16.04 / Jewel      / kernel 4.4  :  61.57 MB/s
> > >>     >
> > >>     >     If needed, I have the raw output of "ceph tell osd.* bench"
> > >>     >
> > >>     >     Best regards    
> > >>
> > >>    
> > >
> > >
> > > --
> > > Yoann Moulin
> > > EPFL IC-IT
> > > --
> > > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> > > the body of a message to majordomo at vger.kernel.org
> > > More majordomo info at  http://vger.kernel.org/majordomo-info.html    
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users at lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com  
> 
> 


-- 
Christian Balzer        Network/Systems Engineer                
chibi at gol.com   	Global OnLine Japan/Rakuten Communications
http://www.gol.com/


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux