Re: Poor RBD performance as LIO iSCSI target

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi David,

I've just finished running the 75GB fio test you posted a few days back on
my new test cluster.

The cluster is as follows:-

Single server with 3x hdd and 1 ssd
Ubuntu 14.04 with 3.16.7 kernel
2+1 EC pool on hdds below a 10G ssd cache pool. SSD is also partitioned to
provide journals for hdds.
150G RBD mapped locally

The fio test seemed to run without any problems. I want to run a few more
tests with different settings to see if I can reproduce your problem. I will
let you know if I find anything.

If there is anything you would like me to try, please let me know.

Nick

-----Original Message-----
From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of
David Moreau Simard
Sent: 19 November 2014 10:48
To: Ramakrishna Nishtala (rnishtal)
Cc: ceph-users@xxxxxxxxxxxxxx; Nick Fisk
Subject: Re:  Poor RBD performance as LIO iSCSI target

Rama,

Thanks for your reply.

My end goal is to use iSCSI (with LIO/targetcli) to export rbd block
devices.

I was encountering issues with iSCSI which are explained in my previous
emails.
I ended up being able to reproduce the problem at will on various Kernel and
OS combinations, even on raw RBD devices - thus ruling out the hypothesis
that it was a problem with iSCSI but rather with Ceph.
I'm even running 0.88 now and the issue is still there.

I haven't isolated the issue just yet.
My next tests involve disabling the cache tiering.

I do have client krbd cache as well, i'll try to disable it too if cache
tiering isn't enough.
--
David Moreau Simard


> On Nov 18, 2014, at 8:10 PM, Ramakrishna Nishtala (rnishtal)
<rnishtal@xxxxxxxxx> wrote:
> 
> Hi Dave
> Did you say iscsi only? The tracker issue does not say though.
> I am on giant, with both client and ceph on RHEL 7 and seems to work ok,
unless I am missing something here. RBD on baremetal with kmod-rbd and
caching disabled.
>  
> [root@compute4 ~]# time fio --name=writefile --size=100G 
> --filesize=100G --filename=/dev/rbd0 --bs=1M --nrfiles=1 --direct=1 
> --sync=0 --randrepeat=0 --rw=write --refill_buffers --end_fsync=1 
> --iodepth=200 --ioengine=libaio
> writefile: (g=0): rw=write, bs=1M-1M/1M-1M/1M-1M, ioengine=libaio, 
> iodepth=200
> fio-2.1.11
> Starting 1 process
> Jobs: 1 (f=1): [W(1)] [100.0% done] [0KB/853.0MB/0KB /s] [0/853/0 
> iops] [eta 00m:00s] ...
> Disk stats (read/write):
>   rbd0: ios=184/204800, merge=0/0, ticks=70/16164931, 
> in_queue=16164942, util=99.98%
>  
> real    1m56.175s
> user    0m18.115s
> sys     0m10.430s
>  
> Regards,
> 
> Rama
>  
>  
> -----Original Message-----
> From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf 
> Of David Moreau Simard
> Sent: Tuesday, November 18, 2014 3:49 PM
> To: Nick Fisk
> Cc: ceph-users@xxxxxxxxxxxxxx
> Subject: Re:  Poor RBD performance as LIO iSCSI target
>  
> Testing without the cache tiering is the next test I want to do when I
have time..
>  
> When it's hanging, there is no activity at all on the cluster.
> Nothing in "ceph -w", nothing in "ceph osd pool stats".
>  
> I'll provide an update when I have a chance to test without tiering.
> --
> David Moreau Simard
>  
>  
> > On Nov 18, 2014, at 3:28 PM, Nick Fisk <nick@xxxxxxxxxx> wrote:
> > 
> > Hi David,
> > 
> > Have you tried on a normal replicated pool with no cache? I've seen 
> > a number of threads recently where caching is causing various things to
block/hang.
> > It would be interesting to see if this still happens without the 
> > caching layer, at least it would rule it out.
> > 
> > Also is there any sign that as the test passes ~50GB that the cache 
> > might start flushing to the backing pool causing slow performance?
> > 
> > I am planning a deployment very similar to yours so I am following 
> > this with great interest. I'm hoping to build a single node test 
> > "cluster" shortly, so I might be in a position to work with you on 
> > this issue and hopefully get it resolved.
> > 
> > Nick
> > 
> > -----Original Message-----
> > From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On 
> > Behalf Of David Moreau Simard
> > Sent: 18 November 2014 19:58
> > To: Mike Christie
> > Cc: ceph-users@xxxxxxxxxxxxxx; Christopher Spearman
> > Subject: Re:  Poor RBD performance as LIO iSCSI target
> > 
> > Thanks guys. I looked at http://tracker.ceph.com/issues/8818 and 
> > chatted with "dis" on #ceph-devel.
> > 
> > I ran a LOT of tests on a LOT of comabination of kernels (sometimes 
> > with tunables legacy). I haven't found a magical combination in 
> > which the following test does not hang:
> > fio --name=writefile --size=100G --filesize=100G 
> > --filename=/dev/rbd0 --bs=1M --nrfiles=1 --direct=1 --sync=0 
> > --randrepeat=0 --rw=write --refill_buffers --end_fsync=1 
> > --iodepth=200 --ioengine=libaio
> > 
> > Either directly on a mapped rbd device, on a mounted filesystem 
> > (over rbd), exported through iSCSI.. nothing.
> > I guess that rules out a potential issue with iSCSI overhead.
> > 
> > Now, something I noticed out of pure luck is that I am unable to 
> > reproduce the issue if I drop the size of the test to 50GB. Tests 
> > will complete in under 2 minutes.
> > 75GB will hang right at the end and take more than 10 minutes.
> > 
> > TL;DR of tests:
> > - 3x fio --name=writefile --size=50G --filesize=50G
> > --filename=/dev/rbd0 --bs=1M --nrfiles=1 --direct=1 --sync=0
> > --randrepeat=0 --rw=write --refill_buffers --end_fsync=1 
> > --iodepth=200 --ioengine=libaio
> > -- 1m44s, 1m49s, 1m40s
> > 
> > - 3x fio --name=writefile --size=75G --filesize=75G
> > --filename=/dev/rbd0 --bs=1M --nrfiles=1 --direct=1 --sync=0
> > --randrepeat=0 --rw=write --refill_buffers --end_fsync=1 
> > --iodepth=200 --ioengine=libaio
> > -- 10m12s, 10m11s, 10m13s
> > 
> > Details of tests here: http://pastebin.com/raw.php?i=3v9wMtYP
> > 
> > Does that ring you guys a bell ?
> > 
> > --
> > David Moreau Simard
> > 
> > 
> >> On Nov 13, 2014, at 3:31 PM, Mike Christie <mchristi@xxxxxxxxxx> wrote:
> >> 
> >> On 11/13/2014 10:17 AM, David Moreau Simard wrote:
> >>> Running into weird issues here as well in a test environment. I 
> >>> don't
> > have a solution either but perhaps we can find some things in common..
> >>> 
> >>> Setup in a nutshell:
> >>> - Ceph cluster: Ubuntu 14.04, Kernel 3.16.7, Ceph 0.87-1 (OSDs 
> >>> with separate public/cluster network in 10 Gbps)
> >>> - iSCSI Proxy node (targetcli/LIO): Ubuntu 14.04, Kernel 3.16.7, 
> >>> Ceph
> >>> 0.87-1 (10 Gbps)
> >>> - Client node: Ubuntu 12.04, Kernel 3.11 (10 Gbps)
> >>> 
> >>> Relevant cluster config: Writeback cache tiering with NVME PCI-E 
> >>> cards (2
> > replica) in front of a erasure coded pool (k=3,m=2) backed by spindles.
> >>> 
> >>> I'm following the instructions here: 
> >>> http://www.hastexo.com/resources/hints-and-kinks/turning-ceph-rbd-
> >>> im a ges-san-storage-devices No issues with creating and mapping a 
> >>> 100GB RBD image and then creating the target.
> >>> 
> >>> I'm interested in finding out the overhead/performance impact of
> > re-exporting through iSCSI so the idea is to run benchmarks.
> >>> Here's a fio test I'm trying to run on the client node on the 
> >>> mounted
> > iscsi device:
> >>> fio --name=writefile --size=100G --filesize=100G 
> >>> --filename=/dev/sdu --bs=1M --nrfiles=1 --direct=1 --sync=0 
> >>> --randrepeat=0 --rw=write --refill_buffers --end_fsync=1 
> >>> --iodepth=200 --ioengine=libaio
> >>> 
> >>> The benchmark will eventually hang towards the end of the test for 
> >>> some
> > long seconds before completing.
> >>> On the proxy node, the kernel complains with iscsi portal login
> >>> timeout: http://pastebin.com/Q49UnTPr and I also see irqbalance 
> >>> errors in syslog: http://pastebin.com/AiRTWDwR
> >>> 
> >> 
> >> You are hitting a different issue. German Anders is most likely 
> >> correct and you hit the rbd hang. That then caused the iscsi/scsi 
> >> command to timeout which caused the scsi error handler to run. In 
> >> your logs we see the LIO error handler has received a task abort 
> >> from the initiator and that timed out which caused the escalation 
> >> (iscsi portal login related messages).
> > 
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@xxxxxxxxxxxxxx
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > 
> > 
> > 
> > 
>  
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux