Re: RBD as backend for iSCSI SAN Targets

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Wido,

I will have some new hardware for running tests in the next two weeks or
so and will report my findings once I get a chance to run some tests. I
will disable writeback on the target side as I will be attempting to
configure an ssd caching pool of 24 ssd's with writeback for the main pool
with 360 disks with a 5 osd spinners to 1 ssd journal ratio. I will be
running everything through 10Gig SFP+ Ethernet interfaces with a dedicated
cluster network interface, dedicated public ceph interface and a separate
iscsi network also with 10 gig interfaces for the target machines.

I am ideally looking for a 20,000 to 60,000 IOPS from this system if I can
get the caching pool configuration right. The application has a 30ms max
latency requirement for the storage.

In my current tests I have only spinners with SAS 10K disks, 4.2ms write
latency on the disks with separate journaling on SAS 15K disks with a
3.3ms write latency. With 20 OSDs and 4 Journals I am only concerned with
the overall operation apply latency that I have been seeing (1-6ms idle is
normal, but up to 60-170ms for a moderate workload using rbd bench-write)
however I am on a network where I am bound to 1500 mtu and I will get to
test jumbo frames with the next setup in addition to the ssd¹s. I suspect
the overall performance will be good in the new test setup and I am
curious to see what my tests will yield.

Thanks for the response!

Karol



On 2014-03-15, 12:18 PM, "Wido den Hollander" <wido@xxxxxxxx> wrote:

>On 03/15/2014 04:11 PM, Karol Kozubal wrote:
>> Hi Everyone,
>>
>> I am just wondering if any of you are running a ceph cluster with an
>> iSCSI target front end? I know this isn¹t available out of the box,
>> unfortunately in one particular use case we are looking at providing
>> iSCSI access and it's a necessity. I am liking the idea of having rbd
>> devices serving block level storage to the iSCSI Target servers while
>> providing a unified backed for native rbd access by openstack and
>> various application servers. On multiple levels this would reduce the
>> complexity of our SAN environment and move us away from expensive
>> proprietary solutions that don¹t scale out.
>>
>> If any of you have deployed any HA iSCSI Targets backed by rbd I would
>> really appreciate your feedback and any thoughts.
>>
>
>I haven't used it in production, but a couple of things which come to
>mind:
>
>- Use TGT so you can run it all in userspace backed by librbd
>- Do not use writeback caching on the targets
>
>You could use multipathing if you don't use writeback caching. Use
>writeback would also cause data loss/corruption in case of multiple
>targets.
>
>It will probably just work with TGT, but I don't know anything about the
>performance.
>
>> Karol
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
>-- 
>Wido den Hollander
>42on B.V.
>
>Phone: +31 (0)20 700 9902
>Skype: contact42on
>_______________________________________________
>ceph-users mailing list
>ceph-users@xxxxxxxxxxxxxx
>http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux