Re: ceph/rados performace sync vs async

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Daniel;

How is your pool configured? Replica, or Erasure-Coded?

I don't actually know any of this, but...

I would expect that a synchronous call to a replica pool (R=3) would look something like this:
Client --> PG Master Host (data)
PG Master Host --> Local Disk (data)
PG Master Host --> PG Replica Host 1 (data)
PG Master Host --> PG Replica Host 2 (data)
PG Replica Host 1 --> Disk (data)
PG Replica Host 2 --> Disk (data)
PG Replica Host 1 --> PG Master Host (ack)
PG Replica Host 2 --> PG Master Host (ack)
PG Master Host --> Client (ack)

Some of that can happen in parallel, for instance the master could be transferring the file to both replica hosts, while also writing it to disk.

You can imagine why a synchronous call could be significantly slower though, yes?

Thank you,

Dominic L. Hilsbos, MBA 
Director – Information Technology 
Perform Air International, Inc.
DHilsbos@xxxxxxxxxxxxxx 
www.PerformAir.com


-----Original Message-----
From: Daniel Mezentsev [mailto:dan@xxxxxxxxxx] 
Sent: Friday, July 17, 2020 3:05 PM
To: ceph-users@xxxxxxx
Subject:  ceph/rados performace sync vs async

Hi All,

  I started a small project related to metrics collection and  
processing, Ceph was chosen as a storage backend. Decided to use rados  
directly, to avoid any additional layers. I got a very simple client -  
it works fine, but performance is very low. Can't get more than  
30-35MBsec. Rados bench shows 200MBsec for my test pool. Should be  
mentioned about the client - I'm using sbcl (yep lisp). Call to rados  
API is just cffi call. 
   
  Did try in async mode. Wow ! Saturated network bandwidth for large  
objects (4Mb and bigger), for small objects - saturated OSD IOPS  
~2.4KIOPS for 8 SAS disks, so ~300 IOPS per disk - that sounds pretty  
reasonable. Bottom line - issue not with the lisp client - im getting  
close to C performance, difference is sync vs async IO.
   
  Why it's so big - sync operations are approx 2-3 times slower then async .
  Daniel Mezentsev, founder
(+1) 604 313 8592.
Soleks Data Group.
Shaping the clouds.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux