Re: ceph luminous 12.2.4 - 2 servers better than 3 ?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks for helping

I thought that with  CEPH,  the more servers you have  the  better the performance
- that is why I am so confused 

Also, I tried to add the 4th server ( still no luck - in fact the rado bench output I included was from 4 servers, one OSD on each, bluestore, replication 2 )
  
Here is the same rados bench but with only 2 servers/ 2 OSDs

rados bench -p rbd 120 write --no-cleanup && rados bench -p rbd 120 seq
hints = 1
Maintaining 16 concurrent writes of 4194304 bytes to objects of size 4194304 for up to 120 seconds or 0 objects
Object prefix: benchmark_data_osd01_384454
  sec Cur ops   started  finished  avg MB/s  cur MB/s last lat(s)  avg lat(s)
    0       0         0         0         0         0           -           0
    1      16       101        85   339.981       340    0.244679    0.160362
    2      16       149       133   265.972       192    0.303997    0.226268
    3      16       197       181   241.307       192    0.179013    0.255609
    4      16       241       225   224.975       176    0.353464    0.272623
    5      16       289       273   218.376       192    0.303821    0.282425
    6      16       338       322   214.643       196    0.326009     0.29105
    7      16       387       371   211.977       196     0.27048    0.296497
    8      16       436       420   209.977       196    0.287188    0.299224
    9      16       479       463   205.755       172    0.380512    0.302272
   10      16       527       511   204.378       192    0.289532    0.306163
   11      16       576       560   203.614       196    0.406783    0.309271
   12      16       624       608   202.645       192    0.282266    0.312167
   13      16       667       651   200.286       172    0.377555    0.313733
   14      16       716       700   199.978       196    0.350938    0.315445
   15      16       764       748   199.445       192    0.183931    0.317474


Here is my iperf3 results

 0.00-1.00   sec  1.15 GBytes  9.92 Gbits/sec    0    850 KBytes
[  4]   1.00-2.00   sec  1.15 GBytes  9.90 Gbits/sec    0    850 KBytes


On Thu, 19 Apr 2018 at 08:28, Marc Roos <M.Roos@xxxxxxxxxxxxxxxxx> wrote:

If I may guess, because with 3 it reads from 3 and with 2 it reads only
from 2. You should be able to verify this with something like dstat -d
-D sda,sdb,sdc,sdd,sde,sdf,sdg not?

With replication of 2, objects are still being stored among the 3 nodes.

I am getting with iperf3 on 10Gbit
[ ID] Interval           Transfer     Bandwidth       Retr  Cwnd
[  4]   0.00-10.00  sec  11.5 GBytes  9.89 Gbits/sec    0   1.31 MBytes
[  4]  10.00-20.00  sec  11.5 GBytes  9.89 Gbits/sec    0   1.79 MBytes



-----Original Message-----
From: Steven Vacaroaia [mailto:stef97@xxxxxxxxx]
Sent: donderdag 19 april 2018 14:11
To: ceph-users
Subject: [ceph-users] ceph luminous 12.2.4 - 2 servers better than 3 ?

Hi,

Any idea why 2 servers with one OSD each will provide better performance
than 3 ?

Servers are identical
Performance  is impacted irrespective if I used SSD for WAL/DB or not
Basically, I am getting lots of cur MB/s zero 

Network is separate 10 GB for public and private I tested it with iperf
and I am getting 9.3 Gbs

I have tried replication by 2 and 3 with same results ( much better for
2 servers than 3 )

reinstalled CEPH multiple times
ceph.conf very simple - no major customization ( see below) I am out of
ideas - any hint will be TRULY appreciated

Steven



auth_cluster_required = cephx
auth_service_required = cephx
auth_client_required = cephx


public_network = 10.10.30.0/24
cluster_network = 192.168.0.0/24


osd_pool_default_size = 2
osd_pool_default_min_size = 1 # Allow writing 1 copy in a degraded state
osd_crush_chooseleaf_type = 1


[mon]
mon_allow_pool_delete = true
mon_osd_min_down_reporters = 1

[osd]
osd_mkfs_type = xfs
osd_mount_options_xfs =
"rw,noatime,nodiratime,attr2,logbufs=8,logbsize=256k,largeio,inode64,swa
lloc,allocsize=4M"
osd_mkfs_options_xfs = "-f -i size=2048"
bluestore_block_db_size = 32212254720
bluestore_block_wal_size = 1073741824

rados bench -p rbd 120 write --no-cleanup && rados bench -p rbd 120 seq
hints = 1 Maintaining 16 concurrent writes of 4194304 bytes to objects
of size 4194304 for up to 120 seconds or 0 objects Object prefix:
benchmark_data_osd01_383626
  sec Cur ops   started  finished  avg MB/s  cur MB/s last lat(s)  avg
lat(s)
    0       0         0         0         0         0           -       
   0
    1      16        57        41   163.991       164    0.197929   
0.065543
    2      16        57        41    81.992         0           -   
0.065543
    3      16        67        51   67.9936        20   0.0164632   
0.249939
    4      16        67        51   50.9951         0           -   
0.249939
    5      16        71        55   43.9958         8   0.0171439   
0.319973
    6      16       181       165   109.989       440   0.0159057   
0.563746
    7      16       182       166   94.8476         4    0.221421   
0.561684
    8      16       182       166   82.9917         0           -   
0.561684
    9      16       240       224   99.5458       116   0.0232989   
0.638292
   10      16       264       248   99.1901        96   0.0222669   
0.583336
   11      16       264       248   90.1729         0           -   
0.583336
   12      16       285       269   89.6579        42   0.0165706   
0.600606
   13      16       285       269   82.7611         0           -   
0.600606
   14      16       310       294   83.9918        50   0.0254241   
0.756351




_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux