Re: Performance optimization

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> 
> 
> At the moment, the nodes look like this:
> 8 Nodes
> Worst CPU: i7-3930K (up to i7-6850K)

> Worst ammount of RAM: 24GB (up to 64GB)
> HDD Layout:
> 1x 1TB
> 4x 2TB
> 1x 6TB
> all sata, some just 5400rpm
> 
> I had to put the OS on the 6TB HDDs, because there are no more sata
> connections on the motherboard.
Why not on the 1TB?

> The servers, which have to be backed up, have mounted the ceph with
> cephfs.
> 99% of the files, that have to be backed up, are harddisk images, so
> sizes from 5GB to 1TB.
> 
> All files are written to an erasure-coded pool with k=6 m=2, compression
> is on passive snappy, default settings.
> 
> I'm getting really bad performace with this setup.
> This is a bench, run with: "rados -p ec_test bench -b 524288 60 write"
> while normal operations:
> 
> Total time run:         63.4957
> Total writes made:      459
> Write size:             524288
> Object size:            524288
> Bandwidth (MB/sec):     3.61442
> Stddev Bandwidth:       3.30073
> Max bandwidth (MB/sec): 16
> Min bandwidth (MB/sec): 0
> Average IOPS:           7
> Stddev IOPS:            6.6061
> Max IOPS:               32
> Min IOPS:               0
> Average Latency(s):     2.151
> Stddev Latency(s):      2.3661
> Max latency(s):         14.0916
> Min latency(s):         0.0420954
> Cleaning up (deleting benchmark objects)
> Removed 459 objects
> Clean up completed and total clean up time :35.6908
> 
> [root@testnode01 ~]# ceph osd perf
> osd  commit_latency(ms)  apply_latency(ms)
>   6                 655                655
>   9                  13                 13
>  11                  15                 15
>   7                  17                 17
>  10                  19                 19
>   8                  12                 12
>  24                 153                153
>  25                  22                 22
>  47                  20                 20
>  46                  23                 23
>  45                  43                 43
>  44                   8                  8
>  16                  26                 26
>  15                  18                 18
>  14                  14                 14
>  13                  23                 23
>  12                  47                 47
>  18                 595                595
>   1                  20                 20
>  38                  25                 25
>  17                  17                 17
>   0                 317                317
>  37                  19                 19
>  19                  14                 14
>   2                  16                 16
>  39                   9                  9
>  20                  16                 16
>   3                  18                 18
>  40                  10                 10
>  21                  23                 23
>   4                  17                 17
>  41                  29                 29
>   5                  18                 18
>  42                  16                 16
>  22                  16                 16
>  23                  13                 13
>  26                  20                 20
>  27                  10                 10
>  28                  28                 28
>  29                  13                 13
>  30                  34                 34
>  31                  10                 10
>  32                  31                 31
>  33                  44                 44
>  34                  21                 21
>  35                  22                 22
>  36                 295                295
>  43                   9                  9
> 
> 
> 
> What do you think is the most obvious Problem?

erasure-coded

> - The one 6TB disk, per node?

You get bad distribution of data, why not move drives around between these to clusters, so you have more the same in each. 

> - The OS on the 6TB disk?

You have the os combined on a disk that also acts as a ceph osd? That is not really a pretty solution. Why just not use the 1TB as os?

> What would you suggest?

Forget about cephfs, that requires a mds, and mine was already eating 12GB. You do not have that memory. Use and rbd image 3x replicated.

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux