Yet another performance tuning for CephFS

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

 

I located and applied almost every different tuning setting/config over the internet. I couldn’t manage to speed up my speed one byte further. It is always same speed whatever I do.

 

I was on jewel, now I tried BlueStore on Luminous. Still exact same speed I gain from cephfs.

 

It doesn’t matter if I disable debug log, or remove [osd] section as below and re-add as below (see .conf). Results are exactly the same. Not a single byte is gained from those tunings. I also did tuning for kernel (sysctl.conf).

 

Basics:

 

I have 2 nodes with 10 OSD each and each OSD is 3TB SATA drive. Each node has 24 cores and 64GB of RAM. Ceph nodes are connected via 10GbE NIC. No FUSE used. But tried that too. Same results.

 

$ dd if=/dev/zero of=/mnt/c/testfile bs=100M count=10 oflag=direct

10+0 records in

10+0 records out

1048576000 bytes (1.0 GB, 1000 MiB) copied, 5.77219 s, 182 MB/s

 

182MB/s. This is the best speed i get so far. Usually 170~MB/s. Hm.. I get much much much higher speeds on different filesystems. Even with glusterfs. Is there anything I can do or try?

 

Read speed is also around 180-220MB/s but not higher.

 

This is What I am using on ceph.conf:

 

[global]

fsid = d7163667-f8c5-466b-88df-8747b26c91df

mon_initial_members = server1

mon_host = 192.168.0.1

auth_cluster_required = cephx

auth_service_required = cephx

auth_client_required = cephx

 

osd mount options = rw,noexec,nodev,noatime,nodiratime,nobarrier

osd mount options xfs = rw,noexec,nodev,noatime,nodiratime,nobarrier

osd_mkfs_type = xfs

 

osd pool default size = 2

enable experimental unrecoverable data corrupting features = bluestore rocksdb

bluestore fsck on mount = true

rbd readahead disable after bytes = 0

rbd readahead max bytes = 4194304

 

log to syslog = false

debug_lockdep = 0/0

debug_context = 0/0

debug_crush = 0/0

debug_buffer = 0/0

debug_timer = 0/0

debug_filer = 0/0

debug_objecter = 0/0

debug_rados = 0/0

debug_rbd = 0/0

debug_journaler = 0/0

debug_objectcatcher = 0/0

debug_client = 0/0

debug_osd = 0/0

debug_optracker = 0/0

debug_objclass = 0/0

debug_filestore = 0/0

debug_journal = 0/0

debug_ms = 0/0

debug_monc = 0/0

debug_tp = 0/0

debug_auth = 0/0

debug_finisher = 0/0

debug_heartbeatmap = 0/0

debug_perfcounter = 0/0

debug_asok = 0/0

debug_throttle = 0/0

debug_mon = 0/0

debug_paxos = 0/0

debug_rgw = 0/0

 

 

[osd]

osd max write size = 512

osd client message size cap = 2147483648

osd mount options xfs = rw,noexec,nodev,noatime,nodiratime,nobarrier

filestore xattr use omap = true

osd_op_threads = 8

osd disk threads = 4

osd map cache size = 1024

filestore_queue_max_ops = 25000

filestore_queue_max_bytes = 10485760

filestore_queue_committing_max_ops = 5000

filestore_queue_committing_max_bytes = 10485760000

journal_max_write_entries = 1000

journal_queue_max_ops = 3000

journal_max_write_bytes = 1048576000

journal_queue_max_bytes = 1048576000

filestore_max_sync_interval = 15

filestore_merge_threshold = 20

filestore_split_multiple = 2

osd_enable_op_tracker = false

filestore_wbthrottle_enable = false

osd_client_message_size_cap = 0

osd_client_message_cap = 0

filestore_fd_cache_size = 64

filestore_fd_cache_shards = 32

filestore_op_threads = 12

 

 

As I stated above, it doesn’t matter if I have this [osd] section or not. Results are same.

 

I am open to all suggestions.

 

Thanks,

Gencer.

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux