speed decrease with size

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I am testing attached volume storage on our openstack cluster which uses ceph for block storage. 
our Ceph nodes have large SSD's for their journals 50+GB for each OSD. I'm thinking some parameter is a little off because with relatively small writes I am seeing drastically reduced write speeds.


we have 2 nodes withs 12 total OSD's each with 50GB SSD Journal.


 here is our Ceph config

[global]
fsid = 19bc15fd-c0cc-4f35-acd2-292a86fbcf7d
mon_initial_members = node-5 node-4 node-3
mon_host = 192.168.0.8 192.168.0.7 192.168.0.13
auth_cluster_required = cephx
auth_service_required = cephx
auth_client_required = cephx
filestore_xattr_use_omap = true
log_to_syslog_level = info
log_to_syslog = True
osd_pool_default_size = 1
osd_pool_default_min_size = 1
osd_pool_default_pg_num = 64
public_network = 192.168.0.0/24
log_to_syslog_facility = LOG_LOCAL0
osd_journal_size = 50000
auth_supported = cephx
osd_pool_default_pgp_num = 64
osd_mkfs_type = xfs
cluster_network = 192.168.1.0/24
osd_recovery_max_active = 1
osd_max_backfills = 1

[client]
rbd_cache = True
rbd_cache_writethrough_until_flush = True

[client.radosgw.gateway]
rgw_keystone_accepted_roles = _member_, Member, admin, swiftoperator
keyring = /etc/ceph/keyring.radosgw.gateway
rgw_socket_path = /tmp/radosgw.sock
rgw_keystone_revocation_interval = 1000000
rgw_keystone_url = 192.168.0.2:35357
rgw_keystone_admin_token = ZBz37Vlv
host = node-3
rgw_dns_name = *.ciminc.com
rgw_print_continue = True
rgw_keystone_token_cache_size = 10
rgw_data = /var/lib/ceph/radosgw
user = www-data

This is the degradation I am speaking of..


dd if=/dev/zero of=/mnt/ext4/output bs=1000k count=1k; rm -f /mnt/ext4/output;
1024+0 records in
1024+0 records out
1048576000 bytes (1.0 GB) copied, 0.887431 s, 1.2 GB/s

dd if=/dev/zero of=/mnt/ext4/output bs=1000k count=2k; rm -f /mnt/ext4/output;
2048+0 records in
2048+0 records out
2097152000 bytes (2.1 GB) copied, 3.75782 s, 558 MB/s

 dd if=/dev/zero of=/mnt/ext4/output bs=1000k count=3k; rm -f /mnt/ext4/output;
3072+0 records in
3072+0 records out
3145728000 bytes (3.1 GB) copied, 10.0054 s, 314 MB/s

dd if=/dev/zero of=/mnt/ext4/output bs=1000k count=5k; rm -f /mnt/ext4/output;
5120+0 records in
5120+0 records out
5242880000 bytes (5.2 GB) copied, 24.1971 s, 217 MB/s

Any suggestions for improving the large write degradation?
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux