Re: ceph-fuse performance about hammer and jewel

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, May 30, 2016 at 10:22 PM, qisy <qisy@xxxxxxxxxxxx> wrote:
> Hi,
>     After jewel released fs product ready version, I upgrade the old hammer
> cluster, but iops droped a lot
>
>     I made a test, with 3 nodes, each one have 8c 16G 1osd, the osd device
> got 15000 iops
>
>     I found ceph-fuse client has better performance on hammer than jewel.
>
>     fio randwrite 4K
>     |                       | jewel server | hammer server |
>     |jewel client      |  480+ iops    |      no test         |
>     |hammer client |  6000+ iops  |   6000+ iops     |

please post the fio config file.

Regards
Yan, Zheng

>
>     ceph-fuse(jewel) mount with jewel server got pity iops, is there any
> special options need to set?
>     If I continue use ceph-fuse(hammer) with jewel server, any problems will
> cause?
>
>     thanks
>
>     my ceph.conf below:
>
> [global]
> fsid = xxxxxxx
> mon_initial_members = xxx, xxx, xxx
> mon_host = 10.0.0.1,10.0.0.2,10.0.0.3
> auth_cluster_required = cephx
> auth_service_required = cephx
> auth_client_required = cephx
>
> filestore_xattr_use_omap = true
> osd_pool_default_size = 2
> osd_pool_default_min_size = 1
> mon_data_avail_warn = 15
> mon_data_avail_crit = 5
> mon_clock_drift_allowed = 0.6
>
> [osd]
> osd_disk_threads = 8
> osd_op_threads = 8
> journal_block_align = true
> journal_dio = true
> journal_aio = true
> journal_force_aio = true
> filestore_journal_writeahead = true
> filestore_max_sync_interval = 15
> filestore_min_sync_interval = 10
> filestore_queue_max_ops = 25000
> filestore_queue_committing_max_ops = 5000
> filestore_op_threads = 32
> osd_journal_size = 20000
> osd_map_cache_size = 1024
> osd_max_write_size = 512
> osd_scrub_load_threshold = 1
> osd_heartbeat_grace = 30
>
> [mds]
> mds_session_timeout = 120
> mds_session_autoclose = 600
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux