Re: Specify omap path for filestore

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Maybe, I figured that the call to DBObjectMap::sync in FileStore::sync
should take care of it though?
-Sam

On Sat, Oct 31, 2015 at 11:41 PM, Chen, Xiaoxi <xiaoxi.chen@xxxxxxxxx> wrote:
> As we use submit_transaction(instead of submit_transaction_sync) in DBObjectmap, and we also don't  use a kv_sync_thread for DB. Seems we need to rely on the syncfs(2) at commit time for persist everything?
>
> If that is the case, moving db out of the same FS as Data  may cause issue?
>
>
>> -----Original Message-----
>> From: ceph-devel-owner@xxxxxxxxxxxxxxx [mailto:ceph-devel-
>> owner@xxxxxxxxxxxxxxx] On Behalf Of Xue, Chendi
>> Sent: Friday, October 30, 2015 10:05 AM
>> To: 'Samuel Just'
>> Cc: ceph-devel@xxxxxxxxxxxxxxx
>> Subject: Specify omap path for filestore
>>
>> Hi, Sam
>>
>> Last week I introduced about how we saw the benefit of moving omap to a
>> separate device.
>>
>> And here is the pull request:
>> https://github.com/ceph/ceph/pull/6421
>>
>> I had tested redeploy and restart ceph cluster at my setup, the codes works
>> fine.
>> one problem is do you think I should *DELETE* all the files under the
>> omap_path firstly? Because I notice if old pg data leaves there, osd daemon
>> may run into chaos. But I am not sure if it should leave to users to DELETE.
>>
>> Any thoughts?
>>
>> Also I paste some data I talked , which is about the rbd and osd write iops
>> ratio when doing randwrite to a rbd device.
>>
>> ======Here is some data=====
>> We uses 4 clients , 35 vm each to test on rbd randwrite.
>> 4 osd physical nodes, each has 10 HDD as osd and 2 ssd as journal
>> 2 replica
>> filestore_max_inline_xattr_xfs=0
>> filestore_max_inline_xattr_size_xfs=0
>>
>> Before moving omap to separate ssd, we saw a frontend and backend iops
>> ratio of 1:5.8, rbd side total iops 1206, hdd total iops 7034 Like we talked, 5.8
>> consists of 2 replica write, inode and omap writes
>> runid         op_size    op_type             QD             engine               serverNum       clie
>> ntNum         rbdNum   runtime             fio_iops         fio_bw               fio_latency
>>            osd_iops           osd_bw             osd_latency
>> 332            4k              randwrite         qd8            qemurbd           4                          4
>>                  140            400 sec              1206.000         4.987 MB/s      884.617
>> msec           7034.975          47.407 MB/s    242.620 msec
>>
>> And after moving omap to a separate ssd, we saw a frontend vs. backend
>> ratio drops to 1:2.6, rbd side total iops 5006, hdd total iops 13089
>> runid         op_size    op_type             QD             engine               serverNum       clie
>> ntNum         rbdNum   runtime             fio_iops         fio_bw               fio_latency
>>            osd_iops           osd_bw             osd_latency
>> 326            4k              randwrite         qd8            qemurbd           4                          4
>>                  140            400 sec              5006.000         19.822 MB/s    222.296
>> msec           13089.020        82.897 MB/s    482.203 msec
>>
>>
>> Best regards,
>> Chendi
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the
>> body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at
>> http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux