Re: MDS performance issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 





On Mon, Jan 21, 2019 at 12:52 AM Yan, Zheng <ukernel@xxxxxxxxx> wrote:
On Mon, Jan 21, 2019 at 12:12 PM Albert Yue <transuranium.yue@xxxxxxxxx> wrote:
>
> Hi Yan Zheng,
>
> 1. mds cache limit is set to 64GB
> 2. we get the size of meta data pool by running `ceph df` and saw meta data pool just used 200MB space.
>

That's very strange. One file uses about 1k metadata storage. 560M
files should use hundreds gigabytes.

That's presumably because OSDs still don't report LevelDB/RocksDB usage up in that view, and all the MDS metadata is stored there?
-Greg
 

> Thanks,
>
>
> On Mon, Jan 21, 2019 at 11:35 AM Yan, Zheng <ukernel@xxxxxxxxx> wrote:
>>
>> On Mon, Jan 21, 2019 at 11:16 AM Albert Yue <transuranium.yue@xxxxxxxxx> wrote:
>> >
>> > Dear Ceph Users,
>> >
>> > We have set up a cephFS cluster with 6 osd machines, each with 16 8TB harddisk. Ceph version is luminous 12.2.5. We created one data pool with these hard disks and created another meta data pool with 3 ssd. We created a MDS with 65GB cache size.
>> >
>> > But our users are keep complaining that cephFS is too slow. What we observed is cephFS is fast when we switch to a new MDS instance, once the cache fills up (which will happen very fast), client became very slow when performing some basic filesystem operation such as `ls`.
>> >
>>
>> what's your mds cache config ?
>>
>> > What we know is our user are putting lots of small files into the cephFS, now there are around 560 Million files. We didn't see high CPU wait on MDS instance and meta data pool just used around 200MB space.
>>
>> It's unlikely.  For output of 'ceph osd df', you should take both both
>> DATA and OMAP into account.
>>
>> >
>> > My question is, what is the relationship between the metadata pool and MDS? Is this performance issue caused by the hardware behind meta data pool? Why the meta data pool only used 200MB space, and we saw 3k iops on each of these three ssds, why can't MDS cache all these 200MB into memory?
>> >
>> > Thanks very much!
>> >
>> >
>> > Best Regards,
>> >
>> > Albert
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@xxxxxxxxxxxxxx
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux