On Thu, 19 Oct 2017, Daniel Pryor wrote:
> Hello Everyone,
>
> We are currently running into two issues.
>
> 1) We are noticing huge pauses during directory creation, but our file write
> times are super fast. The metadata and data pools are on the same
> infrastructure.
> * https://gist.github.com/pryorda/a0d5c37f119c4a320fa4ca9d48c8752b
> * https://gist.github.com/pryorda/ba6e5c2f94f67ca72a744b90cc58024e
Separate metadata onto different (ideally faster) devices is usually a
good idea if you want to protect metadata performance. The stalls you're
seeing could either be MDS requests getting slowed down by the OSDs or it
might be the MDS missing something in it's cache and having to go
fetch or flush something to RADOS. You might see if increasing the MDS
cache size helps.
> 2) Since we were having the issue above, we wanted to possibly move to a
> larger top level directory. Stuff everything in there and later move
> everything out via a batch job. To do this we need to increase the the
> directory limit from 100,000 to 300,000. How do we increase this limit.
I would recommend upgrading to luminous and enabling directory
fragmentation instead of increasing the per-fragment limit on Jewel. Big
fragments have a negative impact on MDS performance (leading to spikes
like you see above) and can also make life harder for the OSDs.
sage
>
>
> dpryor@beta-ceph-node1:~$ dpkg -l |grep ceph
> ii ceph-base 10.2.10-1xenial
> amd64 common ceph daemon libraries and management tools
> ii ceph-common 10.2.10-1xenial
> amd64 common utilities to mount and interact with a ceph storage
> cluster
> ii ceph-deploy 1.5.38
> all Ceph-deploy is an easy to use configuration tool
> ii ceph-mds 10.2.10-1xenial
> amd64 metadata server for the ceph distributed file system
> ii ceph-mon 10.2.10-1xenial
> amd64 monitor server for the ceph storage system
> ii ceph-osd 10.2.10-1xenial
> amd64 OSD server for the ceph storage system
> ii libcephfs1 10.2.10-1xenial
> amd64 Ceph distributed file system client library
> ii python-cephfs 10.2.10-1xenial
> amd64 Python libraries for the Ceph libcephfs library
> dpryor@beta-ceph-node1:~$
>
> Any direction would be appreciated!?
>
> Thanks,
> Daniel
>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com