Re: Large directory block size on XFS may be harmful

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

Thanks for linking to a current update on this problem [1] [2]. I
really hope that new Ceph installations aren't still following that
old advice... it's been known to be a problem for around a year and a
half [3].

That said, the "-n size=64k" wisdom was really prevalent a few years
ago, and I wonder how many old clusters are at risk today. I manage a
sufficiently large enough number of affected OSDs that I'll be willing
to try all other possibilities before reformatting them [4]. Today
they're rock solid stable on EL6 (with hammer), but the jewel release
is getting closer and that's when we'll need to upgrade to EL7. (I've
already upgraded one host to 7 and haven't seen any problems yet, but
that one sample doesn't offer much comfort for the rest.) Anyway, it's
great to hear that there's a patch in the works... Dave deserves
infinite thanks if this gets resolved.

Cheers, Dan

[1] http://tracker.ceph.com/issues/6301
[2] https://bugzilla.redhat.com/show_bug.cgi?id=1278992
[3] https://github.com/redhat-cip/puppet-ceph/commit/b9407efd4a8a25d452e493fb48ea048e4d36e070
[4] https://access.redhat.com/solutions/1597523


On Thu, Feb 18, 2016 at 1:14 PM, Jens Rosenboom <j.rosenboom@xxxxxxxx> wrote:
> Various people have noticed performance problems and sporadic kernel
> log messages like
>
> kernel: XFS: possible memory allocation deadlock in kmem_alloc (mode:0x8250)
>
> with their Ceph clusters. We have seen this in one of our clusters
> ourselves, but not been able to reproduce it in a lab environment
> until recently. While trying to setup a benchmark for comparing the
> effect of varying the bucket shard count, I suddenly started seeing
> the same issues again and they seemed to be reproducible even with the
> latest upstream kernel.
>
> The test setup comprised 8 nodes with 2 SSDs as OSDs each. The
> messages started to appear after writing 16kb sized objects with
> cosbench using 32 workers for about 2 hours and soon after that the
> OSDs started dying because of suicide timeout.
>
> So we went ahead and tried running a kernel patched with [1], but this
> had only partial success, so I posted these results to the XFS mailing
> list. The response by Dave Chinner led to some important result:
> Creating the file system with the option "-n size=64k" was the
> culprit. Repeating the tests with sizes <=16k did not show any issues
> and the performance for this particular test even turned out to be
> better with simply letting the directory block size stay at the
> default value of 4k.
>
> In case you are seeing similar issues, you may want to check the
> directory block size of your file system, you can use xfs_info for
> that. The bad news is that the parameter cannot be changed for an
> existing file system, so you will need to reformat everything.
>
> And the morale is: Do not blindly trust configuration settings to be
> helpful, even if their use seems to be widely spread and is looking
> reasonable at first.
>
> [1] http://oss.sgi.com/pipermail/xfs/2016-January/046308.html
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux