Re: [PATCH 2/2] fs/ext4,jbd2: Add support for passing write-hint with journal.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




On Tuesday 11 December 2018 10:37 AM, Jens Axboe wrote:
On 12/10/18 9:07 PM, Dave Chinner wrote:
On Mon, Dec 10, 2018 at 08:44:32AM -0700, Jens Axboe wrote:
On 12/10/18 8:41 AM, Jan Kara wrote:
On Mon 10-12-18 08:17:18, Jens Axboe wrote:
On 12/10/18 7:12 AM, Jan Kara wrote:
On Mon 10-12-18 18:20:04, Kanchan Joshi wrote:
This patch introduces "j_writehint" in JBD2 journal,
which is set based by Ext4 depending on "journal_writehint"
mount option (inspired from "journal_ioprio").

Thanks for the patch! It would be good to provide the explanation you have
in the cover letter in this patch as well so that it gets recorded in git
logs.

Also I don't like the fact that users have to set the hint via a mount
option for this to be enabled. OTOH the WRITE_FILE_<foo> hints defined in
fs.h are generally supposed to be used by userspace so it's difficult to
pick anything if we don't know what the userspace is going to do. I'd argue
it's even difficult for the sysadmin to pick any good value even if he
actually knows that he might benefit from setting some. Jens, is there
some reasonable way for fs to automatically pick some stream value for its
journal?

I think we have two options here:

1) It's _probably_ safe to assume that journal data is short lived. While
    hints are all relative to the specific use case, the size of the journal
    compared to the rest of the drive is most likely very small. Hence a
    default of WRITE_LIFE_SHORT is probably a good idea.

That's what I was thinking as well. But there are some exceptions like
heavy DB load where there's very little of metadata modified (and thus
almost no journal IO) compared to the DB data. So journal blocks may have
actually longer life time than data blocks. OTOH if there's little journal
IO there's no big benefit in specifying a stream for it so WRITE_LIFE_SHORT
is probably a good default anyway.

Right, that's my probably, it would definitely not work for all cases. But
it only really fails if two uses of the same life time ends up being vastly
different. It doesn't matter if LIFE_SHORT ends up being the longest life
time of them all.

2) We add a specific internal life time hint for fs journals.

#2 makes the most sense to me, but requires a bit more work...

Yeah, #2 would look more natural to me but I guess it needs some mapping to
what the drive offers, doesn't it?

We only used 4 streams, drives generally offer a lot more. So we can expand
it quite easily.

Can we get the number of stream supported from the drive? If we can
get at this at mount time, we can use high numbers down for internal
filesystem stuff, and low numbers up for user data (as already
defined by the fcntl interface).

If the hardware doesn't support streams or doesn't support any more
than the userspace interface covers, then it is probably best not to
use hints at all for metadata...

Yes, we query these values. For instance, if we can't get the current
number of streams we support (4), then we don't use them. We currently
don't export this anywhere for the kernel to see, but that could be
rectified. In terms of values, the NVMe stream space is 16-bits, so
we could allocate from 65535 and down. There are no restrictions on
ordering, so it'd be perfectly fine to use your suggestion of top down
for the kernel.

In terms of hardware support, we assign a number of streams per
namespace, and there's a fixed number of concurrently open streams per
drive. We can add reservations, for instance 8, for each namespace.
That'll give you the 4 user streams, and 4 for the kernel, 65535..65532.

Jens,

Currently hints from two or more independent user-space applications may accidentally fall into same stream as there is no arbiter. By keeping two sets of write-hints, user-space does not get in the way of kernel-space. But I wonder about multiple users of this ,kernel-private, hint-set (say multiple FS) - who gets to use what, and whether we need arbiter for that. And that makes me remember your earlier work when streams were streams (i.e. plain numbers) and block layer managed a bitmap for allocation. OTOH, I doubt likelihood of in-kernel stream-collision in real-world.

Thanks,



[Index of Archives]     [Reiser Filesystem Development]     [Ceph FS]     [Kernel Newbies]     [Security]     [Netfilter]     [Bugtraq]     [Linux FS]     [Yosemite National Park]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Samba]     [Device Mapper]     [Linux Media]

  Powered by Linux