Re: [PATCH RFC] fs: New zonefs file system

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 2019/07/16 1:54, Viacheslav Dubeyko wrote:
[...]
>>> Do you have in mind some special use-case?
>> As the commit message mentions, zonefs is not a traditional file
>> system by any
>> mean and much closer to a raw block device access interface than
>> anything else.
>> This is the entire point of this exercise: allow replacing the raw
>> block device
>> accesses with the easier to use file system API. Raw block device
>> access is also
>> file API so one could argue that this is nonsense. What I mean here
>> is that by
>> abstracting zones with files, the user does not need to do the zone
>> configuration discovery with ioctl(BLKREPORTZONES), does not need to
>> do explicit
>> zone resets with ioctl(BLKRESETZONE), does not have to "start from
>> one sector
>> and write sequentially from there" management for write() calls (i.e.
>> seeks),
>> etc. This is all replaced with the file abstraction: directory entry
>> list
>> replace zone information, truncate() replace zone reset, file current
>> position
>> replaces the application zone write pointer management.
>>
>> This simplifies implementing support of applications for zoned block
>> devices,
>> but only in cases where said applications:
>> 1) Operate with large files
>> 2) have no or only minimal need for random writes
>>
>> A perfect match for this as mentioned in the commit message are LSM-
>> tree based
>> applications such as LevelDB or RocksDB. Other examples, related,
>> include
>> Bluestore distributed object store which uses RocksDB but still has a
>> bluefs
>> layer that could be replaced with zonefs.
>>
>> As an illustration of this, Ting Yao of Huazhong University of
>> Science and
>> Technology (China) and her team modified LevelDB to work with zonefs.
>> The early
>> prototype code is on github here: https://github.com/PDS-Lab/GearDB/t
>> ree/zonefs
>>
>> LSM-Tree applications typically operate on large files, in the same
>> range as
>> zoned block device zone size (e.g. 256 MB or so). While this is
>> generally a
>> parameter that can be changed, the use of zonefs and zoned block
>> device forces
>> using the zone size as the SSTable file maximum size. This can have
>> an impact on
>> the DB performance depending on the device type, but that is another
>> discussion.
>> The point here is the code simplifications that zonefs allows.
>>
>> For more general purpose use cases (small files, lots of random
>> modifications),
>> we already have the dm-zoned device mapper and f2fs support and we
>> are also
>> currently working on btrfs support. These solutions are in my opinion
>> more
>> appropriate than zonefs to address the points you raised.
>>
> 
> Sounds pretty reasonable. But I still have two worries.
> 
> First of all, even modest file system could contain about 100K files on
> a volume. So, if our zone is 256 MB then we need in 24 TB storage
> device for 100K files. Even if we consider some special use-case of
> database, for example, then it's pretty easy to imagine the creation a
> lot of files. So, are we ready to provide such huge storage devices
> (especially, for the case of SSDs)?

The small file use case you are describing is not zonefs target use case. It
does not make any sense to discuss small files in the context of zonefs. If
small file is the use case needed for an application, then a "normal" file
system should be use such as f2fs or btrfs (zoned block device support is being
worked on, see patches posted on btrfs list).

As mentioned previously, zonefs goal is to represent zones of a zoned block
device with files, thus providing a simple abstraction one file == one zone and
simplifying application implementation. And this means that the only sensible
use case for zonefs is applications using large container like files. LSM-tree
based applications being a very good match in this respect.

> Secondly, the allocation scheme is too simplified for my taste and it
> could create a significant fragmentation of a volume. Again, 256 MB is
> pretty big size. So, I assume that, mostly, it will be allocated only
> one zone at first for a created file. If file grows then it means that
> it will need to allocate the two contigous zones and to move the file's
> content. Finally, it sounds for me that it is possible to create a lot
> of holes and to achieve the volume state when it exists a lot of free
> space but files will be unable to grow and it will be impossible to add
> a new data on the volume. Have you made an estimation of the suggested
> allocation scheme?

What do you mean allocation scheme ? There is none ! one file == one zone and
all files are fully provisioned and allocated on mount. zonefs does not allow
the creation of files and there is no dynamic "block allocation". Again, please
do not consider zonefs as a normal file system. It is closer to a raw block
device interface than to a fully featured file system.

Best regards.

-- 
Damien Le Moal
Western Digital Research




[Index of Archives]     [XFS Filesystem Development (older mail)]     [Linux Filesystem Development]     [Linux Audio Users]     [Yosemite Trails]     [Linux Kernel]     [Linux RAID]     [Linux SCSI]


  Powered by Linux