Tux3 Report: How fast can we fsync?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Greetings,

This post is dedicated to Ted, who raised doubts a while back about whether Tux3 can ever have a fast fsync:

  https://lkml.org/lkml/2013/5/11/128
  "Re: Tux3 Report: Faster than tmpfs, what?"

Ted suggested that Tux3's inherently asynchronous design would be a limitation when it comes to synchronous operations like fsync. As he put it, "any advantage of decoupling the front/back end is nullified" because of temporal coupling. We found the opposite to be true: our asynchronous design works as well for synchronous operations as it does for any other, and Tux3 now has a high performance fsync to prove it.

Until now, Tux3 handled fsync as a full delta commit equivalent to syncfs, just like Ext3. This is semantically correct but sometimes commits more data than necessary and creates a bottleneck by serializing all fsyncs. To optimize it, we added a mechanism for committing any subset of dirty inodes separately from a full delta.

Like our normal delta commit, the design is asynchronous: front end tasks produce fsync work and a backend task commits it. We now have two backends, one to commit fsyncs and another to commit deltas, serialized so the combination is single threaded and mostly lockless. Each fsync moves an inode's dirty blocks from the front delta to the back delta, then queues the inode for the backend. The backend spins away committing fsyncs, opportunistically batching them up into group commits. The fsync backend gives priority to the delta backend: whenever a full delta flush is needed because of cache pressure or cache age, it finishes its fsync work and gets out of the way.

This approach required only minimal changes to our existing delta commit mechanism, mainly to support crash consistency. In particular, our delta model did not need to be changed at all, and block forking protects fsync commits in the same way as normal delta commits. That is, an inode can be freely updated (even via mmap) while it is being fsynced. The fsync data is isolated from those changes and the frontend does not stall.

I measured fsync performance using a 7200 RPM disk as a virtual drive under KVM, configured with cache=none so that asynchronous writes are cached and synchronous writes translate into direct writes to the block device. To focus purely on fsync, I wrote a small utility (at the end of this post) that forks a number of tasks, each of which continuously appends to and fsyncs its own file. For a single task doing 1,000 fsyncs of 1K each, we have:

   Ext4:  34.34s
   XFS:   23.63s
   Btrfs: 34.84s
   Tux3:  17.24s

Tux3 has a nice advantage for isolated fsyncs. This is mainly due to writing a small number of blocks per commit, currently just five blocks for a small file. As for a normal delta commit, we do not update bitmap blocks or index nodes, but log logical changes instead, flushing out the primary metadata with occasional "unify" deltas to control the log size and keep replay fast. The 1,000 fsync test typically does about three unifies, so we do in fact update all our metadata and pay that cost, just not on every commit.

The win is bigger than it appears at first glance, because writing the commit block takes about 16.5 ms. That would be similar for all the filesystems tested, so factoring that out, we see that Tux3 must be doing 10 times less work than XFS, 24 times less than Ext4 and 25 times less than Btrfs. Over time, that gap should widen as we reduce our small file commit overhead towards just two blocks.

In this single threaded test, we pay the full price for "communication delays" with our asynchronous backend, which evidently does not amount to much. Given that a context switch is on the order of microseconds while writing the commit block is on the order of milliseconds, it is unsurprising that two extra context switches just disappear in the noise.

Things get more interesting with parallel fsyncs. In this test, each task does ten fsyncs and task count scales from ten to ten thousand. We see that all tested filesystems are able to combine fsyncs into group commits, with varying degrees of success:

   Tasks:   10      100    1,000    10,000
   Ext4:   0.79s   0.98s    4.62s    61.45s
   XFS:    0.75s   1.68s   20.97s   238.23s
   Btrfs   0.53s   0.78s    3.80s    84.34s
   Tux3:   0.27s   0.34s    1.00s     6.86s
             (lower is better)

We expect sub-linear scaling with respect to tasks as opportunities to combine commits increase, then linear scaling as total write traffic begins to dominate. Tux3 still shows sub-linear scaling even at 10,000 tasks. XFS scales poorly, and also suffers from read starvation at the high end, sometimes taking tens of seconds to cat a file or minutes to list a directory. Ext4 and Tux3 exhibit no such issues, remaining completely responsive under all tested loads. The bottom line for this test is, Tux3 is twice as fast at fsync with a modest task count, and the gap widens to nine times faster than its nearest competitor as task count increases.

Is there any practical use for fast parallel fsync of tens of thousands
of tasks? This could be useful for a scalable transaction server that sits directly on the filesystem instead of a database, as is the fashion for big data these days. It certainly can't hurt to know that if you need that kind of scaling, Tux3 will do it.

Of course, a pure fsync load could be viewed as somewhat unnatural. We
also need to know what happens under a realistic load with buffered
operations mixed with fsyncs. We turn to an old friend, dbench:

Dbench -t10

   Tasks:       8           16           32
   Ext4:    35.32 MB/s   34.08 MB/s   39.71 MB/s
   XFS:     32.12 MB/s   25.08 MB/s   30.12 MB/s
   Btrfs:   54.40 MB/s   75.09 MB/s  102.81 MB/s
   Tux3:    85.82 MB/s  133.69 MB/s  159.78 MB/s
                  (higher is better)

Tux3 and Btrfs scale well and are way ahead of Ext4 and XFS, while Ext4 and XFS scale poorly or even negatively. Tux3 is the leader by a wide
margin, beating XFS by more than a factor of 5 at the high end.
Dbench -t10 -s (all file operations synchronous)

   Tasks:       8           16           32
   Ext4:     4.51 MB/s    6.25 MB/s    7.72 MB/s
   XFS:      4.24 MB/s    4.77 MB/s    5.15 MB/s
   Btrfs:    7.98 MB/s   13.87 MB/s   22.87 MB/s
   Tux3:    15.41 MB/s   25.56 MB/s   39.15 MB/s
                  (higher is better)

With a pure synchronous load (O_SYNC) the ranking is not changed but the gaps widen, and Tux3 outperforms XFS by a factor of 7.5.

At the risk of overgeneralizing, a trend seems to be emerging: the new, write-anywhere designs run synchronous operations faster, combine synchronous and asynchronous operations more efficiently, and scale better to high task counts than the traditional journaling designs. If there is to be an Ext5, it might be worth considering the merit of abandoning the journal in favor of something along the lines of Tux3's redirect-on-write and logging combination.

Getting back to Ted's question, perhaps an asynchronous design really is a better idea all round, even for synchronous operations, and perhaps there really is such a thing as an improved design that is not just a different set of tradeoffs.

In the full disclosure department, Tux3 is still not properly optimized in some areas. One of them is fragmentation: it is not very hard to make Tux3 slow down by running long tests. Our current allocation algorithm is completely naive - it just allocates the next available block and wraps at the top of volume. After a few wraps, it makes a big mess. So today we are not claiming victory in the benchmark department, we still have some work to do. Today is just about fsync, for which it is fair to say that Tux3 sets a new performance standard.

Regards,

Daniel

Footnote: While I was doing this work, Hirofumi set about improving our full filesystem sync to support merging parallel full syncs into single delta commits. That one change improved fsync performance so much that I almost abandoned my separate fsync work. However, a true, separate fsync with aggressive group commit eventually proved superior, so now we have both: a high performance fsync and a full filesystem sync that is nearly as fast under many loads.

/*
* syncs.c
* * D.R. Phillips, 2015 * * To build: c99 -Wall syncs.c -o syncs
* To run: ./syncs [<filename> [<syncs> [<tasks>]]]
*/

#include <unistd.h>
#include <stdlib.h>
#include <stdio.h>
#include <fcntl.h>
#include <sys/wait.h>
#include <errno.h>
#include <sys/stat.h>

char text[1024] = { "hello world!\n" };

int main(int argc, const char *argv[]) {
	const char *basename = argc < 1 ? "foo" : argv[1];
	char name[100];
	int steps = argc < 3 ? 1 : atoi(argv[2]);
	int tasks = argc < 4 ? 1 : atoi(argv[3]);
	int err, fd;

	for (int t = 0; t < tasks; t++) {
		snprintf(name, sizeof name, "%s%i", basename, t);
		if (!fork())
			goto child;
	}
	for (int t = 0; t < tasks; t++)
		wait(&err);
	return 0;

child:
	fd = creat(name, S_IRWXU);
	for (int i = 0; i < steps; i++) {
		write(fd, text, sizeof text);
		fsync(fd);
	}
	return 0;
}

--
To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux Ext4 Filesystem]     [Union Filesystem]     [Filesystem Testing]     [Ceph Users]     [Ecryptfs]     [AutoFS]     [Kernel Newbies]     [Share Photos]     [Security]     [Netfilter]     [Bugtraq]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux Cachefs]     [Reiser Filesystem]     [Linux RAID]     [Samba]     [Device Mapper]     [CEPH Development]
  Powered by Linux