On 2020/2/6 0:38, Jeff Layton wrote:
On Wed, 2020-02-05 at 11:24 -0500, Jeff Layton wrote:
On Mon, 2020-02-03 at 20:54 -0500, xiubli@xxxxxxxxxx wrote:
From: Xiubo Li <xiubli@xxxxxxxxxx>
If the direct io couldn't be submit in a single request, for multiple
writers, they may overlap each other.
For example, with the file layout:
ceph.file.layout="stripe_unit=4194304 stripe_count=1 object_size=4194304
fd = open(, O_DIRECT | O_WRONLY, );
Writer1:
posix_memalign(&buffer, 4194304, SIZE);
memset(buffer, 'T', SIZE);
write(fd, buffer, SIZE);
Writer2:
posix_memalign(&buffer, 4194304, SIZE);
memset(buffer, 'X', SIZE);
write(fd, buffer, SIZE);
From the test result, the data in the file possiblly will be:
TTT...TTT <---> object1
XXX...XXX <---> object2
The expected result should be all "XX.." or "TT.." in both object1
and object2.
I really don't see this as broken. If you're using O_DIRECT, I don't
believe there is any expectation that the write operations (or even read
operations) will be atomic wrt to one another.
Basically, when you do this, you're saying "I know what I'm doing", and
need to provide synchronization yourself between competing applications
and clients (typically via file locking).
In fact, here's a discussion about this from a couple of years ago. This
one mostly applies to local filesystems, but the same concepts apply to
CephFS as well:
https://www.spinics.net/lists/linux-fsdevel/msg118115.html
Okay. So for the O_DIRECT write/read, we won't guarantee it will be
atomic in fs layer.
Thanks,