Re: libgfapi zero copy write - application in samba, nfs-ganesha

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Poornima,

I think that the goal would be to add support for SMB specific copy offload command in Samba. NFS also has a protocol specific one. This is a matter of protocol support.

On the Samba server side (or NFS Ganesha side), we would get a client copy offload request that we could handle without shipping data over the wire to the client.

Under Samba or NFS Ganesha on the server side, using copy offload for the local XFS file system is a different problem I think.

Regards,
Ric


On 09/27/2016 11:02 AM, Poornima Gurusiddaiah wrote:
W.r.t Samba consuming this, it requires a great deal of code change in Samba.
Currently samba has no concept of getting buf from the underlying file system,
the filesystem comes into picture only at the last layer(gluster plugin),
where system calls are replaced by libgfapi calls. Hence, this is not readily
consumable by Samba, and i think same will be the case with NFS_Ganesha, will
let the Ganesha folksc comment on the same.


Regards,
Poornima

----- Original Message -----
From: "Ric Wheeler" <ricwheeler@xxxxxxxxx>
To: "Raghavendra Gowdappa" <rgowdapp@xxxxxxxxxx>, "Ric Wheeler" <rwheeler@xxxxxxxxxx>
Cc: "Ben England" <bengland@xxxxxxxxxx>, "Gluster Devel" <gluster-devel@xxxxxxxxxxx>, "Ben Turner"
<bturner@xxxxxxxxxx>
Sent: Tuesday, September 27, 2016 2:25:40 AM
Subject: Re:  libgfapi zero copy write - application in samba, nfs-ganesha

On 09/27/2016 08:53 AM, Raghavendra Gowdappa wrote:
----- Original Message -----
From: "Ric Wheeler" <rwheeler@xxxxxxxxxx>
To: "Raghavendra Gowdappa" <rgowdapp@xxxxxxxxxx>, "Saravanakumar Arumugam"
<sarumuga@xxxxxxxxxx>
Cc: "Gluster Devel" <gluster-devel@xxxxxxxxxxx>, "Ben Turner"
<bturner@xxxxxxxxxx>, "Ben England"
<bengland@xxxxxxxxxx>
Sent: Tuesday, September 27, 2016 10:51:48 AM
Subject: Re:  libgfapi zero copy write - application in
samba, nfs-ganesha

On 09/27/2016 07:56 AM, Raghavendra Gowdappa wrote:
+Manoj, +Ben turner, +Ben England.

@Perf-team,

Do you think the gains are significant enough, so that smb and
nfs-ganesha
team can start thinking about consuming this change?

regards,
Raghavendra
This is a large gain but I think that we might see even larger gains (a
lot
depends on how we implement copy offload :)).
Can you elaborate on what you mean "copy offload"? If it is the way we
avoid a copy in gfapi (from application buffer), following is the
workflow:

<commit>

Work flow of zero copy write operation:
--------------------------------------

1) Application requests a buffer of specific size. A new buffer is
allocated from iobuf pool, and this buffer is passed on to application.
     Achieved using "glfs_get_buffer"

2) Application writes into the received buffer, and passes that to
libgfapi, and libgfapi in turn passes the same buffer to underlying
translators. This avoids a memcpy in glfs write
     Achieved using "glfs_zero_write"

3) Once the write operation is complete, Application must take the
responsibilty of freeing the buffer.
     Achieved using "glfs_free_buffer"

</commit>

Do you've any suggestions/improvements on this? I think Shyam mentioned an
alternative approach (for zero-copy readv I think), let me look up at that
too.

regards,
Raghavendra
Both NFS and SMB support a copy offload that allows a client to produce a new
copy of a file without bringing data over the wire. Both, if I remember
correctly, do a ranged copy within a file.

Yup, also referred to as Server side copy, Niels is working on having this for Gluster.

The key here is that since the data does not move over the wire from server
to
client, we can shift the performance bottleneck to the storage server.

If we have a slow (1GB) link between client and server, we should be able to
do
that copy as if it happened just on the server itself. For a single NFS
server
(not a clustered, scale out server), that usually means we are as fast as the
local file system copy.

Note that there are also servers that simply "reflink" that file, so we have
a
very small amount of time needed on the server to produce that copy.  This
can
be a huge win for say a copy of a virtual machine guest image.

Gluster and other distributed servers won't benefit as much as a local server
would I suspect because of the need to do things internally over our networks
between storage server nodes.

Hope that makes my thoughts clearer?

Here is a link to a brief overview of the new Linux system call:

https://kernelnewbies.org/Linux_4.5#head-6df3d298d8e0afa8e85e1125cc54d5f13b9a0d8c

Note that block devices or pseudo devices can also implement a copy offload.

Regards,

Ric

Worth looking at how we can make use of it.

thanks!

Ric

----- Original Message -----
From: "Saravanakumar Arumugam" <sarumuga@xxxxxxxxxx>
To: "Gluster Devel" <gluster-devel@xxxxxxxxxxx>
Sent: Monday, September 26, 2016 7:18:26 PM
Subject:  libgfapi zero copy write - application in
samba,
	nfs-ganesha

Hi,

I have carried out "basic" performance measurement with zero copy write
APIs.
Throughput of zero copy write is 57 MB/sec vs default write 43 MB/sec.
( I have modified Ben England's gfapi_perf_test.c for this. Attached the
same
for reference )

We would like to hear how samba/ nfs-ganesha who are libgfapi users can
make
use of this.
Please provide your comments. Refer attached results.

Zero copy in write patch: http://review.gluster.org/#/c/14784/

Thanks,
Saravana
_______________________________________________
Gluster-devel mailing list
Gluster-devel@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-devel

_______________________________________________
Gluster-devel mailing list
Gluster-devel@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-devel

_______________________________________________
Gluster-devel mailing list
Gluster-devel@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-devel



_______________________________________________
Gluster-devel mailing list
Gluster-devel@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-devel



[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux