On Thu, Feb 9, 2012 at 1:39 PM, J. Bruce Fields <bfields@xxxxxxxxxxxx> wrote: > On Thu, Feb 09, 2012 at 09:37:22AM +0100, Tigran Mkrtchyan wrote: >> Putting my 'high energy physics community' hat on let me comment on it. >> >> As soon as we trying to use nfs over high latency networks the > > How high is the latency? > >> application efficiency rapidly drops. Efficiency is wallTIme/cpuTime. >> We have solve this in our home grown protocols by adding vector read >> and vector write. Vector is set of offset_length. As most of the our >> files has DB like structure, after reading header (something like >> index) we knew where data is located. This allows us to perform in >> some work loads 100 times better than NFS. >> >> Posix does not provides such interface. But we can simulate that with >> fadvise calls (and we do). Since nfs-4.0 we got compound operations. >> And you can (in theory) build a compound with multiple READ or WRITE >> ops. Nevertheless this does not work for several reasons: maximal >> reply size and you still have to wait for full reply. and some reply >> may be up 100MB in size. >> >> The solution here is to issue multiple requests in parallel. And this >> is possible only if you have enough session slots. Server can reply >> out of order and populate clients file system cache. > > Yep. > > I'm just curious whether Andy or someone's beeing doing experiments with > these patches, and if so, what they look like. > > (The numbers I can find from the one case we worked on at citi (UM to > CERN), were 10 gig * 120ms latency for a 143MB bandwidth-delay product, > so in theory 143 slots would suffice if they were all doing maximum-size The net I tested with has a 127ms delay (~152MB bandwitdth-delay product) - the same ballpark as 120ms. As you mention, you are assuming an rsize/wsize of 1 MB. Think of a server that only supports 64k and you need a lot more slots (143 * 16 = 2288 slots) to fill the 143MB bandwidth-delay product. A 64K r/wsize server with 255 slots on a 10G net could only fill a 13ms latency 10G net. Plus - 10GB nets are old tech! The CERN/UMICH machines I was working with had 40GB NICs. 100GB nets are on the way... -->Andy > IO--but I can't find any results from NFS tests over that link (only > results for a lower-latency 10gig network). > > And, of course, if you're doing smaller operations or have an even > higher-latency network, etc., you could need more slots--I just wondered > abuot the details.) > > --b. > -- > To unsubscribe from this list: send the line "unsubscribe linux-nfs" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe linux-nfs" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html