Re: [Resend] Cross Memory Attach v3 [PATCH]

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, 24 Mar 2011 09:20:41 +1030
Rusty Russell <rusty@xxxxxxxxxxxxxxx> wrote:

> On Wed, 23 Mar 2011 12:52:13 +1030, Christopher Yeoh
> <cyeoh@xxxxxxxxxxx> wrote:
> > On Sun, 20 Mar 2011 18:55:32 -0700
> > Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> wrote:
> > > On Mon, 21 Mar 2011 12:20:18 +1030 Christopher Yeoh
> > > <cyeoh@xxxxxxxxxxx> wrote:
> > > 
> > > > On Thu, 17 Mar 2011 12:54:27 -0700
> > > > Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> wrote:
> > > > > On Thu, 17 Mar 2011 15:40:26 +1030
> > > > > Christopher Yeoh <cyeoh@xxxxxxxxxxx> wrote:
> > > > > 
> > > > > > > Thinking out loud: if we had a way in which a process can
> > > > > > > add and remove a local anonymous page into pagecache then
> > > > > > > other processes could access that page via mmap.  If both
> > > > > > > processes map the file with a nonlinear vma they they can
> > > > > > > happily sit there flipping pages into and out of the
> > > > > > > shared mmap at arbitrary file offsets. The details might
> > > > > > > get hairy ;) We wouldn't want all the regular mmap
> > > > > > > semantics of
> > > > > > 
> > > > > > Yea, its the complexity of trying to do it that way that
> > > > > > eventually lead me to implementing it via a syscall and
> > > > > > get_user_pages instead, trying to keep things as simple as
> > > > > > possible.
> > > > > 
> > > > > The pagecache trick potentially gives zero-copy access,
> > > > > whereas the proposed code is single-copy.  Although the
> > > > > expected benefits of that may not be so great due to TLB
> > > > > manipulation overheads.
> > > > > 
> > > > > I worry that one day someone will come along and implement the
> > > > > pagecache trick, then we're stuck with obsolete code which we
> > > > > have to maintain for ever.
> 
> Since this is for MPI (ie. message passing), they really want copy
> semantics.  If they didn't want copy semantics, they could just
> MAP_SHARED some memory and away they go...
> 
> You don't want to implement copy semantics with page-flipping; you
> would need to COW the outgoing pages, so you end up copying *and*
> trapping.
> 
> If you are allowed to replace "sent" pages with zeroed ones or
> something then you don't have to COW.  Yet even if your messages were
> a few MB, it's still not clear you'd win; in a NUMA world you're
> better off copying into a local page and then working on it.
> 
> Copying just isn't that bad when it's cache-hot on the sender and you
> are about to use it on the receiver, as MPI tends to be.  And it's
> damn simple.
> 
> But we should be able to benchmark an approximation to the
> page-flipping approach anyway, by not copying the data and doing the
> appropriate tlb flushes in the system call.

I've done some hacking on the naturally ordered and randomly ordered
ring bandwidth tests of hpcc to try to simulate what we'd get with a
page flipping approach.

- Modified hpcc so it checksums the data on the receiver. normally it
  just checks the data in a couple of places but the checksum simulates
  the receiver actually using all of the data

- For the page flipping scenario
  - allocate from a shared memory pool for data that is to be
    transferred
  - instead of sending the data via OpenMPI send some control data
    instead which describes where the receiver can read the data in
    shared memory. Thus "zero copy" with just checksum
  - Adds tlb flushing for sender/receiver processes

The results are below (numbers are in MB/s, higher the better). Base
is double copy via shared memory, CMA is single copy.

	                     Num MPI Processes			
Naturally Ordered	4	8	16	32
Base	               1152	929	567	370
CMA	               3682	3071	2753	2548
Zero Copy	       4634	4039	3149	2852
				
                      	Num MPI Processes
Randomly Ordered       	4	8	16	32
Base	                1154	927	588	389
CMA	                3632	3060	2897	2904
Zero Copy	        4668	3970	3077	2962

the benchmarks were run on a 32 way (SMT-off) Power6 machine.

So we can see that on lower numbers of processes there is a gain in
performance between single and zero copy (though the big jump is between
double and single copy), but this reduces as the number of processes
increases. The difference between the single and zero copy
approach reduces to almost nothing for when the number of MPI processes
is equal to the number of processors (for the randomly ordered ring
bandwidth).

Chris
-- 
cyeoh@xxxxxxxxxx

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@xxxxxxxxxx  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Fight unfair telecom internet charges in Canada: sign http://stopthemeter.ca/
Don't email: <a href=mailto:"dont@xxxxxxxxx";> email@xxxxxxxxx </a>


[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [ECOS]     [Asterisk Internet PBX]     [Linux API]