Re: [PATCH rdma-next 0/3] Support out of order data placement

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 7/21/2017 9:50 PM, Parav Pandit wrote:
Hi Tom,

-----Original Message-----
From: Tom Talpey [mailto:tom@xxxxxxxxxx]
Sent: Friday, July 21, 2017 9:29 PM
To: Parav Pandit <parav@xxxxxxxxxxxx>; Jason Gunthorpe
<jgunthorpe@xxxxxxxxxxxxxxxxxxxx>
Cc: Bart Van Assche <Bart.VanAssche@xxxxxxxxxxx>; leon@xxxxxxxxxx;
dledford@xxxxxxxxxx; linux-rdma@xxxxxxxxxxxxxxx; Idan Burstein
<idanb@xxxxxxxxxxxx>
Subject: Re: [PATCH rdma-next 0/3] Support out of order data placement

On 7/18/2017 10:33 PM, Parav Pandit wrote:
Hi Tom, Jason,

Sorry for the late response.
Please find the response inline below.

-----Original Message-----
From: Tom Talpey [mailto:tom@xxxxxxxxxx]
Sent: Monday, June 12, 2017 8:30 PM
To: Parav Pandit <parav@xxxxxxxxxxxx>; Jason Gunthorpe
<jgunthorpe@xxxxxxxxxxxxxxxxxxxx>
Cc: Bart Van Assche <Bart.VanAssche@xxxxxxxxxxx>; leon@xxxxxxxxxx;
dledford@xxxxxxxxxx; linux-rdma@xxxxxxxxxxxxxxx; Idan Burstein
<idanb@xxxxxxxxxxxx>
Subject: Re: [PATCH rdma-next 0/3] Support out of order data
placement


In IB spec, in-order delivery is default.

I don't agree. Requests are sent in-order, and the responder
processes them in- order, but the bytes thenselves are not guaranteed to
appear in-order.
Additionally, if retries occur, this is most definitely not the case.

Section 9.5 Transaction Ordering, I believe, covers these
requirements. Can you tell me where I misunderstand them?
In fact, c9-28 explicitly warns:

     • An application shall not depend upon the order of data writes to
     memory within a message. For example, if an application sets up
     data buffers that overlap, for separate data segments within a
     message, it is not guaranteed that the last sent data will always
     overwrite the earlier.

The IB spec indeed does not imply any ordering in the placement of data into
memory within a single message.

It does guarantee that writes don't bypass writes and reads don't bypass reads
(Table 76), and transport operations are executed in their *message* order (C9-
28):
"A responder shall execute SEND requests, RDMA WRITE requests and
ATOMIC Operation requests in the message order in which they are
received."

Thus, ordering between messages is guaranteed - changes to remote memory
of an RDMA-W will be observed strictly after any changes done by a previous
RDMA-W; changes to local memory of an RDMA-R response will be observed
strictly after any changes done by a previous RDMA-R response.

The proposed feature in this patch set is to relax the memory placement
ordering *across* messages and not within a single message (which is not
mandated by the spec as u noted), such that multiple consecutive RDMA-Ws
may be committed to memory in any order, and similarly for RDMA-R responses.
This changes application semantics whenever multiple-inflight RDMA
operations write to overlapping locations, or when one operation indicates the
completion of the other.
A simple example to clarify: a requestor posted the following work elements in
the written order:
1. RDMA-W(VA=0x1000, value=0x1)
2. RDMA-W(VA=0x1000, value=0x2)
3. Send()
On responder side, following the Send() operation completion, and according
to spec (C9-28), reading from VA=0x1000 will produce the value 2. With the
proposed feature enabled, the read value is not deterministic and dependent on
the order in which the RDMA-W operations were received.

The proposed QP flag allows applications to knowingly indicate this relaxed
data placement, thereby enabling the HCA to place OOO RDMA messages into
memory without buffering them.

You didn't answer my question what is the actual benefit of relaxing the
ordering. Is it performance?

Yes. Performance is better.

And, specifically what applications *can't* use it?
Applications which poll on RDMA-W data at responder side or RDMA-R data on Read requester side, cannot use this.
Because as explained in above example 2nd RDMA-W message can be executed first at responder side.
We cannot break such user space applications deployed in field by enabling this by default and without negotiation with peer.

Those applications ignored the spec, and got away with it only
because the Mellanox (is that who "we" is?) implementation was
strongly ordered. Thats not much of an excuse, in my opinion, to
force change on the well-behaved, spec-observing ULPs in order
that they might take advantage of it.

To me, it appears that most storage upper layers can already use the extension.

Yes. As they don't poll on data and they depend on incoming RDMA Send they can make use of it.

Not without changing their protocols and implementations. I think
you should reconsider your approach to throw the responsibility to
them, and them only.

If it performs better, I expect they will definitely want to enable it. In that case I
believe it should be the *default*, not an opt-in that these upper layers are
newly responsible for.

Verb layer is unware of caller ULPs. At most it knows that its kernel vs user ULP easily - which is good enough.
Verb layer also doesn't know whether remote side support it or not.
Once rdmacm extension is done, all kernel ULPs which uses rdmacm - can be enabled by default.

Well, then this change should wait for that to become available.

This patchset enables user space applications to take immediate benefit of it which doesn't depend on rdmacm.

But it changes the API in a way that we don't want to survive.
Let's get the interface right first.

I have one other question on the Documentation out-of-order.txt.
It states the fence bit can be used to force ordering on a non-strict
connection.
But fence doesn't apply to RDMA Write?
It only applies to operations which produce a reply, such as RDMA
Read or Atomic. Have you changed the semantic?

RDMA-R followed by RDMA-R semantic is changed when proposed QP flag is
set.

Can you explain that statement in more detail please? Also, please clarify on
what operation(s) the fence bit now applies.
Sure.
Let's take example.
A requestor posted the following work elements in the written order:
1. RDMA-R(VA=0x1000, len=0x400)
2. RDMA-R(VA=0x1400, len=0x4)
Currently as per Table-76, RDMA-R read response data of 1st RDMA-R is placed first.
With relaxed data placement attribute, 4 bytes data of 2nd RDMA-R can be placed first.
If user needs a ordering of current Table-76, it needs to set fence on 2nd RDMA-R.
This will ensure that 1st RDMA-R is executed before 2nd RDMA-R.

Oh, that's the same issue a the initial one - polling the "last"
bit was never guaranteed. I dont see this as a change to the
semantic.

But, I take it that the fence bit still applies as before, this is
not a proposal to extended fencing to RDMA Write. Ok.



This translates to In Table 76,
RDMA-R (Row) and RDMA-R(Column) changes from '#' to 'F'.

--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Photo]     [Yosemite News]     [Yosemite Photos]     [Linux Kernel]     [Linux SCSI]     [XFree86]
  Powered by Linux