On Tue, Aug 29, 2017 at 12:01 AM, Florian Haas <florian@xxxxxxxxxxx> wrote: > Sorry, I worded my questions poorly in the last email, so I'm asking > for clarification here: > > On Mon, Aug 28, 2017 at 6:04 PM, Haomai Wang <haomai@xxxxxxxx> wrote: >> On Mon, Aug 28, 2017 at 7:54 AM, Florian Haas <florian@xxxxxxxxxxx> wrote: >>> On Mon, Aug 28, 2017 at 4:21 PM, Haomai Wang <haomai@xxxxxxxx> wrote: >>>> On Wed, Aug 23, 2017 at 1:26 AM, Florian Haas <florian@xxxxxxxxxxx> wrote: >>>>> Hello everyone, >>>>> >>>>> I'm trying to get a handle on the current state of the async messenger's >>>>> RDMA transport in Luminous, and I've noticed that the information >>>>> available is a little bit sparse (I've found >>>>> https://community.mellanox.com/docs/DOC-2693 and >>>>> https://community.mellanox.com/docs/DOC-2721, which are a great start >>>>> but don't look very complete). So I'm kicking off this thread that might >>>>> hopefully bring interested parties and developers together. >>>>> >>>>> Could someone in the know please confirm that the following assumptions >>>>> of mine are accurate: >>>>> >>>>> - RDMA support for the async messenger is available in Luminous. >>>> >>>> to be precious, rdma in luminous is available but lack of memory >>>> control when under pressure. it would be ok to run for test purpose. >>> >>> OK, thanks! Assuming async+rdma will become fully supported some time >>> in the next release or two, are there plans to backport async+rdma >>> related features to Luminous? Or will users likely need to wait for >>> the next release to get a production-grade Ceph/RDMA stack? >> >> I think so > > OK, so just to clarify: > > (1) production RDMA support *will* be in the next LTS. Correct? > > (2) Users should *not* expect production RDMA support in any Luminous > point release. Correct? from my view, we still contribute to a stable rdma impl. I can't make a promise about this. > >>>>> - You enable it globally by setting ms_type to "async+rdma", and by >>>>> setting appropriate values for the various ms_async_rdma* options (most >>>>> importantly, ms_async_rdma_device_name). >>>>> >>>>> - You can also set RDMA messaging just for the public or cluster >>>>> network, via ms_public_type and ms_cluster_type. >>>>> >>>>> - Users have to make a global async+rdma vs. async+posix decision on >>>>> either network. For example, if either ms_type or ms_public_type is >>>>> configured to async+rdma on cluster nodes, then a client configured with >>>>> ms_type = async+posix can't communicate. >>>>> >>>>> Based on those assumptions, I have the following questions: >>>>> >>>>> - What is the current state of RDMA support in kernel libceph? In other >>>>> words, is there currently a way to map RBDs, or mount CephFS, if a Ceph >>>>> cluster uses RDMA messaging? >>>> >>>> no planning on kernel side so far. rbd-nbd, cephfs-fuse should be supported now. >>> >>> Understood — are there plans to support async+rdma in the kernel at >>> all, or is there something in the kernel that precludes this? >> >> no. > > Do you mean: > > (1) There is a plan to support async+rdma in the kernel client eventually, or > > (2) There are no plans to bring async+rdma support to the kernel > client, even though it *would* be possible to implement from a kernel > perspective, or > > (3) There are no plans to bring async+rdma support to the kernel > client *because* something deep in the kernel prevents it? yes, async is totoaly a userspace impl. kernel client need to make use of existing kernel infra > > Thanks again! > > Cheers, > Florian > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com