Re: [PATCH v2 1/8] nvme-rdma: use ib_client API to wrap ib_device

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Jun 5, 2018 at 10:24 AM, Max Gurtovoy <maxg@xxxxxxxxxxxx> wrote:
>
>
> On 6/5/2018 12:38 AM, Jason Gunthorpe wrote:
>>
>> On Mon, Jun 04, 2018 at 02:29:56PM +0200, Roman Pen wrote:
>>>
>>> ib_client API provides a way to wrap an ib_device with a specific ULP
>>> structure.  Using that API local lists and mutexes can be completely
>>> avoided and allocation/removal paths become a bit cleaner.
>>>
>>> Signed-off-by: Roman Pen <roman.penyaev@xxxxxxxxxxxxxxx>
>>> Cc: Christoph Hellwig <hch@xxxxxx>
>>> Cc: Steve Wise <swise@xxxxxxxxxxxxxxxxxxxxx>
>>> Cc: Bart Van Assche <bart.vanassche@xxxxxxxxxxx>
>>> Cc: Sagi Grimberg <sagi@xxxxxxxxxxx>
>>> Cc: Doug Ledford <dledford@xxxxxxxxxx>
>>> Cc: linux-nvme@xxxxxxxxxxxxxxxxxxx
>>>   drivers/nvme/host/rdma.c | 82
>>> ++++++++++++++++++++++--------------------------
>>>   1 file changed, 38 insertions(+), 44 deletions(-)
>>>
>>> diff --git a/drivers/nvme/host/rdma.c b/drivers/nvme/host/rdma.c
>>> index 1eb4438a8763..dd79250c9df4 100644
>>> +++ b/drivers/nvme/host/rdma.c
>>> @@ -46,7 +46,6 @@ struct nvme_rdma_device {
>>>         struct ib_device        *dev;
>>>         struct ib_pd            *pd;
>>>         struct kref             ref;
>>> -       struct list_head        entry;
>>>   };
>>>     struct nvme_rdma_qe {
>>> @@ -124,9 +123,7 @@ static inline struct nvme_rdma_ctrl
>>> *to_rdma_ctrl(struct nvme_ctrl *ctrl)
>>>         return container_of(ctrl, struct nvme_rdma_ctrl, ctrl);
>>>   }
>>>   -static LIST_HEAD(device_list);
>>> -static DEFINE_MUTEX(device_list_mutex);
>>> -
>>> +static struct ib_client nvme_rdma_ib_client;
>>>   static LIST_HEAD(nvme_rdma_ctrl_list);
>>>   static DEFINE_MUTEX(nvme_rdma_ctrl_mutex);
>>>   @@ -325,17 +322,14 @@ static void nvme_rdma_free_dev(struct kref *ref)
>>>         struct nvme_rdma_device *ndev =
>>>                 container_of(ref, struct nvme_rdma_device, ref);
>>>   -     mutex_lock(&device_list_mutex);
>>> -       list_del(&ndev->entry);
>>> -       mutex_unlock(&device_list_mutex);
>>> -
>>> +       ib_set_client_data(ndev->dev, &nvme_rdma_ib_client, NULL);
>>>         ib_dealloc_pd(ndev->pd);
>>>         kfree(ndev);
>>>   }
>>>   -static void nvme_rdma_dev_put(struct nvme_rdma_device *dev)
>>> +static int nvme_rdma_dev_put(struct nvme_rdma_device *dev)
>>>   {
>>> -       kref_put(&dev->ref, nvme_rdma_free_dev);
>>> +       return kref_put(&dev->ref, nvme_rdma_free_dev);
>>>   }
>>>     static int nvme_rdma_dev_get(struct nvme_rdma_device *dev)
>>> @@ -348,43 +342,42 @@ nvme_rdma_find_get_device(struct rdma_cm_id *cm_id)
>>>   {
>>>         struct nvme_rdma_device *ndev;
>>>   -     mutex_lock(&device_list_mutex);
>>> -       list_for_each_entry(ndev, &device_list, entry) {
>>> -               if (ndev->dev->node_guid == cm_id->device->node_guid &&
>>> -                   nvme_rdma_dev_get(ndev))
>>> -                       goto out_unlock;
>>> +       ndev = ib_get_client_data(cm_id->device, &nvme_rdma_ib_client);
>>> +       if (ndev && WARN_ON(!nvme_rdma_dev_get(ndev)))
>>> +               ndev = NULL;
>>
>>
>> I think this is a much better idea to use the client data than
>> maintaining an internal list - this is what client data was for..
>>
>> But I wonder if the allocation of the client data should be deferred
>> until the ulp actually needs to use the device?
>
>
> I agree with Jason here. You will create ndev for every ib_device and this
> is redundant.

It seems this is a common pattern: wrap all ib devices in *add_one()
and use them calling ib_get_client_data().  But of course ulp dev
can be created on demand, when it is not found.  I sent a chunk of
the code in reply to Jason's email, please take a look.

>
> Also, any reason you use likely/unlikely statements not in the fast path ?

IMO sometimes it is easier to read the code, e.g. error branches better
distinguishable.

--
Roman
--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Photo]     [Yosemite News]     [Yosemite Photos]     [Linux Kernel]     [Linux SCSI]     [XFree86]

  Powered by Linux