Re: [RFC] Netlink and user-space buffer pointers

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Note: We've transitioned off topic. If what this means is "there isn't a good
way except by ioctls (which still isn't easily portable) or system calls",
then that's ok. Then at least we know the limits and can look at other
implementation alternatives.

Mike Christie wrote:
James Smart wrote:
Mike Christie wrote:
For the tasks you want to do for the fc class is performance critical?
No, it should not be.

If not, you could do what the iscsi class (for the netdev people this is
drivers/scsi/scsi_transport_iscsi.c) does and just suffer a couple
copies. For iscsi we do this in userspace to send down a login pdu:

    /*
     * xmitbuf is a buffer that is large enough for the iscsi_event,
     * iscsi pdu (hdr_size) and iscsi pdu data (data_size)
     */
Well, the real difference is that the payload of the "message" is actually
the payload of the SCSI command or ELS/CT Request. Thus, the payload may

I am not sure I follow. For iscsi, everything after the iscsi_event
struct can be the iscsi request that is to be transmitted. The payload
will not normally be Mbytes but it is not a couple if bytes.

True... For a large read/write - it will eventually total what the i/o
request size was, and you did have to push it through the socekt.
What this discussion really comes down to is the difference between initiator
offload and what a target does.

The initiator offloads the "full" i/o from the users - e.g. send command,
get response. In the initiator case, the user isn't aware of each and
every IU that makes up the i/o. As it's on an i/o basis, the LLDD doing
the offload needs the full buffer sitting and ready. DMA is preferred so
the buffer doesn't have to be consuming socket/kernel/driver buffers while
it's pending - plus speed.

In the target case, the target controls each IU and it's size, thus it
only has to have access to as much buffer space as it wants to push the next
IU. The i/o can be "paced" by the target. Unfortunately, this is an entirely
different use model than users of a scsi initiator expect, and it won't map
well into replacing things like our sg_io ioctls.

Instead of netlink for scsi commands and transport requests....

For scsi commands could we just use sg io, or is there something special
about the command you want to send? If you can use sg io for scsi
commands, maybe for transport level requests (in my example iscsi pdu)
we could modify something like sg/bsg/block layer scsi_ioctl.c to send
down transport requests to the classes and encapsulate them in some new
struct transport_requests or use the existing struct request but do that
thing people keep taling about using the request/request_queue for
message passing.

Well - there's 2 parts to this answer:

First : IOCTL's are considered dangerous/bad practice and therefore it would
  be nice to find a replacement mechanism that eliminates them. If that
  mechanism has some of the cool features that netlink does, even better.
  Using sg io, in the manner you indicate, wouldn't remove the ioctl use.
  Note: I have OEMs/users that are very confused about the community's statement
  about ioctls. They've heard they are bad, should never be allowed, will no
  be longer supported, but yet they are at the heart of DM and sg io and other
  subsystems. Other than a "grandfathered" explanation, they don't understand
  why the rules bend for one piece of code but not for another. To them, all
  the features are just as critical regardless of whose providing them.

Second: transport level i/o could be done like you suggest, and we've
  prototyped some of this as well. However, there's something very wrong
  about putting "block device" wrappers and settings around something that
  is not a block device.  In general, it's a heck of a lot of overhead and
  still doesn't solve the real issue - how to portably pass that user buffer
  in to/out of the kernel.


-- james s
-
: send the line "unsubscribe linux-scsi" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[Index of Archives]     [SCSI Target Devel]     [Linux SCSI Target Infrastructure]     [Kernel Newbies]     [IDE]     [Security]     [Git]     [Netfilter]     [Bugtraq]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Linux ATA RAID]     [Linux IIO]     [Samba]     [Device Mapper]
  Powered by Linux