On Wed, Nov 12, 2014 at 05:03:18PM -0500, Alan Stern wrote: > On Wed, 12 Nov 2014, Sean O. Stalley wrote: > > Our plan to support multiple MA devices is to have them all connected > > to the same virtual host controller, so only 1 would be needed. > > > > Would you prefer we have 1 host controller instance per MA device? > > We are definitely open to suggestions on how this should be architected. > > I haven't read the MA USB spec, so I don't know how it's intended to > work. Still, what happens if you create a virtual host controller > with, say, 16 ports, and then someone wants to connect a 17th MA > device? To summarize the spec: MA USB groups a host & connected devices into MA service sets (MSS). The architectural limit is 254 MA devices per MSS. If the host needs to connect more devices than that, It can start a new MSS and connect to 254 more MA devices. Is supporting up to 254 devices on one machine sufficient? Would it make sense (and does the usb stack support) having 254 root ports on one host controller? If so, we could make our host controller instance have 254 ports. I'm guessing the hub driver may have a problem with this (especially for superspeed). If that doesn't make sense (or isn't supported), we can have 1 host controller instance per MA device. Would that be preferred? > Also, I noticed that your patch adds a new bus type for these MA host > controllers. It really seems like overkill to have a whole new bus > type if there's only going to be one device on it. The bus was added when we were quickly trying to replace the platform device code. It's probably not the right thing to do. I'm still not sure why we can't make our hcd a platform device, especially since dummy_hcd & the usbip's hcd are both platform devices. > > If we get rid of these locks, endpoints can't run simultaneously. > > MA USB IN endpoints have to copy data, which could take a while. > > I don't know what you mean by "run simultaneously". Certainly multiple > network packets can be transmitted and received concurrently even if > you use a single spinlock, since your locking won't affect the > networking subsystem. I meant we couldn't have 2 threads in our driver. With one lock, One thread would always have to wait for the other, even though they could be working on 2 different endpoints doing completely independent tasks. > > Couldn't this cause a bottleneck? > > Probably not enough to matter. After all, the other host controller > drivers rely on a single spinlock. And if it did matter, you could > drop the spinlock while copying the data. Good point. We can cut our driver down to using 1 lock. If we find that only having 1 spinlock does cause a bottleneck, we can deal with it then. Thanks, Sean _______________________________________________ devel mailing list devel@xxxxxxxxxxxxxxxxxxxxxx http://driverdev.linuxdriverproject.org/mailman/listinfo/driverdev-devel