On 3/3/22 05:16, Greg KH wrote: > On Wed, Mar 02, 2022 at 02:27:05PM -0800, Iouri Tarassov wrote: >> On 3/2/2022 12:20 PM, Greg KH wrote: >>> On Wed, Mar 02, 2022 at 10:49:15AM -0800, Iouri Tarassov wrote: >>>> On 3/2/2022 3:53 AM, Wei Liu wrote: >>>>> On Wed, Mar 02, 2022 at 08:53:15AM +0100, Greg KH wrote: >>>>>> On Tue, Mar 01, 2022 at 10:23:21PM +0000, Wei Liu wrote: >>>>>>>>> +struct dxgglobal *dxgglobal; >>>>>>>> No, make this per-device, NEVER have a single device for your driver. >>>>>>>> The Linux driver model makes it harder to do it this way than to do it >>>>>>>> correctly. Do it correctly please and have no global structures like >>>>>>>> this. >>>>>>>> >>>>>>> This may not be as big an issue as you thought. The device discovery is >>>>>>> still done via the normal VMBus probing routine. For all intents and >>>>>>> purposes the dxgglobal structure can be broken down into per device >>>>>>> fields and a global structure which contains the protocol versioning >>>>>>> information -- my understanding is there will always be a global >>>>>>> structure to hold information related to the backend, regardless of how >>>>>>> many devices there are. >>>>>> Then that is wrong and needs to be fixed. Drivers should almost never >>>>>> have any global data, that is not how Linux drivers work. What happens >>>>>> when you get a second device in your system for this? Major rework >>>>>> would have to happen and the code will break. Handle that all now as it >>>>>> takes less work to make this per-device than it does to have a global >>>>>> variable. >>>>>> >>>>> It is perhaps easier to draw parallel from an existing driver. I feel >>>>> like we're talking past each other. >>>>> >>>>> Let's look at drivers/iommu/intel/iommu.c. There are a bunch of lists >>>>> like `static LIST_HEAD(dmar_rmrr_units)`. During the probing phase, new >>>>> units will be added to the list. I this the current code is following >>>>> this model. dxgglobal fulfills the role of a list. >>>>> >>>>> Setting aside the question of whether it makes sense to keep a copy of >>>>> the per-VM state in each device instance, I can see the code be changed >>>>> to: >>>>> >>>>> struct mutex device_mutex; /* split out from dxgglobal */ >>>>> static LIST_HEAD(dxglist); >>>>> >>>>> /* Rename struct dxgglobal to struct dxgstate */ >>>>> struct dxgstate { >>>>> struct list_head dxglist; /* link for dxglist */ >>>>> /* ... original fields sans device_mutex */ >>>>> } >>>>> >>>>> /* >>>>> * Provide a bunch of helpers manipulate the list. Called in probe / >>>>> * remove etc. >>>>> */ >>>>> struct dxgstate *find_dxgstate(...); >>>>> void remove_dxgstate(...); >>>>> int add_dxgstate(...); >>>>> >>>>> This model is well understood and used in tree. It is just that it >>>>> doesn't provide much value in doing this now since the list will only >>>>> contain one element. I hope that you're not saying we cannot even use a >>>>> per-module pointer to quickly get the data structure we want to use, >>>>> right? >>>>> >>>>> Are you suggesting Iouri use dev_set_drvdata to stash the dxgstate >>>>> into the device object? I think that can be done too. >>>>> >>>>> The code can be changed as: >>>>> >>>>> /* Rename struct dxgglobal to dxgstate and remove unneeded fields */ >>>>> struct dxgstate { ... }; >>>>> >>>>> static int dxg_probe_vmbus(...) { >>>>> >>>>> /* probe successfully */ >>>>> >>>>> struct dxgstate *state = kmalloc(...); >>>>> /* Fill in dxgstate with information from backend */ >>>>> >>>>> /* hdev->dev is the device object from the core driver framework */ >>>>> dev_set_drvdata(&hdev->dev, state); >>>>> } >>>>> >>>>> static int dxg_remove_vmbus(...) { >>>>> /* Normal stuff here ...*/ >>>>> >>>>> struct dxgstate *state = dev_get_drvdata(...); >>>>> dev_set_drvdata(..., NULL); >>>>> kfree(state); >>>>> } >>>>> >>>>> /* In all other functions */ >>>>> void do_things(...) { >>>>> struct dxgstate *state = dev_get_drvdata(...); >>>>> >>>>> /* Use state in place of where dxgglobal was needed */ >>>>> >>>>> } >>>>> >>>>> Iouri, notice this doesn't change anything regarding how userspace is >>>>> designed. This is about how kernel organises its data. >>>>> >>>>> I hope what I wrote above can bring our understanding closer. >>>>> >>>>> Thanks, >>>>> Wei. >>>> >>>> I can certainly remove dxgglobal and keep the pointer to the global >>>> state in the device object. >>>> >>>> This will require passing of the global pointer to all functions, which >>>> need to access it. >>>> >>>> >>>> Maybe my understanding of the Greg's suggestion was not correct. I >>>> thought the suggestion was >>>> >>>> to have multiple /dev/dxgN devices (one per virtual compute device). >>> You have one device per HV device, as the bus already provides you. >>> That's all you really need, right? Who would be opening the same device >>> node multiple times? >>>> This would change how the user mode >>>> clients enumerate and communicate with compute devices. >>> What does userspace have to do here? It should just open the device >>> node that is present when needed. How will there be multiple userspace >>> clients for a single HV device? >> >> Dxgkrnl creates a single user mode visible device node /dev/dxg. > When you do that, you have a device to put all of your data on. Use > that. > >> It has >> nothing to do with a specific hardware compute device on the host. Its >> purpose is to provide services (IOCTLs) to enumerate and manage virtual >> compute devices, which represent hardware devices on the host. The VMBus >> devices are not used directly by user mode clients in the current design. > That's horrid, why not just export the virtual compute devices properly > through individual device nodes instead? > > In essence, you are creating a new syscall here to manage and handle > devices for just your driver with the use of this ioctl. That's > generally a bad idea. > >> Virtual compute devices are shared between processes. There could be a >> Cuda application, Gimp and a Direct3D12 application working at the same >> time. > Why are all of those applications sharing anything? How are they > sharing anything? If you need to share something across processes, use > the existing inter-process ways of sharing stuff that we have today (12+ > different apis and counting). Don't create yet-another-one just for > your tiny little driver here. That's rude to the rest of the kernel. > >> This is what I mean by saying that there are multiple user mode >> clients who use the /dev/dxg driver interface. Each of this applications >> will open the /dev/dxg device node and enumerate/use virtual compute >> devices. > That seems like an odd model to follow. How many virtual devices do you > support? Is there a limit? Why not just enumerate them all to start > with? Or if there are too many, why not do it like the raw device and > have a single device node that is used to create the virtual devices you > wish to use? > >> If we change the way how the virtual compute devices are visible to user >> mode, the Cuda runtime, Direct3D runtime would need to be changed. > We do not care about existing userspace code at this point in time, you > are trying to get the kernel api correct here. Once you have done that, > then you can go fix up your userspace code to work properly. Anything > that came before today is pointless here, right? :) > >> I think we agreed that I will keep the global driver state in the device >> object as Wei suggested and remove global variables. There still will be >> a single /dev/dxg device node. Correct? > I do not know, as I can't figure out your goals here at all, sorry. > > Please go work with some experienced Linux developers in your company. > They should be the ones answering these questions for you, not me :) > > thanks, > > greg k-h Thanks Greg and folks for all the feedback. The userspace API provided by this driver is specifically designed to match the Windows compute device abstraction in order to light up a large eco-system of graphics and compute APIs in our various virtualized Linux environment. This is a pretty critical design point as this is the only way we can scale to such a vast eco-system of APIs and their related tools. This enables APIs and framework which are available natively on Linux, such as CUDA, OpenGL, OpenCL, OpenVINO, OneAPI/L0, TensorFlow, Pytorch, soon Vulkan and more in the future and allows them to share the host GPU or compute device when run under our hypervisor. Most of these APIs sit directly on top of our userspace API and changing that abstraction would have major implications. Our D3D/DML userspace components are only implementation details for a subset of those APIs. We have millions of developers experiencing and enjoying various Linux distributions through our Windows Subsystem for Linux (WSL) and this gives them a way to get hardware acceleration in a large number of important scenarios. WSL makes it trivial for developers to try out Linux totally risk free on over a billion devices, without having to commit to a native install upfront or understand and manually manage VMs. We think this is a pretty valuable proposition to both the Windows and the Linux community and one which has definitely been quite popular. That being said, we recognize that our virtualization approach makes some folks in the community uncomfortable and the resulting userspace API exposed by our driver will look a bit different than typical Linux driver. Taking into consideration feedback and reception, we decided to pause our upstream effort for this driver. We will keep our driver open source and available in our downstream kernel tree for folks who wants to leverage it in their private kernel. Thanks, Steve