On 3/10/22 02:13, Greg KH wrote: > On Wed, Mar 09, 2022 at 03:14:24PM -0800, Steve Pronovost wrote: >> On 3/3/22 05:16, Greg KH wrote: >>> On Wed, Mar 02, 2022 at 02:27:05PM -0800, Iouri Tarassov wrote: >>>> On 3/2/2022 12:20 PM, Greg KH wrote: >>>>> On Wed, Mar 02, 2022 at 10:49:15AM -0800, Iouri Tarassov wrote: >>>>>> On 3/2/2022 3:53 AM, Wei Liu wrote: >>>>>>> On Wed, Mar 02, 2022 at 08:53:15AM +0100, Greg KH wrote: >>>>>>>> On Tue, Mar 01, 2022 at 10:23:21PM +0000, Wei Liu wrote: >>>>>>>>>>> +struct dxgglobal *dxgglobal; >>>>>>>>>> No, make this per-device, NEVER have a single device for your driver. >>>>>>>>>> The Linux driver model makes it harder to do it this way than to do it >>>>>>>>>> correctly. Do it correctly please and have no global structures like >>>>>>>>>> this. >>>>>>>>>> >>>>>>>>> This may not be as big an issue as you thought. The device discovery is >>>>>>>>> still done via the normal VMBus probing routine. For all intents and >>>>>>>>> purposes the dxgglobal structure can be broken down into per device >>>>>>>>> fields and a global structure which contains the protocol versioning >>>>>>>>> information -- my understanding is there will always be a global >>>>>>>>> structure to hold information related to the backend, regardless of how >>>>>>>>> many devices there are. >>>>>>>> Then that is wrong and needs to be fixed. Drivers should almost never >>>>>>>> have any global data, that is not how Linux drivers work. What happens >>>>>>>> when you get a second device in your system for this? Major rework >>>>>>>> would have to happen and the code will break. Handle that all now as it >>>>>>>> takes less work to make this per-device than it does to have a global >>>>>>>> variable. >>>>>>>> >>>>>>> It is perhaps easier to draw parallel from an existing driver. I feel >>>>>>> like we're talking past each other. >>>>>>> >>>>>>> Let's look at drivers/iommu/intel/iommu.c. There are a bunch of lists >>>>>>> like `static LIST_HEAD(dmar_rmrr_units)`. During the probing phase, new >>>>>>> units will be added to the list. I this the current code is following >>>>>>> this model. dxgglobal fulfills the role of a list. >>>>>>> >>>>>>> Setting aside the question of whether it makes sense to keep a copy of >>>>>>> the per-VM state in each device instance, I can see the code be changed >>>>>>> to: >>>>>>> >>>>>>> struct mutex device_mutex; /* split out from dxgglobal */ >>>>>>> static LIST_HEAD(dxglist); >>>>>>> >>>>>>> /* Rename struct dxgglobal to struct dxgstate */ >>>>>>> struct dxgstate { >>>>>>> struct list_head dxglist; /* link for dxglist */ >>>>>>> /* ... original fields sans device_mutex */ >>>>>>> } >>>>>>> >>>>>>> /* >>>>>>> * Provide a bunch of helpers manipulate the list. Called in probe / >>>>>>> * remove etc. >>>>>>> */ >>>>>>> struct dxgstate *find_dxgstate(...); >>>>>>> void remove_dxgstate(...); >>>>>>> int add_dxgstate(...); >>>>>>> >>>>>>> This model is well understood and used in tree. It is just that it >>>>>>> doesn't provide much value in doing this now since the list will only >>>>>>> contain one element. I hope that you're not saying we cannot even use a >>>>>>> per-module pointer to quickly get the data structure we want to use, >>>>>>> right? >>>>>>> >>>>>>> Are you suggesting Iouri use dev_set_drvdata to stash the dxgstate >>>>>>> into the device object? I think that can be done too. >>>>>>> >>>>>>> The code can be changed as: >>>>>>> >>>>>>> /* Rename struct dxgglobal to dxgstate and remove unneeded fields */ >>>>>>> struct dxgstate { ... }; >>>>>>> >>>>>>> static int dxg_probe_vmbus(...) { >>>>>>> >>>>>>> /* probe successfully */ >>>>>>> >>>>>>> struct dxgstate *state = kmalloc(...); >>>>>>> /* Fill in dxgstate with information from backend */ >>>>>>> >>>>>>> /* hdev->dev is the device object from the core driver framework */ >>>>>>> dev_set_drvdata(&hdev->dev, state); >>>>>>> } >>>>>>> >>>>>>> static int dxg_remove_vmbus(...) { >>>>>>> /* Normal stuff here ...*/ >>>>>>> >>>>>>> struct dxgstate *state = dev_get_drvdata(...); >>>>>>> dev_set_drvdata(..., NULL); >>>>>>> kfree(state); >>>>>>> } >>>>>>> >>>>>>> /* In all other functions */ >>>>>>> void do_things(...) { >>>>>>> struct dxgstate *state = dev_get_drvdata(...); >>>>>>> >>>>>>> /* Use state in place of where dxgglobal was needed */ >>>>>>> >>>>>>> } >>>>>>> >>>>>>> Iouri, notice this doesn't change anything regarding how userspace is >>>>>>> designed. This is about how kernel organises its data. >>>>>>> >>>>>>> I hope what I wrote above can bring our understanding closer. >>>>>>> >>>>>>> Thanks, >>>>>>> Wei. >>>>>> I can certainly remove dxgglobal and keep the pointer to the global >>>>>> state in the device object. >>>>>> >>>>>> This will require passing of the global pointer to all functions, which >>>>>> need to access it. >>>>>> >>>>>> >>>>>> Maybe my understanding of the Greg's suggestion was not correct. I >>>>>> thought the suggestion was >>>>>> >>>>>> to have multiple /dev/dxgN devices (one per virtual compute device). >>>>> You have one device per HV device, as the bus already provides you. >>>>> That's all you really need, right? Who would be opening the same device >>>>> node multiple times? >>>>>> This would change how the user mode >>>>>> clients enumerate and communicate with compute devices. >>>>> What does userspace have to do here? It should just open the device >>>>> node that is present when needed. How will there be multiple userspace >>>>> clients for a single HV device? >>>> Dxgkrnl creates a single user mode visible device node /dev/dxg. >>> When you do that, you have a device to put all of your data on. Use >>> that. >>> >>>> It has >>>> nothing to do with a specific hardware compute device on the host. Its >>>> purpose is to provide services (IOCTLs) to enumerate and manage virtual >>>> compute devices, which represent hardware devices on the host. The VMBus >>>> devices are not used directly by user mode clients in the current design. >>> That's horrid, why not just export the virtual compute devices properly >>> through individual device nodes instead? >>> >>> In essence, you are creating a new syscall here to manage and handle >>> devices for just your driver with the use of this ioctl. That's >>> generally a bad idea. >>> >>>> Virtual compute devices are shared between processes. There could be a >>>> Cuda application, Gimp and a Direct3D12 application working at the same >>>> time. >>> Why are all of those applications sharing anything? How are they >>> sharing anything? If you need to share something across processes, use >>> the existing inter-process ways of sharing stuff that we have today (12+ >>> different apis and counting). Don't create yet-another-one just for >>> your tiny little driver here. That's rude to the rest of the kernel. >>> >>>> This is what I mean by saying that there are multiple user mode >>>> clients who use the /dev/dxg driver interface. Each of this applications >>>> will open the /dev/dxg device node and enumerate/use virtual compute >>>> devices. >>> That seems like an odd model to follow. How many virtual devices do you >>> support? Is there a limit? Why not just enumerate them all to start >>> with? Or if there are too many, why not do it like the raw device and >>> have a single device node that is used to create the virtual devices you >>> wish to use? >>> >>>> If we change the way how the virtual compute devices are visible to user >>>> mode, the Cuda runtime, Direct3D runtime would need to be changed. >>> We do not care about existing userspace code at this point in time, you >>> are trying to get the kernel api correct here. Once you have done that, >>> then you can go fix up your userspace code to work properly. Anything >>> that came before today is pointless here, right? :) >>> >>>> I think we agreed that I will keep the global driver state in the device >>>> object as Wei suggested and remove global variables. There still will be >>>> a single /dev/dxg device node. Correct? >>> I do not know, as I can't figure out your goals here at all, sorry. >>> >>> Please go work with some experienced Linux developers in your company. >>> They should be the ones answering these questions for you, not me :) >>> >>> thanks, >>> >>> greg k-h >> Thanks Greg and folks for all the feedback. The userspace API provided >> by this driver is specifically designed to match the Windows compute >> device abstraction in order to light up a large eco-system of graphics >> and compute APIs in our various virtualized Linux environment. This is >> a pretty critical design point as this is the only way we can scale to >> such a vast eco-system of APIs and their related tools. This enables >> APIs and framework which are available natively on Linux, such as >> CUDA, OpenGL, OpenCL, OpenVINO, OneAPI/L0, TensorFlow, Pytorch, soon >> Vulkan and more in the future and allows them to share the host GPU or >> compute device when run under our hypervisor. Most of these APIs sit >> directly on top of our userspace API and changing that abstraction >> would have major implications. Our D3D/DML userspace components are >> only implementation details for a subset of those APIs. > This feels odd to me as I was only discussing the in-kernel apis, and > how this driver does not follow them correctly at all (global device > state, odd device creation, etc.) That should have nothing to do with > the user/kernel api at all. And surely CUDA and the like never have > defined a specific user/kernal api that must be followed, right? > > Yes, the existing user/kernel api in this driver is a bit odd, but > really, we haven't even started to review that. For you all to just > cut-and-run at the first sign of a detailed in-kernel code review is > strange, why are you giving up now? > >> We have millions of developers experiencing and enjoying various Linux >> distributions through our Windows Subsystem for Linux (WSL) and this >> gives them a way to get hardware acceleration in a large number of >> important scenarios. WSL makes it trivial for developers to try out >> Linux totally risk free on over a billion devices, without having to >> commit to a native install upfront or understand and manually manage >> VMs. We think this is a pretty valuable proposition to both the >> Windows and the Linux community and one which has definitely been >> quite popular. > This has nothing to do with how this tiny driver uses the in-kernel apis > for creating and managing devices. You are discussing userspace > applications here, which is a nice marketing discussion about how many > developers you have, but means nothing when it comes to how the kernel > code works. So why mention it at all? > >> That being said, we recognize that our virtualization approach makes >> some folks in the community uncomfortable and the resulting userspace >> API exposed by our driver will look a bit different than typical Linux >> driver. Taking into consideration feedback and reception, we decided >> to pause our upstream effort for this driver. We will keep our driver >> open source and available in our downstream kernel tree for folks who >> wants to leverage it in their private kernel. > The problems I am pointing out here are how your kernel driver is > interacting with the kernel driver model. It has almost nothing to do > with the user/kernel api at all. > > And even if it did concern the user/kernel api, you already have a > userspace library that has to talk to this driver, all you need to do is > modify that to handle the expected changes that will happen when you > actually get experienced kernel developers to review your code and apis. > > I don't understand what you all were expecting here. Did you not want > to fit into the kernel tree and take advantage of using the proper > functions and interfaces that have been created already? That should > enable your users to be able to rely on a more stable and secure and > widespread driver for everyone to use. > > Instead you are saying that you don't appreciate the time and energy we > spent on trying to help you improve your code for your users, and wish > to just ignore our help? > > By doing this, you are telling all community members that they should > not waste their time in reviewing your company's code submissions as > they are not appreciated or wanted unless they are just a rubber-stamp > approval. That's going to ensure that no one ever reviews your > contributions again in the future, which is an odd request, don't you > think? > > Oh well, it's easy to add a filter to just ignore any future code > submissions from your email domain. I'll recommend that all other > kernel developers and maintainers do the same as we all have changes to > review from companies and developers that actually want our feedback. > > best of luck! > > greg k-h Hey Greg, Apologies, I didn’t mean to imply we don’t appreciate the feedback from you or the community nor were we looking for a rubber stamp approval. You’ve provided great feedback and we’ll continue to evolve our driver toward this and we will continue to review internally with more experienced Linux kernel developers to get our driver into a better shape that can be upstreamed It’s clear our code isn’t ready to be included upstream and we don’t want to waste reviewer’s time on it. We take all of the feedback seriously. Last year in our rfc, when we got the feedback that we really needed an open source userspace, we went to great length to get one. We will continue to integrate all of the feedback we have gotten so far. I totally agree that some of the feedback has nothing to do with the userspace API and we’ll evolve the driver toward that direction, it’s just goodness and will help make our driver simpler and more reliable. We will go back and figure out how to address this feedback, but it will take some time to do it right. The pause below wasn’t meant to imply we’re giving up forever, but that it will take a while before we can attempt again with a patch set that is hopefully better aligned and more mature in the future. My email below was meant to be a thank you for the feedback, we’ll pause iteration on the current patch set since we have a lot of internalizing to do and some things that will take quite a while to address and we didn’t want to just go silent. but obviously I did a very poor job of communicating that. Let me assure you there is no ill feeling on our end and we do appreciate the feedback, it may take some time but we will be back again :-). Thanks, Steve