> Bandan Das <bsd@xxxxxxxxxx> wrote on 03/21/2016 07:43:41 PM: > > "Michael Rapoport" <RAPOPORT@xxxxxxxxxx> writes: > > > > Hi Bandan, > > > >> From: Bandan Das <bsd@xxxxxxxxxx> > >> > >> At Linuxcon last year, based on our presentation "vhost: sharing is > > better" [1], > >> we had briefly discussed the idea of cgroup aware workqueues with Tejun. > > The > >> following patches are a result of the discussion. They are in no way > > complete in > >> that the changes are for unbounded workqueues only, but I just wanted to > > present my > >> unfinished work as RFC and get some feedback. > >> > >> 1/4 and 3/4 are simple cgroup changes and add a helper function. > >> 2/4 is the main implementation. > >> 4/4 changes vhost to use workqueues with support for cgroups. > >> > >> Example: > >> vhost creates a worker thread when invoked for a kvm guest. Since, > >> the guest is a normal process, the kernel thread servicing it should be > >> attached to the vm process' cgroups. > > > > I did some performance evaluation of different threading models in vhost, > > and in most tests replacing vhost kthread's with workqueues degrades the > > Workqueues us kthread_create internally and if calling one over the > other impacts performace, I think we should investigate that. Agree. Didn't have time to do it myself yet... > Which patches did you use ? Note that an earlier version of workqueue patches > that I posted used per-cpu workqueues. I've used your earlier version of workqueue patches, then I modified it to use unbound workqueues, and then I even restored to some extent original vhost workqueue usage. In all the cases I saw performance degradation relatively to the baseline. > > performance. Moreover, having thread management inside the vhost provides > > What exactly is the advantage doing our own thread management ? Do you have > any examples ? (Besides for doing our own scheduling like in the original Elvis > paper which I don't think is gonna happen). Also, note here that there is > a possibility to affect how our work gets executed by using optional switches to > alloc_workqueue() so all is not lost. Well, Elvis is a _theoretical_ example that showed that I/O scheduling in the vhost improves performance. I'm not saying we should take Evlis and try to squeeze it into the vhost, I just want to say that we cannot switch vhost to use workqueues if it causes performance degradation. My opinion is that we need to give it some more thought, much more performance evaluation, so that we can find the best model. > > opportunity for optimization, at least for some workloads... > > That said, I believe that switching vhost to use workqueues is not that > > good idea after all. > > -- Sincerely yours, Mike. -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html