[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-Id: <201603220712.u2M7CFdO004636@d06av03.portsmouth.uk.ibm.com>
Date: Tue, 22 Mar 2016 09:12:08 +0200
From: "Michael Rapoport" <RAPOPORT@...ibm.com>
To: Bandan Das <bsd@...hat.com>
Cc: jiangshanlai@...il.com, kvm@...r.kernel.org,
linux-kernel@...r.kernel.org, mst@...hat.com, tj@...nel.org
Subject: vhost threading model (was: Re: [RFC PATCH 0/4] cgroup aware workqueues)
> Bandan Das <bsd@...hat.com> wrote on 03/21/2016 07:43:41 PM:
> > "Michael Rapoport" <RAPOPORT@...ibm.com> writes:
> >
> > Hi Bandan,
> >
> >> From: Bandan Das <bsd@...hat.com>
> >>
> >> At Linuxcon last year, based on our presentation "vhost: sharing is
> > better" [1],
> >> we had briefly discussed the idea of cgroup aware workqueues with
Tejun.
> > The
> >> following patches are a result of the discussion. They are in no way
> > complete in
> >> that the changes are for unbounded workqueues only, but I just wanted
to
> > present my
> >> unfinished work as RFC and get some feedback.
> >>
> >> 1/4 and 3/4 are simple cgroup changes and add a helper function.
> >> 2/4 is the main implementation.
> >> 4/4 changes vhost to use workqueues with support for cgroups.
> >>
> >> Example:
> >> vhost creates a worker thread when invoked for a kvm guest. Since,
> >> the guest is a normal process, the kernel thread servicing it should
be
> >> attached to the vm process' cgroups.
> >
> > I did some performance evaluation of different threading models in
vhost,
> > and in most tests replacing vhost kthread's with workqueues degrades
the
>
> Workqueues us kthread_create internally and if calling one over the
> other impacts performace, I think we should investigate that.
Agree. Didn't have time to do it myself yet...
> Which patches did you use ? Note that an earlier version of workqueue
patches
> that I posted used per-cpu workqueues.
I've used your earlier version of workqueue patches, then I modified it to
use unbound workqueues, and then I even restored to some extent original
vhost workqueue usage. In all the cases I saw performance degradation
relatively to the baseline.
> > performance. Moreover, having thread management inside the vhost
provides
>
> What exactly is the advantage doing our own thread management ? Do you
have
> any examples ? (Besides for doing our own scheduling like in the
original Elvis
> paper which I don't think is gonna happen). Also, note here that there
is
> a possibility to affect how our work gets executed by using optional
switches to
> alloc_workqueue() so all is not lost.
Well, Elvis is a _theoretical_ example that showed that I/O scheduling in
the vhost improves performance.
I'm not saying we should take Evlis and try to squeeze it into the vhost,
I just want to say that we cannot switch vhost to use workqueues if it
causes performance degradation.
My opinion is that we need to give it some more thought, much more
performance evaluation, so that we can find the best model.
> > opportunity for optimization, at least for some workloads...
> > That said, I believe that switching vhost to use workqueues is not
that
> > good idea after all.
> >
--
Sincerely yours,
Mike.
Powered by blists - more mailing lists