[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20090529133804.GA26962@redhat.com>
Date: Fri, 29 May 2009 09:38:04 -0400
From: Vivek Goyal <vgoyal@...hat.com>
To: Ryo Tsuruta <ryov@...inux.co.jp>
Cc: linux-kernel@...r.kernel.org,
containers@...ts.linux-foundation.org, dm-devel@...hat.com,
jens.axboe@...cle.com, nauman@...gle.com, dpshah@...gle.com,
lizf@...fujitsu.com, mikew@...gle.com, fchecconi@...il.com,
paolo.valente@...more.it, fernando@....ntt.co.jp,
s-uchida@...jp.nec.com, taka@...inux.co.jp,
guijianfeng@...fujitsu.com, jmoyer@...hat.com,
dhaval@...ux.vnet.ibm.com, balbir@...ux.vnet.ibm.com,
righi.andrea@...il.com, m-ikeda@...jp.nec.com, jbaron@...hat.com,
agk@...hat.com, snitzer@...hat.com, akpm@...ux-foundation.org,
peterz@...radead.org
Subject: Re: [PATCH 15/20] io-controller: map async requests to appropriate
cgroup
On Fri, May 29, 2009 at 12:17:37PM +0900, Ryo Tsuruta wrote:
> Hi Vivek,
>
> Vivek Goyal <vgoyal@...hat.com> wrote:
> > On Thu, May 28, 2009 at 06:27:40PM +0900, Ryo Tsuruta wrote:
> > > Hi Vivek,
> > >
> > > > +#ifdef CONFIG_TRACK_ASYNC_CONTEXT
> > > > + if (elv_bio_sync(bio)) {
> > > > + /* sync io. Determine cgroup from submitting task context. */
> > > > + cgroup = task_cgroup(current, io_subsys_id);
> > > > + return cgroup;
> > > > + }
> > > > +
> > > > + /* Async io. Determine cgroup from with cgroup id stored in page */
> > > > + bio_cgroup_id = get_blkio_cgroup_id(bio);
> > > > +
> > > > + if (!bio_cgroup_id)
> > > > + return NULL;
> > > > +
> > > > + cgroup = blkio_cgroup_lookup(bio_cgroup_id);
> > > > +#else
> > > > + cgroup = task_cgroup(current, io_subsys_id);
> > > > +#endif
> > > > + return cgroup;
> > > > +}
> > >
> > > There is a case where a kernel thread (such as device-mapper drivers)
> > > submits a sync IO instead of a task which originates the IO. I think
> > > you should always use get_blkio_cgroup_id() to determine cgroup.
> > >
> >
> > Hi Ryo,
> >
> > Ok. Can you give some examples of drivers which are submitting reads in
> > different context al-together. You mentioned in the past that dm-crypt
> > looks like the one. How does current CFQ takes care of that. So if a
> > BE prio 7 or an RT prio 0, task is submitting a READ, CFQ will not know it
> > and it will put that READ in the queue of the READ submitting device
> > mapper thread (may be BE prio 3 or 4)?
>
> In the case of READ, dm-raid1 submits read IOs in differenct context
> under some conditions. dm-ioband also does it.
>
> > Always determining the cgroup from bio, will make things slower at the
> > same time complicated from the CFQ point of view. Right now cfq creates
> > and caches the queue pointer in the io context of the bio submitting task
> > and assumes sync requests are coming from that task/io context. Currently
> > there can only be one sync queue associated with one context. So if a single
> > thread is submitting reads (may be a worker thread) on behalf of other
> > processes, then we loose the io context information. In fact currently we
> > don't even carry ioprio and io class information in bio.
> >
> > So looks like we need to carry task io context information also in bio
> > to be able to associate the bio to right queue at CFQ level. This makes
> > it bit more complicated. For the time being I will keep it in my TODO
> > list and handle it once other more severe problems have been taken care
> > of.
>
> There is a patchset which makes every bio points the iocontext of the
> process which is originally generated an IO request.
>
> Date Tue, 22 Apr 2008 22:51:31 +0900 (JST)
> Subject [RFC][PATCH 1/10] I/O context inheritance
> From Hirokazu Takahashi <>
> http://lkml.org/lkml/2008/4/22/195
Ok, Thanks. This is good. So once above patches make to upstream, I will
just forward port my patches to make use of this infrastructure.
Thanks
Vivek
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists