[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4C7C4CE0.5080402@cn.fujitsu.com>
Date: Tue, 31 Aug 2010 08:29:20 +0800
From: Gui Jianfeng <guijianfeng@...fujitsu.com>
To: Vivek Goyal <vgoyal@...hat.com>
CC: Jens Axboe <axboe@...nel.dk>, Jeff Moyer <jmoyer@...hat.com>,
Divyesh Shah <dpshah@...gle.com>,
Corrado Zoccolo <czoccolo@...il.com>,
Nauman Rafique <nauman@...gle.com>,
linux kernel mailing list <linux-kernel@...r.kernel.org>
Subject: Re: [RFC] [PATCH] cfq-iosched: add cfq group hierarchical scheduling
support
Vivek Goyal wrote:
> On Mon, Aug 30, 2010 at 02:50:40PM +0800, Gui Jianfeng wrote:
>> Hi All,
>>
>> This patch enables cfq group hierarchical scheduling.
>>
>> With this patch, you can create a cgroup directory deeper than level 1.
>> Now, I/O Bandwidth is distributed in a hierarchy way. For example:
>> We create cgroup directories as following(the number represents weight):
>>
>> Root grp
>> / \
>> grp_1(100) grp_2(400)
>> / \
>> grp_3(200) grp_4(300)
>>
>> If grp_2 grp_3 and grp_4 are contending for I/O Bandwidth,
>> grp_2 will share 80% of total bandwidth.
>> For sub_groups, grp_3 shares 8%(20% * 40%), grp_4 shares 12%(20% * 60%)
>>
>> Design:
>> o Each cfq group has its own group service tree.
>> o Each cfq group contains a "group schedule entity" (gse) that
>> schedules on parent cfq group's service tree.
>> o Each cfq group contains a "queue schedule entity"(qse), it
>> represents all cfqqs located on this cfq group. It schedules
>> on this group's service tree. For the time being, root group
>> qse's weight is 1000, and subgroup qse's weight is 500.
>> o All gses and qse which belones to a same cfq group schedules
>> on the same group service tree.
>
> Hi Gui,
>
> Thanks for the patch. I have few questions.
>
> - So how does the hierarchy look like, w.r.t root group. Something as
> follows?
>
>
> root
> / | \
> q1 q2 G1
>
> Assume there are two processes doin IO in root group and q1 and q2 are
> cfqq queues for those processes and G1 is the cgroup created by user.
>
> If yes, then what algorithm do you use to do scheduling between q1, q2
> and G1? IOW, currently we have two algorithms operating in CFQ. One for
> cfqq and other for groups. Group algorithm does not use the logic of
> cfq_slice_offset().
Hi Vivek,
This patch doesn't break the original sheduling logic. That is cfqg => st => cfqq.
If q1 and q2 in root group, I treat q1 and q2 bundle as a queue sched entity, and
it will schedule on root group service with G1, as following:
root group
/ \
qse(q1,q2) gse(G1)
Thanks,
Gui
>
> The reason being they are different because CFQ wanted to provide ioprio
> service differentiation on SSD where slice idle is effectively zero. But
> that slice_offset() logic is approximate at best and is not suitable
> for group scheduling.
>
> I have yet to go through the patches in detail, so thought will ask you
> that how have you handeled above.
>
> Vivek
>
>> o cfq group allocates in a recursive manner, that means when a cfq
>> group needs to be allocated, the upper level cfq groups are also
>> allocated.
>> o When a cfq group served, not only charge this cfq group but also
>> charge its ancestors.
>>
>> Signed-off-by: Gui Jianfeng <guijianfeng@...fujitsu.com>
>> ---
>> block/blk-cgroup.c | 4 -
>> block/cfq-iosched.c | 477 +++++++++++++++++++++++++++++++++++++-------------
>> 2 files changed, 353 insertions(+), 128 deletions(-)
>>
>> diff --git a/block/blk-cgroup.c b/block/blk-cgroup.c
>> index 2fef1ef..aa06cb3 100644
>> --- a/block/blk-cgroup.c
>> +++ b/block/blk-cgroup.c
>> @@ -964,10 +964,6 @@ blkiocg_create(struct cgroup_subsys *subsys, struct cgroup *cgroup)
>> goto done;
>> }
>>
>> - /* Currently we do not support hierarchy deeper than two level (0,1) */
>> - if (parent != cgroup->top_cgroup)
>> - return ERR_PTR(-EPERM);
>> -
>> blkcg = kzalloc(sizeof(*blkcg), GFP_KERNEL);
>> if (!blkcg)
>> return ERR_PTR(-ENOMEM);
>> diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c
>> index f65c6f0..4660d89 100644
>> --- a/block/cfq-iosched.c
>> +++ b/block/cfq-iosched.c
>> @@ -73,7 +73,8 @@ static DEFINE_IDA(cic_index_ida);
>> #define cfq_class_rt(cfqq) ((cfqq)->ioprio_class == IOPRIO_CLASS_RT)
>>
>> #define sample_valid(samples) ((samples) > 80)
>> -#define rb_entry_cfqg(node) rb_entry((node), struct cfq_group, rb_node)
>> +#define rb_entry_se(node) \
>> + rb_entry((node), struct io_sched_entity, rb_node)
>>
>> /*
>> * Most of our rbtree usage is for sorting with min extraction, so
>> @@ -171,19 +172,40 @@ enum wl_type_t {
>> SYNC_WORKLOAD = 2
>> };
>>
>> -/* This is per cgroup per device grouping structure */
>> -struct cfq_group {
>> +/*
>> + * This's the schedule entity which is scheduled on group service tree.
>> + * It represents shedule structure of a cfq group, or a bundle of cfq
>> + * queues that locate in a cfq group.
>> + */
>> +struct io_sched_entity {
>> + struct cfq_rb_root *st;
>> +
>> /* group service_tree member */
>> struct rb_node rb_node;
>> -
>> - /* group service_tree key */
>> u64 vdisktime;
>> unsigned int weight;
>> bool on_st;
>> + bool is_group_se;
>> + struct io_sched_entity *parent;
>> +};
>> +
>> +/* This is per cgroup per device grouping structure */
>> +struct cfq_group {
>> + /* cfq group sched entity */
>> + struct io_sched_entity group_se;
>> +
>> + /* sched entity for cfqqs */
>> + struct io_sched_entity queue_se;
>> +
>> + /* Service tree for cfq_groups and cfqqs set*/
>> + struct cfq_rb_root grp_service_tree;
>>
>> /* number of cfqq currently on this group */
>> int nr_cfqq;
>>
>> + /* number of sub cfq groups */
>> + int nr_subgp;
>> +
>> /* Per group busy queus average. Useful for workload slice calc. */
>> unsigned int busy_queues_avg[2];
>> /*
>> @@ -210,8 +232,6 @@ struct cfq_group {
>> */
>> struct cfq_data {
>> struct request_queue *queue;
>> - /* Root service tree for cfq_groups */
>> - struct cfq_rb_root grp_service_tree;
>> struct cfq_group root_group;
>>
>> /*
>> @@ -399,6 +419,22 @@ static inline bool iops_mode(struct cfq_data *cfqd)
>> return false;
>> }
>>
>> +static inline struct cfq_group *cfqg_of_gse(struct io_sched_entity *se)
>> +{
>> + if (se->is_group_se)
>> + return container_of(se, struct cfq_group, group_se);
>> + else
>> + return NULL;
>> +}
>> +
>> +static inline struct cfq_group *cfqg_of_qse(struct io_sched_entity *se)
>> +{
>> + if (!se->is_group_se)
>> + return container_of(se, struct cfq_group, queue_se);
>> + else
>> + return NULL;
>> +}
>> +
>> static inline enum wl_prio_t cfqq_prio(struct cfq_queue *cfqq)
>> {
>> if (cfq_class_idle(cfqq))
>> @@ -522,12 +558,13 @@ cfq_prio_to_slice(struct cfq_data *cfqd, struct cfq_queue *cfqq)
>> return cfq_prio_slice(cfqd, cfq_cfqq_sync(cfqq), cfqq->ioprio);
>> }
>>
>> -static inline u64 cfq_scale_slice(unsigned long delta, struct cfq_group *cfqg)
>> +static inline u64 cfq_scale_slice(unsigned long delta,
>> + struct io_sched_entity *se)
>> {
>> u64 d = delta << CFQ_SERVICE_SHIFT;
>>
>> d = d * BLKIO_WEIGHT_DEFAULT;
>> - do_div(d, cfqg->weight);
>> + do_div(d, se->weight);
>> return d;
>> }
>>
>> @@ -552,16 +589,16 @@ static inline u64 min_vdisktime(u64 min_vdisktime, u64 vdisktime)
>> static void update_min_vdisktime(struct cfq_rb_root *st)
>> {
>> u64 vdisktime = st->min_vdisktime;
>> - struct cfq_group *cfqg;
>> + struct io_sched_entity *se;
>>
>> if (st->active) {
>> - cfqg = rb_entry_cfqg(st->active);
>> - vdisktime = cfqg->vdisktime;
>> + se = rb_entry_se(st->active);
>> + vdisktime = se->vdisktime;
>> }
>>
>> if (st->left) {
>> - cfqg = rb_entry_cfqg(st->left);
>> - vdisktime = min_vdisktime(vdisktime, cfqg->vdisktime);
>> + se = rb_entry_se(st->left);
>> + vdisktime = min_vdisktime(vdisktime, se->vdisktime);
>> }
>>
>> st->min_vdisktime = max_vdisktime(st->min_vdisktime, vdisktime);
>> @@ -591,9 +628,10 @@ static inline unsigned cfq_group_get_avg_queues(struct cfq_data *cfqd,
>> static inline unsigned
>> cfq_group_slice(struct cfq_data *cfqd, struct cfq_group *cfqg)
>> {
>> - struct cfq_rb_root *st = &cfqd->grp_service_tree;
>> + struct io_sched_entity *gse = &cfqg->queue_se;
>> + struct cfq_rb_root *st = gse->st;
>>
>> - return cfq_target_latency * cfqg->weight / st->total_weight;
>> + return cfq_target_latency * gse->weight / st->total_weight;
>> }
>>
>> static inline void
>> @@ -756,13 +794,13 @@ static struct cfq_queue *cfq_rb_first(struct cfq_rb_root *root)
>> return NULL;
>> }
>>
>> -static struct cfq_group *cfq_rb_first_group(struct cfq_rb_root *root)
>> +static struct io_sched_entity *cfq_rb_first_se(struct cfq_rb_root *root)
>> {
>> if (!root->left)
>> root->left = rb_first(&root->rb);
>>
>> if (root->left)
>> - return rb_entry_cfqg(root->left);
>> + return rb_entry_se(root->left);
>>
>> return NULL;
>> }
>> @@ -819,25 +857,25 @@ static unsigned long cfq_slice_offset(struct cfq_data *cfqd,
>> }
>>
>> static inline s64
>> -cfqg_key(struct cfq_rb_root *st, struct cfq_group *cfqg)
>> +se_key(struct cfq_rb_root *st, struct io_sched_entity *se)
>> {
>> - return cfqg->vdisktime - st->min_vdisktime;
>> + return se->vdisktime - st->min_vdisktime;
>> }
>>
>> static void
>> -__cfq_group_service_tree_add(struct cfq_rb_root *st, struct cfq_group *cfqg)
>> +__io_sched_entity_add(struct cfq_rb_root *st, struct io_sched_entity *se)
>> {
>> struct rb_node **node = &st->rb.rb_node;
>> struct rb_node *parent = NULL;
>> - struct cfq_group *__cfqg;
>> - s64 key = cfqg_key(st, cfqg);
>> + struct io_sched_entity *__se;
>> + s64 key = se_key(st, se);
>> int left = 1;
>>
>> while (*node != NULL) {
>> parent = *node;
>> - __cfqg = rb_entry_cfqg(parent);
>> + __se = rb_entry_se(parent);
>>
>> - if (key < cfqg_key(st, __cfqg))
>> + if (key < se_key(st, __se))
>> node = &parent->rb_left;
>> else {
>> node = &parent->rb_right;
>> @@ -846,47 +884,80 @@ __cfq_group_service_tree_add(struct cfq_rb_root *st, struct cfq_group *cfqg)
>> }
>>
>> if (left)
>> - st->left = &cfqg->rb_node;
>> + st->left = &se->rb_node;
>>
>> - rb_link_node(&cfqg->rb_node, parent, node);
>> - rb_insert_color(&cfqg->rb_node, &st->rb);
>> + rb_link_node(&se->rb_node, parent, node);
>> + rb_insert_color(&se->rb_node, &st->rb);
>> }
>>
>> -static void
>> -cfq_group_service_tree_add(struct cfq_data *cfqd, struct cfq_group *cfqg)
>> +static void io_sched_entity_add(struct cfq_rb_root *st,
>> + struct io_sched_entity *se)
>> {
>> - struct cfq_rb_root *st = &cfqd->grp_service_tree;
>> - struct cfq_group *__cfqg;
>> struct rb_node *n;
>> + struct io_sched_entity *__se;
>>
>> - cfqg->nr_cfqq++;
>> - if (cfqg->on_st)
>> - return;
>>
>> + if (se->on_st)
>> + return;
>> /*
>> * Currently put the group at the end. Later implement something
>> * so that groups get lesser vtime based on their weights, so that
>> * if group does not loose all if it was not continously backlogged.
>> */
>> - n = rb_last(&st->rb);
>> + n = rb_last(&se->st->rb);
>> if (n) {
>> - __cfqg = rb_entry_cfqg(n);
>> - cfqg->vdisktime = __cfqg->vdisktime + CFQ_IDLE_DELAY;
>> + __se = rb_entry_se(n);
>> + se->vdisktime = __se->vdisktime + CFQ_IDLE_DELAY;
>> } else
>> - cfqg->vdisktime = st->min_vdisktime;
>> + se->vdisktime = st->min_vdisktime;
>>
>> - __cfq_group_service_tree_add(st, cfqg);
>> - cfqg->on_st = true;
>> - st->total_weight += cfqg->weight;
>> + __io_sched_entity_add(se->st, se);
>> + se->on_st = true;
>> + st->total_weight += se->weight;
>> +}
>> +
>> +static void
>> +cfq_group_service_tree_add(struct cfq_data *cfqd, struct cfq_group *cfqg)
>> +{
>> + struct io_sched_entity *gse = &cfqg->group_se;
>> + struct io_sched_entity *qse = &cfqg->queue_se;
>> + struct cfq_group *__cfqg;
>> +
>> + cfqg->nr_cfqq++;
>> +
>> + io_sched_entity_add(qse->st, qse);
>> +
>> + while (gse && gse->parent) {
>> + if (gse->on_st)
>> + return;
>> + io_sched_entity_add(gse->st, gse);
>> + gse = gse->parent;
>> + __cfqg = cfqg_of_gse(gse);
>> + __cfqg->nr_subgp++;
>> + }
>> +}
>> +
>> +static void io_sched_entity_del(struct io_sched_entity *se)
>> +{
>> + if (!RB_EMPTY_NODE(&se->rb_node))
>> + cfq_rb_erase(&se->rb_node, se->st);
>> +
>> + se->on_st = false;
>> + se->st->total_weight -= se->weight;
>> }
>>
>> static void
>> cfq_group_service_tree_del(struct cfq_data *cfqd, struct cfq_group *cfqg)
>> {
>> - struct cfq_rb_root *st = &cfqd->grp_service_tree;
>> + struct io_sched_entity *gse = &cfqg->group_se;
>> + struct io_sched_entity *qse = &cfqg->queue_se;
>> + struct cfq_group *__cfqg, *p_cfqg;
>> +
>> + if (gse->st && gse->st->active == &gse->rb_node)
>> + gse->st->active = NULL;
>>
>> - if (st->active == &cfqg->rb_node)
>> - st->active = NULL;
>> + if (qse->st && qse->st->active == &qse->rb_node)
>> + qse->st->active = NULL;
>>
>> BUG_ON(cfqg->nr_cfqq < 1);
>> cfqg->nr_cfqq--;
>> @@ -895,13 +966,25 @@ cfq_group_service_tree_del(struct cfq_data *cfqd, struct cfq_group *cfqg)
>> if (cfqg->nr_cfqq)
>> return;
>>
>> - cfq_log_cfqg(cfqd, cfqg, "del_from_rr group");
>> - cfqg->on_st = false;
>> - st->total_weight -= cfqg->weight;
>> - if (!RB_EMPTY_NODE(&cfqg->rb_node))
>> - cfq_rb_erase(&cfqg->rb_node, st);
>> - cfqg->saved_workload_slice = 0;
>> - cfq_blkiocg_update_dequeue_stats(&cfqg->blkg, 1);
>> + /* dequeue queue se from group */
>> + io_sched_entity_del(qse);
>> +
>> + if (cfqg->nr_subgp)
>> + return;
>> +
>> + /* prevent from dequeuing root group */
>> + while (gse && gse->parent) {
>> + __cfqg = cfqg_of_gse(gse);
>> + p_cfqg = cfqg_of_gse(gse->parent);
>> + io_sched_entity_del(gse);
>> + cfq_blkiocg_update_dequeue_stats(&__cfqg->blkg, 1);
>> + cfq_log_cfqg(cfqd, __cfqg, "del_from_rr group");
>> + __cfqg->saved_workload_slice = 0;
>> + gse = gse->parent;
>> + p_cfqg->nr_subgp--;
>> + if (p_cfqg->nr_cfqq || p_cfqg->nr_subgp)
>> + return;
>> + }
>> }
>>
>> static inline unsigned int cfq_cfqq_slice_usage(struct cfq_queue *cfqq)
>> @@ -933,7 +1016,8 @@ static inline unsigned int cfq_cfqq_slice_usage(struct cfq_queue *cfqq)
>> static void cfq_group_served(struct cfq_data *cfqd, struct cfq_group *cfqg,
>> struct cfq_queue *cfqq)
>> {
>> - struct cfq_rb_root *st = &cfqd->grp_service_tree;
>> + struct io_sched_entity *gse = &cfqg->group_se;
>> + struct io_sched_entity *qse = &cfqg->queue_se;
>> unsigned int used_sl, charge;
>> int nr_sync = cfqg->nr_cfqq - cfqg_busy_async_queues(cfqd, cfqg)
>> - cfqg->service_tree_idle.count;
>> @@ -946,10 +1030,25 @@ static void cfq_group_served(struct cfq_data *cfqd, struct cfq_group *cfqg,
>> else if (!cfq_cfqq_sync(cfqq) && !nr_sync)
>> charge = cfqq->allocated_slice;
>>
>> - /* Can't update vdisktime while group is on service tree */
>> - cfq_rb_erase(&cfqg->rb_node, st);
>> - cfqg->vdisktime += cfq_scale_slice(charge, cfqg);
>> - __cfq_group_service_tree_add(st, cfqg);
>> + /*
>> + * update queue se's vdisktime.
>> + * Can't update vdisktime while group is on service tree.
>> + */
>> +
>> + cfq_rb_erase(&qse->rb_node, qse->st);
>> + qse->vdisktime += cfq_scale_slice(charge, qse);
>> + __io_sched_entity_add(qse->st, qse);
>> + if (&qse->rb_node == qse->st->active)
>> + qse->st->active = NULL;
>> +
>> + while (gse && gse->parent) {
>> + cfq_rb_erase(&gse->rb_node, gse->st);
>> + gse->vdisktime += cfq_scale_slice(charge, gse);
>> + __io_sched_entity_add(gse->st, gse);
>> + if (&gse->rb_node == gse->st->active)
>> + gse->st->active = NULL;
>> + gse = gse->parent;
>> + }
>>
>> /* This group is being expired. Save the context */
>> if (time_after(cfqd->workload_expires, jiffies)) {
>> @@ -960,8 +1059,9 @@ static void cfq_group_served(struct cfq_data *cfqd, struct cfq_group *cfqg,
>> } else
>> cfqg->saved_workload_slice = 0;
>>
>> - cfq_log_cfqg(cfqd, cfqg, "served: vt=%llu min_vt=%llu", cfqg->vdisktime,
>> - st->min_vdisktime);
>> + cfq_log_cfqg(cfqd, cfqg, "served: vt=%llu min_vt=%llu", gse->vdisktime,
>> + gse->st->min_vdisktime);
>> +
>> cfq_log_cfqq(cfqq->cfqd, cfqq, "sl_used=%u disp=%u charge=%u iops=%u"
>> " sect=%u", used_sl, cfqq->slice_dispatch, charge,
>> iops_mode(cfqd), cfqq->nr_sectors);
>> @@ -977,39 +1077,84 @@ static inline struct cfq_group *cfqg_of_blkg(struct blkio_group *blkg)
>> return NULL;
>> }
>>
>> +static void cfq_put_cfqg(struct cfq_group *cfqg)
>> +{
>> + struct cfq_rb_root *st;
>> + int i, j;
>> + struct io_sched_entity *gse;
>> + struct cfq_group *p_cfqg;
>> +
>> + BUG_ON(atomic_read(&cfqg->ref) <= 0);
>> + if (!atomic_dec_and_test(&cfqg->ref))
>> + return;
>> + for_each_cfqg_st(cfqg, i, j, st)
>> + BUG_ON(!RB_EMPTY_ROOT(&st->rb) || st->active != NULL);
>> +
>> + gse = &cfqg->group_se;
>> + if (gse->parent) {
>> + p_cfqg = cfqg_of_gse(gse->parent);
>> + /* Drop the reference taken by children */
>> + atomic_dec(&p_cfqg->ref);
>> + }
>> +
>> + kfree(cfqg);
>> +}
>> +
>> +static void cfq_destroy_cfqg(struct cfq_data *cfqd, struct cfq_group *cfqg)
>> +{
>> + /* Something wrong if we are trying to remove same group twice */
>> + BUG_ON(hlist_unhashed(&cfqg->cfqd_node));
>> +
>> + hlist_del_init(&cfqg->cfqd_node);
>> +
>> + /*
>> + * Put the reference taken at the time of creation so that when all
>> + * queues are gone, group can be destroyed.
>> + */
>> + cfq_put_cfqg(cfqg);
>> +}
>> +
>> void
>> cfq_update_blkio_group_weight(struct blkio_group *blkg, unsigned int weight)
>> {
>> - cfqg_of_blkg(blkg)->weight = weight;
>> + struct cfq_group *cfqg;
>> + struct io_sched_entity *gse;
>> +
>> + cfqg = cfqg_of_blkg(blkg);
>> + gse = &cfqg->group_se;
>> + gse->weight = weight;
>> }
>>
>> -static struct cfq_group *
>> -cfq_find_alloc_cfqg(struct cfq_data *cfqd, struct cgroup *cgroup, int create)
>> +static void init_group_queue_entity(struct blkio_cgroup *blkcg,
>> + struct cfq_group *cfqg)
>> +{
>> + struct io_sched_entity *gse = &cfqg->group_se;
>> + struct io_sched_entity *qse = &cfqg->queue_se;
>> +
>> + gse->weight = blkcg_get_weight(blkcg, cfqg->blkg.dev);
>> + RB_CLEAR_NODE(&gse->rb_node);
>> + gse->is_group_se = true;
>> + gse->on_st = false;
>> +
>> + /* Set to 500 for the time being */
>> + qse->weight = 500;
>> + RB_CLEAR_NODE(&qse->rb_node);
>> + qse->is_group_se = false;
>> + qse->on_st = false;
>> +}
>> +
>> +static void init_cfqg(struct cfq_data *cfqd, struct blkio_cgroup *blkcg,
>> + struct cfq_group *cfqg)
>> {
>> - struct blkio_cgroup *blkcg = cgroup_to_blkio_cgroup(cgroup);
>> - struct cfq_group *cfqg = NULL;
>> - void *key = cfqd;
>> int i, j;
>> struct cfq_rb_root *st;
>> - struct backing_dev_info *bdi = &cfqd->queue->backing_dev_info;
>> unsigned int major, minor;
>> + struct backing_dev_info *bdi = &cfqd->queue->backing_dev_info;
>>
>> - cfqg = cfqg_of_blkg(blkiocg_lookup_group(blkcg, key));
>> - if (cfqg && !cfqg->blkg.dev && bdi->dev && dev_name(bdi->dev)) {
>> - sscanf(dev_name(bdi->dev), "%u:%u", &major, &minor);
>> - cfqg->blkg.dev = MKDEV(major, minor);
>> - goto done;
>> - }
>> - if (cfqg || !create)
>> - goto done;
>> -
>> - cfqg = kzalloc_node(sizeof(*cfqg), GFP_ATOMIC, cfqd->queue->node);
>> - if (!cfqg)
>> - goto done;
>> + cfqg->grp_service_tree = CFQ_RB_ROOT;
>>
>> for_each_cfqg_st(cfqg, i, j, st)
>> *st = CFQ_RB_ROOT;
>> - RB_CLEAR_NODE(&cfqg->rb_node);
>>
>> /*
>> * Take the initial reference that will be released on destroy
>> @@ -1023,11 +1168,102 @@ cfq_find_alloc_cfqg(struct cfq_data *cfqd, struct cgroup *cgroup, int create)
>> sscanf(dev_name(bdi->dev), "%u:%u", &major, &minor);
>> cfq_blkiocg_add_blkio_group(blkcg, &cfqg->blkg, (void *)cfqd,
>> MKDEV(major, minor));
>> - cfqg->weight = blkcg_get_weight(blkcg, cfqg->blkg.dev);
>> -
>> + init_group_queue_entity(blkcg, cfqg);
>> /* Add group on cfqd list */
>> hlist_add_head(&cfqg->cfqd_node, &cfqd->cfqg_list);
>> +}
>> +
>> +static void uninit_cfqg(struct cfq_data *cfqd, struct cfq_group *cfqg)
>> +{
>> + if (!cfq_blkiocg_del_blkio_group(&cfqg->blkg))
>> + cfq_destroy_cfqg(cfqd, cfqg);
>> +}
>> +
>> +static void cfqg_set_parent(struct cfq_group *cfqg, struct cfq_group *p_cfqg)
>> +{
>> + struct io_sched_entity *gse = &cfqg->group_se;
>> + struct io_sched_entity *qse = &cfqg->queue_se;
>> + struct io_sched_entity *p_gse = &p_cfqg->group_se;
>> +
>> + gse->parent = p_gse;
>> + gse->st = &p_cfqg->grp_service_tree;
>> +
>> + qse->parent = gse;
>> + qse->st = &cfqg->grp_service_tree;
>> +
>> + /* child reference */
>> + atomic_inc(&p_cfqg->ref);
>> +}
>> +
>> +int cfqg_chain_alloc(struct cfq_data *cfqd, struct cgroup *cgroup)
>> +{
>> + struct blkio_cgroup *blkcg = cgroup_to_blkio_cgroup(cgroup);
>> + struct blkio_cgroup *p_blkcg;
>> + struct backing_dev_info *bdi = &cfqd->queue->backing_dev_info;
>> + unsigned int major, minor;
>> + struct cfq_group *cfqg, *p_cfqg;
>> + void *key = cfqd;
>> + int ret;
>> +
>> + cfqg = cfqg_of_blkg(blkiocg_lookup_group(blkcg, key));
>> + if (cfqg) {
>> + if (!cfqg->blkg.dev && bdi->dev && dev_name(bdi->dev)) {
>> + sscanf(dev_name(bdi->dev), "%u:%u", &major, &minor);
>> + cfqg->blkg.dev = MKDEV(major, minor);
>> + }
>> + /* chain has already been built */
>> + return 0;
>> + }
>> +
>> + cfqg = kzalloc_node(sizeof(*cfqg), GFP_ATOMIC, cfqd->queue->node);
>> + if (!cfqg)
>> + return -1;
>> +
>> + init_cfqg(cfqd, blkcg, cfqg);
>> +
>> + /* Already to the top group */
>> + if (!cgroup->parent)
>> + return 0;
>> +
>> + ret = cfqg_chain_alloc(cfqd, cgroup->parent);
>> + if (ret == -1) {
>> + uninit_cfqg(cfqd, cfqg);
>> + return -1;
>> + }
>> +
>> + p_blkcg = cgroup_to_blkio_cgroup(cgroup->parent);
>> + p_cfqg = cfqg_of_blkg(blkiocg_lookup_group(p_blkcg, key));
>> + BUG_ON(p_cfqg == NULL);
>> +
>> + cfqg_set_parent(cfqg, p_cfqg);
>> + return 0;
>> +}
>> +
>> +static struct cfq_group *
>> +cfq_find_alloc_cfqg(struct cfq_data *cfqd, struct cgroup *cgroup, int create)
>> +{
>> + struct blkio_cgroup *blkcg = cgroup_to_blkio_cgroup(cgroup);
>> + struct cfq_group *cfqg = NULL;
>> + void *key = cfqd;
>> + struct backing_dev_info *bdi = &cfqd->queue->backing_dev_info;
>> + unsigned int major, minor;
>> + int ret;
>>
>> + cfqg = cfqg_of_blkg(blkiocg_lookup_group(blkcg, key));
>> + if (cfqg && !cfqg->blkg.dev && bdi->dev && dev_name(bdi->dev)) {
>> + sscanf(dev_name(bdi->dev), "%u:%u", &major, &minor);
>> + cfqg->blkg.dev = MKDEV(major, minor);
>> + goto done;
>> + }
>> + if (cfqg || !create)
>> + goto done;
>> +
>> + ret = cfqg_chain_alloc(cfqd, cgroup);
>> + if (!ret) {
>> + cfqg = cfqg_of_blkg(blkiocg_lookup_group(blkcg, key));
>> + BUG_ON(cfqg == NULL);
>> + goto done;
>> + }
>> done:
>> return cfqg;
>> }
>> @@ -1067,33 +1303,6 @@ static void cfq_link_cfqq_cfqg(struct cfq_queue *cfqq, struct cfq_group *cfqg)
>> atomic_inc(&cfqq->cfqg->ref);
>> }
>>
>> -static void cfq_put_cfqg(struct cfq_group *cfqg)
>> -{
>> - struct cfq_rb_root *st;
>> - int i, j;
>> -
>> - BUG_ON(atomic_read(&cfqg->ref) <= 0);
>> - if (!atomic_dec_and_test(&cfqg->ref))
>> - return;
>> - for_each_cfqg_st(cfqg, i, j, st)
>> - BUG_ON(!RB_EMPTY_ROOT(&st->rb) || st->active != NULL);
>> - kfree(cfqg);
>> -}
>> -
>> -static void cfq_destroy_cfqg(struct cfq_data *cfqd, struct cfq_group *cfqg)
>> -{
>> - /* Something wrong if we are trying to remove same group twice */
>> - BUG_ON(hlist_unhashed(&cfqg->cfqd_node));
>> -
>> - hlist_del_init(&cfqg->cfqd_node);
>> -
>> - /*
>> - * Put the reference taken at the time of creation so that when all
>> - * queues are gone, group can be destroyed.
>> - */
>> - cfq_put_cfqg(cfqg);
>> -}
>> -
>> static void cfq_release_cfq_groups(struct cfq_data *cfqd)
>> {
>> struct hlist_node *pos, *n;
>> @@ -1668,9 +1877,6 @@ __cfq_slice_expired(struct cfq_data *cfqd, struct cfq_queue *cfqq,
>> if (cfqq == cfqd->active_queue)
>> cfqd->active_queue = NULL;
>>
>> - if (&cfqq->cfqg->rb_node == cfqd->grp_service_tree.active)
>> - cfqd->grp_service_tree.active = NULL;
>> -
>> if (cfqd->active_cic) {
>> put_io_context(cfqd->active_cic->ioc);
>> cfqd->active_cic = NULL;
>> @@ -2173,17 +2379,26 @@ static void choose_service_tree(struct cfq_data *cfqd, struct cfq_group *cfqg)
>> cfqd->noidle_tree_requires_idle = false;
>> }
>>
>> +
>> static struct cfq_group *cfq_get_next_cfqg(struct cfq_data *cfqd)
>> {
>> - struct cfq_rb_root *st = &cfqd->grp_service_tree;
>> + struct cfq_group *root_group = &cfqd->root_group;
>> + struct cfq_rb_root *st = &root_group->grp_service_tree;
>> struct cfq_group *cfqg;
>> + struct io_sched_entity *se;
>>
>> - if (RB_EMPTY_ROOT(&st->rb))
>> - return NULL;
>> - cfqg = cfq_rb_first_group(st);
>> - st->active = &cfqg->rb_node;
>> - update_min_vdisktime(st);
>> - return cfqg;
>> + do {
>> + se = cfq_rb_first_se(st);
>> + if (!se)
>> + return NULL;
>> + st->active = &se->rb_node;
>> + update_min_vdisktime(st);
>> + cfqg = cfqg_of_qse(se);
>> + if (cfqg)
>> + return cfqg;
>> + cfqg = cfqg_of_gse(se);
>> + st = &cfqg->grp_service_tree;
>> + } while (1);
>> }
>>
>> static void cfq_choose_cfqg(struct cfq_data *cfqd)
>> @@ -2215,15 +2430,18 @@ static struct cfq_queue *cfq_select_queue(struct cfq_data *cfqd)
>> if (!cfqq)
>> goto new_queue;
>>
>> +
>> if (!cfqd->rq_queued)
>> return NULL;
>>
>> +
>> /*
>> * We were waiting for group to get backlogged. Expire the queue
>> */
>> if (cfq_cfqq_wait_busy(cfqq) && !RB_EMPTY_ROOT(&cfqq->sort_list))
>> goto expire;
>>
>> +
>> /*
>> * The active queue has run out of time, expire it and select new.
>> */
>> @@ -2245,6 +2463,7 @@ static struct cfq_queue *cfq_select_queue(struct cfq_data *cfqd)
>> goto check_group_idle;
>> }
>>
>> +
>> /*
>> * The active queue has requests and isn't expired, allow it to
>> * dispatch.
>> @@ -2252,6 +2471,7 @@ static struct cfq_queue *cfq_select_queue(struct cfq_data *cfqd)
>> if (!RB_EMPTY_ROOT(&cfqq->sort_list))
>> goto keep_queue;
>>
>> +
>> /*
>> * If another queue has a request waiting within our mean seek
>> * distance, let it run. The expire code will check for close
>> @@ -2505,6 +2725,7 @@ static int cfq_dispatch_requests(struct request_queue *q, int force)
>> }
>>
>> cfq_log_cfqq(cfqd, cfqq, "dispatched a request");
>> +
>> return 1;
>> }
>>
>> @@ -3854,17 +4075,25 @@ static void *cfq_init_queue(struct request_queue *q)
>>
>> cfqd->cic_index = i;
>>
>> - /* Init root service tree */
>> - cfqd->grp_service_tree = CFQ_RB_ROOT;
>> -
>> /* Init root group */
>> cfqg = &cfqd->root_group;
>> + cfqg->grp_service_tree = CFQ_RB_ROOT;
>> for_each_cfqg_st(cfqg, i, j, st)
>> *st = CFQ_RB_ROOT;
>> - RB_CLEAR_NODE(&cfqg->rb_node);
>> -
>> + cfqg->group_se.is_group_se = true;
>> + RB_CLEAR_NODE(&cfqg->group_se.rb_node);
>> + cfqg->group_se.on_st = false;
>> /* Give preference to root group over other groups */
>> - cfqg->weight = 2*BLKIO_WEIGHT_DEFAULT;
>> + cfqg->group_se.weight = 2*BLKIO_WEIGHT_DEFAULT;
>> + cfqg->group_se.parent = NULL;
>> + cfqg->group_se.st = NULL;
>> +
>> + cfqg->queue_se.is_group_se = false;
>> + RB_CLEAR_NODE(&cfqg->queue_se.rb_node);
>> + cfqg->queue_se.on_st = false;
>> + cfqg->queue_se.weight = 2*BLKIO_WEIGHT_DEFAULT;
>> + cfqg->queue_se.parent = &cfqg->group_se;
>> + cfqg->queue_se.st = &cfqg->grp_service_tree;
>>
>> #ifdef CONFIG_CFQ_GROUP_IOSCHED
>> /*
>> --
>> 1.6.5.2
>>
>>
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists