lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4C7C4CE0.5080402@cn.fujitsu.com>
Date:	Tue, 31 Aug 2010 08:29:20 +0800
From:	Gui Jianfeng <guijianfeng@...fujitsu.com>
To:	Vivek Goyal <vgoyal@...hat.com>
CC:	Jens Axboe <axboe@...nel.dk>, Jeff Moyer <jmoyer@...hat.com>,
	Divyesh Shah <dpshah@...gle.com>,
	Corrado Zoccolo <czoccolo@...il.com>,
	Nauman Rafique <nauman@...gle.com>,
	linux kernel mailing list <linux-kernel@...r.kernel.org>
Subject: Re: [RFC] [PATCH] cfq-iosched: add cfq group hierarchical scheduling
 support

Vivek Goyal wrote:
> On Mon, Aug 30, 2010 at 02:50:40PM +0800, Gui Jianfeng wrote:
>> Hi All,
>>
>> This patch enables cfq group hierarchical scheduling.
>>
>> With this patch, you can create a cgroup directory deeper than level 1.
>> Now, I/O Bandwidth is distributed in a hierarchy way. For example:
>> We create cgroup directories as following(the number represents weight):
>>
>>             Root grp
>>            /       \
>>        grp_1(100) grp_2(400)
>>        /    \ 
>>   grp_3(200) grp_4(300)
>>
>> If grp_2 grp_3 and grp_4 are contending for I/O Bandwidth,
>> grp_2 will share 80% of total bandwidth.
>> For sub_groups, grp_3 shares 8%(20% * 40%), grp_4 shares 12%(20% * 60%)
>>
>> Design:
>>   o Each cfq group has its own group service tree. 
>>   o Each cfq group contains a "group schedule entity" (gse) that 
>>     schedules on parent cfq group's service tree.
>>   o Each cfq group contains a "queue schedule entity"(qse), it
>>     represents all cfqqs located on this cfq group. It schedules
>>     on this group's service tree. For the time being, root group
>>     qse's weight is 1000, and subgroup qse's weight is 500.
>>   o All gses and qse which belones to a same cfq group schedules
>>     on the same group service tree.
> 
> Hi Gui,
> 
> Thanks for the patch. I have few questions.
> 
> - So how does the hierarchy look like, w.r.t root group. Something as
>   follows?
> 
> 
> 			root
> 		       / | \
> 		     q1  q2 G1
> 
> Assume there are two processes doin IO in root group and q1 and q2 are
> cfqq queues for those processes and G1 is the cgroup created by user.
> 
> If yes, then what algorithm do you use to do scheduling between q1, q2
> and G1? IOW, currently we have two algorithms operating in CFQ. One for
> cfqq and other for groups. Group algorithm does not use the logic of
> cfq_slice_offset().

Hi Vivek,

This patch doesn't break the original sheduling logic. That is cfqg => st => cfqq.
If q1 and q2 in root group, I treat q1 and q2 bundle as a queue sched entity, and
it will schedule on root group service with G1, as following:

                         root group
                        /         \
                    qse(q1,q2)    gse(G1)

Thanks,
Gui

> 
> The reason being they are different because CFQ wanted to provide ioprio
> service differentiation on SSD where slice idle is effectively zero. But
> that slice_offset() logic is approximate at best and is not suitable
> for group scheduling.
> 
> I have yet to go through the patches in detail, so thought will ask you
> that how have you handeled above.
> 
> Vivek
> 
>>   o cfq group allocates in a recursive manner, that means when a cfq 
>>     group needs to be allocated, the upper level cfq groups are also
>>     allocated.
>>   o When a cfq group served, not only charge this cfq group but also
>>     charge its ancestors.
>>
>> Signed-off-by: Gui Jianfeng <guijianfeng@...fujitsu.com>
>> ---
>>  block/blk-cgroup.c  |    4 -
>>  block/cfq-iosched.c |  477 +++++++++++++++++++++++++++++++++++++-------------
>>  2 files changed, 353 insertions(+), 128 deletions(-)
>>
>> diff --git a/block/blk-cgroup.c b/block/blk-cgroup.c
>> index 2fef1ef..aa06cb3 100644
>> --- a/block/blk-cgroup.c
>> +++ b/block/blk-cgroup.c
>> @@ -964,10 +964,6 @@ blkiocg_create(struct cgroup_subsys *subsys, struct cgroup *cgroup)
>>  		goto done;
>>  	}
>>  
>> -	/* Currently we do not support hierarchy deeper than two level (0,1) */
>> -	if (parent != cgroup->top_cgroup)
>> -		return ERR_PTR(-EPERM);
>> -
>>  	blkcg = kzalloc(sizeof(*blkcg), GFP_KERNEL);
>>  	if (!blkcg)
>>  		return ERR_PTR(-ENOMEM);
>> diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c
>> index f65c6f0..4660d89 100644
>> --- a/block/cfq-iosched.c
>> +++ b/block/cfq-iosched.c
>> @@ -73,7 +73,8 @@ static DEFINE_IDA(cic_index_ida);
>>  #define cfq_class_rt(cfqq)	((cfqq)->ioprio_class == IOPRIO_CLASS_RT)
>>  
>>  #define sample_valid(samples)	((samples) > 80)
>> -#define rb_entry_cfqg(node)	rb_entry((node), struct cfq_group, rb_node)
>> +#define rb_entry_se(node)	\
>> +	rb_entry((node), struct io_sched_entity, rb_node)
>>  
>>  /*
>>   * Most of our rbtree usage is for sorting with min extraction, so
>> @@ -171,19 +172,40 @@ enum wl_type_t {
>>  	SYNC_WORKLOAD = 2
>>  };
>>  
>> -/* This is per cgroup per device grouping structure */
>> -struct cfq_group {
>> +/* 
>> + * This's the schedule entity which is scheduled on group service tree.
>> + * It represents shedule structure of a cfq group, or a bundle of cfq
>> + * queues that locate in a cfq group.
>> + */
>> +struct io_sched_entity {
>> +	struct cfq_rb_root *st;
>> +
>>  	/* group service_tree member */
>>  	struct rb_node rb_node;
>> -
>> -	/* group service_tree key */
>>  	u64 vdisktime;
>>  	unsigned int weight;
>>  	bool on_st;
>> +	bool is_group_se;
>> +	struct io_sched_entity *parent;
>> +};
>> +
>> +/* This is per cgroup per device grouping structure */
>> +struct cfq_group {
>> +	/* cfq group sched entity */
>> +	struct io_sched_entity group_se;
>> +
>> +	/* sched entity for cfqqs */
>> +	struct io_sched_entity queue_se;
>> +
>> +	/* Service tree for cfq_groups and cfqqs set*/
>> +	struct cfq_rb_root grp_service_tree;
>>  
>>  	/* number of cfqq currently on this group */
>>  	int nr_cfqq;
>>  
>> +	/* number of sub cfq groups */
>> +	int nr_subgp;
>> +
>>  	/* Per group busy queus average. Useful for workload slice calc. */
>>  	unsigned int busy_queues_avg[2];
>>  	/*
>> @@ -210,8 +232,6 @@ struct cfq_group {
>>   */
>>  struct cfq_data {
>>  	struct request_queue *queue;
>> -	/* Root service tree for cfq_groups */
>> -	struct cfq_rb_root grp_service_tree;
>>  	struct cfq_group root_group;
>>  
>>  	/*
>> @@ -399,6 +419,22 @@ static inline bool iops_mode(struct cfq_data *cfqd)
>>  		return false;
>>  }
>>  
>> +static inline struct cfq_group *cfqg_of_gse(struct io_sched_entity *se)
>> +{
>> +	if (se->is_group_se)
>> +		return container_of(se, struct cfq_group, group_se);
>> +	else
>> +		return NULL;
>> +}
>> +
>> +static inline struct cfq_group *cfqg_of_qse(struct io_sched_entity *se)
>> +{
>> +	if (!se->is_group_se)
>> +		return container_of(se, struct cfq_group, queue_se);
>> +	else
>> +		return NULL;
>> +}
>> +
>>  static inline enum wl_prio_t cfqq_prio(struct cfq_queue *cfqq)
>>  {
>>  	if (cfq_class_idle(cfqq))
>> @@ -522,12 +558,13 @@ cfq_prio_to_slice(struct cfq_data *cfqd, struct cfq_queue *cfqq)
>>  	return cfq_prio_slice(cfqd, cfq_cfqq_sync(cfqq), cfqq->ioprio);
>>  }
>>  
>> -static inline u64 cfq_scale_slice(unsigned long delta, struct cfq_group *cfqg)
>> +static inline u64 cfq_scale_slice(unsigned long delta,
>> +				  struct io_sched_entity *se)
>>  {
>>  	u64 d = delta << CFQ_SERVICE_SHIFT;
>>  
>>  	d = d * BLKIO_WEIGHT_DEFAULT;
>> -	do_div(d, cfqg->weight);
>> +	do_div(d, se->weight);
>>  	return d;
>>  }
>>  
>> @@ -552,16 +589,16 @@ static inline u64 min_vdisktime(u64 min_vdisktime, u64 vdisktime)
>>  static void update_min_vdisktime(struct cfq_rb_root *st)
>>  {
>>  	u64 vdisktime = st->min_vdisktime;
>> -	struct cfq_group *cfqg;
>> +	struct io_sched_entity *se;
>>  
>>  	if (st->active) {
>> -		cfqg = rb_entry_cfqg(st->active);
>> -		vdisktime = cfqg->vdisktime;
>> +		se = rb_entry_se(st->active);
>> +		vdisktime = se->vdisktime;
>>  	}
>>  
>>  	if (st->left) {
>> -		cfqg = rb_entry_cfqg(st->left);
>> -		vdisktime = min_vdisktime(vdisktime, cfqg->vdisktime);
>> +		se = rb_entry_se(st->left);
>> +		vdisktime = min_vdisktime(vdisktime, se->vdisktime);
>>  	}
>>  
>>  	st->min_vdisktime = max_vdisktime(st->min_vdisktime, vdisktime);
>> @@ -591,9 +628,10 @@ static inline unsigned cfq_group_get_avg_queues(struct cfq_data *cfqd,
>>  static inline unsigned
>>  cfq_group_slice(struct cfq_data *cfqd, struct cfq_group *cfqg)
>>  {
>> -	struct cfq_rb_root *st = &cfqd->grp_service_tree;
>> +	struct io_sched_entity *gse = &cfqg->queue_se;
>> +	struct cfq_rb_root *st = gse->st;
>>  
>> -	return cfq_target_latency * cfqg->weight / st->total_weight;
>> +	return cfq_target_latency * gse->weight / st->total_weight;
>>  }
>>  
>>  static inline void
>> @@ -756,13 +794,13 @@ static struct cfq_queue *cfq_rb_first(struct cfq_rb_root *root)
>>  	return NULL;
>>  }
>>  
>> -static struct cfq_group *cfq_rb_first_group(struct cfq_rb_root *root)
>> +static struct io_sched_entity *cfq_rb_first_se(struct cfq_rb_root *root)
>>  {
>>  	if (!root->left)
>>  		root->left = rb_first(&root->rb);
>>  
>>  	if (root->left)
>> -		return rb_entry_cfqg(root->left);
>> +		return rb_entry_se(root->left);
>>  
>>  	return NULL;
>>  }
>> @@ -819,25 +857,25 @@ static unsigned long cfq_slice_offset(struct cfq_data *cfqd,
>>  }
>>  
>>  static inline s64
>> -cfqg_key(struct cfq_rb_root *st, struct cfq_group *cfqg)
>> +se_key(struct cfq_rb_root *st, struct io_sched_entity *se)
>>  {
>> -	return cfqg->vdisktime - st->min_vdisktime;
>> +	return se->vdisktime - st->min_vdisktime;
>>  }
>>  
>>  static void
>> -__cfq_group_service_tree_add(struct cfq_rb_root *st, struct cfq_group *cfqg)
>> +__io_sched_entity_add(struct cfq_rb_root *st, struct io_sched_entity *se)
>>  {
>>  	struct rb_node **node = &st->rb.rb_node;
>>  	struct rb_node *parent = NULL;
>> -	struct cfq_group *__cfqg;
>> -	s64 key = cfqg_key(st, cfqg);
>> +	struct io_sched_entity *__se;
>> +	s64 key = se_key(st, se);
>>  	int left = 1;
>>  
>>  	while (*node != NULL) {
>>  		parent = *node;
>> -		__cfqg = rb_entry_cfqg(parent);
>> +		__se = rb_entry_se(parent);
>>  
>> -		if (key < cfqg_key(st, __cfqg))
>> +		if (key < se_key(st, __se))
>>  			node = &parent->rb_left;
>>  		else {
>>  			node = &parent->rb_right;
>> @@ -846,47 +884,80 @@ __cfq_group_service_tree_add(struct cfq_rb_root *st, struct cfq_group *cfqg)
>>  	}
>>  
>>  	if (left)
>> -		st->left = &cfqg->rb_node;
>> +		st->left = &se->rb_node;
>>  
>> -	rb_link_node(&cfqg->rb_node, parent, node);
>> -	rb_insert_color(&cfqg->rb_node, &st->rb);
>> +	rb_link_node(&se->rb_node, parent, node);
>> +	rb_insert_color(&se->rb_node, &st->rb);
>>  }
>>  
>> -static void
>> -cfq_group_service_tree_add(struct cfq_data *cfqd, struct cfq_group *cfqg)
>> +static void io_sched_entity_add(struct cfq_rb_root *st,
>> +				struct io_sched_entity *se)
>>  {
>> -	struct cfq_rb_root *st = &cfqd->grp_service_tree;
>> -	struct cfq_group *__cfqg;
>>  	struct rb_node *n;
>> +	struct io_sched_entity *__se;
>>  
>> -	cfqg->nr_cfqq++;
>> -	if (cfqg->on_st)
>> -		return;
>>  
>> +	if (se->on_st)
>> +		return;
>>  	/*
>>  	 * Currently put the group at the end. Later implement something
>>  	 * so that groups get lesser vtime based on their weights, so that
>>  	 * if group does not loose all if it was not continously backlogged.
>>  	 */
>> -	n = rb_last(&st->rb);
>> +	n = rb_last(&se->st->rb);
>>  	if (n) {
>> -		__cfqg = rb_entry_cfqg(n);
>> -		cfqg->vdisktime = __cfqg->vdisktime + CFQ_IDLE_DELAY;
>> +		__se = rb_entry_se(n);
>> +		se->vdisktime = __se->vdisktime + CFQ_IDLE_DELAY;
>>  	} else
>> -		cfqg->vdisktime = st->min_vdisktime;
>> +		se->vdisktime = st->min_vdisktime;
>>  
>> -	__cfq_group_service_tree_add(st, cfqg);
>> -	cfqg->on_st = true;
>> -	st->total_weight += cfqg->weight;
>> +	__io_sched_entity_add(se->st, se);
>> +	se->on_st = true;
>> +	st->total_weight += se->weight;
>> +}
>> +
>> +static void
>> +cfq_group_service_tree_add(struct cfq_data *cfqd, struct cfq_group *cfqg)
>> +{
>> +	struct io_sched_entity *gse = &cfqg->group_se;
>> +	struct io_sched_entity *qse = &cfqg->queue_se;
>> +	struct cfq_group *__cfqg;
>> +
>> +	cfqg->nr_cfqq++;
>> +
>> +	io_sched_entity_add(qse->st, qse);
>> +
>> +	while (gse && gse->parent) {
>> +		if (gse->on_st)
>> +			return;
>> +		io_sched_entity_add(gse->st, gse);
>> +		gse = gse->parent;
>> +		__cfqg = cfqg_of_gse(gse);
>> +		__cfqg->nr_subgp++;
>> +	}
>> +}
>> +
>> +static void io_sched_entity_del(struct io_sched_entity *se)
>> +{
>> +	if (!RB_EMPTY_NODE(&se->rb_node))
>> +		cfq_rb_erase(&se->rb_node, se->st);
>> +
>> +	se->on_st = false;
>> +	se->st->total_weight -= se->weight;
>>  }
>>  
>>  static void
>>  cfq_group_service_tree_del(struct cfq_data *cfqd, struct cfq_group *cfqg)
>>  {
>> -	struct cfq_rb_root *st = &cfqd->grp_service_tree;
>> +	struct io_sched_entity *gse = &cfqg->group_se;
>> +	struct io_sched_entity *qse = &cfqg->queue_se;
>> +	struct cfq_group *__cfqg, *p_cfqg;
>> +
>> +	if (gse->st && gse->st->active == &gse->rb_node)
>> +		gse->st->active = NULL;
>>  
>> -	if (st->active == &cfqg->rb_node)
>> -		st->active = NULL;
>> +	if (qse->st && qse->st->active == &qse->rb_node)
>> +		qse->st->active = NULL;
>>  
>>  	BUG_ON(cfqg->nr_cfqq < 1);
>>  	cfqg->nr_cfqq--;
>> @@ -895,13 +966,25 @@ cfq_group_service_tree_del(struct cfq_data *cfqd, struct cfq_group *cfqg)
>>  	if (cfqg->nr_cfqq)
>>  		return;
>>  
>> -	cfq_log_cfqg(cfqd, cfqg, "del_from_rr group");
>> -	cfqg->on_st = false;
>> -	st->total_weight -= cfqg->weight;
>> -	if (!RB_EMPTY_NODE(&cfqg->rb_node))
>> -		cfq_rb_erase(&cfqg->rb_node, st);
>> -	cfqg->saved_workload_slice = 0;
>> -	cfq_blkiocg_update_dequeue_stats(&cfqg->blkg, 1);
>> +	/* dequeue queue se from group */
>> +	io_sched_entity_del(qse);
>> +
>> +	if (cfqg->nr_subgp)
>> +		return;
>> +
>> +	/* prevent from dequeuing root group */
>> +	while (gse && gse->parent) {
>> +		__cfqg = cfqg_of_gse(gse);
>> +		p_cfqg = cfqg_of_gse(gse->parent);
>> +		io_sched_entity_del(gse);
>> +		cfq_blkiocg_update_dequeue_stats(&__cfqg->blkg, 1);
>> +		cfq_log_cfqg(cfqd, __cfqg, "del_from_rr group");
>> +		__cfqg->saved_workload_slice = 0;
>> +		gse = gse->parent;
>> +		p_cfqg->nr_subgp--;
>> +		if (p_cfqg->nr_cfqq || p_cfqg->nr_subgp)
>> +			return;
>> +	}
>>  }
>>  
>>  static inline unsigned int cfq_cfqq_slice_usage(struct cfq_queue *cfqq)
>> @@ -933,7 +1016,8 @@ static inline unsigned int cfq_cfqq_slice_usage(struct cfq_queue *cfqq)
>>  static void cfq_group_served(struct cfq_data *cfqd, struct cfq_group *cfqg,
>>  				struct cfq_queue *cfqq)
>>  {
>> -	struct cfq_rb_root *st = &cfqd->grp_service_tree;
>> +	struct io_sched_entity *gse = &cfqg->group_se;
>> +	struct io_sched_entity *qse = &cfqg->queue_se;
>>  	unsigned int used_sl, charge;
>>  	int nr_sync = cfqg->nr_cfqq - cfqg_busy_async_queues(cfqd, cfqg)
>>  			- cfqg->service_tree_idle.count;
>> @@ -946,10 +1030,25 @@ static void cfq_group_served(struct cfq_data *cfqd, struct cfq_group *cfqg,
>>  	else if (!cfq_cfqq_sync(cfqq) && !nr_sync)
>>  		charge = cfqq->allocated_slice;
>>  
>> -	/* Can't update vdisktime while group is on service tree */
>> -	cfq_rb_erase(&cfqg->rb_node, st);
>> -	cfqg->vdisktime += cfq_scale_slice(charge, cfqg);
>> -	__cfq_group_service_tree_add(st, cfqg);
>> +	/*
>> +	 *  update queue se's vdisktime.
>> +	 *  Can't update vdisktime while group is on service tree.
>> +	 */
>> +
>> +	cfq_rb_erase(&qse->rb_node, qse->st);
>> +	qse->vdisktime += cfq_scale_slice(charge, qse);
>> +	__io_sched_entity_add(qse->st, qse);
>> +	if (&qse->rb_node == qse->st->active)
>> +		qse->st->active = NULL;
>> +
>> +	while (gse && gse->parent) {
>> +		cfq_rb_erase(&gse->rb_node, gse->st);
>> +		gse->vdisktime += cfq_scale_slice(charge, gse);
>> +		__io_sched_entity_add(gse->st, gse);
>> +		if (&gse->rb_node == gse->st->active)
>> +			gse->st->active = NULL;
>> +		gse = gse->parent;
>> +	}
>>  
>>  	/* This group is being expired. Save the context */
>>  	if (time_after(cfqd->workload_expires, jiffies)) {
>> @@ -960,8 +1059,9 @@ static void cfq_group_served(struct cfq_data *cfqd, struct cfq_group *cfqg,
>>  	} else
>>  		cfqg->saved_workload_slice = 0;
>>  
>> -	cfq_log_cfqg(cfqd, cfqg, "served: vt=%llu min_vt=%llu", cfqg->vdisktime,
>> -					st->min_vdisktime);
>> +	cfq_log_cfqg(cfqd, cfqg, "served: vt=%llu min_vt=%llu", gse->vdisktime,
>> +		     gse->st->min_vdisktime);
>> +
>>  	cfq_log_cfqq(cfqq->cfqd, cfqq, "sl_used=%u disp=%u charge=%u iops=%u"
>>  			" sect=%u", used_sl, cfqq->slice_dispatch, charge,
>>  			iops_mode(cfqd), cfqq->nr_sectors);
>> @@ -977,39 +1077,84 @@ static inline struct cfq_group *cfqg_of_blkg(struct blkio_group *blkg)
>>  	return NULL;
>>  }
>>  
>> +static void cfq_put_cfqg(struct cfq_group *cfqg)
>> +{
>> +	struct cfq_rb_root *st;
>> +	int i, j;
>> +	struct io_sched_entity *gse;
>> +	struct cfq_group *p_cfqg;
>> +
>> +	BUG_ON(atomic_read(&cfqg->ref) <= 0);
>> +	if (!atomic_dec_and_test(&cfqg->ref))
>> +		return;
>> +	for_each_cfqg_st(cfqg, i, j, st)
>> +		BUG_ON(!RB_EMPTY_ROOT(&st->rb) || st->active != NULL);
>> +
>> +	gse = &cfqg->group_se;
>> +	if (gse->parent) {
>> +		p_cfqg = cfqg_of_gse(gse->parent);
>> +		/* Drop the reference taken by children */
>> +		atomic_dec(&p_cfqg->ref);
>> +	}
>> +
>> +	kfree(cfqg);
>> +}
>> +
>> +static void cfq_destroy_cfqg(struct cfq_data *cfqd, struct cfq_group *cfqg)
>> +{
>> +	/* Something wrong if we are trying to remove same group twice */
>> +	BUG_ON(hlist_unhashed(&cfqg->cfqd_node));
>> +
>> +	hlist_del_init(&cfqg->cfqd_node);
>> +
>> +	/*
>> +	 * Put the reference taken at the time of creation so that when all
>> +	 * queues are gone, group can be destroyed.
>> +	 */
>> +	cfq_put_cfqg(cfqg);
>> +}
>> +
>>  void
>>  cfq_update_blkio_group_weight(struct blkio_group *blkg, unsigned int weight)
>>  {
>> -	cfqg_of_blkg(blkg)->weight = weight;
>> +	struct cfq_group *cfqg;
>> +	struct io_sched_entity *gse;
>> +
>> +	cfqg = cfqg_of_blkg(blkg);
>> +	gse = &cfqg->group_se;
>> +	gse->weight = weight;
>>  }
>>  
>> -static struct cfq_group *
>> -cfq_find_alloc_cfqg(struct cfq_data *cfqd, struct cgroup *cgroup, int create)
>> +static void init_group_queue_entity(struct blkio_cgroup *blkcg,
>> +				   struct cfq_group *cfqg)
>> +{
>> +	struct io_sched_entity *gse = &cfqg->group_se;
>> +	struct io_sched_entity *qse = &cfqg->queue_se;
>> +
>> +	gse->weight = blkcg_get_weight(blkcg, cfqg->blkg.dev);
>> +	RB_CLEAR_NODE(&gse->rb_node);
>> +	gse->is_group_se = true;
>> +	gse->on_st = false;
>> +
>> +	/* Set to 500 for the time being */
>> +	qse->weight = 500;
>> +	RB_CLEAR_NODE(&qse->rb_node);
>> +	qse->is_group_se = false;
>> +	qse->on_st = false;
>> +}
>> +
>> +static void init_cfqg(struct cfq_data *cfqd, struct blkio_cgroup *blkcg,
>> +		      struct cfq_group *cfqg)
>>  {
>> -	struct blkio_cgroup *blkcg = cgroup_to_blkio_cgroup(cgroup);
>> -	struct cfq_group *cfqg = NULL;
>> -	void *key = cfqd;
>>  	int i, j;
>>  	struct cfq_rb_root *st;
>> -	struct backing_dev_info *bdi = &cfqd->queue->backing_dev_info;
>>  	unsigned int major, minor;
>> +	struct backing_dev_info *bdi = &cfqd->queue->backing_dev_info;
>>  
>> -	cfqg = cfqg_of_blkg(blkiocg_lookup_group(blkcg, key));
>> -	if (cfqg && !cfqg->blkg.dev && bdi->dev && dev_name(bdi->dev)) {
>> -		sscanf(dev_name(bdi->dev), "%u:%u", &major, &minor);
>> -		cfqg->blkg.dev = MKDEV(major, minor);
>> -		goto done;
>> -	}
>> -	if (cfqg || !create)
>> -		goto done;
>> -
>> -	cfqg = kzalloc_node(sizeof(*cfqg), GFP_ATOMIC, cfqd->queue->node);
>> -	if (!cfqg)
>> -		goto done;
>> +	cfqg->grp_service_tree = CFQ_RB_ROOT;
>>  
>>  	for_each_cfqg_st(cfqg, i, j, st)
>>  		*st = CFQ_RB_ROOT;
>> -	RB_CLEAR_NODE(&cfqg->rb_node);
>>  
>>  	/*
>>  	 * Take the initial reference that will be released on destroy
>> @@ -1023,11 +1168,102 @@ cfq_find_alloc_cfqg(struct cfq_data *cfqd, struct cgroup *cgroup, int create)
>>  	sscanf(dev_name(bdi->dev), "%u:%u", &major, &minor);
>>  	cfq_blkiocg_add_blkio_group(blkcg, &cfqg->blkg, (void *)cfqd,
>>  					MKDEV(major, minor));
>> -	cfqg->weight = blkcg_get_weight(blkcg, cfqg->blkg.dev);
>> -
>> +	init_group_queue_entity(blkcg, cfqg);
>>  	/* Add group on cfqd list */
>>  	hlist_add_head(&cfqg->cfqd_node, &cfqd->cfqg_list);
>> +}
>> +
>> +static void uninit_cfqg(struct cfq_data *cfqd, struct cfq_group *cfqg)
>> +{
>> +	if (!cfq_blkiocg_del_blkio_group(&cfqg->blkg))
>> +		cfq_destroy_cfqg(cfqd, cfqg);
>> +}
>> +
>> +static void cfqg_set_parent(struct cfq_group *cfqg, struct cfq_group *p_cfqg)
>> +{
>> +	struct io_sched_entity *gse = &cfqg->group_se;
>> +	struct io_sched_entity *qse = &cfqg->queue_se;
>> +	struct io_sched_entity *p_gse = &p_cfqg->group_se;
>> +
>> +	gse->parent = p_gse;
>> +	gse->st = &p_cfqg->grp_service_tree;
>> +
>> +	qse->parent = gse;
>> +	qse->st = &cfqg->grp_service_tree;
>> +
>> +	/* child reference */
>> +	atomic_inc(&p_cfqg->ref);
>> +}
>> +
>> +int cfqg_chain_alloc(struct cfq_data *cfqd, struct cgroup *cgroup)
>> +{
>> +	struct blkio_cgroup *blkcg = cgroup_to_blkio_cgroup(cgroup);
>> +	struct blkio_cgroup *p_blkcg;
>> +	struct backing_dev_info *bdi = &cfqd->queue->backing_dev_info;
>> +	unsigned int major, minor;
>> +	struct cfq_group *cfqg, *p_cfqg;
>> +	void *key = cfqd;
>> +	int ret;
>> +
>> +	cfqg = cfqg_of_blkg(blkiocg_lookup_group(blkcg, key));
>> +	if (cfqg) {
>> +		if (!cfqg->blkg.dev && bdi->dev && dev_name(bdi->dev)) {
>> +			sscanf(dev_name(bdi->dev), "%u:%u", &major, &minor);
>> +			cfqg->blkg.dev = MKDEV(major, minor);
>> +		}
>> +		/* chain has already been built */
>> +		return 0;
>> +	}
>> +
>> +	cfqg = kzalloc_node(sizeof(*cfqg), GFP_ATOMIC, cfqd->queue->node);
>> +	if (!cfqg)
>> +		return -1;
>> +
>> +	init_cfqg(cfqd, blkcg, cfqg);
>> +
>> +	/* Already to the top group */
>> +	if (!cgroup->parent)
>> +		return 0;
>> +
>> +	ret = cfqg_chain_alloc(cfqd, cgroup->parent);
>> +	if (ret == -1) {
>> +		uninit_cfqg(cfqd, cfqg);
>> +		return -1;
>> +	}
>> +
>> +	p_blkcg = cgroup_to_blkio_cgroup(cgroup->parent);
>> +	p_cfqg = cfqg_of_blkg(blkiocg_lookup_group(p_blkcg, key));
>> +	BUG_ON(p_cfqg == NULL);
>> +
>> +	cfqg_set_parent(cfqg, p_cfqg);
>> +	return 0;
>> +}
>> +
>> +static struct cfq_group *
>> +cfq_find_alloc_cfqg(struct cfq_data *cfqd, struct cgroup *cgroup, int create)
>> +{
>> +	struct blkio_cgroup *blkcg = cgroup_to_blkio_cgroup(cgroup);
>> +	struct cfq_group *cfqg = NULL;
>> +	void *key = cfqd;
>> +	struct backing_dev_info *bdi = &cfqd->queue->backing_dev_info;
>> +	unsigned int major, minor;
>> +	int ret;
>>  
>> +	cfqg = cfqg_of_blkg(blkiocg_lookup_group(blkcg, key));
>> +	if (cfqg && !cfqg->blkg.dev && bdi->dev && dev_name(bdi->dev)) {
>> +		sscanf(dev_name(bdi->dev), "%u:%u", &major, &minor);
>> +		cfqg->blkg.dev = MKDEV(major, minor);
>> +		goto done;
>> +	}
>> +	if (cfqg || !create)
>> +		goto done;
>> +
>> +	ret = cfqg_chain_alloc(cfqd, cgroup);
>> +	if (!ret) {
>> +		cfqg = cfqg_of_blkg(blkiocg_lookup_group(blkcg, key));
>> +		BUG_ON(cfqg == NULL);
>> +		goto done;
>> +	}
>>  done:
>>  	return cfqg;
>>  }
>> @@ -1067,33 +1303,6 @@ static void cfq_link_cfqq_cfqg(struct cfq_queue *cfqq, struct cfq_group *cfqg)
>>  	atomic_inc(&cfqq->cfqg->ref);
>>  }
>>  
>> -static void cfq_put_cfqg(struct cfq_group *cfqg)
>> -{
>> -	struct cfq_rb_root *st;
>> -	int i, j;
>> -
>> -	BUG_ON(atomic_read(&cfqg->ref) <= 0);
>> -	if (!atomic_dec_and_test(&cfqg->ref))
>> -		return;
>> -	for_each_cfqg_st(cfqg, i, j, st)
>> -		BUG_ON(!RB_EMPTY_ROOT(&st->rb) || st->active != NULL);
>> -	kfree(cfqg);
>> -}
>> -
>> -static void cfq_destroy_cfqg(struct cfq_data *cfqd, struct cfq_group *cfqg)
>> -{
>> -	/* Something wrong if we are trying to remove same group twice */
>> -	BUG_ON(hlist_unhashed(&cfqg->cfqd_node));
>> -
>> -	hlist_del_init(&cfqg->cfqd_node);
>> -
>> -	/*
>> -	 * Put the reference taken at the time of creation so that when all
>> -	 * queues are gone, group can be destroyed.
>> -	 */
>> -	cfq_put_cfqg(cfqg);
>> -}
>> -
>>  static void cfq_release_cfq_groups(struct cfq_data *cfqd)
>>  {
>>  	struct hlist_node *pos, *n;
>> @@ -1668,9 +1877,6 @@ __cfq_slice_expired(struct cfq_data *cfqd, struct cfq_queue *cfqq,
>>  	if (cfqq == cfqd->active_queue)
>>  		cfqd->active_queue = NULL;
>>  
>> -	if (&cfqq->cfqg->rb_node == cfqd->grp_service_tree.active)
>> -		cfqd->grp_service_tree.active = NULL;
>> -
>>  	if (cfqd->active_cic) {
>>  		put_io_context(cfqd->active_cic->ioc);
>>  		cfqd->active_cic = NULL;
>> @@ -2173,17 +2379,26 @@ static void choose_service_tree(struct cfq_data *cfqd, struct cfq_group *cfqg)
>>  	cfqd->noidle_tree_requires_idle = false;
>>  }
>>  
>> +
>>  static struct cfq_group *cfq_get_next_cfqg(struct cfq_data *cfqd)
>>  {
>> -	struct cfq_rb_root *st = &cfqd->grp_service_tree;
>> +	struct cfq_group *root_group = &cfqd->root_group;
>> +	struct cfq_rb_root *st = &root_group->grp_service_tree;
>>  	struct cfq_group *cfqg;
>> +	struct io_sched_entity *se;
>>  
>> -	if (RB_EMPTY_ROOT(&st->rb))
>> -		return NULL;
>> -	cfqg = cfq_rb_first_group(st);
>> -	st->active = &cfqg->rb_node;
>> -	update_min_vdisktime(st);
>> -	return cfqg;
>> +	do {
>> +		se = cfq_rb_first_se(st);
>> +		if (!se)
>> +			return NULL;
>> +		st->active = &se->rb_node;
>> +		update_min_vdisktime(st);
>> +		cfqg = cfqg_of_qse(se);
>> +		if (cfqg)
>> +			return cfqg;
>> +		cfqg = cfqg_of_gse(se);
>> +		st = &cfqg->grp_service_tree;
>> +	} while (1);
>>  }
>>  
>>  static void cfq_choose_cfqg(struct cfq_data *cfqd)
>> @@ -2215,15 +2430,18 @@ static struct cfq_queue *cfq_select_queue(struct cfq_data *cfqd)
>>  	if (!cfqq)
>>  		goto new_queue;
>>  
>> +
>>  	if (!cfqd->rq_queued)
>>  		return NULL;
>>  
>> +
>>  	/*
>>  	 * We were waiting for group to get backlogged. Expire the queue
>>  	 */
>>  	if (cfq_cfqq_wait_busy(cfqq) && !RB_EMPTY_ROOT(&cfqq->sort_list))
>>  		goto expire;
>>  
>> +
>>  	/*
>>  	 * The active queue has run out of time, expire it and select new.
>>  	 */
>> @@ -2245,6 +2463,7 @@ static struct cfq_queue *cfq_select_queue(struct cfq_data *cfqd)
>>  			goto check_group_idle;
>>  	}
>>  
>> +
>>  	/*
>>  	 * The active queue has requests and isn't expired, allow it to
>>  	 * dispatch.
>> @@ -2252,6 +2471,7 @@ static struct cfq_queue *cfq_select_queue(struct cfq_data *cfqd)
>>  	if (!RB_EMPTY_ROOT(&cfqq->sort_list))
>>  		goto keep_queue;
>>  
>> +
>>  	/*
>>  	 * If another queue has a request waiting within our mean seek
>>  	 * distance, let it run.  The expire code will check for close
>> @@ -2505,6 +2725,7 @@ static int cfq_dispatch_requests(struct request_queue *q, int force)
>>  	}
>>  
>>  	cfq_log_cfqq(cfqd, cfqq, "dispatched a request");
>> +
>>  	return 1;
>>  }
>>  
>> @@ -3854,17 +4075,25 @@ static void *cfq_init_queue(struct request_queue *q)
>>  
>>  	cfqd->cic_index = i;
>>  
>> -	/* Init root service tree */
>> -	cfqd->grp_service_tree = CFQ_RB_ROOT;
>> -
>>  	/* Init root group */
>>  	cfqg = &cfqd->root_group;
>> +	cfqg->grp_service_tree = CFQ_RB_ROOT;
>>  	for_each_cfqg_st(cfqg, i, j, st)
>>  		*st = CFQ_RB_ROOT;
>> -	RB_CLEAR_NODE(&cfqg->rb_node);
>> -
>> +	cfqg->group_se.is_group_se = true;
>> +	RB_CLEAR_NODE(&cfqg->group_se.rb_node);
>> +	cfqg->group_se.on_st = false;
>>  	/* Give preference to root group over other groups */
>> -	cfqg->weight = 2*BLKIO_WEIGHT_DEFAULT;
>> +	cfqg->group_se.weight = 2*BLKIO_WEIGHT_DEFAULT;
>> +	cfqg->group_se.parent = NULL;
>> +	cfqg->group_se.st = NULL;
>> +
>> +	cfqg->queue_se.is_group_se = false;
>> +	RB_CLEAR_NODE(&cfqg->queue_se.rb_node);
>> +	cfqg->queue_se.on_st = false;
>> +	cfqg->queue_se.weight = 2*BLKIO_WEIGHT_DEFAULT;
>> +	cfqg->queue_se.parent = &cfqg->group_se;
>> +	cfqg->queue_se.st = &cfqg->grp_service_tree;
>>  
>>  #ifdef CONFIG_CFQ_GROUP_IOSCHED
>>  	/*
>> -- 
>> 1.6.5.2
>>
>>
> 
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ