[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20100426171800.GB3372@redhat.com>
Date: Mon, 26 Apr 2010 13:18:00 -0400
From: Vivek Goyal <vgoyal@...hat.com>
To: linux kernel mailing list <linux-kernel@...r.kernel.org>,
Jens Axboe <jens.axboe@...cle.com>
Cc: Divyesh Shah <dpshah@...gle.com>,
Nauman Rafique <nauman@...gle.com>,
Gui Jianfeng <guijianfeng@...fujitsu.com>
Subject: Re: [PATCH] blkio: Fix another BUG_ON() crash due to cfqq movement
across groups
On Wed, Apr 21, 2010 at 06:19:06PM -0400, Vivek Goyal wrote:
> o Once in a while, I was hitting a BUG_ON() in blkio code. empty_time was
> assuming that upon slice expiry, group can't be marked empty already (except
> forced dispatch).
>
Hi Jens,
Do you have any concerns with this patch? If not, can you please include
it.
Thanks
Vivek
> But this assumption is broken if cfqq can move (group_isolation=0) across
> groups after receiving a request.
>
> I think most likely in this case we got a request in a cfqq and accounted
> the rq in one group, later while adding the cfqq to tree, we moved the queue
> to a different group which was already marked empty and after dispatch from
> slice we found group already marked empty and raised alarm.
>
> This patch does not error out if group is already marked empty. This can
> introduce some empty_time stat error only in case of group_isolation=0. This
> is better than crashing. In case of group_isolation=1 we should still get
> same stats as before this patch.
>
> [ 222.308546] ------------[ cut here ]------------
> [ 222.309311] kernel BUG at block/blk-cgroup.c:236!
> [ 222.309311] invalid opcode: 0000 [#1] SMP
> [ 222.309311] last sysfs file: /sys/devices/virtual/block/dm-3/queue/scheduler
> [ 222.309311] CPU 1
> [ 222.309311] Modules linked in: dm_round_robin dm_multipath qla2xxx scsi_transport_fc dm_zero dm_mirror dm_region_hash dm_log dm_mod [last unloaded: scsi_wait_scan]
> [ 222.309311]
> [ 222.309311] Pid: 4780, comm: fio Not tainted 2.6.34-rc4-blkio-config #68 0A98h/HP xw8600 Workstation
> [ 222.309311] RIP: 0010:[<ffffffff8121ad88>] [<ffffffff8121ad88>] blkiocg_set_start_empty_time+0x50/0x83
> [ 222.309311] RSP: 0018:ffff8800ba6e79f8 EFLAGS: 00010002
> [ 222.309311] RAX: 0000000000000082 RBX: ffff8800a13b7990 RCX: ffff8800a13b7808
> [ 222.309311] RDX: 0000000000002121 RSI: 0000000000000082 RDI: ffff8800a13b7a30
> [ 222.309311] RBP: ffff8800ba6e7a18 R08: 0000000000000000 R09: 0000000000000001
> [ 222.309311] R10: 000000000002f8c8 R11: ffff8800ba6e7ad8 R12: ffff8800a13b78ff
> [ 222.309311] R13: ffff8800a13b7990 R14: 0000000000000001 R15: ffff8800a13b7808
> [ 222.309311] FS: 00007f3beec476f0(0000) GS:ffff880001e40000(0000) knlGS:0000000000000000
> [ 222.309311] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> [ 222.309311] CR2: 000000000040e7f0 CR3: 00000000a12d5000 CR4: 00000000000006e0
> [ 222.309311] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
> [ 222.309311] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
> [ 222.309311] Process fio (pid: 4780, threadinfo ffff8800ba6e6000, task ffff8800b3d6bf00)
> [ 222.309311] Stack:
> [ 222.309311] 0000000000000001 ffff8800bab17a48 ffff8800bab17a48 ffff8800a13b7800
> [ 222.309311] <0> ffff8800ba6e7a68 ffffffff8121da35 ffff880000000001 00ff8800ba5c5698
> [ 222.309311] <0> ffff8800ba6e7a68 ffff8800a13b7800 0000000000000000 ffff8800bab17a48
> [ 222.309311] Call Trace:
> [ 222.309311] [<ffffffff8121da35>] __cfq_slice_expired+0x2af/0x3ec
> [ 222.309311] [<ffffffff8121fd7b>] cfq_dispatch_requests+0x2c8/0x8e8
> [ 222.309311] [<ffffffff8120f1cd>] ? spin_unlock_irqrestore+0xe/0x10
> [ 222.309311] [<ffffffff8120fb1a>] ? blk_insert_cloned_request+0x70/0x7b
> [ 222.309311] [<ffffffff81210461>] blk_peek_request+0x191/0x1a7
> [ 222.309311] [<ffffffffa0002799>] dm_request_fn+0x38/0x14c [dm_mod]
> [ 222.309311] [<ffffffff810ae61f>] ? sync_page_killable+0x0/0x35
> [ 222.309311] [<ffffffff81210fd4>] __generic_unplug_device+0x32/0x37
> [ 222.309311] [<ffffffff81211274>] generic_unplug_device+0x2e/0x3c
> [ 222.309311] [<ffffffffa00011a6>] dm_unplug_all+0x42/0x5b [dm_mod]
> [ 222.309311] [<ffffffff8120ca37>] blk_unplug+0x29/0x2d
> [ 222.309311] [<ffffffff8120ca4d>] blk_backing_dev_unplug+0x12/0x14
> [ 222.309311] [<ffffffff81109a7a>] block_sync_page+0x35/0x39
> [ 222.309311] [<ffffffff810ae616>] sync_page+0x41/0x4a
> [ 222.309311] [<ffffffff810ae62d>] sync_page_killable+0xe/0x35
> [ 222.309311] [<ffffffff8158aa59>] __wait_on_bit_lock+0x46/0x8f
> [ 222.309311] [<ffffffff810ae4f5>] __lock_page_killable+0x66/0x6d
> [ 222.309311] [<ffffffff81056f9c>] ? wake_bit_function+0x0/0x33
> [ 222.309311] [<ffffffff810ae528>] lock_page_killable+0x2c/0x2e
> [ 222.309311] [<ffffffff810afbc5>] generic_file_aio_read+0x361/0x4f0
> [ 222.309311] [<ffffffff810ea044>] do_sync_read+0xcb/0x108
> [ 222.309311] [<ffffffff811e42f7>] ? security_file_permission+0x16/0x18
> [ 222.309311] [<ffffffff810ea6ab>] vfs_read+0xab/0x108
> [ 222.309311] [<ffffffff810ea7c8>] sys_read+0x4a/0x6e
> [ 222.309311] [<ffffffff81002b5b>] system_call_fastpath+0x16/0x1b
> [ 222.309311] Code: 58 01 00 00 00 48 89 c6 75 0a 48 83 bb 60 01 00 00 00 74 09 48 8d bb a0 00 00 00 eb 35 41 fe cc 74 0d f6 83 c0 01 00 00 04 74 04 <0f> 0b eb fe 48 89 75 e8 e8 be e0 de ff 66 83 8b c0 01 00 00 04
> [ 222.309311] RIP [<ffffffff8121ad88>] blkiocg_set_start_empty_time+0x50/0x83
> [ 222.309311] RSP <ffff8800ba6e79f8>
> [ 222.309311] ---[ end trace 32b4f71dffc15712 ]---
>
> Signed-off-by: Vivek Goyal <vgoyal@...hat.com>
> ---
> block/blk-cgroup.c | 15 +++++++++------
> block/blk-cgroup.h | 5 ++---
> block/cfq-iosched.c | 33 ++++++++++++++++-----------------
> 3 files changed, 27 insertions(+), 26 deletions(-)
>
> diff --git a/block/blk-cgroup.c b/block/blk-cgroup.c
> index 83930f6..af42efb 100644
> --- a/block/blk-cgroup.c
> +++ b/block/blk-cgroup.c
> @@ -213,7 +213,7 @@ void blkiocg_update_avg_queue_size_stats(struct blkio_group *blkg)
> }
> EXPORT_SYMBOL_GPL(blkiocg_update_avg_queue_size_stats);
>
> -void blkiocg_set_start_empty_time(struct blkio_group *blkg, bool ignore)
> +void blkiocg_set_start_empty_time(struct blkio_group *blkg)
> {
> unsigned long flags;
> struct blkio_group_stats *stats;
> @@ -228,12 +228,15 @@ void blkiocg_set_start_empty_time(struct blkio_group *blkg, bool ignore)
> }
>
> /*
> - * If ignore is set, we do not panic on the empty flag being set
> - * already. This is to avoid cases where there are superfluous timeslice
> - * complete events (for eg., forced_dispatch in CFQ) when no IOs are
> - * served which could result in triggering the empty check incorrectly.
> + * group is already marked empty. This can happen if cfqq got new
> + * request in parent group and moved to this group while being added
> + * to service tree. Just ignore the event and move on.
> */
> - BUG_ON(!ignore && blkio_blkg_empty(stats));
> + if(blkio_blkg_empty(stats)) {
> + spin_unlock_irqrestore(&blkg->stats_lock, flags);
> + return;
> + }
> +
> stats->start_empty_time = sched_clock();
> blkio_mark_blkg_empty(stats);
> spin_unlock_irqrestore(&blkg->stats_lock, flags);
> diff --git a/block/blk-cgroup.h b/block/blk-cgroup.h
> index 2c956a0..a491a6d 100644
> --- a/block/blk-cgroup.h
> +++ b/block/blk-cgroup.h
> @@ -174,7 +174,7 @@ void blkiocg_update_dequeue_stats(struct blkio_group *blkg,
> unsigned long dequeue);
> void blkiocg_update_set_idle_time_stats(struct blkio_group *blkg);
> void blkiocg_update_idle_time_stats(struct blkio_group *blkg);
> -void blkiocg_set_start_empty_time(struct blkio_group *blkg, bool ignore);
> +void blkiocg_set_start_empty_time(struct blkio_group *blkg);
>
> #define BLKG_FLAG_FNS(name) \
> static inline void blkio_mark_blkg_##name( \
> @@ -205,8 +205,7 @@ static inline void blkiocg_update_dequeue_stats(struct blkio_group *blkg,
> static inline void blkiocg_update_set_idle_time_stats(struct blkio_group *blkg)
> {}
> static inline void blkiocg_update_idle_time_stats(struct blkio_group *blkg) {}
> -static inline void blkiocg_set_start_empty_time(struct blkio_group *blkg,
> - bool ignore) {}
> +static inline void blkiocg_set_start_empty_time(struct blkio_group *blkg) {}
> #endif
>
> #if defined(CONFIG_BLK_CGROUP) || defined(CONFIG_BLK_CGROUP_MODULE)
> diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c
> index d5927b5..002a5b6 100644
> --- a/block/cfq-iosched.c
> +++ b/block/cfq-iosched.c
> @@ -888,7 +888,7 @@ static inline unsigned int cfq_cfqq_slice_usage(struct cfq_queue *cfqq)
> }
>
> static void cfq_group_served(struct cfq_data *cfqd, struct cfq_group *cfqg,
> - struct cfq_queue *cfqq, bool forced)
> + struct cfq_queue *cfqq)
> {
> struct cfq_rb_root *st = &cfqd->grp_service_tree;
> unsigned int used_sl, charge_sl;
> @@ -918,7 +918,7 @@ static void cfq_group_served(struct cfq_data *cfqd, struct cfq_group *cfqg,
> cfq_log_cfqg(cfqd, cfqg, "served: vt=%llu min_vt=%llu", cfqg->vdisktime,
> st->min_vdisktime);
> blkiocg_update_timeslice_used(&cfqg->blkg, used_sl);
> - blkiocg_set_start_empty_time(&cfqg->blkg, forced);
> + blkiocg_set_start_empty_time(&cfqg->blkg);
> }
>
> #ifdef CONFIG_CFQ_GROUP_IOSCHED
> @@ -1582,7 +1582,7 @@ static void __cfq_set_active_queue(struct cfq_data *cfqd,
> */
> static void
> __cfq_slice_expired(struct cfq_data *cfqd, struct cfq_queue *cfqq,
> - bool timed_out, bool forced)
> + bool timed_out)
> {
> cfq_log_cfqq(cfqd, cfqq, "slice expired t=%d", timed_out);
>
> @@ -1609,7 +1609,7 @@ __cfq_slice_expired(struct cfq_data *cfqd, struct cfq_queue *cfqq,
> cfq_log_cfqq(cfqd, cfqq, "resid=%ld", cfqq->slice_resid);
> }
>
> - cfq_group_served(cfqd, cfqq->cfqg, cfqq, forced);
> + cfq_group_served(cfqd, cfqq->cfqg, cfqq);
>
> if (cfq_cfqq_on_rr(cfqq) && RB_EMPTY_ROOT(&cfqq->sort_list))
> cfq_del_cfqq_rr(cfqd, cfqq);
> @@ -1628,13 +1628,12 @@ __cfq_slice_expired(struct cfq_data *cfqd, struct cfq_queue *cfqq,
> }
> }
>
> -static inline void cfq_slice_expired(struct cfq_data *cfqd, bool timed_out,
> - bool forced)
> +static inline void cfq_slice_expired(struct cfq_data *cfqd, bool timed_out)
> {
> struct cfq_queue *cfqq = cfqd->active_queue;
>
> if (cfqq)
> - __cfq_slice_expired(cfqd, cfqq, timed_out, forced);
> + __cfq_slice_expired(cfqd, cfqq, timed_out);
> }
>
> /*
> @@ -2202,7 +2201,7 @@ static struct cfq_queue *cfq_select_queue(struct cfq_data *cfqd)
> }
>
> expire:
> - cfq_slice_expired(cfqd, 0, false);
> + cfq_slice_expired(cfqd, 0);
> new_queue:
> /*
> * Current queue expired. Check if we have to switch to a new
> @@ -2228,7 +2227,7 @@ static int __cfq_forced_dispatch_cfqq(struct cfq_queue *cfqq)
> BUG_ON(!list_empty(&cfqq->fifo));
>
> /* By default cfqq is not expired if it is empty. Do it explicitly */
> - __cfq_slice_expired(cfqq->cfqd, cfqq, 0, true);
> + __cfq_slice_expired(cfqq->cfqd, cfqq, 0);
> return dispatched;
> }
>
> @@ -2242,7 +2241,7 @@ static int cfq_forced_dispatch(struct cfq_data *cfqd)
> int dispatched = 0;
>
> /* Expire the timeslice of the current active queue first */
> - cfq_slice_expired(cfqd, 0, true);
> + cfq_slice_expired(cfqd, 0);
> while ((cfqq = cfq_get_next_queue_forced(cfqd)) != NULL) {
> __cfq_set_active_queue(cfqd, cfqq);
> dispatched += __cfq_forced_dispatch_cfqq(cfqq);
> @@ -2411,7 +2410,7 @@ static int cfq_dispatch_requests(struct request_queue *q, int force)
> cfqq->slice_dispatch >= cfq_prio_to_maxrq(cfqd, cfqq)) ||
> cfq_class_idle(cfqq))) {
> cfqq->slice_end = jiffies + 1;
> - cfq_slice_expired(cfqd, 0, false);
> + cfq_slice_expired(cfqd, 0);
> }
>
> cfq_log_cfqq(cfqd, cfqq, "dispatched a request");
> @@ -2442,7 +2441,7 @@ static void cfq_put_queue(struct cfq_queue *cfqq)
> orig_cfqg = cfqq->orig_cfqg;
>
> if (unlikely(cfqd->active_queue == cfqq)) {
> - __cfq_slice_expired(cfqd, cfqq, 0, false);
> + __cfq_slice_expired(cfqd, cfqq, 0);
> cfq_schedule_dispatch(cfqd);
> }
>
> @@ -2543,7 +2542,7 @@ static void cfq_exit_cfqq(struct cfq_data *cfqd, struct cfq_queue *cfqq)
> struct cfq_queue *__cfqq, *next;
>
> if (unlikely(cfqq == cfqd->active_queue)) {
> - __cfq_slice_expired(cfqd, cfqq, 0, false);
> + __cfq_slice_expired(cfqd, cfqq, 0);
> cfq_schedule_dispatch(cfqd);
> }
>
> @@ -3172,7 +3171,7 @@ cfq_should_preempt(struct cfq_data *cfqd, struct cfq_queue *new_cfqq,
> static void cfq_preempt_queue(struct cfq_data *cfqd, struct cfq_queue *cfqq)
> {
> cfq_log_cfqq(cfqd, cfqq, "preempt");
> - cfq_slice_expired(cfqd, 1, false);
> + cfq_slice_expired(cfqd, 1);
>
> /*
> * Put the new queue at the front of the of the current list,
> @@ -3383,7 +3382,7 @@ static void cfq_completed_request(struct request_queue *q, struct request *rq)
> * - when there is a close cooperator
> */
> if (cfq_slice_used(cfqq) || cfq_class_idle(cfqq))
> - cfq_slice_expired(cfqd, 1, false);
> + cfq_slice_expired(cfqd, 1);
> else if (sync && cfqq_empty &&
> !cfq_close_cooperator(cfqd, cfqq)) {
> cfqd->noidle_tree_requires_idle |= !rq_noidle(rq);
> @@ -3648,7 +3647,7 @@ static void cfq_idle_slice_timer(unsigned long data)
> cfq_clear_cfqq_deep(cfqq);
> }
> expire:
> - cfq_slice_expired(cfqd, timed_out, false);
> + cfq_slice_expired(cfqd, timed_out);
> out_kick:
> cfq_schedule_dispatch(cfqd);
> out_cont:
> @@ -3691,7 +3690,7 @@ static void cfq_exit_queue(struct elevator_queue *e)
> spin_lock_irq(q->queue_lock);
>
> if (cfqd->active_queue)
> - __cfq_slice_expired(cfqd, cfqd->active_queue, 0, false);
> + __cfq_slice_expired(cfqd, cfqd->active_queue, 0);
>
> while (!list_empty(&cfqd->cic_list)) {
> struct cfq_io_context *cic = list_entry(cfqd->cic_list.next,
> --
> 1.6.2.5
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists