[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20150420153257.GE17020@quack.suse.cz>
Date: Mon, 20 Apr 2015 17:32:57 +0200
From: Jan Kara <jack@...e.cz>
To: Tejun Heo <tj@...nel.org>
Cc: axboe@...nel.dk, linux-kernel@...r.kernel.org, jack@...e.cz,
hch@...radead.org, hannes@...xchg.org,
linux-fsdevel@...r.kernel.org, vgoyal@...hat.com,
lizefan@...wei.com, cgroups@...r.kernel.org, linux-mm@...ck.org,
mhocko@...e.cz, clm@...com, fengguang.wu@...el.com,
david@...morbit.com, gthelen@...gle.com, drbd-dev@...ts.linbit.com,
Neil Brown <neilb@...e.de>, Alasdair Kergon <agk@...hat.com>,
Mike Snitzer <snitzer@...hat.com>
Subject: Re: [PATCH 11/49] writeback: move backing_dev_info->state into
bdi_writeback
On Mon 06-04-15 15:58:00, Tejun Heo wrote:
> Currently, a bdi (backing_dev_info) embeds single wb (bdi_writeback)
> and the role of the separation is unclear. For cgroup support for
> writeback IOs, a bdi will be updated to host multiple wb's where each
> wb serves writeback IOs of a different cgroup on the bdi. To achieve
> that, a wb should carry all states necessary for servicing writeback
> IOs for a cgroup independently.
>
> This patch moves bdi->state into wb.
>
> * enum bdi_state is renamed to wb_state and the prefix of all enums is
> changed from BDI_ to WB_.
>
> * Explicit zeroing of bdi->state is removed without adding zeoring of
> wb->state as the whole data structure is zeroed on init anyway.
>
> * As there's still only one bdi_writeback per backing_dev_info, all
> uses of bdi->state are mechanically replaced with bdi->wb.state
> introducing no behavior changes.
Looks good. You can add:
Reviewed-by: Jan Kara <jack@...e.cz>
Honza
>
> Signed-off-by: Tejun Heo <tj@...nel.org>
> Cc: Jens Axboe <axboe@...nel.dk>
> Cc: Jan Kara <jack@...e.cz>
> Cc: Wu Fengguang <fengguang.wu@...el.com>
> Cc: drbd-dev@...ts.linbit.com
> Cc: Neil Brown <neilb@...e.de>
> Cc: Alasdair Kergon <agk@...hat.com>
> Cc: Mike Snitzer <snitzer@...hat.com>
> ---
> block/blk-core.c | 1 -
> drivers/block/drbd/drbd_main.c | 10 +++++-----
> drivers/md/dm.c | 2 +-
> drivers/md/raid1.c | 4 ++--
> drivers/md/raid10.c | 2 +-
> fs/fs-writeback.c | 14 +++++++-------
> include/linux/backing-dev.h | 24 ++++++++++++------------
> mm/backing-dev.c | 20 ++++++++++----------
> 8 files changed, 38 insertions(+), 39 deletions(-)
>
> diff --git a/block/blk-core.c b/block/blk-core.c
> index 56da125..fa1314e 100644
> --- a/block/blk-core.c
> +++ b/block/blk-core.c
> @@ -606,7 +606,6 @@ struct request_queue *blk_alloc_queue_node(gfp_t gfp_mask, int node_id)
>
> q->backing_dev_info.ra_pages =
> (VM_MAX_READAHEAD * 1024) / PAGE_CACHE_SIZE;
> - q->backing_dev_info.state = 0;
> q->backing_dev_info.capabilities = 0;
> q->backing_dev_info.name = "block";
> q->node = node_id;
> diff --git a/drivers/block/drbd/drbd_main.c b/drivers/block/drbd/drbd_main.c
> index 1fc8342..61b00aa 100644
> --- a/drivers/block/drbd/drbd_main.c
> +++ b/drivers/block/drbd/drbd_main.c
> @@ -2360,7 +2360,7 @@ static void drbd_cleanup(void)
> * @congested_data: User data
> * @bdi_bits: Bits the BDI flusher thread is currently interested in
> *
> - * Returns 1<<BDI_async_congested and/or 1<<BDI_sync_congested if we are congested.
> + * Returns 1<<WB_async_congested and/or 1<<WB_sync_congested if we are congested.
> */
> static int drbd_congested(void *congested_data, int bdi_bits)
> {
> @@ -2377,14 +2377,14 @@ static int drbd_congested(void *congested_data, int bdi_bits)
> }
>
> if (test_bit(CALLBACK_PENDING, &first_peer_device(device)->connection->flags)) {
> - r |= (1 << BDI_async_congested);
> + r |= (1 << WB_async_congested);
> /* Without good local data, we would need to read from remote,
> * and that would need the worker thread as well, which is
> * currently blocked waiting for that usermode helper to
> * finish.
> */
> if (!get_ldev_if_state(device, D_UP_TO_DATE))
> - r |= (1 << BDI_sync_congested);
> + r |= (1 << WB_sync_congested);
> else
> put_ldev(device);
> r &= bdi_bits;
> @@ -2400,9 +2400,9 @@ static int drbd_congested(void *congested_data, int bdi_bits)
> reason = 'b';
> }
>
> - if (bdi_bits & (1 << BDI_async_congested) &&
> + if (bdi_bits & (1 << WB_async_congested) &&
> test_bit(NET_CONGESTED, &first_peer_device(device)->connection->flags)) {
> - r |= (1 << BDI_async_congested);
> + r |= (1 << WB_async_congested);
> reason = reason == 'b' ? 'a' : 'n';
> }
>
> diff --git a/drivers/md/dm.c b/drivers/md/dm.c
> index 73f2880..c076982 100644
> --- a/drivers/md/dm.c
> +++ b/drivers/md/dm.c
> @@ -2039,7 +2039,7 @@ static int dm_any_congested(void *congested_data, int bdi_bits)
> * the query about congestion status of request_queue
> */
> if (dm_request_based(md))
> - r = md->queue->backing_dev_info.state &
> + r = md->queue->backing_dev_info.wb.state &
> bdi_bits;
> else
> r = dm_table_any_congested(map, bdi_bits);
> diff --git a/drivers/md/raid1.c b/drivers/md/raid1.c
> index d34e238..2fca392 100644
> --- a/drivers/md/raid1.c
> +++ b/drivers/md/raid1.c
> @@ -739,7 +739,7 @@ static int raid1_congested(struct mddev *mddev, int bits)
> struct r1conf *conf = mddev->private;
> int i, ret = 0;
>
> - if ((bits & (1 << BDI_async_congested)) &&
> + if ((bits & (1 << WB_async_congested)) &&
> conf->pending_count >= max_queued_requests)
> return 1;
>
> @@ -754,7 +754,7 @@ static int raid1_congested(struct mddev *mddev, int bits)
> /* Note the '|| 1' - when read_balance prefers
> * non-congested targets, it can be removed
> */
> - if ((bits & (1<<BDI_async_congested)) || 1)
> + if ((bits & (1 << WB_async_congested)) || 1)
> ret |= bdi_congested(&q->backing_dev_info, bits);
> else
> ret &= bdi_congested(&q->backing_dev_info, bits);
> diff --git a/drivers/md/raid10.c b/drivers/md/raid10.c
> index a7196c4..8d38857 100644
> --- a/drivers/md/raid10.c
> +++ b/drivers/md/raid10.c
> @@ -914,7 +914,7 @@ static int raid10_congested(struct mddev *mddev, int bits)
> struct r10conf *conf = mddev->private;
> int i, ret = 0;
>
> - if ((bits & (1 << BDI_async_congested)) &&
> + if ((bits & (1 << WB_async_congested)) &&
> conf->pending_count >= max_queued_requests)
> return 1;
>
> diff --git a/fs/fs-writeback.c b/fs/fs-writeback.c
> index e907052..fea13fe 100644
> --- a/fs/fs-writeback.c
> +++ b/fs/fs-writeback.c
> @@ -62,7 +62,7 @@ struct wb_writeback_work {
> */
> int writeback_in_progress(struct backing_dev_info *bdi)
> {
> - return test_bit(BDI_writeback_running, &bdi->state);
> + return test_bit(WB_writeback_running, &bdi->wb.state);
> }
> EXPORT_SYMBOL(writeback_in_progress);
>
> @@ -100,7 +100,7 @@ EXPORT_TRACEPOINT_SYMBOL_GPL(wbc_writepage);
> static void bdi_wakeup_thread(struct backing_dev_info *bdi)
> {
> spin_lock_bh(&bdi->wb_lock);
> - if (test_bit(BDI_registered, &bdi->state))
> + if (test_bit(WB_registered, &bdi->wb.state))
> mod_delayed_work(bdi_wq, &bdi->wb.dwork, 0);
> spin_unlock_bh(&bdi->wb_lock);
> }
> @@ -111,7 +111,7 @@ static void bdi_queue_work(struct backing_dev_info *bdi,
> trace_writeback_queue(bdi, work);
>
> spin_lock_bh(&bdi->wb_lock);
> - if (!test_bit(BDI_registered, &bdi->state)) {
> + if (!test_bit(WB_registered, &bdi->wb.state)) {
> if (work->done)
> complete(work->done);
> goto out_unlock;
> @@ -1039,7 +1039,7 @@ static long wb_do_writeback(struct bdi_writeback *wb)
> struct wb_writeback_work *work;
> long wrote = 0;
>
> - set_bit(BDI_writeback_running, &wb->bdi->state);
> + set_bit(WB_writeback_running, &wb->state);
> while ((work = get_next_work_item(bdi)) != NULL) {
>
> trace_writeback_exec(bdi, work);
> @@ -1061,7 +1061,7 @@ static long wb_do_writeback(struct bdi_writeback *wb)
> */
> wrote += wb_check_old_data_flush(wb);
> wrote += wb_check_background_flush(wb);
> - clear_bit(BDI_writeback_running, &wb->bdi->state);
> + clear_bit(WB_writeback_running, &wb->state);
>
> return wrote;
> }
> @@ -1081,7 +1081,7 @@ void bdi_writeback_workfn(struct work_struct *work)
> current->flags |= PF_SWAPWRITE;
>
> if (likely(!current_is_workqueue_rescuer() ||
> - !test_bit(BDI_registered, &bdi->state))) {
> + !test_bit(WB_registered, &wb->state))) {
> /*
> * The normal path. Keep writing back @bdi until its
> * work_list is empty. Note that this path is also taken
> @@ -1255,7 +1255,7 @@ void __mark_inode_dirty(struct inode *inode, int flags)
> spin_unlock(&inode->i_lock);
> spin_lock(&bdi->wb.list_lock);
> if (bdi_cap_writeback_dirty(bdi)) {
> - WARN(!test_bit(BDI_registered, &bdi->state),
> + WARN(!test_bit(WB_registered, &bdi->wb.state),
> "bdi-%s not registered\n", bdi->name);
>
> /*
> diff --git a/include/linux/backing-dev.h b/include/linux/backing-dev.h
> index aff923a..eb14f98 100644
> --- a/include/linux/backing-dev.h
> +++ b/include/linux/backing-dev.h
> @@ -25,13 +25,13 @@ struct device;
> struct dentry;
>
> /*
> - * Bits in backing_dev_info.state
> + * Bits in bdi_writeback.state
> */
> -enum bdi_state {
> - BDI_async_congested, /* The async (write) queue is getting full */
> - BDI_sync_congested, /* The sync queue is getting full */
> - BDI_registered, /* bdi_register() was done */
> - BDI_writeback_running, /* Writeback is in progress */
> +enum wb_state {
> + WB_async_congested, /* The async (write) queue is getting full */
> + WB_sync_congested, /* The sync queue is getting full */
> + WB_registered, /* bdi_register() was done */
> + WB_writeback_running, /* Writeback is in progress */
> };
>
> typedef int (congested_fn)(void *, int);
> @@ -49,6 +49,7 @@ enum bdi_stat_item {
> struct bdi_writeback {
> struct backing_dev_info *bdi; /* our parent bdi */
>
> + unsigned long state; /* Always use atomic bitops on this */
> unsigned long last_old_flush; /* last old data flush */
>
> struct delayed_work dwork; /* work item used for writeback */
> @@ -62,7 +63,6 @@ struct bdi_writeback {
> struct backing_dev_info {
> struct list_head bdi_list;
> unsigned long ra_pages; /* max readahead in PAGE_CACHE_SIZE units */
> - unsigned long state; /* Always use atomic bitops on this */
> unsigned int capabilities; /* Device capabilities */
> congested_fn *congested_fn; /* Function pointer if device is md/dm */
> void *congested_data; /* Pointer to aux data for congested func */
> @@ -250,23 +250,23 @@ static inline int bdi_congested(struct backing_dev_info *bdi, int bdi_bits)
> {
> if (bdi->congested_fn)
> return bdi->congested_fn(bdi->congested_data, bdi_bits);
> - return (bdi->state & bdi_bits);
> + return (bdi->wb.state & bdi_bits);
> }
>
> static inline int bdi_read_congested(struct backing_dev_info *bdi)
> {
> - return bdi_congested(bdi, 1 << BDI_sync_congested);
> + return bdi_congested(bdi, 1 << WB_sync_congested);
> }
>
> static inline int bdi_write_congested(struct backing_dev_info *bdi)
> {
> - return bdi_congested(bdi, 1 << BDI_async_congested);
> + return bdi_congested(bdi, 1 << WB_async_congested);
> }
>
> static inline int bdi_rw_congested(struct backing_dev_info *bdi)
> {
> - return bdi_congested(bdi, (1 << BDI_sync_congested) |
> - (1 << BDI_async_congested));
> + return bdi_congested(bdi, (1 << WB_sync_congested) |
> + (1 << WB_async_congested));
> }
>
> enum {
> diff --git a/mm/backing-dev.c b/mm/backing-dev.c
> index 6dc4580..b23cf0e 100644
> --- a/mm/backing-dev.c
> +++ b/mm/backing-dev.c
> @@ -96,7 +96,7 @@ static int bdi_debug_stats_show(struct seq_file *m, void *v)
> nr_io,
> nr_more_io,
> nr_dirty_time,
> - !list_empty(&bdi->bdi_list), bdi->state);
> + !list_empty(&bdi->bdi_list), bdi->wb.state);
> #undef K
>
> return 0;
> @@ -280,7 +280,7 @@ void bdi_wakeup_thread_delayed(struct backing_dev_info *bdi)
>
> timeout = msecs_to_jiffies(dirty_writeback_interval * 10);
> spin_lock_bh(&bdi->wb_lock);
> - if (test_bit(BDI_registered, &bdi->state))
> + if (test_bit(WB_registered, &bdi->wb.state))
> queue_delayed_work(bdi_wq, &bdi->wb.dwork, timeout);
> spin_unlock_bh(&bdi->wb_lock);
> }
> @@ -315,7 +315,7 @@ int bdi_register(struct backing_dev_info *bdi, struct device *parent,
> bdi->dev = dev;
>
> bdi_debug_register(bdi, dev_name(dev));
> - set_bit(BDI_registered, &bdi->state);
> + set_bit(WB_registered, &bdi->wb.state);
>
> spin_lock_bh(&bdi_lock);
> list_add_tail_rcu(&bdi->bdi_list, &bdi_list);
> @@ -339,7 +339,7 @@ static void bdi_wb_shutdown(struct backing_dev_info *bdi)
> {
> /* Make sure nobody queues further work */
> spin_lock_bh(&bdi->wb_lock);
> - if (!test_and_clear_bit(BDI_registered, &bdi->state)) {
> + if (!test_and_clear_bit(WB_registered, &bdi->wb.state)) {
> spin_unlock_bh(&bdi->wb_lock);
> return;
> }
> @@ -492,11 +492,11 @@ static atomic_t nr_bdi_congested[2];
>
> void clear_bdi_congested(struct backing_dev_info *bdi, int sync)
> {
> - enum bdi_state bit;
> + enum wb_state bit;
> wait_queue_head_t *wqh = &congestion_wqh[sync];
>
> - bit = sync ? BDI_sync_congested : BDI_async_congested;
> - if (test_and_clear_bit(bit, &bdi->state))
> + bit = sync ? WB_sync_congested : WB_async_congested;
> + if (test_and_clear_bit(bit, &bdi->wb.state))
> atomic_dec(&nr_bdi_congested[sync]);
> smp_mb__after_atomic();
> if (waitqueue_active(wqh))
> @@ -506,10 +506,10 @@ EXPORT_SYMBOL(clear_bdi_congested);
>
> void set_bdi_congested(struct backing_dev_info *bdi, int sync)
> {
> - enum bdi_state bit;
> + enum wb_state bit;
>
> - bit = sync ? BDI_sync_congested : BDI_async_congested;
> - if (!test_and_set_bit(bit, &bdi->state))
> + bit = sync ? WB_sync_congested : WB_async_congested;
> + if (!test_and_set_bit(bit, &bdi->wb.state))
> atomic_inc(&nr_bdi_congested[sync]);
> }
> EXPORT_SYMBOL(set_bdi_congested);
> --
> 2.1.0
>
--
Jan Kara <jack@...e.cz>
SUSE Labs, CR
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists