Try to limit the dirty throttle pause time in range (10ms, 100ms), by controlling how many pages are dirtied before doing a throttle pause. The dirty count will be directly billed to the task struct. Slow start and quick back off is employed, so that the stable range will be biased towards 10ms. Another intention is for fine timing control of slow devices, which may need to pause for 100ms for several pages. Signed-off-by: Wu Fengguang --- include/linux/sched.h | 7 +++++++ mm/page-writeback.c | 31 ++++++++++++++----------------- 2 files changed, 21 insertions(+), 17 deletions(-) --- linux-next.orig/include/linux/sched.h 2010-09-12 13:10:54.000000000 +0800 +++ linux-next/include/linux/sched.h 2010-09-12 13:16:20.000000000 +0800 @@ -1455,6 +1455,13 @@ struct task_struct { int make_it_fail; #endif struct prop_local_single dirties; + /* + * when (nr_dirtied >= nr_dirtied_pause), it's time to call + * balance_dirty_pages() for some dirty throttling pause + */ + int nr_dirtied; + int nr_dirtied_pause; + #ifdef CONFIG_LATENCYTOP int latency_record_count; struct latency_record latency_record[LT_SAVECOUNT]; --- linux-next.orig/mm/page-writeback.c 2010-09-12 13:10:54.000000000 +0800 +++ linux-next/mm/page-writeback.c 2010-09-12 13:12:48.000000000 +0800 @@ -529,6 +529,12 @@ static void balance_dirty_pages(struct a pause = HZ * (pages_dirtied << PAGE_CACHE_SHIFT) / (bw + 1); pause = clamp_val(pause, 1, HZ/5); + if (pause > HZ/10) { + current->nr_dirtied_pause >>= 1; + current->nr_dirtied_pause++; + } else if (pause < HZ/100) + current->nr_dirtied_pause++; + __set_current_state(TASK_INTERRUPTIBLE); io_schedule_timeout(pause); @@ -570,8 +576,6 @@ void set_page_dirty_balance(struct page } } -static DEFINE_PER_CPU(unsigned long, bdp_ratelimits) = 0; - /** * balance_dirty_pages_ratelimited_nr - balance dirty memory state * @mapping: address_space which was dirtied @@ -589,28 +593,21 @@ static DEFINE_PER_CPU(unsigned long, bdp void balance_dirty_pages_ratelimited_nr(struct address_space *mapping, unsigned long nr_pages_dirtied) { - unsigned long ratelimit; - unsigned long *p; + if (!current->nr_dirtied_pause) + current->nr_dirtied_pause = + mapping->backing_dev_info->dirty_exceeded ? + 8 : ratelimit_pages; - ratelimit = ratelimit_pages; - if (mapping->backing_dev_info->dirty_exceeded) - ratelimit = 8; + current->nr_dirtied += nr_pages_dirtied; /* * Check the rate limiting. Also, we do not want to throttle real-time * tasks in balance_dirty_pages(). Period. */ - preempt_disable(); - p = &__get_cpu_var(bdp_ratelimits); - *p += nr_pages_dirtied; - if (unlikely(*p >= ratelimit)) { - ratelimit = *p; - *p = 0; - preempt_enable(); - balance_dirty_pages(mapping, ratelimit); - return; + if (current->nr_dirtied >= current->nr_dirtied_pause) { + balance_dirty_pages(mapping, current->nr_dirtied); + current->nr_dirtied = 0; } - preempt_enable(); } EXPORT_SYMBOL(balance_dirty_pages_ratelimited_nr); -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/