lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <xr93ipw7qvkt.fsf@gthelen.mtv.corp.google.com>
Date:	Fri, 25 Feb 2011 20:16:34 -0800
From:	Greg Thelen <gthelen@...gle.com>
To:	Andrew Morton <akpm@...ux-foundation.org>
Cc:	linux-kernel@...r.kernel.org, linux-mm@...ck.org,
	containers@...ts.osdl.org, Andrea Righi <arighi@...eler.com>,
	Balbir Singh <balbir@...ux.vnet.ibm.com>,
	KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>,
	Daisuke Nishimura <nishimura@....nes.nec.co.jp>,
	Minchan Kim <minchan.kim@...il.com>,
	Ciju Rajan K <ciju@...ux.vnet.ibm.com>,
	David Rientjes <rientjes@...gle.com>,
	Wu Fengguang <fengguang.wu@...el.com>,
	Chad Talbott <ctalbott@...gle.com>,
	Justin TerAvest <teravest@...gle.com>,
	Vivek Goyal <vgoyal@...hat.com>
Subject: Re: [PATCH v5 0/9] memcg: per cgroup dirty page accounting

Greg Thelen <gthelen@...gle.com> writes:

> Changes since v4:
> - Moved documentation changes to start of series to provide a better
>   introduction to the series.
> - Added support for hierarchical dirty limits.
> - Incorporated bug fixes previously found in v4.
> - Include a new patch "writeback: convert variables to unsigned" to provide a
>   clearer transition to the the new dirty_info structure (patch "writeback:
>   create dirty_info structure").
> - Within the new dirty_info structure, replaced nr_reclaimable with
>   nr_file_dirty and nr_unstable_nfs to give callers finer grain dirty usage
>   information also added dirty_info_reclaimable().
> - Rebased the series to mmotm-2011-02-10-16-26 with two pending mmotm patches:
>   memcg: break out event counters from other stats
>     https://lkml.org/lkml/2011/2/17/415
>   memcg: use native word page statistics counters
>     https://lkml.org/lkml/2011/2/17/413

For those who prefer a monolithic patch, which applies to
mmotm-2011-02-10-16-26:
  http://www.kernel.org/pub/linux/kernel/people/gthelen/memcg/memcg-dirty-limits-v5-on-mmotm-2011-02-10-16-26.patch

The monolithic patch includes this -v4 series and the two above
mentioned memcg prerequisite patches.

> Changes since v3:
> - Refactored balance_dirty_pages() dirtying checking to use new struct
>   dirty_info, which is used to compare both system and memcg dirty limits
>   against usage.
> - Disabled memcg dirty limits when memory.use_hierarchy=1.  An enhancement is
>   needed to check the chain of parents to ensure that no dirty limit is
>   exceeded.
> - Ported to mmotm-2010-10-22-16-36.
>
> Changes since v2:
> - Rather than disabling softirq in lock_page_cgroup(), introduce a separate lock
>   to synchronize between memcg page accounting and migration.  This only affects
>   patch 4 of the series.  Patch 4 used to disable softirq, now it introduces the
>   new lock.
>
> Changes since v1:
> - Renamed "nfs"/"total_nfs" to "nfs_unstable"/"total_nfs_unstable" in per cgroup
>   memory.stat to match /proc/meminfo.
> - Avoid lockdep warnings by using rcu_read_[un]lock() in
>   mem_cgroup_has_dirty_limit().
> - Fixed lockdep issue in mem_cgroup_read_stat() which is exposed by these
>   patches.
> - Remove redundant comments.
> - Rename (for clarity):
>   - mem_cgroup_write_page_stat_item -> mem_cgroup_page_stat_item
>   - mem_cgroup_read_page_stat_item -> mem_cgroup_nr_pages_item
> - Renamed newly created proc files:
>   - memory.dirty_bytes -> memory.dirty_limit_in_bytes
>   - memory.dirty_background_bytes -> memory.dirty_background_limit_in_bytes
> - Removed unnecessary get_ prefix from get_xxx() functions.
> - Allow [kKmMgG] suffixes for newly created dirty limit value cgroupfs files.
> - Disable softirq rather than hardirq in lock_page_cgroup()
> - Made mem_cgroup_move_account_page_stat() inline.
> - Ported patches to mmotm-2010-10-13-17-13.
>
> This patch set provides the ability for each cgroup to have independent dirty
> page limits.
>
> Limiting dirty memory is like fixing the max amount of dirty (hard to reclaim)
> page cache used by a cgroup.  So, in case of multiple cgroup writers, they will
> not be able to consume more than their designated share of dirty pages and will
> be throttled if they cross that limit.
>
> Example use case:
>   #!/bin/bash
>   #
>   # Here is a test script that shows a situation where memcg dirty limits are
>   # beneficial.
>   #
>   # The script runs two programs:
>   # 1) a dirty page background antagonist (dd)
>   # 2) an interactive foreground process (tar).
>   #
>   # If the script's argument is false, then both processes are limited by the
>   # classic global dirty limits.  If the script is given a true argument, then a
>   # per-cgroup dirty limit is used to contain dd dirty page consumption.  The
>   # cgroup isolates the dd dirty memory consumption from the rest of the system
>   # processes (tar in this case).
>   #
>   # The time used by the tar process is printed (lower is better).
>   #
>   # When dd is run within a dirty limiting cgroup, the tar process had faster
>   # and more predictable performance.  memcg dirty ratios might be useful to
>   # serve different task classes (interactive vs batch).  A past discussion
>   # touched on this: http://lkml.org/lkml/2010/5/20/136
>   #
>   # When called with 'false' (using memcg without dirty isolation):
>   #   tar finished in 7.0s
>   #   dd reports 92 MB/s  
>   #
>   # When called with 'true' (using memcg for dirty isolation):
>   #   tar finished in 2.5s
>   #   dd reports 82 MB/s
>
>   echo memcg_dirty_limits: $1
>   
>   # set system dirty limits.
>   echo $((1<<30)) > /proc/sys/vm/dirty_bytes
>   echo $((1<<29)) > /proc/sys/vm/dirty_background_bytes
>   
>   mkdir /dev/cgroup/memory/A
>   
>   if $1; then    # if using cgroup to contain 'dd'...
>     echo 100M > /dev/cgroup/memory/A/memory.dirty_limit_in_bytes
>   fi
>   
>   # run antagonist (dd) in cgroup A
>   (echo $BASHPID > /dev/cgroup/memory/A/tasks; \
>    dd if=/dev/zero of=/disk1/big.file count=10k bs=1M) &
>   
>   # let antagonist (dd) get warmed up
>   sleep 10
>   
>   # time interactive job
>   time tar -C /disk2 -xzf linux.tar.gz
>   
>   wait
>   sleep 10
>   rmdir /dev/cgroup/memory/A
>
> The patches are based on a series proposed by Andrea Righi in Mar 2010.
>
> Overview:
> - Add page_cgroup flags to record when pages are dirty, in writeback, or nfs
>   unstable.
>
> - Extend mem_cgroup to record the total number of pages in each of the 
>   interesting dirty states (dirty, writeback, unstable_nfs).  
>
> - Add dirty parameters similar to the system-wide /proc/sys/vm/dirty_* limits to
>   mem_cgroup.  The mem_cgroup dirty parameters are accessible via cgroupfs
>   control files.
>
> - Consider both system and per-memcg dirty limits in page writeback when
>   deciding to queue background writeback or throttle dirty memory production.
>
> Known shortcomings (see the patch 1 update to Documentation/cgroups/memory.txt
> for more details):
> - When a cgroup dirty limit is exceeded, then bdi writeback is employed to
>   writeback dirty inodes.  Bdi writeback considers inodes from any cgroup, not
>   just inodes contributing dirty pages to the cgroup exceeding its limit.  
>
> - A cgroup may exceed its dirty limit if the memory is dirtied by a process in a
>   different memcg.
>
> Performance data:
> - A page fault microbenchmark workload was used to measure performance, which
>   can be called in read or write mode:
>         f = open(foo. $cpu)
>         truncate(f, 4096)
>         alarm(60)
>         while (1) {
>                 p = mmap(f, 4096)
>                 if (write)
> 			*p = 1
> 		else
> 			x = *p
>                 munmap(p)
>         }
>
> - The workload was called for several points in the patch series in different
>   modes:
>   - s_read is a single threaded reader
>   - s_write is a single threaded writer
>   - p_read is a 16 thread reader, each operating on a different file
>   - p_write is a 16 thread writer, each operating on a different file
>
> - Measurements were collected on a 16 core non-numa system using "perf stat
>   --repeat 3".
>
> - All numbers are page fault rate (M/sec).  Higher is better.
>
> - To compare the performance of a kernel without memcg compare the first and
>   last rows - neither has memcg configured.  The first row does not include any
>   of these memcg dirty limit patches.
>
> - To compare the performance of using memcg dirty limits, compare the memcg
>   baseline (2nd row titled "mmotm w/ memcg") with the 3rd row (memcg enabled
>   with all patches).
>
>                           root_cgroup                     child_cgroup
>                  s_read s_write p_read p_write   s_read s_write p_read p_write
> mmotm w/o memcg   0.359  0.312   0.357  0.312
> mmotm w/ memcg    0.366  0.316   0.342  0.301     0.368  0.309   0.347  0.301
> all patches       0.347  0.322   0.327  0.303     0.342  0.323   0.327  0.305
> all patches       0.358  0.322   0.357  0.316
>   w/o memcg
>
> Greg Thelen (9):
>   memcg: document cgroup dirty memory interfaces
>   memcg: add page_cgroup flags for dirty page tracking
>   writeback: convert variables to unsigned
>   writeback: create dirty_info structure
>   memcg: add dirty page accounting infrastructure
>   memcg: add kernel calls for memcg dirty page stats
>   memcg: add dirty limits to mem_cgroup
>   memcg: add cgroupfs interface to memcg dirty limits
>   memcg: check memcg dirty limits in page writeback
>
>  Documentation/cgroups/memory.txt |   80 +++++++
>  fs/fs-writeback.c                |    7 +-
>  fs/nfs/write.c                   |    4 +
>  include/linux/memcontrol.h       |   33 +++-
>  include/linux/page_cgroup.h      |   23 ++
>  include/linux/writeback.h        |   18 ++-
>  mm/backing-dev.c                 |   18 +-
>  mm/filemap.c                     |    1 +
>  mm/memcontrol.c                  |  470 +++++++++++++++++++++++++++++++++++++-
>  mm/page-writeback.c              |  150 +++++++++----
>  mm/truncate.c                    |    1 +
>  mm/vmscan.c                      |    2 +-
>  mm/vmstat.c                      |    6 +-
>  13 files changed, 742 insertions(+), 71 deletions(-)
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ