[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-Id: <1587970985-21629-1-git-send-email-alex.shi@linux.alibaba.com>
Date: Mon, 27 Apr 2020 15:02:49 +0800
From: Alex Shi <alex.shi@...ux.alibaba.com>
To: akpm@...ux-foundation.org, mgorman@...hsingularity.net,
tj@...nel.org, hughd@...gle.com, khlebnikov@...dex-team.ru,
daniel.m.jordan@...cle.com, yang.shi@...ux.alibaba.com,
willy@...radead.org, hannes@...xchg.org, lkp@...el.com,
linux-mm@...ck.org, linux-kernel@...r.kernel.org,
cgroups@...r.kernel.org, shakeelb@...gle.com,
iamjoonsoo.kim@....com, richard.weiyang@...il.com
Cc: Alex Shi <alex.shi@...ux.alibaba.com>
Subject: [PATCH v10 00/15] per memcg lru lock
This is a new version which bases on Johannes new patchset
"mm: memcontrol: charge swapin pages on instantiation"
https://lkml.org/lkml/2020/4/21/266
Johannes Weiner has suggested:
"So here is a crazy idea that may be worth exploring:
Right now, pgdat->lru_lock protects both PageLRU *and* the lruvec's
linked list.
Can we make PageLRU atomic and use it to stabilize the lru_lock
instead, and then use the lru_lock only serialize list operations?
..."
With the cleaning memcg charge path and this suggestion, we could isolate
LRU pages to exclusive visit them in compaction, page migration, reclaim,
memcg move_accunt, huge page split etc scenarios while keeping pages'
memcg stable. Then possible to change per node lru locking to per memcg
lru locking. As to pagevec_lru_move_fn funcs, it would be safe to let
pages remain on lru list, lru lock could guard them for list integrity.
This is version safely pass Hugh Dickins's swapping kernel building
testcase, Thanks for the great case! I want to send out a bit early
for more testing and review while people's memory is still hot with
Johannes new memcg charge patch. :) I will do more testing beside.
The patchset includes 3 parts:
1, some code cleanup and minimum optimization as a preparation.
2, use TestCleanPageLRU as page isolation's precondition
3, replace per node lru_lock with per memcg per node lru_lock
The 3rd part moves per node lru_lock into lruvec, thus bring a lru_lock for
each of memcg per node. So on a large machine, each of memcg don't
have to suffer from per node pgdat->lru_lock competition. They could go
fast with their self lru_lock
Following Daniel Jordan's suggestion, I have run 208 'dd' with on 104
containers on a 2s * 26cores * HT box with a modefied case:
https://git.kernel.org/pub/scm/linux/kernel/git/wfg/vm-scalability.git/tree/case-lru-file-readtwice
With this patchset, the readtwice performance increased about 80%
in concurrent containers.
Thanks Hugh Dickins and Konstantin Khlebnikov, they both brought this
idea 8 years ago, and others who give comments as well: Daniel Jordan,
Mel Gorman, Shakeel Butt, Matthew Wilcox etc.
Thanks for Testing support from Intel 0day and Rong Chen, Fengguang Wu,
and Yun Wang.
Alex Shi (13):
mm/swap: use vmf clean up swapin funcs parameters
mm/vmscan: remove unnecessary lruvec adding
mm/page_idle: no unlikely double check for idle page counting
mm/thp: move lru_add_page_tail func to huge_memory.c
mm/thp: clean up lru_add_page_tail
mm/thp: narrow lru locking
mm/memcg: add debug checking in lock_page_memcg
mm/lru: introduce TestClearPageLRU
mm/compaction: do page isolation first in compaction
mm/mlock: ClearPageLRU before get lru lock in munlock page isolation
mm/lru: replace pgdat lru_lock with lruvec lock
mm/lru: introduce the relock_page_lruvec function
mm/pgdat: remove pgdat lru_lock
Hugh Dickins (2):
mm/vmscan: use relock for move_pages_to_lru
mm/lru: revise the comments of lru_lock
Documentation/admin-guide/cgroup-v1/memcg_test.rst | 15 +-
Documentation/admin-guide/cgroup-v1/memory.rst | 8 +-
Documentation/trace/events-kmem.rst | 2 +-
Documentation/vm/unevictable-lru.rst | 22 +--
include/linux/memcontrol.h | 92 +++++++++++
include/linux/mm_types.h | 2 +-
include/linux/mmzone.h | 5 +-
include/linux/page-flags.h | 1 +
include/linux/swap.h | 12 +-
mm/compaction.c | 85 +++++++----
mm/filemap.c | 4 +-
mm/huge_memory.c | 55 +++++--
mm/madvise.c | 11 +-
mm/memcontrol.c | 87 ++++++++++-
mm/mlock.c | 93 ++++++------
mm/mmzone.c | 1 +
mm/page_alloc.c | 1 -
mm/page_idle.c | 8 -
mm/rmap.c | 2 +-
mm/swap.c | 119 ++++-----------
mm/swap_state.c | 23 ++-
mm/swapfile.c | 8 +-
mm/vmscan.c | 168 +++++++++++----------
mm/zswap.c | 3 +-
24 files changed, 497 insertions(+), 330 deletions(-)
--
1.8.3.1
Powered by blists - more mailing lists