[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-Id: <20231206-zswap-lock-optimize-v1-0-e25b059f9c3a@bytedance.com>
Date: Wed, 06 Dec 2023 09:46:23 +0000
From: Chengming Zhou <zhouchengming@...edance.com>
To: Vitaly Wool <vitaly.wool@...sulko.com>,
Nhat Pham <nphamcs@...il.com>,
Johannes Weiner <hannes@...xchg.org>,
Michal Hocko <mhocko@...nel.org>,
Seth Jennings <sjenning@...hat.com>,
Dan Streetman <ddstreet@...e.org>,
Andrew Morton <akpm@...ux-foundation.org>,
Yosry Ahmed <yosryahmed@...gle.com>
Cc: linux-mm@...ck.org, linux-kernel@...r.kernel.org,
Chengming Zhou <zhouchengming@...edance.com>
Subject: [PATCH 0/7] mm/zswap: optimize the scalability of zswap rb-tree
Hi everyone,
This patch series is based on the linux-next 20231205, which depends on
the "workload-specific and memory pressure-driven zswap writeback" series
from Nhat Pham.
When testing the zswap performance by using kernel build -j32 in a tmpfs
directory, I found the scalability of zswap rb-tree is not good, which
is protected by the only spinlock. That would cause heavy lock contention
if multiple tasks zswap_store/load concurrently.
So a simple solution is to split the only one zswap rb-tree into multiple
rb-trees, each corresponds to SWAP_ADDRESS_SPACE_PAGES (64M). This idea is
from the commit 4b3ef9daa4fc ("mm/swap: split swap cache into 64MB trunks").
Although this method can't solve the spinlock contention completely, it
can mitigate much of that contention.
Another problem when testing the zswap using our default zsmalloc is that
zswap_load() and zswap_writeback_entry() have to malloc a temporary memory
to support !zpool_can_sleep_mapped().
Optimize it by reusing the percpu crypto_acomp_ctx->dstmem, which is also
used by zswap_store() and protected by the same percpu crypto_acomp_ctx->mutex.
Thanks for review and comment!
To: Andrew Morton <akpm@...ux-foundation.org>
To: Seth Jennings <sjenning@...hat.com>
To: Dan Streetman <ddstreet@...e.org>
To: Vitaly Wool <vitaly.wool@...sulko.com>
To: Nhat Pham <nphamcs@...il.com>
To: Johannes Weiner <hannes@...xchg.org>
To: Yosry Ahmed <yosryahmed@...gle.com>
To: Michal Hocko <mhocko@...nel.org>
Cc: linux-kernel@...r.kernel.org
Cc: linux-mm@...ck.org
Signed-off-by: Chengming Zhou <zhouchengming@...edance.com>
---
Chengming Zhou (7):
mm/zswap: make sure each swapfile always have zswap rb-tree
mm/zswap: split zswap rb-tree
mm/zswap: reuse dstmem when decompress
mm/zswap: change dstmem size to one page
mm/zswap: refactor out __zswap_load()
mm/zswap: cleanup zswap_load()
mm/zswap: cleanup zswap_reclaim_entry()
include/linux/zswap.h | 4 +-
mm/swapfile.c | 10 ++-
mm/zswap.c | 233 +++++++++++++++++++++-----------------------------
3 files changed, 106 insertions(+), 141 deletions(-)
---
base-commit: 0f5f12ac05f36f117e793656c3f560625e927f1b
change-id: 20231206-zswap-lock-optimize-06f45683b02b
Best regards,
--
Chengming Zhou <zhouchengming@...edance.com>
Powered by blists - more mailing lists