lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20231107003151.724860-1-nphamcs@gmail.com>
Date:   Mon,  6 Nov 2023 16:31:51 -0800
From:   Nhat Pham <nphamcs@...il.com>
To:     akpm@...ux-foundation.org
Cc:     hannes@...xchg.org, cerasuolodomenico@...il.com,
        yosryahmed@...gle.com, sjenning@...hat.com, ddstreet@...e.org,
        vitaly.wool@...sulko.com, mhocko@...nel.org,
        roman.gushchin@...ux.dev, shakeelb@...gle.com,
        muchun.song@...ux.dev, chrisl@...nel.org, linux-mm@...ck.org,
        kernel-team@...a.com, linux-kernel@...r.kernel.org,
        cgroups@...r.kernel.org, linux-doc@...r.kernel.org,
        linux-kselftest@...r.kernel.org, shuah@...nel.org
Subject: [PATCH v5 3/6] zswap: make shrinking memcg-aware (fix)

Rename get_mem_cgroup_from_entry() to mem_cgroup_from_entry() and use
the zswap_pools_lock for next_shrink update.

Suggested-by: Yosry Ahmed <yosryahmed@...gle.com>
Signed-off-by: Nhat Pham <nphamcs@...il.com>
---
 mm/zswap.c | 22 +++++++++-------------
 1 file changed, 9 insertions(+), 13 deletions(-)

diff --git a/mm/zswap.c b/mm/zswap.c
index 2654b0d214cc..f1998dbd3519 100644
--- a/mm/zswap.c
+++ b/mm/zswap.c
@@ -174,7 +174,6 @@ struct zswap_pool {
 	struct hlist_node node;
 	char tfm_name[CRYPTO_MAX_ALG_NAME];
 	struct list_lru list_lru;
-	spinlock_t next_shrink_lock;
 	struct mem_cgroup *next_shrink;
 };
 
@@ -292,7 +291,7 @@ static void zswap_update_total_size(void)
 }
 
 /* should be called under RCU */
-static inline struct mem_cgroup *get_mem_cgroup_from_entry(struct zswap_entry *entry)
+static inline struct mem_cgroup *mem_cgroup_from_entry(struct zswap_entry *entry)
 {
 	return entry->objcg ? obj_cgroup_memcg(entry->objcg) : NULL;
 }
@@ -309,11 +308,9 @@ void zswap_memcg_offline_cleanup(struct mem_cgroup *memcg)
 	/* lock out zswap pools list modification */
 	spin_lock(&zswap_pools_lock);
 	list_for_each_entry(pool, &zswap_pools, list) {
-		spin_lock(&pool->next_shrink_lock);
 		if (pool->next_shrink == memcg)
 			pool->next_shrink =
 				mem_cgroup_iter(NULL, pool->next_shrink, NULL, true);
-		spin_unlock(&pool->next_shrink_lock);
 	}
 	spin_unlock(&zswap_pools_lock);
 }
@@ -360,7 +357,7 @@ static void zswap_lru_add(struct list_lru *list_lru, struct zswap_entry *entry)
 	 * Similar reasoning holds for list_lru_del() and list_lru_putback().
 	 */
 	rcu_read_lock();
-	memcg = get_mem_cgroup_from_entry(entry);
+	memcg = mem_cgroup_from_entry(entry);
 	/* will always succeed */
 	list_lru_add(list_lru, &entry->lru, nid, memcg);
 	rcu_read_unlock();
@@ -372,7 +369,7 @@ static void zswap_lru_del(struct list_lru *list_lru, struct zswap_entry *entry)
 	struct mem_cgroup *memcg;
 
 	rcu_read_lock();
-	memcg = get_mem_cgroup_from_entry(entry);
+	memcg = mem_cgroup_from_entry(entry);
 	/* will always succeed */
 	list_lru_del(list_lru, &entry->lru, nid, memcg);
 	rcu_read_unlock();
@@ -386,7 +383,7 @@ static void zswap_lru_putback(struct list_lru *list_lru,
 	struct mem_cgroup *memcg;
 
 	rcu_read_lock();
-	memcg = get_mem_cgroup_from_entry(entry);
+	memcg = mem_cgroup_from_entry(entry);
 	spin_lock(lock);
 	/* we cannot use list_lru_add here, because it increments node's lru count */
 	list_lru_putback(list_lru, &entry->lru, nid, memcg);
@@ -806,13 +803,13 @@ static void shrink_worker(struct work_struct *w)
 
 	/* global reclaim will select cgroup in a round-robin fashion. */
 	do {
-		spin_lock(&pool->next_shrink_lock);
+		spin_lock(&zswap_pools_lock);
 		memcg = pool->next_shrink =
 			mem_cgroup_iter(NULL, pool->next_shrink, NULL, true);
 
 		/* full round trip */
 		if (!memcg) {
-			spin_unlock(&pool->next_shrink_lock);
+			spin_unlock(&zswap_pools_lock);
 			if (++failures == MAX_RECLAIM_RETRIES)
 				break;
 
@@ -824,7 +821,7 @@ static void shrink_worker(struct work_struct *w)
 		 * original reference is dropped by the zswap offlining callback.
 		 */
 		css_get(&memcg->css);
-		spin_unlock(&pool->next_shrink_lock);
+		spin_unlock(&zswap_pools_lock);
 
 		ret = shrink_memcg(memcg);
 		mem_cgroup_put(memcg);
@@ -898,7 +895,6 @@ static struct zswap_pool *zswap_pool_create(char *type, char *compressor)
 	INIT_WORK(&pool->shrink_work, shrink_worker);
 
 	zswap_pool_debug("created", pool);
-	spin_lock_init(&pool->next_shrink_lock);
 
 	return pool;
 
@@ -963,10 +959,10 @@ static void zswap_pool_destroy(struct zswap_pool *pool)
 	free_percpu(pool->acomp_ctx);
 	list_lru_destroy(&pool->list_lru);
 
-	spin_lock(&pool->next_shrink_lock);
+	spin_lock(&zswap_pools_lock);
 	mem_cgroup_put(pool->next_shrink);
 	pool->next_shrink = NULL;
-	spin_unlock(&pool->next_shrink_lock);
+	spin_unlock(&zswap_pools_lock);
 
 	for (i = 0; i < ZSWAP_NR_ZPOOLS; i++)
 		zpool_destroy_pool(pool->zpools[i]);
-- 
2.34.1

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ