[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-Id: <1392115590-15260-2-git-send-email-sj38.park@gmail.com>
Date: Tue, 11 Feb 2014 19:46:30 +0900
From: SeongJae Park <sj38.park@...il.com>
To: sjenning@...ux.vnet.ibm.com, trivial@...nel.org
Cc: linux-mm@...ck.org, linux-kernel@...r.kernel.org,
SeongJae Park <sj38.park@...il.com>
Subject: [PATCH 2/2] mm/zswap: update zsmalloc in comment to zbud
zswap used zsmalloc before and now using zbud. But, some comments
saying it use zsmalloc yet. Fix the trivial problems.
Signed-off-by: SeongJae Park <sj38.park@...il.com>
---
mm/zswap.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
diff --git a/mm/zswap.c b/mm/zswap.c
index 5b22453..25312eb 100644
--- a/mm/zswap.c
+++ b/mm/zswap.c
@@ -165,7 +165,7 @@ static void zswap_comp_exit(void)
* be held while changing the refcount. Since the lock must
* be held, there is no reason to also make refcount atomic.
* offset - the swap offset for the entry. Index into the red-black tree.
- * handle - zsmalloc allocation handle that stores the compressed page data
+ * handle - zbud allocation handle that stores the compressed page data
* length - the length in bytes of the compressed page data. Needed during
* decompression
*/
@@ -282,7 +282,7 @@ static void zswap_rb_erase(struct rb_root *root, struct zswap_entry *entry)
}
/*
- * Carries out the common pattern of freeing and entry's zsmalloc allocation,
+ * Carries out the common pattern of freeing and entry's zbud allocation,
* freeing the entry itself, and decrementing the number of stored pages.
*/
static void zswap_free_entry(struct zswap_tree *tree,
--
1.8.1.2
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists