[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-Id: <11e1ead211eb7d141efa0eb75a46ee2096ee63f8.1603267572.git.xuyu@linux.alibaba.com>
Date: Wed, 21 Oct 2020 16:09:39 +0800
From: Xu Yu <xuyu@...ux.alibaba.com>
To: linux-mm@...ck.org
Cc: hughd@...gle.com, akpm@...ux-foundation.org,
linux-kernel@...r.kernel.org
Subject: [PATCH] mm/shmem: fix up gfpmask for shmem hugepage allocation
Currently, the gfpmask used in shmem_alloc_hugepage is fixed, i.e.,
gfp | __GFP_COMP | __GFP_NORETRY | __GFP_NOWARN, where gfp comes from
inode mapping, usually GFP_HIGHUSER_MOVABLE. This will introduce direct
or kswapd reclaim when fast path of shmem hugepage allocation fails,
which is unexpected sometimes.
This applies the effect of defrag option of anonymous hugepage to shmem
hugepage too. By doing so, we can control the defrag behavior of both
kinds of THP.
This also explicitly adds the SHMEM_HUGE_ALWAYS case in
shmem_getpage_gfp, for better code reading.
Signed-off-by: Xu Yu <xuyu@...ux.alibaba.com>
---
mm/shmem.c | 44 ++++++++++++++++++++++++++++++++++++++++++++
1 file changed, 44 insertions(+)
diff --git a/mm/shmem.c b/mm/shmem.c
index 537c137698f8..a0f5d02e479b 100644
--- a/mm/shmem.c
+++ b/mm/shmem.c
@@ -1780,6 +1780,47 @@ static int shmem_swapin_page(struct inode *inode, pgoff_t index,
return error;
}
+#ifdef CONFIG_TRANSPARENT_HUGEPAGE
+static inline gfp_t shmem_hugepage_gfpmask_fixup(gfp_t gfp,
+ enum sgp_type sgp_huge)
+{
+ const bool vma_madvised = sgp_huge == SGP_HUGE;
+
+ gfp |= __GFP_NOMEMALLOC;
+ gfp &= ~__GFP_RECLAIM;
+
+ /* Force do synchronous compaction */
+ if (shmem_huge == SHMEM_HUGE_FORCE)
+ return gfp | __GFP_DIRECT_RECLAIM;
+
+ /* Always do synchronous compaction */
+ if (test_bit(TRANSPARENT_HUGEPAGE_DEFRAG_DIRECT_FLAG, &transparent_hugepage_flags))
+ return gfp | __GFP_DIRECT_RECLAIM | (vma_madvised ? 0 : __GFP_NORETRY);
+
+ /* Kick kcompactd and fail quickly */
+ if (test_bit(TRANSPARENT_HUGEPAGE_DEFRAG_KSWAPD_FLAG, &transparent_hugepage_flags))
+ return gfp | __GFP_KSWAPD_RECLAIM;
+
+ /* Synchronous compaction if madvised, otherwise kick kcompactd */
+ if (test_bit(TRANSPARENT_HUGEPAGE_DEFRAG_KSWAPD_OR_MADV_FLAG, &transparent_hugepage_flags))
+ return gfp |
+ (vma_madvised ? __GFP_DIRECT_RECLAIM :
+ __GFP_KSWAPD_RECLAIM);
+
+ /* Only do synchronous compaction if madvised */
+ if (test_bit(TRANSPARENT_HUGEPAGE_DEFRAG_REQ_MADV_FLAG, &transparent_hugepage_flags))
+ return gfp | (vma_madvised ? __GFP_DIRECT_RECLAIM : 0);
+
+ return gfp;
+}
+#else
+static inline gfp_t shmem_hugepage_gfpmask_fixup(gfp_t gfp,
+ enum sgp_type sgp_huge)
+{
+ return gfp;
+}
+#endif /* CONFIG_TRANSPARENT_HUGEPAGE */
+
/*
* shmem_getpage_gfp - find page in cache, or get from swap, or allocate
*
@@ -1867,6 +1908,8 @@ static int shmem_getpage_gfp(struct inode *inode, pgoff_t index,
switch (sbinfo->huge) {
case SHMEM_HUGE_NEVER:
goto alloc_nohuge;
+ case SHMEM_HUGE_ALWAYS:
+ goto alloc_huge;
case SHMEM_HUGE_WITHIN_SIZE: {
loff_t i_size;
pgoff_t off;
@@ -1887,6 +1930,7 @@ static int shmem_getpage_gfp(struct inode *inode, pgoff_t index,
}
alloc_huge:
+ gfp = shmem_hugepage_gfpmask_fixup(gfp, sgp_huge);
page = shmem_alloc_and_acct_page(gfp, inode, index, true);
if (IS_ERR(page)) {
alloc_nohuge:
--
2.20.1.2432.ga663e714
Powered by blists - more mailing lists