[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <9aea8b6f-0d1a-b700-efe4-8efcebc5d16d@linux.vnet.ibm.com>
Date: Wed, 24 Jan 2018 15:05:26 +0530
From: "Aneesh Kumar K.V" <aneesh.kumar@...ux.vnet.ibm.com>
To: Christophe LEROY <christophe.leroy@....fr>,
Benjamin Herrenschmidt <benh@...nel.crashing.org>,
Paul Mackerras <paulus@...ba.org>,
Michael Ellerman <mpe@...erman.id.au>,
Scott Wood <oss@...error.net>
Cc: linux-kernel@...r.kernel.org, linuxppc-dev@...ts.ozlabs.org
Subject: Re: [PATCH v3 5/5] powerpc/mm: Fix growth direction for hugepages
mmaps with slice
On 01/24/2018 02:57 PM, Christophe LEROY wrote:
>
>
> Le 24/01/2018 à 10:15, Aneesh Kumar K.V a écrit :
>>
>>
>> On 01/24/2018 02:32 PM, Christophe Leroy wrote:
>>> An application running with libhugetlbfs fails to allocate
>>> additional pages to HEAP due to the hugemap being done
>>> inconditionally as topdown mapping:
>>>
>>> mmap(0x10080000, 1572864, PROT_READ|PROT_WRITE,
>>> MAP_PRIVATE|MAP_ANONYMOUS|0x40000, -1, 0) = 0x73e80000
>>> [...]
>>> mmap(0x74000000, 1048576, PROT_READ|PROT_WRITE,
>>> MAP_PRIVATE|MAP_ANONYMOUS|0x40000, -1, 0x180000) = 0x73d80000
>>> munmap(0x73d80000, 1048576) = 0
>>> [...]
>>> mmap(0x74000000, 1572864, PROT_READ|PROT_WRITE,
>>> MAP_PRIVATE|MAP_ANONYMOUS|0x40000, -1, 0x180000) = 0x73d00000
>>> munmap(0x73d00000, 1572864) = 0
>>> [...]
>>> mmap(0x74000000, 1572864, PROT_READ|PROT_WRITE,
>>> MAP_PRIVATE|MAP_ANONYMOUS|0x40000, -1, 0x180000) = 0x73d00000
>>> munmap(0x73d00000, 1572864) = 0
>>> [...]
>>>
>>> As one can see from the above strace log, mmap() allocates further
>>> pages below the initial one because no space is available on top of it.
>>>
>>> This patch fixes it by requesting bottomup mapping as the non
>>> generic hugetlb_get_unmapped_area() does
>>>
>>> Fixes: d0f13e3c20b6f ("[POWERPC] Introduce address space "slices" ")
>>> Signed-off-by: Christophe Leroy <christophe.leroy@....fr>
>>> ---
>>> v3: Was a standalone patch before, but conflicts with this serie.
>>>
>>> arch/powerpc/mm/hugetlbpage.c | 2 +-
>>> 1 file changed, 1 insertion(+), 1 deletion(-)
>>>
>>> diff --git a/arch/powerpc/mm/hugetlbpage.c
>>> b/arch/powerpc/mm/hugetlbpage.c
>>> index 79e1378ee303..368ea6b248ad 100644
>>> --- a/arch/powerpc/mm/hugetlbpage.c
>>> +++ b/arch/powerpc/mm/hugetlbpage.c
>>> @@ -558,7 +558,7 @@ unsigned long hugetlb_get_unmapped_area(struct
>>> file *file, unsigned long addr,
>>> return radix__hugetlb_get_unmapped_area(file, addr, len,
>>> pgoff, flags);
>>> #endif
>>> - return slice_get_unmapped_area(addr, len, flags, mmu_psize, 1);
>>> + return slice_get_unmapped_area(addr, len, flags, mmu_psize, 0);
>>> }
>>> #endif
>>
>> Why make this change also for PPC64? Can you do this #ifdef 8xx?.You
>> can ideally move hugetlb_get_unmapped_area to slice.h and then make
>> this much simpler for 8xxx?
>>
>
> Did you try with HUGETLB_MORECORE_HEAPBASE=0x11000000 on PPC64 as I
> suggested in my last email on this subject (22/01/2018 9:22) ?
yes. The test ran fine for me
kvaneesh@...tulc6a-p1:[~]$ HUGETLB_MORECORE=yes
HUGETLB_MORECORE_HEAPBASE=0x30000000 ./a.out
10000000-10010000 r-xp 00000000 fc:00 9044312
/home/kvaneesh/a.out
10010000-10020000 r--p 00000000 fc:00 9044312
/home/kvaneesh/a.out
10020000-10030000 rw-p 00010000 fc:00 9044312
/home/kvaneesh/a.out
30000000-33000000 rw-p 00000000 00:0d 1062697
/anon_hugepage (deleted)
33000000-35000000 rw-p 03000000 00:0d 1062698
/anon_hugepage (deleted)
35000000-37000000 rw-p 05000000 00:0d 1062699
/anon_hugepage (deleted)
7ffff7d60000-7ffff7f10000 r-xp 00000000 fc:00 9250090
/lib/powerpc64le-linux-gnu/libc-2.23.so
7ffff7f10000-7ffff7f20000 r--p 001a0000 fc:00 9250090
/lib/powerpc64le-linux-gnu/libc-2.23.so
7ffff7f20000-7ffff7f30000 rw-p 001b0000 fc:00 9250090
/lib/powerpc64le-linux-gnu/libc-2.23.so
7ffff7f40000-7ffff7f60000 r-xp 00000000 fc:00 10754812
/usr/lib/libhugetlbfs.so.0
7ffff7f60000-7ffff7f70000 r--p 00010000 fc:00 10754812
/usr/lib/libhugetlbfs.so.0
7ffff7f70000-7ffff7f80000 rw-p 00020000 fc:00 10754812
/usr/lib/libhugetlbfs.so.0
7ffff7f80000-7ffff7fa0000 r-xp 00000000 00:00 0
[vdso]
7ffff7fa0000-7ffff7fe0000 r-xp 00000000 fc:00 9250107
/lib/powerpc64le-linux-gnu/ld-2.23.so
7ffff7fe0000-7ffff7ff0000 r--p 00030000 fc:00 9250107
/lib/powerpc64le-linux-gnu/ld-2.23.so
7ffff7ff0000-7ffff8000000 rw-p 00040000 fc:00 9250107
/lib/powerpc64le-linux-gnu/ld-2.23.so
7ffffffd0000-800000000000 rw-p 00000000 00:00 0
[stack]
>
> Before doing anything specific to the PPC32/8xx, I'd like to be sure the
> issue is definitly only on PPC32.
>
I am not sure I understand the problem correctly. If there is a free
space in the required range, both topdown/bottomup search should be able
to find it. Unless topdown found another free area suitable for hugetlb
allocation above. My take is we should not change the topdown to
bottomup without really understanding the failure scenarios.
-aneesh
Powered by blists - more mailing lists