[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <41fc5421-051d-f041-2cf6-b7c07e05127a@infradead.org>
Date: Thu, 6 Apr 2017 10:04:43 -0700
From: Randy Dunlap <rdunlap@...radead.org>
To: SeongJae Park <sj38.park@...il.com>, corbet@....net
Cc: akpm@...ux-foundation.org, rientjes@...gle.com,
linux-doc@...r.kernel.org, linux-kernel@...r.kernel.org
Subject: Re: [PATCH] docs/vm/transhuge: Fix few trivial typos
On 04/05/17 14:02, SeongJae Park wrote:
> Signed-off-by: SeongJae Park <sj38.park@...il.com>
> ---
> Documentation/vm/transhuge.txt | 10 +++++-----
> 1 file changed, 5 insertions(+), 5 deletions(-)
>
> diff --git a/Documentation/vm/transhuge.txt b/Documentation/vm/transhuge.txt
> index cd28d5ee5273..4e22578e50d3 100644
> --- a/Documentation/vm/transhuge.txt
> +++ b/Documentation/vm/transhuge.txt
> @@ -266,7 +266,7 @@ for each mapping.
>
> The number of file transparent huge pages mapped to userspace is available
> by reading ShmemPmdMapped and ShmemHugePages fields in /proc/meminfo.
> -To identify what applications are mapping file transparent huge pages, it
> +To identify what applications are mapping file transparent huge pages, it
> is necessary to read /proc/PID/smaps and count the FileHugeMapped fields
> for each mapping.
>
> @@ -292,7 +292,7 @@ thp_collapse_alloc_failed is incremented if khugepaged found a range
> the allocation.
>
> thp_file_alloc is incremented every time a file huge page is successfully
> -i allocated.
> + allocated.
>
> thp_file_mapped is incremented every time a file huge page is mapped into
> user address space.
> @@ -501,7 +501,7 @@ scanner can get reference to a page is get_page_unless_zero().
>
> All tail pages have zero ->_refcount until atomic_add(). This prevents the
> scanner from getting a reference to the tail page up to that point. After the
> -atomic_add() we don't care about the ->_refcount value. We already known how
> +atomic_add() we don't care about the ->_refcount value. We already known how
> many references should be uncharged from the head page.
>
> For head page get_page_unless_zero() will succeed and we don't mind. It's
> @@ -519,8 +519,8 @@ comes. Splitting will free up unused subpages.
>
> Splitting the page right away is not an option due to locking context in
> the place where we can detect partial unmap. It's also might be
> -counterproductive since in many cases partial unmap unmap happens during
> -exit(2) if an THP crosses VMA boundary.
> +counterproductive since in many cases partial unmap happens during exit(2) if
> +an THP crosses VMA boundary.
a THP
and maybe:
crosses a VMA boundary.
>
> Function deferred_split_huge_page() is used to queue page for splitting.
> The splitting itself will happen when we get memory pressure via shrinker
>
--
~Randy
Powered by blists - more mailing lists