lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <YeA5oP/iaxtVPHb3@casper.infradead.org>
Date:   Thu, 13 Jan 2022 14:39:28 +0000
From:   Matthew Wilcox <willy@...radead.org>
To:     Liang Zhang <zhangliang5@...wei.com>
Cc:     akpm@...ux-foundation.org, linux-mm@...ck.org,
        linux-kernel@...r.kernel.org, wangzhigang17@...wei.com,
        Linus Torvalds <torvalds@...ux-foundation.org>,
        David Hildenbrand <david@...hat.com>
Subject: Re: [PATCH] mm: reuse the unshared swapcache page in do_wp_page

On Thu, Jan 13, 2022 at 10:03:18PM +0800, Liang Zhang wrote:
> In current implementation, process's read requestions will fault in pages
> with WP flags in PTEs. Next, if process emit a write requestion will go
> into do_wp_page() and copy data to a new allocated page from the old one
> due to refcount > 1 (page table mapped and swapcache), which could be
> result in performance degradation. In fact, this page is exclusively owned
> by this process and the duplication from old to a new allocated page is
> really unnecessary.
> 
> So In this situation, these unshared pages can be reused by its process.

Let's bring Linus in on this, but I think this reintroduces all of the
mapcount problems that we've been discussing recently.

How about this as an alternative?

+++ b/mm/memory.c
@@ -3291,11 +3291,11 @@ static vm_fault_t do_wp_page(struct vm_fault *vmf)
                struct page *page = vmf->page;

                /* PageKsm() doesn't necessarily raise the page refcount */
-               if (PageKsm(page) || page_count(page) != 1)
+               if (PageKsm(page) || page_count(page) != 1 + PageSwapCache(page))
                        goto copy;
                if (!trylock_page(page))
                        goto copy;
-               if (PageKsm(page) || page_mapcount(page) != 1 || page_count(page) != 1) {
+               if (PageKsm(page) || page_mapcount(page) != 1 || page_count(page) != 1 + PageSwapCache(page)) {
                        unlock_page(page);
                        goto copy;
                }


> Signed-off-by: Liang Zhang <zhangliang5@...wei.com>
> ---
> This patch has been tested with redis benchmark. Here is the test
> result.
> 
> Hardware
> ========
> Memory (GB): 512G
> CPU (total #): 88
> NVMe SSD (GB): 1024
> 
> OS
> ==
> kernel 5.10.0
> 
> Testcase
> ========
> step 1:
>   Run 16 VMs (4U8G), each running with redis-server, in a cgroup 
>   limiting memory.limit_in_bytes to 100G. 
> step 2:
>   Run memtier_bemchmark in host with params "--threads=1 --clients=1 \
> --pipeline=256 --data-size=2048 --requests=allkeys --key-minimum=1 \
> --key-maximum=30000000 --key-prefix=memtier-benchmark-prefix-redistests"
>   to test every VM concurrently.
> 
> Workset size
> ============
> cat memory.memsw.usage_in_bytes
> 125403303936
> 
> Result
> ======
> Comparing with Baseline, this patch can achieved 41% more Ops/sec, 
> 41% more Hits/sec, 41% more Misses/sec, 30% less Latency and 
> 41% more KB/sec. 
> 
>   Index(average)        Baseline kernel        Patched kernel
>   Ops/sec               109497                 155428
>   Hits/sec              8653                   12283
>   Misses/sec            90889                  129014
>   Latency               2.297                  1.603
>   KB/sec                44569                  63186
> 
> 
>  mm/memory.c | 9 ++++++++-
>  1 file changed, 8 insertions(+), 1 deletion(-)
> 
> diff --git a/mm/memory.c b/mm/memory.c
> index 23f2f1300d42..fd4d868b1c2d 100644
> --- a/mm/memory.c
> +++ b/mm/memory.c
> @@ -3291,10 +3291,16 @@ static vm_fault_t do_wp_page(struct vm_fault *vmf)
>  		struct page *page = vmf->page;
>  
>  		/* PageKsm() doesn't necessarily raise the page refcount */
> -		if (PageKsm(page) || page_count(page) != 1)
> +		if (PageKsm(page))
>  			goto copy;
>  		if (!trylock_page(page))
>  			goto copy;
> +
> +		/* reuse the unshared swapcache page */
> +		if (PageSwapCache(page) && reuse_swap_page(page, NULL)) {
> +			goto reuse;
> +		}
> +
>  		if (PageKsm(page) || page_mapcount(page) != 1 || page_count(page) != 1) {
>  			unlock_page(page);
>  			goto copy;
> @@ -3304,6 +3310,7 @@ static vm_fault_t do_wp_page(struct vm_fault *vmf)
>  		 * page count reference, and the page is locked,
>  		 * it's dark out, and we're wearing sunglasses. Hit it.
>  		 */
> +reuse:
>  		unlock_page(page);
>  		wp_page_reuse(vmf);
>  		return VM_FAULT_WRITE;
> -- 
> 2.30.0
> 
> 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ