lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <CA+2MQi-yLWdoP0goa8ZzGGJmEJrQoO_jHN8gYS=gvKAt3JAswg@mail.gmail.com>
Date:   Thu, 7 Jan 2021 11:29:22 +0800
From:   Liang Li <liliang324@...il.com>
To:     Alexander Duyck <alexander.duyck@...il.com>
Cc:     Alexander Duyck <alexander.h.duyck@...ux.intel.com>,
        Mel Gorman <mgorman@...hsingularity.net>,
        Andrew Morton <akpm@...ux-foundation.org>,
        Andrea Arcangeli <aarcange@...hat.com>,
        Dan Williams <dan.j.williams@...el.com>,
        "Michael S. Tsirkin" <mst@...hat.com>,
        David Hildenbrand <david@...hat.com>,
        Jason Wang <jasowang@...hat.com>,
        Dave Hansen <dave.hansen@...el.com>,
        Michal Hocko <mhocko@...e.com>,
        Liang Li <liliangleo@...iglobal.com>,
        Mike Kravetz <mike.kravetz@...cle.com>,
        linux-mm <linux-mm@...ck.org>,
        LKML <linux-kernel@...r.kernel.org>,
        virtualization@...ts.linux-foundation.org
Subject: Re: [PATCH 2/6] mm: let user decide page reporting option

> >  enum {
> >         PAGE_REPORTING_IDLE = 0,
> > @@ -44,7 +45,7 @@ __page_reporting_request(struct page_reporting_dev_info *prdev)
> >          * now we are limiting this to running no more than once every
> >          * couple of seconds.
> >          */
> > -       schedule_delayed_work(&prdev->work, PAGE_REPORTING_DELAY);
> > +       schedule_delayed_work(&prdev->work, prdev->delay_jiffies);
> >  }
> >
>
> So this ends up being the reason why you needed to add the batch size
> value. However I don't really see it working as expected since you
> could essentially have 1 page freed 4M times that could trigger your
> page zeroing logic. So for example if a NIC is processing frames and
> ends up freeing and then reallocating some small batch of pages this
> could would be running often even though there isn't really all that
> many pages that needed zeroing.

Good catch, it works not like batch size means.

> >  /* notify prdev of free page reporting request */
> > @@ -230,7 +231,7 @@ page_reporting_process_zone(struct page_reporting_dev_info *prdev,
> >
> >         /* Generate minimum watermark to be able to guarantee progress */
> >         watermark = low_wmark_pages(zone) +
> > -                   (PAGE_REPORTING_CAPACITY << PAGE_REPORTING_MIN_ORDER);
> > +                   (PAGE_REPORTING_CAPACITY << prdev->mini_order);
> >
> >         /*
> >          * Cancel request if insufficient free memory or if we failed
>
> With the page order being able to be greatly reduced this could have a
> significant impact on if this code really has any value. Previously we
> were able to guarantee a pretty significant number of higher order
> pages free. With this we might only be guaranteeing something like 32
> 4K pages which is pretty small compared to what can end up being
> pulled out at the higher end.

I have dropped the 'buddy free page pre zero'  patch, so the mini order will
not change to a small value.

> > @@ -240,7 +241,7 @@ page_reporting_process_zone(struct page_reporting_dev_info *prdev,
> >                 return err;
> >
> >         /* Process each free list starting from lowest order/mt */
> > -       for (order = PAGE_REPORTING_MIN_ORDER; order < MAX_ORDER; order++) {
> > +       for (order = prdev->mini_order; order < MAX_ORDER; order++) {
> >                 for (mt = 0; mt < MIGRATE_TYPES; mt++) {
> >                         /* We do not pull pages from the isolate free list */
> >                         if (is_migrate_isolate(mt))
> > @@ -307,7 +308,7 @@ static void page_reporting_process(struct work_struct *work)
> >          */
> >         state = atomic_cmpxchg(&prdev->state, state, PAGE_REPORTING_IDLE);
> >         if (state == PAGE_REPORTING_REQUESTED)
> > -               schedule_delayed_work(&prdev->work, PAGE_REPORTING_DELAY);
> > +               schedule_delayed_work(&prdev->work, prdev->delay_jiffies);
> >  }
> >
> >  static DEFINE_MUTEX(page_reporting_mutex);
> > @@ -335,6 +336,8 @@ int page_reporting_register(struct page_reporting_dev_info *prdev)
> >         /* Assign device to allow notifications */
> >         rcu_assign_pointer(pr_dev_info, prdev);
> >
> > +       page_report_mini_order = prdev->mini_order;
> > +       page_report_batch_size = prdev->batch_size;
> >         /* enable page reporting notification */
> >         if (!static_key_enabled(&page_reporting_enabled)) {
> >                 static_branch_enable(&page_reporting_enabled);
> > @@ -352,6 +355,8 @@ void page_reporting_unregister(struct page_reporting_dev_info *prdev)
> >         mutex_lock(&page_reporting_mutex);
> >
> >         if (rcu_access_pointer(pr_dev_info) == prdev) {
> > +               if (static_key_enabled(&page_reporting_enabled))
> > +                       static_branch_disable(&page_reporting_enabled);
> >                 /* Disable page reporting notification */
> >                 RCU_INIT_POINTER(pr_dev_info, NULL);
> >                 synchronize_rcu();
>
> If we are going to use this we are using it. Once we NULL out the
> prdev that should stop page reporting from running. We shouldn't be
> relying on the static key.

The benefits for this is that the function call of '__page_reporting_notify' in
'page_reporting_notify_free' can be skipped, it helps to save some
cycles.

> > diff --git a/mm/page_reporting.h b/mm/page_reporting.h
> > index b8fb3bbb345f..86ac6ffad970 100644
> > --- a/mm/page_reporting.h
> > +++ b/mm/page_reporting.h
> > @@ -9,9 +9,9 @@
> >  #include <linux/slab.h>
> >  #include <linux/pgtable.h>
> >  #include <linux/scatterlist.h>
> > +#include <linux/page_reporting.h>
> >
> > -#define PAGE_REPORTING_MIN_ORDER       pageblock_order
> > -
> > +extern int page_report_mini_order;
> >  extern unsigned long page_report_batch_size;
> >
> >  #ifdef CONFIG_PAGE_REPORTING
> > @@ -42,7 +42,7 @@ static inline void page_reporting_notify_free(unsigned int order)
> >                 return;
> >
> >         /* Determine if we have crossed reporting threshold */
> > -       if (order < PAGE_REPORTING_MIN_ORDER)
> > +       if (order < page_report_mini_order)
> >                 return;
>
> This is another case where it might make sense to look at placing the
> code in __page_reporting_notify if we are going to allow this to be a
> dynamically configurable value.

I will change it in the next revision. Thanks!

Liang

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ