lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <CALZtOND_44Y=AJ24gtVv8kFK6j54tqhXp=A7+QVMno=YGDHrRA@mail.gmail.com>
Date:   Thu, 26 Sep 2019 06:51:55 -0400
From:   Dan Streetman <ddstreet@...e.org>
To:     Hui Zhu <teawaterz@...ux.alibaba.com>,
        Andrew Morton <akpm@...ux-foundation.org>
Cc:     Seth Jennings <sjenning@...hat.com>,
        Michal Hocko <mhocko@...e.com>,
        Matthew Wilcox <willy@...radead.org>, chris@...is-wilson.co.uk,
        Johannes Weiner <hannes@...xchg.org>, ziqian.lzq@...fin.com,
        osandov@...com, Huang Ying <ying.huang@...el.com>,
        aryabinin@...tuozzo.com, vovoy@...omium.org,
        richard.weiyang@...il.com, jgg@...pe.ca, dan.j.williams@...el.com,
        rppt@...ux.ibm.com, jglisse@...hat.com, b.zolnierkie@...sung.com,
        axboe@...nel.dk, dennis@...nel.org,
        Josef Bacik <josef@...icpanda.com>, tj@...nel.org,
        oleg@...hat.com, Randy Dunlap <rdunlap@...radead.org>,
        linux-kernel <linux-kernel@...r.kernel.org>,
        Linux-MM <linux-mm@...ck.org>
Subject: Re: [RFC v3] zswap: Add CONFIG_ZSWAP_IO_SWITCH to handle swap IO issue

On Mon, Sep 23, 2019 at 4:14 PM Dan Streetman <ddstreet@...e.org> wrote:
>
> On Sun, Sep 22, 2019 at 11:32 PM Hui Zhu <teawaterz@...ux.alibaba.com> wrote:
> >
> > This is the third version of this patch.  The first and second version
> > is in [1] and [2].
> > This verion is updated according to the comments from Randy Dunlap
> > in [3].
> >
> > Currently, I use a VM that has 2 CPUs, 4G memory and 4G swap file.
> > I found that swap will affect the IO performance when it is running.
> > So I open zswap to handle it because it just use CPU cycles but not
> > disk IO.
> >
> > It work OK but I found that zswap is slower than normal swap in this
> > VM.  zswap is about 300M/s and normal swap is about 500M/s. (The reason
> > is disk inside VM has fscache in host machine.)
>
> I must be missing something here - if zswap in the guest is *slower*
> than real swap, why are you using zswap?
>
> Also, I don't see why zswap is slower than normal swap, unless you
> mean that your zswap is full, since once zswap fills up any additional
> swap will absolutely be slower than not having zswap at all.
>
> > So open zswap is make memory shrinker slower but good for IO performance
> > in this VM.
> > So I just want zswap work when the disk of the swap file is under high
> > IO load.
> >
> > This commit is designed for this idea.
> > It add two parameters read_in_flight_limit and write_in_flight_limit to
> > zswap.
> > In zswap_frontswap_store, pages will be stored to zswap only when
> > the IO in flight number of swap device is bigger than
> > zswap_read_in_flight_limit or zswap_write_in_flight_limit
> > when zswap is enabled.
> > Then the zswap just work when the IO in flight number of swap device
> > is low.
>
> Ok, so maybe I understand what you mean, your disk I/O is normally
> very fast, but once your host-side cache is full it starts actually
> writing to your host physical disk, and your guest swap I/O drops way
> down (since caching pages in host memory is much faster than writing
> to a host physical disk).  Is that what's going on?  That was not
> clear at all to me from the commit description...
>
> In general I think the description of this commit, as well as the docs
> and even user interface of how to use it, is very confusing.  I can
> see how it would be beneficial in this specific situation, but I'm not
> a fan of the implementation, and I'm very concerned that nobody will
> be able to understand how to use it properly - when should they enable
> it?  What limit values should they use?  Why are there separate read
> and write limits?  None of that is clear to me, and I'm fairly
> certainly it would not be clear to other normal users.
>
> Is there a better way this can be done?
>
> >
> > [1] https://lkml.org/lkml/2019/9/11/935
> > [2] https://lkml.org/lkml/2019/9/20/90
> > [3] https://lkml.org/lkml/2019/9/20/1076
> >
> > Signed-off-by: Hui Zhu <teawaterz@...ux.alibaba.com>

Nacked-by: Dan Streetman <ddstreet@...e.org>

due to my concerns that I emailed before


> > ---
> >  include/linux/swap.h |  3 +++
> >  mm/Kconfig           | 18 ++++++++++++++++
> >  mm/page_io.c         | 16 +++++++++++++++
> >  mm/zswap.c           | 58 ++++++++++++++++++++++++++++++++++++++++++++++++++++
> >  4 files changed, 95 insertions(+)
> >
> > diff --git a/include/linux/swap.h b/include/linux/swap.h
> > index de2c67a..82b621f 100644
> > --- a/include/linux/swap.h
> > +++ b/include/linux/swap.h
> > @@ -389,6 +389,9 @@ extern void end_swap_bio_write(struct bio *bio);
> >  extern int __swap_writepage(struct page *page, struct writeback_control *wbc,
> >         bio_end_io_t end_write_func);
> >  extern int swap_set_page_dirty(struct page *page);
> > +#ifdef CONFIG_ZSWAP_IO_SWITCH
> > +extern void swap_io_in_flight(struct page *page, unsigned int inflight[2]);
> > +#endif
> >
> >  int add_swap_extent(struct swap_info_struct *sis, unsigned long start_page,
> >                 unsigned long nr_pages, sector_t start_block);
> > diff --git a/mm/Kconfig b/mm/Kconfig
> > index 56cec63..387c3b5 100644
> > --- a/mm/Kconfig
> > +++ b/mm/Kconfig
> > @@ -546,6 +546,24 @@ config ZSWAP
> >           they have not be fully explored on the large set of potential
> >           configurations and workloads that exist.
> >
> > +config ZSWAP_IO_SWITCH
> > +       bool "Compressed cache for swap pages according to the IO status"
> > +       depends on ZSWAP
> > +       help
> > +         This function helps the system that normal swap speed is higher
> > +         than zswap speed to handle the swap IO issue.
> > +         For example, a VM where the disk device is not set cache config or
> > +         set cache=writeback.
> > +
> > +         This function makes zswap just work when the disk of the swap file
> > +         is under high IO load.
> > +         It add two parameters (read_in_flight_limit and
> > +         write_in_flight_limit) to zswap.  When zswap is enabled, pages will
> > +         be stored to zswap only when the IO in flight number of swap device
> > +         is bigger than zswap_read_in_flight_limit or
> > +         zswap_write_in_flight_limit.
> > +         If unsure, say "n".
> > +
> >  config ZPOOL
> >         tristate "Common API for compressed memory storage"
> >         help
> > diff --git a/mm/page_io.c b/mm/page_io.c
> > index 24ee600..e66b050 100644
> > --- a/mm/page_io.c
> > +++ b/mm/page_io.c
> > @@ -434,3 +434,19 @@ int swap_set_page_dirty(struct page *page)
> >                 return __set_page_dirty_no_writeback(page);
> >         }
> >  }
> > +
> > +#ifdef CONFIG_ZSWAP_IO_SWITCH
> > +void swap_io_in_flight(struct page *page, unsigned int inflight[2])
> > +{
> > +       struct swap_info_struct *sis = page_swap_info(page);
> > +
> > +       if (!sis->bdev) {
> > +               inflight[0] = 0;
> > +               inflight[1] = 0;
> > +               return;
> > +       }
> > +
> > +       part_in_flight_rw(bdev_get_queue(sis->bdev), sis->bdev->bd_part,
> > +                                         inflight);
>
> this potentially will read inflight stats info from all possible cpus,
> that's not something I'm a big fan of adding to every single page swap
> call...it's not awful, but there might be scaling issues for systems
> with lots of cpus.
>
> > +}
> > +#endif
> > diff --git a/mm/zswap.c b/mm/zswap.c
> > index 0e22744..0190b2d 100644
> > --- a/mm/zswap.c
> > +++ b/mm/zswap.c
> > @@ -62,6 +62,14 @@ static u64 zswap_reject_compress_poor;
> >  static u64 zswap_reject_alloc_fail;
> >  /* Store failed because the entry metadata could not be allocated (rare) */
> >  static u64 zswap_reject_kmemcache_fail;
> > +#ifdef CONFIG_ZSWAP_IO_SWITCH
> > +/*
> > + * Store failed because zswap_read_in_flight_limit or
> > + * zswap_write_in_flight_limit is bigger than IO in flight number of
> > + * swap device
> > + */
> > +static u64 zswap_reject_io;
> > +#endif
> >  /* Duplicate store was encountered (rare) */
> >  static u64 zswap_duplicate_entry;
> >
> > @@ -114,6 +122,24 @@ static bool zswap_same_filled_pages_enabled = true;
> >  module_param_named(same_filled_pages_enabled, zswap_same_filled_pages_enabled,
> >                    bool, 0644);
> >
> > +#ifdef CONFIG_ZSWAP_IO_SWITCH
> > +/*
> > + * zswap will not try to store the page if zswap_read_in_flight_limit is
> > + * bigger than IO read in flight number of swap device
> > + */
> > +static unsigned int zswap_read_in_flight_limit;
> > +module_param_named(read_in_flight_limit, zswap_read_in_flight_limit,
> > +                  uint, 0644);
> > +
> > +/*
> > + * zswap will not try to store the page if zswap_write_in_flight_limit is
> > + * bigger than IO write in flight number of swap device
> > + */
> > +static unsigned int zswap_write_in_flight_limit;
> > +module_param_named(write_in_flight_limit, zswap_write_in_flight_limit,
> > +                  uint, 0644);
> > +#endif
> > +
> >  /*********************************
> >  * data structures
> >  **********************************/
> > @@ -1009,6 +1035,34 @@ static int zswap_frontswap_store(unsigned type, pgoff_t offset,
> >                 goto reject;
> >         }
> >
> > +#ifdef CONFIG_ZSWAP_IO_SWITCH
> > +       if (zswap_read_in_flight_limit || zswap_write_in_flight_limit) {
> > +               unsigned int inflight[2];
> > +               bool should_swap = false;
> > +
> > +               swap_io_in_flight(page, inflight);
> > +
> > +               if (zswap_write_in_flight_limit &&
> > +                       inflight[1] < zswap_write_in_flight_limit)
> > +                       should_swap = true;
> > +
> > +               if (zswap_read_in_flight_limit &&
> > +                       (should_swap ||
> > +                        (!should_swap && !zswap_write_in_flight_limit))) {
> > +                       if (inflight[0] < zswap_read_in_flight_limit)
> > +                               should_swap = true;
> > +                       else
> > +                               should_swap = false;
> > +               }
> > +
> > +               if (should_swap) {
> > +                       zswap_reject_io++;
> > +                       ret = -EIO;
> > +                       goto reject;
> > +               }
> > +       }
> > +#endif
> > +
> >         /* reclaim space if needed */
> >         if (zswap_is_full()) {
> >                 zswap_pool_limit_hit++;
> > @@ -1264,6 +1318,10 @@ static int __init zswap_debugfs_init(void)
> >                            zswap_debugfs_root, &zswap_reject_kmemcache_fail);
> >         debugfs_create_u64("reject_compress_poor", 0444,
> >                            zswap_debugfs_root, &zswap_reject_compress_poor);
> > +#ifdef CONFIG_ZSWAP_IO_SWITCH
> > +       debugfs_create_u64("reject_io", 0444,
>
> "reject_io" is not very clear about why it was rejected; I think most
> people will assume this means pages were rejected because of I/O
> errors, not because the I/O inflight page count was lower than the set
> limit.
>
> > +                          zswap_debugfs_root, &zswap_reject_io);
> > +#endif
> >         debugfs_create_u64("written_back_pages", 0444,
> >                            zswap_debugfs_root, &zswap_written_back_pages);
> >         debugfs_create_u64("duplicate_entry", 0444,
> > --
> > 2.7.4
> >

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ