lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20180806155852.7jvudjpzzq6fdp33@mac>
Date:   Mon, 6 Aug 2018 17:58:52 +0200
From:   Roger Pau Monné <roger.pau@...rix.com>
To:     Juergen Gross <jgross@...e.com>
CC:     <linux-kernel@...r.kernel.org>, <xen-devel@...ts.xenproject.org>,
        <linux-block@...r.kernel.org>, <konrad.wilk@...cle.com>,
        <axboe@...nel.dk>, <boris.ostrovsky@...cle.com>
Subject: Re: [PATCH 1/4] xen/blkback: don't keep persistent grants too long

On Mon, Aug 06, 2018 at 01:33:59PM +0200, Juergen Gross wrote:
> Persistent grants are allocated until a threshold per ring is being
> reached. Those grants won't be freed until the ring is being destroyed
> meaning there will be resources kept busy which might no longer be
> used.
> 
> Instead of freeing only persistent grants until the threshold is
> reached add a timestamp and remove all persistent grants not having
> been in use for a minute.
> 
> Signed-off-by: Juergen Gross <jgross@...e.com>
> ---
>  drivers/block/xen-blkback/blkback.c | 77 +++++++++++++++++++++++--------------
>  drivers/block/xen-blkback/common.h  |  1 +
>  2 files changed, 50 insertions(+), 28 deletions(-)

You should document this new parameter in
Documentation/ABI/testing/sysfs-driver-xen-blkback.

> 
> diff --git a/drivers/block/xen-blkback/blkback.c b/drivers/block/xen-blkback/blkback.c
> index b55b245e8052..485e3ecab144 100644
> --- a/drivers/block/xen-blkback/blkback.c
> +++ b/drivers/block/xen-blkback/blkback.c
> @@ -84,6 +84,18 @@ MODULE_PARM_DESC(max_persistent_grants,
>                   "Maximum number of grants to map persistently");
>  
>  /*
> + * How long a persistent grant is allowed to remain allocated without being in
> + * use. The time is in seconds, 0 means indefinitely long.
> + */
> +
> +unsigned int xen_blkif_pgrant_timeout = 60;
> +module_param_named(persistent_grant_unused_seconds, xen_blkif_pgrant_timeout,
> +		   uint, 0644);
> +MODULE_PARM_DESC(persistent_grant_unused_seconds,
> +		 "Time in seconds an unused persistent grant is allowed to "
> +		 "remain allocated. Default is 60, 0 means unlimited.");
> +
> +/*
>   * Maximum number of rings/queues blkback supports, allow as many queues as there
>   * are CPUs if user has not specified a value.
>   */
> @@ -123,6 +135,13 @@ module_param(log_stats, int, 0644);
>  /* Number of free pages to remove on each call to gnttab_free_pages */
>  #define NUM_BATCH_FREE_PAGES 10
>  
> +static inline bool persistent_gnt_timeout(struct persistent_gnt *persistent_gnt)
> +{
> +	return xen_blkif_pgrant_timeout &&
> +	       (jiffies - persistent_gnt->last_used >=
> +		HZ * xen_blkif_pgrant_timeout);
> +}
> +
>  static inline int get_free_page(struct xen_blkif_ring *ring, struct page **page)
>  {
>  	unsigned long flags;
> @@ -278,6 +297,7 @@ static void put_persistent_gnt(struct xen_blkif_ring *ring,
>  {
>  	if(!test_bit(PERSISTENT_GNT_ACTIVE, persistent_gnt->flags))
>  		pr_alert_ratelimited("freeing a grant already unused\n");
> +	persistent_gnt->last_used = jiffies;
>  	set_bit(PERSISTENT_GNT_WAS_ACTIVE, persistent_gnt->flags);
>  	clear_bit(PERSISTENT_GNT_ACTIVE, persistent_gnt->flags);
>  	atomic_dec(&ring->persistent_gnt_in_use);
> @@ -374,23 +394,23 @@ static void purge_persistent_gnt(struct xen_blkif_ring *ring)
>  	bool scan_used = false, clean_used = false;
>  	struct rb_root *root;
>  
> -	if (ring->persistent_gnt_c < xen_blkif_max_pgrants ||
> -	    (ring->persistent_gnt_c == xen_blkif_max_pgrants &&
> -	    !ring->blkif->vbd.overflow_max_grants)) {
> -		goto out;
> -	}
> -
>  	if (work_busy(&ring->persistent_purge_work)) {
>  		pr_alert_ratelimited("Scheduled work from previous purge is still busy, cannot purge list\n");
>  		goto out;
>  	}
>  
> -	num_clean = (xen_blkif_max_pgrants / 100) * LRU_PERCENT_CLEAN;
> -	num_clean = ring->persistent_gnt_c - xen_blkif_max_pgrants + num_clean;
> -	num_clean = min(ring->persistent_gnt_c, num_clean);
> -	if ((num_clean == 0) ||
> -	    (num_clean > (ring->persistent_gnt_c - atomic_read(&ring->persistent_gnt_in_use))))
> -		goto out;
> +	if (ring->persistent_gnt_c < xen_blkif_max_pgrants ||
> +	    (ring->persistent_gnt_c == xen_blkif_max_pgrants &&
> +	    !ring->blkif->vbd.overflow_max_grants)) {
> +		num_clean = 0;
> +	} else {
> +		num_clean = (xen_blkif_max_pgrants / 100) * LRU_PERCENT_CLEAN;
> +		num_clean = ring->persistent_gnt_c - xen_blkif_max_pgrants +
> +			    num_clean;
> +		num_clean = min(ring->persistent_gnt_c, num_clean);
> +		pr_debug("Going to purge at least %u persistent grants\n",
> +			 num_clean);
> +	}
>  
>  	/*
>  	 * At this point, we can assure that there will be no calls
> @@ -401,9 +421,7 @@ static void purge_persistent_gnt(struct xen_blkif_ring *ring)
>           * number of grants.
>  	 */
>  
> -	total = num_clean;
> -
> -	pr_debug("Going to purge %u persistent grants\n", num_clean);
> +	total = 0;
>  
>  	BUG_ON(!list_empty(&ring->persistent_purge_list));
>  	root = &ring->persistent_gnts;
> @@ -419,39 +437,42 @@ static void purge_persistent_gnt(struct xen_blkif_ring *ring)
>  
>  		if (test_bit(PERSISTENT_GNT_ACTIVE, persistent_gnt->flags))
>  			continue;
> -		if (!scan_used &&
> +		if (!scan_used && !persistent_gnt_timeout(persistent_gnt) &&
>  		    (test_bit(PERSISTENT_GNT_WAS_ACTIVE, persistent_gnt->flags)))

If you store the jiffies of the time when the grant was last used it
seems like we could get rid of the PERSISTENT_GNT_WAS_ACTIVE flag and
instead use the per-grant jiffies and the jiffies from the last scan
in order to decide which grants to remove?

Thanks, Roger.

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ