lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <200905071418.05035.rjw@sisk.pl>
Date:	Thu, 7 May 2009 14:18:04 +0200
From:	"Rafael J. Wysocki" <rjw@...k.pl>
To:	nigel@...onice.net
Cc:	pm list <linux-pm@...ts.linux-foundation.org>,
	Wu Fengguang <fengguang.wu@...el.com>,
	Andrew Morton <akpm@...ux-foundation.org>,
	LKML <linux-kernel@...r.kernel.org>, Pavel Machek <pavel@....cz>
Subject: Re: [RFC][PATCH 4/5] PM/Hibernate: Rework shrinking of memory

On Thursday 07 May 2009, Nigel Cunningham wrote:
> Hi.

Hi,

> On Thu, 2009-05-07 at 00:44 +0200, Rafael J. Wysocki wrote:
> > From: Rafael J. Wysocki <rjw@...k.pl>
> > 
> > Rework swsusp_shrink_memory() so that it calls shrink_all_memory()
> > just once to make some room for the image and then allocates memory
> > to apply more pressure to the memory management subsystem, if
> > necessary.
> > 
> > Unfortunately, we don't seem to be able to drop shrink_all_memory()
> > entirely just yet, because that would lead to huge performance
> > regressions in some test cases.
> 
> I know it doesn't fit with your current way of doing things, but have
> you considered trying larger order allocations as a means of getting
> memory freed?

Actually, I was thinking about that.  What's your experience with this
approach?

> I have code in tuxonice_prepare_image.c (look for extra_pages_allocated) that
> might be helpful for this purpose.

OK, thanks.  I'll have a look at it.

> > Signed-off-by: Rafael J. Wysocki <rjw@...k.pl>
> > ---
> >  kernel/power/snapshot.c |  145 ++++++++++++++++++++++++++++++++----------------
> >  1 file changed, 98 insertions(+), 47 deletions(-)
> > 
> > Index: linux-2.6/kernel/power/snapshot.c
> > ===================================================================
> > --- linux-2.6.orig/kernel/power/snapshot.c
> > +++ linux-2.6/kernel/power/snapshot.c
> > @@ -1066,69 +1066,120 @@ void swsusp_free(void)
> >  	buffer = NULL;
> >  }
> >  
> > +/* Helper functions used for the shrinking of memory. */
> > +
> >  /**
> > - *	swsusp_shrink_memory -  Try to free as much memory as needed
> > - *
> > - *	... but do not OOM-kill anyone
> > + * preallocate_image_memory - Allocate given number of page frames
> > + * @nr_pages: Number of page frames to allocate
> >   *
> > - *	Notice: all userland should be stopped before it is called, or
> > - *	livelock is possible.
> > + * Return value: Number of page frames actually allocated
> >   */
> > -
> > -#define SHRINK_BITE	10000
> > -static inline unsigned long __shrink_memory(long tmp)
> > +static unsigned long preallocate_image_memory(unsigned long nr_pages)
> >  {
> > -	if (tmp > SHRINK_BITE)
> > -		tmp = SHRINK_BITE;
> > -	return shrink_all_memory(tmp);
> > +	unsigned long nr_alloc = 0;
> > +
> > +	while (nr_pages-- > 0) {
> > +		struct page *page;
> > +
> > +		page = alloc_image_page(GFP_KERNEL | __GFP_NOWARN);
> 
> Ah... now I see you're using __GFP_NOWARN already :)
> 
> > +		if (!page)
> > +			break;
> > +		nr_alloc++;
> > +	}
> > +
> > +	return nr_alloc;
> >  }
> >  
> > +/**
> > + * swsusp_shrink_memory -  Make the kernel release as much memory as needed
> > + *
> > + * To create a hibernation image it is necessary to make a copy of every page
> > + * frame in use.  We also need a number of page frames to be free during
> > + * hibernation for allocations made while saving the image and for device
> > + * drivers, in case they need to allocate memory from their hibernation
> > + * callbacks (these two numbers are given by PAGES_FOR_IO and SPARE_PAGES,
> > + * respectively, both of which are rough estimates).  To make this happen, we
> > + * compute the total number of available page frames and allocate at least
> > + *
> > + * ([page frames total] + PAGES_FOR_IO + [metadata pages]) / 2 + 2 * SPARE_PAGES
> > + *
> > + * of them, which corresponds to the maximum size of a hibernation image.
> > + *
> > + * If image_size is set below the number following from the above formula,
> > + * the preallocation of memory is continued until the total number of page
> > + * frames in use is below the requested image size or it is impossible to
> > + * allocate more memory, whichever happens first.
> > + */
> 
> You should also be taking into account how much storage is available
> here - that would make things more reliable. If compression is begin
> used, you could also apply an 'expected compression ratio' so that you
> don't unnecessarily free memory that will fit once compressed.

Currently compression is only done in user space so I don't know in advance
whether or not it's going to be used.

> >  int swsusp_shrink_memory(void)
> >  {
> > -	long tmp;
> >  	struct zone *zone;
> > -	unsigned long pages = 0;
> > -	unsigned int i = 0;
> > -	char *p = "-\\|/";
> > +	unsigned long saveable, size, max_size, count, pages = 0;
> >  	struct timeval start, stop;
> > +	int error = 0;
> >  
> > -	printk(KERN_INFO "PM: Shrinking memory...  ");
> > +	printk(KERN_INFO "PM: Shrinking memory ... ");
> 
> Without the space is normal, at least to my mind.

OK

> >  	do_gettimeofday(&start);
> > -	do {
> > -		long size, highmem_size;
> >  
> > -		highmem_size = count_highmem_pages();
> > -		size = count_data_pages() + PAGES_FOR_IO + SPARE_PAGES;
> > -		tmp = size;
> > -		size += highmem_size;
> > -		for_each_populated_zone(zone) {
> > -			tmp += snapshot_additional_pages(zone);
> > -			if (is_highmem(zone)) {
> > -				highmem_size -=
> > -					zone_page_state(zone, NR_FREE_PAGES);
> > -			} else {
> > -				tmp -= zone_page_state(zone, NR_FREE_PAGES);
> > -				tmp += zone->lowmem_reserve[ZONE_NORMAL];
> > -			}
> > -		}
> > +	/* Count the number of saveable data pages. */
> > +	saveable = count_data_pages() + count_highmem_pages();
> > +
> > +	/*
> > +	 * Compute the total number of page frames we can use (count) and the
> > +	 * number of pages needed for image metadata (size).
> > +	 */
> > +	count = saveable;
> > +	size = 0;
> > +	for_each_populated_zone(zone) {
> > +		size += snapshot_additional_pages(zone);
> > +		count += zone_page_state(zone, NR_FREE_PAGES);
> > +		if (!is_highmem(zone))
> > +			count -= zone->lowmem_reserve[ZONE_NORMAL];
> > +	}
> >  
> > -		if (highmem_size < 0)
> > -			highmem_size = 0;
> 
> You're not taking watermarks into account here - that isn't a problem
> with shrink_all_memory because it usually frees more than you ask for
> (or has done in the past), but if you're getting exactly what you ask
> for, you might run into trouble if more than half of memory is in use to
> start with.

Hmm, why exactly?

> > +	/* Compute the maximum number of saveable pages to leave in memory. */
> > +	max_size = (count - (size + PAGES_FOR_IO)) / 2 - 2 * SPARE_PAGES;
> > +	size = DIV_ROUND_UP(image_size, PAGE_SIZE);
> > +	if (size > max_size)
> > +		size = max_size;
> > +	/*
> > +	 * If the maximum is not lesser than the current number of saveable
> 
> s/lesser/less/

Right, thanks.

> > +	 * pages in memory, we don't need to do anything more.
> > +	 */
> > +	if (size >= saveable)
> > +		goto out;
> >  
> > -		tmp += highmem_size;
> > -		if (tmp > 0) {
> > -			tmp = __shrink_memory(tmp);
> > -			if (!tmp)
> > -				return -ENOMEM;
> > -			pages += tmp;
> > -		} else if (size > image_size / PAGE_SIZE) {
> > -			tmp = __shrink_memory(size - (image_size / PAGE_SIZE));
> > -			pages += tmp;
> > -		}
> > -		printk("\b%c", p[i++%4]);
> > -	} while (tmp > 0);
> > +	/*
> > +	 * Let the memory management subsystem know that we're going to need a
> > +	 * large number of page frames to allocate and make it free some memory.
> > +	 * NOTE: If this is not done, performance is heavily affected in some
> > +	 * test cases.
> > +	 */
> > +	shrink_all_memory(saveable - size);
> > +
> > +	/*
> > +	 * The number of saveable pages in memory was too high, so apply some
> > +	 * pressure to decrease it.  First, make room for the largest possible
> > +	 * image and fail if that doesn't work.  Next, try to decrease the size
> > +	 * of the image as much as indicated by image_size.
> > +	 */
> > +	count -= max_size;
> > +	pages = preallocate_image_memory(count);
> > +	if (pages < count)
> > +		error = -ENOMEM;
> > +	else
> > +		pages += preallocate_image_memory(max_size - size);
> > +
> > +	/* Release all of the preallocated page frames. */
> > +	swsusp_free();
> > +
> > +	if (error) {
> > +		printk(KERN_CONT "\n");
> > +		return error;
> > +	}
> > +
> > + out:
> >  	do_gettimeofday(&stop);
> > -	printk("\bdone (%lu pages freed)\n", pages);
> > +	printk(KERN_CONT "done (preallocated %lu free pages)\n", pages);
> >  	swsusp_show_speed(&start, &stop, pages, "Freed");
> >  
> >  	return 0;

Best,
Rafael
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ