lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Wed, 25 Jan 2012 10:27:28 -0500
From:	Rik van Riel <riel@...hat.com>
To:	Mel Gorman <mel@....ul.ie>
CC:	linux-mm@...ck.org, lkml <linux-kernel@...r.kernel.org>,
	Andrea Arcangeli <aarcange@...hat.com>,
	Johannes Weiner <hannes@...xchg.org>,
	Andrew Morton <akpm@...ux-foundation.org>,
	Minchan Kim <minchan.kim@...il.com>,
	KOSAKI Motohiro <kosaki.motohiro@...il.com>
Subject: Re: [PATCH v2 -mm 1/3] mm: reclaim at order 0 when compaction is
 enabled

On 01/25/2012 10:00 AM, Mel Gorman wrote:
> On Tue, Jan 24, 2012 at 01:21:36PM -0500, Rik van Riel wrote:
>> When built with CONFIG_COMPACTION, kswapd does not try to free
>> contiguous pages.
>
> balance_pgdat() gets its order from wakeup_kswapd(). This does not apply
> to THP because kswapd does not get woken for THP but it should be woken
> up for allocations like jumbo frames or order-1.

In the kernel I run at home, I wake up kswapd for THP
as well. This is a larger change, which Andrea asked
me to delay submitting upstream for a bit.

So far there seem to be no ill effects. I'll continue
watching for them.

> As kswapd does no memory compaction itself, this patch still makes
> sense but I found the changelog misleading.

Fair enough.  I will adjust the changelog.

>> diff --git a/mm/vmscan.c b/mm/vmscan.c
>> index 2880396..0398fab 100644
>> --- a/mm/vmscan.c
>> +++ b/mm/vmscan.c
>> @@ -1512,6 +1512,7 @@ shrink_inactive_list(unsigned long nr_to_scan, struct mem_cgroup_zone *mz,
>>   	unsigned long nr_writeback = 0;
>>   	isolate_mode_t reclaim_mode = ISOLATE_INACTIVE;
>>   	struct zone *zone = mz->zone;
>> +	int order = 0;
>>
>>   	while (unlikely(too_many_isolated(zone, file, sc))) {
>>   		congestion_wait(BLK_RW_ASYNC, HZ/10);
>> @@ -1522,8 +1523,10 @@ shrink_inactive_list(unsigned long nr_to_scan, struct mem_cgroup_zone *mz,
>>   	}
>>
>>   	set_reclaim_mode(priority, sc, false);
>> -	if (sc->reclaim_mode&  RECLAIM_MODE_LUMPYRECLAIM)
>> +	if (sc->reclaim_mode&  RECLAIM_MODE_LUMPYRECLAIM) {
>>   		reclaim_mode |= ISOLATE_ACTIVE;
>> +		order = sc->order;
>> +	}
>>
>>   	lru_add_drain();
>>
>
> This is a nit-pick but I would far prefer if you did not bypass
> sc->order like this and instead changes isolate_lru_pages to do a
>
> if (!order || !(sc->reclaim_mode&  RECLAIM_MODE_LUMPYRECLAIM))
> 	continue;
>
> That would very clearly mark where LUMPYRECLAIM takes effect in
> isolate_lru_pages() and makes deleting LUMPYRECLAIM easier in the
> future.

OK, I will do things that way instead.

> The second effect of this change is a non-obvious side-effect. kswapd
> will now isolate fewer pages per cycle because it will isolate
> SWAP_CLUSTER_MAX pages instead of SWAP_CLUSTER_MAX<<order which it
> potentially does currently. This is not wrong as such and may be
> desirable to limit how much reclaim kswapd does but potentially it
> impacts success rates for compaction. As this does not apply to THP,
> it will be difficult to detect but bear in mind if we see an increase
> in high-order allocation failures after this patch is merged. I am
> not suggesting a change here but it would be nice to note in the
> changelog if there is a new version of this patch.

Good point.  I am running with THP waking up kswapd, and
things seem to behave (and compaction seems to succeed),
but we might indeed want to change balance_pgdat to free
more pages for higher order allocations.

Maybe this is the place to check (in balanced_pgdat) ?

                 /*
                  * We do this so kswapd doesn't build up large 
priorities for
                  * example when it is freeing in parallel with 
allocators. It
                  * matches the direct reclaim path behaviour in terms 
of impact
                  * on zone->*_priority.
                  */
                 if (sc.nr_reclaimed >= SWAP_CLUSTER_MAX)
                         break;

>> @@ -2922,8 +2939,6 @@ out:
>>
>>   			/* If balanced, clear the congested flag */
>>   			zone_clear_flag(zone, ZONE_CONGESTED);
>> -			if (i<= *classzone_idx)
>> -				balanced += zone->present_pages;
>>   		}
>
> Why is this being deleted? It is still used by pgdat_balanced().

This is outside of the big while loop and is not used again
in the function.  This final for loop does not appear to
use the variable balanced at all, except for incrementing
it.

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ