[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20070516061259.GZ19966@holomorphy.com>
Date: Tue, 15 May 2007 23:12:59 -0700
From: William Lee Irwin III <wli@...omorphy.com>
To: dean gaudet <dean@...tic.org>
Cc: linux-kernel@...r.kernel.org
Subject: Re: 2.6.21 numa policy and huge pages not working
On Tue, May 15, 2007 at 10:41:06PM -0700, dean gaudet wrote:
> prior to 2.6.21 i could "numactl --interleave=all" and use SHM_HUGETLB and
> the interleave policy would be respected. as of 2.6.21 it doesn't seem to
> respect the policy on SHM_HUGETLB request.
> see test program below.
> output from pre-2.6.21:
> 2ab196200000 interleave=0-3 file=/2\040(deleted) huge dirty=32 N0=8 N1=8 N2=8 N3=8
> 2ab19a200000 default file=/SYSV00000000\040(deleted) dirty=16384 active=0 N0=4096 N1=4096 N2=4096 N3=4096
> output from 2.6.21:
> 2b49b1c00000 default file=/10\040(deleted) huge dirty=32 N3=32
> 2b49b5c00000 default file=/SYSV00000000\040(deleted) dirty=16384 active=0 N0=4096 N1=4096 N2=4096 N3=4096
> was this an intentional behaviour change? it seems to be only affecting
> SHM_HUGETLB allocations. (i haven't tested hugetlbfs yet.)
> run with "numactl --interleave=all ./shmtest"
This was not intentional. I'll search for where it broke.
-- wli
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists