[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <Pine.LNX.4.64.0705211358210.28867@schroedinger.engr.sgi.com>
Date: Mon, 21 May 2007 14:04:33 -0700 (PDT)
From: Christoph Lameter <clameter@....com>
To: Peter Zijlstra <a.p.zijlstra@...llo.nl>
cc: Matt Mackall <mpm@...enic.com>, linux-kernel@...r.kernel.org,
linux-mm@...ck.org, Thomas Graf <tgraf@...g.ch>,
David Miller <davem@...emloft.net>,
Andrew Morton <akpm@...ux-foundation.org>,
Daniel Phillips <phillips@...gle.com>,
Pekka Enberg <penberg@...helsinki.fi>,
Paul Jackson <pj@....com>, npiggin@...e.de
Subject: Re: [PATCH 0/5] make slab gfp fair
On Mon, 21 May 2007, Peter Zijlstra wrote:
> > Yes sure if we do not have a context then no restrictions originating
> > there can be enforced. So you want to restrict the logic now to
> > interrupt allocs? I.e. GFP_ATOMIC?
>
> No, any kernel alloc.
Then we have the problem again.
> > Correct. That is an optimization but it may be called anytime from the
> > perspective of an execution thread and that may cause problems with your
> > approach.
>
> I'm not seeing how this would interfere; if the alloc can be handled
> from a partial slab, that is fine.
There is no guarantee that a partial slab is available.
> > In the case
> > of a memory policy we may have limited the allocations to a single node
> > where there is no escape (the zonelist does *not* contain zones of other
> > nodes).
>
> Ah, this is the point I was missing; I assumed each zonelist would
> always include all zones, but would just continue/break the loop using
> things like cpuset_zone_allwed_*().
>
> This might indeed foil the game.
>
> I could 'fix' this by doing the PF_MEMALLOC allocation from the regular
> node zonelist instead of from the one handed down....
I wonder if this makes any sense at all given that the only point of
what you are doing is to help to decide which alloc should fail...
> /me thinks out loud.. since direct reclaim runs in whatever process
> context was handed out we're stuck with whatever policy we started from;
> but since the allocations are kernel allocs - not userspace allocs, and
> we're in dire straights, it makes sense to violate the tasks restraints
> in order to keep the machine up.
The memory policy constraints may have been setup to cage in an
application. It was setup to *stop* the application from using memory on
other nodes. If you now allow that then the semantics of memory policies
are significantly changed. The cpuset constraints are sometimes not that
hard but I better let Paul speak for them.
> memory policies are the only ones with 'short' zonelists, right? CPU
> sets are on top of whatever zonelist is handed out, and the normal
> zonelists include all nodes - ordered by distance
GFP_THISNODE can have a similar effect.
> > The only chance to bypass this is by only dealing with allocations
> > during interrupt that have no allocation context.
>
> But you just said that interrupts are not exempt from memory policies,
> and policies are the only ones that have 'short' zonelists. /me
> confused.
No I said that in an interrupt allocation we have no process context and
therefore no cpuset or memory policy context. Thus no policies or cpusets
are applied to an allocation. You can allocate without restrictions.
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists