[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <496F4AD4.9080201@cosmosbay.com>
Date: Thu, 15 Jan 2009 15:40:20 +0100
From: Eric Dumazet <dada1@...mosbay.com>
To: Pekka Pietikainen <pp@...oulu.fi>
CC: Mihai Donțu <mihai.dontu@...il.com>,
Herbert Xu <herbert@...dor.apana.org.au>,
linux-kernel@...r.kernel.org, netdev@...r.kernel.org
Subject: Re: b44 - swapper: page allocation failure
Pekka Pietikainen a écrit :
> On Thu, Jan 15, 2009 at 03:27:43PM +0200, Mihai Donțu wrote:
>>>> A friend of mine just booted 2.6.28 and when someone else tried to send
>>>> something to him over ssh (scp), the transfer failed and the following
>>>> appeared in dmesg:
>>> You missed the all important order line before the stack trace.
>> Right, sorry for that. Here it is:
>> swapper: page allocation failure. order:0, mode:0x21
>>> If it's order 0, then you're just out of memory, if it's greater
>>> than order 0, then either you're using jumbo frames or b44 is
>>> broken.
>> It is order zero, but he's not out of memory. He has just booted 2.6.28 (no X,
>> no thing), has 3GB of RAM and 2GB of swap.
> b44 uses GFP_DMA bounce buffers in some situations (x86_64 with > 1GB of
> memory would do the trick I think, as does x86 with 4:4 memory split),
> and that's a very limited resource (Memory < 16MB ).
>
> Lovely hardware feature requiring nasty workarounds
> (the chip can't do DMA for addresses > 1GB).
>
> Maybe something else sucks up GFP_DMA memory with 2.6.28?
Normal tcp queues (in this case, since a single scp is enough
to trigger the bug)
If a rx buffer was allocated using GFP_DMA, then b44_rx()
should copy it to a new skb, (like done by copybreak algo),
regardless of frame size, to not consume DMA memory in protocol
queues.
Better be slow (because of extra copy) then risking exhausting DMA zone
easy way to test would be to redefine RX_COPY_THRESHOLD to 16000
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists