[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <Pine.LNX.4.64.0804062213230.18148@schroedinger.engr.sgi.com>
Date: Sun, 6 Apr 2008 22:26:54 -0700 (PDT)
From: Christoph Lameter <clameter@....com>
To: Hugh Dickins <hugh@...itas.com>
cc: James Bottomley <James.Bottomley@...senPartnership.com>,
Linus Torvalds <torvalds@...ux-foundation.org>,
Andrew Morton <akpm@...ux-foundation.org>,
FUJITA Tomonori <fujita.tomonori@....ntt.co.jp>,
Jens Axboe <jens.axboe@...cle.com>,
Pekka Enberg <penberg@...helsinki.fi>,
Peter Zijlstra <a.p.ziljstra@...llo.nl>,
"Rafael J. Wysocki" <rjw@...k.pl>, linux-kernel@...r.kernel.org
Subject: Re: [PATCH] scsi: fix sense_slab/bio swapping livelock
On Sun, 6 Apr 2008, Hugh Dickins wrote:
> __scsi_get_command has a (one item) free_list to protect against
> this, but rc1's [SCSI] use dynamically allocated sense buffer
> de25deb18016f66dcdede165d07654559bb332bc upset that slightly.
> When it fails to allocate from the separate sense_slab, instead
> of giving up, it must fall back to the command free_list, which
> is sure to have a sense_buffer attached.
>
> Either my earlier -rc testing missed this, or there's some recent
> contributory factor. One very significant factor is SLUB, which
> merges slab caches when it can, and on 64-bit happens to merge
> both bio cache and sense_slab cache into kmalloc's 128-byte cache:
> so that under this swapping load, bios above are liable to gobble
> up all the slots needed for scsi_cmnd sense_buffers below.
A reliance on free slots that the slab allocator may provide? That is a
rather bad dependency since it is up to the slab allocator to implement
the storage layout for the objects and thus the availability of slots may
vary depending on the layout for the objects chosen by the allocator.
Looking at mempool_alloc: Mempools may be used to do atomic allocations
until they fail thereby exhausting reserves and available object in the
partial lists of slab caches?
In order to make this a significant factor we need to have already
exhausted reserves right? Thus we are already operating at the boundary of
what memory there is. Any non atomic alloc will then allocate a new page
with N elements in order to get one object. The mempool_allocs from the
atomic context will then gooble up the N-1 remaining objects? So the
nonatomic alloc will then have to hit the page allocator again...
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists