[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <Pine.LNX.4.64.0707301114060.743@schroedinger.engr.sgi.com>
Date: Mon, 30 Jul 2007 11:20:04 -0700 (PDT)
From: Christoph Lameter <clameter@....com>
To: "Siddha, Suresh B" <suresh.b.siddha@...el.com>
cc: linux-kernel@...r.kernel.org, arjan@...ux.intel.com, mingo@...e.hu,
npiggin@...e.de, ak@...e.de, jens.axboe@...cle.com,
James.Bottomley@...elEye.com, andrea@...e.de,
akpm@...ux-foundation.org, andrew.vasquez@...gic.com
Subject: Re: [rfc] direct IO submission and completion scalability issues
On Fri, 27 Jul 2007, Siddha, Suresh B wrote:
> We have been looking into the linux kernel direct IO scalability issues with
> database workloads. Comments and suggestions on our below experiments are
> welcome.
This was on an SMP system? These issues are much more pronounced on a NUMA
system. There the locality of the device may be a prime issue.
> In the linux kernel, direct IO requests are not batched at the block layer.
> i.e, as a new request comes in, the request get directly submitted to the
> IO controller on the same cpu that the request originates. And the IO completion
> likely happens on a different cpu which is processing interrupts. This results
> in cacheline bouncing of some of the hot kernel cachelines (like timers, scsi
> cmds, slab, sched, etc) and is becoming an important scalability issue
> as the number of cpus and distance between them increase with multi-core
> and numa.
Yes. The issue is even worse if the submission comes from a remote node.
F.e. If we have a system with a scsi controller on node 2. Now I/O
submission on node 1 and completion on node 2. In that case the
cacheline has to be transferred across the NUMA interlink.
However, you cannot avoid running the completion on the node where the
device sits. The device has all sorts of control structures and if you
would handle the completion on node 1 then it would have to transfer lots
of cachelines that contain device state to node 1.
I think it is better to leave things as is. Or have the I/O submission be
relocated to the node of the device.
> Second experiment which we did was migrating the IO submission to the
> IO completion cpu. Instead of submitting the IO on the same cpu where the
> request arrived, in this experiment the IO submission gets migrated to the
> cpu that is processing IO completions(interrupt). This will minimize the
> access to remote cachelines (that happens in timers, slab, scsi layers). The
> IO submission request is forwarded to the kblockd thread on the cpu receiving
> the interrupts. As part of this, we also made kblockd thread on each cpu as the
> highest priority thread, so that IO gets submitted as soon as possible on the
> interrupt cpu with out any delay. On x86_64 SMP platform with 16 cores, this
> resulted in 2% performance improvement and 3.3% improvement on two node ia64
> platform.
I think that is the right approach. This will also help in cases where I/O
devices can only be accessed from a certain node (NUMA device address
restrictions on some systems may not allow remote cacheline access!)
> Observation #2: This introduces some migration overhead during IO submission.
> With the current prototype, every incoming IO request results in an IPI and
> context switch(to kblockd thread) on the interrupt processing cpu.
> This issue needs to be addressed and main challenge to address is
> the efficient mechanism of doing this IO migration(how much batching to do and
> when to send the migrate request?), so that we don't delay the IO much and at
> the same point, don't cause much overhead during migration.
Right.
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists