[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <Pine.LNX.4.44L0.0702251147500.25265-100000@netrider.rowland.org>
Date: Sun, 25 Feb 2007 11:51:46 -0500 (EST)
From: Alan Stern <stern@...land.harvard.edu>
To: Kernel development list <linux-kernel@...r.kernel.org>
cc: Sarah Bailey <saharabeara@...il.com>,
<usb-hacking@...s.cs.pdx.edu>,
USB development list <linux-usb-devel@...ts.sourceforge.net>
Subject: [linux-usb-devel] usbfs2: Why asynchronous I/O?
This deserves to be discussed on LKML.
Alan Stern
---------- Forwarded message ----------
Date: Sun, 25 Feb 2007 00:57:55 -0800
From: Sarah Bailey <saharabeara@...il.com>
To: David Brownell <david-b@...bell.net>, greg@...ah.com
Cc: usb-hacking@...s.cs.pdx.edu, linux-usb-devel@...ts.sourceforge.net
Subject: [linux-usb-devel] usbfs2: Why asynchronous I/O?
I've been doing some research into how asynchronous I/O is implemented,
and I'm beginning to doubt the usefulness of implementing aio_read
and aio_write in usbfs2. More detail on what I've learned can be found
at http://wiki.cs.pdx.edu/usb/asyncdebate.html
It was a surprise to me that aio_read(3) and aio_write(3) don't actually
end up calling aio_read and aio_write file operations. Instead, GNU
libc spawns threads that call the blocking read and write file
operations.
I haven't seen any evidence that the kernel-side aio is substantially
more efficient than the GNU libc implementation, so it seems like it
would be better to leave the complexity in userspace. I also doubt that
most userspace application writers know they aren't getting kernel-side
aio when they use aio_read(3) and aio_write(3). Why implement something
that isn't going to be used?
There are few examples in the kernel where the aio API is implemented in
a truly asynchronous manner, and that leads me to wonder if the aio
system has been thoroughly tested. The majority of aio_read and
aio_write file operations simply block and wait for their transactions
to complete.
The only "proper" async examples I could find were gadgetfs, NFS, and
block devices. NFS and block devices only use aio when the O_DIRECT
flag is set, so that code may not be well tested. I just found a bug in
gadgetfs that has been there for six months that means the code wasn't
tested for when io_submit is called in readv mode.
So, why do I want a non-blocking aio_read and aio_write file operation?
It's useful to have read and readv implemented automatically from
aio_read, but I see no substantial benefit to implementing aio_read in a
non-blocking way.
Sarah Bailey
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists