[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <tkrat.ca8b7763ca31e24d@s5r6.in-berlin.de>
Date: Sat, 17 Jan 2009 22:45:54 +0100 (CET)
From: Stefan Richter <stefanr@...6.in-berlin.de>
To: linux1394-devel@...ts.sourceforge.net
cc: linux-kernel@...r.kernel.org
Subject: [PATCH 0/4] firewire: cope with temporary connection loss
In the following series, the 1st patch enables the firewire stack to
tolerate brief disappearance of nodes similarly as the ieee1394 stack
has been taught in Linux 2.6.28.
This finally allows me to push the patch
Date: Wed, 19 Mar 2008 22:02:40 +0100 (CET)
From: Stefan Richter <stefanr@...6.in-berlin.de>
Subject: firewire: insist on successive self ID complete events
into the mainline, which is a fix that potentially prevents kernel
panics. This fix had the unfortunate side effect that the firewire-core
topology code destroyed and re-created node representations more often
than before. Until now, this caused high-level drivers (firewire-sbp2
and userspace drivers) to be unbound and rebound, so that for example
mounted filesystems were lost.
Besides the condition which triggered the "insist on successive self ID
complete events" precautions, there are also other occasions at which
nodes vanish briefly and reappear. See the changelog of patch 1/4.
firewire-core is now able to keep drivers bound until the nodes come
back, if they come back quickly.
The latter 3 patches are cleanups.
[PATCH 1/4] firewire: keep highlevel drivers attached during brief connection loss
[PATCH 2/4] firewire: core: clean up includes
[PATCH 3/4] firewire: core: move some functions
[PATCH 4/4] firewire: core: remove condition which is always false
drivers/firewire/fw-device.c | 231 +++++++++++++++++++++++------------
drivers/firewire/fw-device.h | 14 +-
2 files changed, 166 insertions(+), 79 deletions(-)
--
Stefan Richter
-=====-==--= ---= =---=
http://arcgraph.de/sr/
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists