[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20190824021254.GB127465@google.com>
Date: Fri, 23 Aug 2019 21:12:54 -0500
From: Bjorn Helgaas <helgaas@...nel.org>
To: Mika Westerberg <mika.westerberg@...ux.intel.com>
Cc: "Rafael J. Wysocki" <rjw@...ysocki.net>,
Len Brown <lenb@...nel.org>, Lukas Wunner <lukas@...ner.de>,
Keith Busch <keith.busch@...el.com>,
Alex Williamson <alex.williamson@...hat.com>,
Alexandru Gagniuc <mr.nuke.me@...il.com>,
Kai-Heng Feng <kai.heng.feng@...onical.com>,
Matthias Andree <matthias.andree@....de>,
Paul Menzel <pmenzel@...gen.mpg.de>,
Nicholas Johnson <nicholas.johnson-opensource@...look.com.au>,
Justin Forbes <jmforbes@...uxtx.org>,
linux-pci@...r.kernel.org, linux-kernel@...r.kernel.org
Subject: Re: [PATCH] PCI: Add missing link delays required by the PCIe spec
Hi Mika,
I'm trying to figure out specifically why we need this and where it
should go. Questions below.
On Wed, Aug 21, 2019 at 03:45:19PM +0300, Mika Westerberg wrote:
> Currently Linux does not follow PCIe spec regarding the required delays
> after reset. A concrete example is a Thunderbolt add-in-card that
> consists of a PCIe switch and two PCIe endpoints:
>
> +-1b.0-[01-6b]----00.0-[02-6b]--+-00.0-[03]----00.0 TBT controller
> +-01.0-[04-36]-- DS hotplug port
> +-02.0-[37]----00.0 xHCI controller
> \-04.0-[38-6b]-- DS hotplug port
>
> The root port (1b.0) and the PCIe switch downstream ports are all PCIe
> gen3 so they support 8GT/s link speeds.
>
> We wait for the PCIe hierarchy to enter D3cold (runtime):
>
> pcieport 0000:00:1b.0: power state changed by ACPI to D3cold
>
> When it wakes up from D3cold, according to the PCIe 4.0 section 5.8 the
> PCIe switch is put to reset and its power is re-applied. This means that
> we must follow the rules in PCIe 4.0 section 6.6.1.
>
> For the PCIe gen3 ports we are dealing with here, the following applies:
>
> With a Downstream Port that supports Link speeds greater than 5.0
> GT/s, software must wait a minimum of 100 ms after Link training
> completes before sending a Configuration Request to the device
> immediately below that Port. Software can determine when Link training
> completes by polling the Data Link Layer Link Active bit or by setting
> up an associated interrupt (see Section 6.7.3.3).
>
> Translating this into the above topology we would need to do this (DLLLA
> stands for Data Link Layer Link Active):
>
> pcieport 0000:00:1b.0: wait for 100ms after DLLLA is set before access to 0000:01:00.0
> pcieport 0000:02:00.0: wait for 100ms after DLLLA is set before access to 0000:03:00.0
> pcieport 0000:02:02.0: wait for 100ms after DLLLA is set before access to 0000:37:00.0
>
> I've instrumented the kernel with additional logging so we can see the
> actual delays the kernel performs:
>
> pcieport 0000:00:1b.0: power state changed by ACPI to D0
> pcieport 0000:00:1b.0: waiting for D3cold delay of 100 ms
> pcieport 0000:00:1b.0: waking up bus
> pcieport 0000:00:1b.0: waiting for D3hot delay of 10 ms
> pcieport 0000:00:1b.0: restoring config space at offset 0x2c (was 0x60, writing 0x60)
> ...
> pcieport 0000:00:1b.0: PME# disabled
> pcieport 0000:01:00.0: restoring config space at offset 0x3c (was 0x1ff, writing 0x201ff)
> ...
> pcieport 0000:01:00.0: PME# disabled
> pcieport 0000:02:00.0: restoring config space at offset 0x3c (was 0x1ff, writing 0x201ff)
> ...
> pcieport 0000:02:00.0: PME# disabled
> pcieport 0000:02:01.0: restoring config space at offset 0x3c (was 0x1ff, writing 0x201ff)
> ...
> pcieport 0000:02:01.0: restoring config space at offset 0x4 (was 0x100000, writing 0x100407)
> pcieport 0000:02:01.0: PME# disabled
> pcieport 0000:02:02.0: restoring config space at offset 0x3c (was 0x1ff, writing 0x201ff)
> ...
> pcieport 0000:02:02.0: PME# disabled
> pcieport 0000:02:04.0: restoring config space at offset 0x3c (was 0x1ff, writing 0x201ff)
> ...
> pcieport 0000:02:04.0: PME# disabled
> pcieport 0000:02:01.0: PME# enabled
> pcieport 0000:02:01.0: waiting for D3hot delay of 10 ms
> pcieport 0000:02:04.0: PME# enabled
> pcieport 0000:02:04.0: waiting for D3hot delay of 10 ms
> thunderbolt 0000:03:00.0: restoring config space at offset 0x14 (was 0x0, writing 0x8a040000)
> ...
> thunderbolt 0000:03:00.0: PME# disabled
> xhci_hcd 0000:37:00.0: restoring config space at offset 0x10 (was 0x0, writing 0x73f00000)
> ...
> xhci_hcd 0000:37:00.0: PME# disabled
>
> For the switch upstream port (01:00.0) we wait for 100ms but not taking
> into account the DLLLA requirement. We then wait 10ms for D3hot -> D0
> transition of the root port and the two downstream hotplug ports. This
> means that we deviate from what the spec requires.
>
> Performing the same check for system sleep (s2idle) transitions we can
> see following when resuming from s2idle:
>
> pcieport 0000:00:1b.0: power state changed by ACPI to D0
> pcieport 0000:00:1b.0: restoring config space at offset 0x2c (was 0x60, writing 0x60)
> ...
> pcieport 0000:01:00.0: restoring config space at offset 0x3c (was 0x1ff, writing 0x201ff)
> ...
I think the important thing in all the above logging is that it
doesn't show any delay, right? If that's the case, you can just say
that in one line; I trust you even without 40 lines of config space
restore debug output :)
> xhci_hcd 0000:37:00.0: restoring config space at offset 0x10 (was 0x0, writing 0x73f00000)
> ...
> thunderbolt 0000:03:00.0: restoring config space at offset 0x14 (was 0x0, writing 0x8a040000)
>
> This is even worse. None of the mandatory delays are performed. If this
> would be S3 instead of s2idle then according to PCI FW spec 3.2 section
> 4.6.8. there is a specific _DSM that allows the OS to skip the delays
> but this platform does not provide the _DSM and does not go to S3 anyway
> so no firmware is involved that could already handle these delays.
>
> In this particular Intel Coffee Lake platform these delays are not
> actually needed because there is an additional delay as part of the ACPI
> power resource that is used to turn on power to the hierarchy but since
> that additional delay is not required by any of standards (PCIe, ACPI)
So it sounds like this Coffee Lake accidentally works because of
unrelated firmware delay that's not actually required, or at least not
related to the delay required by PCIe?
I did notice that we don't implement all of _DSM function 9 and the
parts we're missing look like they could be relevant.
> it is not present in the Intel Ice Lake, for example where missing the
> mandatory delays causes pciehp to start tearing down the stack too early
> (links are not yet trained).
I'm guessing the Coffee Lake/Ice Lake distinction is not really
relevant and the important thing is that something about Ice Lake is
faster and reduces the accidental delay to the point that things stop
working.
So probably the same thing could happen on Coffee Lake or any other
system if it had different firmware.
It *would* be interesting to mention what the pciehp/link training
issue looks like to a user, just to help people who see it find this
fix.
> There is also one reported case (see the bugzilla link below) where the
> missing delay causes xHCI on a Titan Ridge controller fail to runtime
> resume when USB-C dock is plugged.
>
> For this reason, introduce a new function pcie_wait_downstream_accessible()
> that is called on PCI core resume and runtime resume paths accordingly
> if downstream/root port with device below entered D3cold.
>
> This is second attempt to add the missing delays. The previous solution
> in commit c2bf1fc212f7 ("PCI: Add missing link delays required by the
> PCIe spec") was reverted because of two issues it caused:
c2bf1fc212f7 was merged for v5.3 (it appeared in v5.3-rc1). I *guess*
it addressed https://bugzilla.kernel.org/show_bug.cgi?id=203885, which
was reported on v5.2-rc4, though c2bf1fc212f7 doesn't actually mention
the bugzilla?
0617bdede511 ("Revert "PCI: Add missing link delays required by the
PCIe spec"") appeared in v5.3-rc4. If c2bf1fc212f7 was supposed to
address BZ 203885, I'm confused about why you asked Kai-Heng to test
v5.3-rc4, where c2bf1fc212f7 had already been reverted.
Or maybe c2bf1fc212f7 wasn't connected with BZ 203885 in the first
place?
The net result is that I think v5.3-rc4 is equivalent to v5.2 with
respect to this issue. It *is* a pretty serious usability issue, no
question, but it's not completely obvious that this fix needs to be in
v5.3 since it's not something we broke during the v5.3 merge window.
So if we *do* want it in v5.3, we need to think about how to justify
that, e.g., if this issue affects shipping systems that are likely to
run an upstream kernel, etc.
> 1. One system become unresponsive after S3 resume due to PME service
> spinning in pcie_pme_work_fn(). The root port in question reports
> that the xHCI sent PME but the xHCI device itself does not have PME
> status set. The PME status bit is never cleared in the root port
> resulting the indefinite loop in pcie_pme_work_fn().
I don't see the connection between PME and either c2bf1fc212f7 or this
patch. Is there a fix for this pcie_pme_work_fn() infinite loop?
I do see that BZ 203885 mentions a flood of "PME: Spurious native
interrupt!" messages, but I don't see how we've fixed that.
> 2. Slows down resume if the root/downstream port does not support
> Data Link Layer Active Reporting because pcie_wait_for_link_delay()
> waits 1100ms in that case.
I don't see the slowdown mentioned in BZ 203885; is there another link
to these reports?
> This version should avoid the above issues because we restrict the delay
> to happen only if the port went into D3cold (so it goes through reset)
> and only when there is no firmware involved on resume path (so the
> kernel is responsible for all the delays).
>
> Bugzilla: https://bugzilla.kernel.org/show_bug.cgi?id=203885
> Reported-by: Kai-Heng Feng <kai.heng.feng@...onical.com>
> Tested-by: Kai-Heng Feng <kai.heng.feng@...onical.com>
> Signed-off-by: Mika Westerberg <mika.westerberg@...ux.intel.com>
> ---
> Hi all,
>
> As the changelog says this is reworked version that tries to avoid reported
> issues while at the same time fix the missing delays so we can get ICL
> systems and at least the one system with Titan Ridge controller working
> properly.
>
> @Matthias, @Paul and @Nicholas: it would be great if you could try the
> patch on top of v5.4-rc5+ and verify that it does not cause any issues on
> your systems.
>
> drivers/pci/pci-driver.c | 19 ++++++
> drivers/pci/pci.c | 127 ++++++++++++++++++++++++++++++++++++---
> drivers/pci/pci.h | 1 +
> 3 files changed, 137 insertions(+), 10 deletions(-)
>
> diff --git a/drivers/pci/pci-driver.c b/drivers/pci/pci-driver.c
> index a8124e47bf6e..9aec78ed8907 100644
> --- a/drivers/pci/pci-driver.c
> +++ b/drivers/pci/pci-driver.c
> @@ -918,6 +918,7 @@ static int pci_pm_resume_noirq(struct device *dev)
> {
> struct pci_dev *pci_dev = to_pci_dev(dev);
> struct device_driver *drv = dev->driver;
> + pci_power_t state = pci_dev->current_state;
> int error = 0;
>
> if (dev_pm_may_skip_resume(dev))
> @@ -947,6 +948,15 @@ static int pci_pm_resume_noirq(struct device *dev)
>
> pcie_pme_root_status_cleanup(pci_dev);
>
> + /*
> + * If resume involves firmware assume it takes care of any delays
> + * for now. For suspend-to-idle case we need to do that here before
> + * resuming PCIe port services to keep pciehp from tearing down the
> + * downstream devices too early.
> + */
> + if (state == PCI_D3cold && pm_suspend_no_platform())
> + pcie_wait_downstream_accessible(pci_dev);
Aren't these paths used for Conventional PCI devices as well as PCIe
devices?
I think the D3cold and pm_suspend_no_platform() checks should move
inside the new interface, whatever it's called. I'm not sure what
that means for the fact that you don't check pm_suspend_no_platform()
in the runtime-resume path; maybe it needs a flag or something.
But the "wait downstream" part seems a little too specific to be at
the .resume_noirq and .runtime_resume level.
Do we descend the hierarchy and call .resume_noirq and .runtime_resume
for the children of the bridge, too?
> if (drv && drv->pm && drv->pm->resume_noirq)
> error = drv->pm->resume_noirq(dev);
>
> @@ -1329,6 +1339,7 @@ static int pci_pm_runtime_resume(struct device *dev)
> int rc = 0;
> struct pci_dev *pci_dev = to_pci_dev(dev);
> const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL;
> + pci_power_t state = pci_dev->current_state;
>
> /*
> * Restoring config space is necessary even if the device is not bound
> @@ -1344,6 +1355,14 @@ static int pci_pm_runtime_resume(struct device *dev)
> pci_enable_wake(pci_dev, PCI_D0, false);
> pci_fixup_device(pci_fixup_resume, pci_dev);
>
> + /*
> + * If the hierarcy went into D3cold wait for the link to be
> + * reactivated before resuming PCIe port services to keep pciehp
> + * from tearing down the downstream devices too early.
s/hierarcy/hierarchy/
> + */
> + if (state == PCI_D3cold)
> + pcie_wait_downstream_accessible(pci_dev);
> +
> if (pm && pm->runtime_resume)
> rc = pm->runtime_resume(dev);
>
> diff --git a/drivers/pci/pci.c b/drivers/pci/pci.c
> index 1b27b5af3d55..9ac50710f1d4 100644
> --- a/drivers/pci/pci.c
> +++ b/drivers/pci/pci.c
> @@ -1025,15 +1025,11 @@ static void __pci_start_power_transition(struct pci_dev *dev, pci_power_t state)
> if (state == PCI_D0) {
> pci_platform_power_transition(dev, PCI_D0);
> /*
> - * Mandatory power management transition delays, see
> - * PCI Express Base Specification Revision 2.0 Section
> - * 6.6.1: Conventional Reset. Do not delay for
> - * devices powered on/off by corresponding bridge,
> - * because have already delayed for the bridge.
> + * Mandatory power management transition delays are handled
> + * in pci_pm_runtime_resume() of the corresponding
> + * downstream/root port.
> */
> if (dev->runtime_d3cold) {
> - if (dev->d3cold_delay && !dev->imm_ready)
> - msleep(dev->d3cold_delay);
> /*
> * When powering on a bridge from D3cold, the
> * whole hierarchy may be powered on into
> @@ -4607,14 +4603,17 @@ static int pci_pm_reset(struct pci_dev *dev, int probe)
>
> return pci_dev_wait(dev, "PM D3->D0", PCIE_RESET_READY_POLL_MS);
> }
> +
> /**
> - * pcie_wait_for_link - Wait until link is active or inactive
> + * pcie_wait_for_link_delay - Wait until link is active or inactive
> * @pdev: Bridge device
> * @active: waiting for active or inactive?
> + * @delay: Delay to wait after link has become active (in ms)
> *
> * Use this to wait till link becomes active or inactive.
> */
> -bool pcie_wait_for_link(struct pci_dev *pdev, bool active)
> +static bool pcie_wait_for_link_delay(struct pci_dev *pdev, bool active,
> + int delay)
> {
> int timeout = 1000;
> bool ret;
> @@ -4651,13 +4650,121 @@ bool pcie_wait_for_link(struct pci_dev *pdev, bool active)
> timeout -= 10;
> }
> if (active && ret)
> - msleep(100);
> + msleep(delay);
> else if (ret != active)
> pci_info(pdev, "Data Link Layer Link Active not %s in 1000 msec\n",
> active ? "set" : "cleared");
> return ret == active;
> }
>
> +/**
> + * pcie_wait_for_link - Wait until link is active or inactive
> + * @pdev: Bridge device
> + * @active: waiting for active or inactive?
> + *
> + * Use this to wait till link becomes active or inactive.
> + */
> +bool pcie_wait_for_link(struct pci_dev *pdev, bool active)
> +{
> + return pcie_wait_for_link_delay(pdev, active, 100);
> +}
This part (adding pcie_wait_for_link_delay() to make the delay time
configurable) could be a separate patch to reduce the size of this
patch and make the important pieces a little more obvious.
> +static int pcie_get_downstream_delay(struct pci_bus *bus)
> +{
> + struct pci_dev *pdev;
> + int min_delay = 100;
> + int max_delay = 0;
> +
> + list_for_each_entry(pdev, &bus->devices, bus_list) {
> + if (pdev->imm_ready)
> + min_delay = 0;
> + else if (pdev->d3cold_delay < min_delay)
> + min_delay = pdev->d3cold_delay;
> + if (pdev->d3cold_delay > max_delay)
> + max_delay = pdev->d3cold_delay;
> + }
> +
> + return max(min_delay, max_delay);
> +}
> +
> +/**
> + * pcie_wait_downstream_accessible - Wait downstream device to be accessible
> + * @pdev: PCIe port whose downstream device is waited
> + *
> + * Handle delays according to PCIe 4.0 section 6.6.1 before configuration
> + * access to the downstream device is permitted. If the port does not have
> + * any devices connected, does nothing.
> + *
> + * This is needed if the hierarchy below @pdev went through reset (after
> + * exit from D3cold back to D0uninitialized).
D3cold -> D0 is a cold reset by definition, isn't it?
> + */
> +void pcie_wait_downstream_accessible(struct pci_dev *pdev)
> +{
> + struct pci_dev *child;
> + struct pci_bus *bus;
> + int delay;
IIUC pdev is not guaranteed to be a PCIe device here.
Do we need to observe the Trhfa requirements for Conventional PCI and
PCI-X devices here? If we don't do it here, where is it observed?
> + if (pci_pcie_type(pdev) != PCI_EXP_TYPE_ROOT_PORT &&
> + pci_pcie_type(pdev) != PCI_EXP_TYPE_DOWNSTREAM)
> + return;
> +
> + if (pci_dev_is_disconnected(pdev))
> + return;
> +
> + if (!pdev->bridge_d3)
> + return;
> +
> + bus = pdev->subordinate;
> + if (!bus)
> + return;
> +
> + child = list_first_entry_or_null(&bus->devices, struct pci_dev,
> + bus_list);
> + if (!child)
> + return;
I'm not convinced yet about skipping this if there's no subordinate
bus or no children. Don't we have to assume that the caller may
immediately *probe* for children as soon as we return?
> + delay = pcie_get_downstream_delay(bus);
> + if (!delay)
> + return;
I'm not sold on the idea that this delay depends on what's *below* the
bridge. We're using sec 6.6.1 to justify the delay, and that section
doesn't say anything about downstream devices.
If we call .resume_noirq/.runtime_resume for the downstream devices
themselves, could we use d3cold_delay *there*?
> + /*
> + * If downstream port does not support speeds greater than 5 GT/s
> + * need to wait minimum 100ms. For higher speeds (gen3) we need to
> + * wait first for the data link layer to become active.
> + *
> + * However, 100ms is the minimum and the spec says that the
> + * software must allow at least 1s before it can determine that the
> + * device that did not respond is a broken device. Also there is
> + * evidence that the 100ms is not always enough so what we do here
This sort of "there is evidence" statement needs a specific reference.
Otherwise it ends up meaning "we don't know why it matters and we
can't test any changes in this area".
> + * is that we wait for the minimum 100ms (possibly after link
> + * training completes) and then probe for the device presence once.
> + * If we still don't get response we wait for another 100ms just to
> + * give it some additional time to complete its initialization.
> + */
> + if (pcie_get_speed_cap(pdev) <= PCIE_SPEED_5_0GT) {
> + dev_dbg(&pdev->dev, "waiting downstream link for %d ms\n",
pci_dbg()
s/waiting downstream link for %d ms/waiting %d ms for downstream link/
> + delay);
> + msleep(delay);
> + } else {
> + dev_dbg(&pdev->dev,
> + "waiting downstream link for %d ms after activation\n",
> + delay);
> + if (!pcie_wait_for_link_delay(pdev, true, delay)) {
> + /*
> + * If the link did not train, no need to wait
> + * further the device is probably gone.
> + */
> + return;
> + }
> + }
> +
> + if (!pci_device_is_present(child)) {
> + dev_dbg(&child->dev,
> + "waiting for additional 100 ms for the device to become accessible\n");
> + msleep(100);
This seems like magic. Can we relate this back to anything in the
spec? If we need this, maybe we're seeing CRRS? I suppose this is
related to the note above about evidence that 100ms may not be enough?
> + }
> +}
> +
> void pci_reset_secondary_bus(struct pci_dev *dev)
> {
> u16 ctrl;
> diff --git a/drivers/pci/pci.h b/drivers/pci/pci.h
> index d22d1b807701..9a83fcf612ca 100644
> --- a/drivers/pci/pci.h
> +++ b/drivers/pci/pci.h
> @@ -498,6 +498,7 @@ void pcie_do_recovery(struct pci_dev *dev, enum pci_channel_state state,
> u32 service);
>
> bool pcie_wait_for_link(struct pci_dev *pdev, bool active);
> +void pcie_wait_downstream_accessible(struct pci_dev *pdev);
> #ifdef CONFIG_PCIEASPM
> void pcie_aspm_init_link_state(struct pci_dev *pdev);
> void pcie_aspm_exit_link_state(struct pci_dev *pdev);
> --
> 2.23.0.rc1
>
Powered by blists - more mailing lists