lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <2341d40b-326f-0fbd-bcf0-9069e9f27ff3@linux.intel.com>
Date:   Thu, 15 Oct 2020 13:12:28 -0700
From:   "Kuppuswamy, Sathyanarayanan" 
        <sathyanarayanan.kuppuswamy@...ux.intel.com>
To:     Ethan Zhao <xerces.zhao@...il.com>,
        Kuppuswamy Sathyanarayanan 
        <sathyanarayanan.nkuppuswamy@...il.com>
Cc:     Bjorn Helgaas <bhelgaas@...gle.com>, Sinan Kaya <okaya@...nel.org>,
        linux-pci <linux-pci@...r.kernel.org>,
        Linux Kernel Mailing List <linux-kernel@...r.kernel.org>,
        Ashok Raj <ashok.raj@...el.com>
Subject: Re: [PATCH v6 2/2] PCI/ERR: Split the fatal and non-fatal error
 recovery handling



On 10/15/20 6:55 AM, Ethan Zhao wrote:
> On Wed, Oct 14, 2020 at 5:00 PM Kuppuswamy Sathyanarayanan
> <sathyanarayanan.nkuppuswamy@...il.com> wrote:
>>
>> Commit bdb5ac85777d ("PCI/ERR: Handle fatal error recovery")
>> merged fatal and non-fatal error recovery paths, and also made
>> recovery code depend on hotplug handler for "remove affected
>> device + rescan" support. But this change also complicated the
>> error recovery path and which in turn led to the following
>> issues.
>>
>> 1. We depend on hotplug handler for removing the affected
>> devices/drivers on DLLSC LINK down event (on DPC event
>> trigger) and DPC handler for handling the error recovery. Since
>> both handlers operate on same set of affected devices, it leads
>> to race condition, which in turn leads to  NULL pointer
>> exceptions or error recovery failures.You can find more details
>> about this issue in following link.
>>
>> https://lore.kernel.org/linux-pci/20201007113158.48933-1-haifeng.zhao@intel.com/T/#t
>>
>> 2. For non-hotplug capable devices fatal (DPC) error recovery
>> is currently broken. Current fatal error recovery implementation
>> relies on PCIe hotplug (pciehp) handler for detaching and
>> re-enumerating the affected devices/drivers. So when dealing with
>> non-hotplug capable devices, recovery code does not restore the state
>> of the affected devices correctly. You can find more details about
>> this issue in the following links.
>>
>> https://lore.kernel.org/linux-pci/20200527083130.4137-1-Zhiqiang.Hou@nxp.com/
>> https://lore.kernel.org/linux-pci/12115.1588207324@famine/
>> https://lore.kernel.org/linux-pci/0e6f89cd6b9e4a72293cc90fafe93487d7c2d295.1585000084.git.sathyanarayanan.kuppuswamy@linux.intel.com/
>>
>> In order to fix the above two issues, we should stop relying on hotplug
>> handler for cleaning the affected devices/drivers and let error recovery
>> handler own this functionality. So this patch reverts Commit bdb5ac85777d
>> ("PCI/ERR: Handle fatal error recovery") and re-introduce the  "remove
>> affected device + rescan"  functionality in fatal error recovery handler.
>>
>> Also holding pci_lock_rescan_remove() will prevent the race between hotplug
>> and DPC handler.
>>
>> Fixes: bdb5ac85777d ("PCI/ERR: Handle fatal error recovery")
>> Signed-off-by: Kuppuswamy Sathyanarayanan <sathyanarayanan.kuppuswamy@...ux.intel.com>
>> Reviewed-by: Sinan Kaya <okaya@...nel.org>
>> ---
>>   Changes since v4:
>>    * Added new interfaces for error recovery (pcie_do_fatal_recovery()
>>      and pcie_do_nonfatal_recovery()).
>>
>>   Changes since v5:
>>    * Fixed static/non-static declartion issue.
>>
>>   Documentation/PCI/pci-error-recovery.rst | 47 ++++++++++------
>>   Documentation/PCI/pcieaer-howto.rst      |  2 +-
>>   drivers/pci/pci.h                        |  5 +-
>>   drivers/pci/pcie/aer.c                   | 10 ++--
>>   drivers/pci/pcie/dpc.c                   |  2 +-
>>   drivers/pci/pcie/edr.c                   |  2 +-
>>   drivers/pci/pcie/err.c                   | 70 ++++++++++++++++++------
>>   7 files changed, 94 insertions(+), 44 deletions(-)
>>
>> diff --git a/Documentation/PCI/pci-error-recovery.rst b/Documentation/PCI/pci-error-recovery.rst
>> index 84ceebb08cac..830c8af5838b 100644
>> --- a/Documentation/PCI/pci-error-recovery.rst
>> +++ b/Documentation/PCI/pci-error-recovery.rst
>> @@ -115,7 +115,7 @@ The actual steps taken by a platform to recover from a PCI error
>>   event will be platform-dependent, but will follow the general
>>   sequence described below.
>>
>> -STEP 0: Error Event
>> +STEP 0: Error Event: ERR_NONFATAL
>>   -------------------
>>   A PCI bus error is detected by the PCI hardware.  On powerpc, the slot
>>   is isolated, in that all I/O is blocked: all reads return 0xffffffff,
>> @@ -160,10 +160,10 @@ particular, if the platform doesn't isolate slots), and recovery
>>   proceeds to STEP 2 (MMIO Enable).
>>
>>   If any driver requested a slot reset (by returning PCI_ERS_RESULT_NEED_RESET),
>> -then recovery proceeds to STEP 4 (Slot Reset).
>> +then recovery proceeds to STEP 3 (Slot Reset).
>>
>>   If the platform is unable to recover the slot, the next step
>> -is STEP 6 (Permanent Failure).
>> +is STEP 5 (Permanent Failure).
>>
>>   .. note::
>>
>> @@ -198,7 +198,7 @@ reset or some such, but not restart operations. This callback is made if
>>   all drivers on a segment agree that they can try to recover and if no automatic
>>   link reset was performed by the HW. If the platform can't just re-enable IOs
>>   without a slot reset or a link reset, it will not call this callback, and
>> -instead will have gone directly to STEP 3 (Link Reset) or STEP 4 (Slot Reset)
>> +instead will have gone directly to STEP 3 (Slot Reset)
>>
>>   .. note::
>>
>> @@ -233,18 +233,12 @@ The driver should return one of the following result codes:
>>
>>   The next step taken depends on the results returned by the drivers.
>>   If all drivers returned PCI_ERS_RESULT_RECOVERED, then the platform
>> -proceeds to either STEP3 (Link Reset) or to STEP 5 (Resume Operations).
>> +proceeds to STEP 4 (Resume Operations).
>>
>>   If any driver returned PCI_ERS_RESULT_NEED_RESET, then the platform
>> -proceeds to STEP 4 (Slot Reset)
>> +proceeds to STEP 3 (Slot Reset)
>>
>> -STEP 3: Link Reset
>> -------------------
>> -The platform resets the link.  This is a PCI-Express specific step
>> -and is done whenever a fatal error has been detected that can be
>> -"solved" by resetting the link.
>> -
>> -STEP 4: Slot Reset
>> +STEP 3: Slot Reset
>>   ------------------
>>
>>   In response to a return value of PCI_ERS_RESULT_NEED_RESET, the
>> @@ -322,7 +316,7 @@ PCI card types::
>>          +               pdev->needs_freset = 1;
>>          +
>>
>> -Platform proceeds either to STEP 5 (Resume Operations) or STEP 6 (Permanent
>> +Platform proceeds either to STEP 4 (Resume Operations) or STEP 5 (Permanent
>>   Failure).
>>
>>   .. note::
>> @@ -332,7 +326,7 @@ Failure).
>>      However, it probably should.
>>
>>
>> -STEP 5: Resume Operations
>> +STEP 4: Resume Operations
>>   -------------------------
>>   The platform will call the resume() callback on all affected device
>>   drivers if all drivers on the segment have returned
>> @@ -344,7 +338,7 @@ a result code.
>>   At this point, if a new error happens, the platform will restart
>>   a new error recovery sequence.
>>
>> -STEP 6: Permanent Failure
>> +STEP 5: Permanent Failure
>>   -------------------------
>>   A "permanent failure" has occurred, and the platform cannot recover
>>   the device.  The platform will call error_detected() with a
>> @@ -367,6 +361,27 @@ errors. See the discussion in powerpc/eeh-pci-error-recovery.txt
>>   for additional detail on real-life experience of the causes of
>>   software errors.
>>
>> +STEP 0: Error Event: ERR_FATAL
>> +--------------------
>> +PCI bus error is detected by the PCI hardware. On powerpc, the slot is
>> +isolated, in that all I/O is blocked: all reads return 0xffffffff, all
>> +writes are ignored.
>> +
>> +STEP 1: Remove devices
>> +---------------------
>> +Platform removes the devices depending on the error agent, it could be
>> +this port for all subordinates or upstream component (likely downstream
>> +port)
>> +
>> +STEP 2: Reset link
>> +---------------------
>> +The platform resets the link.  This is a PCI-Express specific step and is
>> +done whenever a fatal error has been detected that can be "solved" by
>> +resetting the link.
>> +
>> +STEP 3: Re-enumerate the devices
>> +---------------------
>> +Initiates the re-enumeration.
>>
>>   Conclusion; General Remarks
>>   ---------------------------
>> diff --git a/Documentation/PCI/pcieaer-howto.rst b/Documentation/PCI/pcieaer-howto.rst
>> index 0b36b9ebfa4b..9528cfd9449b 100644
>> --- a/Documentation/PCI/pcieaer-howto.rst
>> +++ b/Documentation/PCI/pcieaer-howto.rst
>> @@ -206,7 +206,7 @@ error_detected(dev, pci_channel_io_frozen) to all drivers within
>>   a hierarchy in question. Then, performing link reset at upstream is
>>   necessary. As different kinds of devices might use different approaches
>>   to reset link, AER port service driver is required to provide the
>> -function to reset link via callback parameter of pcie_do_recovery()
>> +function to reset link via callback parameter of pcie_do_fatal_recovery()
>>   function. If reset_link is not NULL, recovery function will use it
>>   to reset the link. If error_detected returns PCI_ERS_RESULT_CAN_RECOVER
>>   and reset_link returns PCI_ERS_RESULT_RECOVERED, the error handling goes
>> diff --git a/drivers/pci/pci.h b/drivers/pci/pci.h
>> index fa12f7cbc1a0..b3e1571107b5 100644
>> --- a/drivers/pci/pci.h
>> +++ b/drivers/pci/pci.h
>> @@ -555,8 +555,9 @@ static inline int pci_dev_specific_disable_acs_redir(struct pci_dev *dev)
>>   #endif
>>
>>   /* PCI error reporting and recovery */
>> -pci_ers_result_t pcie_do_recovery(struct pci_dev *dev,
>> -                       pci_channel_state_t state,
>> +pci_ers_result_t pcie_do_nonfatal_recovery(struct pci_dev *dev);
>> +
>> +pci_ers_result_t pcie_do_fatal_recovery(struct pci_dev *dev,
>>                          pci_ers_result_t (*reset_link)(struct pci_dev *pdev));
>>
>>   bool pcie_wait_for_link(struct pci_dev *pdev, bool active);
>> diff --git a/drivers/pci/pcie/aer.c b/drivers/pci/pcie/aer.c
>> index 65dff5f3457a..f3e70bb9b30d 100644
>> --- a/drivers/pci/pcie/aer.c
>> +++ b/drivers/pci/pcie/aer.c
>> @@ -947,9 +947,9 @@ static void handle_error_source(struct pci_dev *dev, struct aer_err_info *info)
>>                  if (pcie_aer_is_native(dev))
>>                          pcie_clear_device_status(dev);
>>          } else if (info->severity == AER_NONFATAL)
>> -               pcie_do_recovery(dev, pci_channel_io_normal, aer_root_reset);
>> +               pcie_do_nonfatal_recovery(dev);
>>          else if (info->severity == AER_FATAL)
>> -               pcie_do_recovery(dev, pci_channel_io_frozen, aer_root_reset);
>> +               pcie_do_fatal_recovery(dev, aer_root_reset);
>>          pci_dev_put(dev);
>>   }
>>
>> @@ -985,11 +985,9 @@ static void aer_recover_work_func(struct work_struct *work)
>>                  }
>>                  cper_print_aer(pdev, entry.severity, entry.regs);
>>                  if (entry.severity == AER_NONFATAL)
>> -                       pcie_do_recovery(pdev, pci_channel_io_normal,
>> -                                        aer_root_reset);
>> +                       pcie_do_nonfatal_recovery(pdev);
>>                  else if (entry.severity == AER_FATAL)
>> -                       pcie_do_recovery(pdev, pci_channel_io_frozen,
>> -                                        aer_root_reset);
>> +                       pcie_do_fatal_recovery(pdev, aer_root_reset);
>>                  pci_dev_put(pdev);
>>          }
>>   }
>> diff --git a/drivers/pci/pcie/dpc.c b/drivers/pci/pcie/dpc.c
>> index daa9a4153776..74e7d1da3cf0 100644
>> --- a/drivers/pci/pcie/dpc.c
>> +++ b/drivers/pci/pcie/dpc.c
>> @@ -233,7 +233,7 @@ static irqreturn_t dpc_handler(int irq, void *context)
>>          dpc_process_error(pdev);
>>
>>          /* We configure DPC so it only triggers on ERR_FATAL */
>> -       pcie_do_recovery(pdev, pci_channel_io_frozen, dpc_reset_link);
>> +       pcie_do_fatal_recovery(pdev, dpc_reset_link);
>>
>>          return IRQ_HANDLED;
>>   }
>> diff --git a/drivers/pci/pcie/edr.c b/drivers/pci/pcie/edr.c
>> index a6b9b479b97a..87379bc566f6 100644
>> --- a/drivers/pci/pcie/edr.c
>> +++ b/drivers/pci/pcie/edr.c
>> @@ -183,7 +183,7 @@ static void edr_handle_event(acpi_handle handle, u32 event, void *data)
>>           * or ERR_NONFATAL, since the link is already down, use the FATAL
>>           * error recovery path for both cases.
>>           */
>> -       estate = pcie_do_recovery(edev, pci_channel_io_frozen, dpc_reset_link);
>> +       estate = pcie_do_fatal_recovery(edev, dpc_reset_link);
>>
>>   send_ost:
>>
>> diff --git a/drivers/pci/pcie/err.c b/drivers/pci/pcie/err.c
>> index 315a4d559c4c..fa50366a9632 100644
>> --- a/drivers/pci/pcie/err.c
>> +++ b/drivers/pci/pcie/err.c
>> @@ -79,11 +79,6 @@ static int report_error_detected(struct pci_dev *dev,
>>          return 0;
>>   }
>>
>> -static int report_frozen_detected(struct pci_dev *dev, void *data)
>> -{
>> -       return report_error_detected(dev, pci_channel_io_frozen, data);
>> -}
>> -
>>   static int report_normal_detected(struct pci_dev *dev, void *data)
>>   {
>>          return report_error_detected(dev, pci_channel_io_normal, data);
>> @@ -146,9 +141,59 @@ static int report_resume(struct pci_dev *dev, void *data)
>>          return 0;
>>   }
>>
>> -pci_ers_result_t pcie_do_recovery(struct pci_dev *dev,
>> -                       pci_channel_state_t state,
>> +pci_ers_result_t pcie_do_fatal_recovery(struct pci_dev *dev,
>>                          pci_ers_result_t (*reset_link)(struct pci_dev *pdev))
>> +{
>> +       struct pci_dev *udev;
>> +       struct pci_bus *parent;
>> +       struct pci_dev *pdev, *temp;
>> +       pci_ers_result_t result;
>> +
>> +       if (dev->hdr_type == PCI_HEADER_TYPE_BRIDGE)
>> +               udev = dev;
>> +       else
>> +               udev = dev->bus->self;
>> +
>> +       parent = udev->subordinate;
>> +       pci_walk_bus(parent, pci_dev_set_disconnected, NULL);
>> +
>> +        pci_lock_rescan_remove();
>> +        pci_dev_get(dev);
>> +        list_for_each_entry_safe_reverse(pdev, temp, &parent->devices,
>> +                                        bus_list) {
>> +               pci_stop_and_remove_bus_device(pdev);
> 
> Again , pciehp driver wouldn't naturally ignore the DLLSC/PDC event,
> or because you do
> pci_stop_and_remove_bus_device(pdev) here, pciehp driver wouldn't do
> pcie_disable_slot()
> anymore,  when DPC driver get its interrupt and run into here, pciehp
> driver also gets DLLSC
> /PDC and may run into SUPPRISE_REMOVAL first.  so there would be duplicated
>   pci_stop_and_remove_bus_device(pdev);

If you check the implementation of pci_stop_and_remove_bus_device(), all it
does is to go over the list of sub-devices and bus device, and then do pci_stop_dev() and 
pci_destroy_dev() operations. My point is, once the device is stopped or removed then
in other thread which executes pci_stop_and_remove_bus_device(), the same device will
not be present. And hence we are not repeating the core operation. Ofcourse there are
still some duplication (like DLLSC/PDC event handling). But I am not trying to solve
that problem here.

I am just making error recovery handler own this functionality to solve,

1. recovery problem of non-hotplug capapble devices.
2. remove dependency on hotplug handler for handline remove and rescan functionality.
3. Make pcie_do_recovery() less complex.

> 
>   Don't say it would naturally be okay. check it.
> 
>> +       }
>> +
>> +       result = reset_link(udev);
>> +
>> +       if (dev->hdr_type == PCI_HEADER_TYPE_BRIDGE) {
>> +               /*
>> +                * If the error is reported by a bridge, we think this error
>> +                * is related to the downstream link of the bridge, so we
>> +                * do error recovery on all subordinates of the bridge instead
>> +                * of the bridge and clear the error status of the bridge.
>> +                */
>> +               pci_aer_clear_fatal_status(dev);
>> +               if (pcie_aer_is_native(dev))
>> +                       pcie_clear_device_status(dev);
>> +       }
>> +
>> +       if (result == PCI_ERS_RESULT_RECOVERED) {
>> +               if (pcie_wait_for_link(udev, true))
>> +                       pci_rescan_bus(udev->bus);
>          So does to this pci_rescan_bus(udev->bus);
>          You said you won't rely on pciehp handler, but how do you eliminate the
>          handler's duplicated actions.
> 
>   Thanks,
>   Ethan
>> +               pci_info(dev, "Device recovery from fatal error successful\n");
>> +        } else {
>> +               pci_uevent_ers(dev, PCI_ERS_RESULT_DISCONNECT);
>> +               pci_info(dev, "Device recovery from fatal error failed\n");
>> +        }
>> +
>> +       pci_dev_put(dev);
>> +       pci_unlock_rescan_remove();
>> +
>> +       return result;
>> +}
>> +
>> +pci_ers_result_t pcie_do_nonfatal_recovery(struct pci_dev *dev)
>>   {
>>          pci_ers_result_t status = PCI_ERS_RESULT_CAN_RECOVER;
>>          struct pci_bus *bus;
>> @@ -164,16 +209,7 @@ pci_ers_result_t pcie_do_recovery(struct pci_dev *dev,
>>          bus = dev->subordinate;
>>
>>          pci_dbg(dev, "broadcast error_detected message\n");
>> -       if (state == pci_channel_io_frozen) {
>> -               pci_walk_bus(bus, report_frozen_detected, &status);
>> -               status = reset_link(dev);
>> -               if (status != PCI_ERS_RESULT_RECOVERED) {
>> -                       pci_warn(dev, "link reset failed\n");
>> -                       goto failed;
>> -               }
>> -       } else {
>> -               pci_walk_bus(bus, report_normal_detected, &status);
>> -       }
>> +       pci_walk_bus(bus, report_normal_detected, &status);
>>
>>          if (status == PCI_ERS_RESULT_CAN_RECOVER) {
>>                  status = PCI_ERS_RESULT_RECOVERED;
>> --
>> 2.17.1
>>

-- 
Sathyanarayanan Kuppuswamy
Linux Kernel Developer

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ