[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <DM5PR2101MB10300B0494E43D69A20C86B3DCDA0@DM5PR2101MB1030.namprd21.prod.outlook.com>
Date: Mon, 5 Mar 2018 23:47:32 +0000
From: "Michael Kelley (EOSG)" <Michael.H.Kelley@...rosoft.com>
To: Dexuan Cui <decui@...rosoft.com>,
"bhelgaas@...gle.com" <bhelgaas@...gle.com>,
"linux-pci@...r.kernel.org" <linux-pci@...r.kernel.org>,
KY Srinivasan <kys@...rosoft.com>,
Stephen Hemminger <sthemmin@...rosoft.com>,
"olaf@...fle.de" <olaf@...fle.de>,
"apw@...onical.com" <apw@...onical.com>,
"jasowang@...hat.com" <jasowang@...hat.com>
CC: "linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
"driverdev-devel@...uxdriverproject.org"
<driverdev-devel@...uxdriverproject.org>,
Haiyang Zhang <haiyangz@...rosoft.com>,
"vkuznets@...hat.com" <vkuznets@...hat.com>,
"marcelo.cerri@...onical.com" <marcelo.cerri@...onical.com>,
Jack Morgenstein <jackm@...lanox.com>,
"stable@...r.kernel.org" <stable@...r.kernel.org>
Subject: RE: [PATCH v2 5/6] PCI: hv: hv_pci_devices_present(): only queue a
new work when necessary
> -----Original Message-----
> From: Dexuan Cui
> Sent: Monday, March 5, 2018 11:22 AM
> To: bhelgaas@...gle.com; linux-pci@...r.kernel.org; KY Srinivasan <kys@...rosoft.com>;
> Stephen Hemminger <sthemmin@...rosoft.com>; olaf@...fle.de; apw@...onical.com;
> jasowang@...hat.com
> Cc: linux-kernel@...r.kernel.org; driverdev-devel@...uxdriverproject.org; Haiyang Zhang
> <haiyangz@...rosoft.com>; vkuznets@...hat.com; marcelo.cerri@...onical.com; Michael
> Kelley (EOSG) <Michael.H.Kelley@...rosoft.com>; Dexuan Cui <decui@...rosoft.com>; Jack
> Morgenstein <jackm@...lanox.com>; stable@...r.kernel.org
> Subject: [PATCH v2 5/6] PCI: hv: hv_pci_devices_present(): only queue a new work when
> necessary
>
> If there is a pending work, we just need to add the new dr into
> the dr_list.
>
> This is suggested by Michael Kelley.
>
> Signed-off-by: Dexuan Cui <decui@...rosoft.com>
> Cc: Vitaly Kuznetsov <vkuznets@...hat.com>
> Cc: Jack Morgenstein <jackm@...lanox.com>
> Cc: stable@...r.kernel.org
> Cc: Stephen Hemminger <sthemmin@...rosoft.com>
> Cc: K. Y. Srinivasan <kys@...rosoft.com>
> Cc: Michael Kelley (EOSG) <Michael.H.Kelley@...rosoft.com>
> ---
> drivers/pci/host/pci-hyperv.c | 19 ++++++++++++++++---
> 1 file changed, 16 insertions(+), 3 deletions(-)
>
> diff --git a/drivers/pci/host/pci-hyperv.c b/drivers/pci/host/pci-hyperv.c
> index 3a385212f666..d3aa6736a9bb 100644
> --- a/drivers/pci/host/pci-hyperv.c
> +++ b/drivers/pci/host/pci-hyperv.c
> @@ -1733,6 +1733,7 @@ static void hv_pci_devices_present(struct hv_pcibus_device *hbus,
> struct hv_dr_state *dr;
> struct hv_dr_work *dr_wrk;
> unsigned long flags;
> + bool pending_dr;
>
> dr_wrk = kzalloc(sizeof(*dr_wrk), GFP_NOWAIT);
> if (!dr_wrk)
> @@ -1756,11 +1757,23 @@ static void hv_pci_devices_present(struct hv_pcibus_device
> *hbus,
> }
>
> spin_lock_irqsave(&hbus->device_list_lock, flags);
> +
> + /*
> + * If pending_dr is true, we have already queued a work,
> + * which will see the new dr. Otherwise, we need to
> + * queue a new work.
> + */
> + pending_dr = !list_empty(&hbus->dr_list);
> list_add_tail(&dr->list_entry, &hbus->dr_list);
> - spin_unlock_irqrestore(&hbus->device_list_lock, flags);
A minor point: The spin_unlock_irqrestore() call can
stay here. Once we have the list status in a local variable
and the new entry is added to the list, nothing bad can
happen if we drop the spin lock. At worst, and very unlikely,
we'll queue work when some other thread has already queued
work to process the list entry, but that's no big deal. I'd argue
for keeping the code covered by a spin lock as small as possible.
Michael
>
> - get_hvpcibus(hbus);
> - queue_work(hbus->wq, &dr_wrk->wrk);
> + if (pending_dr) {
> + kfree(dr_wrk);
> + } else {
> + get_hvpcibus(hbus);
> + queue_work(hbus->wq, &dr_wrk->wrk);
> + }
> +
> + spin_unlock_irqrestore(&hbus->device_list_lock, flags);
> }
>
> /**
> --
> 2.7.4
Powered by blists - more mailing lists