[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20190207222245.GR7268@google.com>
Date: Thu, 7 Feb 2019 16:22:46 -0600
From: Bjorn Helgaas <helgaas@...nel.org>
To: Ming Lei <ming.lei@...hat.com>
Cc: Christoph Hellwig <hch@....de>,
Thomas Gleixner <tglx@...utronix.de>,
Jens Axboe <axboe@...nel.dk>, linux-block@...r.kernel.org,
Sagi Grimberg <sagi@...mberg.me>,
linux-nvme@...ts.infradead.org, linux-kernel@...r.kernel.org,
linux-pci@...r.kernel.org
Subject: Re: [PATCH 5/5] genirq/affinity: remove support for allocating
interrupt sets
On Fri, Jan 25, 2019 at 05:53:47PM +0800, Ming Lei wrote:
> Now allocating interrupt sets can be done via .setup_affinity()
> easily, so remove the support for allocating interrupt sets.
>
> With this change, we don't need the limit of 'minvec == maxvec'
> any more in pci_alloc_irq_vectors_affinity().
>
> Meantime irq_create_affinity_masks() gets simplified a lot.
>
> Signed-off-by: Ming Lei <ming.lei@...hat.com>
Acked-by: Bjorn Helgaas <bhelgaas@...gle.com> # pci/msi.c parts
> ---
> drivers/pci/msi.c | 14 -------------
> include/linux/interrupt.h | 4 ----
> kernel/irq/affinity.c | 52 +++++++++++------------------------------------
> 3 files changed, 12 insertions(+), 58 deletions(-)
>
> diff --git a/drivers/pci/msi.c b/drivers/pci/msi.c
> index 4c0b47867258..331483de1294 100644
> --- a/drivers/pci/msi.c
> +++ b/drivers/pci/msi.c
> @@ -1035,13 +1035,6 @@ static int __pci_enable_msi_range(struct pci_dev *dev, int minvec, int maxvec,
> if (maxvec < minvec)
> return -ERANGE;
>
> - /*
> - * If the caller is passing in sets, we can't support a range of
> - * vectors. The caller needs to handle that.
> - */
> - if (affd && affd->nr_sets && minvec != maxvec)
> - return -EINVAL;
> -
> if (WARN_ON_ONCE(dev->msi_enabled))
> return -EINVAL;
>
> @@ -1093,13 +1086,6 @@ static int __pci_enable_msix_range(struct pci_dev *dev,
> if (maxvec < minvec)
> return -ERANGE;
>
> - /*
> - * If the caller is passing in sets, we can't support a range of
> - * supported vectors. The caller needs to handle that.
> - */
> - if (affd && affd->nr_sets && minvec != maxvec)
> - return -EINVAL;
> -
> if (WARN_ON_ONCE(dev->msix_enabled))
> return -EINVAL;
>
> diff --git a/include/linux/interrupt.h b/include/linux/interrupt.h
> index b820b07f3b55..a035e165f405 100644
> --- a/include/linux/interrupt.h
> +++ b/include/linux/interrupt.h
> @@ -260,8 +260,6 @@ struct irq_affinity_desc {
> * and driver has to handle pre_vectors & post_vectors
> * correctly, set 'is_managed' flag correct too
> * @priv: Private data of @setup_affinity
> - * @nr_sets: Length of passed in *sets array
> - * @sets: Number of affinitized sets
> */
> struct irq_affinity {
> int pre_vectors;
> @@ -270,8 +268,6 @@ struct irq_affinity {
> struct irq_affinity_desc *,
> unsigned int);
> void *priv;
> - int nr_sets;
> - int *sets;
> };
>
> #if defined(CONFIG_SMP)
> diff --git a/kernel/irq/affinity.c b/kernel/irq/affinity.c
> index 524fdcda9f85..e8fea65325d9 100644
> --- a/kernel/irq/affinity.c
> +++ b/kernel/irq/affinity.c
> @@ -269,9 +269,9 @@ struct irq_affinity_desc *
> irq_create_affinity_masks(int nvecs, const struct irq_affinity *affd)
> {
> int affvecs = nvecs - affd->pre_vectors - affd->post_vectors;
> - int curvec, usedvecs;
> + int curvec;
> struct irq_affinity_desc *masks = NULL;
> - int i, nr_sets;
> + int i;
>
> /*
> * If there aren't any vectors left after applying the pre/post
> @@ -293,34 +293,14 @@ irq_create_affinity_masks(int nvecs, const struct irq_affinity *affd)
> /* Fill out vectors at the beginning that don't need affinity */
> for (curvec = 0; curvec < affd->pre_vectors; curvec++)
> cpumask_copy(&masks[curvec].mask, irq_default_affinity);
> - /*
> - * Spread on present CPUs starting from affd->pre_vectors. If we
> - * have multiple sets, build each sets affinity mask separately.
> - */
> - nr_sets = affd->nr_sets;
> - if (!nr_sets)
> - nr_sets = 1;
> -
> - for (i = 0, usedvecs = 0; i < nr_sets; i++) {
> - int this_vecs = affd->sets ? affd->sets[i] : affvecs;
> - int ret;
> -
> - ret = irq_build_affinity_masks(affd, curvec, this_vecs,
> - curvec, masks);
> - if (ret) {
> - kfree(masks);
> - return NULL;
> - }
> - curvec += this_vecs;
> - usedvecs += this_vecs;
> +
> + if (irq_build_affinity_masks(affd, curvec, affvecs, curvec, masks)) {
> + kfree(masks);
> + return NULL;
> }
>
> /* Fill out vectors at the end that don't need affinity */
> - if (usedvecs >= affvecs)
> - curvec = affd->pre_vectors + affvecs;
> - else
> - curvec = affd->pre_vectors + usedvecs;
> - for (; curvec < nvecs; curvec++)
> + for (curvec = affd->pre_vectors + affvecs; curvec < nvecs; curvec++)
> cpumask_copy(&masks[curvec].mask, irq_default_affinity);
>
> /* Mark the managed interrupts */
> @@ -340,21 +320,13 @@ int irq_calc_affinity_vectors(int minvec, int maxvec, const struct irq_affinity
> {
> int resv = affd->pre_vectors + affd->post_vectors;
> int vecs = maxvec - resv;
> - int set_vecs;
> + int ret;
>
> if (resv > minvec)
> return 0;
>
> - if (affd->nr_sets) {
> - int i;
> -
> - for (i = 0, set_vecs = 0; i < affd->nr_sets; i++)
> - set_vecs += affd->sets[i];
> - } else {
> - get_online_cpus();
> - set_vecs = cpumask_weight(cpu_possible_mask);
> - put_online_cpus();
> - }
> -
> - return resv + min(set_vecs, vecs);
> + get_online_cpus();
> + ret = min_t(int, cpumask_weight(cpu_possible_mask), vecs) + resv;
> + put_online_cpus();
> + return ret;
> }
> --
> 2.9.5
>
Powered by blists - more mailing lists