lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Wed, 6 Dec 2017 10:23:11 -0800
From:   "Raj, Ashok" <ashok.raj@...el.com>
To:     Jerry Snitselaar <jsnitsel@...hat.com>
Cc:     iommu@...ts.linux-foundation.org, linux-kernel@...r.kernel.org,
        Alex Williamson <alex.williamson@...hat.com>,
        Joerg Roedel <joro@...tes.org>, Ashok Raj <ashok.raj@...el.com>
Subject: Re: [PATCH] iommu/vt-d: clean up pr_irq if request_threaded_irq fails

On Wed, Dec 06, 2017 at 09:49:59AM -0700, Jerry Snitselaar wrote:
> It is unlikely request_threaded_irq will fail, but if it does for some
> reason we should clear iommu->pr_irq in the error path. Also
> intel_svm_finish_prq shouldn't try to clean up the page request
> interrupt if pr_irq is 0. Without these, if request_threaded_irq were
> to fail the following occurs:

Looks good.

Reviewed-by: Ashok Raj <ashok.raj@...el.com>

Cheers,
Ashok

> 
> Cc: Alex Williamson <alex.williamson@...hat.com>
> Cc: Joerg Roedel <joro@...tes.org>
> Cc: Ashok Raj <ashok.raj@...el.com>
> Signed-off-by: Jerry Snitselaar <jsnitsel@...hat.com>
> ---
>  drivers/iommu/intel-svm.c | 9 ++++++---
>  1 file changed, 6 insertions(+), 3 deletions(-)
> 
> diff --git a/drivers/iommu/intel-svm.c b/drivers/iommu/intel-svm.c
> index ed1cf7c5a43b..6643277e321e 100644
> --- a/drivers/iommu/intel-svm.c
> +++ b/drivers/iommu/intel-svm.c
> @@ -129,6 +129,7 @@ int intel_svm_enable_prq(struct intel_iommu *iommu)
>  		pr_err("IOMMU: %s: Failed to request IRQ for page request queue\n",
>  		       iommu->name);
>  		dmar_free_hwirq(irq);
> +		iommu->pr_irq = 0;
>  		goto err;
>  	}
>  	dmar_writeq(iommu->reg + DMAR_PQH_REG, 0ULL);
> @@ -144,9 +145,11 @@ int intel_svm_finish_prq(struct intel_iommu *iommu)
>  	dmar_writeq(iommu->reg + DMAR_PQT_REG, 0ULL);
>  	dmar_writeq(iommu->reg + DMAR_PQA_REG, 0ULL);
>  
> -	free_irq(iommu->pr_irq, iommu);
> -	dmar_free_hwirq(iommu->pr_irq);
> -	iommu->pr_irq = 0;
> +	if (iommu->pr_irq) {
> +		free_irq(iommu->pr_irq, iommu);
> +		dmar_free_hwirq(iommu->pr_irq);
> +		iommu->pr_irq = 0;
> +	}
>  
>  	free_pages((unsigned long)iommu->prq, PRQ_ORDER);
>  	iommu->prq = NULL;
> -- 
> 2.14.3
> 

Powered by blists - more mailing lists