lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:	Wed, 17 Oct 2007 17:42:30 -0700
From:	"Yinghai Lu" <yhlu.kernel@...il.com>
To:	"Manfred Spraul" <manfred@...orfullife.com>
Cc:	jgarzik@...ox.com, AAbdulla@...dia.com, akpm@...ux-foundation.org,
	linux-kernel@...r.kernel.org
Subject: Re: [PATCH] forcedeth msi bugfix

On 10/17/07, Manfred Spraul <manfred@...orfullife.com> wrote:
> pci_enable_msi() replaces the INTx irq number in pci_dev->irq with the
> new MSI irq number.
> The forcedeth driver did not update the copy in netdevice->irq and
> parts of the driver used the stale copy.
> See bugzilla.kernel.org, bug 9047.
>
> The patch
> - updates netdevice->irq
> - replaces all accesses to netdevice->irq with pci_dev->irq.
>
> The patch is against 2.6.23.1. IMHO suitable for both 2.6.23 and 2.6.24
>
> Signed-Off-By: Manfred Spraul <manfred@...orfullife.com>
>
> --- 2.6/drivers/net/forcedeth.c 2007-10-17 21:36:36.000000000 +0200
> +++ build-2.6/drivers/net/forcedeth.c   2007-10-17 21:54:58.000000000 +0200
> @@ -128,7 +128,7 @@
>  #else
>  #define DRIVERNAPI
>  #endif
> -#define FORCEDETH_VERSION              "0.60"
> +#define FORCEDETH_VERSION              "0.61"
>  #define DRV_NAME                       "forcedeth"
>
>  #include <linux/module.h>
> @@ -988,7 +988,7 @@
>                 if (np->msi_flags & NV_MSI_X_ENABLED)
>                         enable_irq(np->msi_x_entry[NV_MSI_X_VECTOR_ALL].vector);
>                 else
> -                       enable_irq(dev->irq);
> +                       enable_irq(np->pci_dev->irq);
>         } else {
>                 enable_irq(np->msi_x_entry[NV_MSI_X_VECTOR_RX].vector);
>                 enable_irq(np->msi_x_entry[NV_MSI_X_VECTOR_TX].vector);
> @@ -1004,7 +1004,7 @@
>                 if (np->msi_flags & NV_MSI_X_ENABLED)
>                         disable_irq(np->msi_x_entry[NV_MSI_X_VECTOR_ALL].vector);
>                 else
> -                       disable_irq(dev->irq);
> +                       disable_irq(np->pci_dev->irq);
>         } else {
>                 disable_irq(np->msi_x_entry[NV_MSI_X_VECTOR_RX].vector);
>                 disable_irq(np->msi_x_entry[NV_MSI_X_VECTOR_TX].vector);
> @@ -1601,7 +1601,7 @@
>                 if (np->msi_flags & NV_MSI_X_ENABLED)
>                         disable_irq(np->msi_x_entry[NV_MSI_X_VECTOR_ALL].vector);
>                 else
> -                       disable_irq(dev->irq);
> +                       disable_irq(np->pci_dev->irq);
>         } else {
>                 disable_irq(np->msi_x_entry[NV_MSI_X_VECTOR_RX].vector);
>         }
> @@ -1619,7 +1619,7 @@
>                 if (np->msi_flags & NV_MSI_X_ENABLED)
>                         enable_irq(np->msi_x_entry[NV_MSI_X_VECTOR_ALL].vector);
>                 else
> -                       enable_irq(dev->irq);
> +                       enable_irq(np->pci_dev->irq);
>         } else {
>                 enable_irq(np->msi_x_entry[NV_MSI_X_VECTOR_RX].vector);
>         }
> @@ -3557,10 +3557,12 @@
>         if (ret != 0 && np->msi_flags & NV_MSI_CAPABLE) {
>                 if ((ret = pci_enable_msi(np->pci_dev)) == 0) {
>                         np->msi_flags |= NV_MSI_ENABLED;

> +                       dev->irq = np->pci_dev->irq;

move to after request_irq?

>                         if (request_irq(np->pci_dev->irq, handler, IRQF_SHARED, dev->name, dev) != 0) {
>                                 printk(KERN_INFO "forcedeth: request_irq failed %d\n", ret);
>                                 pci_disable_msi(np->pci_dev);
>                                 np->msi_flags &= ~NV_MSI_ENABLED;
> +                               dev->irq = np->pci_dev->irq;
>                                 goto out_err;
>                         }

put
 +                       dev->irq = np->pci_dev->irq;
here instead

>
> @@ -3623,7 +3625,7 @@
>                 if (np->msi_flags & NV_MSI_X_ENABLED)
>                         disable_irq_lockdep(np->msi_x_entry[NV_MSI_X_VECTOR_ALL].vector);
>                 else
> -                       disable_irq_lockdep(dev->irq);
> +                       disable_irq_lockdep(np->pci_dev->irq);
>                 mask = np->irqmask;
>         } else {
>                 if (np->nic_poll_irq & NVREG_IRQ_RX_ALL) {
> @@ -3641,6 +3643,8 @@
>         }
>         np->nic_poll_irq = 0;
>
> +       /* disable_irq() contains synchronize_irq, thus no irq handler can run now */
> +
>         if (np->recover_error) {
>                 np->recover_error = 0;
>                 printk(KERN_INFO "forcedeth: MAC in recoverable error state\n");
> @@ -3677,7 +3681,6 @@
>                 }
>         }
>
> -       /* FIXME: Do we need synchronize_irq(dev->irq) here? */
>
>         writel(mask, base + NvRegIrqMask);
>         pci_push(base);
> @@ -3690,7 +3693,7 @@
>                 if (np->msi_flags & NV_MSI_X_ENABLED)
>                        enable_irq_lockdep(np->msi_x_entry[NV_MSI_X_VECTOR_ALL].vector);
>                 else
> -                       enable_irq_lockdep(dev->irq);
> +                       enable_irq_lockdep(np->pci_dev->irq);
>         } else {
>                 if (np->nic_poll_irq & NVREG_IRQ_RX_ALL) {
>                         nv_nic_irq_rx(0, dev);
> @@ -4943,7 +4946,7 @@
>         np->in_shutdown = 1;
>         spin_unlock_irq(&np->lock);
>         netif_poll_disable(dev);
> -       synchronize_irq(dev->irq);
> +       synchronize_irq(np->pci_dev->irq);
that is in nv_close, how about if is MSI-X.

YH
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ