lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Fri, 18 Jul 2008 04:37:08 -0700 (PDT) From: David Miller <davem@...emloft.net> To: jeffrey.t.kirsher@...el.com Cc: netdev@...r.kernel.org Subject: Re: [PATCH] igb/ixgbe/e1000e: resolve tx multiqueue bug From: Jeff Kirsher <jeffrey.t.kirsher@...el.com> Date: Fri, 18 Jul 2008 04:02:25 -0700 > With the recent changes to tx mutiqueue, igb/ixgbe/e1000e was not calling > netif_tx_start_all_queues() before calling netif_tx_wake_all_queues(). > This causes an issue during loading of the driver. > > In addition, updated e1000e to use the updated tx mutliqueue api. > > Signed-off-by: Jeff Kirsher <jeffrey.t.kirsher@...el.com> Thanks Jeff. I've applied this to kill the BUG() triggers you were seeing. Background for everyone else: 1) These drivers have a watchdog handler that watches the PHY state and wakes the TX queue when the link comes up. 2) Since ->open() was not doing a start queue, we would do a wake queue while attached to the noop_qdisc This causes the BUG check in netif_schedule() to trigger. After some discussion with Jeff I've advised him that it's not kosher for the driver to manage the TX queue wake state like this in respose to link up and down events, but rather simply maintaining the netif_carrier state is all that is needed. When netif_carrier_ok() is false, we're attached to the noop_qdisc which eats all packets and frees them up, never sending them down to the device. When netif_carrier_ok() is true, we attach to the configured normal qdisc and packets begin flowing to the device again. Jeff told me he will take a look more deeply into this. -- To unsubscribe from this list: send the line "unsubscribe netdev" in the body of a message to majordomo@...r.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists