diff --git a/drivers/net/tg3.c b/drivers/net/tg3.c index 5a864bd..9aafb78 100644 --- a/drivers/net/tg3.c +++ b/drivers/net/tg3.c @@ -3103,6 +3103,13 @@ static inline u32 tg3_tx_avail(struct tg3 *tp) ((tp->tx_prod - tp->tx_cons) & (TG3_TX_RING_SIZE - 1))); } +static inline void tg3_set_win(struct tg3 *tp) +{ + tp->dev->xmit_win = tg3_tx_avail(tp) - (MAX_SKB_FRAGS + 1); + if (tp->dev->xmit_win < 1) + tp->dev->xmit_win = 1; +} + /* Tigon3 never reports partial packet sends. So we do not * need special logic to handle SKBs that have not had all * of their frags sent yet, like SunGEM does. @@ -3165,8 +3172,10 @@ static void tg3_tx(struct tg3 *tp) (tg3_tx_avail(tp) > TG3_TX_WAKEUP_THRESH(tp)))) { netif_tx_lock(tp->dev); if (netif_queue_stopped(tp->dev) && - (tg3_tx_avail(tp) > TG3_TX_WAKEUP_THRESH(tp))) + (tg3_tx_avail(tp) > TG3_TX_WAKEUP_THRESH(tp))) { + tg3_set_win(tp); netif_wake_queue(tp->dev); + } netif_tx_unlock(tp->dev); } } @@ -4007,8 +4016,13 @@ void tg3_kick_DMA(struct net_device *dev) if (unlikely(tg3_tx_avail(tp) <= (MAX_SKB_FRAGS + 1))) { netif_stop_queue(dev); - if (tg3_tx_avail(tp) >= TG3_TX_WAKEUP_THRESH(tp)) + dev->xmit_win = 1; + if (tg3_tx_avail(tp) >= TG3_TX_WAKEUP_THRESH(tp)) { + tg3_set_win(tp); netif_wake_queue(dev); + } + } else { + tg3_set_win(tp); } mmiowb(); @@ -4085,6 +4099,7 @@ static int tg3_start_xmit(struct sk_buff *skb, struct net_device *dev) if (unlikely(tg3_tx_avail(tp) <= (skb_shinfo(skb)->nr_frags + 1))) { if (!netif_queue_stopped(dev)) { netif_stop_queue(dev); + tp->dev->xmit_win = 1; /* This is a hard error, log it. */ printk(KERN_ERR PFX "%s: BUG! Tx Ring full when " @@ -4100,6 +4115,25 @@ static int tg3_start_xmit(struct sk_buff *skb, struct net_device *dev) return ret; } +static int tg3_start_bxmit(struct sk_buff *skb, struct net_device *dev) +{ + struct tg3 *tp = netdev_priv(dev); + + if (unlikely(tg3_tx_avail(tp) <= (skb_shinfo(skb)->nr_frags + 1))) { + if (!netif_queue_stopped(dev)) { + netif_stop_queue(dev); + dev->xmit_win = 1; + + /* This is a hard error, log it. */ + printk(KERN_ERR PFX "%s: BUG! Tx Ring full when " + "queue awake!\n", dev->name); + } + return NETDEV_TX_BUSY; + } + + return tg3_enqueue(skb, dev); +} + static int tg3_start_xmit_dma_bug(struct sk_buff *, struct net_device *); /* Use GSO to workaround a rare TSO bug that may be triggered when the @@ -4112,9 +4146,11 @@ static int tg3_tso_bug(struct tg3 *tp, struct sk_buff *skb) /* Estimate the number of fragments in the worst case */ if (unlikely(tg3_tx_avail(tp) <= (skb_shinfo(skb)->gso_segs * 3))) { netif_stop_queue(tp->dev); + tp->dev->xmit_win = 1; if (tg3_tx_avail(tp) <= (skb_shinfo(skb)->gso_segs * 3)) return NETDEV_TX_BUSY; + tg3_set_win(tp); netif_wake_queue(tp->dev); } @@ -4267,6 +4303,25 @@ static int tg3_enqueue_buggy(struct sk_buff *skb, struct net_device *dev) return NETDEV_TX_OK; } +static int tg3_start_bxmit_dma_bug(struct sk_buff *skb, struct net_device *dev) +{ + struct tg3 *tp = netdev_priv(dev); + + if (unlikely(tg3_tx_avail(tp) <= (skb_shinfo(skb)->nr_frags + 1))) { + if (!netif_queue_stopped(dev)) { + netif_stop_queue(dev); + dev->xmit_win = 1; + + /* This is a hard error, log it. */ + printk(KERN_ERR PFX "%s: BUG! Tx Ring full when " + "queue awake!\n", dev->name); + } + return NETDEV_TX_BUSY; + } + + return tg3_enqueue_buggy(skb, dev); +} + static int tg3_start_xmit_dma_bug(struct sk_buff *skb, struct net_device *dev) { struct tg3 *tp = netdev_priv(dev); @@ -4283,6 +4338,7 @@ static int tg3_start_xmit_dma_bug(struct sk_buff *skb, struct net_device *dev) if (unlikely(tg3_tx_avail(tp) <= (skb_shinfo(skb)->nr_frags + 1))) { if (!netif_queue_stopped(dev)) { netif_stop_queue(dev); + dev->xmit_win = 1; /* This is a hard error, log it. */ printk(KERN_ERR PFX "%s: BUG! Tx Ring full when " @@ -11099,15 +11155,19 @@ static int __devinit tg3_get_invariants(struct tg3 *tp) else tp->tg3_flags &= ~TG3_FLAG_POLL_SERDES; + tp->dev->hard_end_xmit = tg3_kick_DMA; /* All chips before 5787 can get confused if TX buffers * straddle the 4GB address boundary in some cases. */ if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5755 || GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5787 || - GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5906) - tp->dev->hard_start_xmit = tg3_start_xmit; - else - tp->dev->hard_start_xmit = tg3_start_xmit_dma_bug; + GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5906) { + tp->dev->hard_start_xmit = tg3_start_bxmit; + tp->dev->hard_prep_xmit = tg3_prep_frame; + } else { + tp->dev->hard_start_xmit = tg3_start_bxmit_dma_bug; + tp->dev->hard_prep_xmit = tg3_prep_bug_frame; + } tp->rx_offset = 2; if (GET_ASIC_REV(tp->pci_chip_rev_id) == ASIC_REV_5701 && @@ -11955,6 +12015,8 @@ static int __devinit tg3_init_one(struct pci_dev *pdev, dev->watchdog_timeo = TG3_TX_TIMEOUT; dev->change_mtu = tg3_change_mtu; dev->irq = pdev->irq; + dev->features |= NETIF_F_BTX; + dev->xmit_win = tp->tx_pending >> 2; #ifdef CONFIG_NET_POLL_CONTROLLER dev->poll_controller = tg3_poll_controller; #endif