[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <EC64E4604C8024438259B683BB2514E1082B1CB1B1@SC-VEXCH1.marvell.com>
Date: Fri, 6 Aug 2010 08:18:19 -0700
From: Sachin Sanap <ssanap@...vell.com>
To: "20100804104750.GA9652@...vell.com"
<20100804104750.GA9652@...vell.com>,
"buytenh@...tstofly.org" <buytenh@...tstofly.org>
CC: "netdev@...r.kernel.org" <netdev@...r.kernel.org>,
Ashish Karkare <akarkare@...vell.com>,
Prabhanjan Sarnaik <sarnaik@...vell.com>,
"eric.y.miao@...il.com" <eric.y.miao@...il.com>,
Philip Rakity <prakity@...vell.com>,
Mark Brown <markb@...vell.com>
Subject: RE: [PATCH] net: add Fast Ethernet driver for PXA168.
Sorry for resending the patch as new thread. Discard this patch. I will send the patch again as a reply to previous thread.
-Sachin
> -----Original Message-----
> From: Sachin Sanap [mailto:ssanap@...vell.com]
> Sent: Saturday, August 07, 2010 1:10 AM
> To: buytenh@...tstofly.org
> Cc: netdev@...r.kernel.org; Ashish Karkare; Prabhanjan Sarnaik;
> eric.y.miao@...il.com; Philip Rakity; Mark Brown; Sachin Sanap
> Subject: [PATCH] net: add Fast Ethernet driver for PXA168.
>
> Signed-off-by: Sachin Sanap <ssanap@...vell.com>
> ---
> arch/arm/mach-mmp/aspenite.c | 3 +-
> arch/arm/mach-mmp/pxa168.c | 4 +-
> drivers/net/Kconfig | 10 +
> drivers/net/Makefile | 1 +
> drivers/net/pxa168_eth.c | 1592
> ++++++++++++++++++++++++++++++++++++++++++
> include/linux/pxa168_eth.h | 20 +
> 6 files changed, 1626 insertions(+), 4 deletions(-)
> create mode 100644 drivers/net/pxa168_eth.c
> create mode 100644 include/linux/pxa168_eth.h
>
> diff --git a/arch/arm/mach-mmp/aspenite.c b/arch/arm/mach-mmp/aspenite.c
> index 61a6d6a..cec505f 100644
> --- a/arch/arm/mach-mmp/aspenite.c
> +++ b/arch/arm/mach-mmp/aspenite.c
> @@ -200,8 +200,7 @@ static int pxa168_eth_init(void)
> }
>
> static struct pxa168_eth_platform_data pxa168_eth_data = {
> - .phy_addr = 0, /* phy addr depends on boards */
> - .port_number = 0,
> + .phy_addr = 0,
> .init = pxa168_eth_init,
> };
> #endif
> diff --git a/arch/arm/mach-mmp/pxa168.c b/arch/arm/mach-mmp/pxa168.c
> index f7d1158..6fcaf0c 100644
> --- a/arch/arm/mach-mmp/pxa168.c
> +++ b/arch/arm/mach-mmp/pxa168.c
> @@ -97,7 +97,7 @@ static struct clk_lookup pxa168_clkregs[] = {
> INIT_CLKREG(&clk_ssp4, "pxa168-ssp.3", NULL),
> INIT_CLKREG(&clk_ssp5, "pxa168-ssp.4", NULL),
> INIT_CLKREG(&clk_nand, "pxa3xx-nand", NULL),
> - INIT_CLKREG(&clk_mfu, "pxa168-mfu", "MFUCLK"),
> + INIT_CLKREG(&clk_mfu, "pxa168-eth", "MFUCLK"),
> };
>
> static int __init pxa168_init(void)
> @@ -149,4 +149,4 @@ PXA168_DEVICE(ssp2, "pxa168-ssp", 1, SSP2, 0xd401c000,
> 0x40, 54, 55);
> PXA168_DEVICE(ssp3, "pxa168-ssp", 2, SSP3, 0xd401f000, 0x40, 56, 57);
> PXA168_DEVICE(ssp4, "pxa168-ssp", 3, SSP4, 0xd4020000, 0x40, 58, 59);
> PXA168_DEVICE(ssp5, "pxa168-ssp", 4, SSP5, 0xd4021000, 0x40, 60, 61);
> -PXA168_DEVICE(mfu, "pxa168-mfu", -1, MFU, 0xc0800000, 0x0fff);
> +PXA168_DEVICE(mfu, "pxa168-eth", -1, MFU, 0xc0800000, 0x0fff);
> diff --git a/drivers/net/Kconfig b/drivers/net/Kconfig
> index ce2fcdd..78cd7e8 100644
> --- a/drivers/net/Kconfig
> +++ b/drivers/net/Kconfig
> @@ -927,6 +927,16 @@ config SMC91X
> The module will be called smc91x. If you want to compile it as a
> module, say M here and read
> <file:Documentation/kbuild/modules.txt>.
>
> +config PXA168_ETH
> + tristate "Marvell pxa168 ethernet support"
> + depends on CPU_PXA168
> + select PHYLIB
> + help
> + This driver supports the pxa168 Ethernet ports.
> +
> + To compile this driver as a module, choose M here. The module
> + will be called pxa168_eth.
> +
> config NET_NETX
> tristate "NetX Ethernet support"
> select MII
> diff --git a/drivers/net/Makefile b/drivers/net/Makefile
> index 0a0512a..a42d437 100644
> --- a/drivers/net/Makefile
> +++ b/drivers/net/Makefile
> @@ -244,6 +244,7 @@ obj-$(CONFIG_MYRI10GE) += myri10ge/
> obj-$(CONFIG_SMC91X) += smc91x.o
> obj-$(CONFIG_SMC911X) += smc911x.o
> obj-$(CONFIG_SMSC911X) += smsc911x.o
> +obj-$(CONFIG_PXA168_ETH) += pxa168_eth.o
> obj-$(CONFIG_BFIN_MAC) += bfin_mac.o
> obj-$(CONFIG_DM9000) += dm9000.o
> obj-$(CONFIG_PASEMI_MAC) += pasemi_mac_driver.o
> diff --git a/drivers/net/pxa168_eth.c b/drivers/net/pxa168_eth.c
> new file mode 100644
> index 0000000..c00fc3c
> --- /dev/null
> +++ b/drivers/net/pxa168_eth.c
> @@ -0,0 +1,1592 @@
> +/*
> + * PXA168 ethernet driver.
> + * Most of the code is derived from mv643xx ethernet driver.
> + *
> + * Copyright (C) 2010 Marvell International Ltd.
> + * Philip Rakity <prakity@...vell.com>
> + * Mark Brown <markb@...vell.com>
> + * Sachin Sanap <ssanap@...vell.com>
> + *
> + * This program is free software; you can redistribute it and/or
> + * modify it under the terms of the GNU General Public License
> + * as published by the Free Software Foundation; either version 2
> + * of the License, or (at your option) any later version.
> + *
> + * This program is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
> + * GNU General Public License for more details.
> + *
> + * You should have received a copy of the GNU General Public License
> + * along with this program; if not, write to the Free Software
> + * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307,
> USA.
> + */
> +
> +#include <linux/init.h>
> +#include <linux/dma-mapping.h>
> +#include <linux/in.h>
> +#include <linux/ip.h>
> +#include <linux/tcp.h>
> +#include <linux/udp.h>
> +#include <linux/etherdevice.h>
> +#include <linux/bitops.h>
> +#include <linux/delay.h>
> +#include <linux/ethtool.h>
> +#include <linux/platform_device.h>
> +#include <linux/module.h>
> +#include <linux/kernel.h>
> +#include <linux/workqueue.h>
> +#include <linux/clk.h>
> +#include <linux/phy.h>
> +#include <linux/io.h>
> +#include <linux/types.h>
> +#include <asm/pgtable.h>
> +#include <asm/system.h>
> +#include <linux/delay.h>
> +#include <linux/dma-mapping.h>
> +#include <asm/cacheflush.h>
> +#include <linux/pxa168_eth.h>
> +
> +#define DRIVER_NAME "pxa168-eth"
> +#define DRIVER_VERSION "0.3"
> +
> +/*
> + * Registers
> + */
> +
> +#define PHY_ADDRESS 0x0000
> +#define SMI 0x0010
> +#define PORT_CONFIG 0x0400
> +#define PORT_CONFIG_EXT 0x0408
> +#define PORT_COMMAND 0x0410
> +#define PORT_STATUS 0x0418
> +#define HTPR 0x0428
> +#define SDMA_CONFIG 0x0440
> +#define SDMA_CMD 0x0448
> +#define INT_CAUSE 0x0450
> +#define INT_W_CLEAR 0x0454
> +#define INT_MASK 0x0458
> +#define ETH_F_RX_DESC_0 0x0480
> +#define ETH_C_RX_DESC_0 0x04A0
> +#define ETH_C_TX_DESC_1 0x04E4
> +
> +/* smi register */
> +#define SMI_BUSY (1 << 28) /* 0 - Write, 1 - Read */
> +#define SMI_R_VALID (1 << 27) /* 0 - Write, 1 - Read */
> +#define SMI_OP_W (0 << 26) /* Write operation */
> +#define SMI_OP_R (1 << 26) /* Read operation */
> +
> +#define PHY_WAIT_ITERATIONS 500
> +
> +#define PXA168_ETH_PHY_ADDR_DEFAULT 0
> +/* RX & TX descriptor command */
> +#define BUF_OWNED_BY_DMA (1 << 31)
> +
> +/* RX descriptor status */
> +#define RX_EN_INT (1 << 23)
> +#define RX_FIRST_DESC (1 << 17)
> +#define RX_LAST_DESC (1 << 16)
> +#define RX_ERROR (1 << 15)
> +
> +/* TX descriptor command */
> +#define TX_EN_INT (1 << 23)
> +#define TX_GEN_CRC (1 << 22)
> +#define TX_ZERO_PADDING (1 << 18)
> +#define TX_FIRST_DESC (1 << 17)
> +#define TX_LAST_DESC (1 << 16)
> +#define TX_ERROR (1 << 15)
> +
> +/* SDMA_CMD */
> +#define SDMA_CMD_AT (1 << 31)
> +#define SDMA_CMD_TXDL (1 << 24)
> +#define SDMA_CMD_TXDH (1 << 23)
> +#define SDMA_CMD_AR (1 << 15)
> +#define SDMA_CMD_ERD (1 << 7)
> +
> +/* Bit definitions of the Port Config Reg */
> +#define PCR_HS (1 << 12)
> +#define PCR_EN (1 << 7)
> +#define PCR_PM (1 << 0)
> +
> +/* Bit definitions of the Port Config Extend Reg */
> +#define PCXR_2BSM (1 << 28)
> +#define PCXR_DSCP_EN (1 << 21)
> +#define PCXR_MFL_1518 (0 << 14)
> +#define PCXR_MFL_1536 (1 << 14)
> +#define PCXR_MFL_2048 (2 << 14)
> +#define PCXR_MFL_64K (3 << 14)
> +#define PCXR_FLP (1 << 11)
> +#define PCXR_PRIO_TX_OFF 3
> +#define PCXR_TX_HIGH_PRI (7 << PCXR_PRIO_TX_OFF)
> +
> +/* Bit definitions of the SDMA Config Reg */
> +#define SDCR_BSZ_OFF 12
> +#define SDCR_BSZ8 (3 << SDCR_BSZ_OFF)
> +#define SDCR_BSZ4 (2 << SDCR_BSZ_OFF)
> +#define SDCR_BSZ2 (1 << SDCR_BSZ_OFF)
> +#define SDCR_BSZ1 (0 << SDCR_BSZ_OFF)
> +#define SDCR_BLMR (1 << 6)
> +#define SDCR_BLMT (1 << 7)
> +#define SDCR_RIFB (1 << 9)
> +#define SDCR_RC_OFF 2
> +#define SDCR_RC_MAX_RETRANS (0xf << SDCR_RC_OFF)
> +
> +/*
> + * Bit definitions of the Interrupt Cause Reg
> + * and Interrupt MASK Reg is the same
> + */
> +#define ICR_RXBUF (1 << 0)
> +#define ICR_TXBUF_H (1 << 2)
> +#define ICR_TXBUF_L (1 << 3)
> +#define ICR_TXEND_H (1 << 6)
> +#define ICR_TXEND_L (1 << 7)
> +#define ICR_RXERR (1 << 8)
> +#define ICR_TXERR_H (1 << 10)
> +#define ICR_TXERR_L (1 << 11)
> +#define ICR_TX_UDR (1 << 13)
> +#define ICR_MII_CH (1 << 28)
> +
> +#define ALL_INTS (ICR_TXBUF_H | ICR_TXBUF_L | ICR_TX_UDR |\
> + ICR_TXERR_H | ICR_TXERR_L |\
> + ICR_TXEND_H | ICR_TXEND_L |\
> + ICR_RXBUF | ICR_RXERR | ICR_MII_CH)
> +
> +#define ETH_HW_IP_ALIGN 2 /* hw aligns IP header */
> +#define ETH_EXTRA_HEADER (6+6+2+4) /* dest+src addr+protocol id+crc
> */
> +#define ETH_DATA_LEN 1500
> +#define MAX_PKT_SIZE 1518
> +
> +#define NUM_RX_DESCS 64
> +#define NUM_TX_DESCS 64
> +#define MAX_DESCS_PER_HIGH (60)
> +#define TX_DESC_COUNT_LOW (10)
> +
> +#define HASH_ADD 0
> +#define HASH_DELETE 1
> +#define HASH_ADDR_TABLE_SIZE 0x4000 /* 16K (1/2K address - PCR_HS ==
> 1) */
> +#define HOP_NUMBER 12
> +
> +/* Bit definitions for Port status */
> +#define PORT_SPEED_100 (1 << 0)
> +#define FULL_DUPLEX (1 << 1)
> +#define FLOW_CONTROL_ENABLED (1 << 2)
> +#define LINK_UP (1 << 3)
> +
> +/* Bit definitions for work to be done */
> +#define WORK_LINK (1 << 0)
> +#define WORK_TX_DONE (1 << 1)
> +
> +#define TX_DONE_INTERVAL 30
> +
> +struct rx_desc {
> + u32 cmd_sts; /* Descriptor command status */
> + u16 byte_cnt; /* Descriptor buffer byte count */
> + u16 buf_size; /* Buffer size */
> + u32 buf_ptr; /* Descriptor buffer pointer */
> + u32 next_desc_ptr; /* Next descriptor pointer */
> +};
> +
> +struct tx_desc {
> + u32 cmd_sts; /* Command/status field */
> + u16 reserved;
> + u16 byte_cnt; /* buffer byte count */
> + u32 buf_ptr; /* pointer to buffer for this descriptor */
> + u32 next_desc_ptr; /* Pointer to next descriptor */
> +};
> +
> +struct pxa168_eth_private {
> + int rx_resource_err; /* Rx ring resource error flag */
> +
> + /* Next available and first returning Rx resource */
> + int rx_curr_desc_q, rx_used_desc_q;
> +
> + /* Next available and first returning Tx resource */
> + int tx_curr_desc_q, tx_used_desc_q;
> +
> + struct rx_desc *p_rx_desc_area;
> + dma_addr_t rx_desc_dma;
> + int rx_desc_area_size;
> + struct sk_buff **rx_skb;
> +
> + struct tx_desc *p_tx_desc_area;
> + dma_addr_t tx_desc_dma;
> + int tx_desc_area_size;
> + struct sk_buff **tx_skb;
> +
> + struct work_struct tx_timeout_task;
> +
> + struct net_device *dev;
> + struct napi_struct napi;
> + u8 work_todo;
> +
> + struct net_device_stats stats;
> + /* Size of Tx Ring per queue */
> + int tx_ring_size;
> + /* Number of tx descriptors in use */
> + int tx_desc_count;
> + /* Size of Rx Ring per queue */
> + int rx_ring_size;
> + /* Number of rx descriptors in use */
> + int rx_desc_count;
> +
> + /*
> + * Used in case RX Ring is empty, which can occur when
> + * system does not have resources (skb's)
> + */
> + struct timer_list timeout;
> + struct mii_bus *smi_bus;
> + struct phy_device *phy;
> +
> + /* clock */
> + struct clk *clk;
> + struct pxa168_eth_platform_data *pd;
> + /*
> + * Ethernet controller base address.
> + */
> + void __iomem *base;
> +
> + /* Pointer to the hardware address filter table */
> + void *htpr;
> + dma_addr_t htpr_dma;
> +};
> +
> +struct addr_table_entry {
> + u32 lo;
> + u32 hi;
> +};
> +
> +/* Bit fields of a Hash Table Entry */
> +enum hash_table_entry {
> + HASH_ENTRY_VALID = 1,
> + SKIP = 2,
> + HASH_ENTRY_RECEIVE_DISCARD = 4,
> + HASH_ENTRY_RECEIVE_DISCARD_BIT = 2
> +};
> +
> +static int pxa168_get_settings(struct net_device *dev, struct ethtool_cmd
> *cmd);
> +static int pxa168_set_settings(struct net_device *dev, struct ethtool_cmd
> *cmd);
> +static int pxa168_init_hw(struct pxa168_eth_private *pep);
> +static void eth_port_reset(struct net_device *dev);
> +static void eth_port_start(struct net_device *dev);
> +static int pxa168_eth_open(struct net_device *dev);
> +static int pxa168_eth_stop(struct net_device *dev);
> +static int ethernet_phy_setup(struct net_device *dev);
> +
> +static inline u32 rdl(struct pxa168_eth_private *pep, int offset)
> +{
> + return readl(pep->base + offset);
> +}
> +
> +static inline void wrl(struct pxa168_eth_private *pep, int offset, u32
> data)
> +{
> + writel(data, pep->base + offset);
> +}
> +
> +static void abort_dma(struct pxa168_eth_private *pep)
> +{
> + int delay;
> + int max_retries = 40;
> +
> + do {
> + wrl(pep, SDMA_CMD, SDMA_CMD_AR | SDMA_CMD_AT);
> + udelay(100);
> +
> + delay = 10;
> + while ((rdl(pep, SDMA_CMD) & (SDMA_CMD_AR | SDMA_CMD_AT))
> + && delay-- > 0) {
> + udelay(10);
> + }
> + } while (max_retries-- > 0 && delay <= 0);
> +
> + if (max_retries <= 0)
> + printk(KERN_ERR "%s : DMA Stuck\n", __func__);
> +}
> +
> +static int ethernet_phy_get(struct pxa168_eth_private *pep)
> +{
> + unsigned int reg_data;
> +
> + reg_data = rdl(pep, PHY_ADDRESS);
> +
> + return reg_data & 0x1f;
> +}
> +
> +static void ethernet_phy_set_addr(struct pxa168_eth_private *pep, int
> phy_addr)
> +{
> + u32 reg_data;
> +
> + reg_data = rdl(pep, PHY_ADDRESS);
> + reg_data &= ~(0x1f);
> + reg_data |= phy_addr & 0x1f;
> + wrl(pep, PHY_ADDRESS, reg_data);
> +}
> +
> +static void ethernet_phy_reset(struct pxa168_eth_private *pep)
> +{
> + int data;
> +
> + data = phy_read(pep->phy, MII_BMCR);
> + if (data < 0)
> + return;
> +
> + data |= BMCR_RESET;
> + if (phy_write(pep->phy, MII_BMCR, data) < 0)
> + return;
> +
> + do {
> + data = phy_read(pep->phy, MII_BMCR);
> + } while (data >= 0 && data & BMCR_RESET);
> +}
> +
> +static void rxq_refill(struct net_device *dev)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + struct sk_buff *skb;
> + struct rx_desc *p_used_rx_desc;
> + int used_rx_desc;
> +
> + while (pep->rx_desc_count < pep->rx_ring_size) {
> + skb = dev_alloc_skb(MAX_PKT_SIZE + ETH_HW_IP_ALIGN);
> + if (!skb)
> + break;
> + pep->rx_desc_count++;
> + /* Get 'used' Rx descriptor */
> + used_rx_desc = pep->rx_used_desc_q;
> + p_used_rx_desc = &pep->p_rx_desc_area[used_rx_desc];
> + p_used_rx_desc->buf_ptr = dma_map_single(NULL,
> + skb->data,
> + MAX_PKT_SIZE +
> + ETH_HW_IP_ALIGN,
> + DMA_FROM_DEVICE);
> + p_used_rx_desc->buf_size = MAX_PKT_SIZE + ETH_HW_IP_ALIGN;
> + pep->rx_skb[used_rx_desc] = skb;
> +
> + /* Return the descriptor to DMA ownership */
> + wmb();
> + p_used_rx_desc->cmd_sts = BUF_OWNED_BY_DMA | RX_EN_INT;
> + wmb();
> +
> + /* Move the used descriptor pointer to the next descriptor */
> + pep->rx_used_desc_q = (used_rx_desc + 1) % pep->rx_ring_size;
> +
> + /* Any Rx return cancels the Rx resource error status */
> + pep->rx_resource_err = 0;
> +
> + skb_reserve(skb, ETH_HW_IP_ALIGN);
> + }
> +
> + /*
> + * If RX ring is empty of SKB, set a timer to try allocating
> + * again at a later time.
> + */
> + if (pep->rx_desc_count == 0) {
> + pep->timeout.expires = jiffies + (HZ / 10); /* 100 mSec */
> + add_timer(&pep->timeout);
> + }
> +}
> +
> +static inline void rxq_refill_timer_wrapper(unsigned long data)
> +{
> + struct pxa168_eth_private *pep = (void *)data;
> + napi_schedule(&pep->napi);
> +}
> +
> +static inline u32 nibble_swapping_32_bit(u32 x)
> +{
> + return (((x) & 0xf0f0f0f0) >> 4) | (((x) & 0x0f0f0f0f) << 4);
> +}
> +
> +static inline u32 nibble_swapping_16_bit(u32 x)
> +{
> + return (((x) & 0x0000f0f0) >> 4) | (((x) & 0x00000f0f) << 4);
> +}
> +
> +static inline u32 flip_4_bits(u32 x)
> +{
> + return (((x) & 0x01) << 3) | (((x) & 0x002) << 1)
> + | (((x) & 0x04) >> 1) | (((x) & 0x008) >> 3);
> +}
> +
> +/*
> + * ----------------------------------------------------------------------
> ------
> + * This function will calculate the hash function of the address.
> + * depends on the hash mode and hash size.
> + * Inputs
> + * mac_high - the 2 most significant bytes of the MAC address.
> + * mac_low - the 4 least significant bytes of the MAC address.
> + * Outputs
> + * return the calculated entry.
> + */
> +static u32 hash_function(u32 mac_high, u32 mac_low)
> +{
> + u32 hash_result;
> + u32 addr_high;
> + u32 addr_low;
> + u32 addr0;
> + u32 addr1;
> + u32 addr2;
> + u32 addr3;
> + u32 addr_high_swapped;
> + u32 addr_low_swapped;
> +
> + addr_high = nibble_swapping_16_bit(mac_high);
> + addr_low = nibble_swapping_32_bit(mac_low);
> +
> + addr_high_swapped = flip_4_bits(addr_high & 0xf)
> + + ((flip_4_bits((addr_high >> 4) & 0xf)) << 4)
> + + ((flip_4_bits((addr_high >> 8) & 0xf)) << 8)
> + + ((flip_4_bits((addr_high >> 12) & 0xf)) << 12);
> +
> + addr_low_swapped = flip_4_bits(addr_low & 0xf)
> + + ((flip_4_bits((addr_low >> 4) & 0xf)) << 4)
> + + ((flip_4_bits((addr_low >> 8) & 0xf)) << 8)
> + + ((flip_4_bits((addr_low >> 12) & 0xf)) << 12)
> + + ((flip_4_bits((addr_low >> 16) & 0xf)) << 16)
> + + ((flip_4_bits((addr_low >> 20) & 0xf)) << 20)
> + + ((flip_4_bits((addr_low >> 24) & 0xf)) << 24)
> + + ((flip_4_bits((addr_low >> 28) & 0xf)) << 28);
> +
> + addr_high = addr_high_swapped;
> + addr_low = addr_low_swapped;
> +
> + addr0 = (addr_low >> 2) & 0x03f;
> + addr1 = (addr_low & 0x003) | ((addr_low >> 8) & 0x7f) << 2;
> + addr2 = (addr_low >> 15) & 0x1ff;
> + addr3 = ((addr_low >> 24) & 0x0ff) | ((addr_high & 1) << 8);
> +
> + hash_result = (addr0 << 9) | (addr1 ^ addr2 ^ addr3);
> + hash_result = hash_result & 0x07ff;
> + return hash_result;
> +}
> +
> +/*
> + * ----------------------------------------------------------------------
> ------
> + * This function will add/del an entry to the address table.
> + * Inputs
> + * pep - ETHERNET .
> + * mac_high - the 2 most significant bytes of the MAC address.
> + * mac_low - the 4 least significant bytes of the MAC address.
> + * skip - if 1, skip this address.Used in case of deleting an entry which
> is a
> + * part of chain in the hash table.We cant just delete the entry
> since
> + * that will break the chain.We need to defragment the tables time to
> + * time.
> + * rd - 0 Discard packet upon match.
> + * - 1 Receive packet upon match.
> + * Outputs
> + * address table entry is added/deleted.
> + * 0 if success.
> + * -ENOSPC if table full
> + */
> +static int add_del_hash_entry(struct pxa168_eth_private *pep, u32
> mac_high,
> + u32 mac_low, u32 rd, u32 skip, int del)
> +{
> + struct addr_table_entry *entry, *start;
> + u32 new_high;
> + u32 new_low;
> + u32 i;
> +
> + new_low = (((mac_high >> 4) & 0xf) << 15)
> + | (((mac_high >> 0) & 0xf) << 11)
> + | (((mac_high >> 12) & 0xf) << 7)
> + | (((mac_high >> 8) & 0xf) << 3)
> + | (((mac_low >> 20) & 0x1) << 31)
> + | (((mac_low >> 16) & 0xf) << 27)
> + | (((mac_low >> 28) & 0xf) << 23)
> + | (((mac_low >> 24) & 0xf) << 19)
> + | (skip << SKIP) | (rd << HASH_ENTRY_RECEIVE_DISCARD_BIT)
> + | HASH_ENTRY_VALID;
> +
> + new_high = (((mac_low >> 4) & 0xf) << 15)
> + | (((mac_low >> 0) & 0xf) << 11)
> + | (((mac_low >> 12) & 0xf) << 7)
> + | (((mac_low >> 8) & 0xf) << 3)
> + | (((mac_low >> 21) & 0x7) << 0);
> +
> + /*
> + * Pick the appropriate table, start scanning for free/reusable
> + * entries at the index obtained by hashing the specified MAC
> address
> + */
> + start = (struct addr_table_entry *)(pep->htpr);
> + entry = start + hash_function(mac_high, mac_low);
> + for (i = 0; i < HOP_NUMBER; i++) {
> + if (!(entry->lo & HASH_ENTRY_VALID)) {
> + break;
> + } else {
> + /* if same address put in same position */
> + if (((entry->lo & 0xfffffff8) == (new_low & 0xfffffff8))
> + && (entry->hi == new_high)) {
> + break;
> + }
> + }
> + if (entry == start + 0x7ff)
> + entry = start;
> + else
> + entry++;
> + }
> +
> + if (((entry->lo & 0xfffffff8) != (new_low & 0xfffffff8)) &&
> + (entry->hi != new_high) && del)
> + return 0;
> +
> + if (i == HOP_NUMBER) {
> + if (!del) {
> + printk(KERN_INFO "%s: table section is full\n",
> + __FILE__);
> + return -ENOSPC;
> + } else
> + return 0;
> + }
> +
> + /*
> + * Update the selected entry
> + */
> + if (del) {
> + entry->hi = 0;
> + entry->lo = 0;
> + } else {
> + entry->hi = cpu_to_le32(new_high);
> + entry->lo = cpu_to_le32(new_low);
> + }
> +
> + return 0;
> +}
> +
> +/*
> + * ----------------------------------------------------------------------
> ------
> + * Create an addressTable entry from MAC address info
> + * found in the specifed net_device struct
> + *
> + * Input : pointer to ethernet interface network device structure
> + * Output : N/A
> + */
> +static void update_hash_table_mac_address(struct pxa168_eth_private *pep,
> + u8 *oaddr, u8 *addr)
> +{
> + u32 mac_high;
> + u32 mac_low;
> +
> + /* Delete old entry */
> + if (oaddr) {
> + mac_high = cpu_to_le32((oaddr[0] << 8) | oaddr[1]);
> + mac_low = cpu_to_le32((oaddr[2] << 24) | (oaddr[3] << 16) |
> + (oaddr[4] << 8) | oaddr[5]);
> + add_del_hash_entry(pep, mac_high, mac_low, 1, 0, HASH_DELETE);
> + }
> + /* Add new entry */
> + mac_high = cpu_to_le32((addr[0] << 8) | addr[1]);
> + mac_low = cpu_to_le32((addr[2] << 24) | (addr[3] << 16) |
> + (addr[4] << 8) | addr[5]);
> + add_del_hash_entry(pep, mac_high, mac_low, 1, 0, HASH_ADD);
> +}
> +
> +static int init_hash_table(struct pxa168_eth_private *pep)
> +{
> + /*
> + * Hardware expects CPU to build a hash table based on a predefined
> + * hash function and populate it based on hardware address. The
> + * location of the hash table is identified by 32-bit pointer stored
> + * in HTPR internal register. Two possible sizes exists for the hash
> + * table 256kB and 16kB.We currently only support 16kB.
> + */
> + /* TODO: Add support for 256kB hash table */
> + if (pep->htpr == NULL) {
> + pep->htpr = dma_alloc_coherent(NULL,
> + HASH_ADDR_TABLE_SIZE,
> + &pep->htpr_dma, GFP_KERNEL);
> + if (pep->htpr == NULL)
> + return -ENOMEM;
> + }
> + memset(pep->htpr, 0, HASH_ADDR_TABLE_SIZE);
> + wrl(pep, HTPR, pep->htpr_dma);
> + return 0;
> +}
> +
> +static void pxa168_eth_set_rx_mode(struct net_device *dev)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + struct netdev_hw_addr *ha;
> + u32 val;
> +
> + val = rdl(pep, PORT_CONFIG);
> + if (dev->flags & IFF_PROMISC)
> + val |= PCR_PM;
> + else
> + val &= ~PCR_PM;
> + wrl(pep, PORT_CONFIG, val);
> + netdev_for_each_mc_addr(ha, dev)
> + update_hash_table_mac_address(pep, NULL, ha->addr);
> +}
> +
> +static int pxa168_eth_set_mac_address(struct net_device *dev, void *addr)
> +{
> + struct sockaddr *sa = addr;
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + unsigned char oldMac[ETH_ALEN];
> +
> + if (!is_valid_ether_addr(sa->sa_data))
> + return -EINVAL;
> + memcpy(oldMac, dev->dev_addr, ETH_ALEN);
> + memcpy(dev->dev_addr, sa->sa_data, ETH_ALEN);
> + netif_addr_lock_bh(dev);
> + update_hash_table_mac_address(pep, oldMac, dev->dev_addr);
> + netif_addr_unlock_bh(dev);
> + return 0;
> +}
> +
> +static void eth_port_start(struct net_device *dev)
> +{
> + unsigned int val = 0;
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + int tx_curr_desc, rx_curr_desc;
> +
> + /* Perform PHY reset, if there is a PHY. */
> + if (pep->phy != NULL) {
> + struct ethtool_cmd cmd;
> +
> + pxa168_get_settings(pep->dev, &cmd);
> + ethernet_phy_reset(pep);
> + pxa168_set_settings(pep->dev, &cmd);
> + }
> +
> + /* Assignment of Tx CTRP of given queue */
> + tx_curr_desc = pep->tx_curr_desc_q;
> + wrl(pep, ETH_C_TX_DESC_1,
> + (u32) ((struct tx_desc *)pep->tx_desc_dma + tx_curr_desc));
> +
> + /* Assignment of Rx CRDP of given queue */
> + rx_curr_desc = pep->rx_curr_desc_q;
> + wrl(pep, ETH_C_RX_DESC_0,
> + (u32) ((struct rx_desc *)pep->rx_desc_dma + rx_curr_desc));
> +
> + wrl(pep, ETH_F_RX_DESC_0,
> + (u32) ((struct rx_desc *)pep->rx_desc_dma + rx_curr_desc));
> +
> + /* Clear all interrupts */
> + wrl(pep, INT_CAUSE, 0);
> +
> + /* Enable all interrupts for receive, transmit and error. */
> + wrl(pep, INT_MASK, ALL_INTS);
> +
> + val = rdl(pep, PORT_CONFIG);
> + val |= PCR_EN;
> + wrl(pep, PORT_CONFIG, val);
> +
> + /* Start RX DMA engine */
> + val = rdl(pep, SDMA_CMD);
> + val |= SDMA_CMD_ERD;
> + wrl(pep, SDMA_CMD, val);
> +}
> +
> +static void eth_port_reset(struct net_device *dev)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + unsigned int val = 0;
> +
> + /* Stop all interrupts for receive, transmit and error. */
> + wrl(pep, INT_MASK, 0);
> +
> + /* Clear all interrupts */
> + wrl(pep, INT_CAUSE, 0);
> +
> + /* Stop RX DMA */
> + val = rdl(pep, SDMA_CMD);
> + val &= ~SDMA_CMD_ERD; /* abort dma command */
> +
> + /* Abort any transmit and receive operations and put DMA
> + * in idle state.
> + */
> + abort_dma(pep);
> +
> + /* Disable port */
> + val = rdl(pep, PORT_CONFIG);
> + val &= ~PCR_EN;
> + wrl(pep, PORT_CONFIG, val);
> +}
> +
> +/*
> + * txq_reclaim - Free the tx desc data for completed descriptors
> + * If force is non-zero, frees uncompleted descriptors as well
> + */
> +static int txq_reclaim(struct net_device *dev, int force)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + struct tx_desc *desc;
> + u32 cmd_sts;
> + struct sk_buff *skb;
> + int tx_index;
> + dma_addr_t addr;
> + int count;
> + int released = 0;
> +
> + netif_tx_lock(dev);
> +
> + pep->work_todo &= ~(WORK_TX_DONE);
> + while (pep->tx_desc_count > 0) {
> + tx_index = pep->tx_used_desc_q;
> + desc = &pep->p_tx_desc_area[tx_index];
> + cmd_sts = desc->cmd_sts;
> + if (!force && (cmd_sts & BUF_OWNED_BY_DMA)) {
> + if (released > 0) {
> + goto txq_reclaim_end;
> + } else {
> + released = -1;
> + goto txq_reclaim_end;
> + }
> + }
> + pep->tx_used_desc_q = (tx_index + 1) % pep->tx_ring_size;
> + pep->tx_desc_count--;
> + addr = desc->buf_ptr;
> + count = desc->byte_cnt;
> + skb = pep->tx_skb[tx_index];
> + if (skb)
> + pep->tx_skb[tx_index] = NULL;
> +
> + if (cmd_sts & TX_ERROR) {
> + if (net_ratelimit())
> + printk(KERN_ERR "%s: Error in TX\n", dev->name);
> + dev->stats.tx_errors++;
> + }
> + dma_unmap_single(NULL, addr, count, DMA_TO_DEVICE);
> + if (skb)
> + dev_kfree_skb_irq(skb);
> + released++;
> + }
> +txq_reclaim_end:
> + netif_tx_unlock(dev);
> + return released;
> +}
> +
> +static void pxa168_eth_tx_timeout(struct net_device *dev)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> +
> + printk(KERN_INFO "%s: TX timeout desc_count %d\n",
> + dev->name, pep->tx_desc_count);
> +
> + schedule_work(&pep->tx_timeout_task);
> +}
> +
> +static void pxa168_eth_tx_timeout_task(struct work_struct *work)
> +{
> + struct pxa168_eth_private *pep = container_of(work,
> + struct pxa168_eth_private,
> + tx_timeout_task);
> + struct net_device *dev = pep->dev;
> + pxa168_eth_stop(dev);
> + pxa168_eth_open(dev);
> +}
> +
> +static int rxq_process(struct net_device *dev, int budget)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + struct net_device_stats *stats = &dev->stats;
> + unsigned int received_packets = 0;
> + struct sk_buff *skb;
> +
> + while (budget-- > 0) {
> +
> + int rx_next_curr_desc, rx_curr_desc, rx_used_desc;
> + struct rx_desc *rx_desc;
> + unsigned int cmd_sts;
> +
> + /* Do not process Rx ring in case of Rx ring resource error */
> + if (pep->rx_resource_err)
> + break;
> + rx_curr_desc = pep->rx_curr_desc_q;
> + rx_used_desc = pep->rx_used_desc_q;
> + rx_desc = &pep->p_rx_desc_area[rx_curr_desc];
> + cmd_sts = rx_desc->cmd_sts;
> + rmb();
> + if (cmd_sts & (BUF_OWNED_BY_DMA))
> + break;
> + skb = pep->rx_skb[rx_curr_desc];
> + pep->rx_skb[rx_curr_desc] = NULL;
> +
> + rx_next_curr_desc = (rx_curr_desc + 1) % pep->rx_ring_size;
> + pep->rx_curr_desc_q = rx_next_curr_desc;
> +
> + /* Rx descriptors exhausted. */
> + /* Set the Rx ring resource error flag */
> + if (rx_next_curr_desc == rx_used_desc)
> + pep->rx_resource_err = 1;
> + pep->rx_desc_count--;
> + dma_unmap_single(NULL, rx_desc->buf_ptr,
> + MAX_PKT_SIZE + ETH_HW_IP_ALIGN,
> + DMA_FROM_DEVICE);
> + received_packets++;
> + /*
> + * Update statistics.
> + * Note byte count includes 4 byte CRC count
> + */
> + stats->rx_packets++;
> + stats->rx_bytes += rx_desc->byte_cnt;
> + /*
> + * In case received a packet without first / last bits on OR
> + * the error summary bit is on, the packets needs to be droped.
> + */
> + if (((cmd_sts & (RX_FIRST_DESC | RX_LAST_DESC)) !=
> + (RX_FIRST_DESC | RX_LAST_DESC))
> + || (cmd_sts & RX_ERROR)) {
> +
> + stats->rx_dropped++;
> + if ((cmd_sts & (RX_FIRST_DESC | RX_LAST_DESC)) !=
> + (RX_FIRST_DESC | RX_LAST_DESC)) {
> + if (net_ratelimit())
> + printk(KERN_ERR
> + "%s: Rx pkt on multiple desc\n",
> + dev->name);
> + }
> + if (cmd_sts & RX_ERROR)
> + stats->rx_errors++;
> + dev_kfree_skb_irq(skb);
> + } else {
> + /*
> + * The -4 is for the CRC in the trailer of the
> + * received packet
> + */
> + skb_put(skb, rx_desc->byte_cnt - 4);
> + skb->protocol = eth_type_trans(skb, dev);
> + netif_receive_skb(skb);
> + }
> + dev->last_rx = jiffies;
> + }
> + /* Fill RX ring with skb's */
> + rxq_refill(dev);
> + return received_packets;
> +}
> +
> +static int pxa168_eth_collect_events(struct pxa168_eth_private *pep,
> + struct net_device *dev)
> +{
> + u32 icr;
> + int ret = 0;
> +
> + icr = rdl(pep, INT_CAUSE);
> + if (0x00 == icr)
> + return IRQ_NONE;
> +
> + wrl(pep, INT_CAUSE, icr ^ 0xffffffff);
> + if (icr & (ICR_TXBUF_H | ICR_TXBUF_L)) {
> + pep->work_todo |= WORK_TX_DONE;
> + ret = 1;
> + }
> + if (icr & ICR_RXBUF)
> + ret = 1;
> + if (icr & ICR_MII_CH) {
> + pep->work_todo |= WORK_LINK;
> + ret = 1;
> + }
> + return ret;
> +}
> +
> +static void handle_link_event(struct pxa168_eth_private *pep)
> +{
> + struct net_device *dev = pep->dev;
> + u32 port_status;
> + int speed;
> + int duplex;
> + int fc;
> +
> + port_status = rdl(pep, PORT_STATUS);
> + if (!(port_status & LINK_UP)) {
> + if (netif_carrier_ok(dev)) {
> + printk(KERN_INFO "%s: link down\n", dev->name);
> + netif_carrier_off(dev);
> + txq_reclaim(dev, 1);
> + }
> + return;
> + }
> + if (port_status & PORT_SPEED_100)
> + speed = 100;
> + else
> + speed = 10;
> +
> + duplex = (port_status & FULL_DUPLEX) ? 1 : 0;
> + fc = (port_status & FLOW_CONTROL_ENABLED) ? 1 : 0;
> + printk(KERN_INFO "%s: link up, %d Mb/s, %s duplex, "
> + "flow control %sabled\n", dev->name,
> + speed, duplex ? "full" : "half", fc ? "en" : "dis");
> + if (!netif_carrier_ok(dev))
> + netif_carrier_on(dev);
> +}
> +
> +static irqreturn_t pxa168_eth_int_handler(int irq, void *dev_id)
> +{
> + struct net_device *dev = (struct net_device *)dev_id;
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> +
> + if (unlikely(!pxa168_eth_collect_events(pep, dev)))
> + return IRQ_NONE;
> + /* Disable interrupts */
> + wrl(pep, INT_MASK, 0);
> + napi_schedule(&pep->napi);
> + return IRQ_HANDLED;
> +}
> +
> +static int set_port_config_ext(struct pxa168_eth_private *pep, int mtu)
> +{
> + int mtu_size;
> +
> + if ((mtu > ETH_DATA_LEN) || (mtu < 64))
> + return -EINVAL;
> +
> + mtu_size = PCXR_MFL_1518;
> + /* Extended Port Configuration */
> + wrl(pep,
> + PORT_CONFIG_EXT, PCXR_2BSM | /* Two byte suffix aligns IP hdr */
> + PCXR_DSCP_EN | /* Enable DSCP in IP */
> + mtu_size | PCXR_FLP | /* do not force link pass */
> + PCXR_TX_HIGH_PRI); /* Transmit - high priority queue */
> +
> + (pep->dev)->mtu = mtu;
> + return 0;
> +}
> +
> +static int pxa168_init_hw(struct pxa168_eth_private *pep)
> +{
> + int err = 0;
> +
> + /* Disable interrupts */
> + wrl(pep, INT_MASK, 0);
> + wrl(pep, INT_CAUSE, 0);
> + /* Write to ICR to clear interrupts. */
> + wrl(pep, INT_W_CLEAR, 0);
> + /* Abort any transmit and receive operations and put DMA
> + * in idle state.
> + */
> + abort_dma(pep);
> + /* Initialize address hash table */
> + err = init_hash_table(pep);
> + if (err)
> + return err;
> + /* SDMA configuration */
> + wrl(pep, SDMA_CONFIG, SDCR_BSZ8 | /* Burst size = 32 bytes */
> + SDCR_RIFB | /* Rx interrupt on frame */
> + SDCR_BLMT | /* Little endian transmit */
> + SDCR_BLMR | /* Little endian receive */
> + SDCR_RC_MAX_RETRANS); /* Max retransmit count */
> + /* Port Configuration */
> + wrl(pep, PORT_CONFIG, PCR_HS); /* Hash size is 1/2kb */
> + set_port_config_ext(pep, (pep->dev)->mtu);
> + return err;
> +}
> +
> +static int rxq_init(struct net_device *dev)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + struct rx_desc *p_rx_desc;
> + int size = 0, i = 0;
> + int rx_desc_num = pep->rx_ring_size;
> +
> + /* Allocate RX skb rings */
> + pep->rx_skb = kmalloc(sizeof(*pep->rx_skb) * pep->rx_ring_size,
> + GFP_KERNEL);
> + if (!pep->rx_skb) {
> + printk(KERN_ERR "%s: Cannot alloc RX skb ring\n", dev->name);
> + return -ENOMEM;
> + }
> + /* Allocate RX ring */
> + pep->rx_desc_count = 0;
> + size = pep->rx_ring_size * sizeof(struct rx_desc);
> + pep->rx_desc_area_size = size;
> + pep->p_rx_desc_area = dma_alloc_coherent(NULL, size,
> + &pep->rx_desc_dma, GFP_KERNEL);
> + if (!pep->p_rx_desc_area) {
> + printk(KERN_ERR "%s: Cannot alloc RX ring (size %d bytes)\n",
> + dev->name, size);
> + goto out;
> + }
> + memset((void *)pep->p_rx_desc_area, 0, size);
> + /* initialize the next_desc_ptr links in the Rx descriptors ring */
> + p_rx_desc = (struct rx_desc *)pep->p_rx_desc_area;
> + for (i = 0; i < rx_desc_num; i++) {
> + p_rx_desc[i].next_desc_ptr = pep->rx_desc_dma +
> + ((i + 1) % rx_desc_num) * sizeof(struct rx_desc);
> + }
> + /* Save Rx desc pointer to driver struct. */
> + pep->rx_curr_desc_q = 0;
> + pep->rx_used_desc_q = 0;
> + pep->rx_desc_area_size = rx_desc_num * sizeof(struct rx_desc);
> + return 0;
> +out:
> + kfree(pep->rx_skb);
> + return -ENOMEM;
> +}
> +
> +static void rxq_deinit(struct net_device *dev)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + int curr;
> +
> + /* Free preallocated skb's on RX rings */
> + for (curr = 0; pep->rx_desc_count && curr < pep->rx_ring_size;
> curr++) {
> + if (pep->rx_skb[curr]) {
> + dev_kfree_skb(pep->rx_skb[curr]);
> + pep->rx_desc_count--;
> + }
> + }
> + if (pep->rx_desc_count)
> + printk(KERN_ERR
> + "Error in freeing Rx Ring. %d skb's still\n",
> + pep->rx_desc_count);
> + /* Free RX ring */
> + if (pep->p_rx_desc_area)
> + dma_free_coherent(NULL, pep->rx_desc_area_size,
> + pep->p_rx_desc_area, pep->rx_desc_dma);
> + kfree(pep->rx_skb);
> +}
> +
> +static int txq_init(struct net_device *dev)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + struct tx_desc *p_tx_desc;
> + int size = 0, i = 0;
> + int tx_desc_num = pep->tx_ring_size;
> +
> + pep->tx_skb = kmalloc(sizeof(*pep->tx_skb) * pep->tx_ring_size,
> + GFP_KERNEL);
> + if (!pep->tx_skb) {
> + printk(KERN_ERR "%s: Cannot alloc TX skb ring\n", dev->name);
> + return -ENOMEM;
> + }
> + /* Allocate TX ring */
> + pep->tx_desc_count = 0;
> + size = pep->tx_ring_size * sizeof(struct tx_desc);
> + pep->tx_desc_area_size = size;
> + pep->p_tx_desc_area = dma_alloc_coherent(NULL, size,
> + &pep->tx_desc_dma, GFP_KERNEL);
> + if (!pep->p_tx_desc_area) {
> + printk(KERN_ERR "%s: Cannot allocate Tx Ring (size %d
> bytes)\n",
> + dev->name, size);
> + goto out;
> + }
> + memset((void *)pep->p_tx_desc_area, 0, pep->tx_desc_area_size);
> + /* Initialize the next_desc_ptr links in the Tx descriptors ring */
> + p_tx_desc = (struct tx_desc *)pep->p_tx_desc_area;
> + for (i = 0; i < tx_desc_num; i++) {
> + p_tx_desc[i].next_desc_ptr = pep->tx_desc_dma +
> + ((i + 1) % tx_desc_num) * sizeof(struct tx_desc);
> + }
> + pep->tx_curr_desc_q = 0;
> + pep->tx_used_desc_q = 0;
> + pep->tx_desc_area_size = tx_desc_num * sizeof(struct tx_desc);
> + return 0;
> +out:
> + kfree(pep->tx_skb);
> + return -ENOMEM;
> +}
> +
> +static void txq_deinit(struct net_device *dev)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> +
> + /* Free outstanding skb's on TX ring */
> + txq_reclaim(dev, 1);
> + BUG_ON(pep->tx_used_desc_q != pep->tx_curr_desc_q);
> + /* Free TX ring */
> + if (pep->p_tx_desc_area)
> + dma_free_coherent(NULL, pep->tx_desc_area_size,
> + pep->p_tx_desc_area, pep->tx_desc_dma);
> + kfree(pep->tx_skb);
> +}
> +
> +static int pxa168_eth_open(struct net_device *dev)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + int err;
> +
> + err = request_irq(dev->irq, pxa168_eth_int_handler,
> + IRQF_DISABLED , dev->name, dev);
> + if (err) {
> + dev_printk(KERN_ERR, &dev->dev, "can't assign irq\n");
> + return -EAGAIN;
> + }
> + pep->rx_resource_err = 0;
> + err = rxq_init(dev);
> + if (err != 0)
> + goto out_free_irq;
> + err = txq_init(dev);
> + if (err != 0)
> + goto out_free_rx_skb;
> + pep->rx_used_desc_q = 0;
> + pep->rx_curr_desc_q = 0;
> +
> + /* Fill RX ring with skb's */
> + rxq_refill(dev);
> + pep->rx_used_desc_q = 0;
> + pep->rx_curr_desc_q = 0;
> + netif_carrier_off(dev);
> + eth_port_start(dev);
> + napi_enable(&pep->napi);
> + return 0;
> +out_free_rx_skb:
> + rxq_deinit(dev);
> +out_free_irq:
> + free_irq(dev->irq, dev);
> + return err;
> +}
> +
> +static int pxa168_eth_stop(struct net_device *dev)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + eth_port_reset(dev);
> +
> + /* Disable interrupts */
> + wrl(pep, INT_MASK, 0);
> + wrl(pep, INT_CAUSE, 0);
> + /* Write to ICR to clear interrupts. */
> + wrl(pep, INT_W_CLEAR, 0);
> + napi_disable(&pep->napi);
> + del_timer_sync(&pep->timeout);
> + netif_carrier_off(dev);
> + free_irq(dev->irq, dev);
> + rxq_deinit(dev);
> + txq_deinit(dev);
> + return 0;
> +}
> +
> +static int pxa168_eth_change_mtu(struct net_device *dev, int mtu)
> +{
> + if ((mtu > ETH_DATA_LEN) || (mtu < 64))
> + return -EINVAL;
> + dev->mtu = mtu;
> + return 0;
> +}
> +
> +static int eth_alloc_tx_desc_index(struct pxa168_eth_private *pep)
> +{
> + int tx_desc_curr;
> +
> + tx_desc_curr = pep->tx_curr_desc_q;
> + pep->tx_curr_desc_q = (tx_desc_curr + 1) % pep->tx_ring_size;
> + BUG_ON(pep->tx_curr_desc_q == pep->tx_used_desc_q);
> + pep->tx_desc_count++;
> + return tx_desc_curr;
> +}
> +
> +static void eth_tx_submit_descs_for_skb(struct pxa168_eth_private *pep,
> + struct sk_buff *skb)
> +{
> + int tx_index;
> + struct tx_desc *desc;
> + int length;
> +
> + tx_index = eth_alloc_tx_desc_index(pep);
> + desc = &pep->p_tx_desc_area[tx_index];
> + length = skb->len;
> + pep->tx_skb[tx_index] = skb;
> + desc->byte_cnt = length;
> + desc->buf_ptr = dma_map_single(NULL, skb->data, length,
> DMA_TO_DEVICE);
> + wmb();
> + desc->cmd_sts = BUF_OWNED_BY_DMA | TX_GEN_CRC | TX_FIRST_DESC |
> + TX_ZERO_PADDING | TX_LAST_DESC;
> + if (unlikely(!(pep->tx_desc_count % TX_DONE_INTERVAL)))
> + desc->cmd_sts |= TX_EN_INT;
> + wmb();
> + wrl(pep, SDMA_CMD, SDMA_CMD_TXDH | SDMA_CMD_ERD);
> +}
> +
> +static int pxa168_rx_poll(struct napi_struct *napi, int budget)
> +{
> + struct pxa168_eth_private *pep =
> + container_of(napi, struct pxa168_eth_private, napi);
> + struct net_device *dev = pep->dev;
> + int work_done = 0;
> +
> + if (unlikely(pep->work_todo & WORK_LINK)) {
> + pep->work_todo &= ~(WORK_LINK);
> + handle_link_event(pep);
> + }
> + /*
> + * We call txq_reclaim every time since in NAPI interupts are
> disabled
> + * and due to this we miss the TX_DONE interrupt,which is not
> updated in
> + * interrupt status register.
> + */
> + txq_reclaim(dev, 0);
> + if (netif_queue_stopped(dev)
> + && pep->tx_ring_size - pep->tx_desc_count > 1) {
> + netif_wake_queue(dev);
> + }
> + work_done = rxq_process(dev, budget);
> + if (work_done < budget) {
> + napi_complete(napi);
> + wrl(pep, INT_MASK, ALL_INTS);
> + }
> + return work_done;
> +}
> +
> +static int pxa168_eth_start_xmit(struct sk_buff *skb, struct net_device
> *dev)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + struct net_device_stats *stats = &dev->stats;
> +
> + eth_tx_submit_descs_for_skb(pep, skb);
> + stats->tx_bytes += skb->len;
> + stats->tx_packets++;
> + dev->trans_start = jiffies;
> + if (pep->tx_ring_size - pep->tx_desc_count <= 1) {
> + /* We handled the current skb, but now we are out of space.*/
> + netif_stop_queue(dev);
> + }
> + return NETDEV_TX_OK;
> +}
> +
> +static int pxa168_smi_read(struct mii_bus *bus, int phy_addr, int regnum)
> +{
> + int val;
> + struct pxa168_eth_private *pep = bus->priv;
> + int i = 0;
> +
> + /* wait for the SMI register to become available */
> + for (i = 0; (val = rdl(pep, SMI)) & SMI_BUSY; i++) {
> + if (i == PHY_WAIT_ITERATIONS) {
> + printk(KERN_ERR
> + "pxa168 PHY timeout, val=0x%x\n", val);
> + return -ETIMEDOUT;
> + }
> + udelay(1);
> + }
> + wrl(pep, SMI, (phy_addr << 16) | (regnum << 21) | SMI_OP_R);
> + /* now wait for the data to be valid */
> + for (i = 0; !((val = rdl(pep, SMI)) & SMI_R_VALID); i++) {
> + if (i == PHY_WAIT_ITERATIONS) {
> + printk(KERN_ERR
> + "pxa168 PHY RD timeout, val=0x%x\n", val);
> + return -ETIMEDOUT;
> + }
> + udelay(1);
> + }
> + return val & 0xffff;
> +}
> +
> +static int pxa168_smi_write(struct mii_bus *bus, int phy_addr, int regnum,
> + u16 value)
> +{
> + struct pxa168_eth_private *pep = bus->priv;
> + int i;
> +
> + /* wait for the SMI register to become available */
> + for (i = 0; rdl(pep, SMI) & SMI_BUSY; i++) {
> + if (i == PHY_WAIT_ITERATIONS) {
> + printk(KERN_ERR "pxa168 PHY busy timeout.\n");
> + return -ETIMEDOUT;
> + }
> + udelay(1);
> + }
> + wrl(pep, SMI, (phy_addr << 16) | (regnum << 21) |
> + SMI_OP_W | (value & 0xffff));
> +
> + return 0;
> +}
> +
> +static int pxa168_eth_do_ioctl(struct net_device *dev, struct ifreq *ifr,
> + int cmd)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + if (pep->phy != NULL)
> + return phy_mii_ioctl(pep->phy, if_mii(ifr), cmd);
> + return -EOPNOTSUPP;
> +}
> +
> +static struct phy_device *phy_scan(struct pxa168_eth_private *pep, int
> phy_addr)
> +{
> + struct mii_bus *bus = pep->smi_bus;
> + struct phy_device *phydev;
> + int start;
> + int num;
> + int i;
> +
> + if (phy_addr == PXA168_ETH_PHY_ADDR_DEFAULT) {
> + /* Scan entire range */
> + start = ethernet_phy_get(pep);
> + num = 32;
> + } else {
> + /* Use phy addr specific to platform */
> + start = phy_addr & 0x1f;
> + num = 1;
> + }
> + phydev = NULL;
> + for (i = 0; i < num; i++) {
> + int addr = (start + i) & 0x1f;
> + if (bus->phy_map[addr] == NULL)
> + mdiobus_scan(bus, addr);
> +
> + if (phydev == NULL) {
> + phydev = bus->phy_map[addr];
> + if (phydev != NULL)
> + ethernet_phy_set_addr(pep, addr);
> + }
> + }
> +
> + return phydev;
> +}
> +
> +static void phy_init(struct pxa168_eth_private *pep, int speed, int
> duplex)
> +{
> + struct phy_device *phy = pep->phy;
> + ethernet_phy_reset(pep);
> +
> + phy_attach(pep->dev, dev_name(&phy->dev), 0,
> PHY_INTERFACE_MODE_MII);
> +
> + if (speed == 0) {
> + phy->autoneg = AUTONEG_ENABLE;
> + phy->speed = 0;
> + phy->duplex = 0;
> + phy->supported &= PHY_BASIC_FEATURES;
> + phy->advertising = phy->supported | ADVERTISED_Autoneg;
> + } else {
> + phy->autoneg = AUTONEG_DISABLE;
> + phy->advertising = 0;
> + phy->speed = speed;
> + phy->duplex = duplex;
> + }
> + phy_start_aneg(phy);
> +}
> +
> +static int ethernet_phy_setup(struct net_device *dev)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> +
> + if (pep->pd != NULL) {
> + if (pep->pd->init)
> + pep->pd->init();
> + }
> + pep->phy = phy_scan(pep, pep->pd->phy_addr & 0x1f);
> + if (pep->phy != NULL)
> + phy_init(pep, pep->pd->speed, pep->pd->duplex);
> + update_hash_table_mac_address(pep, NULL, dev->dev_addr);
> + return 0;
> +}
> +
> +static int get_random_mac_addr(struct net_device *dev)
> +{
> + printk(KERN_INFO "%s:Using random mac address\n", dev->name);
> + random_ether_addr(dev->dev_addr);
> + return 0;
> +}
> +
> +static int pxa168_get_settings(struct net_device *dev, struct ethtool_cmd
> *cmd)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> + int err;
> +
> + err = phy_read_status(pep->phy);
> + if (err == 0)
> + err = phy_ethtool_gset(pep->phy, cmd);
> + return err;
> +}
> +
> +static int pxa168_set_settings(struct net_device *dev, struct ethtool_cmd
> *cmd)
> +{
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> +
> + return phy_ethtool_sset(pep->phy, cmd);
> +}
> +
> +static void pxa168_get_drvinfo(struct net_device *dev,
> + struct ethtool_drvinfo *info)
> +{
> + strncpy(info->driver, DRIVER_NAME, 32);
> + strncpy(info->version, DRIVER_VERSION, 32);
> + strncpy(info->fw_version, "N/A", 32);
> + strncpy(info->bus_info, "N/A", 32);
> +}
> +
> +static u32 pxa168_get_link(struct net_device *dev)
> +{
> + return !!netif_carrier_ok(dev);
> +}
> +
> +static const struct ethtool_ops pxa168_ethtool_ops = {
> + .get_settings = pxa168_get_settings,
> + .set_settings = pxa168_set_settings,
> + .get_drvinfo = pxa168_get_drvinfo,
> + .get_link = pxa168_get_link,
> +};
> +
> +static const struct net_device_ops pxa168_eth_netdev_ops = {
> + .ndo_open = pxa168_eth_open,
> + .ndo_stop = pxa168_eth_stop,
> + .ndo_start_xmit = pxa168_eth_start_xmit,
> + .ndo_set_rx_mode = pxa168_eth_set_rx_mode,
> + .ndo_set_mac_address = pxa168_eth_set_mac_address,
> + .ndo_validate_addr = eth_validate_addr,
> + .ndo_do_ioctl = pxa168_eth_do_ioctl,
> + .ndo_change_mtu = pxa168_eth_change_mtu,
> + .ndo_tx_timeout = pxa168_eth_tx_timeout,
> +};
> +
> +static int pxa168_eth_probe(struct platform_device *pdev)
> +{
> + struct pxa168_eth_private *pep = NULL;
> + struct net_device *dev = NULL;
> + struct resource *res;
> + struct clk *clk;
> + int err;
> +
> + printk(KERN_NOTICE "PXA168 10/100 Ethernet Driver\n");
> +
> + clk = clk_get(&pdev->dev, "MFUCLK");
> + if (IS_ERR(clk)) {
> + printk(KERN_ERR "fast Ethernet failed to get clock\n");
> + return -ENODEV;
> + }
> + clk_enable(clk);
> +
> + dev = alloc_etherdev(sizeof(struct pxa168_eth_private));
> + if (!dev) {
> + err = -ENOMEM;
> + goto out;
> + }
> +
> + platform_set_drvdata(pdev, dev);
> + pep = netdev_priv(dev);
> + pep->dev = dev;
> + pep->clk = clk;
> + res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
> + if (res == NULL) {
> + err = -ENODEV;
> + goto out;
> + }
> + pep->base = ioremap(res->start, res->end - res->start + 1);
> + if (pep->base == NULL) {
> + err = -ENOMEM;
> + goto out;
> + }
> + res = platform_get_resource(pdev, IORESOURCE_IRQ, 0);
> + BUG_ON(!res);
> + dev->irq = res->start;
> + dev->netdev_ops = &pxa168_eth_netdev_ops;
> + dev->watchdog_timeo = 8 * HZ;
> + dev->base_addr = 0;
> + SET_ETHTOOL_OPS(dev, &pxa168_ethtool_ops);
> +
> + INIT_WORK(&pep->tx_timeout_task, pxa168_eth_tx_timeout_task);
> +
> + pep->rx_ring_size = NUM_RX_DESCS;
> + pep->tx_ring_size = NUM_TX_DESCS;
> + get_random_mac_addr(dev);
> + pep->pd = pdev->dev.platform_data;
> + netif_napi_add(dev, &pep->napi, pxa168_rx_poll, pep->rx_ring_size);
> +
> + memset(&pep->timeout, 0, sizeof(struct timer_list));
> + init_timer(&pep->timeout);
> + pep->timeout.function = rxq_refill_timer_wrapper;
> + pep->timeout.data = (unsigned long)pep;
> +
> + pep->smi_bus = mdiobus_alloc();
> + if (pep->smi_bus == NULL) {
> + err = -ENOMEM;
> + goto out;
> + }
> + pep->smi_bus->priv = pep;
> + pep->smi_bus->name = "pxa168_eth smi";
> + pep->smi_bus->read = pxa168_smi_read;
> + pep->smi_bus->write = pxa168_smi_write;
> + snprintf(pep->smi_bus->id, MII_BUS_ID_SIZE, "%d", pdev->id);
> + pep->smi_bus->parent = &pdev->dev;
> + pep->smi_bus->phy_mask = 0xffffffff;
> + if (mdiobus_register(pep->smi_bus) < 0) {
> + err = -ENOMEM;
> + goto out;
> + }
> + pxa168_init_hw(pep);
> + err = ethernet_phy_setup(dev);
> + if (err)
> + goto out;
> + SET_NETDEV_DEV(dev, &pdev->dev);
> + err = register_netdev(dev);
> + if (err)
> + goto out;
> + return 0;
> +out:
> + if (pep->clk) {
> + clk_disable(pep->clk);
> + clk_put(pep->clk);
> + pep->clk = NULL;
> + }
> + if (pep->base) {
> + iounmap(pep->base);
> + pep->base = NULL;
> + }
> + if (dev)
> + free_netdev(dev);
> + return err;
> +}
> +
> +static int pxa168_eth_remove(struct platform_device *pdev)
> +{
> + struct net_device *dev = platform_get_drvdata(pdev);
> + struct pxa168_eth_private *pep = netdev_priv(dev);
> +
> + if (pep->htpr) {
> + dma_free_coherent(NULL, HASH_ADDR_TABLE_SIZE + 7,
> + pep->htpr, pep->htpr_dma);
> + pep->htpr = NULL;
> + }
> + if (pep->clk) {
> + clk_disable(pep->clk);
> + clk_put(pep->clk);
> + pep->clk = NULL;
> + }
> + if (pep->phy != NULL)
> + phy_detach(pep->phy);
> +
> + iounmap(pep->base);
> + pep->base = NULL;
> + unregister_netdev(dev);
> + flush_scheduled_work();
> + free_netdev(dev);
> + platform_set_drvdata(pdev, NULL);
> + return 0;
> +}
> +
> +static void pxa168_eth_shutdown(struct platform_device *pdev)
> +{
> + struct net_device *dev = platform_get_drvdata(pdev);
> + eth_port_reset(dev);
> +}
> +
> +#ifdef CONFIG_PM
> +static int pxa168_eth_resume(struct platform_device *pdev)
> +{
> + return -ENOSYS;
> +}
> +
> +static int pxa168_eth_suspend(struct platform_device *pdev, pm_message_t
> state)
> +{
> + return -ENOSYS;
> +}
> +
> +#else
> +#define pxa168_eth_resume NULL
> +#define pxa168_eth_suspend NULL
> +#endif
> +
> +static struct platform_driver pxa168_eth_driver = {
> + .probe = pxa168_eth_probe,
> + .remove = pxa168_eth_remove,
> + .shutdown = pxa168_eth_shutdown,
> + .resume = pxa168_eth_resume,
> + .suspend = pxa168_eth_suspend,
> + .driver = {
> + .name = DRIVER_NAME,
> + },
> +};
> +
> +static int __init pxa168_init_module(void)
> +{
> + return platform_driver_register(&pxa168_eth_driver);
> +}
> +
> +static void __exit pxa168_cleanup_module(void)
> +{
> + platform_driver_unregister(&pxa168_eth_driver);
> +}
> +
> +module_init(pxa168_init_module);
> +module_exit(pxa168_cleanup_module);
> +
> +MODULE_LICENSE("GPL");
> +MODULE_DESCRIPTION("Ethernet driver for Marvell PXA168");
> +MODULE_ALIAS("platform:pxa168_eth");
> diff --git a/include/linux/pxa168_eth.h b/include/linux/pxa168_eth.h
> new file mode 100644
> index 0000000..152981c
> --- /dev/null
> +++ b/include/linux/pxa168_eth.h
> @@ -0,0 +1,20 @@
> +/*
> + *pxa168 ethernet platform device data definition file.
> + */
> +#ifndef __LINUX_PXA168_ETH_H
> +#define __LINUX_PXA168_ETH_H
> +
> +struct pxa168_eth_platform_data {
> + int phy_addr;
> +
> + /* If speed is 0, then speed and duplex are autonegotiated. */
> + int speed; /* 0, SPEED_10, SPEED_100 */
> + int duplex; /* DUPLEX_HALF or DUPLEX_FULL */
> +
> + /* init callback is used for board specific initialization
> + * e.g on Aspenite its used to initialize the PHY transceiver.
> + */
> + int (*init)(void);
> +};
> +
> +#endif /* __LINUX_PXA168_ETH_H */
> --
> 1.5.3.3
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists