diff options
Diffstat (limited to 'target/linux/brcm63xx')
4 files changed, 1062 insertions, 2 deletions
diff --git a/target/linux/brcm63xx/patches-2.6.39/080-bcm6345_enet.patch b/target/linux/brcm63xx/patches-2.6.39/080-bcm6345_enet.patch new file mode 100644 index 0000000..ff9ddbd --- /dev/null +++ b/target/linux/brcm63xx/patches-2.6.39/080-bcm6345_enet.patch @@ -0,0 +1,530 @@ +--- a/arch/mips/include/asm/mach-bcm63xx/bcm63xx_regs.h ++++ b/arch/mips/include/asm/mach-bcm63xx/bcm63xx_regs.h +@@ -29,16 +29,15 @@ + CKCTL_6338_SAR_EN | \ + CKCTL_6338_SPI_EN) + +-#define CKCTL_6345_CPU_EN (1 << 0) +-#define CKCTL_6345_BUS_EN (1 << 1) +-#define CKCTL_6345_EBI_EN (1 << 2) +-#define CKCTL_6345_UART_EN (1 << 3) +-#define CKCTL_6345_ADSLPHY_EN (1 << 4) +-#define CKCTL_6345_ENET_EN (1 << 7) +-#define CKCTL_6345_USBH_EN (1 << 8) ++#define CKCTL_6345_CPU_EN (1 << 16) ++#define CKCTL_6345_BUS_EN (1 << 17) ++#define CKCTL_6345_EBI_EN (1 << 18) ++#define CKCTL_6345_UART_EN (1 << 19) ++#define CKCTL_6345_ADSLPHY_EN (1 << 20) ++#define CKCTL_6345_ENET_EN (1 << 23) ++#define CKCTL_6345_USBH_EN (1 << 24) + +-#define CKCTL_6345_ALL_SAFE_EN (CKCTL_6345_ENET_EN | \ +- CKCTL_6345_USBH_EN | \ ++#define CKCTL_6345_ALL_SAFE_EN (CKCTL_6345_USBH_EN | \ + CKCTL_6345_ADSLPHY_EN) + + #define CKCTL_6348_ADSLPHY_EN (1 << 0) +@@ -548,6 +547,39 @@ + #define ENETDMA_SRAM4_REG(x) (0x20c + (x) * 0x10) + + ++/* Broadcom 6345 ENET DMA definitions */ ++#define ENETDMA_6345_CHANCFG_REG(x) (0x00 + (x) * 0x40) ++#define ENETDMA_6345_CHANCFG_EN_SHIFT 0 ++#define ENETDMA_6345_CHANCFG_EN_MASK (1 << ENETDMA_6345_CHANCFG_EN_SHIFT) ++#define ENETDMA_6345_PKTHALT_SHIFT 1 ++#define ENETDMA_6345_PKTHALT_MASK (1 << ENETDMA_6345_PKTHALT_SHIFT) ++#define ENETDMA_6345_CHAINING_SHIFT 2 ++#define ENETDMA_6345_CHAINING_MASK (1 << ENETDMA_6345_CHAINING_SHIFT) ++#define ENETDMA_6345_WRAP_EN_SHIFT 3 ++#define ENETDMA_6345_WRAP_EN_MASK (1 << ENETDMA_6345_WRAP_EN_SHIFT) ++#define ENETDMA_6345_FLOWC_EN_SHIFT 4 ++#define ENETDMA_6345_FLOWC_EN_MASK (1 << ENETDMA_6345_FLOWC_EN_SHIFT) ++ ++#define ENETDMA_6345_MAXBURST_REG(x) (0x04 + (x) * 0x40) ++ ++#define ENETDMA_6345_RSTART_REG(x) (0x08 + (x) * 0x40) ++ ++#define ENETDMA_6345_LEN_REG(x) (0x0C + (x) * 0x40) ++ ++#define ENETDMA_6345_BSTAT_REG(x) (0x10 + (x) * 0x40) ++ ++#define ENETDMA_6345_IR_REG(x) (0x14 + (x) * 0x40) ++#define ENETDMA_6345_IR_BUFDONE_MASK (1 << 0) ++#define ENETDMA_6345_IR_PKTDONE_MASK (1 << 1) ++#define ENETDMA_6345_IR_NOTOWNER_MASK (1 << 2) ++ ++#define ENETDMA_6345_IRMASK_REG(x) (0x18 + (x) * 0x40) ++ ++#define ENETDMA_6345_FC_REG(x) (0x1C + (x) * 0x40) ++ ++#define ENETDMA_6345_BUFALLOC_REG(x) (0x20 + (x) * 0x40) ++ ++ + /************************************************************************* + * _REG relative to RSET_OHCI_PRIV + *************************************************************************/ +--- a/drivers/net/bcm63xx_enet.c ++++ b/drivers/net/bcm63xx_enet.c +@@ -31,6 +31,7 @@ + #include <linux/if_vlan.h> + + #include <bcm63xx_dev_enet.h> ++#include <bcm63xx_cpu.h> + #include "bcm63xx_enet.h" + + static char bcm_enet_driver_name[] = "bcm63xx_enet"; +@@ -177,6 +178,7 @@ static void bcm_enet_mdio_write_mii(stru + static int bcm_enet_refill_rx(struct net_device *dev) + { + struct bcm_enet_priv *priv; ++ unsigned int desc_shift = BCMCPU_IS_6345() ? DMADESC_6345_SHIFT : 0; + + priv = netdev_priv(dev); + +@@ -205,7 +207,7 @@ static int bcm_enet_refill_rx(struct net + len_stat = priv->rx_skb_size << DMADESC_LENGTH_SHIFT; + len_stat |= DMADESC_OWNER_MASK; + if (priv->rx_dirty_desc == priv->rx_ring_size - 1) { +- len_stat |= DMADESC_WRAP_MASK; ++ len_stat |= (DMADESC_WRAP_MASK >> desc_shift); + priv->rx_dirty_desc = 0; + } else { + priv->rx_dirty_desc++; +@@ -216,7 +218,10 @@ static int bcm_enet_refill_rx(struct net + priv->rx_desc_count++; + + /* tell dma engine we allocated one buffer */ +- enet_dma_writel(priv, 1, ENETDMA_BUFALLOC_REG(priv->rx_chan)); ++ if (!BCMCPU_IS_6345()) ++ enet_dma_writel(priv, 1, ENETDMA_BUFALLOC_REG(priv->rx_chan)); ++ else ++ enet_dma_writel(priv, 1, ENETDMA_6345_BUFALLOC_REG(priv->rx_chan)); + } + + /* If rx ring is still empty, set a timer to try allocating +@@ -254,6 +259,7 @@ static int bcm_enet_receive_queue(struct + struct bcm_enet_priv *priv; + struct device *kdev; + int processed; ++ unsigned int desc_shift = BCMCPU_IS_6345() ? DMADESC_6345_SHIFT : 0; + + priv = netdev_priv(dev); + kdev = &priv->pdev->dev; +@@ -292,7 +298,7 @@ static int bcm_enet_receive_queue(struct + + /* if the packet does not have start of packet _and_ + * end of packet flag set, then just recycle it */ +- if ((len_stat & DMADESC_ESOP_MASK) != DMADESC_ESOP_MASK) { ++ if ((len_stat & (DMADESC_ESOP_MASK >> desc_shift)) != (DMADESC_ESOP_MASK >> desc_shift)) { + dev->stats.rx_dropped++; + continue; + } +@@ -352,8 +358,15 @@ static int bcm_enet_receive_queue(struct + bcm_enet_refill_rx(dev); + + /* kick rx dma */ +- enet_dma_writel(priv, ENETDMA_CHANCFG_EN_MASK, +- ENETDMA_CHANCFG_REG(priv->rx_chan)); ++ if (!BCMCPU_IS_6345()) ++ enet_dma_writel(priv, ENETDMA_CHANCFG_EN_MASK, ++ ENETDMA_CHANCFG_REG(priv->rx_chan)); ++ else ++ enet_dma_writel(priv, ENETDMA_6345_CHANCFG_EN_MASK | ++ ENETDMA_6345_CHAINING_MASK | ++ ENETDMA_6345_WRAP_EN_MASK | ++ ENETDMA_6345_FLOWC_EN_MASK, ++ ENETDMA_6345_CHANCFG_REG(priv->rx_chan)); + } + + return processed; +@@ -428,10 +441,21 @@ static int bcm_enet_poll(struct napi_str + dev = priv->net_dev; + + /* ack interrupts */ +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IR_REG(priv->rx_chan)); +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IR_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IR_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IR_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IR_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IR_REG(priv->tx_chan)); ++ } + + /* reclaim sent skb */ + tx_work_done = bcm_enet_tx_reclaim(dev, 0); +@@ -450,10 +474,21 @@ static int bcm_enet_poll(struct napi_str + napi_complete(napi); + + /* restore rx/tx interrupt */ +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IRMASK_REG(priv->rx_chan)); +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IRMASK_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IRMASK_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IRMASK_REG(priv->tx_chan)); ++ } + + return rx_work_done; + } +@@ -496,8 +531,13 @@ static irqreturn_t bcm_enet_isr_dma(int + priv = netdev_priv(dev); + + /* mask rx/tx interrupts */ +- enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->rx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, 0, ENETDMA_6345_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_6345_IRMASK_REG(priv->tx_chan)); ++ } + + napi_schedule(&priv->napi); + +@@ -513,6 +553,7 @@ static int bcm_enet_start_xmit(struct sk + struct bcm_enet_desc *desc; + u32 len_stat; + int ret; ++ unsigned int desc_shift = BCMCPU_IS_6345() ? DMADESC_6345_SHIFT : 0; + + priv = netdev_priv(dev); + +@@ -538,14 +579,13 @@ static int bcm_enet_start_xmit(struct sk + DMA_TO_DEVICE); + + len_stat = (skb->len << DMADESC_LENGTH_SHIFT) & DMADESC_LENGTH_MASK; +- len_stat |= DMADESC_ESOP_MASK | +- DMADESC_APPEND_CRC | +- DMADESC_OWNER_MASK; ++ len_stat |= ((DMADESC_ESOP_MASK >> desc_shift) | ++ DMADESC_APPEND_CRC | DMADESC_OWNER_MASK); + + priv->tx_curr_desc++; + if (priv->tx_curr_desc == priv->tx_ring_size) { + priv->tx_curr_desc = 0; +- len_stat |= DMADESC_WRAP_MASK; ++ len_stat |= (DMADESC_WRAP_MASK >> desc_shift); + } + priv->tx_desc_count--; + +@@ -556,8 +596,15 @@ static int bcm_enet_start_xmit(struct sk + wmb(); + + /* kick tx dma */ +- enet_dma_writel(priv, ENETDMA_CHANCFG_EN_MASK, +- ENETDMA_CHANCFG_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) ++ enet_dma_writel(priv, ENETDMA_CHANCFG_EN_MASK, ++ ENETDMA_CHANCFG_REG(priv->tx_chan)); ++ else ++ enet_dma_writel(priv, ENETDMA_6345_CHANCFG_EN_MASK | ++ ENETDMA_6345_CHAINING_MASK | ++ ENETDMA_6345_WRAP_EN_MASK | ++ ENETDMA_6345_FLOWC_EN_MASK, ++ ENETDMA_6345_CHANCFG_REG(priv->tx_chan)); + + /* stop queue if no more desc available */ + if (!priv->tx_desc_count) +@@ -685,6 +732,9 @@ static void bcm_enet_set_flow(struct bcm + val &= ~ENET_RXCFG_ENFLOW_MASK; + enet_writel(priv, val, ENET_RXCFG_REG); + ++ if (BCMCPU_IS_6345()) ++ return; ++ + /* tx flow control (pause frame generation) */ + val = enet_dma_readl(priv, ENETDMA_CFG_REG); + if (tx_en) +@@ -832,8 +882,13 @@ static int bcm_enet_open(struct net_devi + + /* mask all interrupts and request them */ + enet_writel(priv, 0, ENET_IRMASK_REG); +- enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->rx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, 0, ENETDMA_6345_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_6345_IRMASK_REG(priv->tx_chan)); ++ } + + ret = request_irq(dev->irq, bcm_enet_isr_mac, 0, dev->name, dev); + if (ret) +@@ -912,8 +967,12 @@ static int bcm_enet_open(struct net_devi + priv->rx_curr_desc = 0; + + /* initialize flow control buffer allocation */ +- enet_dma_writel(priv, ENETDMA_BUFALLOC_FORCE_MASK | 0, +- ENETDMA_BUFALLOC_REG(priv->rx_chan)); ++ if (!BCMCPU_IS_6345()) ++ enet_dma_writel(priv, ENETDMA_BUFALLOC_FORCE_MASK | 0, ++ ENETDMA_BUFALLOC_REG(priv->rx_chan)); ++ else ++ enet_dma_writel(priv, ENETDMA_BUFALLOC_FORCE_MASK | 0, ++ ENETDMA_6345_BUFALLOC_REG(priv->rx_chan)); + + if (bcm_enet_refill_rx(dev)) { + dev_err(kdev, "cannot allocate rx skb queue\n"); +@@ -922,37 +981,62 @@ static int bcm_enet_open(struct net_devi + } + + /* write rx & tx ring addresses */ +- enet_dma_writel(priv, priv->rx_desc_dma, +- ENETDMA_RSTART_REG(priv->rx_chan)); +- enet_dma_writel(priv, priv->tx_desc_dma, +- ENETDMA_RSTART_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, priv->rx_desc_dma, ++ ENETDMA_RSTART_REG(priv->rx_chan)); ++ enet_dma_writel(priv, priv->tx_desc_dma, ++ ENETDMA_RSTART_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, priv->rx_desc_dma, ++ ENETDMA_6345_RSTART_REG(priv->rx_chan)); ++ enet_dma_writel(priv, priv->tx_desc_dma, ++ ENETDMA_6345_RSTART_REG(priv->tx_chan)); ++ } + + /* clear remaining state ram for rx & tx channel */ +- enet_dma_writel(priv, 0, ENETDMA_SRAM2_REG(priv->rx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_SRAM2_REG(priv->tx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_SRAM3_REG(priv->rx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_SRAM3_REG(priv->tx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_SRAM4_REG(priv->rx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_SRAM4_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, 0, ENETDMA_SRAM2_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_SRAM2_REG(priv->tx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_SRAM3_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_SRAM3_REG(priv->tx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_SRAM4_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_SRAM4_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, 0, ENETDMA_6345_FC_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_6345_FC_REG(priv->tx_chan)); ++ } + + /* set max rx/tx length */ + enet_writel(priv, priv->hw_mtu, ENET_RXMAXLEN_REG); + enet_writel(priv, priv->hw_mtu, ENET_TXMAXLEN_REG); + + /* set dma maximum burst len */ +- enet_dma_writel(priv, BCMENET_DMA_MAXBURST, +- ENETDMA_MAXBURST_REG(priv->rx_chan)); +- enet_dma_writel(priv, BCMENET_DMA_MAXBURST, +- ENETDMA_MAXBURST_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, BCMENET_DMA_MAXBURST, ++ ENETDMA_MAXBURST_REG(priv->rx_chan)); ++ enet_dma_writel(priv, BCMENET_DMA_MAXBURST, ++ ENETDMA_MAXBURST_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, BCMENET_DMA_MAXBURST, ++ ENETDMA_6345_MAXBURST_REG(priv->rx_chan)); ++ enet_dma_writel(priv, BCMENET_DMA_MAXBURST, ++ ENETDMA_6345_MAXBURST_REG(priv->tx_chan)); ++ } + + /* set correct transmit fifo watermark */ + enet_writel(priv, BCMENET_TX_FIFO_TRESH, ENET_TXWMARK_REG); + + /* set flow control low/high threshold to 1/3 / 2/3 */ +- val = priv->rx_ring_size / 3; +- enet_dma_writel(priv, val, ENETDMA_FLOWCL_REG(priv->rx_chan)); +- val = (priv->rx_ring_size * 2) / 3; +- enet_dma_writel(priv, val, ENETDMA_FLOWCH_REG(priv->rx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ val = priv->rx_ring_size / 3; ++ enet_dma_writel(priv, val, ENETDMA_FLOWCL_REG(priv->rx_chan)); ++ val = (priv->rx_ring_size * 2) / 3; ++ enet_dma_writel(priv, val, ENETDMA_FLOWCH_REG(priv->rx_chan)); ++ } else { ++ enet_dma_writel(priv, 5, ENETDMA_6345_FC_REG(priv->rx_chan)); ++ enet_dma_writel(priv, priv->rx_ring_size, ENETDMA_6345_LEN_REG(priv->rx_chan)); ++ enet_dma_writel(priv, priv->tx_ring_size, ENETDMA_6345_LEN_REG(priv->tx_chan)); ++ } + + /* all set, enable mac and interrupts, start dma engine and + * kick rx dma channel */ +@@ -960,27 +1044,58 @@ static int bcm_enet_open(struct net_devi + val = enet_readl(priv, ENET_CTL_REG); + val |= ENET_CTL_ENABLE_MASK; + enet_writel(priv, val, ENET_CTL_REG); +- enet_dma_writel(priv, ENETDMA_CFG_EN_MASK, ENETDMA_CFG_REG); +- enet_dma_writel(priv, ENETDMA_CHANCFG_EN_MASK, +- ENETDMA_CHANCFG_REG(priv->rx_chan)); ++ ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, ENETDMA_CFG_EN_MASK, ENETDMA_CFG_REG); ++ enet_dma_writel(priv, ENETDMA_CHANCFG_EN_MASK, ++ ENETDMA_CHANCFG_REG(priv->rx_chan)); ++ } else { ++ enet_dma_writel(priv, ENETDMA_6345_CHANCFG_EN_MASK | ++ ENETDMA_6345_CHAINING_MASK | ++ ENETDMA_6345_WRAP_EN_MASK | ++ ENETDMA_6345_FLOWC_EN_MASK, ++ ENETDMA_6345_CHANCFG_REG(priv->rx_chan)); ++ } + + /* watch "mib counters about to overflow" interrupt */ + enet_writel(priv, ENET_IR_MIB, ENET_IR_REG); + enet_writel(priv, ENET_IR_MIB, ENET_IRMASK_REG); + + /* watch "packet transferred" interrupt in rx and tx */ +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IR_REG(priv->rx_chan)); +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IR_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IR_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IR_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IR_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IR_REG(priv->tx_chan)); ++ } + + /* make sure we enable napi before rx interrupt */ + napi_enable(&priv->napi); + +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IRMASK_REG(priv->rx_chan)); +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IRMASK_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IRMASK_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IRMASK_REG(priv->tx_chan)); ++ } + + if (priv->has_phy) + phy_start(priv->phydev); +@@ -1060,13 +1175,19 @@ static void bcm_enet_disable_dma(struct + { + int limit; + +- enet_dma_writel(priv, 0, ENETDMA_CHANCFG_REG(chan)); ++ if (!BCMCPU_IS_6345()) ++ enet_dma_writel(priv, 0, ENETDMA_CHANCFG_REG(chan)); ++ else ++ enet_dma_writel(priv, 0, ENETDMA_6345_CHANCFG_REG(chan)); + + limit = 1000; + do { + u32 val; + +- val = enet_dma_readl(priv, ENETDMA_CHANCFG_REG(chan)); ++ if (!BCMCPU_IS_6345()) ++ val = enet_dma_readl(priv, ENETDMA_CHANCFG_REG(chan)); ++ else ++ val = enet_dma_readl(priv, ENETDMA_6345_CHANCFG_REG(chan)); + if (!(val & ENETDMA_CHANCFG_EN_MASK)) + break; + udelay(1); +@@ -1093,8 +1214,13 @@ static int bcm_enet_stop(struct net_devi + + /* mask all interrupts */ + enet_writel(priv, 0, ENET_IRMASK_REG); +- enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->rx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, 0, ENETDMA_6345_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_6345_IRMASK_REG(priv->tx_chan)); ++ } + + /* make sure no mib update is scheduled */ + cancel_work_sync(&priv->mib_update_task); +@@ -1622,6 +1748,7 @@ static int __devinit bcm_enet_probe(stru + const char *clk_name; + unsigned int iomem_size; + int i, ret; ++ unsigned int chan_offset = 0; + + /* stop if shared driver failed, assume driver->probe will be + * called in the same order we register devices (correct ?) */ +@@ -1661,10 +1788,13 @@ static int __devinit bcm_enet_probe(stru + priv->irq_tx = res_irq_tx->start; + priv->mac_id = pdev->id; + ++ if (BCMCPU_IS_6345()) ++ chan_offset = 1; ++ + /* get rx & tx dma channel id for this mac */ + if (priv->mac_id == 0) { +- priv->rx_chan = 0; +- priv->tx_chan = 1; ++ priv->rx_chan = 0 + chan_offset; ++ priv->tx_chan = 1 + chan_offset; + clk_name = "enet0"; + } else { + priv->rx_chan = 2; +--- a/drivers/net/bcm63xx_enet.h ++++ b/drivers/net/bcm63xx_enet.h +@@ -46,6 +46,9 @@ struct bcm_enet_desc { + #define DMADESC_ESOP_MASK (DMADESC_EOP_MASK | DMADESC_SOP_MASK) + #define DMADESC_WRAP_MASK (1 << 12) + ++/* Shift down for EOP, SOP and WRAP bits */ ++#define DMADESC_6345_SHIFT (3) ++ + #define DMADESC_UNDER_MASK (1 << 9) + #define DMADESC_APPEND_CRC (1 << 8) + #define DMADESC_OVSIZE_MASK (1 << 4) +--- a/arch/mips/bcm63xx/dev-enet.c ++++ b/arch/mips/bcm63xx/dev-enet.c +@@ -104,7 +104,7 @@ int __init bcm63xx_enet_register(int uni + if (unit > 1) + return -ENODEV; + +- if (unit == 1 && BCMCPU_IS_6338()) ++ if (unit == 1 && (BCMCPU_IS_6338() || BCMCPU_IS_6345())) + return -ENODEV; + + if (!shared_device_registered) { diff --git a/target/linux/brcm63xx/patches-3.0/080-bcm6345_enet.patch b/target/linux/brcm63xx/patches-3.0/080-bcm6345_enet.patch new file mode 100644 index 0000000..d18bb25 --- /dev/null +++ b/target/linux/brcm63xx/patches-3.0/080-bcm6345_enet.patch @@ -0,0 +1,530 @@ +--- a/arch/mips/include/asm/mach-bcm63xx/bcm63xx_regs.h ++++ b/arch/mips/include/asm/mach-bcm63xx/bcm63xx_regs.h +@@ -29,16 +29,15 @@ + CKCTL_6338_SAR_EN | \ + CKCTL_6338_SPI_EN) + +-#define CKCTL_6345_CPU_EN (1 << 0) +-#define CKCTL_6345_BUS_EN (1 << 1) +-#define CKCTL_6345_EBI_EN (1 << 2) +-#define CKCTL_6345_UART_EN (1 << 3) +-#define CKCTL_6345_ADSLPHY_EN (1 << 4) +-#define CKCTL_6345_ENET_EN (1 << 7) +-#define CKCTL_6345_USBH_EN (1 << 8) ++#define CKCTL_6345_CPU_EN (1 << 16) ++#define CKCTL_6345_BUS_EN (1 << 17) ++#define CKCTL_6345_EBI_EN (1 << 18) ++#define CKCTL_6345_UART_EN (1 << 19) ++#define CKCTL_6345_ADSLPHY_EN (1 << 20) ++#define CKCTL_6345_ENET_EN (1 << 23) ++#define CKCTL_6345_USBH_EN (1 << 24) + +-#define CKCTL_6345_ALL_SAFE_EN (CKCTL_6345_ENET_EN | \ +- CKCTL_6345_USBH_EN | \ ++#define CKCTL_6345_ALL_SAFE_EN (CKCTL_6345_USBH_EN | \ + CKCTL_6345_ADSLPHY_EN) + + #define CKCTL_6348_ADSLPHY_EN (1 << 0) +@@ -548,6 +547,39 @@ + #define ENETDMA_SRAM4_REG(x) (0x20c + (x) * 0x10) + + ++/* Broadcom 6345 ENET DMA definitions */ ++#define ENETDMA_6345_CHANCFG_REG(x) (0x00 + (x) * 0x40) ++#define ENETDMA_6345_CHANCFG_EN_SHIFT 0 ++#define ENETDMA_6345_CHANCFG_EN_MASK (1 << ENETDMA_6345_CHANCFG_EN_SHIFT) ++#define ENETDMA_6345_PKTHALT_SHIFT 1 ++#define ENETDMA_6345_PKTHALT_MASK (1 << ENETDMA_6345_PKTHALT_SHIFT) ++#define ENETDMA_6345_CHAINING_SHIFT 2 ++#define ENETDMA_6345_CHAINING_MASK (1 << ENETDMA_6345_CHAINING_SHIFT) ++#define ENETDMA_6345_WRAP_EN_SHIFT 3 ++#define ENETDMA_6345_WRAP_EN_MASK (1 << ENETDMA_6345_WRAP_EN_SHIFT) ++#define ENETDMA_6345_FLOWC_EN_SHIFT 4 ++#define ENETDMA_6345_FLOWC_EN_MASK (1 << ENETDMA_6345_FLOWC_EN_SHIFT) ++ ++#define ENETDMA_6345_MAXBURST_REG(x) (0x04 + (x) * 0x40) ++ ++#define ENETDMA_6345_RSTART_REG(x) (0x08 + (x) * 0x40) ++ ++#define ENETDMA_6345_LEN_REG(x) (0x0C + (x) * 0x40) ++ ++#define ENETDMA_6345_BSTAT_REG(x) (0x10 + (x) * 0x40) ++ ++#define ENETDMA_6345_IR_REG(x) (0x14 + (x) * 0x40) ++#define ENETDMA_6345_IR_BUFDONE_MASK (1 << 0) ++#define ENETDMA_6345_IR_PKTDONE_MASK (1 << 1) ++#define ENETDMA_6345_IR_NOTOWNER_MASK (1 << 2) ++ ++#define ENETDMA_6345_IRMASK_REG(x) (0x18 + (x) * 0x40) ++ ++#define ENETDMA_6345_FC_REG(x) (0x1C + (x) * 0x40) ++ ++#define ENETDMA_6345_BUFALLOC_REG(x) (0x20 + (x) * 0x40) ++ ++ + /************************************************************************* + * _REG relative to RSET_OHCI_PRIV + *************************************************************************/ +--- a/drivers/net/bcm63xx_enet.c ++++ b/drivers/net/bcm63xx_enet.c +@@ -31,6 +31,7 @@ + #include <linux/if_vlan.h> + + #include <bcm63xx_dev_enet.h> ++#include <bcm63xx_cpu.h> + #include "bcm63xx_enet.h" + + static char bcm_enet_driver_name[] = "bcm63xx_enet"; +@@ -177,6 +178,7 @@ static void bcm_enet_mdio_write_mii(stru + static int bcm_enet_refill_rx(struct net_device *dev) + { + struct bcm_enet_priv *priv; ++ unsigned int desc_shift = BCMCPU_IS_6345() ? DMADESC_6345_SHIFT : 0; + + priv = netdev_priv(dev); + +@@ -205,7 +207,7 @@ static int bcm_enet_refill_rx(struct net + len_stat = priv->rx_skb_size << DMADESC_LENGTH_SHIFT; + len_stat |= DMADESC_OWNER_MASK; + if (priv->rx_dirty_desc == priv->rx_ring_size - 1) { +- len_stat |= DMADESC_WRAP_MASK; ++ len_stat |= (DMADESC_WRAP_MASK >> desc_shift); + priv->rx_dirty_desc = 0; + } else { + priv->rx_dirty_desc++; +@@ -216,7 +218,10 @@ static int bcm_enet_refill_rx(struct net + priv->rx_desc_count++; + + /* tell dma engine we allocated one buffer */ +- enet_dma_writel(priv, 1, ENETDMA_BUFALLOC_REG(priv->rx_chan)); ++ if (!BCMCPU_IS_6345()) ++ enet_dma_writel(priv, 1, ENETDMA_BUFALLOC_REG(priv->rx_chan)); ++ else ++ enet_dma_writel(priv, 1, ENETDMA_6345_BUFALLOC_REG(priv->rx_chan)); + } + + /* If rx ring is still empty, set a timer to try allocating +@@ -254,6 +259,7 @@ static int bcm_enet_receive_queue(struct + struct bcm_enet_priv *priv; + struct device *kdev; + int processed; ++ unsigned int desc_shift = BCMCPU_IS_6345() ? DMADESC_6345_SHIFT : 0; + + priv = netdev_priv(dev); + kdev = &priv->pdev->dev; +@@ -292,7 +298,7 @@ static int bcm_enet_receive_queue(struct + + /* if the packet does not have start of packet _and_ + * end of packet flag set, then just recycle it */ +- if ((len_stat & DMADESC_ESOP_MASK) != DMADESC_ESOP_MASK) { ++ if ((len_stat & (DMADESC_ESOP_MASK >> desc_shift)) != (DMADESC_ESOP_MASK >> desc_shift)) { + dev->stats.rx_dropped++; + continue; + } +@@ -352,8 +358,15 @@ static int bcm_enet_receive_queue(struct + bcm_enet_refill_rx(dev); + + /* kick rx dma */ +- enet_dma_writel(priv, ENETDMA_CHANCFG_EN_MASK, +- ENETDMA_CHANCFG_REG(priv->rx_chan)); ++ if (!BCMCPU_IS_6345()) ++ enet_dma_writel(priv, ENETDMA_CHANCFG_EN_MASK, ++ ENETDMA_CHANCFG_REG(priv->rx_chan)); ++ else ++ enet_dma_writel(priv, ENETDMA_6345_CHANCFG_EN_MASK | ++ ENETDMA_6345_CHAINING_MASK | ++ ENETDMA_6345_WRAP_EN_MASK | ++ ENETDMA_6345_FLOWC_EN_MASK, ++ ENETDMA_6345_CHANCFG_REG(priv->rx_chan)); + } + + return processed; +@@ -428,10 +441,21 @@ static int bcm_enet_poll(struct napi_str + dev = priv->net_dev; + + /* ack interrupts */ +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IR_REG(priv->rx_chan)); +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IR_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IR_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IR_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IR_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IR_REG(priv->tx_chan)); ++ } + + /* reclaim sent skb */ + tx_work_done = bcm_enet_tx_reclaim(dev, 0); +@@ -450,10 +474,21 @@ static int bcm_enet_poll(struct napi_str + napi_complete(napi); + + /* restore rx/tx interrupt */ +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IRMASK_REG(priv->rx_chan)); +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IRMASK_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IRMASK_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IRMASK_REG(priv->tx_chan)); ++ } + + return rx_work_done; + } +@@ -496,8 +531,13 @@ static irqreturn_t bcm_enet_isr_dma(int + priv = netdev_priv(dev); + + /* mask rx/tx interrupts */ +- enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->rx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, 0, ENETDMA_6345_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_6345_IRMASK_REG(priv->tx_chan)); ++ } + + napi_schedule(&priv->napi); + +@@ -513,6 +553,7 @@ static int bcm_enet_start_xmit(struct sk + struct bcm_enet_desc *desc; + u32 len_stat; + int ret; ++ unsigned int desc_shift = BCMCPU_IS_6345() ? DMADESC_6345_SHIFT : 0; + + priv = netdev_priv(dev); + +@@ -538,14 +579,13 @@ static int bcm_enet_start_xmit(struct sk + DMA_TO_DEVICE); + + len_stat = (skb->len << DMADESC_LENGTH_SHIFT) & DMADESC_LENGTH_MASK; +- len_stat |= DMADESC_ESOP_MASK | +- DMADESC_APPEND_CRC | +- DMADESC_OWNER_MASK; ++ len_stat |= ((DMADESC_ESOP_MASK >> desc_shift) | ++ DMADESC_APPEND_CRC | DMADESC_OWNER_MASK); + + priv->tx_curr_desc++; + if (priv->tx_curr_desc == priv->tx_ring_size) { + priv->tx_curr_desc = 0; +- len_stat |= DMADESC_WRAP_MASK; ++ len_stat |= (DMADESC_WRAP_MASK >> desc_shift); + } + priv->tx_desc_count--; + +@@ -556,8 +596,15 @@ static int bcm_enet_start_xmit(struct sk + wmb(); + + /* kick tx dma */ +- enet_dma_writel(priv, ENETDMA_CHANCFG_EN_MASK, +- ENETDMA_CHANCFG_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) ++ enet_dma_writel(priv, ENETDMA_CHANCFG_EN_MASK, ++ ENETDMA_CHANCFG_REG(priv->tx_chan)); ++ else ++ enet_dma_writel(priv, ENETDMA_6345_CHANCFG_EN_MASK | ++ ENETDMA_6345_CHAINING_MASK | ++ ENETDMA_6345_WRAP_EN_MASK | ++ ENETDMA_6345_FLOWC_EN_MASK, ++ ENETDMA_6345_CHANCFG_REG(priv->tx_chan)); + + /* stop queue if no more desc available */ + if (!priv->tx_desc_count) +@@ -685,6 +732,9 @@ static void bcm_enet_set_flow(struct bcm + val &= ~ENET_RXCFG_ENFLOW_MASK; + enet_writel(priv, val, ENET_RXCFG_REG); + ++ if (BCMCPU_IS_6345()) ++ return; ++ + /* tx flow control (pause frame generation) */ + val = enet_dma_readl(priv, ENETDMA_CFG_REG); + if (tx_en) +@@ -832,8 +882,13 @@ static int bcm_enet_open(struct net_devi + + /* mask all interrupts and request them */ + enet_writel(priv, 0, ENET_IRMASK_REG); +- enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->rx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, 0, ENETDMA_6345_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_6345_IRMASK_REG(priv->tx_chan)); ++ } + + ret = request_irq(dev->irq, bcm_enet_isr_mac, 0, dev->name, dev); + if (ret) +@@ -912,8 +967,12 @@ static int bcm_enet_open(struct net_devi + priv->rx_curr_desc = 0; + + /* initialize flow control buffer allocation */ +- enet_dma_writel(priv, ENETDMA_BUFALLOC_FORCE_MASK | 0, +- ENETDMA_BUFALLOC_REG(priv->rx_chan)); ++ if (!BCMCPU_IS_6345()) ++ enet_dma_writel(priv, ENETDMA_BUFALLOC_FORCE_MASK | 0, ++ ENETDMA_BUFALLOC_REG(priv->rx_chan)); ++ else ++ enet_dma_writel(priv, ENETDMA_BUFALLOC_FORCE_MASK | 0, ++ ENETDMA_6345_BUFALLOC_REG(priv->rx_chan)); + + if (bcm_enet_refill_rx(dev)) { + dev_err(kdev, "cannot allocate rx skb queue\n"); +@@ -922,37 +981,62 @@ static int bcm_enet_open(struct net_devi + } + + /* write rx & tx ring addresses */ +- enet_dma_writel(priv, priv->rx_desc_dma, +- ENETDMA_RSTART_REG(priv->rx_chan)); +- enet_dma_writel(priv, priv->tx_desc_dma, +- ENETDMA_RSTART_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, priv->rx_desc_dma, ++ ENETDMA_RSTART_REG(priv->rx_chan)); ++ enet_dma_writel(priv, priv->tx_desc_dma, ++ ENETDMA_RSTART_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, priv->rx_desc_dma, ++ ENETDMA_6345_RSTART_REG(priv->rx_chan)); ++ enet_dma_writel(priv, priv->tx_desc_dma, ++ ENETDMA_6345_RSTART_REG(priv->tx_chan)); ++ } + + /* clear remaining state ram for rx & tx channel */ +- enet_dma_writel(priv, 0, ENETDMA_SRAM2_REG(priv->rx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_SRAM2_REG(priv->tx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_SRAM3_REG(priv->rx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_SRAM3_REG(priv->tx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_SRAM4_REG(priv->rx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_SRAM4_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, 0, ENETDMA_SRAM2_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_SRAM2_REG(priv->tx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_SRAM3_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_SRAM3_REG(priv->tx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_SRAM4_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_SRAM4_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, 0, ENETDMA_6345_FC_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_6345_FC_REG(priv->tx_chan)); ++ } + + /* set max rx/tx length */ + enet_writel(priv, priv->hw_mtu, ENET_RXMAXLEN_REG); + enet_writel(priv, priv->hw_mtu, ENET_TXMAXLEN_REG); + + /* set dma maximum burst len */ +- enet_dma_writel(priv, BCMENET_DMA_MAXBURST, +- ENETDMA_MAXBURST_REG(priv->rx_chan)); +- enet_dma_writel(priv, BCMENET_DMA_MAXBURST, +- ENETDMA_MAXBURST_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, BCMENET_DMA_MAXBURST, ++ ENETDMA_MAXBURST_REG(priv->rx_chan)); ++ enet_dma_writel(priv, BCMENET_DMA_MAXBURST, ++ ENETDMA_MAXBURST_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, BCMENET_DMA_MAXBURST, ++ ENETDMA_6345_MAXBURST_REG(priv->rx_chan)); ++ enet_dma_writel(priv, BCMENET_DMA_MAXBURST, ++ ENETDMA_6345_MAXBURST_REG(priv->tx_chan)); ++ } + + /* set correct transmit fifo watermark */ + enet_writel(priv, BCMENET_TX_FIFO_TRESH, ENET_TXWMARK_REG); + + /* set flow control low/high threshold to 1/3 / 2/3 */ +- val = priv->rx_ring_size / 3; +- enet_dma_writel(priv, val, ENETDMA_FLOWCL_REG(priv->rx_chan)); +- val = (priv->rx_ring_size * 2) / 3; +- enet_dma_writel(priv, val, ENETDMA_FLOWCH_REG(priv->rx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ val = priv->rx_ring_size / 3; ++ enet_dma_writel(priv, val, ENETDMA_FLOWCL_REG(priv->rx_chan)); ++ val = (priv->rx_ring_size * 2) / 3; ++ enet_dma_writel(priv, val, ENETDMA_FLOWCH_REG(priv->rx_chan)); ++ } else { ++ enet_dma_writel(priv, 5, ENETDMA_6345_FC_REG(priv->rx_chan)); ++ enet_dma_writel(priv, priv->rx_ring_size, ENETDMA_6345_LEN_REG(priv->rx_chan)); ++ enet_dma_writel(priv, priv->tx_ring_size, ENETDMA_6345_LEN_REG(priv->tx_chan)); ++ } + + /* all set, enable mac and interrupts, start dma engine and + * kick rx dma channel */ +@@ -960,27 +1044,58 @@ static int bcm_enet_open(struct net_devi + val = enet_readl(priv, ENET_CTL_REG); + val |= ENET_CTL_ENABLE_MASK; + enet_writel(priv, val, ENET_CTL_REG); +- enet_dma_writel(priv, ENETDMA_CFG_EN_MASK, ENETDMA_CFG_REG); +- enet_dma_writel(priv, ENETDMA_CHANCFG_EN_MASK, +- ENETDMA_CHANCFG_REG(priv->rx_chan)); ++ ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, ENETDMA_CFG_EN_MASK, ENETDMA_CFG_REG); ++ enet_dma_writel(priv, ENETDMA_CHANCFG_EN_MASK, ++ ENETDMA_CHANCFG_REG(priv->rx_chan)); ++ } else { ++ enet_dma_writel(priv, ENETDMA_6345_CHANCFG_EN_MASK | ++ ENETDMA_6345_CHAINING_MASK | ++ ENETDMA_6345_WRAP_EN_MASK | ++ ENETDMA_6345_FLOWC_EN_MASK, ++ ENETDMA_6345_CHANCFG_REG(priv->rx_chan)); ++ } + + /* watch "mib counters about to overflow" interrupt */ + enet_writel(priv, ENET_IR_MIB, ENET_IR_REG); + enet_writel(priv, ENET_IR_MIB, ENET_IRMASK_REG); + + /* watch "packet transferred" interrupt in rx and tx */ +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IR_REG(priv->rx_chan)); +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IR_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IR_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IR_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IR_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IR_REG(priv->tx_chan)); ++ } + + /* make sure we enable napi before rx interrupt */ + napi_enable(&priv->napi); + +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IRMASK_REG(priv->rx_chan)); +- enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, +- ENETDMA_IRMASK_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_PKTDONE_MASK, ++ ENETDMA_IRMASK_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, ENETDMA_IR_BUFDONE_MASK | ++ ENETDMA_IR_PKTDONE_MASK | ++ ENETDMA_IR_NOTOWNER_MASK, ++ ENETDMA_6345_IRMASK_REG(priv->tx_chan)); ++ } + + if (priv->has_phy) + phy_start(priv->phydev); +@@ -1060,13 +1175,19 @@ static void bcm_enet_disable_dma(struct + { + int limit; + +- enet_dma_writel(priv, 0, ENETDMA_CHANCFG_REG(chan)); ++ if (!BCMCPU_IS_6345()) ++ enet_dma_writel(priv, 0, ENETDMA_CHANCFG_REG(chan)); ++ else ++ enet_dma_writel(priv, 0, ENETDMA_6345_CHANCFG_REG(chan)); + + limit = 1000; + do { + u32 val; + +- val = enet_dma_readl(priv, ENETDMA_CHANCFG_REG(chan)); ++ if (!BCMCPU_IS_6345()) ++ val = enet_dma_readl(priv, ENETDMA_CHANCFG_REG(chan)); ++ else ++ val = enet_dma_readl(priv, ENETDMA_6345_CHANCFG_REG(chan)); + if (!(val & ENETDMA_CHANCFG_EN_MASK)) + break; + udelay(1); +@@ -1093,8 +1214,13 @@ static int bcm_enet_stop(struct net_devi + + /* mask all interrupts */ + enet_writel(priv, 0, ENET_IRMASK_REG); +- enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->rx_chan)); +- enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->tx_chan)); ++ if (!BCMCPU_IS_6345()) { ++ enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_IRMASK_REG(priv->tx_chan)); ++ } else { ++ enet_dma_writel(priv, 0, ENETDMA_6345_IRMASK_REG(priv->rx_chan)); ++ enet_dma_writel(priv, 0, ENETDMA_6345_IRMASK_REG(priv->tx_chan)); ++ } + + /* make sure no mib update is scheduled */ + cancel_work_sync(&priv->mib_update_task); +@@ -1623,6 +1749,7 @@ static int __devinit bcm_enet_probe(stru + const char *clk_name; + unsigned int iomem_size; + int i, ret; ++ unsigned int chan_offset = 0; + + /* stop if shared driver failed, assume driver->probe will be + * called in the same order we register devices (correct ?) */ +@@ -1662,10 +1789,13 @@ static int __devinit bcm_enet_probe(stru + priv->irq_tx = res_irq_tx->start; + priv->mac_id = pdev->id; + ++ if (BCMCPU_IS_6345()) ++ chan_offset = 1; ++ + /* get rx & tx dma channel id for this mac */ + if (priv->mac_id == 0) { +- priv->rx_chan = 0; +- priv->tx_chan = 1; ++ priv->rx_chan = 0 + chan_offset; ++ priv->tx_chan = 1 + chan_offset; + clk_name = "enet0"; + } else { + priv->rx_chan = 2; +--- a/drivers/net/bcm63xx_enet.h ++++ b/drivers/net/bcm63xx_enet.h +@@ -46,6 +46,9 @@ struct bcm_enet_desc { + #define DMADESC_ESOP_MASK (DMADESC_EOP_MASK | DMADESC_SOP_MASK) + #define DMADESC_WRAP_MASK (1 << 12) + ++/* Shift down for EOP, SOP and WRAP bits */ ++#define DMADESC_6345_SHIFT (3) ++ + #define DMADESC_UNDER_MASK (1 << 9) + #define DMADESC_APPEND_CRC (1 << 8) + #define DMADESC_OVSIZE_MASK (1 << 4) +--- a/arch/mips/bcm63xx/dev-enet.c ++++ b/arch/mips/bcm63xx/dev-enet.c +@@ -104,7 +104,7 @@ int __init bcm63xx_enet_register(int uni + if (unit > 1) + return -ENODEV; + +- if (unit == 1 && BCMCPU_IS_6338()) ++ if (unit == 1 && (BCMCPU_IS_6338() || BCMCPU_IS_6345())) + return -ENODEV; + + if (!shared_device_registered) { diff --git a/target/linux/brcm63xx/patches-3.0/230-6358-enet1-external-mii-clk.patch b/target/linux/brcm63xx/patches-3.0/230-6358-enet1-external-mii-clk.patch index 8aba413..9b242d2 100644 --- a/target/linux/brcm63xx/patches-3.0/230-6358-enet1-external-mii-clk.patch +++ b/target/linux/brcm63xx/patches-3.0/230-6358-enet1-external-mii-clk.patch @@ -11,7 +11,7 @@ bcm_gpio_writel(val, GPIO_MODE_REG); --- a/arch/mips/include/asm/mach-bcm63xx/bcm63xx_regs.h +++ b/arch/mips/include/asm/mach-bcm63xx/bcm63xx_regs.h -@@ -402,6 +402,8 @@ +@@ -401,6 +401,8 @@ #define GPIO_MODE_6358_EXTRA_SPI_SS (1 << 7) #define GPIO_MODE_6358_SERIAL_LED (1 << 10) #define GPIO_MODE_6358_UTOPIA (1 << 12) diff --git a/target/linux/brcm63xx/patches-3.0/240-spi.patch b/target/linux/brcm63xx/patches-3.0/240-spi.patch index 73d1048..4c5cfa6 100644 --- a/target/linux/brcm63xx/patches-3.0/240-spi.patch +++ b/target/linux/brcm63xx/patches-3.0/240-spi.patch @@ -170,7 +170,7 @@ #define BCM_6358_OHCI0_IRQ (IRQ_INTERNAL_BASE + 5) --- a/arch/mips/include/asm/mach-bcm63xx/bcm63xx_regs.h +++ b/arch/mips/include/asm/mach-bcm63xx/bcm63xx_regs.h -@@ -773,4 +773,116 @@ +@@ -805,4 +805,116 @@ #define DMIPSPLLCFG_N2_SHIFT 29 #define DMIPSPLLCFG_N2_MASK (0x7 << DMIPSPLLCFG_N2_SHIFT) |