v1.5 branch refresh based upon upstream master @ c8677ca89e53e3be7988d54280fce166cc894a7e
[librecmc/librecmc.git] / target / linux / ar71xx / files / drivers / net / ethernet / atheros / ag71xx / ag71xx_main.c
index 566e9513d8b7c6ef101902ae7d281dcc1c233893..54ec8e67b36cd917f025975387bfdcf102597144 100644 (file)
 
 #include "ag71xx.h"
 
-#if LINUX_VERSION_CODE < KERNEL_VERSION(4,2,0)
-static inline void skb_free_frag(void *data)
-{
-       put_page(virt_to_head_page(data));
-}
-#endif
-
 #define AG71XX_DEFAULT_MSG_ENABLE      \
        (NETIF_MSG_DRV                  \
        | NETIF_MSG_PROBE               \
@@ -97,41 +90,6 @@ static inline void ag71xx_dump_intr(struct ag71xx *ag, char *label, u32 intr)
                (intr & AG71XX_INT_RX_BE) ? "RXBE " : "");
 }
 
-static void ag71xx_ring_free(struct ag71xx_ring *ring)
-{
-       int ring_size = BIT(ring->order);
-       kfree(ring->buf);
-
-       if (ring->descs_cpu)
-               dma_free_coherent(NULL, ring_size * AG71XX_DESC_SIZE,
-                                 ring->descs_cpu, ring->descs_dma);
-}
-
-static int ag71xx_ring_alloc(struct ag71xx_ring *ring)
-{
-       int ring_size = BIT(ring->order);
-       int err;
-
-       ring->descs_cpu = dma_alloc_coherent(NULL, ring_size * AG71XX_DESC_SIZE,
-                                            &ring->descs_dma, GFP_ATOMIC);
-       if (!ring->descs_cpu) {
-               err = -ENOMEM;
-               goto err;
-       }
-
-
-       ring->buf = kzalloc(ring_size * sizeof(*ring->buf), GFP_KERNEL);
-       if (!ring->buf) {
-               err = -ENOMEM;
-               goto err;
-       }
-
-       return 0;
-
-err:
-       return err;
-}
-
 static void ag71xx_ring_tx_clean(struct ag71xx *ag)
 {
        struct ag71xx_ring *ring = &ag->tx_ring;
@@ -322,30 +280,56 @@ static int ag71xx_ring_rx_refill(struct ag71xx *ag)
 
 static int ag71xx_rings_init(struct ag71xx *ag)
 {
-       int ret;
+       struct ag71xx_ring *tx = &ag->tx_ring;
+       struct ag71xx_ring *rx = &ag->rx_ring;
+       int ring_size = BIT(tx->order) + BIT(rx->order);
+       int tx_size = BIT(tx->order);
+
+       tx->buf = kzalloc(ring_size * sizeof(*tx->buf), GFP_KERNEL);
+       if (!tx->buf)
+               return -ENOMEM;
+
+       tx->descs_cpu = dma_alloc_coherent(NULL, ring_size * AG71XX_DESC_SIZE,
+                                          &tx->descs_dma, GFP_ATOMIC);
+       if (!tx->descs_cpu) {
+               kfree(tx->buf);
+               tx->buf = NULL;
+               return -ENOMEM;
+       }
 
-       ret = ag71xx_ring_alloc(&ag->tx_ring);
-       if (ret)
-               return ret;
+       rx->buf = &tx->buf[BIT(tx->order)];
+       rx->descs_cpu = ((void *)tx->descs_cpu) + tx_size * AG71XX_DESC_SIZE;
+       rx->descs_dma = tx->descs_dma + tx_size * AG71XX_DESC_SIZE;
 
        ag71xx_ring_tx_init(ag);
+       return ag71xx_ring_rx_init(ag);
+}
 
-       ret = ag71xx_ring_alloc(&ag->rx_ring);
-       if (ret)
-               return ret;
+static void ag71xx_rings_free(struct ag71xx *ag)
+{
+       struct ag71xx_ring *tx = &ag->tx_ring;
+       struct ag71xx_ring *rx = &ag->rx_ring;
+       int ring_size = BIT(tx->order) + BIT(rx->order);
 
-       ret = ag71xx_ring_rx_init(ag);
-       return ret;
+       if (tx->descs_cpu)
+               dma_free_coherent(NULL, ring_size * AG71XX_DESC_SIZE,
+                                 tx->descs_cpu, tx->descs_dma);
+
+       kfree(tx->buf);
+
+       tx->descs_cpu = NULL;
+       rx->descs_cpu = NULL;
+       tx->buf = NULL;
+       rx->buf = NULL;
 }
 
 static void ag71xx_rings_cleanup(struct ag71xx *ag)
 {
        ag71xx_ring_rx_clean(ag);
-       ag71xx_ring_free(&ag->rx_ring);
-
        ag71xx_ring_tx_clean(ag);
+       ag71xx_rings_free(ag);
+
        netdev_reset_queue(ag->dev);
-       ag71xx_ring_free(&ag->tx_ring);
 }
 
 static unsigned char *ag71xx_speed_str(struct ag71xx *ag)
@@ -469,8 +453,8 @@ static void ag71xx_hw_setup(struct ag71xx *ag)
        /* setup FIFO configuration registers */
        ag71xx_wr(ag, AG71XX_REG_FIFO_CFG0, FIFO_CFG0_INIT);
        if (pdata->is_ar724x) {
-               ag71xx_wr(ag, AG71XX_REG_FIFO_CFG1, pdata->fifo_cfg1);
-               ag71xx_wr(ag, AG71XX_REG_FIFO_CFG2, pdata->fifo_cfg2);
+               ag71xx_wr(ag, AG71XX_REG_FIFO_CFG1, 0x0010ffff);
+               ag71xx_wr(ag, AG71XX_REG_FIFO_CFG2, 0x015500aa);
        } else {
                ag71xx_wr(ag, AG71XX_REG_FIFO_CFG1, 0x0fff0000);
                ag71xx_wr(ag, AG71XX_REG_FIFO_CFG2, 0x00001fff);
@@ -612,7 +596,7 @@ __ag71xx_link_adjust(struct ag71xx *ag, bool update)
        if (pdata->is_ar91xx)
                fifo3 = 0x00780fff;
        else if (pdata->is_ar724x)
-               fifo3 = pdata->fifo_cfg3;
+               fifo3 = 0x01f00140;
        else
                fifo3 = 0x008001ff;
 
@@ -629,6 +613,22 @@ __ag71xx_link_adjust(struct ag71xx *ag, bool update)
        ag71xx_wr(ag, AG71XX_REG_MAC_CFG2, cfg2);
        ag71xx_wr(ag, AG71XX_REG_FIFO_CFG5, fifo5);
        ag71xx_wr(ag, AG71XX_REG_MAC_IFCTL, ifctl);
+
+       if (pdata->disable_inline_checksum_engine) {
+               /*
+                * The rx ring buffer can stall on small packets on QCA953x and
+                * QCA956x. Disabling the inline checksum engine fixes the stall.
+                * The wr, rr functions cannot be used since this hidden register
+                * is outside of the normal ag71xx register block.
+                */
+               void __iomem *dam = ioremap_nocache(0xb90001bc, 0x4);
+               if (dam) {
+                       __raw_writel(__raw_readl(dam) & ~BIT(27), dam);
+                       (void)__raw_readl(dam);
+                       iounmap(dam);
+               }
+       }
+
        ag71xx_hw_start(ag);
 
        netif_carrier_on(ag->dev);
@@ -816,7 +816,6 @@ static netdev_tx_t ag71xx_hard_start_xmit(struct sk_buff *skb,
        i = (ring->curr + n - 1) & ring_mask;
        ring->buf[i].len = skb->len;
        ring->buf[i].skb = skb;
-       ring->buf[i].timestamp = jiffies;
 
        netdev_sent_queue(dev, skb->len);
 
@@ -926,10 +925,12 @@ static void ag71xx_restart_work_func(struct work_struct *work)
        rtnl_unlock();
 }
 
-static bool ag71xx_check_dma_stuck(struct ag71xx *ag, unsigned long timestamp)
+static bool ag71xx_check_dma_stuck(struct ag71xx *ag)
 {
+       unsigned long timestamp;
        u32 rx_sm, tx_sm, rx_fd;
 
+       timestamp = netdev_get_tx_queue(ag->dev, 0)->trans_start;
        if (likely(time_before(jiffies, timestamp + HZ/10)))
                return false;
 
@@ -969,7 +970,7 @@ static int ag71xx_tx_packets(struct ag71xx *ag, bool flush)
 
                if (!flush && !ag71xx_desc_empty(desc)) {
                        if (pdata->is_ar724x &&
-                           ag71xx_check_dma_stuck(ag, ring->buf[i].timestamp)) {
+                           ag71xx_check_dma_stuck(ag)) {
                                schedule_delayed_work(&ag->restart_work, HZ / 2);
                                dma_stuck = true;
                        }
@@ -999,12 +1000,12 @@ static int ag71xx_tx_packets(struct ag71xx *ag, bool flush)
 
        DBG("%s: %d packets sent out\n", ag->dev->name, sent);
 
-       ag->dev->stats.tx_bytes += bytes_compl;
-       ag->dev->stats.tx_packets += sent;
-
        if (!sent)
                return 0;
 
+       ag->dev->stats.tx_bytes += bytes_compl;
+       ag->dev->stats.tx_packets += sent;
+
        netdev_completed_queue(ag->dev, sent, bytes_compl);
        if ((ring->curr - ring->dirty) < (ring_size * 3) / 4)
                netif_wake_queue(ag->dev);