Linux-libre 5.3.12-gnu
[librecmc/linux-libre.git] / drivers / net / ethernet / ibm / emac / core.c
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * drivers/net/ethernet/ibm/emac/core.c
4  *
5  * Driver for PowerPC 4xx on-chip ethernet controller.
6  *
7  * Copyright 2007 Benjamin Herrenschmidt, IBM Corp.
8  *                <benh@kernel.crashing.org>
9  *
10  * Based on the arch/ppc version of the driver:
11  *
12  * Copyright (c) 2004, 2005 Zultys Technologies.
13  * Eugene Surovegin <eugene.surovegin@zultys.com> or <ebs@ebshome.net>
14  *
15  * Based on original work by
16  *      Matt Porter <mporter@kernel.crashing.org>
17  *      (c) 2003 Benjamin Herrenschmidt <benh@kernel.crashing.org>
18  *      Armin Kuster <akuster@mvista.com>
19  *      Johnnie Peters <jpeters@mvista.com>
20  */
21
22 #include <linux/module.h>
23 #include <linux/sched.h>
24 #include <linux/string.h>
25 #include <linux/errno.h>
26 #include <linux/delay.h>
27 #include <linux/types.h>
28 #include <linux/pci.h>
29 #include <linux/etherdevice.h>
30 #include <linux/skbuff.h>
31 #include <linux/crc32.h>
32 #include <linux/ethtool.h>
33 #include <linux/mii.h>
34 #include <linux/bitops.h>
35 #include <linux/workqueue.h>
36 #include <linux/of.h>
37 #include <linux/of_address.h>
38 #include <linux/of_irq.h>
39 #include <linux/of_net.h>
40 #include <linux/of_mdio.h>
41 #include <linux/slab.h>
42
43 #include <asm/processor.h>
44 #include <asm/io.h>
45 #include <asm/dma.h>
46 #include <linux/uaccess.h>
47 #include <asm/dcr.h>
48 #include <asm/dcr-regs.h>
49
50 #include "core.h"
51
52 /*
53  * Lack of dma_unmap_???? calls is intentional.
54  *
55  * API-correct usage requires additional support state information to be
56  * maintained for every RX and TX buffer descriptor (BD). Unfortunately, due to
57  * EMAC design (e.g. TX buffer passed from network stack can be split into
58  * several BDs, dma_map_single/dma_map_page can be used to map particular BD),
59  * maintaining such information will add additional overhead.
60  * Current DMA API implementation for 4xx processors only ensures cache coherency
61  * and dma_unmap_???? routines are empty and are likely to stay this way.
62  * I decided to omit dma_unmap_??? calls because I don't want to add additional
63  * complexity just for the sake of following some abstract API, when it doesn't
64  * add any real benefit to the driver. I understand that this decision maybe
65  * controversial, but I really tried to make code API-correct and efficient
66  * at the same time and didn't come up with code I liked :(.                --ebs
67  */
68
69 #define DRV_NAME        "emac"
70 #define DRV_VERSION     "3.54"
71 #define DRV_DESC        "PPC 4xx OCP EMAC driver"
72
73 MODULE_DESCRIPTION(DRV_DESC);
74 MODULE_AUTHOR
75     ("Eugene Surovegin <eugene.surovegin@zultys.com> or <ebs@ebshome.net>");
76 MODULE_LICENSE("GPL");
77
78 /* minimum number of free TX descriptors required to wake up TX process */
79 #define EMAC_TX_WAKEUP_THRESH           (NUM_TX_BUFF / 4)
80
81 /* If packet size is less than this number, we allocate small skb and copy packet
82  * contents into it instead of just sending original big skb up
83  */
84 #define EMAC_RX_COPY_THRESH             CONFIG_IBM_EMAC_RX_COPY_THRESHOLD
85
86 /* Since multiple EMACs share MDIO lines in various ways, we need
87  * to avoid re-using the same PHY ID in cases where the arch didn't
88  * setup precise phy_map entries
89  *
90  * XXX This is something that needs to be reworked as we can have multiple
91  * EMAC "sets" (multiple ASICs containing several EMACs) though we can
92  * probably require in that case to have explicit PHY IDs in the device-tree
93  */
94 static u32 busy_phy_map;
95 static DEFINE_MUTEX(emac_phy_map_lock);
96
97 /* This is the wait queue used to wait on any event related to probe, that
98  * is discovery of MALs, other EMACs, ZMII/RGMIIs, etc...
99  */
100 static DECLARE_WAIT_QUEUE_HEAD(emac_probe_wait);
101
102 /* Having stable interface names is a doomed idea. However, it would be nice
103  * if we didn't have completely random interface names at boot too :-) It's
104  * just a matter of making everybody's life easier. Since we are doing
105  * threaded probing, it's a bit harder though. The base idea here is that
106  * we make up a list of all emacs in the device-tree before we register the
107  * driver. Every emac will then wait for the previous one in the list to
108  * initialize before itself. We should also keep that list ordered by
109  * cell_index.
110  * That list is only 4 entries long, meaning that additional EMACs don't
111  * get ordering guarantees unless EMAC_BOOT_LIST_SIZE is increased.
112  */
113
114 #define EMAC_BOOT_LIST_SIZE     4
115 static struct device_node *emac_boot_list[EMAC_BOOT_LIST_SIZE];
116
117 /* How long should I wait for dependent devices ? */
118 #define EMAC_PROBE_DEP_TIMEOUT  (HZ * 5)
119
120 /* I don't want to litter system log with timeout errors
121  * when we have brain-damaged PHY.
122  */
123 static inline void emac_report_timeout_error(struct emac_instance *dev,
124                                              const char *error)
125 {
126         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX |
127                                   EMAC_FTR_460EX_PHY_CLK_FIX |
128                                   EMAC_FTR_440EP_PHY_CLK_FIX))
129                 DBG(dev, "%s" NL, error);
130         else if (net_ratelimit())
131                 printk(KERN_ERR "%pOF: %s\n", dev->ofdev->dev.of_node, error);
132 }
133
134 /* EMAC PHY clock workaround:
135  * 440EP/440GR has more sane SDR0_MFR register implementation than 440GX,
136  * which allows controlling each EMAC clock
137  */
138 static inline void emac_rx_clk_tx(struct emac_instance *dev)
139 {
140 #ifdef CONFIG_PPC_DCR_NATIVE
141         if (emac_has_feature(dev, EMAC_FTR_440EP_PHY_CLK_FIX))
142                 dcri_clrset(SDR0, SDR0_MFR,
143                             0, SDR0_MFR_ECS >> dev->cell_index);
144 #endif
145 }
146
147 static inline void emac_rx_clk_default(struct emac_instance *dev)
148 {
149 #ifdef CONFIG_PPC_DCR_NATIVE
150         if (emac_has_feature(dev, EMAC_FTR_440EP_PHY_CLK_FIX))
151                 dcri_clrset(SDR0, SDR0_MFR,
152                             SDR0_MFR_ECS >> dev->cell_index, 0);
153 #endif
154 }
155
156 /* PHY polling intervals */
157 #define PHY_POLL_LINK_ON        HZ
158 #define PHY_POLL_LINK_OFF       (HZ / 5)
159
160 /* Graceful stop timeouts in us.
161  * We should allow up to 1 frame time (full-duplex, ignoring collisions)
162  */
163 #define STOP_TIMEOUT_10         1230
164 #define STOP_TIMEOUT_100        124
165 #define STOP_TIMEOUT_1000       13
166 #define STOP_TIMEOUT_1000_JUMBO 73
167
168 static unsigned char default_mcast_addr[] = {
169         0x01, 0x80, 0xC2, 0x00, 0x00, 0x01
170 };
171
172 /* Please, keep in sync with struct ibm_emac_stats/ibm_emac_error_stats */
173 static const char emac_stats_keys[EMAC_ETHTOOL_STATS_COUNT][ETH_GSTRING_LEN] = {
174         "rx_packets", "rx_bytes", "tx_packets", "tx_bytes", "rx_packets_csum",
175         "tx_packets_csum", "tx_undo", "rx_dropped_stack", "rx_dropped_oom",
176         "rx_dropped_error", "rx_dropped_resize", "rx_dropped_mtu",
177         "rx_stopped", "rx_bd_errors", "rx_bd_overrun", "rx_bd_bad_packet",
178         "rx_bd_runt_packet", "rx_bd_short_event", "rx_bd_alignment_error",
179         "rx_bd_bad_fcs", "rx_bd_packet_too_long", "rx_bd_out_of_range",
180         "rx_bd_in_range", "rx_parity", "rx_fifo_overrun", "rx_overrun",
181         "rx_bad_packet", "rx_runt_packet", "rx_short_event",
182         "rx_alignment_error", "rx_bad_fcs", "rx_packet_too_long",
183         "rx_out_of_range", "rx_in_range", "tx_dropped", "tx_bd_errors",
184         "tx_bd_bad_fcs", "tx_bd_carrier_loss", "tx_bd_excessive_deferral",
185         "tx_bd_excessive_collisions", "tx_bd_late_collision",
186         "tx_bd_multple_collisions", "tx_bd_single_collision",
187         "tx_bd_underrun", "tx_bd_sqe", "tx_parity", "tx_underrun", "tx_sqe",
188         "tx_errors"
189 };
190
191 static irqreturn_t emac_irq(int irq, void *dev_instance);
192 static void emac_clean_tx_ring(struct emac_instance *dev);
193 static void __emac_set_multicast_list(struct emac_instance *dev);
194
195 static inline int emac_phy_supports_gige(int phy_mode)
196 {
197         return  phy_interface_mode_is_rgmii(phy_mode) ||
198                 phy_mode == PHY_INTERFACE_MODE_GMII ||
199                 phy_mode == PHY_INTERFACE_MODE_SGMII ||
200                 phy_mode == PHY_INTERFACE_MODE_TBI ||
201                 phy_mode == PHY_INTERFACE_MODE_RTBI;
202 }
203
204 static inline int emac_phy_gpcs(int phy_mode)
205 {
206         return  phy_mode == PHY_INTERFACE_MODE_SGMII ||
207                 phy_mode == PHY_INTERFACE_MODE_TBI ||
208                 phy_mode == PHY_INTERFACE_MODE_RTBI;
209 }
210
211 static inline void emac_tx_enable(struct emac_instance *dev)
212 {
213         struct emac_regs __iomem *p = dev->emacp;
214         u32 r;
215
216         DBG(dev, "tx_enable" NL);
217
218         r = in_be32(&p->mr0);
219         if (!(r & EMAC_MR0_TXE))
220                 out_be32(&p->mr0, r | EMAC_MR0_TXE);
221 }
222
223 static void emac_tx_disable(struct emac_instance *dev)
224 {
225         struct emac_regs __iomem *p = dev->emacp;
226         u32 r;
227
228         DBG(dev, "tx_disable" NL);
229
230         r = in_be32(&p->mr0);
231         if (r & EMAC_MR0_TXE) {
232                 int n = dev->stop_timeout;
233                 out_be32(&p->mr0, r & ~EMAC_MR0_TXE);
234                 while (!(in_be32(&p->mr0) & EMAC_MR0_TXI) && n) {
235                         udelay(1);
236                         --n;
237                 }
238                 if (unlikely(!n))
239                         emac_report_timeout_error(dev, "TX disable timeout");
240         }
241 }
242
243 static void emac_rx_enable(struct emac_instance *dev)
244 {
245         struct emac_regs __iomem *p = dev->emacp;
246         u32 r;
247
248         if (unlikely(test_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags)))
249                 goto out;
250
251         DBG(dev, "rx_enable" NL);
252
253         r = in_be32(&p->mr0);
254         if (!(r & EMAC_MR0_RXE)) {
255                 if (unlikely(!(r & EMAC_MR0_RXI))) {
256                         /* Wait if previous async disable is still in progress */
257                         int n = dev->stop_timeout;
258                         while (!(r = in_be32(&p->mr0) & EMAC_MR0_RXI) && n) {
259                                 udelay(1);
260                                 --n;
261                         }
262                         if (unlikely(!n))
263                                 emac_report_timeout_error(dev,
264                                                           "RX disable timeout");
265                 }
266                 out_be32(&p->mr0, r | EMAC_MR0_RXE);
267         }
268  out:
269         ;
270 }
271
272 static void emac_rx_disable(struct emac_instance *dev)
273 {
274         struct emac_regs __iomem *p = dev->emacp;
275         u32 r;
276
277         DBG(dev, "rx_disable" NL);
278
279         r = in_be32(&p->mr0);
280         if (r & EMAC_MR0_RXE) {
281                 int n = dev->stop_timeout;
282                 out_be32(&p->mr0, r & ~EMAC_MR0_RXE);
283                 while (!(in_be32(&p->mr0) & EMAC_MR0_RXI) && n) {
284                         udelay(1);
285                         --n;
286                 }
287                 if (unlikely(!n))
288                         emac_report_timeout_error(dev, "RX disable timeout");
289         }
290 }
291
292 static inline void emac_netif_stop(struct emac_instance *dev)
293 {
294         netif_tx_lock_bh(dev->ndev);
295         netif_addr_lock(dev->ndev);
296         dev->no_mcast = 1;
297         netif_addr_unlock(dev->ndev);
298         netif_tx_unlock_bh(dev->ndev);
299         netif_trans_update(dev->ndev);  /* prevent tx timeout */
300         mal_poll_disable(dev->mal, &dev->commac);
301         netif_tx_disable(dev->ndev);
302 }
303
304 static inline void emac_netif_start(struct emac_instance *dev)
305 {
306         netif_tx_lock_bh(dev->ndev);
307         netif_addr_lock(dev->ndev);
308         dev->no_mcast = 0;
309         if (dev->mcast_pending && netif_running(dev->ndev))
310                 __emac_set_multicast_list(dev);
311         netif_addr_unlock(dev->ndev);
312         netif_tx_unlock_bh(dev->ndev);
313
314         netif_wake_queue(dev->ndev);
315
316         /* NOTE: unconditional netif_wake_queue is only appropriate
317          * so long as all callers are assured to have free tx slots
318          * (taken from tg3... though the case where that is wrong is
319          *  not terribly harmful)
320          */
321         mal_poll_enable(dev->mal, &dev->commac);
322 }
323
324 static inline void emac_rx_disable_async(struct emac_instance *dev)
325 {
326         struct emac_regs __iomem *p = dev->emacp;
327         u32 r;
328
329         DBG(dev, "rx_disable_async" NL);
330
331         r = in_be32(&p->mr0);
332         if (r & EMAC_MR0_RXE)
333                 out_be32(&p->mr0, r & ~EMAC_MR0_RXE);
334 }
335
336 static int emac_reset(struct emac_instance *dev)
337 {
338         struct emac_regs __iomem *p = dev->emacp;
339         int n = 20;
340         bool __maybe_unused try_internal_clock = false;
341
342         DBG(dev, "reset" NL);
343
344         if (!dev->reset_failed) {
345                 /* 40x erratum suggests stopping RX channel before reset,
346                  * we stop TX as well
347                  */
348                 emac_rx_disable(dev);
349                 emac_tx_disable(dev);
350         }
351
352 #ifdef CONFIG_PPC_DCR_NATIVE
353 do_retry:
354         /*
355          * PPC460EX/GT Embedded Processor Advanced User's Manual
356          * section 28.10.1 Mode Register 0 (EMACx_MR0) states:
357          * Note: The PHY must provide a TX Clk in order to perform a soft reset
358          * of the EMAC. If none is present, select the internal clock
359          * (SDR0_ETH_CFG[EMACx_PHY_CLK] = 1).
360          * After a soft reset, select the external clock.
361          *
362          * The AR8035-A PHY Meraki MR24 does not provide a TX Clk if the
363          * ethernet cable is not attached. This causes the reset to timeout
364          * and the PHY detection code in emac_init_phy() is unable to
365          * communicate and detect the AR8035-A PHY. As a result, the emac
366          * driver bails out early and the user has no ethernet.
367          * In order to stay compatible with existing configurations, the
368          * driver will temporarily switch to the internal clock, after
369          * the first reset fails.
370          */
371         if (emac_has_feature(dev, EMAC_FTR_460EX_PHY_CLK_FIX)) {
372                 if (try_internal_clock || (dev->phy_address == 0xffffffff &&
373                                            dev->phy_map == 0xffffffff)) {
374                         /* No PHY: select internal loop clock before reset */
375                         dcri_clrset(SDR0, SDR0_ETH_CFG,
376                                     0, SDR0_ETH_CFG_ECS << dev->cell_index);
377                 } else {
378                         /* PHY present: select external clock before reset */
379                         dcri_clrset(SDR0, SDR0_ETH_CFG,
380                                     SDR0_ETH_CFG_ECS << dev->cell_index, 0);
381                 }
382         }
383 #endif
384
385         out_be32(&p->mr0, EMAC_MR0_SRST);
386         while ((in_be32(&p->mr0) & EMAC_MR0_SRST) && n)
387                 --n;
388
389 #ifdef CONFIG_PPC_DCR_NATIVE
390         if (emac_has_feature(dev, EMAC_FTR_460EX_PHY_CLK_FIX)) {
391                 if (!n && !try_internal_clock) {
392                         /* first attempt has timed out. */
393                         n = 20;
394                         try_internal_clock = true;
395                         goto do_retry;
396                 }
397
398                 if (try_internal_clock || (dev->phy_address == 0xffffffff &&
399                                            dev->phy_map == 0xffffffff)) {
400                         /* No PHY: restore external clock source after reset */
401                         dcri_clrset(SDR0, SDR0_ETH_CFG,
402                                     SDR0_ETH_CFG_ECS << dev->cell_index, 0);
403                 }
404         }
405 #endif
406
407         if (n) {
408                 dev->reset_failed = 0;
409                 return 0;
410         } else {
411                 emac_report_timeout_error(dev, "reset timeout");
412                 dev->reset_failed = 1;
413                 return -ETIMEDOUT;
414         }
415 }
416
417 static void emac_hash_mc(struct emac_instance *dev)
418 {
419         const int regs = EMAC_XAHT_REGS(dev);
420         u32 *gaht_base = emac_gaht_base(dev);
421         u32 gaht_temp[EMAC_XAHT_MAX_REGS];
422         struct netdev_hw_addr *ha;
423         int i;
424
425         DBG(dev, "hash_mc %d" NL, netdev_mc_count(dev->ndev));
426
427         memset(gaht_temp, 0, sizeof (gaht_temp));
428
429         netdev_for_each_mc_addr(ha, dev->ndev) {
430                 int slot, reg, mask;
431                 DBG2(dev, "mc %pM" NL, ha->addr);
432
433                 slot = EMAC_XAHT_CRC_TO_SLOT(dev,
434                                              ether_crc(ETH_ALEN, ha->addr));
435                 reg = EMAC_XAHT_SLOT_TO_REG(dev, slot);
436                 mask = EMAC_XAHT_SLOT_TO_MASK(dev, slot);
437
438                 gaht_temp[reg] |= mask;
439         }
440
441         for (i = 0; i < regs; i++)
442                 out_be32(gaht_base + i, gaht_temp[i]);
443 }
444
445 static inline u32 emac_iff2rmr(struct net_device *ndev)
446 {
447         struct emac_instance *dev = netdev_priv(ndev);
448         u32 r;
449
450         r = EMAC_RMR_SP | EMAC_RMR_SFCS | EMAC_RMR_IAE | EMAC_RMR_BAE;
451
452         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
453             r |= EMAC4_RMR_BASE;
454         else
455             r |= EMAC_RMR_BASE;
456
457         if (ndev->flags & IFF_PROMISC)
458                 r |= EMAC_RMR_PME;
459         else if (ndev->flags & IFF_ALLMULTI ||
460                          (netdev_mc_count(ndev) > EMAC_XAHT_SLOTS(dev)))
461                 r |= EMAC_RMR_PMME;
462         else if (!netdev_mc_empty(ndev))
463                 r |= EMAC_RMR_MAE;
464
465         if (emac_has_feature(dev, EMAC_APM821XX_REQ_JUMBO_FRAME_SIZE)) {
466                 r &= ~EMAC4_RMR_MJS_MASK;
467                 r |= EMAC4_RMR_MJS(ndev->mtu);
468         }
469
470         return r;
471 }
472
473 static u32 __emac_calc_base_mr1(struct emac_instance *dev, int tx_size, int rx_size)
474 {
475         u32 ret = EMAC_MR1_VLE | EMAC_MR1_IST | EMAC_MR1_TR0_MULT;
476
477         DBG2(dev, "__emac_calc_base_mr1" NL);
478
479         switch(tx_size) {
480         case 2048:
481                 ret |= EMAC_MR1_TFS_2K;
482                 break;
483         default:
484                 printk(KERN_WARNING "%s: Unknown Tx FIFO size %d\n",
485                        dev->ndev->name, tx_size);
486         }
487
488         switch(rx_size) {
489         case 16384:
490                 ret |= EMAC_MR1_RFS_16K;
491                 break;
492         case 4096:
493                 ret |= EMAC_MR1_RFS_4K;
494                 break;
495         default:
496                 printk(KERN_WARNING "%s: Unknown Rx FIFO size %d\n",
497                        dev->ndev->name, rx_size);
498         }
499
500         return ret;
501 }
502
503 static u32 __emac4_calc_base_mr1(struct emac_instance *dev, int tx_size, int rx_size)
504 {
505         u32 ret = EMAC_MR1_VLE | EMAC_MR1_IST | EMAC4_MR1_TR |
506                 EMAC4_MR1_OBCI(dev->opb_bus_freq / 1000000);
507
508         DBG2(dev, "__emac4_calc_base_mr1" NL);
509
510         switch(tx_size) {
511         case 16384:
512                 ret |= EMAC4_MR1_TFS_16K;
513                 break;
514         case 8192:
515                 ret |= EMAC4_MR1_TFS_8K;
516                 break;
517         case 4096:
518                 ret |= EMAC4_MR1_TFS_4K;
519                 break;
520         case 2048:
521                 ret |= EMAC4_MR1_TFS_2K;
522                 break;
523         default:
524                 printk(KERN_WARNING "%s: Unknown Tx FIFO size %d\n",
525                        dev->ndev->name, tx_size);
526         }
527
528         switch(rx_size) {
529         case 16384:
530                 ret |= EMAC4_MR1_RFS_16K;
531                 break;
532         case 8192:
533                 ret |= EMAC4_MR1_RFS_8K;
534                 break;
535         case 4096:
536                 ret |= EMAC4_MR1_RFS_4K;
537                 break;
538         case 2048:
539                 ret |= EMAC4_MR1_RFS_2K;
540                 break;
541         default:
542                 printk(KERN_WARNING "%s: Unknown Rx FIFO size %d\n",
543                        dev->ndev->name, rx_size);
544         }
545
546         return ret;
547 }
548
549 static u32 emac_calc_base_mr1(struct emac_instance *dev, int tx_size, int rx_size)
550 {
551         return emac_has_feature(dev, EMAC_FTR_EMAC4) ?
552                 __emac4_calc_base_mr1(dev, tx_size, rx_size) :
553                 __emac_calc_base_mr1(dev, tx_size, rx_size);
554 }
555
556 static inline u32 emac_calc_trtr(struct emac_instance *dev, unsigned int size)
557 {
558         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
559                 return ((size >> 6) - 1) << EMAC_TRTR_SHIFT_EMAC4;
560         else
561                 return ((size >> 6) - 1) << EMAC_TRTR_SHIFT;
562 }
563
564 static inline u32 emac_calc_rwmr(struct emac_instance *dev,
565                                  unsigned int low, unsigned int high)
566 {
567         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
568                 return (low << 22) | ( (high & 0x3ff) << 6);
569         else
570                 return (low << 23) | ( (high & 0x1ff) << 7);
571 }
572
573 static int emac_configure(struct emac_instance *dev)
574 {
575         struct emac_regs __iomem *p = dev->emacp;
576         struct net_device *ndev = dev->ndev;
577         int tx_size, rx_size, link = netif_carrier_ok(dev->ndev);
578         u32 r, mr1 = 0;
579
580         DBG(dev, "configure" NL);
581
582         if (!link) {
583                 out_be32(&p->mr1, in_be32(&p->mr1)
584                          | EMAC_MR1_FDE | EMAC_MR1_ILE);
585                 udelay(100);
586         } else if (emac_reset(dev) < 0)
587                 return -ETIMEDOUT;
588
589         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
590                 tah_reset(dev->tah_dev);
591
592         DBG(dev, " link = %d duplex = %d, pause = %d, asym_pause = %d\n",
593             link, dev->phy.duplex, dev->phy.pause, dev->phy.asym_pause);
594
595         /* Default fifo sizes */
596         tx_size = dev->tx_fifo_size;
597         rx_size = dev->rx_fifo_size;
598
599         /* No link, force loopback */
600         if (!link)
601                 mr1 = EMAC_MR1_FDE | EMAC_MR1_ILE;
602
603         /* Check for full duplex */
604         else if (dev->phy.duplex == DUPLEX_FULL)
605                 mr1 |= EMAC_MR1_FDE | EMAC_MR1_MWSW_001;
606
607         /* Adjust fifo sizes, mr1 and timeouts based on link speed */
608         dev->stop_timeout = STOP_TIMEOUT_10;
609         switch (dev->phy.speed) {
610         case SPEED_1000:
611                 if (emac_phy_gpcs(dev->phy.mode)) {
612                         mr1 |= EMAC_MR1_MF_1000GPCS | EMAC_MR1_MF_IPPA(
613                                 (dev->phy.gpcs_address != 0xffffffff) ?
614                                  dev->phy.gpcs_address : dev->phy.address);
615
616                         /* Put some arbitrary OUI, Manuf & Rev IDs so we can
617                          * identify this GPCS PHY later.
618                          */
619                         out_be32(&p->u1.emac4.ipcr, 0xdeadbeef);
620                 } else
621                         mr1 |= EMAC_MR1_MF_1000;
622
623                 /* Extended fifo sizes */
624                 tx_size = dev->tx_fifo_size_gige;
625                 rx_size = dev->rx_fifo_size_gige;
626
627                 if (dev->ndev->mtu > ETH_DATA_LEN) {
628                         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
629                                 mr1 |= EMAC4_MR1_JPSM;
630                         else
631                                 mr1 |= EMAC_MR1_JPSM;
632                         dev->stop_timeout = STOP_TIMEOUT_1000_JUMBO;
633                 } else
634                         dev->stop_timeout = STOP_TIMEOUT_1000;
635                 break;
636         case SPEED_100:
637                 mr1 |= EMAC_MR1_MF_100;
638                 dev->stop_timeout = STOP_TIMEOUT_100;
639                 break;
640         default: /* make gcc happy */
641                 break;
642         }
643
644         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
645                 rgmii_set_speed(dev->rgmii_dev, dev->rgmii_port,
646                                 dev->phy.speed);
647         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
648                 zmii_set_speed(dev->zmii_dev, dev->zmii_port, dev->phy.speed);
649
650         /* on 40x erratum forces us to NOT use integrated flow control,
651          * let's hope it works on 44x ;)
652          */
653         if (!emac_has_feature(dev, EMAC_FTR_NO_FLOW_CONTROL_40x) &&
654             dev->phy.duplex == DUPLEX_FULL) {
655                 if (dev->phy.pause)
656                         mr1 |= EMAC_MR1_EIFC | EMAC_MR1_APP;
657                 else if (dev->phy.asym_pause)
658                         mr1 |= EMAC_MR1_APP;
659         }
660
661         /* Add base settings & fifo sizes & program MR1 */
662         mr1 |= emac_calc_base_mr1(dev, tx_size, rx_size);
663         out_be32(&p->mr1, mr1);
664
665         /* Set individual MAC address */
666         out_be32(&p->iahr, (ndev->dev_addr[0] << 8) | ndev->dev_addr[1]);
667         out_be32(&p->ialr, (ndev->dev_addr[2] << 24) |
668                  (ndev->dev_addr[3] << 16) | (ndev->dev_addr[4] << 8) |
669                  ndev->dev_addr[5]);
670
671         /* VLAN Tag Protocol ID */
672         out_be32(&p->vtpid, 0x8100);
673
674         /* Receive mode register */
675         r = emac_iff2rmr(ndev);
676         if (r & EMAC_RMR_MAE)
677                 emac_hash_mc(dev);
678         out_be32(&p->rmr, r);
679
680         /* FIFOs thresholds */
681         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
682                 r = EMAC4_TMR1((dev->mal_burst_size / dev->fifo_entry_size) + 1,
683                                tx_size / 2 / dev->fifo_entry_size);
684         else
685                 r = EMAC_TMR1((dev->mal_burst_size / dev->fifo_entry_size) + 1,
686                               tx_size / 2 / dev->fifo_entry_size);
687         out_be32(&p->tmr1, r);
688         out_be32(&p->trtr, emac_calc_trtr(dev, tx_size / 2));
689
690         /* PAUSE frame is sent when RX FIFO reaches its high-water mark,
691            there should be still enough space in FIFO to allow the our link
692            partner time to process this frame and also time to send PAUSE
693            frame itself.
694
695            Here is the worst case scenario for the RX FIFO "headroom"
696            (from "The Switch Book") (100Mbps, without preamble, inter-frame gap):
697
698            1) One maximum-length frame on TX                    1522 bytes
699            2) One PAUSE frame time                                64 bytes
700            3) PAUSE frame decode time allowance                   64 bytes
701            4) One maximum-length frame on RX                    1522 bytes
702            5) Round-trip propagation delay of the link (100Mb)    15 bytes
703            ----------
704            3187 bytes
705
706            I chose to set high-water mark to RX_FIFO_SIZE / 4 (1024 bytes)
707            low-water mark  to RX_FIFO_SIZE / 8 (512 bytes)
708          */
709         r = emac_calc_rwmr(dev, rx_size / 8 / dev->fifo_entry_size,
710                            rx_size / 4 / dev->fifo_entry_size);
711         out_be32(&p->rwmr, r);
712
713         /* Set PAUSE timer to the maximum */
714         out_be32(&p->ptr, 0xffff);
715
716         /* IRQ sources */
717         r = EMAC_ISR_OVR | EMAC_ISR_BP | EMAC_ISR_SE |
718                 EMAC_ISR_ALE | EMAC_ISR_BFCS | EMAC_ISR_PTLE | EMAC_ISR_ORE |
719                 EMAC_ISR_IRE | EMAC_ISR_TE;
720         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
721             r |= EMAC4_ISR_TXPE | EMAC4_ISR_RXPE /* | EMAC4_ISR_TXUE |
722                                                   EMAC4_ISR_RXOE | */;
723         out_be32(&p->iser,  r);
724
725         /* We need to take GPCS PHY out of isolate mode after EMAC reset */
726         if (emac_phy_gpcs(dev->phy.mode)) {
727                 if (dev->phy.gpcs_address != 0xffffffff)
728                         emac_mii_reset_gpcs(&dev->phy);
729                 else
730                         emac_mii_reset_phy(&dev->phy);
731         }
732
733         return 0;
734 }
735
736 static void emac_reinitialize(struct emac_instance *dev)
737 {
738         DBG(dev, "reinitialize" NL);
739
740         emac_netif_stop(dev);
741         if (!emac_configure(dev)) {
742                 emac_tx_enable(dev);
743                 emac_rx_enable(dev);
744         }
745         emac_netif_start(dev);
746 }
747
748 static void emac_full_tx_reset(struct emac_instance *dev)
749 {
750         DBG(dev, "full_tx_reset" NL);
751
752         emac_tx_disable(dev);
753         mal_disable_tx_channel(dev->mal, dev->mal_tx_chan);
754         emac_clean_tx_ring(dev);
755         dev->tx_cnt = dev->tx_slot = dev->ack_slot = 0;
756
757         emac_configure(dev);
758
759         mal_enable_tx_channel(dev->mal, dev->mal_tx_chan);
760         emac_tx_enable(dev);
761         emac_rx_enable(dev);
762 }
763
764 static void emac_reset_work(struct work_struct *work)
765 {
766         struct emac_instance *dev = container_of(work, struct emac_instance, reset_work);
767
768         DBG(dev, "reset_work" NL);
769
770         mutex_lock(&dev->link_lock);
771         if (dev->opened) {
772                 emac_netif_stop(dev);
773                 emac_full_tx_reset(dev);
774                 emac_netif_start(dev);
775         }
776         mutex_unlock(&dev->link_lock);
777 }
778
779 static void emac_tx_timeout(struct net_device *ndev)
780 {
781         struct emac_instance *dev = netdev_priv(ndev);
782
783         DBG(dev, "tx_timeout" NL);
784
785         schedule_work(&dev->reset_work);
786 }
787
788
789 static inline int emac_phy_done(struct emac_instance *dev, u32 stacr)
790 {
791         int done = !!(stacr & EMAC_STACR_OC);
792
793         if (emac_has_feature(dev, EMAC_FTR_STACR_OC_INVERT))
794                 done = !done;
795
796         return done;
797 };
798
799 static int __emac_mdio_read(struct emac_instance *dev, u8 id, u8 reg)
800 {
801         struct emac_regs __iomem *p = dev->emacp;
802         u32 r = 0;
803         int n, err = -ETIMEDOUT;
804
805         mutex_lock(&dev->mdio_lock);
806
807         DBG2(dev, "mdio_read(%02x,%02x)" NL, id, reg);
808
809         /* Enable proper MDIO port */
810         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
811                 zmii_get_mdio(dev->zmii_dev, dev->zmii_port);
812         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
813                 rgmii_get_mdio(dev->rgmii_dev, dev->rgmii_port);
814
815         /* Wait for management interface to become idle */
816         n = 20;
817         while (!emac_phy_done(dev, in_be32(&p->stacr))) {
818                 udelay(1);
819                 if (!--n) {
820                         DBG2(dev, " -> timeout wait idle\n");
821                         goto bail;
822                 }
823         }
824
825         /* Issue read command */
826         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
827                 r = EMAC4_STACR_BASE(dev->opb_bus_freq);
828         else
829                 r = EMAC_STACR_BASE(dev->opb_bus_freq);
830         if (emac_has_feature(dev, EMAC_FTR_STACR_OC_INVERT))
831                 r |= EMAC_STACR_OC;
832         if (emac_has_feature(dev, EMAC_FTR_HAS_NEW_STACR))
833                 r |= EMACX_STACR_STAC_READ;
834         else
835                 r |= EMAC_STACR_STAC_READ;
836         r |= (reg & EMAC_STACR_PRA_MASK)
837                 | ((id & EMAC_STACR_PCDA_MASK) << EMAC_STACR_PCDA_SHIFT);
838         out_be32(&p->stacr, r);
839
840         /* Wait for read to complete */
841         n = 200;
842         while (!emac_phy_done(dev, (r = in_be32(&p->stacr)))) {
843                 udelay(1);
844                 if (!--n) {
845                         DBG2(dev, " -> timeout wait complete\n");
846                         goto bail;
847                 }
848         }
849
850         if (unlikely(r & EMAC_STACR_PHYE)) {
851                 DBG(dev, "mdio_read(%02x, %02x) failed" NL, id, reg);
852                 err = -EREMOTEIO;
853                 goto bail;
854         }
855
856         r = ((r >> EMAC_STACR_PHYD_SHIFT) & EMAC_STACR_PHYD_MASK);
857
858         DBG2(dev, "mdio_read -> %04x" NL, r);
859         err = 0;
860  bail:
861         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
862                 rgmii_put_mdio(dev->rgmii_dev, dev->rgmii_port);
863         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
864                 zmii_put_mdio(dev->zmii_dev, dev->zmii_port);
865         mutex_unlock(&dev->mdio_lock);
866
867         return err == 0 ? r : err;
868 }
869
870 static void __emac_mdio_write(struct emac_instance *dev, u8 id, u8 reg,
871                               u16 val)
872 {
873         struct emac_regs __iomem *p = dev->emacp;
874         u32 r = 0;
875         int n, err = -ETIMEDOUT;
876
877         mutex_lock(&dev->mdio_lock);
878
879         DBG2(dev, "mdio_write(%02x,%02x,%04x)" NL, id, reg, val);
880
881         /* Enable proper MDIO port */
882         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
883                 zmii_get_mdio(dev->zmii_dev, dev->zmii_port);
884         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
885                 rgmii_get_mdio(dev->rgmii_dev, dev->rgmii_port);
886
887         /* Wait for management interface to be idle */
888         n = 20;
889         while (!emac_phy_done(dev, in_be32(&p->stacr))) {
890                 udelay(1);
891                 if (!--n) {
892                         DBG2(dev, " -> timeout wait idle\n");
893                         goto bail;
894                 }
895         }
896
897         /* Issue write command */
898         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
899                 r = EMAC4_STACR_BASE(dev->opb_bus_freq);
900         else
901                 r = EMAC_STACR_BASE(dev->opb_bus_freq);
902         if (emac_has_feature(dev, EMAC_FTR_STACR_OC_INVERT))
903                 r |= EMAC_STACR_OC;
904         if (emac_has_feature(dev, EMAC_FTR_HAS_NEW_STACR))
905                 r |= EMACX_STACR_STAC_WRITE;
906         else
907                 r |= EMAC_STACR_STAC_WRITE;
908         r |= (reg & EMAC_STACR_PRA_MASK) |
909                 ((id & EMAC_STACR_PCDA_MASK) << EMAC_STACR_PCDA_SHIFT) |
910                 (val << EMAC_STACR_PHYD_SHIFT);
911         out_be32(&p->stacr, r);
912
913         /* Wait for write to complete */
914         n = 200;
915         while (!emac_phy_done(dev, in_be32(&p->stacr))) {
916                 udelay(1);
917                 if (!--n) {
918                         DBG2(dev, " -> timeout wait complete\n");
919                         goto bail;
920                 }
921         }
922         err = 0;
923  bail:
924         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
925                 rgmii_put_mdio(dev->rgmii_dev, dev->rgmii_port);
926         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
927                 zmii_put_mdio(dev->zmii_dev, dev->zmii_port);
928         mutex_unlock(&dev->mdio_lock);
929 }
930
931 static int emac_mdio_read(struct net_device *ndev, int id, int reg)
932 {
933         struct emac_instance *dev = netdev_priv(ndev);
934         int res;
935
936         res = __emac_mdio_read((dev->mdio_instance &&
937                                 dev->phy.gpcs_address != id) ?
938                                 dev->mdio_instance : dev,
939                                (u8) id, (u8) reg);
940         return res;
941 }
942
943 static void emac_mdio_write(struct net_device *ndev, int id, int reg, int val)
944 {
945         struct emac_instance *dev = netdev_priv(ndev);
946
947         __emac_mdio_write((dev->mdio_instance &&
948                            dev->phy.gpcs_address != id) ?
949                            dev->mdio_instance : dev,
950                           (u8) id, (u8) reg, (u16) val);
951 }
952
953 /* Tx lock BH */
954 static void __emac_set_multicast_list(struct emac_instance *dev)
955 {
956         struct emac_regs __iomem *p = dev->emacp;
957         u32 rmr = emac_iff2rmr(dev->ndev);
958
959         DBG(dev, "__multicast %08x" NL, rmr);
960
961         /* I decided to relax register access rules here to avoid
962          * full EMAC reset.
963          *
964          * There is a real problem with EMAC4 core if we use MWSW_001 bit
965          * in MR1 register and do a full EMAC reset.
966          * One TX BD status update is delayed and, after EMAC reset, it
967          * never happens, resulting in TX hung (it'll be recovered by TX
968          * timeout handler eventually, but this is just gross).
969          * So we either have to do full TX reset or try to cheat here :)
970          *
971          * The only required change is to RX mode register, so I *think* all
972          * we need is just to stop RX channel. This seems to work on all
973          * tested SoCs.                                                --ebs
974          *
975          * If we need the full reset, we might just trigger the workqueue
976          * and do it async... a bit nasty but should work --BenH
977          */
978         dev->mcast_pending = 0;
979         emac_rx_disable(dev);
980         if (rmr & EMAC_RMR_MAE)
981                 emac_hash_mc(dev);
982         out_be32(&p->rmr, rmr);
983         emac_rx_enable(dev);
984 }
985
986 /* Tx lock BH */
987 static void emac_set_multicast_list(struct net_device *ndev)
988 {
989         struct emac_instance *dev = netdev_priv(ndev);
990
991         DBG(dev, "multicast" NL);
992
993         BUG_ON(!netif_running(dev->ndev));
994
995         if (dev->no_mcast) {
996                 dev->mcast_pending = 1;
997                 return;
998         }
999
1000         mutex_lock(&dev->link_lock);
1001         __emac_set_multicast_list(dev);
1002         mutex_unlock(&dev->link_lock);
1003 }
1004
1005 static int emac_set_mac_address(struct net_device *ndev, void *sa)
1006 {
1007         struct emac_instance *dev = netdev_priv(ndev);
1008         struct sockaddr *addr = sa;
1009         struct emac_regs __iomem *p = dev->emacp;
1010
1011         if (!is_valid_ether_addr(addr->sa_data))
1012                return -EADDRNOTAVAIL;
1013
1014         mutex_lock(&dev->link_lock);
1015
1016         memcpy(ndev->dev_addr, addr->sa_data, ndev->addr_len);
1017
1018         emac_rx_disable(dev);
1019         emac_tx_disable(dev);
1020         out_be32(&p->iahr, (ndev->dev_addr[0] << 8) | ndev->dev_addr[1]);
1021         out_be32(&p->ialr, (ndev->dev_addr[2] << 24) |
1022                 (ndev->dev_addr[3] << 16) | (ndev->dev_addr[4] << 8) |
1023                 ndev->dev_addr[5]);
1024         emac_tx_enable(dev);
1025         emac_rx_enable(dev);
1026
1027         mutex_unlock(&dev->link_lock);
1028
1029         return 0;
1030 }
1031
1032 static int emac_resize_rx_ring(struct emac_instance *dev, int new_mtu)
1033 {
1034         int rx_sync_size = emac_rx_sync_size(new_mtu);
1035         int rx_skb_size = emac_rx_skb_size(new_mtu);
1036         int i, ret = 0;
1037         int mr1_jumbo_bit_change = 0;
1038
1039         mutex_lock(&dev->link_lock);
1040         emac_netif_stop(dev);
1041         emac_rx_disable(dev);
1042         mal_disable_rx_channel(dev->mal, dev->mal_rx_chan);
1043
1044         if (dev->rx_sg_skb) {
1045                 ++dev->estats.rx_dropped_resize;
1046                 dev_kfree_skb(dev->rx_sg_skb);
1047                 dev->rx_sg_skb = NULL;
1048         }
1049
1050         /* Make a first pass over RX ring and mark BDs ready, dropping
1051          * non-processed packets on the way. We need this as a separate pass
1052          * to simplify error recovery in the case of allocation failure later.
1053          */
1054         for (i = 0; i < NUM_RX_BUFF; ++i) {
1055                 if (dev->rx_desc[i].ctrl & MAL_RX_CTRL_FIRST)
1056                         ++dev->estats.rx_dropped_resize;
1057
1058                 dev->rx_desc[i].data_len = 0;
1059                 dev->rx_desc[i].ctrl = MAL_RX_CTRL_EMPTY |
1060                     (i == (NUM_RX_BUFF - 1) ? MAL_RX_CTRL_WRAP : 0);
1061         }
1062
1063         /* Reallocate RX ring only if bigger skb buffers are required */
1064         if (rx_skb_size <= dev->rx_skb_size)
1065                 goto skip;
1066
1067         /* Second pass, allocate new skbs */
1068         for (i = 0; i < NUM_RX_BUFF; ++i) {
1069                 struct sk_buff *skb;
1070
1071                 skb = netdev_alloc_skb_ip_align(dev->ndev, rx_skb_size);
1072                 if (!skb) {
1073                         ret = -ENOMEM;
1074                         goto oom;
1075                 }
1076
1077                 BUG_ON(!dev->rx_skb[i]);
1078                 dev_kfree_skb(dev->rx_skb[i]);
1079
1080                 dev->rx_desc[i].data_ptr =
1081                     dma_map_single(&dev->ofdev->dev, skb->data - NET_IP_ALIGN,
1082                                    rx_sync_size, DMA_FROM_DEVICE)
1083                                    + NET_IP_ALIGN;
1084                 dev->rx_skb[i] = skb;
1085         }
1086  skip:
1087         /* Check if we need to change "Jumbo" bit in MR1 */
1088         if (emac_has_feature(dev, EMAC_APM821XX_REQ_JUMBO_FRAME_SIZE)) {
1089                 mr1_jumbo_bit_change = (new_mtu > ETH_DATA_LEN) ||
1090                                 (dev->ndev->mtu > ETH_DATA_LEN);
1091         } else {
1092                 mr1_jumbo_bit_change = (new_mtu > ETH_DATA_LEN) ^
1093                                 (dev->ndev->mtu > ETH_DATA_LEN);
1094         }
1095
1096         if (mr1_jumbo_bit_change) {
1097                 /* This is to prevent starting RX channel in emac_rx_enable() */
1098                 set_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1099
1100                 dev->ndev->mtu = new_mtu;
1101                 emac_full_tx_reset(dev);
1102         }
1103
1104         mal_set_rcbs(dev->mal, dev->mal_rx_chan, emac_rx_size(new_mtu));
1105  oom:
1106         /* Restart RX */
1107         clear_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1108         dev->rx_slot = 0;
1109         mal_enable_rx_channel(dev->mal, dev->mal_rx_chan);
1110         emac_rx_enable(dev);
1111         emac_netif_start(dev);
1112         mutex_unlock(&dev->link_lock);
1113
1114         return ret;
1115 }
1116
1117 /* Process ctx, rtnl_lock semaphore */
1118 static int emac_change_mtu(struct net_device *ndev, int new_mtu)
1119 {
1120         struct emac_instance *dev = netdev_priv(ndev);
1121         int ret = 0;
1122
1123         DBG(dev, "change_mtu(%d)" NL, new_mtu);
1124
1125         if (netif_running(ndev)) {
1126                 /* Check if we really need to reinitialize RX ring */
1127                 if (emac_rx_skb_size(ndev->mtu) != emac_rx_skb_size(new_mtu))
1128                         ret = emac_resize_rx_ring(dev, new_mtu);
1129         }
1130
1131         if (!ret) {
1132                 ndev->mtu = new_mtu;
1133                 dev->rx_skb_size = emac_rx_skb_size(new_mtu);
1134                 dev->rx_sync_size = emac_rx_sync_size(new_mtu);
1135         }
1136
1137         return ret;
1138 }
1139
1140 static void emac_clean_tx_ring(struct emac_instance *dev)
1141 {
1142         int i;
1143
1144         for (i = 0; i < NUM_TX_BUFF; ++i) {
1145                 if (dev->tx_skb[i]) {
1146                         dev_kfree_skb(dev->tx_skb[i]);
1147                         dev->tx_skb[i] = NULL;
1148                         if (dev->tx_desc[i].ctrl & MAL_TX_CTRL_READY)
1149                                 ++dev->estats.tx_dropped;
1150                 }
1151                 dev->tx_desc[i].ctrl = 0;
1152                 dev->tx_desc[i].data_ptr = 0;
1153         }
1154 }
1155
1156 static void emac_clean_rx_ring(struct emac_instance *dev)
1157 {
1158         int i;
1159
1160         for (i = 0; i < NUM_RX_BUFF; ++i)
1161                 if (dev->rx_skb[i]) {
1162                         dev->rx_desc[i].ctrl = 0;
1163                         dev_kfree_skb(dev->rx_skb[i]);
1164                         dev->rx_skb[i] = NULL;
1165                         dev->rx_desc[i].data_ptr = 0;
1166                 }
1167
1168         if (dev->rx_sg_skb) {
1169                 dev_kfree_skb(dev->rx_sg_skb);
1170                 dev->rx_sg_skb = NULL;
1171         }
1172 }
1173
1174 static int
1175 __emac_prepare_rx_skb(struct sk_buff *skb, struct emac_instance *dev, int slot)
1176 {
1177         if (unlikely(!skb))
1178                 return -ENOMEM;
1179
1180         dev->rx_skb[slot] = skb;
1181         dev->rx_desc[slot].data_len = 0;
1182
1183         dev->rx_desc[slot].data_ptr =
1184             dma_map_single(&dev->ofdev->dev, skb->data - NET_IP_ALIGN,
1185                            dev->rx_sync_size, DMA_FROM_DEVICE) + NET_IP_ALIGN;
1186         wmb();
1187         dev->rx_desc[slot].ctrl = MAL_RX_CTRL_EMPTY |
1188             (slot == (NUM_RX_BUFF - 1) ? MAL_RX_CTRL_WRAP : 0);
1189
1190         return 0;
1191 }
1192
1193 static int
1194 emac_alloc_rx_skb(struct emac_instance *dev, int slot)
1195 {
1196         struct sk_buff *skb;
1197
1198         skb = __netdev_alloc_skb_ip_align(dev->ndev, dev->rx_skb_size,
1199                                           GFP_KERNEL);
1200
1201         return __emac_prepare_rx_skb(skb, dev, slot);
1202 }
1203
1204 static int
1205 emac_alloc_rx_skb_napi(struct emac_instance *dev, int slot)
1206 {
1207         struct sk_buff *skb;
1208
1209         skb = napi_alloc_skb(&dev->mal->napi, dev->rx_skb_size);
1210
1211         return __emac_prepare_rx_skb(skb, dev, slot);
1212 }
1213
1214 static void emac_print_link_status(struct emac_instance *dev)
1215 {
1216         if (netif_carrier_ok(dev->ndev))
1217                 printk(KERN_INFO "%s: link is up, %d %s%s\n",
1218                        dev->ndev->name, dev->phy.speed,
1219                        dev->phy.duplex == DUPLEX_FULL ? "FDX" : "HDX",
1220                        dev->phy.pause ? ", pause enabled" :
1221                        dev->phy.asym_pause ? ", asymmetric pause enabled" : "");
1222         else
1223                 printk(KERN_INFO "%s: link is down\n", dev->ndev->name);
1224 }
1225
1226 /* Process ctx, rtnl_lock semaphore */
1227 static int emac_open(struct net_device *ndev)
1228 {
1229         struct emac_instance *dev = netdev_priv(ndev);
1230         int err, i;
1231
1232         DBG(dev, "open" NL);
1233
1234         /* Setup error IRQ handler */
1235         err = request_irq(dev->emac_irq, emac_irq, 0, "EMAC", dev);
1236         if (err) {
1237                 printk(KERN_ERR "%s: failed to request IRQ %d\n",
1238                        ndev->name, dev->emac_irq);
1239                 return err;
1240         }
1241
1242         /* Allocate RX ring */
1243         for (i = 0; i < NUM_RX_BUFF; ++i)
1244                 if (emac_alloc_rx_skb(dev, i)) {
1245                         printk(KERN_ERR "%s: failed to allocate RX ring\n",
1246                                ndev->name);
1247                         goto oom;
1248                 }
1249
1250         dev->tx_cnt = dev->tx_slot = dev->ack_slot = dev->rx_slot = 0;
1251         clear_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1252         dev->rx_sg_skb = NULL;
1253
1254         mutex_lock(&dev->link_lock);
1255         dev->opened = 1;
1256
1257         /* Start PHY polling now.
1258          */
1259         if (dev->phy.address >= 0) {
1260                 int link_poll_interval;
1261                 if (dev->phy.def->ops->poll_link(&dev->phy)) {
1262                         dev->phy.def->ops->read_link(&dev->phy);
1263                         emac_rx_clk_default(dev);
1264                         netif_carrier_on(dev->ndev);
1265                         link_poll_interval = PHY_POLL_LINK_ON;
1266                 } else {
1267                         emac_rx_clk_tx(dev);
1268                         netif_carrier_off(dev->ndev);
1269                         link_poll_interval = PHY_POLL_LINK_OFF;
1270                 }
1271                 dev->link_polling = 1;
1272                 wmb();
1273                 schedule_delayed_work(&dev->link_work, link_poll_interval);
1274                 emac_print_link_status(dev);
1275         } else
1276                 netif_carrier_on(dev->ndev);
1277
1278         /* Required for Pause packet support in EMAC */
1279         dev_mc_add_global(ndev, default_mcast_addr);
1280
1281         emac_configure(dev);
1282         mal_poll_add(dev->mal, &dev->commac);
1283         mal_enable_tx_channel(dev->mal, dev->mal_tx_chan);
1284         mal_set_rcbs(dev->mal, dev->mal_rx_chan, emac_rx_size(ndev->mtu));
1285         mal_enable_rx_channel(dev->mal, dev->mal_rx_chan);
1286         emac_tx_enable(dev);
1287         emac_rx_enable(dev);
1288         emac_netif_start(dev);
1289
1290         mutex_unlock(&dev->link_lock);
1291
1292         return 0;
1293  oom:
1294         emac_clean_rx_ring(dev);
1295         free_irq(dev->emac_irq, dev);
1296
1297         return -ENOMEM;
1298 }
1299
1300 /* BHs disabled */
1301 #if 0
1302 static int emac_link_differs(struct emac_instance *dev)
1303 {
1304         u32 r = in_be32(&dev->emacp->mr1);
1305
1306         int duplex = r & EMAC_MR1_FDE ? DUPLEX_FULL : DUPLEX_HALF;
1307         int speed, pause, asym_pause;
1308
1309         if (r & EMAC_MR1_MF_1000)
1310                 speed = SPEED_1000;
1311         else if (r & EMAC_MR1_MF_100)
1312                 speed = SPEED_100;
1313         else
1314                 speed = SPEED_10;
1315
1316         switch (r & (EMAC_MR1_EIFC | EMAC_MR1_APP)) {
1317         case (EMAC_MR1_EIFC | EMAC_MR1_APP):
1318                 pause = 1;
1319                 asym_pause = 0;
1320                 break;
1321         case EMAC_MR1_APP:
1322                 pause = 0;
1323                 asym_pause = 1;
1324                 break;
1325         default:
1326                 pause = asym_pause = 0;
1327         }
1328         return speed != dev->phy.speed || duplex != dev->phy.duplex ||
1329             pause != dev->phy.pause || asym_pause != dev->phy.asym_pause;
1330 }
1331 #endif
1332
1333 static void emac_link_timer(struct work_struct *work)
1334 {
1335         struct emac_instance *dev =
1336                 container_of(to_delayed_work(work),
1337                              struct emac_instance, link_work);
1338         int link_poll_interval;
1339
1340         mutex_lock(&dev->link_lock);
1341         DBG2(dev, "link timer" NL);
1342
1343         if (!dev->opened)
1344                 goto bail;
1345
1346         if (dev->phy.def->ops->poll_link(&dev->phy)) {
1347                 if (!netif_carrier_ok(dev->ndev)) {
1348                         emac_rx_clk_default(dev);
1349                         /* Get new link parameters */
1350                         dev->phy.def->ops->read_link(&dev->phy);
1351
1352                         netif_carrier_on(dev->ndev);
1353                         emac_netif_stop(dev);
1354                         emac_full_tx_reset(dev);
1355                         emac_netif_start(dev);
1356                         emac_print_link_status(dev);
1357                 }
1358                 link_poll_interval = PHY_POLL_LINK_ON;
1359         } else {
1360                 if (netif_carrier_ok(dev->ndev)) {
1361                         emac_rx_clk_tx(dev);
1362                         netif_carrier_off(dev->ndev);
1363                         netif_tx_disable(dev->ndev);
1364                         emac_reinitialize(dev);
1365                         emac_print_link_status(dev);
1366                 }
1367                 link_poll_interval = PHY_POLL_LINK_OFF;
1368         }
1369         schedule_delayed_work(&dev->link_work, link_poll_interval);
1370  bail:
1371         mutex_unlock(&dev->link_lock);
1372 }
1373
1374 static void emac_force_link_update(struct emac_instance *dev)
1375 {
1376         netif_carrier_off(dev->ndev);
1377         smp_rmb();
1378         if (dev->link_polling) {
1379                 cancel_delayed_work_sync(&dev->link_work);
1380                 if (dev->link_polling)
1381                         schedule_delayed_work(&dev->link_work,  PHY_POLL_LINK_OFF);
1382         }
1383 }
1384
1385 /* Process ctx, rtnl_lock semaphore */
1386 static int emac_close(struct net_device *ndev)
1387 {
1388         struct emac_instance *dev = netdev_priv(ndev);
1389
1390         DBG(dev, "close" NL);
1391
1392         if (dev->phy.address >= 0) {
1393                 dev->link_polling = 0;
1394                 cancel_delayed_work_sync(&dev->link_work);
1395         }
1396         mutex_lock(&dev->link_lock);
1397         emac_netif_stop(dev);
1398         dev->opened = 0;
1399         mutex_unlock(&dev->link_lock);
1400
1401         emac_rx_disable(dev);
1402         emac_tx_disable(dev);
1403         mal_disable_rx_channel(dev->mal, dev->mal_rx_chan);
1404         mal_disable_tx_channel(dev->mal, dev->mal_tx_chan);
1405         mal_poll_del(dev->mal, &dev->commac);
1406
1407         emac_clean_tx_ring(dev);
1408         emac_clean_rx_ring(dev);
1409
1410         free_irq(dev->emac_irq, dev);
1411
1412         netif_carrier_off(ndev);
1413
1414         return 0;
1415 }
1416
1417 static inline u16 emac_tx_csum(struct emac_instance *dev,
1418                                struct sk_buff *skb)
1419 {
1420         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH) &&
1421                 (skb->ip_summed == CHECKSUM_PARTIAL)) {
1422                 ++dev->stats.tx_packets_csum;
1423                 return EMAC_TX_CTRL_TAH_CSUM;
1424         }
1425         return 0;
1426 }
1427
1428 static inline netdev_tx_t emac_xmit_finish(struct emac_instance *dev, int len)
1429 {
1430         struct emac_regs __iomem *p = dev->emacp;
1431         struct net_device *ndev = dev->ndev;
1432
1433         /* Send the packet out. If the if makes a significant perf
1434          * difference, then we can store the TMR0 value in "dev"
1435          * instead
1436          */
1437         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
1438                 out_be32(&p->tmr0, EMAC4_TMR0_XMIT);
1439         else
1440                 out_be32(&p->tmr0, EMAC_TMR0_XMIT);
1441
1442         if (unlikely(++dev->tx_cnt == NUM_TX_BUFF)) {
1443                 netif_stop_queue(ndev);
1444                 DBG2(dev, "stopped TX queue" NL);
1445         }
1446
1447         netif_trans_update(ndev);
1448         ++dev->stats.tx_packets;
1449         dev->stats.tx_bytes += len;
1450
1451         return NETDEV_TX_OK;
1452 }
1453
1454 /* Tx lock BH */
1455 static netdev_tx_t emac_start_xmit(struct sk_buff *skb, struct net_device *ndev)
1456 {
1457         struct emac_instance *dev = netdev_priv(ndev);
1458         unsigned int len = skb->len;
1459         int slot;
1460
1461         u16 ctrl = EMAC_TX_CTRL_GFCS | EMAC_TX_CTRL_GP | MAL_TX_CTRL_READY |
1462             MAL_TX_CTRL_LAST | emac_tx_csum(dev, skb);
1463
1464         slot = dev->tx_slot++;
1465         if (dev->tx_slot == NUM_TX_BUFF) {
1466                 dev->tx_slot = 0;
1467                 ctrl |= MAL_TX_CTRL_WRAP;
1468         }
1469
1470         DBG2(dev, "xmit(%u) %d" NL, len, slot);
1471
1472         dev->tx_skb[slot] = skb;
1473         dev->tx_desc[slot].data_ptr = dma_map_single(&dev->ofdev->dev,
1474                                                      skb->data, len,
1475                                                      DMA_TO_DEVICE);
1476         dev->tx_desc[slot].data_len = (u16) len;
1477         wmb();
1478         dev->tx_desc[slot].ctrl = ctrl;
1479
1480         return emac_xmit_finish(dev, len);
1481 }
1482
1483 static inline int emac_xmit_split(struct emac_instance *dev, int slot,
1484                                   u32 pd, int len, int last, u16 base_ctrl)
1485 {
1486         while (1) {
1487                 u16 ctrl = base_ctrl;
1488                 int chunk = min(len, MAL_MAX_TX_SIZE);
1489                 len -= chunk;
1490
1491                 slot = (slot + 1) % NUM_TX_BUFF;
1492
1493                 if (last && !len)
1494                         ctrl |= MAL_TX_CTRL_LAST;
1495                 if (slot == NUM_TX_BUFF - 1)
1496                         ctrl |= MAL_TX_CTRL_WRAP;
1497
1498                 dev->tx_skb[slot] = NULL;
1499                 dev->tx_desc[slot].data_ptr = pd;
1500                 dev->tx_desc[slot].data_len = (u16) chunk;
1501                 dev->tx_desc[slot].ctrl = ctrl;
1502                 ++dev->tx_cnt;
1503
1504                 if (!len)
1505                         break;
1506
1507                 pd += chunk;
1508         }
1509         return slot;
1510 }
1511
1512 /* Tx lock BH disabled (SG version for TAH equipped EMACs) */
1513 static netdev_tx_t
1514 emac_start_xmit_sg(struct sk_buff *skb, struct net_device *ndev)
1515 {
1516         struct emac_instance *dev = netdev_priv(ndev);
1517         int nr_frags = skb_shinfo(skb)->nr_frags;
1518         int len = skb->len, chunk;
1519         int slot, i;
1520         u16 ctrl;
1521         u32 pd;
1522
1523         /* This is common "fast" path */
1524         if (likely(!nr_frags && len <= MAL_MAX_TX_SIZE))
1525                 return emac_start_xmit(skb, ndev);
1526
1527         len -= skb->data_len;
1528
1529         /* Note, this is only an *estimation*, we can still run out of empty
1530          * slots because of the additional fragmentation into
1531          * MAL_MAX_TX_SIZE-sized chunks
1532          */
1533         if (unlikely(dev->tx_cnt + nr_frags + mal_tx_chunks(len) > NUM_TX_BUFF))
1534                 goto stop_queue;
1535
1536         ctrl = EMAC_TX_CTRL_GFCS | EMAC_TX_CTRL_GP | MAL_TX_CTRL_READY |
1537             emac_tx_csum(dev, skb);
1538         slot = dev->tx_slot;
1539
1540         /* skb data */
1541         dev->tx_skb[slot] = NULL;
1542         chunk = min(len, MAL_MAX_TX_SIZE);
1543         dev->tx_desc[slot].data_ptr = pd =
1544             dma_map_single(&dev->ofdev->dev, skb->data, len, DMA_TO_DEVICE);
1545         dev->tx_desc[slot].data_len = (u16) chunk;
1546         len -= chunk;
1547         if (unlikely(len))
1548                 slot = emac_xmit_split(dev, slot, pd + chunk, len, !nr_frags,
1549                                        ctrl);
1550         /* skb fragments */
1551         for (i = 0; i < nr_frags; ++i) {
1552                 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i];
1553                 len = skb_frag_size(frag);
1554
1555                 if (unlikely(dev->tx_cnt + mal_tx_chunks(len) >= NUM_TX_BUFF))
1556                         goto undo_frame;
1557
1558                 pd = skb_frag_dma_map(&dev->ofdev->dev, frag, 0, len,
1559                                       DMA_TO_DEVICE);
1560
1561                 slot = emac_xmit_split(dev, slot, pd, len, i == nr_frags - 1,
1562                                        ctrl);
1563         }
1564
1565         DBG2(dev, "xmit_sg(%u) %d - %d" NL, skb->len, dev->tx_slot, slot);
1566
1567         /* Attach skb to the last slot so we don't release it too early */
1568         dev->tx_skb[slot] = skb;
1569
1570         /* Send the packet out */
1571         if (dev->tx_slot == NUM_TX_BUFF - 1)
1572                 ctrl |= MAL_TX_CTRL_WRAP;
1573         wmb();
1574         dev->tx_desc[dev->tx_slot].ctrl = ctrl;
1575         dev->tx_slot = (slot + 1) % NUM_TX_BUFF;
1576
1577         return emac_xmit_finish(dev, skb->len);
1578
1579  undo_frame:
1580         /* Well, too bad. Our previous estimation was overly optimistic.
1581          * Undo everything.
1582          */
1583         while (slot != dev->tx_slot) {
1584                 dev->tx_desc[slot].ctrl = 0;
1585                 --dev->tx_cnt;
1586                 if (--slot < 0)
1587                         slot = NUM_TX_BUFF - 1;
1588         }
1589         ++dev->estats.tx_undo;
1590
1591  stop_queue:
1592         netif_stop_queue(ndev);
1593         DBG2(dev, "stopped TX queue" NL);
1594         return NETDEV_TX_BUSY;
1595 }
1596
1597 /* Tx lock BHs */
1598 static void emac_parse_tx_error(struct emac_instance *dev, u16 ctrl)
1599 {
1600         struct emac_error_stats *st = &dev->estats;
1601
1602         DBG(dev, "BD TX error %04x" NL, ctrl);
1603
1604         ++st->tx_bd_errors;
1605         if (ctrl & EMAC_TX_ST_BFCS)
1606                 ++st->tx_bd_bad_fcs;
1607         if (ctrl & EMAC_TX_ST_LCS)
1608                 ++st->tx_bd_carrier_loss;
1609         if (ctrl & EMAC_TX_ST_ED)
1610                 ++st->tx_bd_excessive_deferral;
1611         if (ctrl & EMAC_TX_ST_EC)
1612                 ++st->tx_bd_excessive_collisions;
1613         if (ctrl & EMAC_TX_ST_LC)
1614                 ++st->tx_bd_late_collision;
1615         if (ctrl & EMAC_TX_ST_MC)
1616                 ++st->tx_bd_multple_collisions;
1617         if (ctrl & EMAC_TX_ST_SC)
1618                 ++st->tx_bd_single_collision;
1619         if (ctrl & EMAC_TX_ST_UR)
1620                 ++st->tx_bd_underrun;
1621         if (ctrl & EMAC_TX_ST_SQE)
1622                 ++st->tx_bd_sqe;
1623 }
1624
1625 static void emac_poll_tx(void *param)
1626 {
1627         struct emac_instance *dev = param;
1628         u32 bad_mask;
1629
1630         DBG2(dev, "poll_tx, %d %d" NL, dev->tx_cnt, dev->ack_slot);
1631
1632         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
1633                 bad_mask = EMAC_IS_BAD_TX_TAH;
1634         else
1635                 bad_mask = EMAC_IS_BAD_TX;
1636
1637         netif_tx_lock_bh(dev->ndev);
1638         if (dev->tx_cnt) {
1639                 u16 ctrl;
1640                 int slot = dev->ack_slot, n = 0;
1641         again:
1642                 ctrl = dev->tx_desc[slot].ctrl;
1643                 if (!(ctrl & MAL_TX_CTRL_READY)) {
1644                         struct sk_buff *skb = dev->tx_skb[slot];
1645                         ++n;
1646
1647                         if (skb) {
1648                                 dev_kfree_skb(skb);
1649                                 dev->tx_skb[slot] = NULL;
1650                         }
1651                         slot = (slot + 1) % NUM_TX_BUFF;
1652
1653                         if (unlikely(ctrl & bad_mask))
1654                                 emac_parse_tx_error(dev, ctrl);
1655
1656                         if (--dev->tx_cnt)
1657                                 goto again;
1658                 }
1659                 if (n) {
1660                         dev->ack_slot = slot;
1661                         if (netif_queue_stopped(dev->ndev) &&
1662                             dev->tx_cnt < EMAC_TX_WAKEUP_THRESH)
1663                                 netif_wake_queue(dev->ndev);
1664
1665                         DBG2(dev, "tx %d pkts" NL, n);
1666                 }
1667         }
1668         netif_tx_unlock_bh(dev->ndev);
1669 }
1670
1671 static inline void emac_recycle_rx_skb(struct emac_instance *dev, int slot,
1672                                        int len)
1673 {
1674         struct sk_buff *skb = dev->rx_skb[slot];
1675
1676         DBG2(dev, "recycle %d %d" NL, slot, len);
1677
1678         if (len)
1679                 dma_map_single(&dev->ofdev->dev, skb->data - NET_IP_ALIGN,
1680                                SKB_DATA_ALIGN(len + NET_IP_ALIGN),
1681                                DMA_FROM_DEVICE);
1682
1683         dev->rx_desc[slot].data_len = 0;
1684         wmb();
1685         dev->rx_desc[slot].ctrl = MAL_RX_CTRL_EMPTY |
1686             (slot == (NUM_RX_BUFF - 1) ? MAL_RX_CTRL_WRAP : 0);
1687 }
1688
1689 static void emac_parse_rx_error(struct emac_instance *dev, u16 ctrl)
1690 {
1691         struct emac_error_stats *st = &dev->estats;
1692
1693         DBG(dev, "BD RX error %04x" NL, ctrl);
1694
1695         ++st->rx_bd_errors;
1696         if (ctrl & EMAC_RX_ST_OE)
1697                 ++st->rx_bd_overrun;
1698         if (ctrl & EMAC_RX_ST_BP)
1699                 ++st->rx_bd_bad_packet;
1700         if (ctrl & EMAC_RX_ST_RP)
1701                 ++st->rx_bd_runt_packet;
1702         if (ctrl & EMAC_RX_ST_SE)
1703                 ++st->rx_bd_short_event;
1704         if (ctrl & EMAC_RX_ST_AE)
1705                 ++st->rx_bd_alignment_error;
1706         if (ctrl & EMAC_RX_ST_BFCS)
1707                 ++st->rx_bd_bad_fcs;
1708         if (ctrl & EMAC_RX_ST_PTL)
1709                 ++st->rx_bd_packet_too_long;
1710         if (ctrl & EMAC_RX_ST_ORE)
1711                 ++st->rx_bd_out_of_range;
1712         if (ctrl & EMAC_RX_ST_IRE)
1713                 ++st->rx_bd_in_range;
1714 }
1715
1716 static inline void emac_rx_csum(struct emac_instance *dev,
1717                                 struct sk_buff *skb, u16 ctrl)
1718 {
1719 #ifdef CONFIG_IBM_EMAC_TAH
1720         if (!ctrl && dev->tah_dev) {
1721                 skb->ip_summed = CHECKSUM_UNNECESSARY;
1722                 ++dev->stats.rx_packets_csum;
1723         }
1724 #endif
1725 }
1726
1727 static inline int emac_rx_sg_append(struct emac_instance *dev, int slot)
1728 {
1729         if (likely(dev->rx_sg_skb != NULL)) {
1730                 int len = dev->rx_desc[slot].data_len;
1731                 int tot_len = dev->rx_sg_skb->len + len;
1732
1733                 if (unlikely(tot_len + NET_IP_ALIGN > dev->rx_skb_size)) {
1734                         ++dev->estats.rx_dropped_mtu;
1735                         dev_kfree_skb(dev->rx_sg_skb);
1736                         dev->rx_sg_skb = NULL;
1737                 } else {
1738                         memcpy(skb_tail_pointer(dev->rx_sg_skb),
1739                                          dev->rx_skb[slot]->data, len);
1740                         skb_put(dev->rx_sg_skb, len);
1741                         emac_recycle_rx_skb(dev, slot, len);
1742                         return 0;
1743                 }
1744         }
1745         emac_recycle_rx_skb(dev, slot, 0);
1746         return -1;
1747 }
1748
1749 /* NAPI poll context */
1750 static int emac_poll_rx(void *param, int budget)
1751 {
1752         struct emac_instance *dev = param;
1753         int slot = dev->rx_slot, received = 0;
1754
1755         DBG2(dev, "poll_rx(%d)" NL, budget);
1756
1757  again:
1758         while (budget > 0) {
1759                 int len;
1760                 struct sk_buff *skb;
1761                 u16 ctrl = dev->rx_desc[slot].ctrl;
1762
1763                 if (ctrl & MAL_RX_CTRL_EMPTY)
1764                         break;
1765
1766                 skb = dev->rx_skb[slot];
1767                 mb();
1768                 len = dev->rx_desc[slot].data_len;
1769
1770                 if (unlikely(!MAL_IS_SINGLE_RX(ctrl)))
1771                         goto sg;
1772
1773                 ctrl &= EMAC_BAD_RX_MASK;
1774                 if (unlikely(ctrl && ctrl != EMAC_RX_TAH_BAD_CSUM)) {
1775                         emac_parse_rx_error(dev, ctrl);
1776                         ++dev->estats.rx_dropped_error;
1777                         emac_recycle_rx_skb(dev, slot, 0);
1778                         len = 0;
1779                         goto next;
1780                 }
1781
1782                 if (len < ETH_HLEN) {
1783                         ++dev->estats.rx_dropped_stack;
1784                         emac_recycle_rx_skb(dev, slot, len);
1785                         goto next;
1786                 }
1787
1788                 if (len && len < EMAC_RX_COPY_THRESH) {
1789                         struct sk_buff *copy_skb;
1790
1791                         copy_skb = napi_alloc_skb(&dev->mal->napi, len);
1792                         if (unlikely(!copy_skb))
1793                                 goto oom;
1794
1795                         memcpy(copy_skb->data - NET_IP_ALIGN,
1796                                skb->data - NET_IP_ALIGN,
1797                                len + NET_IP_ALIGN);
1798                         emac_recycle_rx_skb(dev, slot, len);
1799                         skb = copy_skb;
1800                 } else if (unlikely(emac_alloc_rx_skb_napi(dev, slot)))
1801                         goto oom;
1802
1803                 skb_put(skb, len);
1804         push_packet:
1805                 skb->protocol = eth_type_trans(skb, dev->ndev);
1806                 emac_rx_csum(dev, skb, ctrl);
1807
1808                 if (unlikely(netif_receive_skb(skb) == NET_RX_DROP))
1809                         ++dev->estats.rx_dropped_stack;
1810         next:
1811                 ++dev->stats.rx_packets;
1812         skip:
1813                 dev->stats.rx_bytes += len;
1814                 slot = (slot + 1) % NUM_RX_BUFF;
1815                 --budget;
1816                 ++received;
1817                 continue;
1818         sg:
1819                 if (ctrl & MAL_RX_CTRL_FIRST) {
1820                         BUG_ON(dev->rx_sg_skb);
1821                         if (unlikely(emac_alloc_rx_skb_napi(dev, slot))) {
1822                                 DBG(dev, "rx OOM %d" NL, slot);
1823                                 ++dev->estats.rx_dropped_oom;
1824                                 emac_recycle_rx_skb(dev, slot, 0);
1825                         } else {
1826                                 dev->rx_sg_skb = skb;
1827                                 skb_put(skb, len);
1828                         }
1829                 } else if (!emac_rx_sg_append(dev, slot) &&
1830                            (ctrl & MAL_RX_CTRL_LAST)) {
1831
1832                         skb = dev->rx_sg_skb;
1833                         dev->rx_sg_skb = NULL;
1834
1835                         ctrl &= EMAC_BAD_RX_MASK;
1836                         if (unlikely(ctrl && ctrl != EMAC_RX_TAH_BAD_CSUM)) {
1837                                 emac_parse_rx_error(dev, ctrl);
1838                                 ++dev->estats.rx_dropped_error;
1839                                 dev_kfree_skb(skb);
1840                                 len = 0;
1841                         } else
1842                                 goto push_packet;
1843                 }
1844                 goto skip;
1845         oom:
1846                 DBG(dev, "rx OOM %d" NL, slot);
1847                 /* Drop the packet and recycle skb */
1848                 ++dev->estats.rx_dropped_oom;
1849                 emac_recycle_rx_skb(dev, slot, 0);
1850                 goto next;
1851         }
1852
1853         if (received) {
1854                 DBG2(dev, "rx %d BDs" NL, received);
1855                 dev->rx_slot = slot;
1856         }
1857
1858         if (unlikely(budget && test_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags))) {
1859                 mb();
1860                 if (!(dev->rx_desc[slot].ctrl & MAL_RX_CTRL_EMPTY)) {
1861                         DBG2(dev, "rx restart" NL);
1862                         received = 0;
1863                         goto again;
1864                 }
1865
1866                 if (dev->rx_sg_skb) {
1867                         DBG2(dev, "dropping partial rx packet" NL);
1868                         ++dev->estats.rx_dropped_error;
1869                         dev_kfree_skb(dev->rx_sg_skb);
1870                         dev->rx_sg_skb = NULL;
1871                 }
1872
1873                 clear_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1874                 mal_enable_rx_channel(dev->mal, dev->mal_rx_chan);
1875                 emac_rx_enable(dev);
1876                 dev->rx_slot = 0;
1877         }
1878         return received;
1879 }
1880
1881 /* NAPI poll context */
1882 static int emac_peek_rx(void *param)
1883 {
1884         struct emac_instance *dev = param;
1885
1886         return !(dev->rx_desc[dev->rx_slot].ctrl & MAL_RX_CTRL_EMPTY);
1887 }
1888
1889 /* NAPI poll context */
1890 static int emac_peek_rx_sg(void *param)
1891 {
1892         struct emac_instance *dev = param;
1893
1894         int slot = dev->rx_slot;
1895         while (1) {
1896                 u16 ctrl = dev->rx_desc[slot].ctrl;
1897                 if (ctrl & MAL_RX_CTRL_EMPTY)
1898                         return 0;
1899                 else if (ctrl & MAL_RX_CTRL_LAST)
1900                         return 1;
1901
1902                 slot = (slot + 1) % NUM_RX_BUFF;
1903
1904                 /* I'm just being paranoid here :) */
1905                 if (unlikely(slot == dev->rx_slot))
1906                         return 0;
1907         }
1908 }
1909
1910 /* Hard IRQ */
1911 static void emac_rxde(void *param)
1912 {
1913         struct emac_instance *dev = param;
1914
1915         ++dev->estats.rx_stopped;
1916         emac_rx_disable_async(dev);
1917 }
1918
1919 /* Hard IRQ */
1920 static irqreturn_t emac_irq(int irq, void *dev_instance)
1921 {
1922         struct emac_instance *dev = dev_instance;
1923         struct emac_regs __iomem *p = dev->emacp;
1924         struct emac_error_stats *st = &dev->estats;
1925         u32 isr;
1926
1927         spin_lock(&dev->lock);
1928
1929         isr = in_be32(&p->isr);
1930         out_be32(&p->isr, isr);
1931
1932         DBG(dev, "isr = %08x" NL, isr);
1933
1934         if (isr & EMAC4_ISR_TXPE)
1935                 ++st->tx_parity;
1936         if (isr & EMAC4_ISR_RXPE)
1937                 ++st->rx_parity;
1938         if (isr & EMAC4_ISR_TXUE)
1939                 ++st->tx_underrun;
1940         if (isr & EMAC4_ISR_RXOE)
1941                 ++st->rx_fifo_overrun;
1942         if (isr & EMAC_ISR_OVR)
1943                 ++st->rx_overrun;
1944         if (isr & EMAC_ISR_BP)
1945                 ++st->rx_bad_packet;
1946         if (isr & EMAC_ISR_RP)
1947                 ++st->rx_runt_packet;
1948         if (isr & EMAC_ISR_SE)
1949                 ++st->rx_short_event;
1950         if (isr & EMAC_ISR_ALE)
1951                 ++st->rx_alignment_error;
1952         if (isr & EMAC_ISR_BFCS)
1953                 ++st->rx_bad_fcs;
1954         if (isr & EMAC_ISR_PTLE)
1955                 ++st->rx_packet_too_long;
1956         if (isr & EMAC_ISR_ORE)
1957                 ++st->rx_out_of_range;
1958         if (isr & EMAC_ISR_IRE)
1959                 ++st->rx_in_range;
1960         if (isr & EMAC_ISR_SQE)
1961                 ++st->tx_sqe;
1962         if (isr & EMAC_ISR_TE)
1963                 ++st->tx_errors;
1964
1965         spin_unlock(&dev->lock);
1966
1967         return IRQ_HANDLED;
1968 }
1969
1970 static struct net_device_stats *emac_stats(struct net_device *ndev)
1971 {
1972         struct emac_instance *dev = netdev_priv(ndev);
1973         struct emac_stats *st = &dev->stats;
1974         struct emac_error_stats *est = &dev->estats;
1975         struct net_device_stats *nst = &ndev->stats;
1976         unsigned long flags;
1977
1978         DBG2(dev, "stats" NL);
1979
1980         /* Compute "legacy" statistics */
1981         spin_lock_irqsave(&dev->lock, flags);
1982         nst->rx_packets = (unsigned long)st->rx_packets;
1983         nst->rx_bytes = (unsigned long)st->rx_bytes;
1984         nst->tx_packets = (unsigned long)st->tx_packets;
1985         nst->tx_bytes = (unsigned long)st->tx_bytes;
1986         nst->rx_dropped = (unsigned long)(est->rx_dropped_oom +
1987                                           est->rx_dropped_error +
1988                                           est->rx_dropped_resize +
1989                                           est->rx_dropped_mtu);
1990         nst->tx_dropped = (unsigned long)est->tx_dropped;
1991
1992         nst->rx_errors = (unsigned long)est->rx_bd_errors;
1993         nst->rx_fifo_errors = (unsigned long)(est->rx_bd_overrun +
1994                                               est->rx_fifo_overrun +
1995                                               est->rx_overrun);
1996         nst->rx_frame_errors = (unsigned long)(est->rx_bd_alignment_error +
1997                                                est->rx_alignment_error);
1998         nst->rx_crc_errors = (unsigned long)(est->rx_bd_bad_fcs +
1999                                              est->rx_bad_fcs);
2000         nst->rx_length_errors = (unsigned long)(est->rx_bd_runt_packet +
2001                                                 est->rx_bd_short_event +
2002                                                 est->rx_bd_packet_too_long +
2003                                                 est->rx_bd_out_of_range +
2004                                                 est->rx_bd_in_range +
2005                                                 est->rx_runt_packet +
2006                                                 est->rx_short_event +
2007                                                 est->rx_packet_too_long +
2008                                                 est->rx_out_of_range +
2009                                                 est->rx_in_range);
2010
2011         nst->tx_errors = (unsigned long)(est->tx_bd_errors + est->tx_errors);
2012         nst->tx_fifo_errors = (unsigned long)(est->tx_bd_underrun +
2013                                               est->tx_underrun);
2014         nst->tx_carrier_errors = (unsigned long)est->tx_bd_carrier_loss;
2015         nst->collisions = (unsigned long)(est->tx_bd_excessive_deferral +
2016                                           est->tx_bd_excessive_collisions +
2017                                           est->tx_bd_late_collision +
2018                                           est->tx_bd_multple_collisions);
2019         spin_unlock_irqrestore(&dev->lock, flags);
2020         return nst;
2021 }
2022
2023 static struct mal_commac_ops emac_commac_ops = {
2024         .poll_tx = &emac_poll_tx,
2025         .poll_rx = &emac_poll_rx,
2026         .peek_rx = &emac_peek_rx,
2027         .rxde = &emac_rxde,
2028 };
2029
2030 static struct mal_commac_ops emac_commac_sg_ops = {
2031         .poll_tx = &emac_poll_tx,
2032         .poll_rx = &emac_poll_rx,
2033         .peek_rx = &emac_peek_rx_sg,
2034         .rxde = &emac_rxde,
2035 };
2036
2037 /* Ethtool support */
2038 static int emac_ethtool_get_link_ksettings(struct net_device *ndev,
2039                                            struct ethtool_link_ksettings *cmd)
2040 {
2041         struct emac_instance *dev = netdev_priv(ndev);
2042         u32 supported, advertising;
2043
2044         supported = dev->phy.features;
2045         cmd->base.port = PORT_MII;
2046         cmd->base.phy_address = dev->phy.address;
2047
2048         mutex_lock(&dev->link_lock);
2049         advertising = dev->phy.advertising;
2050         cmd->base.autoneg = dev->phy.autoneg;
2051         cmd->base.speed = dev->phy.speed;
2052         cmd->base.duplex = dev->phy.duplex;
2053         mutex_unlock(&dev->link_lock);
2054
2055         ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.supported,
2056                                                 supported);
2057         ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.advertising,
2058                                                 advertising);
2059
2060         return 0;
2061 }
2062
2063 static int
2064 emac_ethtool_set_link_ksettings(struct net_device *ndev,
2065                                 const struct ethtool_link_ksettings *cmd)
2066 {
2067         struct emac_instance *dev = netdev_priv(ndev);
2068         u32 f = dev->phy.features;
2069         u32 advertising;
2070
2071         ethtool_convert_link_mode_to_legacy_u32(&advertising,
2072                                                 cmd->link_modes.advertising);
2073
2074         DBG(dev, "set_settings(%d, %d, %d, 0x%08x)" NL,
2075             cmd->base.autoneg, cmd->base.speed, cmd->base.duplex, advertising);
2076
2077         /* Basic sanity checks */
2078         if (dev->phy.address < 0)
2079                 return -EOPNOTSUPP;
2080         if (cmd->base.autoneg != AUTONEG_ENABLE &&
2081             cmd->base.autoneg != AUTONEG_DISABLE)
2082                 return -EINVAL;
2083         if (cmd->base.autoneg == AUTONEG_ENABLE && advertising == 0)
2084                 return -EINVAL;
2085         if (cmd->base.duplex != DUPLEX_HALF && cmd->base.duplex != DUPLEX_FULL)
2086                 return -EINVAL;
2087
2088         if (cmd->base.autoneg == AUTONEG_DISABLE) {
2089                 switch (cmd->base.speed) {
2090                 case SPEED_10:
2091                         if (cmd->base.duplex == DUPLEX_HALF &&
2092                             !(f & SUPPORTED_10baseT_Half))
2093                                 return -EINVAL;
2094                         if (cmd->base.duplex == DUPLEX_FULL &&
2095                             !(f & SUPPORTED_10baseT_Full))
2096                                 return -EINVAL;
2097                         break;
2098                 case SPEED_100:
2099                         if (cmd->base.duplex == DUPLEX_HALF &&
2100                             !(f & SUPPORTED_100baseT_Half))
2101                                 return -EINVAL;
2102                         if (cmd->base.duplex == DUPLEX_FULL &&
2103                             !(f & SUPPORTED_100baseT_Full))
2104                                 return -EINVAL;
2105                         break;
2106                 case SPEED_1000:
2107                         if (cmd->base.duplex == DUPLEX_HALF &&
2108                             !(f & SUPPORTED_1000baseT_Half))
2109                                 return -EINVAL;
2110                         if (cmd->base.duplex == DUPLEX_FULL &&
2111                             !(f & SUPPORTED_1000baseT_Full))
2112                                 return -EINVAL;
2113                         break;
2114                 default:
2115                         return -EINVAL;
2116                 }
2117
2118                 mutex_lock(&dev->link_lock);
2119                 dev->phy.def->ops->setup_forced(&dev->phy, cmd->base.speed,
2120                                                 cmd->base.duplex);
2121                 mutex_unlock(&dev->link_lock);
2122
2123         } else {
2124                 if (!(f & SUPPORTED_Autoneg))
2125                         return -EINVAL;
2126
2127                 mutex_lock(&dev->link_lock);
2128                 dev->phy.def->ops->setup_aneg(&dev->phy,
2129                                               (advertising & f) |
2130                                               (dev->phy.advertising &
2131                                                (ADVERTISED_Pause |
2132                                                 ADVERTISED_Asym_Pause)));
2133                 mutex_unlock(&dev->link_lock);
2134         }
2135         emac_force_link_update(dev);
2136
2137         return 0;
2138 }
2139
2140 static void emac_ethtool_get_ringparam(struct net_device *ndev,
2141                                        struct ethtool_ringparam *rp)
2142 {
2143         rp->rx_max_pending = rp->rx_pending = NUM_RX_BUFF;
2144         rp->tx_max_pending = rp->tx_pending = NUM_TX_BUFF;
2145 }
2146
2147 static void emac_ethtool_get_pauseparam(struct net_device *ndev,
2148                                         struct ethtool_pauseparam *pp)
2149 {
2150         struct emac_instance *dev = netdev_priv(ndev);
2151
2152         mutex_lock(&dev->link_lock);
2153         if ((dev->phy.features & SUPPORTED_Autoneg) &&
2154             (dev->phy.advertising & (ADVERTISED_Pause | ADVERTISED_Asym_Pause)))
2155                 pp->autoneg = 1;
2156
2157         if (dev->phy.duplex == DUPLEX_FULL) {
2158                 if (dev->phy.pause)
2159                         pp->rx_pause = pp->tx_pause = 1;
2160                 else if (dev->phy.asym_pause)
2161                         pp->tx_pause = 1;
2162         }
2163         mutex_unlock(&dev->link_lock);
2164 }
2165
2166 static int emac_get_regs_len(struct emac_instance *dev)
2167 {
2168                 return sizeof(struct emac_ethtool_regs_subhdr) +
2169                         sizeof(struct emac_regs);
2170 }
2171
2172 static int emac_ethtool_get_regs_len(struct net_device *ndev)
2173 {
2174         struct emac_instance *dev = netdev_priv(ndev);
2175         int size;
2176
2177         size = sizeof(struct emac_ethtool_regs_hdr) +
2178                 emac_get_regs_len(dev) + mal_get_regs_len(dev->mal);
2179         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
2180                 size += zmii_get_regs_len(dev->zmii_dev);
2181         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
2182                 size += rgmii_get_regs_len(dev->rgmii_dev);
2183         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
2184                 size += tah_get_regs_len(dev->tah_dev);
2185
2186         return size;
2187 }
2188
2189 static void *emac_dump_regs(struct emac_instance *dev, void *buf)
2190 {
2191         struct emac_ethtool_regs_subhdr *hdr = buf;
2192
2193         hdr->index = dev->cell_index;
2194         if (emac_has_feature(dev, EMAC_FTR_EMAC4SYNC)) {
2195                 hdr->version = EMAC4SYNC_ETHTOOL_REGS_VER;
2196         } else if (emac_has_feature(dev, EMAC_FTR_EMAC4)) {
2197                 hdr->version = EMAC4_ETHTOOL_REGS_VER;
2198         } else {
2199                 hdr->version = EMAC_ETHTOOL_REGS_VER;
2200         }
2201         memcpy_fromio(hdr + 1, dev->emacp, sizeof(struct emac_regs));
2202         return (void *)(hdr + 1) + sizeof(struct emac_regs);
2203 }
2204
2205 static void emac_ethtool_get_regs(struct net_device *ndev,
2206                                   struct ethtool_regs *regs, void *buf)
2207 {
2208         struct emac_instance *dev = netdev_priv(ndev);
2209         struct emac_ethtool_regs_hdr *hdr = buf;
2210
2211         hdr->components = 0;
2212         buf = hdr + 1;
2213
2214         buf = mal_dump_regs(dev->mal, buf);
2215         buf = emac_dump_regs(dev, buf);
2216         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII)) {
2217                 hdr->components |= EMAC_ETHTOOL_REGS_ZMII;
2218                 buf = zmii_dump_regs(dev->zmii_dev, buf);
2219         }
2220         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII)) {
2221                 hdr->components |= EMAC_ETHTOOL_REGS_RGMII;
2222                 buf = rgmii_dump_regs(dev->rgmii_dev, buf);
2223         }
2224         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH)) {
2225                 hdr->components |= EMAC_ETHTOOL_REGS_TAH;
2226                 buf = tah_dump_regs(dev->tah_dev, buf);
2227         }
2228 }
2229
2230 static int emac_ethtool_nway_reset(struct net_device *ndev)
2231 {
2232         struct emac_instance *dev = netdev_priv(ndev);
2233         int res = 0;
2234
2235         DBG(dev, "nway_reset" NL);
2236
2237         if (dev->phy.address < 0)
2238                 return -EOPNOTSUPP;
2239
2240         mutex_lock(&dev->link_lock);
2241         if (!dev->phy.autoneg) {
2242                 res = -EINVAL;
2243                 goto out;
2244         }
2245
2246         dev->phy.def->ops->setup_aneg(&dev->phy, dev->phy.advertising);
2247  out:
2248         mutex_unlock(&dev->link_lock);
2249         emac_force_link_update(dev);
2250         return res;
2251 }
2252
2253 static int emac_ethtool_get_sset_count(struct net_device *ndev, int stringset)
2254 {
2255         if (stringset == ETH_SS_STATS)
2256                 return EMAC_ETHTOOL_STATS_COUNT;
2257         else
2258                 return -EINVAL;
2259 }
2260
2261 static void emac_ethtool_get_strings(struct net_device *ndev, u32 stringset,
2262                                      u8 * buf)
2263 {
2264         if (stringset == ETH_SS_STATS)
2265                 memcpy(buf, &emac_stats_keys, sizeof(emac_stats_keys));
2266 }
2267
2268 static void emac_ethtool_get_ethtool_stats(struct net_device *ndev,
2269                                            struct ethtool_stats *estats,
2270                                            u64 * tmp_stats)
2271 {
2272         struct emac_instance *dev = netdev_priv(ndev);
2273
2274         memcpy(tmp_stats, &dev->stats, sizeof(dev->stats));
2275         tmp_stats += sizeof(dev->stats) / sizeof(u64);
2276         memcpy(tmp_stats, &dev->estats, sizeof(dev->estats));
2277 }
2278
2279 static void emac_ethtool_get_drvinfo(struct net_device *ndev,
2280                                      struct ethtool_drvinfo *info)
2281 {
2282         struct emac_instance *dev = netdev_priv(ndev);
2283
2284         strlcpy(info->driver, "ibm_emac", sizeof(info->driver));
2285         strlcpy(info->version, DRV_VERSION, sizeof(info->version));
2286         snprintf(info->bus_info, sizeof(info->bus_info), "PPC 4xx EMAC-%d %pOF",
2287                  dev->cell_index, dev->ofdev->dev.of_node);
2288 }
2289
2290 static const struct ethtool_ops emac_ethtool_ops = {
2291         .get_drvinfo = emac_ethtool_get_drvinfo,
2292
2293         .get_regs_len = emac_ethtool_get_regs_len,
2294         .get_regs = emac_ethtool_get_regs,
2295
2296         .nway_reset = emac_ethtool_nway_reset,
2297
2298         .get_ringparam = emac_ethtool_get_ringparam,
2299         .get_pauseparam = emac_ethtool_get_pauseparam,
2300
2301         .get_strings = emac_ethtool_get_strings,
2302         .get_sset_count = emac_ethtool_get_sset_count,
2303         .get_ethtool_stats = emac_ethtool_get_ethtool_stats,
2304
2305         .get_link = ethtool_op_get_link,
2306         .get_link_ksettings = emac_ethtool_get_link_ksettings,
2307         .set_link_ksettings = emac_ethtool_set_link_ksettings,
2308 };
2309
2310 static int emac_ioctl(struct net_device *ndev, struct ifreq *rq, int cmd)
2311 {
2312         struct emac_instance *dev = netdev_priv(ndev);
2313         struct mii_ioctl_data *data = if_mii(rq);
2314
2315         DBG(dev, "ioctl %08x" NL, cmd);
2316
2317         if (dev->phy.address < 0)
2318                 return -EOPNOTSUPP;
2319
2320         switch (cmd) {
2321         case SIOCGMIIPHY:
2322                 data->phy_id = dev->phy.address;
2323                 /* Fall through */
2324         case SIOCGMIIREG:
2325                 data->val_out = emac_mdio_read(ndev, dev->phy.address,
2326                                                data->reg_num);
2327                 return 0;
2328
2329         case SIOCSMIIREG:
2330                 emac_mdio_write(ndev, dev->phy.address, data->reg_num,
2331                                 data->val_in);
2332                 return 0;
2333         default:
2334                 return -EOPNOTSUPP;
2335         }
2336 }
2337
2338 struct emac_depentry {
2339         u32                     phandle;
2340         struct device_node      *node;
2341         struct platform_device  *ofdev;
2342         void                    *drvdata;
2343 };
2344
2345 #define EMAC_DEP_MAL_IDX        0
2346 #define EMAC_DEP_ZMII_IDX       1
2347 #define EMAC_DEP_RGMII_IDX      2
2348 #define EMAC_DEP_TAH_IDX        3
2349 #define EMAC_DEP_MDIO_IDX       4
2350 #define EMAC_DEP_PREV_IDX       5
2351 #define EMAC_DEP_COUNT          6
2352
2353 static int emac_check_deps(struct emac_instance *dev,
2354                            struct emac_depentry *deps)
2355 {
2356         int i, there = 0;
2357         struct device_node *np;
2358
2359         for (i = 0; i < EMAC_DEP_COUNT; i++) {
2360                 /* no dependency on that item, allright */
2361                 if (deps[i].phandle == 0) {
2362                         there++;
2363                         continue;
2364                 }
2365                 /* special case for blist as the dependency might go away */
2366                 if (i == EMAC_DEP_PREV_IDX) {
2367                         np = *(dev->blist - 1);
2368                         if (np == NULL) {
2369                                 deps[i].phandle = 0;
2370                                 there++;
2371                                 continue;
2372                         }
2373                         if (deps[i].node == NULL)
2374                                 deps[i].node = of_node_get(np);
2375                 }
2376                 if (deps[i].node == NULL)
2377                         deps[i].node = of_find_node_by_phandle(deps[i].phandle);
2378                 if (deps[i].node == NULL)
2379                         continue;
2380                 if (deps[i].ofdev == NULL)
2381                         deps[i].ofdev = of_find_device_by_node(deps[i].node);
2382                 if (deps[i].ofdev == NULL)
2383                         continue;
2384                 if (deps[i].drvdata == NULL)
2385                         deps[i].drvdata = platform_get_drvdata(deps[i].ofdev);
2386                 if (deps[i].drvdata != NULL)
2387                         there++;
2388         }
2389         return there == EMAC_DEP_COUNT;
2390 }
2391
2392 static void emac_put_deps(struct emac_instance *dev)
2393 {
2394         of_dev_put(dev->mal_dev);
2395         of_dev_put(dev->zmii_dev);
2396         of_dev_put(dev->rgmii_dev);
2397         of_dev_put(dev->mdio_dev);
2398         of_dev_put(dev->tah_dev);
2399 }
2400
2401 static int emac_of_bus_notify(struct notifier_block *nb, unsigned long action,
2402                               void *data)
2403 {
2404         /* We are only intereted in device addition */
2405         if (action == BUS_NOTIFY_BOUND_DRIVER)
2406                 wake_up_all(&emac_probe_wait);
2407         return 0;
2408 }
2409
2410 static struct notifier_block emac_of_bus_notifier = {
2411         .notifier_call = emac_of_bus_notify
2412 };
2413
2414 static int emac_wait_deps(struct emac_instance *dev)
2415 {
2416         struct emac_depentry deps[EMAC_DEP_COUNT];
2417         int i, err;
2418
2419         memset(&deps, 0, sizeof(deps));
2420
2421         deps[EMAC_DEP_MAL_IDX].phandle = dev->mal_ph;
2422         deps[EMAC_DEP_ZMII_IDX].phandle = dev->zmii_ph;
2423         deps[EMAC_DEP_RGMII_IDX].phandle = dev->rgmii_ph;
2424         if (dev->tah_ph)
2425                 deps[EMAC_DEP_TAH_IDX].phandle = dev->tah_ph;
2426         if (dev->mdio_ph)
2427                 deps[EMAC_DEP_MDIO_IDX].phandle = dev->mdio_ph;
2428         if (dev->blist && dev->blist > emac_boot_list)
2429                 deps[EMAC_DEP_PREV_IDX].phandle = 0xffffffffu;
2430         bus_register_notifier(&platform_bus_type, &emac_of_bus_notifier);
2431         wait_event_timeout(emac_probe_wait,
2432                            emac_check_deps(dev, deps),
2433                            EMAC_PROBE_DEP_TIMEOUT);
2434         bus_unregister_notifier(&platform_bus_type, &emac_of_bus_notifier);
2435         err = emac_check_deps(dev, deps) ? 0 : -ENODEV;
2436         for (i = 0; i < EMAC_DEP_COUNT; i++) {
2437                 of_node_put(deps[i].node);
2438                 if (err)
2439                         of_dev_put(deps[i].ofdev);
2440         }
2441         if (err == 0) {
2442                 dev->mal_dev = deps[EMAC_DEP_MAL_IDX].ofdev;
2443                 dev->zmii_dev = deps[EMAC_DEP_ZMII_IDX].ofdev;
2444                 dev->rgmii_dev = deps[EMAC_DEP_RGMII_IDX].ofdev;
2445                 dev->tah_dev = deps[EMAC_DEP_TAH_IDX].ofdev;
2446                 dev->mdio_dev = deps[EMAC_DEP_MDIO_IDX].ofdev;
2447         }
2448         of_dev_put(deps[EMAC_DEP_PREV_IDX].ofdev);
2449         return err;
2450 }
2451
2452 static int emac_read_uint_prop(struct device_node *np, const char *name,
2453                                u32 *val, int fatal)
2454 {
2455         int len;
2456         const u32 *prop = of_get_property(np, name, &len);
2457         if (prop == NULL || len < sizeof(u32)) {
2458                 if (fatal)
2459                         printk(KERN_ERR "%pOF: missing %s property\n",
2460                                np, name);
2461                 return -ENODEV;
2462         }
2463         *val = *prop;
2464         return 0;
2465 }
2466
2467 static void emac_adjust_link(struct net_device *ndev)
2468 {
2469         struct emac_instance *dev = netdev_priv(ndev);
2470         struct phy_device *phy = dev->phy_dev;
2471
2472         dev->phy.autoneg = phy->autoneg;
2473         dev->phy.speed = phy->speed;
2474         dev->phy.duplex = phy->duplex;
2475         dev->phy.pause = phy->pause;
2476         dev->phy.asym_pause = phy->asym_pause;
2477         ethtool_convert_link_mode_to_legacy_u32(&dev->phy.advertising,
2478                                                 phy->advertising);
2479 }
2480
2481 static int emac_mii_bus_read(struct mii_bus *bus, int addr, int regnum)
2482 {
2483         int ret = emac_mdio_read(bus->priv, addr, regnum);
2484         /* This is a workaround for powered down ports/phys.
2485          * In the wild, this was seen on the Cisco Meraki MX60(W).
2486          * This hardware disables ports as part of the handoff
2487          * procedure. Accessing the ports will lead to errors
2488          * (-ETIMEDOUT, -EREMOTEIO) that do more harm than good.
2489          */
2490         return ret < 0 ? 0xffff : ret;
2491 }
2492
2493 static int emac_mii_bus_write(struct mii_bus *bus, int addr,
2494                               int regnum, u16 val)
2495 {
2496         emac_mdio_write(bus->priv, addr, regnum, val);
2497         return 0;
2498 }
2499
2500 static int emac_mii_bus_reset(struct mii_bus *bus)
2501 {
2502         struct emac_instance *dev = netdev_priv(bus->priv);
2503
2504         return emac_reset(dev);
2505 }
2506
2507 static int emac_mdio_phy_start_aneg(struct mii_phy *phy,
2508                                     struct phy_device *phy_dev)
2509 {
2510         phy_dev->autoneg = phy->autoneg;
2511         phy_dev->speed = phy->speed;
2512         phy_dev->duplex = phy->duplex;
2513         ethtool_convert_legacy_u32_to_link_mode(phy_dev->advertising,
2514                                                 phy->advertising);
2515         return phy_start_aneg(phy_dev);
2516 }
2517
2518 static int emac_mdio_setup_aneg(struct mii_phy *phy, u32 advertise)
2519 {
2520         struct net_device *ndev = phy->dev;
2521         struct emac_instance *dev = netdev_priv(ndev);
2522
2523         phy->autoneg = AUTONEG_ENABLE;
2524         phy->advertising = advertise;
2525         return emac_mdio_phy_start_aneg(phy, dev->phy_dev);
2526 }
2527
2528 static int emac_mdio_setup_forced(struct mii_phy *phy, int speed, int fd)
2529 {
2530         struct net_device *ndev = phy->dev;
2531         struct emac_instance *dev = netdev_priv(ndev);
2532
2533         phy->autoneg = AUTONEG_DISABLE;
2534         phy->speed = speed;
2535         phy->duplex = fd;
2536         return emac_mdio_phy_start_aneg(phy, dev->phy_dev);
2537 }
2538
2539 static int emac_mdio_poll_link(struct mii_phy *phy)
2540 {
2541         struct net_device *ndev = phy->dev;
2542         struct emac_instance *dev = netdev_priv(ndev);
2543         int res;
2544
2545         res = phy_read_status(dev->phy_dev);
2546         if (res) {
2547                 dev_err(&dev->ofdev->dev, "link update failed (%d).", res);
2548                 return ethtool_op_get_link(ndev);
2549         }
2550
2551         return dev->phy_dev->link;
2552 }
2553
2554 static int emac_mdio_read_link(struct mii_phy *phy)
2555 {
2556         struct net_device *ndev = phy->dev;
2557         struct emac_instance *dev = netdev_priv(ndev);
2558         struct phy_device *phy_dev = dev->phy_dev;
2559         int res;
2560
2561         res = phy_read_status(phy_dev);
2562         if (res)
2563                 return res;
2564
2565         phy->speed = phy_dev->speed;
2566         phy->duplex = phy_dev->duplex;
2567         phy->pause = phy_dev->pause;
2568         phy->asym_pause = phy_dev->asym_pause;
2569         return 0;
2570 }
2571
2572 static int emac_mdio_init_phy(struct mii_phy *phy)
2573 {
2574         struct net_device *ndev = phy->dev;
2575         struct emac_instance *dev = netdev_priv(ndev);
2576
2577         phy_start(dev->phy_dev);
2578         return phy_init_hw(dev->phy_dev);
2579 }
2580
2581 static const struct mii_phy_ops emac_dt_mdio_phy_ops = {
2582         .init           = emac_mdio_init_phy,
2583         .setup_aneg     = emac_mdio_setup_aneg,
2584         .setup_forced   = emac_mdio_setup_forced,
2585         .poll_link      = emac_mdio_poll_link,
2586         .read_link      = emac_mdio_read_link,
2587 };
2588
2589 static int emac_dt_mdio_probe(struct emac_instance *dev)
2590 {
2591         struct device_node *mii_np;
2592         int res;
2593
2594         mii_np = of_get_child_by_name(dev->ofdev->dev.of_node, "mdio");
2595         if (!mii_np) {
2596                 dev_err(&dev->ofdev->dev, "no mdio definition found.");
2597                 return -ENODEV;
2598         }
2599
2600         if (!of_device_is_available(mii_np)) {
2601                 res = -ENODEV;
2602                 goto put_node;
2603         }
2604
2605         dev->mii_bus = devm_mdiobus_alloc(&dev->ofdev->dev);
2606         if (!dev->mii_bus) {
2607                 res = -ENOMEM;
2608                 goto put_node;
2609         }
2610
2611         dev->mii_bus->priv = dev->ndev;
2612         dev->mii_bus->parent = dev->ndev->dev.parent;
2613         dev->mii_bus->name = "emac_mdio";
2614         dev->mii_bus->read = &emac_mii_bus_read;
2615         dev->mii_bus->write = &emac_mii_bus_write;
2616         dev->mii_bus->reset = &emac_mii_bus_reset;
2617         snprintf(dev->mii_bus->id, MII_BUS_ID_SIZE, "%s", dev->ofdev->name);
2618         res = of_mdiobus_register(dev->mii_bus, mii_np);
2619         if (res) {
2620                 dev_err(&dev->ofdev->dev, "cannot register MDIO bus %s (%d)",
2621                         dev->mii_bus->name, res);
2622         }
2623
2624  put_node:
2625         of_node_put(mii_np);
2626         return res;
2627 }
2628
2629 static int emac_dt_phy_connect(struct emac_instance *dev,
2630                                struct device_node *phy_handle)
2631 {
2632         dev->phy.def = devm_kzalloc(&dev->ofdev->dev, sizeof(*dev->phy.def),
2633                                     GFP_KERNEL);
2634         if (!dev->phy.def)
2635                 return -ENOMEM;
2636
2637         dev->phy_dev = of_phy_connect(dev->ndev, phy_handle, &emac_adjust_link,
2638                                       0, dev->phy_mode);
2639         if (!dev->phy_dev) {
2640                 dev_err(&dev->ofdev->dev, "failed to connect to PHY.\n");
2641                 return -ENODEV;
2642         }
2643
2644         dev->phy.def->phy_id = dev->phy_dev->drv->phy_id;
2645         dev->phy.def->phy_id_mask = dev->phy_dev->drv->phy_id_mask;
2646         dev->phy.def->name = dev->phy_dev->drv->name;
2647         dev->phy.def->ops = &emac_dt_mdio_phy_ops;
2648         ethtool_convert_link_mode_to_legacy_u32(&dev->phy.features,
2649                                                 dev->phy_dev->supported);
2650         dev->phy.address = dev->phy_dev->mdio.addr;
2651         dev->phy.mode = dev->phy_dev->interface;
2652         return 0;
2653 }
2654
2655 static int emac_dt_phy_probe(struct emac_instance *dev)
2656 {
2657         struct device_node *np = dev->ofdev->dev.of_node;
2658         struct device_node *phy_handle;
2659         int res = 1;
2660
2661         phy_handle = of_parse_phandle(np, "phy-handle", 0);
2662
2663         if (phy_handle) {
2664                 res = emac_dt_mdio_probe(dev);
2665                 if (!res) {
2666                         res = emac_dt_phy_connect(dev, phy_handle);
2667                         if (res)
2668                                 mdiobus_unregister(dev->mii_bus);
2669                 }
2670         }
2671
2672         of_node_put(phy_handle);
2673         return res;
2674 }
2675
2676 static int emac_init_phy(struct emac_instance *dev)
2677 {
2678         struct device_node *np = dev->ofdev->dev.of_node;
2679         struct net_device *ndev = dev->ndev;
2680         u32 phy_map, adv;
2681         int i;
2682
2683         dev->phy.dev = ndev;
2684         dev->phy.mode = dev->phy_mode;
2685
2686         /* PHY-less configuration. */
2687         if ((dev->phy_address == 0xffffffff && dev->phy_map == 0xffffffff) ||
2688             of_phy_is_fixed_link(np)) {
2689                 emac_reset(dev);
2690
2691                 /* PHY-less configuration. */
2692                 dev->phy.address = -1;
2693                 dev->phy.features = SUPPORTED_MII;
2694                 if (emac_phy_supports_gige(dev->phy_mode))
2695                         dev->phy.features |= SUPPORTED_1000baseT_Full;
2696                 else
2697                         dev->phy.features |= SUPPORTED_100baseT_Full;
2698                 dev->phy.pause = 1;
2699
2700                 if (of_phy_is_fixed_link(np)) {
2701                         int res = emac_dt_mdio_probe(dev);
2702
2703                         if (res)
2704                                 return res;
2705
2706                         res = of_phy_register_fixed_link(np);
2707                         dev->phy_dev = of_phy_find_device(np);
2708                         if (res || !dev->phy_dev) {
2709                                 mdiobus_unregister(dev->mii_bus);
2710                                 return res ? res : -EINVAL;
2711                         }
2712                         emac_adjust_link(dev->ndev);
2713                         put_device(&dev->phy_dev->mdio.dev);
2714                 }
2715                 return 0;
2716         }
2717
2718         mutex_lock(&emac_phy_map_lock);
2719         phy_map = dev->phy_map | busy_phy_map;
2720
2721         DBG(dev, "PHY maps %08x %08x" NL, dev->phy_map, busy_phy_map);
2722
2723         dev->phy.mdio_read = emac_mdio_read;
2724         dev->phy.mdio_write = emac_mdio_write;
2725
2726         /* Enable internal clock source */
2727 #ifdef CONFIG_PPC_DCR_NATIVE
2728         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX))
2729                 dcri_clrset(SDR0, SDR0_MFR, 0, SDR0_MFR_ECS);
2730 #endif
2731         /* PHY clock workaround */
2732         emac_rx_clk_tx(dev);
2733
2734         /* Enable internal clock source on 440GX*/
2735 #ifdef CONFIG_PPC_DCR_NATIVE
2736         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX))
2737                 dcri_clrset(SDR0, SDR0_MFR, 0, SDR0_MFR_ECS);
2738 #endif
2739         /* Configure EMAC with defaults so we can at least use MDIO
2740          * This is needed mostly for 440GX
2741          */
2742         if (emac_phy_gpcs(dev->phy.mode)) {
2743                 /* XXX
2744                  * Make GPCS PHY address equal to EMAC index.
2745                  * We probably should take into account busy_phy_map
2746                  * and/or phy_map here.
2747                  *
2748                  * Note that the busy_phy_map is currently global
2749                  * while it should probably be per-ASIC...
2750                  */
2751                 dev->phy.gpcs_address = dev->gpcs_address;
2752                 if (dev->phy.gpcs_address == 0xffffffff)
2753                         dev->phy.address = dev->cell_index;
2754         }
2755
2756         emac_configure(dev);
2757
2758         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII)) {
2759                 int res = emac_dt_phy_probe(dev);
2760
2761                 switch (res) {
2762                 case 1:
2763                         /* No phy-handle property configured.
2764                          * Continue with the existing phy probe
2765                          * and setup code.
2766                          */
2767                         break;
2768
2769                 case 0:
2770                         mutex_unlock(&emac_phy_map_lock);
2771                         goto init_phy;
2772
2773                 default:
2774                         mutex_unlock(&emac_phy_map_lock);
2775                         dev_err(&dev->ofdev->dev, "failed to attach dt phy (%d).\n",
2776                                 res);
2777                         return res;
2778                 }
2779         }
2780
2781         if (dev->phy_address != 0xffffffff)
2782                 phy_map = ~(1 << dev->phy_address);
2783
2784         for (i = 0; i < 0x20; phy_map >>= 1, ++i)
2785                 if (!(phy_map & 1)) {
2786                         int r;
2787                         busy_phy_map |= 1 << i;
2788
2789                         /* Quick check if there is a PHY at the address */
2790                         r = emac_mdio_read(dev->ndev, i, MII_BMCR);
2791                         if (r == 0xffff || r < 0)
2792                                 continue;
2793                         if (!emac_mii_phy_probe(&dev->phy, i))
2794                                 break;
2795                 }
2796
2797         /* Enable external clock source */
2798 #ifdef CONFIG_PPC_DCR_NATIVE
2799         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX))
2800                 dcri_clrset(SDR0, SDR0_MFR, SDR0_MFR_ECS, 0);
2801 #endif
2802         mutex_unlock(&emac_phy_map_lock);
2803         if (i == 0x20) {
2804                 printk(KERN_WARNING "%pOF: can't find PHY!\n", np);
2805                 return -ENXIO;
2806         }
2807
2808  init_phy:
2809         /* Init PHY */
2810         if (dev->phy.def->ops->init)
2811                 dev->phy.def->ops->init(&dev->phy);
2812
2813         /* Disable any PHY features not supported by the platform */
2814         dev->phy.def->features &= ~dev->phy_feat_exc;
2815         dev->phy.features &= ~dev->phy_feat_exc;
2816
2817         /* Setup initial link parameters */
2818         if (dev->phy.features & SUPPORTED_Autoneg) {
2819                 adv = dev->phy.features;
2820                 if (!emac_has_feature(dev, EMAC_FTR_NO_FLOW_CONTROL_40x))
2821                         adv |= ADVERTISED_Pause | ADVERTISED_Asym_Pause;
2822                 /* Restart autonegotiation */
2823                 dev->phy.def->ops->setup_aneg(&dev->phy, adv);
2824         } else {
2825                 u32 f = dev->phy.def->features;
2826                 int speed = SPEED_10, fd = DUPLEX_HALF;
2827
2828                 /* Select highest supported speed/duplex */
2829                 if (f & SUPPORTED_1000baseT_Full) {
2830                         speed = SPEED_1000;
2831                         fd = DUPLEX_FULL;
2832                 } else if (f & SUPPORTED_1000baseT_Half)
2833                         speed = SPEED_1000;
2834                 else if (f & SUPPORTED_100baseT_Full) {
2835                         speed = SPEED_100;
2836                         fd = DUPLEX_FULL;
2837                 } else if (f & SUPPORTED_100baseT_Half)
2838                         speed = SPEED_100;
2839                 else if (f & SUPPORTED_10baseT_Full)
2840                         fd = DUPLEX_FULL;
2841
2842                 /* Force link parameters */
2843                 dev->phy.def->ops->setup_forced(&dev->phy, speed, fd);
2844         }
2845         return 0;
2846 }
2847
2848 static int emac_init_config(struct emac_instance *dev)
2849 {
2850         struct device_node *np = dev->ofdev->dev.of_node;
2851         const void *p;
2852
2853         /* Read config from device-tree */
2854         if (emac_read_uint_prop(np, "mal-device", &dev->mal_ph, 1))
2855                 return -ENXIO;
2856         if (emac_read_uint_prop(np, "mal-tx-channel", &dev->mal_tx_chan, 1))
2857                 return -ENXIO;
2858         if (emac_read_uint_prop(np, "mal-rx-channel", &dev->mal_rx_chan, 1))
2859                 return -ENXIO;
2860         if (emac_read_uint_prop(np, "cell-index", &dev->cell_index, 1))
2861                 return -ENXIO;
2862         if (emac_read_uint_prop(np, "max-frame-size", &dev->max_mtu, 0))
2863                 dev->max_mtu = ETH_DATA_LEN;
2864         if (emac_read_uint_prop(np, "rx-fifo-size", &dev->rx_fifo_size, 0))
2865                 dev->rx_fifo_size = 2048;
2866         if (emac_read_uint_prop(np, "tx-fifo-size", &dev->tx_fifo_size, 0))
2867                 dev->tx_fifo_size = 2048;
2868         if (emac_read_uint_prop(np, "rx-fifo-size-gige", &dev->rx_fifo_size_gige, 0))
2869                 dev->rx_fifo_size_gige = dev->rx_fifo_size;
2870         if (emac_read_uint_prop(np, "tx-fifo-size-gige", &dev->tx_fifo_size_gige, 0))
2871                 dev->tx_fifo_size_gige = dev->tx_fifo_size;
2872         if (emac_read_uint_prop(np, "phy-address", &dev->phy_address, 0))
2873                 dev->phy_address = 0xffffffff;
2874         if (emac_read_uint_prop(np, "phy-map", &dev->phy_map, 0))
2875                 dev->phy_map = 0xffffffff;
2876         if (emac_read_uint_prop(np, "gpcs-address", &dev->gpcs_address, 0))
2877                 dev->gpcs_address = 0xffffffff;
2878         if (emac_read_uint_prop(np->parent, "clock-frequency", &dev->opb_bus_freq, 1))
2879                 return -ENXIO;
2880         if (emac_read_uint_prop(np, "tah-device", &dev->tah_ph, 0))
2881                 dev->tah_ph = 0;
2882         if (emac_read_uint_prop(np, "tah-channel", &dev->tah_port, 0))
2883                 dev->tah_port = 0;
2884         if (emac_read_uint_prop(np, "mdio-device", &dev->mdio_ph, 0))
2885                 dev->mdio_ph = 0;
2886         if (emac_read_uint_prop(np, "zmii-device", &dev->zmii_ph, 0))
2887                 dev->zmii_ph = 0;
2888         if (emac_read_uint_prop(np, "zmii-channel", &dev->zmii_port, 0))
2889                 dev->zmii_port = 0xffffffff;
2890         if (emac_read_uint_prop(np, "rgmii-device", &dev->rgmii_ph, 0))
2891                 dev->rgmii_ph = 0;
2892         if (emac_read_uint_prop(np, "rgmii-channel", &dev->rgmii_port, 0))
2893                 dev->rgmii_port = 0xffffffff;
2894         if (emac_read_uint_prop(np, "fifo-entry-size", &dev->fifo_entry_size, 0))
2895                 dev->fifo_entry_size = 16;
2896         if (emac_read_uint_prop(np, "mal-burst-size", &dev->mal_burst_size, 0))
2897                 dev->mal_burst_size = 256;
2898
2899         /* PHY mode needs some decoding */
2900         dev->phy_mode = of_get_phy_mode(np);
2901         if (dev->phy_mode < 0)
2902                 dev->phy_mode = PHY_INTERFACE_MODE_NA;
2903
2904         /* Check EMAC version */
2905         if (of_device_is_compatible(np, "ibm,emac4sync")) {
2906                 dev->features |= (EMAC_FTR_EMAC4 | EMAC_FTR_EMAC4SYNC);
2907                 if (of_device_is_compatible(np, "ibm,emac-460ex") ||
2908                     of_device_is_compatible(np, "ibm,emac-460gt"))
2909                         dev->features |= EMAC_FTR_460EX_PHY_CLK_FIX;
2910                 if (of_device_is_compatible(np, "ibm,emac-405ex") ||
2911                     of_device_is_compatible(np, "ibm,emac-405exr"))
2912                         dev->features |= EMAC_FTR_440EP_PHY_CLK_FIX;
2913                 if (of_device_is_compatible(np, "ibm,emac-apm821xx")) {
2914                         dev->features |= (EMAC_APM821XX_REQ_JUMBO_FRAME_SIZE |
2915                                           EMAC_FTR_APM821XX_NO_HALF_DUPLEX |
2916                                           EMAC_FTR_460EX_PHY_CLK_FIX);
2917                 }
2918         } else if (of_device_is_compatible(np, "ibm,emac4")) {
2919                 dev->features |= EMAC_FTR_EMAC4;
2920                 if (of_device_is_compatible(np, "ibm,emac-440gx"))
2921                         dev->features |= EMAC_FTR_440GX_PHY_CLK_FIX;
2922         } else {
2923                 if (of_device_is_compatible(np, "ibm,emac-440ep") ||
2924                     of_device_is_compatible(np, "ibm,emac-440gr"))
2925                         dev->features |= EMAC_FTR_440EP_PHY_CLK_FIX;
2926                 if (of_device_is_compatible(np, "ibm,emac-405ez")) {
2927 #ifdef CONFIG_IBM_EMAC_NO_FLOW_CTRL
2928                         dev->features |= EMAC_FTR_NO_FLOW_CONTROL_40x;
2929 #else
2930                         printk(KERN_ERR "%pOF: Flow control not disabled!\n",
2931                                         np);
2932                         return -ENXIO;
2933 #endif
2934                 }
2935
2936         }
2937
2938         /* Fixup some feature bits based on the device tree */
2939         if (of_get_property(np, "has-inverted-stacr-oc", NULL))
2940                 dev->features |= EMAC_FTR_STACR_OC_INVERT;
2941         if (of_get_property(np, "has-new-stacr-staopc", NULL))
2942                 dev->features |= EMAC_FTR_HAS_NEW_STACR;
2943
2944         /* CAB lacks the appropriate properties */
2945         if (of_device_is_compatible(np, "ibm,emac-axon"))
2946                 dev->features |= EMAC_FTR_HAS_NEW_STACR |
2947                         EMAC_FTR_STACR_OC_INVERT;
2948
2949         /* Enable TAH/ZMII/RGMII features as found */
2950         if (dev->tah_ph != 0) {
2951 #ifdef CONFIG_IBM_EMAC_TAH
2952                 dev->features |= EMAC_FTR_HAS_TAH;
2953 #else
2954                 printk(KERN_ERR "%pOF: TAH support not enabled !\n", np);
2955                 return -ENXIO;
2956 #endif
2957         }
2958
2959         if (dev->zmii_ph != 0) {
2960 #ifdef CONFIG_IBM_EMAC_ZMII
2961                 dev->features |= EMAC_FTR_HAS_ZMII;
2962 #else
2963                 printk(KERN_ERR "%pOF: ZMII support not enabled !\n", np);
2964                 return -ENXIO;
2965 #endif
2966         }
2967
2968         if (dev->rgmii_ph != 0) {
2969 #ifdef CONFIG_IBM_EMAC_RGMII
2970                 dev->features |= EMAC_FTR_HAS_RGMII;
2971 #else
2972                 printk(KERN_ERR "%pOF: RGMII support not enabled !\n", np);
2973                 return -ENXIO;
2974 #endif
2975         }
2976
2977         /* Read MAC-address */
2978         p = of_get_property(np, "local-mac-address", NULL);
2979         if (p == NULL) {
2980                 printk(KERN_ERR "%pOF: Can't find local-mac-address property\n",
2981                        np);
2982                 return -ENXIO;
2983         }
2984         memcpy(dev->ndev->dev_addr, p, ETH_ALEN);
2985
2986         /* IAHT and GAHT filter parameterization */
2987         if (emac_has_feature(dev, EMAC_FTR_EMAC4SYNC)) {
2988                 dev->xaht_slots_shift = EMAC4SYNC_XAHT_SLOTS_SHIFT;
2989                 dev->xaht_width_shift = EMAC4SYNC_XAHT_WIDTH_SHIFT;
2990         } else {
2991                 dev->xaht_slots_shift = EMAC4_XAHT_SLOTS_SHIFT;
2992                 dev->xaht_width_shift = EMAC4_XAHT_WIDTH_SHIFT;
2993         }
2994
2995         /* This should never happen */
2996         if (WARN_ON(EMAC_XAHT_REGS(dev) > EMAC_XAHT_MAX_REGS))
2997                 return -ENXIO;
2998
2999         DBG(dev, "features     : 0x%08x / 0x%08x\n", dev->features, EMAC_FTRS_POSSIBLE);
3000         DBG(dev, "tx_fifo_size : %d (%d gige)\n", dev->tx_fifo_size, dev->tx_fifo_size_gige);
3001         DBG(dev, "rx_fifo_size : %d (%d gige)\n", dev->rx_fifo_size, dev->rx_fifo_size_gige);
3002         DBG(dev, "max_mtu      : %d\n", dev->max_mtu);
3003         DBG(dev, "OPB freq     : %d\n", dev->opb_bus_freq);
3004
3005         return 0;
3006 }
3007
3008 static const struct net_device_ops emac_netdev_ops = {
3009         .ndo_open               = emac_open,
3010         .ndo_stop               = emac_close,
3011         .ndo_get_stats          = emac_stats,
3012         .ndo_set_rx_mode        = emac_set_multicast_list,
3013         .ndo_do_ioctl           = emac_ioctl,
3014         .ndo_tx_timeout         = emac_tx_timeout,
3015         .ndo_validate_addr      = eth_validate_addr,
3016         .ndo_set_mac_address    = emac_set_mac_address,
3017         .ndo_start_xmit         = emac_start_xmit,
3018 };
3019
3020 static const struct net_device_ops emac_gige_netdev_ops = {
3021         .ndo_open               = emac_open,
3022         .ndo_stop               = emac_close,
3023         .ndo_get_stats          = emac_stats,
3024         .ndo_set_rx_mode        = emac_set_multicast_list,
3025         .ndo_do_ioctl           = emac_ioctl,
3026         .ndo_tx_timeout         = emac_tx_timeout,
3027         .ndo_validate_addr      = eth_validate_addr,
3028         .ndo_set_mac_address    = emac_set_mac_address,
3029         .ndo_start_xmit         = emac_start_xmit_sg,
3030         .ndo_change_mtu         = emac_change_mtu,
3031 };
3032
3033 static int emac_probe(struct platform_device *ofdev)
3034 {
3035         struct net_device *ndev;
3036         struct emac_instance *dev;
3037         struct device_node *np = ofdev->dev.of_node;
3038         struct device_node **blist = NULL;
3039         int err, i;
3040
3041         /* Skip unused/unwired EMACS.  We leave the check for an unused
3042          * property here for now, but new flat device trees should set a
3043          * status property to "disabled" instead.
3044          */
3045         if (of_get_property(np, "unused", NULL) || !of_device_is_available(np))
3046                 return -ENODEV;
3047
3048         /* Find ourselves in the bootlist if we are there */
3049         for (i = 0; i < EMAC_BOOT_LIST_SIZE; i++)
3050                 if (emac_boot_list[i] == np)
3051                         blist = &emac_boot_list[i];
3052
3053         /* Allocate our net_device structure */
3054         err = -ENOMEM;
3055         ndev = alloc_etherdev(sizeof(struct emac_instance));
3056         if (!ndev)
3057                 goto err_gone;
3058
3059         dev = netdev_priv(ndev);
3060         dev->ndev = ndev;
3061         dev->ofdev = ofdev;
3062         dev->blist = blist;
3063         SET_NETDEV_DEV(ndev, &ofdev->dev);
3064
3065         /* Initialize some embedded data structures */
3066         mutex_init(&dev->mdio_lock);
3067         mutex_init(&dev->link_lock);
3068         spin_lock_init(&dev->lock);
3069         INIT_WORK(&dev->reset_work, emac_reset_work);
3070
3071         /* Init various config data based on device-tree */
3072         err = emac_init_config(dev);
3073         if (err)
3074                 goto err_free;
3075
3076         /* Get interrupts. EMAC irq is mandatory, WOL irq is optional */
3077         dev->emac_irq = irq_of_parse_and_map(np, 0);
3078         dev->wol_irq = irq_of_parse_and_map(np, 1);
3079         if (!dev->emac_irq) {
3080                 printk(KERN_ERR "%pOF: Can't map main interrupt\n", np);
3081                 err = -ENODEV;
3082                 goto err_free;
3083         }
3084         ndev->irq = dev->emac_irq;
3085
3086         /* Map EMAC regs */
3087         // TODO : platform_get_resource() and devm_ioremap_resource()
3088         dev->emacp = of_iomap(np, 0);
3089         if (dev->emacp == NULL) {
3090                 printk(KERN_ERR "%pOF: Can't map device registers!\n", np);
3091                 err = -ENOMEM;
3092                 goto err_irq_unmap;
3093         }
3094
3095         /* Wait for dependent devices */
3096         err = emac_wait_deps(dev);
3097         if (err) {
3098                 printk(KERN_ERR
3099                        "%pOF: Timeout waiting for dependent devices\n", np);
3100                 /*  display more info about what's missing ? */
3101                 goto err_reg_unmap;
3102         }
3103         dev->mal = platform_get_drvdata(dev->mal_dev);
3104         if (dev->mdio_dev != NULL)
3105                 dev->mdio_instance = platform_get_drvdata(dev->mdio_dev);
3106
3107         /* Register with MAL */
3108         dev->commac.ops = &emac_commac_ops;
3109         dev->commac.dev = dev;
3110         dev->commac.tx_chan_mask = MAL_CHAN_MASK(dev->mal_tx_chan);
3111         dev->commac.rx_chan_mask = MAL_CHAN_MASK(dev->mal_rx_chan);
3112         err = mal_register_commac(dev->mal, &dev->commac);
3113         if (err) {
3114                 printk(KERN_ERR "%pOF: failed to register with mal %pOF!\n",
3115                        np, dev->mal_dev->dev.of_node);
3116                 goto err_rel_deps;
3117         }
3118         dev->rx_skb_size = emac_rx_skb_size(ndev->mtu);
3119         dev->rx_sync_size = emac_rx_sync_size(ndev->mtu);
3120
3121         /* Get pointers to BD rings */
3122         dev->tx_desc =
3123             dev->mal->bd_virt + mal_tx_bd_offset(dev->mal, dev->mal_tx_chan);
3124         dev->rx_desc =
3125             dev->mal->bd_virt + mal_rx_bd_offset(dev->mal, dev->mal_rx_chan);
3126
3127         DBG(dev, "tx_desc %p" NL, dev->tx_desc);
3128         DBG(dev, "rx_desc %p" NL, dev->rx_desc);
3129
3130         /* Clean rings */
3131         memset(dev->tx_desc, 0, NUM_TX_BUFF * sizeof(struct mal_descriptor));
3132         memset(dev->rx_desc, 0, NUM_RX_BUFF * sizeof(struct mal_descriptor));
3133         memset(dev->tx_skb, 0, NUM_TX_BUFF * sizeof(struct sk_buff *));
3134         memset(dev->rx_skb, 0, NUM_RX_BUFF * sizeof(struct sk_buff *));
3135
3136         /* Attach to ZMII, if needed */
3137         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII) &&
3138             (err = zmii_attach(dev->zmii_dev, dev->zmii_port, &dev->phy_mode)) != 0)
3139                 goto err_unreg_commac;
3140
3141         /* Attach to RGMII, if needed */
3142         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII) &&
3143             (err = rgmii_attach(dev->rgmii_dev, dev->rgmii_port, dev->phy_mode)) != 0)
3144                 goto err_detach_zmii;
3145
3146         /* Attach to TAH, if needed */
3147         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH) &&
3148             (err = tah_attach(dev->tah_dev, dev->tah_port)) != 0)
3149                 goto err_detach_rgmii;
3150
3151         /* Set some link defaults before we can find out real parameters */
3152         dev->phy.speed = SPEED_100;
3153         dev->phy.duplex = DUPLEX_FULL;
3154         dev->phy.autoneg = AUTONEG_DISABLE;
3155         dev->phy.pause = dev->phy.asym_pause = 0;
3156         dev->stop_timeout = STOP_TIMEOUT_100;
3157         INIT_DELAYED_WORK(&dev->link_work, emac_link_timer);
3158
3159         /* Some SoCs like APM821xx does not support Half Duplex mode. */
3160         if (emac_has_feature(dev, EMAC_FTR_APM821XX_NO_HALF_DUPLEX)) {
3161                 dev->phy_feat_exc = (SUPPORTED_1000baseT_Half |
3162                                      SUPPORTED_100baseT_Half |
3163                                      SUPPORTED_10baseT_Half);
3164         }
3165
3166         /* Find PHY if any */
3167         err = emac_init_phy(dev);
3168         if (err != 0)
3169                 goto err_detach_tah;
3170
3171         if (dev->tah_dev) {
3172                 ndev->hw_features = NETIF_F_IP_CSUM | NETIF_F_SG;
3173                 ndev->features |= ndev->hw_features | NETIF_F_RXCSUM;
3174         }
3175         ndev->watchdog_timeo = 5 * HZ;
3176         if (emac_phy_supports_gige(dev->phy_mode)) {
3177                 ndev->netdev_ops = &emac_gige_netdev_ops;
3178                 dev->commac.ops = &emac_commac_sg_ops;
3179         } else
3180                 ndev->netdev_ops = &emac_netdev_ops;
3181         ndev->ethtool_ops = &emac_ethtool_ops;
3182
3183         /* MTU range: 46 - 1500 or whatever is in OF */
3184         ndev->min_mtu = EMAC_MIN_MTU;
3185         ndev->max_mtu = dev->max_mtu;
3186
3187         netif_carrier_off(ndev);
3188
3189         err = register_netdev(ndev);
3190         if (err) {
3191                 printk(KERN_ERR "%pOF: failed to register net device (%d)!\n",
3192                        np, err);
3193                 goto err_detach_tah;
3194         }
3195
3196         /* Set our drvdata last as we don't want them visible until we are
3197          * fully initialized
3198          */
3199         wmb();
3200         platform_set_drvdata(ofdev, dev);
3201
3202         /* There's a new kid in town ! Let's tell everybody */
3203         wake_up_all(&emac_probe_wait);
3204
3205
3206         printk(KERN_INFO "%s: EMAC-%d %pOF, MAC %pM\n",
3207                ndev->name, dev->cell_index, np, ndev->dev_addr);
3208
3209         if (dev->phy_mode == PHY_INTERFACE_MODE_SGMII)
3210                 printk(KERN_NOTICE "%s: in SGMII mode\n", ndev->name);
3211
3212         if (dev->phy.address >= 0)
3213                 printk("%s: found %s PHY (0x%02x)\n", ndev->name,
3214                        dev->phy.def->name, dev->phy.address);
3215
3216         /* Life is good */
3217         return 0;
3218
3219         /* I have a bad feeling about this ... */
3220
3221  err_detach_tah:
3222         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
3223                 tah_detach(dev->tah_dev, dev->tah_port);
3224  err_detach_rgmii:
3225         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
3226                 rgmii_detach(dev->rgmii_dev, dev->rgmii_port);
3227  err_detach_zmii:
3228         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
3229                 zmii_detach(dev->zmii_dev, dev->zmii_port);
3230  err_unreg_commac:
3231         mal_unregister_commac(dev->mal, &dev->commac);
3232  err_rel_deps:
3233         emac_put_deps(dev);
3234  err_reg_unmap:
3235         iounmap(dev->emacp);
3236  err_irq_unmap:
3237         if (dev->wol_irq)
3238                 irq_dispose_mapping(dev->wol_irq);
3239         if (dev->emac_irq)
3240                 irq_dispose_mapping(dev->emac_irq);
3241  err_free:
3242         free_netdev(ndev);
3243  err_gone:
3244         /* if we were on the bootlist, remove us as we won't show up and
3245          * wake up all waiters to notify them in case they were waiting
3246          * on us
3247          */
3248         if (blist) {
3249                 *blist = NULL;
3250                 wake_up_all(&emac_probe_wait);
3251         }
3252         return err;
3253 }
3254
3255 static int emac_remove(struct platform_device *ofdev)
3256 {
3257         struct emac_instance *dev = platform_get_drvdata(ofdev);
3258
3259         DBG(dev, "remove" NL);
3260
3261         unregister_netdev(dev->ndev);
3262
3263         cancel_work_sync(&dev->reset_work);
3264
3265         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
3266                 tah_detach(dev->tah_dev, dev->tah_port);
3267         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
3268                 rgmii_detach(dev->rgmii_dev, dev->rgmii_port);
3269         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
3270                 zmii_detach(dev->zmii_dev, dev->zmii_port);
3271
3272         if (dev->phy_dev)
3273                 phy_disconnect(dev->phy_dev);
3274
3275         if (dev->mii_bus)
3276                 mdiobus_unregister(dev->mii_bus);
3277
3278         busy_phy_map &= ~(1 << dev->phy.address);
3279         DBG(dev, "busy_phy_map now %#x" NL, busy_phy_map);
3280
3281         mal_unregister_commac(dev->mal, &dev->commac);
3282         emac_put_deps(dev);
3283
3284         iounmap(dev->emacp);
3285
3286         if (dev->wol_irq)
3287                 irq_dispose_mapping(dev->wol_irq);
3288         if (dev->emac_irq)
3289                 irq_dispose_mapping(dev->emac_irq);
3290
3291         free_netdev(dev->ndev);
3292
3293         return 0;
3294 }
3295
3296 /* XXX Features in here should be replaced by properties... */
3297 static const struct of_device_id emac_match[] =
3298 {
3299         {
3300                 .type           = "network",
3301                 .compatible     = "ibm,emac",
3302         },
3303         {
3304                 .type           = "network",
3305                 .compatible     = "ibm,emac4",
3306         },
3307         {
3308                 .type           = "network",
3309                 .compatible     = "ibm,emac4sync",
3310         },
3311         {},
3312 };
3313 MODULE_DEVICE_TABLE(of, emac_match);
3314
3315 static struct platform_driver emac_driver = {
3316         .driver = {
3317                 .name = "emac",
3318                 .of_match_table = emac_match,
3319         },
3320         .probe = emac_probe,
3321         .remove = emac_remove,
3322 };
3323
3324 static void __init emac_make_bootlist(void)
3325 {
3326         struct device_node *np = NULL;
3327         int j, max, i = 0;
3328         int cell_indices[EMAC_BOOT_LIST_SIZE];
3329
3330         /* Collect EMACs */
3331         while((np = of_find_all_nodes(np)) != NULL) {
3332                 const u32 *idx;
3333
3334                 if (of_match_node(emac_match, np) == NULL)
3335                         continue;
3336                 if (of_get_property(np, "unused", NULL))
3337                         continue;
3338                 idx = of_get_property(np, "cell-index", NULL);
3339                 if (idx == NULL)
3340                         continue;
3341                 cell_indices[i] = *idx;
3342                 emac_boot_list[i++] = of_node_get(np);
3343                 if (i >= EMAC_BOOT_LIST_SIZE) {
3344                         of_node_put(np);
3345                         break;
3346                 }
3347         }
3348         max = i;
3349
3350         /* Bubble sort them (doh, what a creative algorithm :-) */
3351         for (i = 0; max > 1 && (i < (max - 1)); i++)
3352                 for (j = i; j < max; j++) {
3353                         if (cell_indices[i] > cell_indices[j]) {
3354                                 swap(emac_boot_list[i], emac_boot_list[j]);
3355                                 swap(cell_indices[i], cell_indices[j]);
3356                         }
3357                 }
3358 }
3359
3360 static int __init emac_init(void)
3361 {
3362         int rc;
3363
3364         printk(KERN_INFO DRV_DESC ", version " DRV_VERSION "\n");
3365
3366         /* Build EMAC boot list */
3367         emac_make_bootlist();
3368
3369         /* Init submodules */
3370         rc = mal_init();
3371         if (rc)
3372                 goto err;
3373         rc = zmii_init();
3374         if (rc)
3375                 goto err_mal;
3376         rc = rgmii_init();
3377         if (rc)
3378                 goto err_zmii;
3379         rc = tah_init();
3380         if (rc)
3381                 goto err_rgmii;
3382         rc = platform_driver_register(&emac_driver);
3383         if (rc)
3384                 goto err_tah;
3385
3386         return 0;
3387
3388  err_tah:
3389         tah_exit();
3390  err_rgmii:
3391         rgmii_exit();
3392  err_zmii:
3393         zmii_exit();
3394  err_mal:
3395         mal_exit();
3396  err:
3397         return rc;
3398 }
3399
3400 static void __exit emac_exit(void)
3401 {
3402         int i;
3403
3404         platform_driver_unregister(&emac_driver);
3405
3406         tah_exit();
3407         rgmii_exit();
3408         zmii_exit();
3409         mal_exit();
3410
3411         /* Destroy EMAC boot list */
3412         for (i = 0; i < EMAC_BOOT_LIST_SIZE; i++)
3413                 of_node_put(emac_boot_list[i]);
3414 }
3415
3416 module_init(emac_init);
3417 module_exit(emac_exit);