Merge branch 'drm-for-2.6.35' of git://git.kernel.org/pub/scm/linux/kernel/git/airlie...
[pandora-kernel.git] / drivers / net / ibm_newemac / core.c
1 /*
2  * drivers/net/ibm_newemac/core.c
3  *
4  * Driver for PowerPC 4xx on-chip ethernet controller.
5  *
6  * Copyright 2007 Benjamin Herrenschmidt, IBM Corp.
7  *                <benh@kernel.crashing.org>
8  *
9  * Based on the arch/ppc version of the driver:
10  *
11  * Copyright (c) 2004, 2005 Zultys Technologies.
12  * Eugene Surovegin <eugene.surovegin@zultys.com> or <ebs@ebshome.net>
13  *
14  * Based on original work by
15  *      Matt Porter <mporter@kernel.crashing.org>
16  *      (c) 2003 Benjamin Herrenschmidt <benh@kernel.crashing.org>
17  *      Armin Kuster <akuster@mvista.com>
18  *      Johnnie Peters <jpeters@mvista.com>
19  *
20  * This program is free software; you can redistribute  it and/or modify it
21  * under  the terms of  the GNU General  Public License as published by the
22  * Free Software Foundation;  either version 2 of the  License, or (at your
23  * option) any later version.
24  *
25  */
26
27 #include <linux/module.h>
28 #include <linux/sched.h>
29 #include <linux/string.h>
30 #include <linux/errno.h>
31 #include <linux/delay.h>
32 #include <linux/types.h>
33 #include <linux/pci.h>
34 #include <linux/etherdevice.h>
35 #include <linux/skbuff.h>
36 #include <linux/crc32.h>
37 #include <linux/ethtool.h>
38 #include <linux/mii.h>
39 #include <linux/bitops.h>
40 #include <linux/workqueue.h>
41 #include <linux/of.h>
42 #include <linux/slab.h>
43
44 #include <asm/processor.h>
45 #include <asm/io.h>
46 #include <asm/dma.h>
47 #include <asm/uaccess.h>
48 #include <asm/dcr.h>
49 #include <asm/dcr-regs.h>
50
51 #include "core.h"
52
53 /*
54  * Lack of dma_unmap_???? calls is intentional.
55  *
56  * API-correct usage requires additional support state information to be
57  * maintained for every RX and TX buffer descriptor (BD). Unfortunately, due to
58  * EMAC design (e.g. TX buffer passed from network stack can be split into
59  * several BDs, dma_map_single/dma_map_page can be used to map particular BD),
60  * maintaining such information will add additional overhead.
61  * Current DMA API implementation for 4xx processors only ensures cache coherency
62  * and dma_unmap_???? routines are empty and are likely to stay this way.
63  * I decided to omit dma_unmap_??? calls because I don't want to add additional
64  * complexity just for the sake of following some abstract API, when it doesn't
65  * add any real benefit to the driver. I understand that this decision maybe
66  * controversial, but I really tried to make code API-correct and efficient
67  * at the same time and didn't come up with code I liked :(.                --ebs
68  */
69
70 #define DRV_NAME        "emac"
71 #define DRV_VERSION     "3.54"
72 #define DRV_DESC        "PPC 4xx OCP EMAC driver"
73
74 MODULE_DESCRIPTION(DRV_DESC);
75 MODULE_AUTHOR
76     ("Eugene Surovegin <eugene.surovegin@zultys.com> or <ebs@ebshome.net>");
77 MODULE_LICENSE("GPL");
78
79 /*
80  * PPC64 doesn't (yet) have a cacheable_memcpy
81  */
82 #ifdef CONFIG_PPC64
83 #define cacheable_memcpy(d,s,n) memcpy((d),(s),(n))
84 #endif
85
86 /* minimum number of free TX descriptors required to wake up TX process */
87 #define EMAC_TX_WAKEUP_THRESH           (NUM_TX_BUFF / 4)
88
89 /* If packet size is less than this number, we allocate small skb and copy packet
90  * contents into it instead of just sending original big skb up
91  */
92 #define EMAC_RX_COPY_THRESH             CONFIG_IBM_NEW_EMAC_RX_COPY_THRESHOLD
93
94 /* Since multiple EMACs share MDIO lines in various ways, we need
95  * to avoid re-using the same PHY ID in cases where the arch didn't
96  * setup precise phy_map entries
97  *
98  * XXX This is something that needs to be reworked as we can have multiple
99  * EMAC "sets" (multiple ASICs containing several EMACs) though we can
100  * probably require in that case to have explicit PHY IDs in the device-tree
101  */
102 static u32 busy_phy_map;
103 static DEFINE_MUTEX(emac_phy_map_lock);
104
105 /* This is the wait queue used to wait on any event related to probe, that
106  * is discovery of MALs, other EMACs, ZMII/RGMIIs, etc...
107  */
108 static DECLARE_WAIT_QUEUE_HEAD(emac_probe_wait);
109
110 /* Having stable interface names is a doomed idea. However, it would be nice
111  * if we didn't have completely random interface names at boot too :-) It's
112  * just a matter of making everybody's life easier. Since we are doing
113  * threaded probing, it's a bit harder though. The base idea here is that
114  * we make up a list of all emacs in the device-tree before we register the
115  * driver. Every emac will then wait for the previous one in the list to
116  * initialize before itself. We should also keep that list ordered by
117  * cell_index.
118  * That list is only 4 entries long, meaning that additional EMACs don't
119  * get ordering guarantees unless EMAC_BOOT_LIST_SIZE is increased.
120  */
121
122 #define EMAC_BOOT_LIST_SIZE     4
123 static struct device_node *emac_boot_list[EMAC_BOOT_LIST_SIZE];
124
125 /* How long should I wait for dependent devices ? */
126 #define EMAC_PROBE_DEP_TIMEOUT  (HZ * 5)
127
128 /* I don't want to litter system log with timeout errors
129  * when we have brain-damaged PHY.
130  */
131 static inline void emac_report_timeout_error(struct emac_instance *dev,
132                                              const char *error)
133 {
134         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX |
135                                   EMAC_FTR_460EX_PHY_CLK_FIX |
136                                   EMAC_FTR_440EP_PHY_CLK_FIX))
137                 DBG(dev, "%s" NL, error);
138         else if (net_ratelimit())
139                 printk(KERN_ERR "%s: %s\n", dev->ofdev->node->full_name, error);
140 }
141
142 /* EMAC PHY clock workaround:
143  * 440EP/440GR has more sane SDR0_MFR register implementation than 440GX,
144  * which allows controlling each EMAC clock
145  */
146 static inline void emac_rx_clk_tx(struct emac_instance *dev)
147 {
148 #ifdef CONFIG_PPC_DCR_NATIVE
149         if (emac_has_feature(dev, EMAC_FTR_440EP_PHY_CLK_FIX))
150                 dcri_clrset(SDR0, SDR0_MFR,
151                             0, SDR0_MFR_ECS >> dev->cell_index);
152 #endif
153 }
154
155 static inline void emac_rx_clk_default(struct emac_instance *dev)
156 {
157 #ifdef CONFIG_PPC_DCR_NATIVE
158         if (emac_has_feature(dev, EMAC_FTR_440EP_PHY_CLK_FIX))
159                 dcri_clrset(SDR0, SDR0_MFR,
160                             SDR0_MFR_ECS >> dev->cell_index, 0);
161 #endif
162 }
163
164 /* PHY polling intervals */
165 #define PHY_POLL_LINK_ON        HZ
166 #define PHY_POLL_LINK_OFF       (HZ / 5)
167
168 /* Graceful stop timeouts in us.
169  * We should allow up to 1 frame time (full-duplex, ignoring collisions)
170  */
171 #define STOP_TIMEOUT_10         1230
172 #define STOP_TIMEOUT_100        124
173 #define STOP_TIMEOUT_1000       13
174 #define STOP_TIMEOUT_1000_JUMBO 73
175
176 static unsigned char default_mcast_addr[] = {
177         0x01, 0x80, 0xC2, 0x00, 0x00, 0x01
178 };
179
180 /* Please, keep in sync with struct ibm_emac_stats/ibm_emac_error_stats */
181 static const char emac_stats_keys[EMAC_ETHTOOL_STATS_COUNT][ETH_GSTRING_LEN] = {
182         "rx_packets", "rx_bytes", "tx_packets", "tx_bytes", "rx_packets_csum",
183         "tx_packets_csum", "tx_undo", "rx_dropped_stack", "rx_dropped_oom",
184         "rx_dropped_error", "rx_dropped_resize", "rx_dropped_mtu",
185         "rx_stopped", "rx_bd_errors", "rx_bd_overrun", "rx_bd_bad_packet",
186         "rx_bd_runt_packet", "rx_bd_short_event", "rx_bd_alignment_error",
187         "rx_bd_bad_fcs", "rx_bd_packet_too_long", "rx_bd_out_of_range",
188         "rx_bd_in_range", "rx_parity", "rx_fifo_overrun", "rx_overrun",
189         "rx_bad_packet", "rx_runt_packet", "rx_short_event",
190         "rx_alignment_error", "rx_bad_fcs", "rx_packet_too_long",
191         "rx_out_of_range", "rx_in_range", "tx_dropped", "tx_bd_errors",
192         "tx_bd_bad_fcs", "tx_bd_carrier_loss", "tx_bd_excessive_deferral",
193         "tx_bd_excessive_collisions", "tx_bd_late_collision",
194         "tx_bd_multple_collisions", "tx_bd_single_collision",
195         "tx_bd_underrun", "tx_bd_sqe", "tx_parity", "tx_underrun", "tx_sqe",
196         "tx_errors"
197 };
198
199 static irqreturn_t emac_irq(int irq, void *dev_instance);
200 static void emac_clean_tx_ring(struct emac_instance *dev);
201 static void __emac_set_multicast_list(struct emac_instance *dev);
202
203 static inline int emac_phy_supports_gige(int phy_mode)
204 {
205         return  phy_mode == PHY_MODE_GMII ||
206                 phy_mode == PHY_MODE_RGMII ||
207                 phy_mode == PHY_MODE_SGMII ||
208                 phy_mode == PHY_MODE_TBI ||
209                 phy_mode == PHY_MODE_RTBI;
210 }
211
212 static inline int emac_phy_gpcs(int phy_mode)
213 {
214         return  phy_mode == PHY_MODE_SGMII ||
215                 phy_mode == PHY_MODE_TBI ||
216                 phy_mode == PHY_MODE_RTBI;
217 }
218
219 static inline void emac_tx_enable(struct emac_instance *dev)
220 {
221         struct emac_regs __iomem *p = dev->emacp;
222         u32 r;
223
224         DBG(dev, "tx_enable" NL);
225
226         r = in_be32(&p->mr0);
227         if (!(r & EMAC_MR0_TXE))
228                 out_be32(&p->mr0, r | EMAC_MR0_TXE);
229 }
230
231 static void emac_tx_disable(struct emac_instance *dev)
232 {
233         struct emac_regs __iomem *p = dev->emacp;
234         u32 r;
235
236         DBG(dev, "tx_disable" NL);
237
238         r = in_be32(&p->mr0);
239         if (r & EMAC_MR0_TXE) {
240                 int n = dev->stop_timeout;
241                 out_be32(&p->mr0, r & ~EMAC_MR0_TXE);
242                 while (!(in_be32(&p->mr0) & EMAC_MR0_TXI) && n) {
243                         udelay(1);
244                         --n;
245                 }
246                 if (unlikely(!n))
247                         emac_report_timeout_error(dev, "TX disable timeout");
248         }
249 }
250
251 static void emac_rx_enable(struct emac_instance *dev)
252 {
253         struct emac_regs __iomem *p = dev->emacp;
254         u32 r;
255
256         if (unlikely(test_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags)))
257                 goto out;
258
259         DBG(dev, "rx_enable" NL);
260
261         r = in_be32(&p->mr0);
262         if (!(r & EMAC_MR0_RXE)) {
263                 if (unlikely(!(r & EMAC_MR0_RXI))) {
264                         /* Wait if previous async disable is still in progress */
265                         int n = dev->stop_timeout;
266                         while (!(r = in_be32(&p->mr0) & EMAC_MR0_RXI) && n) {
267                                 udelay(1);
268                                 --n;
269                         }
270                         if (unlikely(!n))
271                                 emac_report_timeout_error(dev,
272                                                           "RX disable timeout");
273                 }
274                 out_be32(&p->mr0, r | EMAC_MR0_RXE);
275         }
276  out:
277         ;
278 }
279
280 static void emac_rx_disable(struct emac_instance *dev)
281 {
282         struct emac_regs __iomem *p = dev->emacp;
283         u32 r;
284
285         DBG(dev, "rx_disable" NL);
286
287         r = in_be32(&p->mr0);
288         if (r & EMAC_MR0_RXE) {
289                 int n = dev->stop_timeout;
290                 out_be32(&p->mr0, r & ~EMAC_MR0_RXE);
291                 while (!(in_be32(&p->mr0) & EMAC_MR0_RXI) && n) {
292                         udelay(1);
293                         --n;
294                 }
295                 if (unlikely(!n))
296                         emac_report_timeout_error(dev, "RX disable timeout");
297         }
298 }
299
300 static inline void emac_netif_stop(struct emac_instance *dev)
301 {
302         netif_tx_lock_bh(dev->ndev);
303         netif_addr_lock(dev->ndev);
304         dev->no_mcast = 1;
305         netif_addr_unlock(dev->ndev);
306         netif_tx_unlock_bh(dev->ndev);
307         dev->ndev->trans_start = jiffies;       /* prevent tx timeout */
308         mal_poll_disable(dev->mal, &dev->commac);
309         netif_tx_disable(dev->ndev);
310 }
311
312 static inline void emac_netif_start(struct emac_instance *dev)
313 {
314         netif_tx_lock_bh(dev->ndev);
315         netif_addr_lock(dev->ndev);
316         dev->no_mcast = 0;
317         if (dev->mcast_pending && netif_running(dev->ndev))
318                 __emac_set_multicast_list(dev);
319         netif_addr_unlock(dev->ndev);
320         netif_tx_unlock_bh(dev->ndev);
321
322         netif_wake_queue(dev->ndev);
323
324         /* NOTE: unconditional netif_wake_queue is only appropriate
325          * so long as all callers are assured to have free tx slots
326          * (taken from tg3... though the case where that is wrong is
327          *  not terribly harmful)
328          */
329         mal_poll_enable(dev->mal, &dev->commac);
330 }
331
332 static inline void emac_rx_disable_async(struct emac_instance *dev)
333 {
334         struct emac_regs __iomem *p = dev->emacp;
335         u32 r;
336
337         DBG(dev, "rx_disable_async" NL);
338
339         r = in_be32(&p->mr0);
340         if (r & EMAC_MR0_RXE)
341                 out_be32(&p->mr0, r & ~EMAC_MR0_RXE);
342 }
343
344 static int emac_reset(struct emac_instance *dev)
345 {
346         struct emac_regs __iomem *p = dev->emacp;
347         int n = 20;
348
349         DBG(dev, "reset" NL);
350
351         if (!dev->reset_failed) {
352                 /* 40x erratum suggests stopping RX channel before reset,
353                  * we stop TX as well
354                  */
355                 emac_rx_disable(dev);
356                 emac_tx_disable(dev);
357         }
358
359 #ifdef CONFIG_PPC_DCR_NATIVE
360         /* Enable internal clock source */
361         if (emac_has_feature(dev, EMAC_FTR_460EX_PHY_CLK_FIX))
362                 dcri_clrset(SDR0, SDR0_ETH_CFG,
363                             0, SDR0_ETH_CFG_ECS << dev->cell_index);
364 #endif
365
366         out_be32(&p->mr0, EMAC_MR0_SRST);
367         while ((in_be32(&p->mr0) & EMAC_MR0_SRST) && n)
368                 --n;
369
370 #ifdef CONFIG_PPC_DCR_NATIVE
371          /* Enable external clock source */
372         if (emac_has_feature(dev, EMAC_FTR_460EX_PHY_CLK_FIX))
373                 dcri_clrset(SDR0, SDR0_ETH_CFG,
374                             SDR0_ETH_CFG_ECS << dev->cell_index, 0);
375 #endif
376
377         if (n) {
378                 dev->reset_failed = 0;
379                 return 0;
380         } else {
381                 emac_report_timeout_error(dev, "reset timeout");
382                 dev->reset_failed = 1;
383                 return -ETIMEDOUT;
384         }
385 }
386
387 static void emac_hash_mc(struct emac_instance *dev)
388 {
389         const int regs = EMAC_XAHT_REGS(dev);
390         u32 *gaht_base = emac_gaht_base(dev);
391         u32 gaht_temp[regs];
392         struct netdev_hw_addr *ha;
393         int i;
394
395         DBG(dev, "hash_mc %d" NL, netdev_mc_count(dev->ndev));
396
397         memset(gaht_temp, 0, sizeof (gaht_temp));
398
399         netdev_for_each_mc_addr(ha, dev->ndev) {
400                 int slot, reg, mask;
401                 DBG2(dev, "mc %pM" NL, ha->addr);
402
403                 slot = EMAC_XAHT_CRC_TO_SLOT(dev,
404                                              ether_crc(ETH_ALEN, ha->addr));
405                 reg = EMAC_XAHT_SLOT_TO_REG(dev, slot);
406                 mask = EMAC_XAHT_SLOT_TO_MASK(dev, slot);
407
408                 gaht_temp[reg] |= mask;
409         }
410
411         for (i = 0; i < regs; i++)
412                 out_be32(gaht_base + i, gaht_temp[i]);
413 }
414
415 static inline u32 emac_iff2rmr(struct net_device *ndev)
416 {
417         struct emac_instance *dev = netdev_priv(ndev);
418         u32 r;
419
420         r = EMAC_RMR_SP | EMAC_RMR_SFCS | EMAC_RMR_IAE | EMAC_RMR_BAE;
421
422         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
423             r |= EMAC4_RMR_BASE;
424         else
425             r |= EMAC_RMR_BASE;
426
427         if (ndev->flags & IFF_PROMISC)
428                 r |= EMAC_RMR_PME;
429         else if (ndev->flags & IFF_ALLMULTI ||
430                          (netdev_mc_count(ndev) > EMAC_XAHT_SLOTS(dev)))
431                 r |= EMAC_RMR_PMME;
432         else if (!netdev_mc_empty(ndev))
433                 r |= EMAC_RMR_MAE;
434
435         return r;
436 }
437
438 static u32 __emac_calc_base_mr1(struct emac_instance *dev, int tx_size, int rx_size)
439 {
440         u32 ret = EMAC_MR1_VLE | EMAC_MR1_IST | EMAC_MR1_TR0_MULT;
441
442         DBG2(dev, "__emac_calc_base_mr1" NL);
443
444         switch(tx_size) {
445         case 2048:
446                 ret |= EMAC_MR1_TFS_2K;
447                 break;
448         default:
449                 printk(KERN_WARNING "%s: Unknown Tx FIFO size %d\n",
450                        dev->ndev->name, tx_size);
451         }
452
453         switch(rx_size) {
454         case 16384:
455                 ret |= EMAC_MR1_RFS_16K;
456                 break;
457         case 4096:
458                 ret |= EMAC_MR1_RFS_4K;
459                 break;
460         default:
461                 printk(KERN_WARNING "%s: Unknown Rx FIFO size %d\n",
462                        dev->ndev->name, rx_size);
463         }
464
465         return ret;
466 }
467
468 static u32 __emac4_calc_base_mr1(struct emac_instance *dev, int tx_size, int rx_size)
469 {
470         u32 ret = EMAC_MR1_VLE | EMAC_MR1_IST | EMAC4_MR1_TR |
471                 EMAC4_MR1_OBCI(dev->opb_bus_freq / 1000000);
472
473         DBG2(dev, "__emac4_calc_base_mr1" NL);
474
475         switch(tx_size) {
476         case 16384:
477                 ret |= EMAC4_MR1_TFS_16K;
478                 break;
479         case 4096:
480                 ret |= EMAC4_MR1_TFS_4K;
481                 break;
482         case 2048:
483                 ret |= EMAC4_MR1_TFS_2K;
484                 break;
485         default:
486                 printk(KERN_WARNING "%s: Unknown Tx FIFO size %d\n",
487                        dev->ndev->name, tx_size);
488         }
489
490         switch(rx_size) {
491         case 16384:
492                 ret |= EMAC4_MR1_RFS_16K;
493                 break;
494         case 4096:
495                 ret |= EMAC4_MR1_RFS_4K;
496                 break;
497         case 2048:
498                 ret |= EMAC4_MR1_RFS_2K;
499                 break;
500         default:
501                 printk(KERN_WARNING "%s: Unknown Rx FIFO size %d\n",
502                        dev->ndev->name, rx_size);
503         }
504
505         return ret;
506 }
507
508 static u32 emac_calc_base_mr1(struct emac_instance *dev, int tx_size, int rx_size)
509 {
510         return emac_has_feature(dev, EMAC_FTR_EMAC4) ?
511                 __emac4_calc_base_mr1(dev, tx_size, rx_size) :
512                 __emac_calc_base_mr1(dev, tx_size, rx_size);
513 }
514
515 static inline u32 emac_calc_trtr(struct emac_instance *dev, unsigned int size)
516 {
517         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
518                 return ((size >> 6) - 1) << EMAC_TRTR_SHIFT_EMAC4;
519         else
520                 return ((size >> 6) - 1) << EMAC_TRTR_SHIFT;
521 }
522
523 static inline u32 emac_calc_rwmr(struct emac_instance *dev,
524                                  unsigned int low, unsigned int high)
525 {
526         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
527                 return (low << 22) | ( (high & 0x3ff) << 6);
528         else
529                 return (low << 23) | ( (high & 0x1ff) << 7);
530 }
531
532 static int emac_configure(struct emac_instance *dev)
533 {
534         struct emac_regs __iomem *p = dev->emacp;
535         struct net_device *ndev = dev->ndev;
536         int tx_size, rx_size, link = netif_carrier_ok(dev->ndev);
537         u32 r, mr1 = 0;
538
539         DBG(dev, "configure" NL);
540
541         if (!link) {
542                 out_be32(&p->mr1, in_be32(&p->mr1)
543                          | EMAC_MR1_FDE | EMAC_MR1_ILE);
544                 udelay(100);
545         } else if (emac_reset(dev) < 0)
546                 return -ETIMEDOUT;
547
548         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
549                 tah_reset(dev->tah_dev);
550
551         DBG(dev, " link = %d duplex = %d, pause = %d, asym_pause = %d\n",
552             link, dev->phy.duplex, dev->phy.pause, dev->phy.asym_pause);
553
554         /* Default fifo sizes */
555         tx_size = dev->tx_fifo_size;
556         rx_size = dev->rx_fifo_size;
557
558         /* No link, force loopback */
559         if (!link)
560                 mr1 = EMAC_MR1_FDE | EMAC_MR1_ILE;
561
562         /* Check for full duplex */
563         else if (dev->phy.duplex == DUPLEX_FULL)
564                 mr1 |= EMAC_MR1_FDE | EMAC_MR1_MWSW_001;
565
566         /* Adjust fifo sizes, mr1 and timeouts based on link speed */
567         dev->stop_timeout = STOP_TIMEOUT_10;
568         switch (dev->phy.speed) {
569         case SPEED_1000:
570                 if (emac_phy_gpcs(dev->phy.mode)) {
571                         mr1 |= EMAC_MR1_MF_1000GPCS | EMAC_MR1_MF_IPPA(
572                                 (dev->phy.gpcs_address != 0xffffffff) ?
573                                  dev->phy.gpcs_address : dev->phy.address);
574
575                         /* Put some arbitrary OUI, Manuf & Rev IDs so we can
576                          * identify this GPCS PHY later.
577                          */
578                         out_be32(&p->u1.emac4.ipcr, 0xdeadbeef);
579                 } else
580                         mr1 |= EMAC_MR1_MF_1000;
581
582                 /* Extended fifo sizes */
583                 tx_size = dev->tx_fifo_size_gige;
584                 rx_size = dev->rx_fifo_size_gige;
585
586                 if (dev->ndev->mtu > ETH_DATA_LEN) {
587                         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
588                                 mr1 |= EMAC4_MR1_JPSM;
589                         else
590                                 mr1 |= EMAC_MR1_JPSM;
591                         dev->stop_timeout = STOP_TIMEOUT_1000_JUMBO;
592                 } else
593                         dev->stop_timeout = STOP_TIMEOUT_1000;
594                 break;
595         case SPEED_100:
596                 mr1 |= EMAC_MR1_MF_100;
597                 dev->stop_timeout = STOP_TIMEOUT_100;
598                 break;
599         default: /* make gcc happy */
600                 break;
601         }
602
603         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
604                 rgmii_set_speed(dev->rgmii_dev, dev->rgmii_port,
605                                 dev->phy.speed);
606         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
607                 zmii_set_speed(dev->zmii_dev, dev->zmii_port, dev->phy.speed);
608
609         /* on 40x erratum forces us to NOT use integrated flow control,
610          * let's hope it works on 44x ;)
611          */
612         if (!emac_has_feature(dev, EMAC_FTR_NO_FLOW_CONTROL_40x) &&
613             dev->phy.duplex == DUPLEX_FULL) {
614                 if (dev->phy.pause)
615                         mr1 |= EMAC_MR1_EIFC | EMAC_MR1_APP;
616                 else if (dev->phy.asym_pause)
617                         mr1 |= EMAC_MR1_APP;
618         }
619
620         /* Add base settings & fifo sizes & program MR1 */
621         mr1 |= emac_calc_base_mr1(dev, tx_size, rx_size);
622         out_be32(&p->mr1, mr1);
623
624         /* Set individual MAC address */
625         out_be32(&p->iahr, (ndev->dev_addr[0] << 8) | ndev->dev_addr[1]);
626         out_be32(&p->ialr, (ndev->dev_addr[2] << 24) |
627                  (ndev->dev_addr[3] << 16) | (ndev->dev_addr[4] << 8) |
628                  ndev->dev_addr[5]);
629
630         /* VLAN Tag Protocol ID */
631         out_be32(&p->vtpid, 0x8100);
632
633         /* Receive mode register */
634         r = emac_iff2rmr(ndev);
635         if (r & EMAC_RMR_MAE)
636                 emac_hash_mc(dev);
637         out_be32(&p->rmr, r);
638
639         /* FIFOs thresholds */
640         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
641                 r = EMAC4_TMR1((dev->mal_burst_size / dev->fifo_entry_size) + 1,
642                                tx_size / 2 / dev->fifo_entry_size);
643         else
644                 r = EMAC_TMR1((dev->mal_burst_size / dev->fifo_entry_size) + 1,
645                               tx_size / 2 / dev->fifo_entry_size);
646         out_be32(&p->tmr1, r);
647         out_be32(&p->trtr, emac_calc_trtr(dev, tx_size / 2));
648
649         /* PAUSE frame is sent when RX FIFO reaches its high-water mark,
650            there should be still enough space in FIFO to allow the our link
651            partner time to process this frame and also time to send PAUSE
652            frame itself.
653
654            Here is the worst case scenario for the RX FIFO "headroom"
655            (from "The Switch Book") (100Mbps, without preamble, inter-frame gap):
656
657            1) One maximum-length frame on TX                    1522 bytes
658            2) One PAUSE frame time                                64 bytes
659            3) PAUSE frame decode time allowance                   64 bytes
660            4) One maximum-length frame on RX                    1522 bytes
661            5) Round-trip propagation delay of the link (100Mb)    15 bytes
662            ----------
663            3187 bytes
664
665            I chose to set high-water mark to RX_FIFO_SIZE / 4 (1024 bytes)
666            low-water mark  to RX_FIFO_SIZE / 8 (512 bytes)
667          */
668         r = emac_calc_rwmr(dev, rx_size / 8 / dev->fifo_entry_size,
669                            rx_size / 4 / dev->fifo_entry_size);
670         out_be32(&p->rwmr, r);
671
672         /* Set PAUSE timer to the maximum */
673         out_be32(&p->ptr, 0xffff);
674
675         /* IRQ sources */
676         r = EMAC_ISR_OVR | EMAC_ISR_BP | EMAC_ISR_SE |
677                 EMAC_ISR_ALE | EMAC_ISR_BFCS | EMAC_ISR_PTLE | EMAC_ISR_ORE |
678                 EMAC_ISR_IRE | EMAC_ISR_TE;
679         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
680             r |= EMAC4_ISR_TXPE | EMAC4_ISR_RXPE /* | EMAC4_ISR_TXUE |
681                                                   EMAC4_ISR_RXOE | */;
682         out_be32(&p->iser,  r);
683
684         /* We need to take GPCS PHY out of isolate mode after EMAC reset */
685         if (emac_phy_gpcs(dev->phy.mode)) {
686                 if (dev->phy.gpcs_address != 0xffffffff)
687                         emac_mii_reset_gpcs(&dev->phy);
688                 else
689                         emac_mii_reset_phy(&dev->phy);
690         }
691
692         return 0;
693 }
694
695 static void emac_reinitialize(struct emac_instance *dev)
696 {
697         DBG(dev, "reinitialize" NL);
698
699         emac_netif_stop(dev);
700         if (!emac_configure(dev)) {
701                 emac_tx_enable(dev);
702                 emac_rx_enable(dev);
703         }
704         emac_netif_start(dev);
705 }
706
707 static void emac_full_tx_reset(struct emac_instance *dev)
708 {
709         DBG(dev, "full_tx_reset" NL);
710
711         emac_tx_disable(dev);
712         mal_disable_tx_channel(dev->mal, dev->mal_tx_chan);
713         emac_clean_tx_ring(dev);
714         dev->tx_cnt = dev->tx_slot = dev->ack_slot = 0;
715
716         emac_configure(dev);
717
718         mal_enable_tx_channel(dev->mal, dev->mal_tx_chan);
719         emac_tx_enable(dev);
720         emac_rx_enable(dev);
721 }
722
723 static void emac_reset_work(struct work_struct *work)
724 {
725         struct emac_instance *dev = container_of(work, struct emac_instance, reset_work);
726
727         DBG(dev, "reset_work" NL);
728
729         mutex_lock(&dev->link_lock);
730         if (dev->opened) {
731                 emac_netif_stop(dev);
732                 emac_full_tx_reset(dev);
733                 emac_netif_start(dev);
734         }
735         mutex_unlock(&dev->link_lock);
736 }
737
738 static void emac_tx_timeout(struct net_device *ndev)
739 {
740         struct emac_instance *dev = netdev_priv(ndev);
741
742         DBG(dev, "tx_timeout" NL);
743
744         schedule_work(&dev->reset_work);
745 }
746
747
748 static inline int emac_phy_done(struct emac_instance *dev, u32 stacr)
749 {
750         int done = !!(stacr & EMAC_STACR_OC);
751
752         if (emac_has_feature(dev, EMAC_FTR_STACR_OC_INVERT))
753                 done = !done;
754
755         return done;
756 };
757
758 static int __emac_mdio_read(struct emac_instance *dev, u8 id, u8 reg)
759 {
760         struct emac_regs __iomem *p = dev->emacp;
761         u32 r = 0;
762         int n, err = -ETIMEDOUT;
763
764         mutex_lock(&dev->mdio_lock);
765
766         DBG2(dev, "mdio_read(%02x,%02x)" NL, id, reg);
767
768         /* Enable proper MDIO port */
769         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
770                 zmii_get_mdio(dev->zmii_dev, dev->zmii_port);
771         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
772                 rgmii_get_mdio(dev->rgmii_dev, dev->rgmii_port);
773
774         /* Wait for management interface to become idle */
775         n = 20;
776         while (!emac_phy_done(dev, in_be32(&p->stacr))) {
777                 udelay(1);
778                 if (!--n) {
779                         DBG2(dev, " -> timeout wait idle\n");
780                         goto bail;
781                 }
782         }
783
784         /* Issue read command */
785         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
786                 r = EMAC4_STACR_BASE(dev->opb_bus_freq);
787         else
788                 r = EMAC_STACR_BASE(dev->opb_bus_freq);
789         if (emac_has_feature(dev, EMAC_FTR_STACR_OC_INVERT))
790                 r |= EMAC_STACR_OC;
791         if (emac_has_feature(dev, EMAC_FTR_HAS_NEW_STACR))
792                 r |= EMACX_STACR_STAC_READ;
793         else
794                 r |= EMAC_STACR_STAC_READ;
795         r |= (reg & EMAC_STACR_PRA_MASK)
796                 | ((id & EMAC_STACR_PCDA_MASK) << EMAC_STACR_PCDA_SHIFT);
797         out_be32(&p->stacr, r);
798
799         /* Wait for read to complete */
800         n = 200;
801         while (!emac_phy_done(dev, (r = in_be32(&p->stacr)))) {
802                 udelay(1);
803                 if (!--n) {
804                         DBG2(dev, " -> timeout wait complete\n");
805                         goto bail;
806                 }
807         }
808
809         if (unlikely(r & EMAC_STACR_PHYE)) {
810                 DBG(dev, "mdio_read(%02x, %02x) failed" NL, id, reg);
811                 err = -EREMOTEIO;
812                 goto bail;
813         }
814
815         r = ((r >> EMAC_STACR_PHYD_SHIFT) & EMAC_STACR_PHYD_MASK);
816
817         DBG2(dev, "mdio_read -> %04x" NL, r);
818         err = 0;
819  bail:
820         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
821                 rgmii_put_mdio(dev->rgmii_dev, dev->rgmii_port);
822         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
823                 zmii_put_mdio(dev->zmii_dev, dev->zmii_port);
824         mutex_unlock(&dev->mdio_lock);
825
826         return err == 0 ? r : err;
827 }
828
829 static void __emac_mdio_write(struct emac_instance *dev, u8 id, u8 reg,
830                               u16 val)
831 {
832         struct emac_regs __iomem *p = dev->emacp;
833         u32 r = 0;
834         int n, err = -ETIMEDOUT;
835
836         mutex_lock(&dev->mdio_lock);
837
838         DBG2(dev, "mdio_write(%02x,%02x,%04x)" NL, id, reg, val);
839
840         /* Enable proper MDIO port */
841         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
842                 zmii_get_mdio(dev->zmii_dev, dev->zmii_port);
843         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
844                 rgmii_get_mdio(dev->rgmii_dev, dev->rgmii_port);
845
846         /* Wait for management interface to be idle */
847         n = 20;
848         while (!emac_phy_done(dev, in_be32(&p->stacr))) {
849                 udelay(1);
850                 if (!--n) {
851                         DBG2(dev, " -> timeout wait idle\n");
852                         goto bail;
853                 }
854         }
855
856         /* Issue write command */
857         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
858                 r = EMAC4_STACR_BASE(dev->opb_bus_freq);
859         else
860                 r = EMAC_STACR_BASE(dev->opb_bus_freq);
861         if (emac_has_feature(dev, EMAC_FTR_STACR_OC_INVERT))
862                 r |= EMAC_STACR_OC;
863         if (emac_has_feature(dev, EMAC_FTR_HAS_NEW_STACR))
864                 r |= EMACX_STACR_STAC_WRITE;
865         else
866                 r |= EMAC_STACR_STAC_WRITE;
867         r |= (reg & EMAC_STACR_PRA_MASK) |
868                 ((id & EMAC_STACR_PCDA_MASK) << EMAC_STACR_PCDA_SHIFT) |
869                 (val << EMAC_STACR_PHYD_SHIFT);
870         out_be32(&p->stacr, r);
871
872         /* Wait for write to complete */
873         n = 200;
874         while (!emac_phy_done(dev, in_be32(&p->stacr))) {
875                 udelay(1);
876                 if (!--n) {
877                         DBG2(dev, " -> timeout wait complete\n");
878                         goto bail;
879                 }
880         }
881         err = 0;
882  bail:
883         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
884                 rgmii_put_mdio(dev->rgmii_dev, dev->rgmii_port);
885         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
886                 zmii_put_mdio(dev->zmii_dev, dev->zmii_port);
887         mutex_unlock(&dev->mdio_lock);
888 }
889
890 static int emac_mdio_read(struct net_device *ndev, int id, int reg)
891 {
892         struct emac_instance *dev = netdev_priv(ndev);
893         int res;
894
895         res = __emac_mdio_read((dev->mdio_instance &&
896                                 dev->phy.gpcs_address != id) ?
897                                 dev->mdio_instance : dev,
898                                (u8) id, (u8) reg);
899         return res;
900 }
901
902 static void emac_mdio_write(struct net_device *ndev, int id, int reg, int val)
903 {
904         struct emac_instance *dev = netdev_priv(ndev);
905
906         __emac_mdio_write((dev->mdio_instance &&
907                            dev->phy.gpcs_address != id) ?
908                            dev->mdio_instance : dev,
909                           (u8) id, (u8) reg, (u16) val);
910 }
911
912 /* Tx lock BH */
913 static void __emac_set_multicast_list(struct emac_instance *dev)
914 {
915         struct emac_regs __iomem *p = dev->emacp;
916         u32 rmr = emac_iff2rmr(dev->ndev);
917
918         DBG(dev, "__multicast %08x" NL, rmr);
919
920         /* I decided to relax register access rules here to avoid
921          * full EMAC reset.
922          *
923          * There is a real problem with EMAC4 core if we use MWSW_001 bit
924          * in MR1 register and do a full EMAC reset.
925          * One TX BD status update is delayed and, after EMAC reset, it
926          * never happens, resulting in TX hung (it'll be recovered by TX
927          * timeout handler eventually, but this is just gross).
928          * So we either have to do full TX reset or try to cheat here :)
929          *
930          * The only required change is to RX mode register, so I *think* all
931          * we need is just to stop RX channel. This seems to work on all
932          * tested SoCs.                                                --ebs
933          *
934          * If we need the full reset, we might just trigger the workqueue
935          * and do it async... a bit nasty but should work --BenH
936          */
937         dev->mcast_pending = 0;
938         emac_rx_disable(dev);
939         if (rmr & EMAC_RMR_MAE)
940                 emac_hash_mc(dev);
941         out_be32(&p->rmr, rmr);
942         emac_rx_enable(dev);
943 }
944
945 /* Tx lock BH */
946 static void emac_set_multicast_list(struct net_device *ndev)
947 {
948         struct emac_instance *dev = netdev_priv(ndev);
949
950         DBG(dev, "multicast" NL);
951
952         BUG_ON(!netif_running(dev->ndev));
953
954         if (dev->no_mcast) {
955                 dev->mcast_pending = 1;
956                 return;
957         }
958         __emac_set_multicast_list(dev);
959 }
960
961 static int emac_resize_rx_ring(struct emac_instance *dev, int new_mtu)
962 {
963         int rx_sync_size = emac_rx_sync_size(new_mtu);
964         int rx_skb_size = emac_rx_skb_size(new_mtu);
965         int i, ret = 0;
966
967         mutex_lock(&dev->link_lock);
968         emac_netif_stop(dev);
969         emac_rx_disable(dev);
970         mal_disable_rx_channel(dev->mal, dev->mal_rx_chan);
971
972         if (dev->rx_sg_skb) {
973                 ++dev->estats.rx_dropped_resize;
974                 dev_kfree_skb(dev->rx_sg_skb);
975                 dev->rx_sg_skb = NULL;
976         }
977
978         /* Make a first pass over RX ring and mark BDs ready, dropping
979          * non-processed packets on the way. We need this as a separate pass
980          * to simplify error recovery in the case of allocation failure later.
981          */
982         for (i = 0; i < NUM_RX_BUFF; ++i) {
983                 if (dev->rx_desc[i].ctrl & MAL_RX_CTRL_FIRST)
984                         ++dev->estats.rx_dropped_resize;
985
986                 dev->rx_desc[i].data_len = 0;
987                 dev->rx_desc[i].ctrl = MAL_RX_CTRL_EMPTY |
988                     (i == (NUM_RX_BUFF - 1) ? MAL_RX_CTRL_WRAP : 0);
989         }
990
991         /* Reallocate RX ring only if bigger skb buffers are required */
992         if (rx_skb_size <= dev->rx_skb_size)
993                 goto skip;
994
995         /* Second pass, allocate new skbs */
996         for (i = 0; i < NUM_RX_BUFF; ++i) {
997                 struct sk_buff *skb = alloc_skb(rx_skb_size, GFP_ATOMIC);
998                 if (!skb) {
999                         ret = -ENOMEM;
1000                         goto oom;
1001                 }
1002
1003                 BUG_ON(!dev->rx_skb[i]);
1004                 dev_kfree_skb(dev->rx_skb[i]);
1005
1006                 skb_reserve(skb, EMAC_RX_SKB_HEADROOM + 2);
1007                 dev->rx_desc[i].data_ptr =
1008                     dma_map_single(&dev->ofdev->dev, skb->data - 2, rx_sync_size,
1009                                    DMA_FROM_DEVICE) + 2;
1010                 dev->rx_skb[i] = skb;
1011         }
1012  skip:
1013         /* Check if we need to change "Jumbo" bit in MR1 */
1014         if ((new_mtu > ETH_DATA_LEN) ^ (dev->ndev->mtu > ETH_DATA_LEN)) {
1015                 /* This is to prevent starting RX channel in emac_rx_enable() */
1016                 set_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1017
1018                 dev->ndev->mtu = new_mtu;
1019                 emac_full_tx_reset(dev);
1020         }
1021
1022         mal_set_rcbs(dev->mal, dev->mal_rx_chan, emac_rx_size(new_mtu));
1023  oom:
1024         /* Restart RX */
1025         clear_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1026         dev->rx_slot = 0;
1027         mal_enable_rx_channel(dev->mal, dev->mal_rx_chan);
1028         emac_rx_enable(dev);
1029         emac_netif_start(dev);
1030         mutex_unlock(&dev->link_lock);
1031
1032         return ret;
1033 }
1034
1035 /* Process ctx, rtnl_lock semaphore */
1036 static int emac_change_mtu(struct net_device *ndev, int new_mtu)
1037 {
1038         struct emac_instance *dev = netdev_priv(ndev);
1039         int ret = 0;
1040
1041         if (new_mtu < EMAC_MIN_MTU || new_mtu > dev->max_mtu)
1042                 return -EINVAL;
1043
1044         DBG(dev, "change_mtu(%d)" NL, new_mtu);
1045
1046         if (netif_running(ndev)) {
1047                 /* Check if we really need to reinitalize RX ring */
1048                 if (emac_rx_skb_size(ndev->mtu) != emac_rx_skb_size(new_mtu))
1049                         ret = emac_resize_rx_ring(dev, new_mtu);
1050         }
1051
1052         if (!ret) {
1053                 ndev->mtu = new_mtu;
1054                 dev->rx_skb_size = emac_rx_skb_size(new_mtu);
1055                 dev->rx_sync_size = emac_rx_sync_size(new_mtu);
1056         }
1057
1058         return ret;
1059 }
1060
1061 static void emac_clean_tx_ring(struct emac_instance *dev)
1062 {
1063         int i;
1064
1065         for (i = 0; i < NUM_TX_BUFF; ++i) {
1066                 if (dev->tx_skb[i]) {
1067                         dev_kfree_skb(dev->tx_skb[i]);
1068                         dev->tx_skb[i] = NULL;
1069                         if (dev->tx_desc[i].ctrl & MAL_TX_CTRL_READY)
1070                                 ++dev->estats.tx_dropped;
1071                 }
1072                 dev->tx_desc[i].ctrl = 0;
1073                 dev->tx_desc[i].data_ptr = 0;
1074         }
1075 }
1076
1077 static void emac_clean_rx_ring(struct emac_instance *dev)
1078 {
1079         int i;
1080
1081         for (i = 0; i < NUM_RX_BUFF; ++i)
1082                 if (dev->rx_skb[i]) {
1083                         dev->rx_desc[i].ctrl = 0;
1084                         dev_kfree_skb(dev->rx_skb[i]);
1085                         dev->rx_skb[i] = NULL;
1086                         dev->rx_desc[i].data_ptr = 0;
1087                 }
1088
1089         if (dev->rx_sg_skb) {
1090                 dev_kfree_skb(dev->rx_sg_skb);
1091                 dev->rx_sg_skb = NULL;
1092         }
1093 }
1094
1095 static inline int emac_alloc_rx_skb(struct emac_instance *dev, int slot,
1096                                     gfp_t flags)
1097 {
1098         struct sk_buff *skb = alloc_skb(dev->rx_skb_size, flags);
1099         if (unlikely(!skb))
1100                 return -ENOMEM;
1101
1102         dev->rx_skb[slot] = skb;
1103         dev->rx_desc[slot].data_len = 0;
1104
1105         skb_reserve(skb, EMAC_RX_SKB_HEADROOM + 2);
1106         dev->rx_desc[slot].data_ptr =
1107             dma_map_single(&dev->ofdev->dev, skb->data - 2, dev->rx_sync_size,
1108                            DMA_FROM_DEVICE) + 2;
1109         wmb();
1110         dev->rx_desc[slot].ctrl = MAL_RX_CTRL_EMPTY |
1111             (slot == (NUM_RX_BUFF - 1) ? MAL_RX_CTRL_WRAP : 0);
1112
1113         return 0;
1114 }
1115
1116 static void emac_print_link_status(struct emac_instance *dev)
1117 {
1118         if (netif_carrier_ok(dev->ndev))
1119                 printk(KERN_INFO "%s: link is up, %d %s%s\n",
1120                        dev->ndev->name, dev->phy.speed,
1121                        dev->phy.duplex == DUPLEX_FULL ? "FDX" : "HDX",
1122                        dev->phy.pause ? ", pause enabled" :
1123                        dev->phy.asym_pause ? ", asymmetric pause enabled" : "");
1124         else
1125                 printk(KERN_INFO "%s: link is down\n", dev->ndev->name);
1126 }
1127
1128 /* Process ctx, rtnl_lock semaphore */
1129 static int emac_open(struct net_device *ndev)
1130 {
1131         struct emac_instance *dev = netdev_priv(ndev);
1132         int err, i;
1133
1134         DBG(dev, "open" NL);
1135
1136         /* Setup error IRQ handler */
1137         err = request_irq(dev->emac_irq, emac_irq, 0, "EMAC", dev);
1138         if (err) {
1139                 printk(KERN_ERR "%s: failed to request IRQ %d\n",
1140                        ndev->name, dev->emac_irq);
1141                 return err;
1142         }
1143
1144         /* Allocate RX ring */
1145         for (i = 0; i < NUM_RX_BUFF; ++i)
1146                 if (emac_alloc_rx_skb(dev, i, GFP_KERNEL)) {
1147                         printk(KERN_ERR "%s: failed to allocate RX ring\n",
1148                                ndev->name);
1149                         goto oom;
1150                 }
1151
1152         dev->tx_cnt = dev->tx_slot = dev->ack_slot = dev->rx_slot = 0;
1153         clear_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1154         dev->rx_sg_skb = NULL;
1155
1156         mutex_lock(&dev->link_lock);
1157         dev->opened = 1;
1158
1159         /* Start PHY polling now.
1160          */
1161         if (dev->phy.address >= 0) {
1162                 int link_poll_interval;
1163                 if (dev->phy.def->ops->poll_link(&dev->phy)) {
1164                         dev->phy.def->ops->read_link(&dev->phy);
1165                         emac_rx_clk_default(dev);
1166                         netif_carrier_on(dev->ndev);
1167                         link_poll_interval = PHY_POLL_LINK_ON;
1168                 } else {
1169                         emac_rx_clk_tx(dev);
1170                         netif_carrier_off(dev->ndev);
1171                         link_poll_interval = PHY_POLL_LINK_OFF;
1172                 }
1173                 dev->link_polling = 1;
1174                 wmb();
1175                 schedule_delayed_work(&dev->link_work, link_poll_interval);
1176                 emac_print_link_status(dev);
1177         } else
1178                 netif_carrier_on(dev->ndev);
1179
1180         /* Required for Pause packet support in EMAC */
1181         dev_mc_add_global(ndev, default_mcast_addr);
1182
1183         emac_configure(dev);
1184         mal_poll_add(dev->mal, &dev->commac);
1185         mal_enable_tx_channel(dev->mal, dev->mal_tx_chan);
1186         mal_set_rcbs(dev->mal, dev->mal_rx_chan, emac_rx_size(ndev->mtu));
1187         mal_enable_rx_channel(dev->mal, dev->mal_rx_chan);
1188         emac_tx_enable(dev);
1189         emac_rx_enable(dev);
1190         emac_netif_start(dev);
1191
1192         mutex_unlock(&dev->link_lock);
1193
1194         return 0;
1195  oom:
1196         emac_clean_rx_ring(dev);
1197         free_irq(dev->emac_irq, dev);
1198
1199         return -ENOMEM;
1200 }
1201
1202 /* BHs disabled */
1203 #if 0
1204 static int emac_link_differs(struct emac_instance *dev)
1205 {
1206         u32 r = in_be32(&dev->emacp->mr1);
1207
1208         int duplex = r & EMAC_MR1_FDE ? DUPLEX_FULL : DUPLEX_HALF;
1209         int speed, pause, asym_pause;
1210
1211         if (r & EMAC_MR1_MF_1000)
1212                 speed = SPEED_1000;
1213         else if (r & EMAC_MR1_MF_100)
1214                 speed = SPEED_100;
1215         else
1216                 speed = SPEED_10;
1217
1218         switch (r & (EMAC_MR1_EIFC | EMAC_MR1_APP)) {
1219         case (EMAC_MR1_EIFC | EMAC_MR1_APP):
1220                 pause = 1;
1221                 asym_pause = 0;
1222                 break;
1223         case EMAC_MR1_APP:
1224                 pause = 0;
1225                 asym_pause = 1;
1226                 break;
1227         default:
1228                 pause = asym_pause = 0;
1229         }
1230         return speed != dev->phy.speed || duplex != dev->phy.duplex ||
1231             pause != dev->phy.pause || asym_pause != dev->phy.asym_pause;
1232 }
1233 #endif
1234
1235 static void emac_link_timer(struct work_struct *work)
1236 {
1237         struct emac_instance *dev =
1238                 container_of(to_delayed_work(work),
1239                              struct emac_instance, link_work);
1240         int link_poll_interval;
1241
1242         mutex_lock(&dev->link_lock);
1243         DBG2(dev, "link timer" NL);
1244
1245         if (!dev->opened)
1246                 goto bail;
1247
1248         if (dev->phy.def->ops->poll_link(&dev->phy)) {
1249                 if (!netif_carrier_ok(dev->ndev)) {
1250                         emac_rx_clk_default(dev);
1251                         /* Get new link parameters */
1252                         dev->phy.def->ops->read_link(&dev->phy);
1253
1254                         netif_carrier_on(dev->ndev);
1255                         emac_netif_stop(dev);
1256                         emac_full_tx_reset(dev);
1257                         emac_netif_start(dev);
1258                         emac_print_link_status(dev);
1259                 }
1260                 link_poll_interval = PHY_POLL_LINK_ON;
1261         } else {
1262                 if (netif_carrier_ok(dev->ndev)) {
1263                         emac_rx_clk_tx(dev);
1264                         netif_carrier_off(dev->ndev);
1265                         netif_tx_disable(dev->ndev);
1266                         emac_reinitialize(dev);
1267                         emac_print_link_status(dev);
1268                 }
1269                 link_poll_interval = PHY_POLL_LINK_OFF;
1270         }
1271         schedule_delayed_work(&dev->link_work, link_poll_interval);
1272  bail:
1273         mutex_unlock(&dev->link_lock);
1274 }
1275
1276 static void emac_force_link_update(struct emac_instance *dev)
1277 {
1278         netif_carrier_off(dev->ndev);
1279         smp_rmb();
1280         if (dev->link_polling) {
1281                 cancel_rearming_delayed_work(&dev->link_work);
1282                 if (dev->link_polling)
1283                         schedule_delayed_work(&dev->link_work,  PHY_POLL_LINK_OFF);
1284         }
1285 }
1286
1287 /* Process ctx, rtnl_lock semaphore */
1288 static int emac_close(struct net_device *ndev)
1289 {
1290         struct emac_instance *dev = netdev_priv(ndev);
1291
1292         DBG(dev, "close" NL);
1293
1294         if (dev->phy.address >= 0) {
1295                 dev->link_polling = 0;
1296                 cancel_rearming_delayed_work(&dev->link_work);
1297         }
1298         mutex_lock(&dev->link_lock);
1299         emac_netif_stop(dev);
1300         dev->opened = 0;
1301         mutex_unlock(&dev->link_lock);
1302
1303         emac_rx_disable(dev);
1304         emac_tx_disable(dev);
1305         mal_disable_rx_channel(dev->mal, dev->mal_rx_chan);
1306         mal_disable_tx_channel(dev->mal, dev->mal_tx_chan);
1307         mal_poll_del(dev->mal, &dev->commac);
1308
1309         emac_clean_tx_ring(dev);
1310         emac_clean_rx_ring(dev);
1311
1312         free_irq(dev->emac_irq, dev);
1313
1314         netif_carrier_off(ndev);
1315
1316         return 0;
1317 }
1318
1319 static inline u16 emac_tx_csum(struct emac_instance *dev,
1320                                struct sk_buff *skb)
1321 {
1322         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH) &&
1323                 (skb->ip_summed == CHECKSUM_PARTIAL)) {
1324                 ++dev->stats.tx_packets_csum;
1325                 return EMAC_TX_CTRL_TAH_CSUM;
1326         }
1327         return 0;
1328 }
1329
1330 static inline int emac_xmit_finish(struct emac_instance *dev, int len)
1331 {
1332         struct emac_regs __iomem *p = dev->emacp;
1333         struct net_device *ndev = dev->ndev;
1334
1335         /* Send the packet out. If the if makes a significant perf
1336          * difference, then we can store the TMR0 value in "dev"
1337          * instead
1338          */
1339         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
1340                 out_be32(&p->tmr0, EMAC4_TMR0_XMIT);
1341         else
1342                 out_be32(&p->tmr0, EMAC_TMR0_XMIT);
1343
1344         if (unlikely(++dev->tx_cnt == NUM_TX_BUFF)) {
1345                 netif_stop_queue(ndev);
1346                 DBG2(dev, "stopped TX queue" NL);
1347         }
1348
1349         ndev->trans_start = jiffies;
1350         ++dev->stats.tx_packets;
1351         dev->stats.tx_bytes += len;
1352
1353         return NETDEV_TX_OK;
1354 }
1355
1356 /* Tx lock BH */
1357 static int emac_start_xmit(struct sk_buff *skb, struct net_device *ndev)
1358 {
1359         struct emac_instance *dev = netdev_priv(ndev);
1360         unsigned int len = skb->len;
1361         int slot;
1362
1363         u16 ctrl = EMAC_TX_CTRL_GFCS | EMAC_TX_CTRL_GP | MAL_TX_CTRL_READY |
1364             MAL_TX_CTRL_LAST | emac_tx_csum(dev, skb);
1365
1366         slot = dev->tx_slot++;
1367         if (dev->tx_slot == NUM_TX_BUFF) {
1368                 dev->tx_slot = 0;
1369                 ctrl |= MAL_TX_CTRL_WRAP;
1370         }
1371
1372         DBG2(dev, "xmit(%u) %d" NL, len, slot);
1373
1374         dev->tx_skb[slot] = skb;
1375         dev->tx_desc[slot].data_ptr = dma_map_single(&dev->ofdev->dev,
1376                                                      skb->data, len,
1377                                                      DMA_TO_DEVICE);
1378         dev->tx_desc[slot].data_len = (u16) len;
1379         wmb();
1380         dev->tx_desc[slot].ctrl = ctrl;
1381
1382         return emac_xmit_finish(dev, len);
1383 }
1384
1385 static inline int emac_xmit_split(struct emac_instance *dev, int slot,
1386                                   u32 pd, int len, int last, u16 base_ctrl)
1387 {
1388         while (1) {
1389                 u16 ctrl = base_ctrl;
1390                 int chunk = min(len, MAL_MAX_TX_SIZE);
1391                 len -= chunk;
1392
1393                 slot = (slot + 1) % NUM_TX_BUFF;
1394
1395                 if (last && !len)
1396                         ctrl |= MAL_TX_CTRL_LAST;
1397                 if (slot == NUM_TX_BUFF - 1)
1398                         ctrl |= MAL_TX_CTRL_WRAP;
1399
1400                 dev->tx_skb[slot] = NULL;
1401                 dev->tx_desc[slot].data_ptr = pd;
1402                 dev->tx_desc[slot].data_len = (u16) chunk;
1403                 dev->tx_desc[slot].ctrl = ctrl;
1404                 ++dev->tx_cnt;
1405
1406                 if (!len)
1407                         break;
1408
1409                 pd += chunk;
1410         }
1411         return slot;
1412 }
1413
1414 /* Tx lock BH disabled (SG version for TAH equipped EMACs) */
1415 static int emac_start_xmit_sg(struct sk_buff *skb, struct net_device *ndev)
1416 {
1417         struct emac_instance *dev = netdev_priv(ndev);
1418         int nr_frags = skb_shinfo(skb)->nr_frags;
1419         int len = skb->len, chunk;
1420         int slot, i;
1421         u16 ctrl;
1422         u32 pd;
1423
1424         /* This is common "fast" path */
1425         if (likely(!nr_frags && len <= MAL_MAX_TX_SIZE))
1426                 return emac_start_xmit(skb, ndev);
1427
1428         len -= skb->data_len;
1429
1430         /* Note, this is only an *estimation*, we can still run out of empty
1431          * slots because of the additional fragmentation into
1432          * MAL_MAX_TX_SIZE-sized chunks
1433          */
1434         if (unlikely(dev->tx_cnt + nr_frags + mal_tx_chunks(len) > NUM_TX_BUFF))
1435                 goto stop_queue;
1436
1437         ctrl = EMAC_TX_CTRL_GFCS | EMAC_TX_CTRL_GP | MAL_TX_CTRL_READY |
1438             emac_tx_csum(dev, skb);
1439         slot = dev->tx_slot;
1440
1441         /* skb data */
1442         dev->tx_skb[slot] = NULL;
1443         chunk = min(len, MAL_MAX_TX_SIZE);
1444         dev->tx_desc[slot].data_ptr = pd =
1445             dma_map_single(&dev->ofdev->dev, skb->data, len, DMA_TO_DEVICE);
1446         dev->tx_desc[slot].data_len = (u16) chunk;
1447         len -= chunk;
1448         if (unlikely(len))
1449                 slot = emac_xmit_split(dev, slot, pd + chunk, len, !nr_frags,
1450                                        ctrl);
1451         /* skb fragments */
1452         for (i = 0; i < nr_frags; ++i) {
1453                 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i];
1454                 len = frag->size;
1455
1456                 if (unlikely(dev->tx_cnt + mal_tx_chunks(len) >= NUM_TX_BUFF))
1457                         goto undo_frame;
1458
1459                 pd = dma_map_page(&dev->ofdev->dev, frag->page, frag->page_offset, len,
1460                                   DMA_TO_DEVICE);
1461
1462                 slot = emac_xmit_split(dev, slot, pd, len, i == nr_frags - 1,
1463                                        ctrl);
1464         }
1465
1466         DBG2(dev, "xmit_sg(%u) %d - %d" NL, skb->len, dev->tx_slot, slot);
1467
1468         /* Attach skb to the last slot so we don't release it too early */
1469         dev->tx_skb[slot] = skb;
1470
1471         /* Send the packet out */
1472         if (dev->tx_slot == NUM_TX_BUFF - 1)
1473                 ctrl |= MAL_TX_CTRL_WRAP;
1474         wmb();
1475         dev->tx_desc[dev->tx_slot].ctrl = ctrl;
1476         dev->tx_slot = (slot + 1) % NUM_TX_BUFF;
1477
1478         return emac_xmit_finish(dev, skb->len);
1479
1480  undo_frame:
1481         /* Well, too bad. Our previous estimation was overly optimistic.
1482          * Undo everything.
1483          */
1484         while (slot != dev->tx_slot) {
1485                 dev->tx_desc[slot].ctrl = 0;
1486                 --dev->tx_cnt;
1487                 if (--slot < 0)
1488                         slot = NUM_TX_BUFF - 1;
1489         }
1490         ++dev->estats.tx_undo;
1491
1492  stop_queue:
1493         netif_stop_queue(ndev);
1494         DBG2(dev, "stopped TX queue" NL);
1495         return NETDEV_TX_BUSY;
1496 }
1497
1498 /* Tx lock BHs */
1499 static void emac_parse_tx_error(struct emac_instance *dev, u16 ctrl)
1500 {
1501         struct emac_error_stats *st = &dev->estats;
1502
1503         DBG(dev, "BD TX error %04x" NL, ctrl);
1504
1505         ++st->tx_bd_errors;
1506         if (ctrl & EMAC_TX_ST_BFCS)
1507                 ++st->tx_bd_bad_fcs;
1508         if (ctrl & EMAC_TX_ST_LCS)
1509                 ++st->tx_bd_carrier_loss;
1510         if (ctrl & EMAC_TX_ST_ED)
1511                 ++st->tx_bd_excessive_deferral;
1512         if (ctrl & EMAC_TX_ST_EC)
1513                 ++st->tx_bd_excessive_collisions;
1514         if (ctrl & EMAC_TX_ST_LC)
1515                 ++st->tx_bd_late_collision;
1516         if (ctrl & EMAC_TX_ST_MC)
1517                 ++st->tx_bd_multple_collisions;
1518         if (ctrl & EMAC_TX_ST_SC)
1519                 ++st->tx_bd_single_collision;
1520         if (ctrl & EMAC_TX_ST_UR)
1521                 ++st->tx_bd_underrun;
1522         if (ctrl & EMAC_TX_ST_SQE)
1523                 ++st->tx_bd_sqe;
1524 }
1525
1526 static void emac_poll_tx(void *param)
1527 {
1528         struct emac_instance *dev = param;
1529         u32 bad_mask;
1530
1531         DBG2(dev, "poll_tx, %d %d" NL, dev->tx_cnt, dev->ack_slot);
1532
1533         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
1534                 bad_mask = EMAC_IS_BAD_TX_TAH;
1535         else
1536                 bad_mask = EMAC_IS_BAD_TX;
1537
1538         netif_tx_lock_bh(dev->ndev);
1539         if (dev->tx_cnt) {
1540                 u16 ctrl;
1541                 int slot = dev->ack_slot, n = 0;
1542         again:
1543                 ctrl = dev->tx_desc[slot].ctrl;
1544                 if (!(ctrl & MAL_TX_CTRL_READY)) {
1545                         struct sk_buff *skb = dev->tx_skb[slot];
1546                         ++n;
1547
1548                         if (skb) {
1549                                 dev_kfree_skb(skb);
1550                                 dev->tx_skb[slot] = NULL;
1551                         }
1552                         slot = (slot + 1) % NUM_TX_BUFF;
1553
1554                         if (unlikely(ctrl & bad_mask))
1555                                 emac_parse_tx_error(dev, ctrl);
1556
1557                         if (--dev->tx_cnt)
1558                                 goto again;
1559                 }
1560                 if (n) {
1561                         dev->ack_slot = slot;
1562                         if (netif_queue_stopped(dev->ndev) &&
1563                             dev->tx_cnt < EMAC_TX_WAKEUP_THRESH)
1564                                 netif_wake_queue(dev->ndev);
1565
1566                         DBG2(dev, "tx %d pkts" NL, n);
1567                 }
1568         }
1569         netif_tx_unlock_bh(dev->ndev);
1570 }
1571
1572 static inline void emac_recycle_rx_skb(struct emac_instance *dev, int slot,
1573                                        int len)
1574 {
1575         struct sk_buff *skb = dev->rx_skb[slot];
1576
1577         DBG2(dev, "recycle %d %d" NL, slot, len);
1578
1579         if (len)
1580                 dma_map_single(&dev->ofdev->dev, skb->data - 2,
1581                                EMAC_DMA_ALIGN(len + 2), DMA_FROM_DEVICE);
1582
1583         dev->rx_desc[slot].data_len = 0;
1584         wmb();
1585         dev->rx_desc[slot].ctrl = MAL_RX_CTRL_EMPTY |
1586             (slot == (NUM_RX_BUFF - 1) ? MAL_RX_CTRL_WRAP : 0);
1587 }
1588
1589 static void emac_parse_rx_error(struct emac_instance *dev, u16 ctrl)
1590 {
1591         struct emac_error_stats *st = &dev->estats;
1592
1593         DBG(dev, "BD RX error %04x" NL, ctrl);
1594
1595         ++st->rx_bd_errors;
1596         if (ctrl & EMAC_RX_ST_OE)
1597                 ++st->rx_bd_overrun;
1598         if (ctrl & EMAC_RX_ST_BP)
1599                 ++st->rx_bd_bad_packet;
1600         if (ctrl & EMAC_RX_ST_RP)
1601                 ++st->rx_bd_runt_packet;
1602         if (ctrl & EMAC_RX_ST_SE)
1603                 ++st->rx_bd_short_event;
1604         if (ctrl & EMAC_RX_ST_AE)
1605                 ++st->rx_bd_alignment_error;
1606         if (ctrl & EMAC_RX_ST_BFCS)
1607                 ++st->rx_bd_bad_fcs;
1608         if (ctrl & EMAC_RX_ST_PTL)
1609                 ++st->rx_bd_packet_too_long;
1610         if (ctrl & EMAC_RX_ST_ORE)
1611                 ++st->rx_bd_out_of_range;
1612         if (ctrl & EMAC_RX_ST_IRE)
1613                 ++st->rx_bd_in_range;
1614 }
1615
1616 static inline void emac_rx_csum(struct emac_instance *dev,
1617                                 struct sk_buff *skb, u16 ctrl)
1618 {
1619 #ifdef CONFIG_IBM_NEW_EMAC_TAH
1620         if (!ctrl && dev->tah_dev) {
1621                 skb->ip_summed = CHECKSUM_UNNECESSARY;
1622                 ++dev->stats.rx_packets_csum;
1623         }
1624 #endif
1625 }
1626
1627 static inline int emac_rx_sg_append(struct emac_instance *dev, int slot)
1628 {
1629         if (likely(dev->rx_sg_skb != NULL)) {
1630                 int len = dev->rx_desc[slot].data_len;
1631                 int tot_len = dev->rx_sg_skb->len + len;
1632
1633                 if (unlikely(tot_len + 2 > dev->rx_skb_size)) {
1634                         ++dev->estats.rx_dropped_mtu;
1635                         dev_kfree_skb(dev->rx_sg_skb);
1636                         dev->rx_sg_skb = NULL;
1637                 } else {
1638                         cacheable_memcpy(skb_tail_pointer(dev->rx_sg_skb),
1639                                          dev->rx_skb[slot]->data, len);
1640                         skb_put(dev->rx_sg_skb, len);
1641                         emac_recycle_rx_skb(dev, slot, len);
1642                         return 0;
1643                 }
1644         }
1645         emac_recycle_rx_skb(dev, slot, 0);
1646         return -1;
1647 }
1648
1649 /* NAPI poll context */
1650 static int emac_poll_rx(void *param, int budget)
1651 {
1652         struct emac_instance *dev = param;
1653         int slot = dev->rx_slot, received = 0;
1654
1655         DBG2(dev, "poll_rx(%d)" NL, budget);
1656
1657  again:
1658         while (budget > 0) {
1659                 int len;
1660                 struct sk_buff *skb;
1661                 u16 ctrl = dev->rx_desc[slot].ctrl;
1662
1663                 if (ctrl & MAL_RX_CTRL_EMPTY)
1664                         break;
1665
1666                 skb = dev->rx_skb[slot];
1667                 mb();
1668                 len = dev->rx_desc[slot].data_len;
1669
1670                 if (unlikely(!MAL_IS_SINGLE_RX(ctrl)))
1671                         goto sg;
1672
1673                 ctrl &= EMAC_BAD_RX_MASK;
1674                 if (unlikely(ctrl && ctrl != EMAC_RX_TAH_BAD_CSUM)) {
1675                         emac_parse_rx_error(dev, ctrl);
1676                         ++dev->estats.rx_dropped_error;
1677                         emac_recycle_rx_skb(dev, slot, 0);
1678                         len = 0;
1679                         goto next;
1680                 }
1681
1682                 if (len < ETH_HLEN) {
1683                         ++dev->estats.rx_dropped_stack;
1684                         emac_recycle_rx_skb(dev, slot, len);
1685                         goto next;
1686                 }
1687
1688                 if (len && len < EMAC_RX_COPY_THRESH) {
1689                         struct sk_buff *copy_skb =
1690                             alloc_skb(len + EMAC_RX_SKB_HEADROOM + 2, GFP_ATOMIC);
1691                         if (unlikely(!copy_skb))
1692                                 goto oom;
1693
1694                         skb_reserve(copy_skb, EMAC_RX_SKB_HEADROOM + 2);
1695                         cacheable_memcpy(copy_skb->data - 2, skb->data - 2,
1696                                          len + 2);
1697                         emac_recycle_rx_skb(dev, slot, len);
1698                         skb = copy_skb;
1699                 } else if (unlikely(emac_alloc_rx_skb(dev, slot, GFP_ATOMIC)))
1700                         goto oom;
1701
1702                 skb_put(skb, len);
1703         push_packet:
1704                 skb->protocol = eth_type_trans(skb, dev->ndev);
1705                 emac_rx_csum(dev, skb, ctrl);
1706
1707                 if (unlikely(netif_receive_skb(skb) == NET_RX_DROP))
1708                         ++dev->estats.rx_dropped_stack;
1709         next:
1710                 ++dev->stats.rx_packets;
1711         skip:
1712                 dev->stats.rx_bytes += len;
1713                 slot = (slot + 1) % NUM_RX_BUFF;
1714                 --budget;
1715                 ++received;
1716                 continue;
1717         sg:
1718                 if (ctrl & MAL_RX_CTRL_FIRST) {
1719                         BUG_ON(dev->rx_sg_skb);
1720                         if (unlikely(emac_alloc_rx_skb(dev, slot, GFP_ATOMIC))) {
1721                                 DBG(dev, "rx OOM %d" NL, slot);
1722                                 ++dev->estats.rx_dropped_oom;
1723                                 emac_recycle_rx_skb(dev, slot, 0);
1724                         } else {
1725                                 dev->rx_sg_skb = skb;
1726                                 skb_put(skb, len);
1727                         }
1728                 } else if (!emac_rx_sg_append(dev, slot) &&
1729                            (ctrl & MAL_RX_CTRL_LAST)) {
1730
1731                         skb = dev->rx_sg_skb;
1732                         dev->rx_sg_skb = NULL;
1733
1734                         ctrl &= EMAC_BAD_RX_MASK;
1735                         if (unlikely(ctrl && ctrl != EMAC_RX_TAH_BAD_CSUM)) {
1736                                 emac_parse_rx_error(dev, ctrl);
1737                                 ++dev->estats.rx_dropped_error;
1738                                 dev_kfree_skb(skb);
1739                                 len = 0;
1740                         } else
1741                                 goto push_packet;
1742                 }
1743                 goto skip;
1744         oom:
1745                 DBG(dev, "rx OOM %d" NL, slot);
1746                 /* Drop the packet and recycle skb */
1747                 ++dev->estats.rx_dropped_oom;
1748                 emac_recycle_rx_skb(dev, slot, 0);
1749                 goto next;
1750         }
1751
1752         if (received) {
1753                 DBG2(dev, "rx %d BDs" NL, received);
1754                 dev->rx_slot = slot;
1755         }
1756
1757         if (unlikely(budget && test_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags))) {
1758                 mb();
1759                 if (!(dev->rx_desc[slot].ctrl & MAL_RX_CTRL_EMPTY)) {
1760                         DBG2(dev, "rx restart" NL);
1761                         received = 0;
1762                         goto again;
1763                 }
1764
1765                 if (dev->rx_sg_skb) {
1766                         DBG2(dev, "dropping partial rx packet" NL);
1767                         ++dev->estats.rx_dropped_error;
1768                         dev_kfree_skb(dev->rx_sg_skb);
1769                         dev->rx_sg_skb = NULL;
1770                 }
1771
1772                 clear_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1773                 mal_enable_rx_channel(dev->mal, dev->mal_rx_chan);
1774                 emac_rx_enable(dev);
1775                 dev->rx_slot = 0;
1776         }
1777         return received;
1778 }
1779
1780 /* NAPI poll context */
1781 static int emac_peek_rx(void *param)
1782 {
1783         struct emac_instance *dev = param;
1784
1785         return !(dev->rx_desc[dev->rx_slot].ctrl & MAL_RX_CTRL_EMPTY);
1786 }
1787
1788 /* NAPI poll context */
1789 static int emac_peek_rx_sg(void *param)
1790 {
1791         struct emac_instance *dev = param;
1792
1793         int slot = dev->rx_slot;
1794         while (1) {
1795                 u16 ctrl = dev->rx_desc[slot].ctrl;
1796                 if (ctrl & MAL_RX_CTRL_EMPTY)
1797                         return 0;
1798                 else if (ctrl & MAL_RX_CTRL_LAST)
1799                         return 1;
1800
1801                 slot = (slot + 1) % NUM_RX_BUFF;
1802
1803                 /* I'm just being paranoid here :) */
1804                 if (unlikely(slot == dev->rx_slot))
1805                         return 0;
1806         }
1807 }
1808
1809 /* Hard IRQ */
1810 static void emac_rxde(void *param)
1811 {
1812         struct emac_instance *dev = param;
1813
1814         ++dev->estats.rx_stopped;
1815         emac_rx_disable_async(dev);
1816 }
1817
1818 /* Hard IRQ */
1819 static irqreturn_t emac_irq(int irq, void *dev_instance)
1820 {
1821         struct emac_instance *dev = dev_instance;
1822         struct emac_regs __iomem *p = dev->emacp;
1823         struct emac_error_stats *st = &dev->estats;
1824         u32 isr;
1825
1826         spin_lock(&dev->lock);
1827
1828         isr = in_be32(&p->isr);
1829         out_be32(&p->isr, isr);
1830
1831         DBG(dev, "isr = %08x" NL, isr);
1832
1833         if (isr & EMAC4_ISR_TXPE)
1834                 ++st->tx_parity;
1835         if (isr & EMAC4_ISR_RXPE)
1836                 ++st->rx_parity;
1837         if (isr & EMAC4_ISR_TXUE)
1838                 ++st->tx_underrun;
1839         if (isr & EMAC4_ISR_RXOE)
1840                 ++st->rx_fifo_overrun;
1841         if (isr & EMAC_ISR_OVR)
1842                 ++st->rx_overrun;
1843         if (isr & EMAC_ISR_BP)
1844                 ++st->rx_bad_packet;
1845         if (isr & EMAC_ISR_RP)
1846                 ++st->rx_runt_packet;
1847         if (isr & EMAC_ISR_SE)
1848                 ++st->rx_short_event;
1849         if (isr & EMAC_ISR_ALE)
1850                 ++st->rx_alignment_error;
1851         if (isr & EMAC_ISR_BFCS)
1852                 ++st->rx_bad_fcs;
1853         if (isr & EMAC_ISR_PTLE)
1854                 ++st->rx_packet_too_long;
1855         if (isr & EMAC_ISR_ORE)
1856                 ++st->rx_out_of_range;
1857         if (isr & EMAC_ISR_IRE)
1858                 ++st->rx_in_range;
1859         if (isr & EMAC_ISR_SQE)
1860                 ++st->tx_sqe;
1861         if (isr & EMAC_ISR_TE)
1862                 ++st->tx_errors;
1863
1864         spin_unlock(&dev->lock);
1865
1866         return IRQ_HANDLED;
1867 }
1868
1869 static struct net_device_stats *emac_stats(struct net_device *ndev)
1870 {
1871         struct emac_instance *dev = netdev_priv(ndev);
1872         struct emac_stats *st = &dev->stats;
1873         struct emac_error_stats *est = &dev->estats;
1874         struct net_device_stats *nst = &dev->nstats;
1875         unsigned long flags;
1876
1877         DBG2(dev, "stats" NL);
1878
1879         /* Compute "legacy" statistics */
1880         spin_lock_irqsave(&dev->lock, flags);
1881         nst->rx_packets = (unsigned long)st->rx_packets;
1882         nst->rx_bytes = (unsigned long)st->rx_bytes;
1883         nst->tx_packets = (unsigned long)st->tx_packets;
1884         nst->tx_bytes = (unsigned long)st->tx_bytes;
1885         nst->rx_dropped = (unsigned long)(est->rx_dropped_oom +
1886                                           est->rx_dropped_error +
1887                                           est->rx_dropped_resize +
1888                                           est->rx_dropped_mtu);
1889         nst->tx_dropped = (unsigned long)est->tx_dropped;
1890
1891         nst->rx_errors = (unsigned long)est->rx_bd_errors;
1892         nst->rx_fifo_errors = (unsigned long)(est->rx_bd_overrun +
1893                                               est->rx_fifo_overrun +
1894                                               est->rx_overrun);
1895         nst->rx_frame_errors = (unsigned long)(est->rx_bd_alignment_error +
1896                                                est->rx_alignment_error);
1897         nst->rx_crc_errors = (unsigned long)(est->rx_bd_bad_fcs +
1898                                              est->rx_bad_fcs);
1899         nst->rx_length_errors = (unsigned long)(est->rx_bd_runt_packet +
1900                                                 est->rx_bd_short_event +
1901                                                 est->rx_bd_packet_too_long +
1902                                                 est->rx_bd_out_of_range +
1903                                                 est->rx_bd_in_range +
1904                                                 est->rx_runt_packet +
1905                                                 est->rx_short_event +
1906                                                 est->rx_packet_too_long +
1907                                                 est->rx_out_of_range +
1908                                                 est->rx_in_range);
1909
1910         nst->tx_errors = (unsigned long)(est->tx_bd_errors + est->tx_errors);
1911         nst->tx_fifo_errors = (unsigned long)(est->tx_bd_underrun +
1912                                               est->tx_underrun);
1913         nst->tx_carrier_errors = (unsigned long)est->tx_bd_carrier_loss;
1914         nst->collisions = (unsigned long)(est->tx_bd_excessive_deferral +
1915                                           est->tx_bd_excessive_collisions +
1916                                           est->tx_bd_late_collision +
1917                                           est->tx_bd_multple_collisions);
1918         spin_unlock_irqrestore(&dev->lock, flags);
1919         return nst;
1920 }
1921
1922 static struct mal_commac_ops emac_commac_ops = {
1923         .poll_tx = &emac_poll_tx,
1924         .poll_rx = &emac_poll_rx,
1925         .peek_rx = &emac_peek_rx,
1926         .rxde = &emac_rxde,
1927 };
1928
1929 static struct mal_commac_ops emac_commac_sg_ops = {
1930         .poll_tx = &emac_poll_tx,
1931         .poll_rx = &emac_poll_rx,
1932         .peek_rx = &emac_peek_rx_sg,
1933         .rxde = &emac_rxde,
1934 };
1935
1936 /* Ethtool support */
1937 static int emac_ethtool_get_settings(struct net_device *ndev,
1938                                      struct ethtool_cmd *cmd)
1939 {
1940         struct emac_instance *dev = netdev_priv(ndev);
1941
1942         cmd->supported = dev->phy.features;
1943         cmd->port = PORT_MII;
1944         cmd->phy_address = dev->phy.address;
1945         cmd->transceiver =
1946             dev->phy.address >= 0 ? XCVR_EXTERNAL : XCVR_INTERNAL;
1947
1948         mutex_lock(&dev->link_lock);
1949         cmd->advertising = dev->phy.advertising;
1950         cmd->autoneg = dev->phy.autoneg;
1951         cmd->speed = dev->phy.speed;
1952         cmd->duplex = dev->phy.duplex;
1953         mutex_unlock(&dev->link_lock);
1954
1955         return 0;
1956 }
1957
1958 static int emac_ethtool_set_settings(struct net_device *ndev,
1959                                      struct ethtool_cmd *cmd)
1960 {
1961         struct emac_instance *dev = netdev_priv(ndev);
1962         u32 f = dev->phy.features;
1963
1964         DBG(dev, "set_settings(%d, %d, %d, 0x%08x)" NL,
1965             cmd->autoneg, cmd->speed, cmd->duplex, cmd->advertising);
1966
1967         /* Basic sanity checks */
1968         if (dev->phy.address < 0)
1969                 return -EOPNOTSUPP;
1970         if (cmd->autoneg != AUTONEG_ENABLE && cmd->autoneg != AUTONEG_DISABLE)
1971                 return -EINVAL;
1972         if (cmd->autoneg == AUTONEG_ENABLE && cmd->advertising == 0)
1973                 return -EINVAL;
1974         if (cmd->duplex != DUPLEX_HALF && cmd->duplex != DUPLEX_FULL)
1975                 return -EINVAL;
1976
1977         if (cmd->autoneg == AUTONEG_DISABLE) {
1978                 switch (cmd->speed) {
1979                 case SPEED_10:
1980                         if (cmd->duplex == DUPLEX_HALF &&
1981                             !(f & SUPPORTED_10baseT_Half))
1982                                 return -EINVAL;
1983                         if (cmd->duplex == DUPLEX_FULL &&
1984                             !(f & SUPPORTED_10baseT_Full))
1985                                 return -EINVAL;
1986                         break;
1987                 case SPEED_100:
1988                         if (cmd->duplex == DUPLEX_HALF &&
1989                             !(f & SUPPORTED_100baseT_Half))
1990                                 return -EINVAL;
1991                         if (cmd->duplex == DUPLEX_FULL &&
1992                             !(f & SUPPORTED_100baseT_Full))
1993                                 return -EINVAL;
1994                         break;
1995                 case SPEED_1000:
1996                         if (cmd->duplex == DUPLEX_HALF &&
1997                             !(f & SUPPORTED_1000baseT_Half))
1998                                 return -EINVAL;
1999                         if (cmd->duplex == DUPLEX_FULL &&
2000                             !(f & SUPPORTED_1000baseT_Full))
2001                                 return -EINVAL;
2002                         break;
2003                 default:
2004                         return -EINVAL;
2005                 }
2006
2007                 mutex_lock(&dev->link_lock);
2008                 dev->phy.def->ops->setup_forced(&dev->phy, cmd->speed,
2009                                                 cmd->duplex);
2010                 mutex_unlock(&dev->link_lock);
2011
2012         } else {
2013                 if (!(f & SUPPORTED_Autoneg))
2014                         return -EINVAL;
2015
2016                 mutex_lock(&dev->link_lock);
2017                 dev->phy.def->ops->setup_aneg(&dev->phy,
2018                                               (cmd->advertising & f) |
2019                                               (dev->phy.advertising &
2020                                                (ADVERTISED_Pause |
2021                                                 ADVERTISED_Asym_Pause)));
2022                 mutex_unlock(&dev->link_lock);
2023         }
2024         emac_force_link_update(dev);
2025
2026         return 0;
2027 }
2028
2029 static void emac_ethtool_get_ringparam(struct net_device *ndev,
2030                                        struct ethtool_ringparam *rp)
2031 {
2032         rp->rx_max_pending = rp->rx_pending = NUM_RX_BUFF;
2033         rp->tx_max_pending = rp->tx_pending = NUM_TX_BUFF;
2034 }
2035
2036 static void emac_ethtool_get_pauseparam(struct net_device *ndev,
2037                                         struct ethtool_pauseparam *pp)
2038 {
2039         struct emac_instance *dev = netdev_priv(ndev);
2040
2041         mutex_lock(&dev->link_lock);
2042         if ((dev->phy.features & SUPPORTED_Autoneg) &&
2043             (dev->phy.advertising & (ADVERTISED_Pause | ADVERTISED_Asym_Pause)))
2044                 pp->autoneg = 1;
2045
2046         if (dev->phy.duplex == DUPLEX_FULL) {
2047                 if (dev->phy.pause)
2048                         pp->rx_pause = pp->tx_pause = 1;
2049                 else if (dev->phy.asym_pause)
2050                         pp->tx_pause = 1;
2051         }
2052         mutex_unlock(&dev->link_lock);
2053 }
2054
2055 static u32 emac_ethtool_get_rx_csum(struct net_device *ndev)
2056 {
2057         struct emac_instance *dev = netdev_priv(ndev);
2058
2059         return dev->tah_dev != NULL;
2060 }
2061
2062 static int emac_get_regs_len(struct emac_instance *dev)
2063 {
2064         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
2065                 return sizeof(struct emac_ethtool_regs_subhdr) +
2066                         EMAC4_ETHTOOL_REGS_SIZE(dev);
2067         else
2068                 return sizeof(struct emac_ethtool_regs_subhdr) +
2069                         EMAC_ETHTOOL_REGS_SIZE(dev);
2070 }
2071
2072 static int emac_ethtool_get_regs_len(struct net_device *ndev)
2073 {
2074         struct emac_instance *dev = netdev_priv(ndev);
2075         int size;
2076
2077         size = sizeof(struct emac_ethtool_regs_hdr) +
2078                 emac_get_regs_len(dev) + mal_get_regs_len(dev->mal);
2079         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
2080                 size += zmii_get_regs_len(dev->zmii_dev);
2081         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
2082                 size += rgmii_get_regs_len(dev->rgmii_dev);
2083         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
2084                 size += tah_get_regs_len(dev->tah_dev);
2085
2086         return size;
2087 }
2088
2089 static void *emac_dump_regs(struct emac_instance *dev, void *buf)
2090 {
2091         struct emac_ethtool_regs_subhdr *hdr = buf;
2092
2093         hdr->index = dev->cell_index;
2094         if (emac_has_feature(dev, EMAC_FTR_EMAC4)) {
2095                 hdr->version = EMAC4_ETHTOOL_REGS_VER;
2096                 memcpy_fromio(hdr + 1, dev->emacp, EMAC4_ETHTOOL_REGS_SIZE(dev));
2097                 return ((void *)(hdr + 1) + EMAC4_ETHTOOL_REGS_SIZE(dev));
2098         } else {
2099                 hdr->version = EMAC_ETHTOOL_REGS_VER;
2100                 memcpy_fromio(hdr + 1, dev->emacp, EMAC_ETHTOOL_REGS_SIZE(dev));
2101                 return ((void *)(hdr + 1) + EMAC_ETHTOOL_REGS_SIZE(dev));
2102         }
2103 }
2104
2105 static void emac_ethtool_get_regs(struct net_device *ndev,
2106                                   struct ethtool_regs *regs, void *buf)
2107 {
2108         struct emac_instance *dev = netdev_priv(ndev);
2109         struct emac_ethtool_regs_hdr *hdr = buf;
2110
2111         hdr->components = 0;
2112         buf = hdr + 1;
2113
2114         buf = mal_dump_regs(dev->mal, buf);
2115         buf = emac_dump_regs(dev, buf);
2116         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII)) {
2117                 hdr->components |= EMAC_ETHTOOL_REGS_ZMII;
2118                 buf = zmii_dump_regs(dev->zmii_dev, buf);
2119         }
2120         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII)) {
2121                 hdr->components |= EMAC_ETHTOOL_REGS_RGMII;
2122                 buf = rgmii_dump_regs(dev->rgmii_dev, buf);
2123         }
2124         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH)) {
2125                 hdr->components |= EMAC_ETHTOOL_REGS_TAH;
2126                 buf = tah_dump_regs(dev->tah_dev, buf);
2127         }
2128 }
2129
2130 static int emac_ethtool_nway_reset(struct net_device *ndev)
2131 {
2132         struct emac_instance *dev = netdev_priv(ndev);
2133         int res = 0;
2134
2135         DBG(dev, "nway_reset" NL);
2136
2137         if (dev->phy.address < 0)
2138                 return -EOPNOTSUPP;
2139
2140         mutex_lock(&dev->link_lock);
2141         if (!dev->phy.autoneg) {
2142                 res = -EINVAL;
2143                 goto out;
2144         }
2145
2146         dev->phy.def->ops->setup_aneg(&dev->phy, dev->phy.advertising);
2147  out:
2148         mutex_unlock(&dev->link_lock);
2149         emac_force_link_update(dev);
2150         return res;
2151 }
2152
2153 static int emac_ethtool_get_sset_count(struct net_device *ndev, int stringset)
2154 {
2155         if (stringset == ETH_SS_STATS)
2156                 return EMAC_ETHTOOL_STATS_COUNT;
2157         else
2158                 return -EINVAL;
2159 }
2160
2161 static void emac_ethtool_get_strings(struct net_device *ndev, u32 stringset,
2162                                      u8 * buf)
2163 {
2164         if (stringset == ETH_SS_STATS)
2165                 memcpy(buf, &emac_stats_keys, sizeof(emac_stats_keys));
2166 }
2167
2168 static void emac_ethtool_get_ethtool_stats(struct net_device *ndev,
2169                                            struct ethtool_stats *estats,
2170                                            u64 * tmp_stats)
2171 {
2172         struct emac_instance *dev = netdev_priv(ndev);
2173
2174         memcpy(tmp_stats, &dev->stats, sizeof(dev->stats));
2175         tmp_stats += sizeof(dev->stats) / sizeof(u64);
2176         memcpy(tmp_stats, &dev->estats, sizeof(dev->estats));
2177 }
2178
2179 static void emac_ethtool_get_drvinfo(struct net_device *ndev,
2180                                      struct ethtool_drvinfo *info)
2181 {
2182         struct emac_instance *dev = netdev_priv(ndev);
2183
2184         strcpy(info->driver, "ibm_emac");
2185         strcpy(info->version, DRV_VERSION);
2186         info->fw_version[0] = '\0';
2187         sprintf(info->bus_info, "PPC 4xx EMAC-%d %s",
2188                 dev->cell_index, dev->ofdev->node->full_name);
2189         info->regdump_len = emac_ethtool_get_regs_len(ndev);
2190 }
2191
2192 static const struct ethtool_ops emac_ethtool_ops = {
2193         .get_settings = emac_ethtool_get_settings,
2194         .set_settings = emac_ethtool_set_settings,
2195         .get_drvinfo = emac_ethtool_get_drvinfo,
2196
2197         .get_regs_len = emac_ethtool_get_regs_len,
2198         .get_regs = emac_ethtool_get_regs,
2199
2200         .nway_reset = emac_ethtool_nway_reset,
2201
2202         .get_ringparam = emac_ethtool_get_ringparam,
2203         .get_pauseparam = emac_ethtool_get_pauseparam,
2204
2205         .get_rx_csum = emac_ethtool_get_rx_csum,
2206
2207         .get_strings = emac_ethtool_get_strings,
2208         .get_sset_count = emac_ethtool_get_sset_count,
2209         .get_ethtool_stats = emac_ethtool_get_ethtool_stats,
2210
2211         .get_link = ethtool_op_get_link,
2212         .get_tx_csum = ethtool_op_get_tx_csum,
2213         .get_sg = ethtool_op_get_sg,
2214 };
2215
2216 static int emac_ioctl(struct net_device *ndev, struct ifreq *rq, int cmd)
2217 {
2218         struct emac_instance *dev = netdev_priv(ndev);
2219         struct mii_ioctl_data *data = if_mii(rq);
2220
2221         DBG(dev, "ioctl %08x" NL, cmd);
2222
2223         if (dev->phy.address < 0)
2224                 return -EOPNOTSUPP;
2225
2226         switch (cmd) {
2227         case SIOCGMIIPHY:
2228                 data->phy_id = dev->phy.address;
2229                 /* Fall through */
2230         case SIOCGMIIREG:
2231                 data->val_out = emac_mdio_read(ndev, dev->phy.address,
2232                                                data->reg_num);
2233                 return 0;
2234
2235         case SIOCSMIIREG:
2236                 emac_mdio_write(ndev, dev->phy.address, data->reg_num,
2237                                 data->val_in);
2238                 return 0;
2239         default:
2240                 return -EOPNOTSUPP;
2241         }
2242 }
2243
2244 struct emac_depentry {
2245         u32                     phandle;
2246         struct device_node      *node;
2247         struct of_device        *ofdev;
2248         void                    *drvdata;
2249 };
2250
2251 #define EMAC_DEP_MAL_IDX        0
2252 #define EMAC_DEP_ZMII_IDX       1
2253 #define EMAC_DEP_RGMII_IDX      2
2254 #define EMAC_DEP_TAH_IDX        3
2255 #define EMAC_DEP_MDIO_IDX       4
2256 #define EMAC_DEP_PREV_IDX       5
2257 #define EMAC_DEP_COUNT          6
2258
2259 static int __devinit emac_check_deps(struct emac_instance *dev,
2260                                      struct emac_depentry *deps)
2261 {
2262         int i, there = 0;
2263         struct device_node *np;
2264
2265         for (i = 0; i < EMAC_DEP_COUNT; i++) {
2266                 /* no dependency on that item, allright */
2267                 if (deps[i].phandle == 0) {
2268                         there++;
2269                         continue;
2270                 }
2271                 /* special case for blist as the dependency might go away */
2272                 if (i == EMAC_DEP_PREV_IDX) {
2273                         np = *(dev->blist - 1);
2274                         if (np == NULL) {
2275                                 deps[i].phandle = 0;
2276                                 there++;
2277                                 continue;
2278                         }
2279                         if (deps[i].node == NULL)
2280                                 deps[i].node = of_node_get(np);
2281                 }
2282                 if (deps[i].node == NULL)
2283                         deps[i].node = of_find_node_by_phandle(deps[i].phandle);
2284                 if (deps[i].node == NULL)
2285                         continue;
2286                 if (deps[i].ofdev == NULL)
2287                         deps[i].ofdev = of_find_device_by_node(deps[i].node);
2288                 if (deps[i].ofdev == NULL)
2289                         continue;
2290                 if (deps[i].drvdata == NULL)
2291                         deps[i].drvdata = dev_get_drvdata(&deps[i].ofdev->dev);
2292                 if (deps[i].drvdata != NULL)
2293                         there++;
2294         }
2295         return (there == EMAC_DEP_COUNT);
2296 }
2297
2298 static void emac_put_deps(struct emac_instance *dev)
2299 {
2300         if (dev->mal_dev)
2301                 of_dev_put(dev->mal_dev);
2302         if (dev->zmii_dev)
2303                 of_dev_put(dev->zmii_dev);
2304         if (dev->rgmii_dev)
2305                 of_dev_put(dev->rgmii_dev);
2306         if (dev->mdio_dev)
2307                 of_dev_put(dev->mdio_dev);
2308         if (dev->tah_dev)
2309                 of_dev_put(dev->tah_dev);
2310 }
2311
2312 static int __devinit emac_of_bus_notify(struct notifier_block *nb,
2313                                         unsigned long action, void *data)
2314 {
2315         /* We are only intereted in device addition */
2316         if (action == BUS_NOTIFY_BOUND_DRIVER)
2317                 wake_up_all(&emac_probe_wait);
2318         return 0;
2319 }
2320
2321 static struct notifier_block emac_of_bus_notifier __devinitdata = {
2322         .notifier_call = emac_of_bus_notify
2323 };
2324
2325 static int __devinit emac_wait_deps(struct emac_instance *dev)
2326 {
2327         struct emac_depentry deps[EMAC_DEP_COUNT];
2328         int i, err;
2329
2330         memset(&deps, 0, sizeof(deps));
2331
2332         deps[EMAC_DEP_MAL_IDX].phandle = dev->mal_ph;
2333         deps[EMAC_DEP_ZMII_IDX].phandle = dev->zmii_ph;
2334         deps[EMAC_DEP_RGMII_IDX].phandle = dev->rgmii_ph;
2335         if (dev->tah_ph)
2336                 deps[EMAC_DEP_TAH_IDX].phandle = dev->tah_ph;
2337         if (dev->mdio_ph)
2338                 deps[EMAC_DEP_MDIO_IDX].phandle = dev->mdio_ph;
2339         if (dev->blist && dev->blist > emac_boot_list)
2340                 deps[EMAC_DEP_PREV_IDX].phandle = 0xffffffffu;
2341         bus_register_notifier(&of_platform_bus_type, &emac_of_bus_notifier);
2342         wait_event_timeout(emac_probe_wait,
2343                            emac_check_deps(dev, deps),
2344                            EMAC_PROBE_DEP_TIMEOUT);
2345         bus_unregister_notifier(&of_platform_bus_type, &emac_of_bus_notifier);
2346         err = emac_check_deps(dev, deps) ? 0 : -ENODEV;
2347         for (i = 0; i < EMAC_DEP_COUNT; i++) {
2348                 if (deps[i].node)
2349                         of_node_put(deps[i].node);
2350                 if (err && deps[i].ofdev)
2351                         of_dev_put(deps[i].ofdev);
2352         }
2353         if (err == 0) {
2354                 dev->mal_dev = deps[EMAC_DEP_MAL_IDX].ofdev;
2355                 dev->zmii_dev = deps[EMAC_DEP_ZMII_IDX].ofdev;
2356                 dev->rgmii_dev = deps[EMAC_DEP_RGMII_IDX].ofdev;
2357                 dev->tah_dev = deps[EMAC_DEP_TAH_IDX].ofdev;
2358                 dev->mdio_dev = deps[EMAC_DEP_MDIO_IDX].ofdev;
2359         }
2360         if (deps[EMAC_DEP_PREV_IDX].ofdev)
2361                 of_dev_put(deps[EMAC_DEP_PREV_IDX].ofdev);
2362         return err;
2363 }
2364
2365 static int __devinit emac_read_uint_prop(struct device_node *np, const char *name,
2366                                          u32 *val, int fatal)
2367 {
2368         int len;
2369         const u32 *prop = of_get_property(np, name, &len);
2370         if (prop == NULL || len < sizeof(u32)) {
2371                 if (fatal)
2372                         printk(KERN_ERR "%s: missing %s property\n",
2373                                np->full_name, name);
2374                 return -ENODEV;
2375         }
2376         *val = *prop;
2377         return 0;
2378 }
2379
2380 static int __devinit emac_init_phy(struct emac_instance *dev)
2381 {
2382         struct device_node *np = dev->ofdev->node;
2383         struct net_device *ndev = dev->ndev;
2384         u32 phy_map, adv;
2385         int i;
2386
2387         dev->phy.dev = ndev;
2388         dev->phy.mode = dev->phy_mode;
2389
2390         /* PHY-less configuration.
2391          * XXX I probably should move these settings to the dev tree
2392          */
2393         if (dev->phy_address == 0xffffffff && dev->phy_map == 0xffffffff) {
2394                 emac_reset(dev);
2395
2396                 /* PHY-less configuration.
2397                  * XXX I probably should move these settings to the dev tree
2398                  */
2399                 dev->phy.address = -1;
2400                 dev->phy.features = SUPPORTED_MII;
2401                 if (emac_phy_supports_gige(dev->phy_mode))
2402                         dev->phy.features |= SUPPORTED_1000baseT_Full;
2403                 else
2404                         dev->phy.features |= SUPPORTED_100baseT_Full;
2405                 dev->phy.pause = 1;
2406
2407                 return 0;
2408         }
2409
2410         mutex_lock(&emac_phy_map_lock);
2411         phy_map = dev->phy_map | busy_phy_map;
2412
2413         DBG(dev, "PHY maps %08x %08x" NL, dev->phy_map, busy_phy_map);
2414
2415         dev->phy.mdio_read = emac_mdio_read;
2416         dev->phy.mdio_write = emac_mdio_write;
2417
2418         /* Enable internal clock source */
2419 #ifdef CONFIG_PPC_DCR_NATIVE
2420         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX))
2421                 dcri_clrset(SDR0, SDR0_MFR, 0, SDR0_MFR_ECS);
2422 #endif
2423         /* PHY clock workaround */
2424         emac_rx_clk_tx(dev);
2425
2426         /* Enable internal clock source on 440GX*/
2427 #ifdef CONFIG_PPC_DCR_NATIVE
2428         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX))
2429                 dcri_clrset(SDR0, SDR0_MFR, 0, SDR0_MFR_ECS);
2430 #endif
2431         /* Configure EMAC with defaults so we can at least use MDIO
2432          * This is needed mostly for 440GX
2433          */
2434         if (emac_phy_gpcs(dev->phy.mode)) {
2435                 /* XXX
2436                  * Make GPCS PHY address equal to EMAC index.
2437                  * We probably should take into account busy_phy_map
2438                  * and/or phy_map here.
2439                  *
2440                  * Note that the busy_phy_map is currently global
2441                  * while it should probably be per-ASIC...
2442                  */
2443                 dev->phy.gpcs_address = dev->gpcs_address;
2444                 if (dev->phy.gpcs_address == 0xffffffff)
2445                         dev->phy.address = dev->cell_index;
2446         }
2447
2448         emac_configure(dev);
2449
2450         if (dev->phy_address != 0xffffffff)
2451                 phy_map = ~(1 << dev->phy_address);
2452
2453         for (i = 0; i < 0x20; phy_map >>= 1, ++i)
2454                 if (!(phy_map & 1)) {
2455                         int r;
2456                         busy_phy_map |= 1 << i;
2457
2458                         /* Quick check if there is a PHY at the address */
2459                         r = emac_mdio_read(dev->ndev, i, MII_BMCR);
2460                         if (r == 0xffff || r < 0)
2461                                 continue;
2462                         if (!emac_mii_phy_probe(&dev->phy, i))
2463                                 break;
2464                 }
2465
2466         /* Enable external clock source */
2467 #ifdef CONFIG_PPC_DCR_NATIVE
2468         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX))
2469                 dcri_clrset(SDR0, SDR0_MFR, SDR0_MFR_ECS, 0);
2470 #endif
2471         mutex_unlock(&emac_phy_map_lock);
2472         if (i == 0x20) {
2473                 printk(KERN_WARNING "%s: can't find PHY!\n", np->full_name);
2474                 return -ENXIO;
2475         }
2476
2477         /* Init PHY */
2478         if (dev->phy.def->ops->init)
2479                 dev->phy.def->ops->init(&dev->phy);
2480
2481         /* Disable any PHY features not supported by the platform */
2482         dev->phy.def->features &= ~dev->phy_feat_exc;
2483
2484         /* Setup initial link parameters */
2485         if (dev->phy.features & SUPPORTED_Autoneg) {
2486                 adv = dev->phy.features;
2487                 if (!emac_has_feature(dev, EMAC_FTR_NO_FLOW_CONTROL_40x))
2488                         adv |= ADVERTISED_Pause | ADVERTISED_Asym_Pause;
2489                 /* Restart autonegotiation */
2490                 dev->phy.def->ops->setup_aneg(&dev->phy, adv);
2491         } else {
2492                 u32 f = dev->phy.def->features;
2493                 int speed = SPEED_10, fd = DUPLEX_HALF;
2494
2495                 /* Select highest supported speed/duplex */
2496                 if (f & SUPPORTED_1000baseT_Full) {
2497                         speed = SPEED_1000;
2498                         fd = DUPLEX_FULL;
2499                 } else if (f & SUPPORTED_1000baseT_Half)
2500                         speed = SPEED_1000;
2501                 else if (f & SUPPORTED_100baseT_Full) {
2502                         speed = SPEED_100;
2503                         fd = DUPLEX_FULL;
2504                 } else if (f & SUPPORTED_100baseT_Half)
2505                         speed = SPEED_100;
2506                 else if (f & SUPPORTED_10baseT_Full)
2507                         fd = DUPLEX_FULL;
2508
2509                 /* Force link parameters */
2510                 dev->phy.def->ops->setup_forced(&dev->phy, speed, fd);
2511         }
2512         return 0;
2513 }
2514
2515 static int __devinit emac_init_config(struct emac_instance *dev)
2516 {
2517         struct device_node *np = dev->ofdev->node;
2518         const void *p;
2519         unsigned int plen;
2520         const char *pm, *phy_modes[] = {
2521                 [PHY_MODE_NA] = "",
2522                 [PHY_MODE_MII] = "mii",
2523                 [PHY_MODE_RMII] = "rmii",
2524                 [PHY_MODE_SMII] = "smii",
2525                 [PHY_MODE_RGMII] = "rgmii",
2526                 [PHY_MODE_TBI] = "tbi",
2527                 [PHY_MODE_GMII] = "gmii",
2528                 [PHY_MODE_RTBI] = "rtbi",
2529                 [PHY_MODE_SGMII] = "sgmii",
2530         };
2531
2532         /* Read config from device-tree */
2533         if (emac_read_uint_prop(np, "mal-device", &dev->mal_ph, 1))
2534                 return -ENXIO;
2535         if (emac_read_uint_prop(np, "mal-tx-channel", &dev->mal_tx_chan, 1))
2536                 return -ENXIO;
2537         if (emac_read_uint_prop(np, "mal-rx-channel", &dev->mal_rx_chan, 1))
2538                 return -ENXIO;
2539         if (emac_read_uint_prop(np, "cell-index", &dev->cell_index, 1))
2540                 return -ENXIO;
2541         if (emac_read_uint_prop(np, "max-frame-size", &dev->max_mtu, 0))
2542                 dev->max_mtu = 1500;
2543         if (emac_read_uint_prop(np, "rx-fifo-size", &dev->rx_fifo_size, 0))
2544                 dev->rx_fifo_size = 2048;
2545         if (emac_read_uint_prop(np, "tx-fifo-size", &dev->tx_fifo_size, 0))
2546                 dev->tx_fifo_size = 2048;
2547         if (emac_read_uint_prop(np, "rx-fifo-size-gige", &dev->rx_fifo_size_gige, 0))
2548                 dev->rx_fifo_size_gige = dev->rx_fifo_size;
2549         if (emac_read_uint_prop(np, "tx-fifo-size-gige", &dev->tx_fifo_size_gige, 0))
2550                 dev->tx_fifo_size_gige = dev->tx_fifo_size;
2551         if (emac_read_uint_prop(np, "phy-address", &dev->phy_address, 0))
2552                 dev->phy_address = 0xffffffff;
2553         if (emac_read_uint_prop(np, "phy-map", &dev->phy_map, 0))
2554                 dev->phy_map = 0xffffffff;
2555         if (emac_read_uint_prop(np, "gpcs-address", &dev->gpcs_address, 0))
2556                 dev->gpcs_address = 0xffffffff;
2557         if (emac_read_uint_prop(np->parent, "clock-frequency", &dev->opb_bus_freq, 1))
2558                 return -ENXIO;
2559         if (emac_read_uint_prop(np, "tah-device", &dev->tah_ph, 0))
2560                 dev->tah_ph = 0;
2561         if (emac_read_uint_prop(np, "tah-channel", &dev->tah_port, 0))
2562                 dev->tah_port = 0;
2563         if (emac_read_uint_prop(np, "mdio-device", &dev->mdio_ph, 0))
2564                 dev->mdio_ph = 0;
2565         if (emac_read_uint_prop(np, "zmii-device", &dev->zmii_ph, 0))
2566                 dev->zmii_ph = 0;
2567         if (emac_read_uint_prop(np, "zmii-channel", &dev->zmii_port, 0))
2568                 dev->zmii_port = 0xffffffff;
2569         if (emac_read_uint_prop(np, "rgmii-device", &dev->rgmii_ph, 0))
2570                 dev->rgmii_ph = 0;
2571         if (emac_read_uint_prop(np, "rgmii-channel", &dev->rgmii_port, 0))
2572                 dev->rgmii_port = 0xffffffff;
2573         if (emac_read_uint_prop(np, "fifo-entry-size", &dev->fifo_entry_size, 0))
2574                 dev->fifo_entry_size = 16;
2575         if (emac_read_uint_prop(np, "mal-burst-size", &dev->mal_burst_size, 0))
2576                 dev->mal_burst_size = 256;
2577
2578         /* PHY mode needs some decoding */
2579         dev->phy_mode = PHY_MODE_NA;
2580         pm = of_get_property(np, "phy-mode", &plen);
2581         if (pm != NULL) {
2582                 int i;
2583                 for (i = 0; i < ARRAY_SIZE(phy_modes); i++)
2584                         if (!strcasecmp(pm, phy_modes[i])) {
2585                                 dev->phy_mode = i;
2586                                 break;
2587                         }
2588         }
2589
2590         /* Backward compat with non-final DT */
2591         if (dev->phy_mode == PHY_MODE_NA && pm != NULL && plen == 4) {
2592                 u32 nmode = *(const u32 *)pm;
2593                 if (nmode > PHY_MODE_NA && nmode <= PHY_MODE_SGMII)
2594                         dev->phy_mode = nmode;
2595         }
2596
2597         /* Check EMAC version */
2598         if (of_device_is_compatible(np, "ibm,emac4sync")) {
2599                 dev->features |= (EMAC_FTR_EMAC4 | EMAC_FTR_EMAC4SYNC);
2600                 if (of_device_is_compatible(np, "ibm,emac-460ex") ||
2601                     of_device_is_compatible(np, "ibm,emac-460gt"))
2602                         dev->features |= EMAC_FTR_460EX_PHY_CLK_FIX;
2603                 if (of_device_is_compatible(np, "ibm,emac-405ex") ||
2604                     of_device_is_compatible(np, "ibm,emac-405exr"))
2605                         dev->features |= EMAC_FTR_440EP_PHY_CLK_FIX;
2606         } else if (of_device_is_compatible(np, "ibm,emac4")) {
2607                 dev->features |= EMAC_FTR_EMAC4;
2608                 if (of_device_is_compatible(np, "ibm,emac-440gx"))
2609                         dev->features |= EMAC_FTR_440GX_PHY_CLK_FIX;
2610         } else {
2611                 if (of_device_is_compatible(np, "ibm,emac-440ep") ||
2612                     of_device_is_compatible(np, "ibm,emac-440gr"))
2613                         dev->features |= EMAC_FTR_440EP_PHY_CLK_FIX;
2614                 if (of_device_is_compatible(np, "ibm,emac-405ez")) {
2615 #ifdef CONFIG_IBM_NEW_EMAC_NO_FLOW_CTRL
2616                         dev->features |= EMAC_FTR_NO_FLOW_CONTROL_40x;
2617 #else
2618                         printk(KERN_ERR "%s: Flow control not disabled!\n",
2619                                         np->full_name);
2620                         return -ENXIO;
2621 #endif
2622                 }
2623
2624         }
2625
2626         /* Fixup some feature bits based on the device tree */
2627         if (of_get_property(np, "has-inverted-stacr-oc", NULL))
2628                 dev->features |= EMAC_FTR_STACR_OC_INVERT;
2629         if (of_get_property(np, "has-new-stacr-staopc", NULL))
2630                 dev->features |= EMAC_FTR_HAS_NEW_STACR;
2631
2632         /* CAB lacks the appropriate properties */
2633         if (of_device_is_compatible(np, "ibm,emac-axon"))
2634                 dev->features |= EMAC_FTR_HAS_NEW_STACR |
2635                         EMAC_FTR_STACR_OC_INVERT;
2636
2637         /* Enable TAH/ZMII/RGMII features as found */
2638         if (dev->tah_ph != 0) {
2639 #ifdef CONFIG_IBM_NEW_EMAC_TAH
2640                 dev->features |= EMAC_FTR_HAS_TAH;
2641 #else
2642                 printk(KERN_ERR "%s: TAH support not enabled !\n",
2643                        np->full_name);
2644                 return -ENXIO;
2645 #endif
2646         }
2647
2648         if (dev->zmii_ph != 0) {
2649 #ifdef CONFIG_IBM_NEW_EMAC_ZMII
2650                 dev->features |= EMAC_FTR_HAS_ZMII;
2651 #else
2652                 printk(KERN_ERR "%s: ZMII support not enabled !\n",
2653                        np->full_name);
2654                 return -ENXIO;
2655 #endif
2656         }
2657
2658         if (dev->rgmii_ph != 0) {
2659 #ifdef CONFIG_IBM_NEW_EMAC_RGMII
2660                 dev->features |= EMAC_FTR_HAS_RGMII;
2661 #else
2662                 printk(KERN_ERR "%s: RGMII support not enabled !\n",
2663                        np->full_name);
2664                 return -ENXIO;
2665 #endif
2666         }
2667
2668         /* Read MAC-address */
2669         p = of_get_property(np, "local-mac-address", NULL);
2670         if (p == NULL) {
2671                 printk(KERN_ERR "%s: Can't find local-mac-address property\n",
2672                        np->full_name);
2673                 return -ENXIO;
2674         }
2675         memcpy(dev->ndev->dev_addr, p, 6);
2676
2677         /* IAHT and GAHT filter parameterization */
2678         if (emac_has_feature(dev, EMAC_FTR_EMAC4SYNC)) {
2679                 dev->xaht_slots_shift = EMAC4SYNC_XAHT_SLOTS_SHIFT;
2680                 dev->xaht_width_shift = EMAC4SYNC_XAHT_WIDTH_SHIFT;
2681         } else {
2682                 dev->xaht_slots_shift = EMAC4_XAHT_SLOTS_SHIFT;
2683                 dev->xaht_width_shift = EMAC4_XAHT_WIDTH_SHIFT;
2684         }
2685
2686         DBG(dev, "features     : 0x%08x / 0x%08x\n", dev->features, EMAC_FTRS_POSSIBLE);
2687         DBG(dev, "tx_fifo_size : %d (%d gige)\n", dev->tx_fifo_size, dev->tx_fifo_size_gige);
2688         DBG(dev, "rx_fifo_size : %d (%d gige)\n", dev->rx_fifo_size, dev->rx_fifo_size_gige);
2689         DBG(dev, "max_mtu      : %d\n", dev->max_mtu);
2690         DBG(dev, "OPB freq     : %d\n", dev->opb_bus_freq);
2691
2692         return 0;
2693 }
2694
2695 static const struct net_device_ops emac_netdev_ops = {
2696         .ndo_open               = emac_open,
2697         .ndo_stop               = emac_close,
2698         .ndo_get_stats          = emac_stats,
2699         .ndo_set_multicast_list = emac_set_multicast_list,
2700         .ndo_do_ioctl           = emac_ioctl,
2701         .ndo_tx_timeout         = emac_tx_timeout,
2702         .ndo_validate_addr      = eth_validate_addr,
2703         .ndo_set_mac_address    = eth_mac_addr,
2704         .ndo_start_xmit         = emac_start_xmit,
2705         .ndo_change_mtu         = eth_change_mtu,
2706 };
2707
2708 static const struct net_device_ops emac_gige_netdev_ops = {
2709         .ndo_open               = emac_open,
2710         .ndo_stop               = emac_close,
2711         .ndo_get_stats          = emac_stats,
2712         .ndo_set_multicast_list = emac_set_multicast_list,
2713         .ndo_do_ioctl           = emac_ioctl,
2714         .ndo_tx_timeout         = emac_tx_timeout,
2715         .ndo_validate_addr      = eth_validate_addr,
2716         .ndo_set_mac_address    = eth_mac_addr,
2717         .ndo_start_xmit         = emac_start_xmit_sg,
2718         .ndo_change_mtu         = emac_change_mtu,
2719 };
2720
2721 static int __devinit emac_probe(struct of_device *ofdev,
2722                                 const struct of_device_id *match)
2723 {
2724         struct net_device *ndev;
2725         struct emac_instance *dev;
2726         struct device_node *np = ofdev->node;
2727         struct device_node **blist = NULL;
2728         int err, i;
2729
2730         /* Skip unused/unwired EMACS.  We leave the check for an unused
2731          * property here for now, but new flat device trees should set a
2732          * status property to "disabled" instead.
2733          */
2734         if (of_get_property(np, "unused", NULL) || !of_device_is_available(np))
2735                 return -ENODEV;
2736
2737         /* Find ourselves in the bootlist if we are there */
2738         for (i = 0; i < EMAC_BOOT_LIST_SIZE; i++)
2739                 if (emac_boot_list[i] == np)
2740                         blist = &emac_boot_list[i];
2741
2742         /* Allocate our net_device structure */
2743         err = -ENOMEM;
2744         ndev = alloc_etherdev(sizeof(struct emac_instance));
2745         if (!ndev) {
2746                 printk(KERN_ERR "%s: could not allocate ethernet device!\n",
2747                        np->full_name);
2748                 goto err_gone;
2749         }
2750         dev = netdev_priv(ndev);
2751         dev->ndev = ndev;
2752         dev->ofdev = ofdev;
2753         dev->blist = blist;
2754         SET_NETDEV_DEV(ndev, &ofdev->dev);
2755
2756         /* Initialize some embedded data structures */
2757         mutex_init(&dev->mdio_lock);
2758         mutex_init(&dev->link_lock);
2759         spin_lock_init(&dev->lock);
2760         INIT_WORK(&dev->reset_work, emac_reset_work);
2761
2762         /* Init various config data based on device-tree */
2763         err = emac_init_config(dev);
2764         if (err != 0)
2765                 goto err_free;
2766
2767         /* Get interrupts. EMAC irq is mandatory, WOL irq is optional */
2768         dev->emac_irq = irq_of_parse_and_map(np, 0);
2769         dev->wol_irq = irq_of_parse_and_map(np, 1);
2770         if (dev->emac_irq == NO_IRQ) {
2771                 printk(KERN_ERR "%s: Can't map main interrupt\n", np->full_name);
2772                 goto err_free;
2773         }
2774         ndev->irq = dev->emac_irq;
2775
2776         /* Map EMAC regs */
2777         if (of_address_to_resource(np, 0, &dev->rsrc_regs)) {
2778                 printk(KERN_ERR "%s: Can't get registers address\n",
2779                        np->full_name);
2780                 goto err_irq_unmap;
2781         }
2782         // TODO : request_mem_region
2783         dev->emacp = ioremap(dev->rsrc_regs.start,
2784                              dev->rsrc_regs.end - dev->rsrc_regs.start + 1);
2785         if (dev->emacp == NULL) {
2786                 printk(KERN_ERR "%s: Can't map device registers!\n",
2787                        np->full_name);
2788                 err = -ENOMEM;
2789                 goto err_irq_unmap;
2790         }
2791
2792         /* Wait for dependent devices */
2793         err = emac_wait_deps(dev);
2794         if (err) {
2795                 printk(KERN_ERR
2796                        "%s: Timeout waiting for dependent devices\n",
2797                        np->full_name);
2798                 /*  display more info about what's missing ? */
2799                 goto err_reg_unmap;
2800         }
2801         dev->mal = dev_get_drvdata(&dev->mal_dev->dev);
2802         if (dev->mdio_dev != NULL)
2803                 dev->mdio_instance = dev_get_drvdata(&dev->mdio_dev->dev);
2804
2805         /* Register with MAL */
2806         dev->commac.ops = &emac_commac_ops;
2807         dev->commac.dev = dev;
2808         dev->commac.tx_chan_mask = MAL_CHAN_MASK(dev->mal_tx_chan);
2809         dev->commac.rx_chan_mask = MAL_CHAN_MASK(dev->mal_rx_chan);
2810         err = mal_register_commac(dev->mal, &dev->commac);
2811         if (err) {
2812                 printk(KERN_ERR "%s: failed to register with mal %s!\n",
2813                        np->full_name, dev->mal_dev->node->full_name);
2814                 goto err_rel_deps;
2815         }
2816         dev->rx_skb_size = emac_rx_skb_size(ndev->mtu);
2817         dev->rx_sync_size = emac_rx_sync_size(ndev->mtu);
2818
2819         /* Get pointers to BD rings */
2820         dev->tx_desc =
2821             dev->mal->bd_virt + mal_tx_bd_offset(dev->mal, dev->mal_tx_chan);
2822         dev->rx_desc =
2823             dev->mal->bd_virt + mal_rx_bd_offset(dev->mal, dev->mal_rx_chan);
2824
2825         DBG(dev, "tx_desc %p" NL, dev->tx_desc);
2826         DBG(dev, "rx_desc %p" NL, dev->rx_desc);
2827
2828         /* Clean rings */
2829         memset(dev->tx_desc, 0, NUM_TX_BUFF * sizeof(struct mal_descriptor));
2830         memset(dev->rx_desc, 0, NUM_RX_BUFF * sizeof(struct mal_descriptor));
2831         memset(dev->tx_skb, 0, NUM_TX_BUFF * sizeof(struct sk_buff *));
2832         memset(dev->rx_skb, 0, NUM_RX_BUFF * sizeof(struct sk_buff *));
2833
2834         /* Attach to ZMII, if needed */
2835         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII) &&
2836             (err = zmii_attach(dev->zmii_dev, dev->zmii_port, &dev->phy_mode)) != 0)
2837                 goto err_unreg_commac;
2838
2839         /* Attach to RGMII, if needed */
2840         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII) &&
2841             (err = rgmii_attach(dev->rgmii_dev, dev->rgmii_port, dev->phy_mode)) != 0)
2842                 goto err_detach_zmii;
2843
2844         /* Attach to TAH, if needed */
2845         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH) &&
2846             (err = tah_attach(dev->tah_dev, dev->tah_port)) != 0)
2847                 goto err_detach_rgmii;
2848
2849         /* Set some link defaults before we can find out real parameters */
2850         dev->phy.speed = SPEED_100;
2851         dev->phy.duplex = DUPLEX_FULL;
2852         dev->phy.autoneg = AUTONEG_DISABLE;
2853         dev->phy.pause = dev->phy.asym_pause = 0;
2854         dev->stop_timeout = STOP_TIMEOUT_100;
2855         INIT_DELAYED_WORK(&dev->link_work, emac_link_timer);
2856
2857         /* Find PHY if any */
2858         err = emac_init_phy(dev);
2859         if (err != 0)
2860                 goto err_detach_tah;
2861
2862         if (dev->tah_dev)
2863                 ndev->features |= NETIF_F_IP_CSUM | NETIF_F_SG;
2864         ndev->watchdog_timeo = 5 * HZ;
2865         if (emac_phy_supports_gige(dev->phy_mode)) {
2866                 ndev->netdev_ops = &emac_gige_netdev_ops;
2867                 dev->commac.ops = &emac_commac_sg_ops;
2868         } else
2869                 ndev->netdev_ops = &emac_netdev_ops;
2870         SET_ETHTOOL_OPS(ndev, &emac_ethtool_ops);
2871
2872         netif_carrier_off(ndev);
2873         netif_stop_queue(ndev);
2874
2875         err = register_netdev(ndev);
2876         if (err) {
2877                 printk(KERN_ERR "%s: failed to register net device (%d)!\n",
2878                        np->full_name, err);
2879                 goto err_detach_tah;
2880         }
2881
2882         /* Set our drvdata last as we don't want them visible until we are
2883          * fully initialized
2884          */
2885         wmb();
2886         dev_set_drvdata(&ofdev->dev, dev);
2887
2888         /* There's a new kid in town ! Let's tell everybody */
2889         wake_up_all(&emac_probe_wait);
2890
2891
2892         printk(KERN_INFO "%s: EMAC-%d %s, MAC %pM\n",
2893                ndev->name, dev->cell_index, np->full_name, ndev->dev_addr);
2894
2895         if (dev->phy_mode == PHY_MODE_SGMII)
2896                 printk(KERN_NOTICE "%s: in SGMII mode\n", ndev->name);
2897
2898         if (dev->phy.address >= 0)
2899                 printk("%s: found %s PHY (0x%02x)\n", ndev->name,
2900                        dev->phy.def->name, dev->phy.address);
2901
2902         emac_dbg_register(dev);
2903
2904         /* Life is good */
2905         return 0;
2906
2907         /* I have a bad feeling about this ... */
2908
2909  err_detach_tah:
2910         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
2911                 tah_detach(dev->tah_dev, dev->tah_port);
2912  err_detach_rgmii:
2913         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
2914                 rgmii_detach(dev->rgmii_dev, dev->rgmii_port);
2915  err_detach_zmii:
2916         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
2917                 zmii_detach(dev->zmii_dev, dev->zmii_port);
2918  err_unreg_commac:
2919         mal_unregister_commac(dev->mal, &dev->commac);
2920  err_rel_deps:
2921         emac_put_deps(dev);
2922  err_reg_unmap:
2923         iounmap(dev->emacp);
2924  err_irq_unmap:
2925         if (dev->wol_irq != NO_IRQ)
2926                 irq_dispose_mapping(dev->wol_irq);
2927         if (dev->emac_irq != NO_IRQ)
2928                 irq_dispose_mapping(dev->emac_irq);
2929  err_free:
2930         kfree(ndev);
2931  err_gone:
2932         /* if we were on the bootlist, remove us as we won't show up and
2933          * wake up all waiters to notify them in case they were waiting
2934          * on us
2935          */
2936         if (blist) {
2937                 *blist = NULL;
2938                 wake_up_all(&emac_probe_wait);
2939         }
2940         return err;
2941 }
2942
2943 static int __devexit emac_remove(struct of_device *ofdev)
2944 {
2945         struct emac_instance *dev = dev_get_drvdata(&ofdev->dev);
2946
2947         DBG(dev, "remove" NL);
2948
2949         dev_set_drvdata(&ofdev->dev, NULL);
2950
2951         unregister_netdev(dev->ndev);
2952
2953         flush_scheduled_work();
2954
2955         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
2956                 tah_detach(dev->tah_dev, dev->tah_port);
2957         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
2958                 rgmii_detach(dev->rgmii_dev, dev->rgmii_port);
2959         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
2960                 zmii_detach(dev->zmii_dev, dev->zmii_port);
2961
2962         mal_unregister_commac(dev->mal, &dev->commac);
2963         emac_put_deps(dev);
2964
2965         emac_dbg_unregister(dev);
2966         iounmap(dev->emacp);
2967
2968         if (dev->wol_irq != NO_IRQ)
2969                 irq_dispose_mapping(dev->wol_irq);
2970         if (dev->emac_irq != NO_IRQ)
2971                 irq_dispose_mapping(dev->emac_irq);
2972
2973         kfree(dev->ndev);
2974
2975         return 0;
2976 }
2977
2978 /* XXX Features in here should be replaced by properties... */
2979 static struct of_device_id emac_match[] =
2980 {
2981         {
2982                 .type           = "network",
2983                 .compatible     = "ibm,emac",
2984         },
2985         {
2986                 .type           = "network",
2987                 .compatible     = "ibm,emac4",
2988         },
2989         {
2990                 .type           = "network",
2991                 .compatible     = "ibm,emac4sync",
2992         },
2993         {},
2994 };
2995 MODULE_DEVICE_TABLE(of, emac_match);
2996
2997 static struct of_platform_driver emac_driver = {
2998         .name = "emac",
2999         .match_table = emac_match,
3000
3001         .probe = emac_probe,
3002         .remove = emac_remove,
3003 };
3004
3005 static void __init emac_make_bootlist(void)
3006 {
3007         struct device_node *np = NULL;
3008         int j, max, i = 0, k;
3009         int cell_indices[EMAC_BOOT_LIST_SIZE];
3010
3011         /* Collect EMACs */
3012         while((np = of_find_all_nodes(np)) != NULL) {
3013                 const u32 *idx;
3014
3015                 if (of_match_node(emac_match, np) == NULL)
3016                         continue;
3017                 if (of_get_property(np, "unused", NULL))
3018                         continue;
3019                 idx = of_get_property(np, "cell-index", NULL);
3020                 if (idx == NULL)
3021                         continue;
3022                 cell_indices[i] = *idx;
3023                 emac_boot_list[i++] = of_node_get(np);
3024                 if (i >= EMAC_BOOT_LIST_SIZE) {
3025                         of_node_put(np);
3026                         break;
3027                 }
3028         }
3029         max = i;
3030
3031         /* Bubble sort them (doh, what a creative algorithm :-) */
3032         for (i = 0; max > 1 && (i < (max - 1)); i++)
3033                 for (j = i; j < max; j++) {
3034                         if (cell_indices[i] > cell_indices[j]) {
3035                                 np = emac_boot_list[i];
3036                                 emac_boot_list[i] = emac_boot_list[j];
3037                                 emac_boot_list[j] = np;
3038                                 k = cell_indices[i];
3039                                 cell_indices[i] = cell_indices[j];
3040                                 cell_indices[j] = k;
3041                         }
3042                 }
3043 }
3044
3045 static int __init emac_init(void)
3046 {
3047         int rc;
3048
3049         printk(KERN_INFO DRV_DESC ", version " DRV_VERSION "\n");
3050
3051         /* Init debug stuff */
3052         emac_init_debug();
3053
3054         /* Build EMAC boot list */
3055         emac_make_bootlist();
3056
3057         /* Init submodules */
3058         rc = mal_init();
3059         if (rc)
3060                 goto err;
3061         rc = zmii_init();
3062         if (rc)
3063                 goto err_mal;
3064         rc = rgmii_init();
3065         if (rc)
3066                 goto err_zmii;
3067         rc = tah_init();
3068         if (rc)
3069                 goto err_rgmii;
3070         rc = of_register_platform_driver(&emac_driver);
3071         if (rc)
3072                 goto err_tah;
3073
3074         return 0;
3075
3076  err_tah:
3077         tah_exit();
3078  err_rgmii:
3079         rgmii_exit();
3080  err_zmii:
3081         zmii_exit();
3082  err_mal:
3083         mal_exit();
3084  err:
3085         return rc;
3086 }
3087
3088 static void __exit emac_exit(void)
3089 {
3090         int i;
3091
3092         of_unregister_platform_driver(&emac_driver);
3093
3094         tah_exit();
3095         rgmii_exit();
3096         zmii_exit();
3097         mal_exit();
3098         emac_fini_debug();
3099
3100         /* Destroy EMAC boot list */
3101         for (i = 0; i < EMAC_BOOT_LIST_SIZE; i++)
3102                 if (emac_boot_list[i])
3103                         of_node_put(emac_boot_list[i]);
3104 }
3105
3106 module_init(emac_init);
3107 module_exit(emac_exit);