1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * drivers/net/ethernet/ibm/emac/mal.c
5 * Memory Access Layer (MAL) support
7 * Copyright 2007 Benjamin Herrenschmidt, IBM Corp.
8 * <benh@kernel.crashing.org>
10 * Based on the arch/ppc version of the driver:
12 * Copyright (c) 2004, 2005 Zultys Technologies.
13 * Eugene Surovegin <eugene.surovegin@zultys.com> or <ebs@ebshome.net>
15 * Based on original work by
16 * Benjamin Herrenschmidt <benh@kernel.crashing.org>,
17 * David Gibson <hermes@gibson.dropbear.id.au>,
19 * Armin Kuster <akuster@mvista.com>
20 * Copyright 2002 MontaVista Softare Inc.
23 #include <linux/delay.h>
24 #include <linux/slab.h>
26 #include <linux/of_irq.h>
27 #include <linux/platform_device.h>
30 #include <asm/dcr-regs.h>
34 int mal_register_commac(struct mal_instance
*mal
, struct mal_commac
*commac
)
38 spin_lock_irqsave(&mal
->lock
, flags
);
40 MAL_DBG(mal
, "reg(%08x, %08x)" NL
,
41 commac
->tx_chan_mask
, commac
->rx_chan_mask
);
43 /* Don't let multiple commacs claim the same channel(s) */
44 if ((mal
->tx_chan_mask
& commac
->tx_chan_mask
) ||
45 (mal
->rx_chan_mask
& commac
->rx_chan_mask
)) {
46 spin_unlock_irqrestore(&mal
->lock
, flags
);
47 printk(KERN_WARNING
"mal%d: COMMAC channels conflict!\n",
52 if (list_empty(&mal
->list
))
53 napi_enable(&mal
->napi
);
54 mal
->tx_chan_mask
|= commac
->tx_chan_mask
;
55 mal
->rx_chan_mask
|= commac
->rx_chan_mask
;
56 list_add(&commac
->list
, &mal
->list
);
58 spin_unlock_irqrestore(&mal
->lock
, flags
);
63 void mal_unregister_commac(struct mal_instance
*mal
,
64 struct mal_commac
*commac
)
68 spin_lock_irqsave(&mal
->lock
, flags
);
70 MAL_DBG(mal
, "unreg(%08x, %08x)" NL
,
71 commac
->tx_chan_mask
, commac
->rx_chan_mask
);
73 mal
->tx_chan_mask
&= ~commac
->tx_chan_mask
;
74 mal
->rx_chan_mask
&= ~commac
->rx_chan_mask
;
75 list_del_init(&commac
->list
);
76 if (list_empty(&mal
->list
))
77 napi_disable(&mal
->napi
);
79 spin_unlock_irqrestore(&mal
->lock
, flags
);
82 int mal_set_rcbs(struct mal_instance
*mal
, int channel
, unsigned long size
)
84 BUG_ON(channel
< 0 || channel
>= mal
->num_rx_chans
||
85 size
> MAL_MAX_RX_SIZE
);
87 MAL_DBG(mal
, "set_rbcs(%d, %lu)" NL
, channel
, size
);
91 "mal%d: incorrect RX size %lu for the channel %d\n",
92 mal
->index
, size
, channel
);
96 set_mal_dcrn(mal
, MAL_RCBS(channel
), size
>> 4);
100 int mal_tx_bd_offset(struct mal_instance
*mal
, int channel
)
102 BUG_ON(channel
< 0 || channel
>= mal
->num_tx_chans
);
104 return channel
* NUM_TX_BUFF
;
107 int mal_rx_bd_offset(struct mal_instance
*mal
, int channel
)
109 BUG_ON(channel
< 0 || channel
>= mal
->num_rx_chans
);
110 return mal
->num_tx_chans
* NUM_TX_BUFF
+ channel
* NUM_RX_BUFF
;
113 void mal_enable_tx_channel(struct mal_instance
*mal
, int channel
)
117 spin_lock_irqsave(&mal
->lock
, flags
);
119 MAL_DBG(mal
, "enable_tx(%d)" NL
, channel
);
121 set_mal_dcrn(mal
, MAL_TXCASR
,
122 get_mal_dcrn(mal
, MAL_TXCASR
) | MAL_CHAN_MASK(channel
));
124 spin_unlock_irqrestore(&mal
->lock
, flags
);
127 void mal_disable_tx_channel(struct mal_instance
*mal
, int channel
)
129 set_mal_dcrn(mal
, MAL_TXCARR
, MAL_CHAN_MASK(channel
));
131 MAL_DBG(mal
, "disable_tx(%d)" NL
, channel
);
134 void mal_enable_rx_channel(struct mal_instance
*mal
, int channel
)
139 * On some 4xx PPC's (e.g. 460EX/GT), the rx channel is a multiple
140 * of 8, but enabling in MAL_RXCASR needs the divided by 8 value
146 spin_lock_irqsave(&mal
->lock
, flags
);
148 MAL_DBG(mal
, "enable_rx(%d)" NL
, channel
);
150 set_mal_dcrn(mal
, MAL_RXCASR
,
151 get_mal_dcrn(mal
, MAL_RXCASR
) | MAL_CHAN_MASK(channel
));
153 spin_unlock_irqrestore(&mal
->lock
, flags
);
156 void mal_disable_rx_channel(struct mal_instance
*mal
, int channel
)
159 * On some 4xx PPC's (e.g. 460EX/GT), the rx channel is a multiple
160 * of 8, but enabling in MAL_RXCASR needs the divided by 8 value
166 set_mal_dcrn(mal
, MAL_RXCARR
, MAL_CHAN_MASK(channel
));
168 MAL_DBG(mal
, "disable_rx(%d)" NL
, channel
);
171 void mal_poll_add(struct mal_instance
*mal
, struct mal_commac
*commac
)
175 spin_lock_irqsave(&mal
->lock
, flags
);
177 MAL_DBG(mal
, "poll_add(%p)" NL
, commac
);
179 /* starts disabled */
180 set_bit(MAL_COMMAC_POLL_DISABLED
, &commac
->flags
);
182 list_add_tail(&commac
->poll_list
, &mal
->poll_list
);
184 spin_unlock_irqrestore(&mal
->lock
, flags
);
187 void mal_poll_del(struct mal_instance
*mal
, struct mal_commac
*commac
)
191 spin_lock_irqsave(&mal
->lock
, flags
);
193 MAL_DBG(mal
, "poll_del(%p)" NL
, commac
);
195 list_del(&commac
->poll_list
);
197 spin_unlock_irqrestore(&mal
->lock
, flags
);
200 /* synchronized by mal_poll() */
201 static inline void mal_enable_eob_irq(struct mal_instance
*mal
)
203 MAL_DBG2(mal
, "enable_irq" NL
);
205 // XXX might want to cache MAL_CFG as the DCR read can be slooooow
206 set_mal_dcrn(mal
, MAL_CFG
, get_mal_dcrn(mal
, MAL_CFG
) | MAL_CFG_EOPIE
);
209 /* synchronized by NAPI state */
210 static inline void mal_disable_eob_irq(struct mal_instance
*mal
)
212 // XXX might want to cache MAL_CFG as the DCR read can be slooooow
213 set_mal_dcrn(mal
, MAL_CFG
, get_mal_dcrn(mal
, MAL_CFG
) & ~MAL_CFG_EOPIE
);
215 MAL_DBG2(mal
, "disable_irq" NL
);
218 static irqreturn_t
mal_serr(int irq
, void *dev_instance
)
220 struct mal_instance
*mal
= dev_instance
;
222 u32 esr
= get_mal_dcrn(mal
, MAL_ESR
);
224 /* Clear the error status register */
225 set_mal_dcrn(mal
, MAL_ESR
, esr
);
227 MAL_DBG(mal
, "SERR %08x" NL
, esr
);
229 if (esr
& MAL_ESR_EVB
) {
230 if (esr
& MAL_ESR_DE
) {
231 /* We ignore Descriptor error,
232 * TXDE or RXDE interrupt will be generated anyway.
237 if (esr
& MAL_ESR_PEIN
) {
238 /* PLB error, it's probably buggy hardware or
239 * incorrect physical address in BD (i.e. bug)
243 "mal%d: system error, "
244 "PLB (ESR = 0x%08x)\n",
249 /* OPB error, it's probably buggy hardware or incorrect
254 "mal%d: system error, OPB (ESR = 0x%08x)\n",
260 static inline void mal_schedule_poll(struct mal_instance
*mal
)
262 if (likely(napi_schedule_prep(&mal
->napi
))) {
263 MAL_DBG2(mal
, "schedule_poll" NL
);
264 spin_lock(&mal
->lock
);
265 mal_disable_eob_irq(mal
);
266 spin_unlock(&mal
->lock
);
267 __napi_schedule(&mal
->napi
);
269 MAL_DBG2(mal
, "already in poll" NL
);
272 static irqreturn_t
mal_txeob(int irq
, void *dev_instance
)
274 struct mal_instance
*mal
= dev_instance
;
276 u32 r
= get_mal_dcrn(mal
, MAL_TXEOBISR
);
278 MAL_DBG2(mal
, "txeob %08x" NL
, r
);
280 mal_schedule_poll(mal
);
281 set_mal_dcrn(mal
, MAL_TXEOBISR
, r
);
283 #ifdef CONFIG_PPC_DCR_NATIVE
284 if (mal_has_feature(mal
, MAL_FTR_CLEAR_ICINTSTAT
))
285 mtdcri(SDR0
, DCRN_SDR_ICINTSTAT
,
286 (mfdcri(SDR0
, DCRN_SDR_ICINTSTAT
) | ICINTSTAT_ICTX
));
292 static irqreturn_t
mal_rxeob(int irq
, void *dev_instance
)
294 struct mal_instance
*mal
= dev_instance
;
296 u32 r
= get_mal_dcrn(mal
, MAL_RXEOBISR
);
298 MAL_DBG2(mal
, "rxeob %08x" NL
, r
);
300 mal_schedule_poll(mal
);
301 set_mal_dcrn(mal
, MAL_RXEOBISR
, r
);
303 #ifdef CONFIG_PPC_DCR_NATIVE
304 if (mal_has_feature(mal
, MAL_FTR_CLEAR_ICINTSTAT
))
305 mtdcri(SDR0
, DCRN_SDR_ICINTSTAT
,
306 (mfdcri(SDR0
, DCRN_SDR_ICINTSTAT
) | ICINTSTAT_ICRX
));
312 static irqreturn_t
mal_txde(int irq
, void *dev_instance
)
314 struct mal_instance
*mal
= dev_instance
;
316 u32 deir
= get_mal_dcrn(mal
, MAL_TXDEIR
);
317 set_mal_dcrn(mal
, MAL_TXDEIR
, deir
);
319 MAL_DBG(mal
, "txde %08x" NL
, deir
);
323 "mal%d: TX descriptor error (TXDEIR = 0x%08x)\n",
329 static irqreturn_t
mal_rxde(int irq
, void *dev_instance
)
331 struct mal_instance
*mal
= dev_instance
;
334 u32 deir
= get_mal_dcrn(mal
, MAL_RXDEIR
);
336 MAL_DBG(mal
, "rxde %08x" NL
, deir
);
338 list_for_each(l
, &mal
->list
) {
339 struct mal_commac
*mc
= list_entry(l
, struct mal_commac
, list
);
340 if (deir
& mc
->rx_chan_mask
) {
341 set_bit(MAL_COMMAC_RX_STOPPED
, &mc
->flags
);
342 mc
->ops
->rxde(mc
->dev
);
346 mal_schedule_poll(mal
);
347 set_mal_dcrn(mal
, MAL_RXDEIR
, deir
);
352 static irqreturn_t
mal_int(int irq
, void *dev_instance
)
354 struct mal_instance
*mal
= dev_instance
;
355 u32 esr
= get_mal_dcrn(mal
, MAL_ESR
);
357 if (esr
& MAL_ESR_EVB
) {
358 /* descriptor error */
359 if (esr
& MAL_ESR_DE
) {
360 if (esr
& MAL_ESR_CIDT
)
361 return mal_rxde(irq
, dev_instance
);
363 return mal_txde(irq
, dev_instance
);
365 return mal_serr(irq
, dev_instance
);
371 void mal_poll_disable(struct mal_instance
*mal
, struct mal_commac
*commac
)
373 /* Spinlock-type semantics: only one caller disable poll at a time */
374 while (test_and_set_bit(MAL_COMMAC_POLL_DISABLED
, &commac
->flags
))
377 /* Synchronize with the MAL NAPI poller */
378 napi_synchronize(&mal
->napi
);
381 void mal_poll_enable(struct mal_instance
*mal
, struct mal_commac
*commac
)
384 clear_bit(MAL_COMMAC_POLL_DISABLED
, &commac
->flags
);
386 /* Feels better to trigger a poll here to catch up with events that
387 * may have happened on this channel while disabled. It will most
388 * probably be delayed until the next interrupt but that's mostly a
389 * non-issue in the context where this is called.
391 napi_schedule(&mal
->napi
);
394 static int mal_poll(struct napi_struct
*napi
, int budget
)
396 struct mal_instance
*mal
= container_of(napi
, struct mal_instance
, napi
);
401 MAL_DBG2(mal
, "poll(%d)" NL
, budget
);
403 /* Process TX skbs */
404 list_for_each(l
, &mal
->poll_list
) {
405 struct mal_commac
*mc
=
406 list_entry(l
, struct mal_commac
, poll_list
);
407 mc
->ops
->poll_tx(mc
->dev
);
412 * We _might_ need something more smart here to enforce polling
415 list_for_each(l
, &mal
->poll_list
) {
416 struct mal_commac
*mc
=
417 list_entry(l
, struct mal_commac
, poll_list
);
419 if (unlikely(test_bit(MAL_COMMAC_POLL_DISABLED
, &mc
->flags
)))
421 n
= mc
->ops
->poll_rx(mc
->dev
, budget
- received
);
424 if (received
>= budget
)
429 if (napi_complete_done(napi
, received
)) {
430 /* We need to disable IRQs to protect from RXDE IRQ here */
431 spin_lock_irqsave(&mal
->lock
, flags
);
432 mal_enable_eob_irq(mal
);
433 spin_unlock_irqrestore(&mal
->lock
, flags
);
436 /* Check for "rotting" packet(s) */
437 list_for_each(l
, &mal
->poll_list
) {
438 struct mal_commac
*mc
=
439 list_entry(l
, struct mal_commac
, poll_list
);
440 if (unlikely(test_bit(MAL_COMMAC_POLL_DISABLED
, &mc
->flags
)))
442 if (unlikely(mc
->ops
->peek_rx(mc
->dev
) ||
443 test_bit(MAL_COMMAC_RX_STOPPED
, &mc
->flags
))) {
444 MAL_DBG2(mal
, "rotting packet" NL
);
445 if (!napi_schedule(napi
))
448 spin_lock_irqsave(&mal
->lock
, flags
);
449 mal_disable_eob_irq(mal
);
450 spin_unlock_irqrestore(&mal
->lock
, flags
);
452 mc
->ops
->poll_tx(mc
->dev
);
456 MAL_DBG2(mal
, "poll() %d <- %d" NL
, budget
, received
);
460 static void mal_reset(struct mal_instance
*mal
)
464 MAL_DBG(mal
, "reset" NL
);
466 set_mal_dcrn(mal
, MAL_CFG
, MAL_CFG_SR
);
468 /* Wait for reset to complete (1 system clock) */
469 while ((get_mal_dcrn(mal
, MAL_CFG
) & MAL_CFG_SR
) && n
)
473 printk(KERN_ERR
"mal%d: reset timeout\n", mal
->index
);
476 int mal_get_regs_len(struct mal_instance
*mal
)
478 return sizeof(struct emac_ethtool_regs_subhdr
) +
479 sizeof(struct mal_regs
);
482 void *mal_dump_regs(struct mal_instance
*mal
, void *buf
)
484 struct emac_ethtool_regs_subhdr
*hdr
= buf
;
485 struct mal_regs
*regs
= (struct mal_regs
*)(hdr
+ 1);
488 hdr
->version
= mal
->version
;
489 hdr
->index
= mal
->index
;
491 regs
->tx_count
= mal
->num_tx_chans
;
492 regs
->rx_count
= mal
->num_rx_chans
;
494 regs
->cfg
= get_mal_dcrn(mal
, MAL_CFG
);
495 regs
->esr
= get_mal_dcrn(mal
, MAL_ESR
);
496 regs
->ier
= get_mal_dcrn(mal
, MAL_IER
);
497 regs
->tx_casr
= get_mal_dcrn(mal
, MAL_TXCASR
);
498 regs
->tx_carr
= get_mal_dcrn(mal
, MAL_TXCARR
);
499 regs
->tx_eobisr
= get_mal_dcrn(mal
, MAL_TXEOBISR
);
500 regs
->tx_deir
= get_mal_dcrn(mal
, MAL_TXDEIR
);
501 regs
->rx_casr
= get_mal_dcrn(mal
, MAL_RXCASR
);
502 regs
->rx_carr
= get_mal_dcrn(mal
, MAL_RXCARR
);
503 regs
->rx_eobisr
= get_mal_dcrn(mal
, MAL_RXEOBISR
);
504 regs
->rx_deir
= get_mal_dcrn(mal
, MAL_RXDEIR
);
506 for (i
= 0; i
< regs
->tx_count
; ++i
)
507 regs
->tx_ctpr
[i
] = get_mal_dcrn(mal
, MAL_TXCTPR(i
));
509 for (i
= 0; i
< regs
->rx_count
; ++i
) {
510 regs
->rx_ctpr
[i
] = get_mal_dcrn(mal
, MAL_RXCTPR(i
));
511 regs
->rcbs
[i
] = get_mal_dcrn(mal
, MAL_RCBS(i
));
516 static int mal_probe(struct platform_device
*ofdev
)
518 struct mal_instance
*mal
;
519 int err
= 0, i
, bd_size
;
520 int index
= mal_count
++;
521 unsigned int dcr_base
;
524 unsigned long irqflags
;
525 irq_handler_t hdlr_serr
, hdlr_txde
, hdlr_rxde
;
527 mal
= devm_kzalloc(&ofdev
->dev
, sizeof(struct mal_instance
),
534 mal
->version
= of_device_is_compatible(ofdev
->dev
.of_node
, "ibm,mcmal2") ? 2 : 1;
536 MAL_DBG(mal
, "probe" NL
);
538 prop
= of_get_property(ofdev
->dev
.of_node
, "num-tx-chans", NULL
);
541 "mal%d: can't find MAL num-tx-chans property!\n",
545 mal
->num_tx_chans
= prop
[0];
547 prop
= of_get_property(ofdev
->dev
.of_node
, "num-rx-chans", NULL
);
550 "mal%d: can't find MAL num-rx-chans property!\n",
554 mal
->num_rx_chans
= prop
[0];
556 dcr_base
= dcr_resource_start(ofdev
->dev
.of_node
, 0);
559 "mal%d: can't find DCR resource!\n", index
);
562 mal
->dcr_host
= dcr_map(ofdev
->dev
.of_node
, dcr_base
, 0x100);
563 if (!DCR_MAP_OK(mal
->dcr_host
)) {
565 "mal%d: failed to map DCRs !\n", index
);
569 if (of_device_is_compatible(ofdev
->dev
.of_node
, "ibm,mcmal-405ez")) {
570 #if defined(CONFIG_IBM_EMAC_MAL_CLR_ICINTSTAT) && \
571 defined(CONFIG_IBM_EMAC_MAL_COMMON_ERR)
572 mal
->features
|= (MAL_FTR_CLEAR_ICINTSTAT
|
573 MAL_FTR_COMMON_ERR_INT
);
575 printk(KERN_ERR
"%pOF: Support for 405EZ not enabled!\n",
582 INIT_LIST_HEAD(&mal
->poll_list
);
583 INIT_LIST_HEAD(&mal
->list
);
584 spin_lock_init(&mal
->lock
);
586 mal
->dummy_dev
= alloc_netdev_dummy(0);
587 if (!mal
->dummy_dev
) {
592 netif_napi_add_weight(mal
->dummy_dev
, &mal
->napi
, mal_poll
,
593 CONFIG_IBM_EMAC_POLL_WEIGHT
);
595 /* Load power-on reset defaults */
598 /* Set the MAL configuration register */
599 cfg
= (mal
->version
== 2) ? MAL2_CFG_DEFAULT
: MAL1_CFG_DEFAULT
;
600 cfg
|= MAL_CFG_PLBB
| MAL_CFG_OPBBL
| MAL_CFG_LEA
;
602 /* Current Axon is not happy with priority being non-0, it can
603 * deadlock, fix it up here
605 if (of_device_is_compatible(ofdev
->dev
.of_node
, "ibm,mcmal-axon"))
606 cfg
&= ~(MAL2_CFG_RPP_10
| MAL2_CFG_WPP_10
);
608 /* Apply configuration */
609 set_mal_dcrn(mal
, MAL_CFG
, cfg
);
611 /* Allocate space for BD rings */
612 BUG_ON(mal
->num_tx_chans
<= 0 || mal
->num_tx_chans
> 32);
613 BUG_ON(mal
->num_rx_chans
<= 0 || mal
->num_rx_chans
> 32);
615 bd_size
= sizeof(struct mal_descriptor
) *
616 (NUM_TX_BUFF
* mal
->num_tx_chans
+
617 NUM_RX_BUFF
* mal
->num_rx_chans
);
618 mal
->bd_virt
= dma_alloc_coherent(&ofdev
->dev
, bd_size
, &mal
->bd_dma
,
620 if (mal
->bd_virt
== NULL
) {
625 for (i
= 0; i
< mal
->num_tx_chans
; ++i
)
626 set_mal_dcrn(mal
, MAL_TXCTPR(i
), mal
->bd_dma
+
627 sizeof(struct mal_descriptor
) *
628 mal_tx_bd_offset(mal
, i
));
630 for (i
= 0; i
< mal
->num_rx_chans
; ++i
)
631 set_mal_dcrn(mal
, MAL_RXCTPR(i
), mal
->bd_dma
+
632 sizeof(struct mal_descriptor
) *
633 mal_rx_bd_offset(mal
, i
));
635 mal
->txeob_irq
= platform_get_irq(ofdev
, 0);
636 mal
->rxeob_irq
= platform_get_irq(ofdev
, 1);
637 mal
->serr_irq
= platform_get_irq(ofdev
, 2);
639 if (mal_has_feature(mal
, MAL_FTR_COMMON_ERR_INT
)) {
640 mal
->txde_irq
= mal
->rxde_irq
= mal
->serr_irq
;
641 irqflags
= IRQF_SHARED
;
642 hdlr_serr
= hdlr_txde
= hdlr_rxde
= mal_int
;
644 mal
->txde_irq
= platform_get_irq(ofdev
, 3);
645 mal
->rxde_irq
= platform_get_irq(ofdev
, 4);
647 hdlr_serr
= mal_serr
;
648 hdlr_txde
= mal_txde
;
649 hdlr_rxde
= mal_rxde
;
652 err
= devm_request_irq(&ofdev
->dev
, mal
->serr_irq
, hdlr_serr
, irqflags
,
656 err
= devm_request_irq(&ofdev
->dev
, mal
->txde_irq
, hdlr_txde
, irqflags
,
660 err
= devm_request_irq(&ofdev
->dev
, mal
->txeob_irq
, mal_txeob
, 0,
664 err
= devm_request_irq(&ofdev
->dev
, mal
->rxde_irq
, hdlr_rxde
, irqflags
,
668 err
= devm_request_irq(&ofdev
->dev
, mal
->rxeob_irq
, mal_rxeob
, 0,
673 /* Enable all MAL SERR interrupt sources */
674 set_mal_dcrn(mal
, MAL_IER
, MAL_IER_EVENTS
);
676 /* Enable EOB interrupt */
677 mal_enable_eob_irq(mal
);
680 "MAL v%d %pOF, %d TX channels, %d RX channels\n",
681 mal
->version
, ofdev
->dev
.of_node
,
682 mal
->num_tx_chans
, mal
->num_rx_chans
);
684 /* Advertise this instance to the rest of the world */
686 platform_set_drvdata(ofdev
, mal
);
691 dma_free_coherent(&ofdev
->dev
, bd_size
, mal
->bd_virt
, mal
->bd_dma
);
693 free_netdev(mal
->dummy_dev
);
695 dcr_unmap(mal
->dcr_host
, 0x100);
699 static void mal_remove(struct platform_device
*ofdev
)
701 struct mal_instance
*mal
= platform_get_drvdata(ofdev
);
703 MAL_DBG(mal
, "remove" NL
);
705 /* Synchronize with scheduled polling */
706 napi_disable(&mal
->napi
);
708 if (!list_empty(&mal
->list
))
709 /* This is *very* bad */
711 "mal%d: commac list is not empty on remove!\n",
716 free_netdev(mal
->dummy_dev
);
718 dcr_unmap(mal
->dcr_host
, 0x100);
720 dma_free_coherent(&ofdev
->dev
,
721 sizeof(struct mal_descriptor
) *
722 (NUM_TX_BUFF
* mal
->num_tx_chans
+
723 NUM_RX_BUFF
* mal
->num_rx_chans
),
724 mal
->bd_virt
, mal
->bd_dma
);
727 static const struct of_device_id mal_platform_match
[] =
730 .compatible
= "ibm,mcmal",
733 .compatible
= "ibm,mcmal2",
735 /* Backward compat */
738 .compatible
= "ibm,mcmal",
742 .compatible
= "ibm,mcmal2",
747 static struct platform_driver mal_of_driver
= {
750 .of_match_table
= mal_platform_match
,
753 .remove
= mal_remove
,
756 int __init
mal_init(void)
758 return platform_driver_register(&mal_of_driver
);
763 platform_driver_unregister(&mal_of_driver
);