1 // SPDX-License-Identifier: GPL-2.0-only
3 * linux/drivers/mmc/host/pxa.c - PXA MMCI driver
5 * Copyright (C) 2003 Russell King, All Rights Reserved.
7 * This hardware is really sick:
8 * - No way to clear interrupts.
9 * - Have to turn off the clock whenever we touch the device.
10 * - Doesn't tell you how many data blocks were transferred.
13 * 1 and 3 byte data transfers not supported
14 * max block length up to 1023
16 #include <linux/module.h>
17 #include <linux/init.h>
18 #include <linux/ioport.h>
19 #include <linux/platform_device.h>
20 #include <linux/delay.h>
21 #include <linux/interrupt.h>
22 #include <linux/dmaengine.h>
23 #include <linux/dma-mapping.h>
24 #include <linux/clk.h>
25 #include <linux/err.h>
26 #include <linux/mmc/host.h>
27 #include <linux/mmc/slot-gpio.h>
29 #include <linux/regulator/consumer.h>
30 #include <linux/gpio/consumer.h>
31 #include <linux/gfp.h>
33 #include <linux/soc/pxa/cpu.h>
35 #include <linux/sizes.h>
37 #include <linux/platform_data/mmc-pxamci.h>
41 #define DRIVER_NAME "pxa2xx-mci"
44 #define CLKRT_OFF (~0)
46 #define mmc_has_26MHz() (cpu_is_pxa300() || cpu_is_pxa310() \
55 unsigned long clkrate
;
59 unsigned int power_mode
;
60 unsigned long detect_delay_ms
;
62 struct gpio_desc
*power
;
63 struct pxamci_platform_data
*pdata
;
65 struct mmc_request
*mrq
;
66 struct mmc_command
*cmd
;
67 struct mmc_data
*data
;
69 struct dma_chan
*dma_chan_rx
;
70 struct dma_chan
*dma_chan_tx
;
71 dma_cookie_t dma_cookie
;
76 static int pxamci_init_ocr(struct pxamci_host
*host
)
78 struct mmc_host
*mmc
= host
->mmc
;
81 ret
= mmc_regulator_get_supply(mmc
);
85 if (IS_ERR(mmc
->supply
.vmmc
)) {
86 /* fall-back to platform data */
87 mmc
->ocr_avail
= host
->pdata
?
88 host
->pdata
->ocr_mask
:
89 MMC_VDD_32_33
| MMC_VDD_33_34
;
95 static inline int pxamci_set_power(struct pxamci_host
*host
,
96 unsigned char power_mode
,
99 struct mmc_host
*mmc
= host
->mmc
;
100 struct regulator
*supply
= mmc
->supply
.vmmc
;
103 return mmc_regulator_set_ocr(mmc
, supply
, vdd
);
106 bool on
= !!((1 << vdd
) & host
->pdata
->ocr_mask
);
107 gpiod_set_value(host
->power
, on
);
110 if (host
->pdata
&& host
->pdata
->setpower
)
111 return host
->pdata
->setpower(mmc_dev(host
->mmc
), vdd
);
116 static void pxamci_stop_clock(struct pxamci_host
*host
)
118 if (readl(host
->base
+ MMC_STAT
) & STAT_CLK_EN
) {
119 unsigned long timeout
= 10000;
122 writel(STOP_CLOCK
, host
->base
+ MMC_STRPCL
);
125 v
= readl(host
->base
+ MMC_STAT
);
126 if (!(v
& STAT_CLK_EN
))
132 dev_err(mmc_dev(host
->mmc
), "unable to stop clock\n");
136 static void pxamci_enable_irq(struct pxamci_host
*host
, unsigned int mask
)
140 spin_lock_irqsave(&host
->lock
, flags
);
141 host
->imask
&= ~mask
;
142 writel(host
->imask
, host
->base
+ MMC_I_MASK
);
143 spin_unlock_irqrestore(&host
->lock
, flags
);
146 static void pxamci_disable_irq(struct pxamci_host
*host
, unsigned int mask
)
150 spin_lock_irqsave(&host
->lock
, flags
);
152 writel(host
->imask
, host
->base
+ MMC_I_MASK
);
153 spin_unlock_irqrestore(&host
->lock
, flags
);
156 static void pxamci_dma_irq(void *param
);
158 static void pxamci_setup_data(struct pxamci_host
*host
, struct mmc_data
*data
)
160 struct dma_async_tx_descriptor
*tx
;
161 enum dma_transfer_direction direction
;
162 struct dma_slave_config config
;
163 struct dma_chan
*chan
;
164 unsigned int nob
= data
->blocks
;
165 unsigned long long clks
;
166 unsigned int timeout
;
171 writel(nob
, host
->base
+ MMC_NOB
);
172 writel(data
->blksz
, host
->base
+ MMC_BLKLEN
);
174 clks
= (unsigned long long)data
->timeout_ns
* host
->clkrate
;
175 do_div(clks
, 1000000000UL);
176 timeout
= (unsigned int)clks
+ (data
->timeout_clks
<< host
->clkrt
);
177 writel((timeout
+ 255) / 256, host
->base
+ MMC_RDTO
);
179 memset(&config
, 0, sizeof(config
));
180 config
.src_addr_width
= DMA_SLAVE_BUSWIDTH_1_BYTE
;
181 config
.dst_addr_width
= DMA_SLAVE_BUSWIDTH_1_BYTE
;
182 config
.src_addr
= host
->res
->start
+ MMC_RXFIFO
;
183 config
.dst_addr
= host
->res
->start
+ MMC_TXFIFO
;
184 config
.src_maxburst
= 32;
185 config
.dst_maxburst
= 32;
187 if (data
->flags
& MMC_DATA_READ
) {
188 host
->dma_dir
= DMA_FROM_DEVICE
;
189 direction
= DMA_DEV_TO_MEM
;
190 chan
= host
->dma_chan_rx
;
192 host
->dma_dir
= DMA_TO_DEVICE
;
193 direction
= DMA_MEM_TO_DEV
;
194 chan
= host
->dma_chan_tx
;
197 config
.direction
= direction
;
199 ret
= dmaengine_slave_config(chan
, &config
);
201 dev_err(mmc_dev(host
->mmc
), "dma slave config failed\n");
205 host
->dma_len
= dma_map_sg(chan
->device
->dev
, data
->sg
, data
->sg_len
,
208 tx
= dmaengine_prep_slave_sg(chan
, data
->sg
, host
->dma_len
, direction
,
211 dev_err(mmc_dev(host
->mmc
), "prep_slave_sg() failed\n");
215 if (!(data
->flags
& MMC_DATA_READ
)) {
216 tx
->callback
= pxamci_dma_irq
;
217 tx
->callback_param
= host
;
220 host
->dma_cookie
= dmaengine_submit(tx
);
223 * workaround for erratum #91:
224 * only start DMA now if we are doing a read,
225 * otherwise we wait until CMD/RESP has finished
226 * before starting DMA.
228 if (!cpu_is_pxa27x() || data
->flags
& MMC_DATA_READ
)
229 dma_async_issue_pending(chan
);
232 static void pxamci_start_cmd(struct pxamci_host
*host
, struct mmc_command
*cmd
, unsigned int cmdat
)
234 WARN_ON(host
->cmd
!= NULL
);
237 if (cmd
->flags
& MMC_RSP_BUSY
)
240 #define RSP_TYPE(x) ((x) & ~(MMC_RSP_BUSY|MMC_RSP_OPCODE))
241 switch (RSP_TYPE(mmc_resp_type(cmd
))) {
242 case RSP_TYPE(MMC_RSP_R1
): /* r1, r1b, r6, r7 */
243 cmdat
|= CMDAT_RESP_SHORT
;
245 case RSP_TYPE(MMC_RSP_R3
):
246 cmdat
|= CMDAT_RESP_R3
;
248 case RSP_TYPE(MMC_RSP_R2
):
249 cmdat
|= CMDAT_RESP_R2
;
255 writel(cmd
->opcode
, host
->base
+ MMC_CMD
);
256 writel(cmd
->arg
>> 16, host
->base
+ MMC_ARGH
);
257 writel(cmd
->arg
& 0xffff, host
->base
+ MMC_ARGL
);
258 writel(cmdat
, host
->base
+ MMC_CMDAT
);
259 writel(host
->clkrt
, host
->base
+ MMC_CLKRT
);
261 writel(START_CLOCK
, host
->base
+ MMC_STRPCL
);
263 pxamci_enable_irq(host
, END_CMD_RES
);
266 static void pxamci_finish_request(struct pxamci_host
*host
, struct mmc_request
*mrq
)
271 mmc_request_done(host
->mmc
, mrq
);
274 static int pxamci_cmd_done(struct pxamci_host
*host
, unsigned int stat
)
276 struct mmc_command
*cmd
= host
->cmd
;
286 * Did I mention this is Sick. We always need to
287 * discard the upper 8 bits of the first 16-bit word.
289 v
= readl(host
->base
+ MMC_RES
) & 0xffff;
290 for (i
= 0; i
< 4; i
++) {
291 u32 w1
= readl(host
->base
+ MMC_RES
) & 0xffff;
292 u32 w2
= readl(host
->base
+ MMC_RES
) & 0xffff;
293 cmd
->resp
[i
] = v
<< 24 | w1
<< 8 | w2
>> 8;
297 if (stat
& STAT_TIME_OUT_RESPONSE
) {
298 cmd
->error
= -ETIMEDOUT
;
299 } else if (stat
& STAT_RES_CRC_ERR
&& cmd
->flags
& MMC_RSP_CRC
) {
301 * workaround for erratum #42:
302 * Intel PXA27x Family Processor Specification Update Rev 001
303 * A bogus CRC error can appear if the msb of a 136 bit
306 if (cpu_is_pxa27x() &&
307 (cmd
->flags
& MMC_RSP_136
&& cmd
->resp
[0] & 0x80000000))
308 pr_debug("ignoring CRC from command %d - *risky*\n", cmd
->opcode
);
310 cmd
->error
= -EILSEQ
;
313 pxamci_disable_irq(host
, END_CMD_RES
);
314 if (host
->data
&& !cmd
->error
) {
315 pxamci_enable_irq(host
, DATA_TRAN_DONE
);
317 * workaround for erratum #91, if doing write
320 if (cpu_is_pxa27x() && host
->data
->flags
& MMC_DATA_WRITE
)
321 dma_async_issue_pending(host
->dma_chan_tx
);
323 pxamci_finish_request(host
, host
->mrq
);
329 static int pxamci_data_done(struct pxamci_host
*host
, unsigned int stat
)
331 struct mmc_data
*data
= host
->data
;
332 struct dma_chan
*chan
;
337 if (data
->flags
& MMC_DATA_READ
)
338 chan
= host
->dma_chan_rx
;
340 chan
= host
->dma_chan_tx
;
341 dma_unmap_sg(chan
->device
->dev
,
342 data
->sg
, data
->sg_len
, host
->dma_dir
);
344 if (stat
& STAT_READ_TIME_OUT
)
345 data
->error
= -ETIMEDOUT
;
346 else if (stat
& (STAT_CRC_READ_ERROR
|STAT_CRC_WRITE_ERROR
))
347 data
->error
= -EILSEQ
;
350 * There appears to be a hardware design bug here. There seems to
351 * be no way to find out how much data was transferred to the card.
352 * This means that if there was an error on any block, we mark all
353 * data blocks as being in error.
356 data
->bytes_xfered
= data
->blocks
* data
->blksz
;
358 data
->bytes_xfered
= 0;
360 pxamci_disable_irq(host
, DATA_TRAN_DONE
);
363 if (host
->mrq
->stop
) {
364 pxamci_stop_clock(host
);
365 pxamci_start_cmd(host
, host
->mrq
->stop
, host
->cmdat
);
367 pxamci_finish_request(host
, host
->mrq
);
373 static irqreturn_t
pxamci_irq(int irq
, void *devid
)
375 struct pxamci_host
*host
= devid
;
379 ireg
= readl(host
->base
+ MMC_I_REG
) & ~readl(host
->base
+ MMC_I_MASK
);
382 unsigned stat
= readl(host
->base
+ MMC_STAT
);
384 pr_debug("PXAMCI: irq %08x stat %08x\n", ireg
, stat
);
386 if (ireg
& END_CMD_RES
)
387 handled
|= pxamci_cmd_done(host
, stat
);
388 if (ireg
& DATA_TRAN_DONE
)
389 handled
|= pxamci_data_done(host
, stat
);
390 if (ireg
& SDIO_INT
) {
391 mmc_signal_sdio_irq(host
->mmc
);
396 return IRQ_RETVAL(handled
);
399 static void pxamci_request(struct mmc_host
*mmc
, struct mmc_request
*mrq
)
401 struct pxamci_host
*host
= mmc_priv(mmc
);
404 WARN_ON(host
->mrq
!= NULL
);
408 pxamci_stop_clock(host
);
411 host
->cmdat
&= ~CMDAT_INIT
;
414 pxamci_setup_data(host
, mrq
->data
);
416 cmdat
&= ~CMDAT_BUSY
;
417 cmdat
|= CMDAT_DATAEN
| CMDAT_DMAEN
;
418 if (mrq
->data
->flags
& MMC_DATA_WRITE
)
419 cmdat
|= CMDAT_WRITE
;
422 pxamci_start_cmd(host
, mrq
->cmd
, cmdat
);
425 static int pxamci_get_ro(struct mmc_host
*mmc
)
427 struct pxamci_host
*host
= mmc_priv(mmc
);
429 if (host
->use_ro_gpio
)
430 return mmc_gpio_get_ro(mmc
);
431 if (host
->pdata
&& host
->pdata
->get_ro
)
432 return !!host
->pdata
->get_ro(mmc_dev(mmc
));
434 * Board doesn't support read only detection; let the mmc core
440 static void pxamci_set_ios(struct mmc_host
*mmc
, struct mmc_ios
*ios
)
442 struct pxamci_host
*host
= mmc_priv(mmc
);
445 unsigned long rate
= host
->clkrate
;
446 unsigned int clk
= rate
/ ios
->clock
;
448 if (host
->clkrt
== CLKRT_OFF
)
449 clk_prepare_enable(host
->clk
);
451 if (ios
->clock
== 26000000) {
452 /* to support 26MHz */
455 /* to handle (19.5MHz, 26MHz) */
460 * clk might result in a lower divisor than we
461 * desire. check for that condition and adjust
464 if (rate
/ clk
> ios
->clock
)
466 host
->clkrt
= fls(clk
) - 1;
470 * we write clkrt on the next command
473 pxamci_stop_clock(host
);
474 if (host
->clkrt
!= CLKRT_OFF
) {
475 host
->clkrt
= CLKRT_OFF
;
476 clk_disable_unprepare(host
->clk
);
480 if (host
->power_mode
!= ios
->power_mode
) {
483 host
->power_mode
= ios
->power_mode
;
485 ret
= pxamci_set_power(host
, ios
->power_mode
, ios
->vdd
);
487 dev_err(mmc_dev(mmc
), "unable to set power\n");
489 * The .set_ios() function in the mmc_host_ops
490 * struct return void, and failing to set the
491 * power should be rare so we print an error and
497 if (ios
->power_mode
== MMC_POWER_ON
)
498 host
->cmdat
|= CMDAT_INIT
;
501 if (ios
->bus_width
== MMC_BUS_WIDTH_4
)
502 host
->cmdat
|= CMDAT_SD_4DAT
;
504 host
->cmdat
&= ~CMDAT_SD_4DAT
;
506 dev_dbg(mmc_dev(mmc
), "PXAMCI: clkrt = %x cmdat = %x\n",
507 host
->clkrt
, host
->cmdat
);
510 static void pxamci_enable_sdio_irq(struct mmc_host
*host
, int enable
)
512 struct pxamci_host
*pxa_host
= mmc_priv(host
);
515 pxamci_enable_irq(pxa_host
, SDIO_INT
);
517 pxamci_disable_irq(pxa_host
, SDIO_INT
);
520 static const struct mmc_host_ops pxamci_ops
= {
521 .request
= pxamci_request
,
522 .get_cd
= mmc_gpio_get_cd
,
523 .get_ro
= pxamci_get_ro
,
524 .set_ios
= pxamci_set_ios
,
525 .enable_sdio_irq
= pxamci_enable_sdio_irq
,
528 static void pxamci_dma_irq(void *param
)
530 struct pxamci_host
*host
= param
;
531 struct dma_tx_state state
;
532 enum dma_status status
;
533 struct dma_chan
*chan
;
536 spin_lock_irqsave(&host
->lock
, flags
);
541 if (host
->data
->flags
& MMC_DATA_READ
)
542 chan
= host
->dma_chan_rx
;
544 chan
= host
->dma_chan_tx
;
546 status
= dmaengine_tx_status(chan
, host
->dma_cookie
, &state
);
548 if (likely(status
== DMA_COMPLETE
)) {
549 writel(BUF_PART_FULL
, host
->base
+ MMC_PRTBUF
);
551 pr_err("%s: DMA error on %s channel\n", mmc_hostname(host
->mmc
),
552 host
->data
->flags
& MMC_DATA_READ
? "rx" : "tx");
553 host
->data
->error
= -EIO
;
554 pxamci_data_done(host
, 0);
558 spin_unlock_irqrestore(&host
->lock
, flags
);
561 static irqreturn_t
pxamci_detect_irq(int irq
, void *devid
)
563 struct pxamci_host
*host
= mmc_priv(devid
);
565 mmc_detect_change(devid
, msecs_to_jiffies(host
->detect_delay_ms
));
570 static const struct of_device_id pxa_mmc_dt_ids
[] = {
571 { .compatible
= "marvell,pxa-mmc" },
575 MODULE_DEVICE_TABLE(of
, pxa_mmc_dt_ids
);
577 static int pxamci_of_init(struct platform_device
*pdev
,
578 struct mmc_host
*mmc
)
580 struct device_node
*np
= pdev
->dev
.of_node
;
581 struct pxamci_host
*host
= mmc_priv(mmc
);
588 /* pxa-mmc specific */
589 if (of_property_read_u32(np
, "pxa-mmc,detect-delay-ms", &tmp
) == 0)
590 host
->detect_delay_ms
= tmp
;
592 ret
= mmc_of_parse(mmc
);
599 static int pxamci_of_init(struct platform_device
*pdev
,
600 struct mmc_host
*mmc
)
606 static int pxamci_probe(struct platform_device
*pdev
)
608 struct mmc_host
*mmc
;
609 struct pxamci_host
*host
= NULL
;
610 struct device
*dev
= &pdev
->dev
;
614 irq
= platform_get_irq(pdev
, 0);
618 mmc
= mmc_alloc_host(sizeof(struct pxamci_host
), dev
);
624 mmc
->ops
= &pxamci_ops
;
627 * We can do SG-DMA, but we don't because we never know how much
628 * data we successfully wrote to the card.
630 mmc
->max_segs
= NR_SG
;
633 * Our hardware DMA can handle a maximum of one page per SG entry.
635 mmc
->max_seg_size
= PAGE_SIZE
;
638 * Block length register is only 10 bits before PXA27x.
640 mmc
->max_blk_size
= cpu_is_pxa25x() ? 1023 : 2048;
643 * Block count register is 16 bits.
645 mmc
->max_blk_count
= 65535;
647 ret
= pxamci_of_init(pdev
, mmc
);
651 host
= mmc_priv(mmc
);
653 host
->pdata
= pdev
->dev
.platform_data
;
654 host
->clkrt
= CLKRT_OFF
;
656 host
->clk
= devm_clk_get(dev
, NULL
);
657 if (IS_ERR(host
->clk
)) {
658 ret
= PTR_ERR(host
->clk
);
663 host
->clkrate
= clk_get_rate(host
->clk
);
666 * Calculate minimum clock rate, rounding up.
668 mmc
->f_min
= (host
->clkrate
+ 63) / 64;
669 mmc
->f_max
= (mmc_has_26MHz()) ? 26000000 : host
->clkrate
;
671 ret
= pxamci_init_ocr(host
);
677 if (!cpu_is_pxa25x()) {
678 mmc
->caps
|= MMC_CAP_4_BIT_DATA
| MMC_CAP_SDIO_IRQ
;
679 host
->cmdat
|= CMDAT_SDIO_INT_EN
;
681 mmc
->caps
|= MMC_CAP_MMC_HIGHSPEED
|
682 MMC_CAP_SD_HIGHSPEED
;
685 spin_lock_init(&host
->lock
);
686 host
->imask
= MMC_I_MASK_ALL
;
688 host
->base
= devm_platform_get_and_ioremap_resource(pdev
, 0, &r
);
689 if (IS_ERR(host
->base
)) {
690 ret
= PTR_ERR(host
->base
);
696 * Ensure that the host controller is shut down, and setup
699 pxamci_stop_clock(host
);
700 writel(0, host
->base
+ MMC_SPI
);
701 writel(64, host
->base
+ MMC_RESTO
);
702 writel(host
->imask
, host
->base
+ MMC_I_MASK
);
704 ret
= devm_request_irq(dev
, irq
, pxamci_irq
, 0,
709 platform_set_drvdata(pdev
, mmc
);
711 host
->dma_chan_rx
= dma_request_chan(dev
, "rx");
712 if (IS_ERR(host
->dma_chan_rx
)) {
713 dev_err(dev
, "unable to request rx dma channel\n");
714 ret
= PTR_ERR(host
->dma_chan_rx
);
715 host
->dma_chan_rx
= NULL
;
719 host
->dma_chan_tx
= dma_request_chan(dev
, "tx");
720 if (IS_ERR(host
->dma_chan_tx
)) {
721 dev_err(dev
, "unable to request tx dma channel\n");
722 ret
= PTR_ERR(host
->dma_chan_tx
);
723 host
->dma_chan_tx
= NULL
;
728 host
->detect_delay_ms
= host
->pdata
->detect_delay_ms
;
730 host
->power
= devm_gpiod_get_optional(dev
, "power", GPIOD_OUT_LOW
);
731 if (IS_ERR(host
->power
)) {
732 ret
= PTR_ERR(host
->power
);
733 dev_err(dev
, "Failed requesting gpio_power\n");
737 /* FIXME: should we pass detection delay to debounce? */
738 ret
= mmc_gpiod_request_cd(mmc
, "cd", 0, false, 0);
739 if (ret
&& ret
!= -ENOENT
) {
740 dev_err(dev
, "Failed requesting gpio_cd\n");
744 if (!host
->pdata
->gpio_card_ro_invert
)
745 mmc
->caps2
|= MMC_CAP2_RO_ACTIVE_HIGH
;
747 ret
= mmc_gpiod_request_ro(mmc
, "wp", 0, 0);
748 if (ret
&& ret
!= -ENOENT
) {
749 dev_err(dev
, "Failed requesting gpio_ro\n");
753 host
->use_ro_gpio
= true;
755 if (host
->pdata
->init
)
756 host
->pdata
->init(dev
, pxamci_detect_irq
, mmc
);
758 if (host
->power
&& host
->pdata
->setpower
)
759 dev_warn(dev
, "gpio_power and setpower() both defined\n");
760 if (host
->use_ro_gpio
&& host
->pdata
->get_ro
)
761 dev_warn(dev
, "gpio_ro and get_ro() both defined\n");
764 ret
= mmc_add_host(mmc
);
766 if (host
->pdata
&& host
->pdata
->exit
)
767 host
->pdata
->exit(dev
, mmc
);
775 if (host
->dma_chan_rx
)
776 dma_release_channel(host
->dma_chan_rx
);
777 if (host
->dma_chan_tx
)
778 dma_release_channel(host
->dma_chan_tx
);
785 static void pxamci_remove(struct platform_device
*pdev
)
787 struct mmc_host
*mmc
= platform_get_drvdata(pdev
);
790 struct pxamci_host
*host
= mmc_priv(mmc
);
792 mmc_remove_host(mmc
);
794 if (host
->pdata
&& host
->pdata
->exit
)
795 host
->pdata
->exit(&pdev
->dev
, mmc
);
797 pxamci_stop_clock(host
);
798 writel(TXFIFO_WR_REQ
|RXFIFO_RD_REQ
|CLK_IS_OFF
|STOP_CMD
|
799 END_CMD_RES
|PRG_DONE
|DATA_TRAN_DONE
,
800 host
->base
+ MMC_I_MASK
);
802 dmaengine_terminate_all(host
->dma_chan_rx
);
803 dmaengine_terminate_all(host
->dma_chan_tx
);
804 dma_release_channel(host
->dma_chan_rx
);
805 dma_release_channel(host
->dma_chan_tx
);
811 static struct platform_driver pxamci_driver
= {
812 .probe
= pxamci_probe
,
813 .remove
= pxamci_remove
,
816 .probe_type
= PROBE_PREFER_ASYNCHRONOUS
,
817 .of_match_table
= of_match_ptr(pxa_mmc_dt_ids
),
821 module_platform_driver(pxamci_driver
);
823 MODULE_DESCRIPTION("PXA Multimedia Card Interface Driver");
824 MODULE_LICENSE("GPL");
825 MODULE_ALIAS("platform:pxa2xx-mci");