2 * linux/drivers/mmc/host/omap.c
4 * Copyright (C) 2004 Nokia Corporation
5 * Written by Tuukka Tikkanen and Juha Yrjölä<juha.yrjola@nokia.com>
6 * Misc hacks here and there by Tony Lindgren <tony@atomide.com>
7 * Other hacks (DMA, SD, etc) by David Brownell
9 * This program is free software; you can redistribute it and/or modify
10 * it under the terms of the GNU General Public License version 2 as
11 * published by the Free Software Foundation.
14 #include <linux/module.h>
15 #include <linux/moduleparam.h>
16 #include <linux/init.h>
17 #include <linux/ioport.h>
18 #include <linux/platform_device.h>
19 #include <linux/interrupt.h>
20 #include <linux/dma-mapping.h>
21 #include <linux/delay.h>
22 #include <linux/spinlock.h>
23 #include <linux/timer.h>
24 #include <linux/mmc/host.h>
25 #include <linux/mmc/card.h>
26 #include <linux/clk.h>
27 #include <linux/scatterlist.h>
28 #include <linux/i2c/tps65010.h>
29 #include <linux/slab.h>
34 #include <plat/board.h>
39 #include <plat/fpga.h>
41 #define OMAP_MMC_REG_CMD 0x00
42 #define OMAP_MMC_REG_ARGL 0x01
43 #define OMAP_MMC_REG_ARGH 0x02
44 #define OMAP_MMC_REG_CON 0x03
45 #define OMAP_MMC_REG_STAT 0x04
46 #define OMAP_MMC_REG_IE 0x05
47 #define OMAP_MMC_REG_CTO 0x06
48 #define OMAP_MMC_REG_DTO 0x07
49 #define OMAP_MMC_REG_DATA 0x08
50 #define OMAP_MMC_REG_BLEN 0x09
51 #define OMAP_MMC_REG_NBLK 0x0a
52 #define OMAP_MMC_REG_BUF 0x0b
53 #define OMAP_MMC_REG_SDIO 0x0d
54 #define OMAP_MMC_REG_REV 0x0f
55 #define OMAP_MMC_REG_RSP0 0x10
56 #define OMAP_MMC_REG_RSP1 0x11
57 #define OMAP_MMC_REG_RSP2 0x12
58 #define OMAP_MMC_REG_RSP3 0x13
59 #define OMAP_MMC_REG_RSP4 0x14
60 #define OMAP_MMC_REG_RSP5 0x15
61 #define OMAP_MMC_REG_RSP6 0x16
62 #define OMAP_MMC_REG_RSP7 0x17
63 #define OMAP_MMC_REG_IOSR 0x18
64 #define OMAP_MMC_REG_SYSC 0x19
65 #define OMAP_MMC_REG_SYSS 0x1a
67 #define OMAP_MMC_STAT_CARD_ERR (1 << 14)
68 #define OMAP_MMC_STAT_CARD_IRQ (1 << 13)
69 #define OMAP_MMC_STAT_OCR_BUSY (1 << 12)
70 #define OMAP_MMC_STAT_A_EMPTY (1 << 11)
71 #define OMAP_MMC_STAT_A_FULL (1 << 10)
72 #define OMAP_MMC_STAT_CMD_CRC (1 << 8)
73 #define OMAP_MMC_STAT_CMD_TOUT (1 << 7)
74 #define OMAP_MMC_STAT_DATA_CRC (1 << 6)
75 #define OMAP_MMC_STAT_DATA_TOUT (1 << 5)
76 #define OMAP_MMC_STAT_END_BUSY (1 << 4)
77 #define OMAP_MMC_STAT_END_OF_DATA (1 << 3)
78 #define OMAP_MMC_STAT_CARD_BUSY (1 << 2)
79 #define OMAP_MMC_STAT_END_OF_CMD (1 << 0)
81 #define OMAP_MMC_REG(host, reg) (OMAP_MMC_REG_##reg << (host)->reg_shift)
82 #define OMAP_MMC_READ(host, reg) __raw_readw((host)->virt_base + OMAP_MMC_REG(host, reg))
83 #define OMAP_MMC_WRITE(host, reg, val) __raw_writew((val), (host)->virt_base + OMAP_MMC_REG(host, reg))
88 #define OMAP_MMC_CMDTYPE_BC 0
89 #define OMAP_MMC_CMDTYPE_BCR 1
90 #define OMAP_MMC_CMDTYPE_AC 2
91 #define OMAP_MMC_CMDTYPE_ADTC 3
94 #define DRIVER_NAME "mmci-omap"
96 /* Specifies how often in millisecs to poll for card status changes
97 * when the cover switch is open */
98 #define OMAP_MMC_COVER_POLL_DELAY 500
100 struct mmc_omap_host
;
102 struct mmc_omap_slot
{
107 unsigned int fclk_freq
;
110 struct tasklet_struct cover_tasklet
;
111 struct timer_list cover_timer
;
114 struct mmc_request
*mrq
;
115 struct mmc_omap_host
*host
;
116 struct mmc_host
*mmc
;
117 struct omap_mmc_slot_data
*pdata
;
120 struct mmc_omap_host
{
123 struct mmc_request
* mrq
;
124 struct mmc_command
* cmd
;
125 struct mmc_data
* data
;
126 struct mmc_host
* mmc
;
128 unsigned char id
; /* 16xx chips have 2 MMC blocks */
131 struct resource
*mem_res
;
132 void __iomem
*virt_base
;
133 unsigned int phys_base
;
135 unsigned char bus_mode
;
136 unsigned char hw_bus_mode
;
137 unsigned int reg_shift
;
139 struct work_struct cmd_abort_work
;
141 struct timer_list cmd_abort_timer
;
143 struct work_struct slot_release_work
;
144 struct mmc_omap_slot
*next_slot
;
145 struct work_struct send_stop_work
;
146 struct mmc_data
*stop_data
;
151 u32 buffer_bytes_left
;
152 u32 total_bytes_left
;
155 unsigned brs_received
:1, dma_done
:1;
156 unsigned dma_is_read
:1;
157 unsigned dma_in_use
:1;
160 struct timer_list dma_timer
;
163 struct mmc_omap_slot
*slots
[OMAP_MMC_MAX_SLOTS
];
164 struct mmc_omap_slot
*current_slot
;
165 spinlock_t slot_lock
;
166 wait_queue_head_t slot_wq
;
169 struct timer_list clk_timer
;
170 spinlock_t clk_lock
; /* for changing enabled state */
171 unsigned int fclk_enabled
:1;
173 struct omap_mmc_platform_data
*pdata
;
176 static struct workqueue_struct
*mmc_omap_wq
;
178 static void mmc_omap_fclk_offdelay(struct mmc_omap_slot
*slot
)
180 unsigned long tick_ns
;
182 if (slot
!= NULL
&& slot
->host
->fclk_enabled
&& slot
->fclk_freq
> 0) {
183 tick_ns
= (1000000000 + slot
->fclk_freq
- 1) / slot
->fclk_freq
;
188 static void mmc_omap_fclk_enable(struct mmc_omap_host
*host
, unsigned int enable
)
192 spin_lock_irqsave(&host
->clk_lock
, flags
);
193 if (host
->fclk_enabled
!= enable
) {
194 host
->fclk_enabled
= enable
;
196 clk_enable(host
->fclk
);
198 clk_disable(host
->fclk
);
200 spin_unlock_irqrestore(&host
->clk_lock
, flags
);
203 static void mmc_omap_select_slot(struct mmc_omap_slot
*slot
, int claimed
)
205 struct mmc_omap_host
*host
= slot
->host
;
210 spin_lock_irqsave(&host
->slot_lock
, flags
);
211 while (host
->mmc
!= NULL
) {
212 spin_unlock_irqrestore(&host
->slot_lock
, flags
);
213 wait_event(host
->slot_wq
, host
->mmc
== NULL
);
214 spin_lock_irqsave(&host
->slot_lock
, flags
);
216 host
->mmc
= slot
->mmc
;
217 spin_unlock_irqrestore(&host
->slot_lock
, flags
);
219 del_timer(&host
->clk_timer
);
220 if (host
->current_slot
!= slot
|| !claimed
)
221 mmc_omap_fclk_offdelay(host
->current_slot
);
223 if (host
->current_slot
!= slot
) {
224 OMAP_MMC_WRITE(host
, CON
, slot
->saved_con
& 0xFC00);
225 if (host
->pdata
->switch_slot
!= NULL
)
226 host
->pdata
->switch_slot(mmc_dev(slot
->mmc
), slot
->id
);
227 host
->current_slot
= slot
;
231 mmc_omap_fclk_enable(host
, 1);
233 /* Doing the dummy read here seems to work around some bug
234 * at least in OMAP24xx silicon where the command would not
235 * start after writing the CMD register. Sigh. */
236 OMAP_MMC_READ(host
, CON
);
238 OMAP_MMC_WRITE(host
, CON
, slot
->saved_con
);
240 mmc_omap_fclk_enable(host
, 0);
243 static void mmc_omap_start_request(struct mmc_omap_host
*host
,
244 struct mmc_request
*req
);
246 static void mmc_omap_slot_release_work(struct work_struct
*work
)
248 struct mmc_omap_host
*host
= container_of(work
, struct mmc_omap_host
,
250 struct mmc_omap_slot
*next_slot
= host
->next_slot
;
251 struct mmc_request
*rq
;
253 host
->next_slot
= NULL
;
254 mmc_omap_select_slot(next_slot
, 1);
257 next_slot
->mrq
= NULL
;
258 mmc_omap_start_request(host
, rq
);
261 static void mmc_omap_release_slot(struct mmc_omap_slot
*slot
, int clk_enabled
)
263 struct mmc_omap_host
*host
= slot
->host
;
267 BUG_ON(slot
== NULL
|| host
->mmc
== NULL
);
270 /* Keeps clock running for at least 8 cycles on valid freq */
271 mod_timer(&host
->clk_timer
, jiffies
+ HZ
/10);
273 del_timer(&host
->clk_timer
);
274 mmc_omap_fclk_offdelay(slot
);
275 mmc_omap_fclk_enable(host
, 0);
278 spin_lock_irqsave(&host
->slot_lock
, flags
);
279 /* Check for any pending requests */
280 for (i
= 0; i
< host
->nr_slots
; i
++) {
281 struct mmc_omap_slot
*new_slot
;
283 if (host
->slots
[i
] == NULL
|| host
->slots
[i
]->mrq
== NULL
)
286 BUG_ON(host
->next_slot
!= NULL
);
287 new_slot
= host
->slots
[i
];
288 /* The current slot should not have a request in queue */
289 BUG_ON(new_slot
== host
->current_slot
);
291 host
->next_slot
= new_slot
;
292 host
->mmc
= new_slot
->mmc
;
293 spin_unlock_irqrestore(&host
->slot_lock
, flags
);
294 queue_work(mmc_omap_wq
, &host
->slot_release_work
);
299 wake_up(&host
->slot_wq
);
300 spin_unlock_irqrestore(&host
->slot_lock
, flags
);
304 int mmc_omap_cover_is_open(struct mmc_omap_slot
*slot
)
306 if (slot
->pdata
->get_cover_state
)
307 return slot
->pdata
->get_cover_state(mmc_dev(slot
->mmc
),
313 mmc_omap_show_cover_switch(struct device
*dev
, struct device_attribute
*attr
,
316 struct mmc_host
*mmc
= container_of(dev
, struct mmc_host
, class_dev
);
317 struct mmc_omap_slot
*slot
= mmc_priv(mmc
);
319 return sprintf(buf
, "%s\n", mmc_omap_cover_is_open(slot
) ? "open" :
323 static DEVICE_ATTR(cover_switch
, S_IRUGO
, mmc_omap_show_cover_switch
, NULL
);
326 mmc_omap_show_slot_name(struct device
*dev
, struct device_attribute
*attr
,
329 struct mmc_host
*mmc
= container_of(dev
, struct mmc_host
, class_dev
);
330 struct mmc_omap_slot
*slot
= mmc_priv(mmc
);
332 return sprintf(buf
, "%s\n", slot
->pdata
->name
);
335 static DEVICE_ATTR(slot_name
, S_IRUGO
, mmc_omap_show_slot_name
, NULL
);
338 mmc_omap_start_command(struct mmc_omap_host
*host
, struct mmc_command
*cmd
)
349 /* Our hardware needs to know exact type */
350 switch (mmc_resp_type(cmd
)) {
355 /* resp 1, 1b, 6, 7 */
365 dev_err(mmc_dev(host
->mmc
), "Invalid response type: %04x\n", mmc_resp_type(cmd
));
369 if (mmc_cmd_type(cmd
) == MMC_CMD_ADTC
) {
370 cmdtype
= OMAP_MMC_CMDTYPE_ADTC
;
371 } else if (mmc_cmd_type(cmd
) == MMC_CMD_BC
) {
372 cmdtype
= OMAP_MMC_CMDTYPE_BC
;
373 } else if (mmc_cmd_type(cmd
) == MMC_CMD_BCR
) {
374 cmdtype
= OMAP_MMC_CMDTYPE_BCR
;
376 cmdtype
= OMAP_MMC_CMDTYPE_AC
;
379 cmdreg
= cmd
->opcode
| (resptype
<< 8) | (cmdtype
<< 12);
381 if (host
->current_slot
->bus_mode
== MMC_BUSMODE_OPENDRAIN
)
384 if (cmd
->flags
& MMC_RSP_BUSY
)
387 if (host
->data
&& !(host
->data
->flags
& MMC_DATA_WRITE
))
390 mod_timer(&host
->cmd_abort_timer
, jiffies
+ HZ
/2);
392 OMAP_MMC_WRITE(host
, CTO
, 200);
393 OMAP_MMC_WRITE(host
, ARGL
, cmd
->arg
& 0xffff);
394 OMAP_MMC_WRITE(host
, ARGH
, cmd
->arg
>> 16);
395 OMAP_MMC_WRITE(host
, IE
,
396 OMAP_MMC_STAT_A_EMPTY
| OMAP_MMC_STAT_A_FULL
|
397 OMAP_MMC_STAT_CMD_CRC
| OMAP_MMC_STAT_CMD_TOUT
|
398 OMAP_MMC_STAT_DATA_CRC
| OMAP_MMC_STAT_DATA_TOUT
|
399 OMAP_MMC_STAT_END_OF_CMD
| OMAP_MMC_STAT_CARD_ERR
|
400 OMAP_MMC_STAT_END_OF_DATA
);
401 OMAP_MMC_WRITE(host
, CMD
, cmdreg
);
405 mmc_omap_release_dma(struct mmc_omap_host
*host
, struct mmc_data
*data
,
408 enum dma_data_direction dma_data_dir
;
410 BUG_ON(host
->dma_ch
< 0);
412 omap_stop_dma(host
->dma_ch
);
413 /* Release DMA channel lazily */
414 mod_timer(&host
->dma_timer
, jiffies
+ HZ
);
415 if (data
->flags
& MMC_DATA_WRITE
)
416 dma_data_dir
= DMA_TO_DEVICE
;
418 dma_data_dir
= DMA_FROM_DEVICE
;
419 dma_unmap_sg(mmc_dev(host
->mmc
), data
->sg
, host
->sg_len
,
423 static void mmc_omap_send_stop_work(struct work_struct
*work
)
425 struct mmc_omap_host
*host
= container_of(work
, struct mmc_omap_host
,
427 struct mmc_omap_slot
*slot
= host
->current_slot
;
428 struct mmc_data
*data
= host
->stop_data
;
429 unsigned long tick_ns
;
431 tick_ns
= (1000000000 + slot
->fclk_freq
- 1)/slot
->fclk_freq
;
434 mmc_omap_start_command(host
, data
->stop
);
438 mmc_omap_xfer_done(struct mmc_omap_host
*host
, struct mmc_data
*data
)
440 if (host
->dma_in_use
)
441 mmc_omap_release_dma(host
, data
, data
->error
);
446 /* NOTE: MMC layer will sometimes poll-wait CMD13 next, issuing
447 * dozens of requests until the card finishes writing data.
448 * It'd be cheaper to just wait till an EOFB interrupt arrives...
452 struct mmc_host
*mmc
;
456 mmc_omap_release_slot(host
->current_slot
, 1);
457 mmc_request_done(mmc
, data
->mrq
);
461 host
->stop_data
= data
;
462 queue_work(mmc_omap_wq
, &host
->send_stop_work
);
466 mmc_omap_send_abort(struct mmc_omap_host
*host
, int maxloops
)
468 struct mmc_omap_slot
*slot
= host
->current_slot
;
469 unsigned int restarts
, passes
, timeout
;
472 /* Sending abort takes 80 clocks. Have some extra and round up */
473 timeout
= (120*1000000 + slot
->fclk_freq
- 1)/slot
->fclk_freq
;
475 while (restarts
< maxloops
) {
476 OMAP_MMC_WRITE(host
, STAT
, 0xFFFF);
477 OMAP_MMC_WRITE(host
, CMD
, (3 << 12) | (1 << 7));
480 while (passes
< timeout
) {
481 stat
= OMAP_MMC_READ(host
, STAT
);
482 if (stat
& OMAP_MMC_STAT_END_OF_CMD
)
491 OMAP_MMC_WRITE(host
, STAT
, stat
);
495 mmc_omap_abort_xfer(struct mmc_omap_host
*host
, struct mmc_data
*data
)
497 if (host
->dma_in_use
)
498 mmc_omap_release_dma(host
, data
, 1);
503 mmc_omap_send_abort(host
, 10000);
507 mmc_omap_end_of_data(struct mmc_omap_host
*host
, struct mmc_data
*data
)
512 if (!host
->dma_in_use
) {
513 mmc_omap_xfer_done(host
, data
);
517 spin_lock_irqsave(&host
->dma_lock
, flags
);
521 host
->brs_received
= 1;
522 spin_unlock_irqrestore(&host
->dma_lock
, flags
);
524 mmc_omap_xfer_done(host
, data
);
528 mmc_omap_dma_timer(unsigned long data
)
530 struct mmc_omap_host
*host
= (struct mmc_omap_host
*) data
;
532 BUG_ON(host
->dma_ch
< 0);
533 omap_free_dma(host
->dma_ch
);
538 mmc_omap_dma_done(struct mmc_omap_host
*host
, struct mmc_data
*data
)
544 spin_lock_irqsave(&host
->dma_lock
, flags
);
545 if (host
->brs_received
)
549 spin_unlock_irqrestore(&host
->dma_lock
, flags
);
551 mmc_omap_xfer_done(host
, data
);
555 mmc_omap_cmd_done(struct mmc_omap_host
*host
, struct mmc_command
*cmd
)
559 del_timer(&host
->cmd_abort_timer
);
561 if (cmd
->flags
& MMC_RSP_PRESENT
) {
562 if (cmd
->flags
& MMC_RSP_136
) {
563 /* response type 2 */
565 OMAP_MMC_READ(host
, RSP0
) |
566 (OMAP_MMC_READ(host
, RSP1
) << 16);
568 OMAP_MMC_READ(host
, RSP2
) |
569 (OMAP_MMC_READ(host
, RSP3
) << 16);
571 OMAP_MMC_READ(host
, RSP4
) |
572 (OMAP_MMC_READ(host
, RSP5
) << 16);
574 OMAP_MMC_READ(host
, RSP6
) |
575 (OMAP_MMC_READ(host
, RSP7
) << 16);
577 /* response types 1, 1b, 3, 4, 5, 6 */
579 OMAP_MMC_READ(host
, RSP6
) |
580 (OMAP_MMC_READ(host
, RSP7
) << 16);
584 if (host
->data
== NULL
|| cmd
->error
) {
585 struct mmc_host
*mmc
;
587 if (host
->data
!= NULL
)
588 mmc_omap_abort_xfer(host
, host
->data
);
591 mmc_omap_release_slot(host
->current_slot
, 1);
592 mmc_request_done(mmc
, cmd
->mrq
);
597 * Abort stuck command. Can occur when card is removed while it is being
600 static void mmc_omap_abort_command(struct work_struct
*work
)
602 struct mmc_omap_host
*host
= container_of(work
, struct mmc_omap_host
,
606 dev_dbg(mmc_dev(host
->mmc
), "Aborting stuck command CMD%d\n",
609 if (host
->cmd
->error
== 0)
610 host
->cmd
->error
= -ETIMEDOUT
;
612 if (host
->data
== NULL
) {
613 struct mmc_command
*cmd
;
614 struct mmc_host
*mmc
;
618 mmc_omap_send_abort(host
, 10000);
622 mmc_omap_release_slot(host
->current_slot
, 1);
623 mmc_request_done(mmc
, cmd
->mrq
);
625 mmc_omap_cmd_done(host
, host
->cmd
);
628 enable_irq(host
->irq
);
632 mmc_omap_cmd_timer(unsigned long data
)
634 struct mmc_omap_host
*host
= (struct mmc_omap_host
*) data
;
637 spin_lock_irqsave(&host
->slot_lock
, flags
);
638 if (host
->cmd
!= NULL
&& !host
->abort
) {
639 OMAP_MMC_WRITE(host
, IE
, 0);
640 disable_irq(host
->irq
);
642 queue_work(mmc_omap_wq
, &host
->cmd_abort_work
);
644 spin_unlock_irqrestore(&host
->slot_lock
, flags
);
649 mmc_omap_sg_to_buf(struct mmc_omap_host
*host
)
651 struct scatterlist
*sg
;
653 sg
= host
->data
->sg
+ host
->sg_idx
;
654 host
->buffer_bytes_left
= sg
->length
;
655 host
->buffer
= sg_virt(sg
);
656 if (host
->buffer_bytes_left
> host
->total_bytes_left
)
657 host
->buffer_bytes_left
= host
->total_bytes_left
;
661 mmc_omap_clk_timer(unsigned long data
)
663 struct mmc_omap_host
*host
= (struct mmc_omap_host
*) data
;
665 mmc_omap_fclk_enable(host
, 0);
670 mmc_omap_xfer_data(struct mmc_omap_host
*host
, int write
)
674 if (host
->buffer_bytes_left
== 0) {
676 BUG_ON(host
->sg_idx
== host
->sg_len
);
677 mmc_omap_sg_to_buf(host
);
680 if (n
> host
->buffer_bytes_left
)
681 n
= host
->buffer_bytes_left
;
682 host
->buffer_bytes_left
-= n
;
683 host
->total_bytes_left
-= n
;
684 host
->data
->bytes_xfered
+= n
;
687 __raw_writesw(host
->virt_base
+ OMAP_MMC_REG(host
, DATA
), host
->buffer
, n
);
689 __raw_readsw(host
->virt_base
+ OMAP_MMC_REG(host
, DATA
), host
->buffer
, n
);
693 static inline void mmc_omap_report_irq(u16 status
)
695 static const char *mmc_omap_status_bits
[] = {
696 "EOC", "CD", "CB", "BRS", "EOFB", "DTO", "DCRC", "CTO",
697 "CCRC", "CRW", "AF", "AE", "OCRB", "CIRQ", "CERR"
701 for (i
= 0; i
< ARRAY_SIZE(mmc_omap_status_bits
); i
++)
702 if (status
& (1 << i
)) {
705 printk("%s", mmc_omap_status_bits
[i
]);
710 static irqreturn_t
mmc_omap_irq(int irq
, void *dev_id
)
712 struct mmc_omap_host
* host
= (struct mmc_omap_host
*)dev_id
;
716 int transfer_error
, cmd_error
;
718 if (host
->cmd
== NULL
&& host
->data
== NULL
) {
719 status
= OMAP_MMC_READ(host
, STAT
);
720 dev_info(mmc_dev(host
->slots
[0]->mmc
),
721 "Spurious IRQ 0x%04x\n", status
);
723 OMAP_MMC_WRITE(host
, STAT
, status
);
724 OMAP_MMC_WRITE(host
, IE
, 0);
734 while ((status
= OMAP_MMC_READ(host
, STAT
)) != 0) {
737 OMAP_MMC_WRITE(host
, STAT
, status
);
738 if (host
->cmd
!= NULL
)
739 cmd
= host
->cmd
->opcode
;
742 #ifdef CONFIG_MMC_DEBUG
743 dev_dbg(mmc_dev(host
->mmc
), "MMC IRQ %04x (CMD %d): ",
745 mmc_omap_report_irq(status
);
748 if (host
->total_bytes_left
) {
749 if ((status
& OMAP_MMC_STAT_A_FULL
) ||
750 (status
& OMAP_MMC_STAT_END_OF_DATA
))
751 mmc_omap_xfer_data(host
, 0);
752 if (status
& OMAP_MMC_STAT_A_EMPTY
)
753 mmc_omap_xfer_data(host
, 1);
756 if (status
& OMAP_MMC_STAT_END_OF_DATA
)
759 if (status
& OMAP_MMC_STAT_DATA_TOUT
) {
760 dev_dbg(mmc_dev(host
->mmc
), "data timeout (CMD%d)\n",
763 host
->data
->error
= -ETIMEDOUT
;
768 if (status
& OMAP_MMC_STAT_DATA_CRC
) {
770 host
->data
->error
= -EILSEQ
;
771 dev_dbg(mmc_dev(host
->mmc
),
772 "data CRC error, bytes left %d\n",
773 host
->total_bytes_left
);
776 dev_dbg(mmc_dev(host
->mmc
), "data CRC error\n");
780 if (status
& OMAP_MMC_STAT_CMD_TOUT
) {
781 /* Timeouts are routine with some commands */
783 struct mmc_omap_slot
*slot
=
786 !mmc_omap_cover_is_open(slot
))
787 dev_err(mmc_dev(host
->mmc
),
788 "command timeout (CMD%d)\n",
790 host
->cmd
->error
= -ETIMEDOUT
;
796 if (status
& OMAP_MMC_STAT_CMD_CRC
) {
798 dev_err(mmc_dev(host
->mmc
),
799 "command CRC error (CMD%d, arg 0x%08x)\n",
800 cmd
, host
->cmd
->arg
);
801 host
->cmd
->error
= -EILSEQ
;
805 dev_err(mmc_dev(host
->mmc
),
806 "command CRC error without cmd?\n");
809 if (status
& OMAP_MMC_STAT_CARD_ERR
) {
810 dev_dbg(mmc_dev(host
->mmc
),
811 "ignoring card status error (CMD%d)\n",
817 * NOTE: On 1610 the END_OF_CMD may come too early when
820 if ((status
& OMAP_MMC_STAT_END_OF_CMD
) &&
821 (!(status
& OMAP_MMC_STAT_A_EMPTY
))) {
826 if (cmd_error
&& host
->data
) {
827 del_timer(&host
->cmd_abort_timer
);
829 OMAP_MMC_WRITE(host
, IE
, 0);
830 disable_irq_nosync(host
->irq
);
831 queue_work(mmc_omap_wq
, &host
->cmd_abort_work
);
835 if (end_command
&& host
->cmd
)
836 mmc_omap_cmd_done(host
, host
->cmd
);
837 if (host
->data
!= NULL
) {
839 mmc_omap_xfer_done(host
, host
->data
);
840 else if (end_transfer
)
841 mmc_omap_end_of_data(host
, host
->data
);
847 void omap_mmc_notify_cover_event(struct device
*dev
, int num
, int is_closed
)
850 struct mmc_omap_host
*host
= dev_get_drvdata(dev
);
851 struct mmc_omap_slot
*slot
= host
->slots
[num
];
853 BUG_ON(num
>= host
->nr_slots
);
855 /* Other subsystems can call in here before we're initialised. */
856 if (host
->nr_slots
== 0 || !host
->slots
[num
])
859 cover_open
= mmc_omap_cover_is_open(slot
);
860 if (cover_open
!= slot
->cover_open
) {
861 slot
->cover_open
= cover_open
;
862 sysfs_notify(&slot
->mmc
->class_dev
.kobj
, NULL
, "cover_switch");
865 tasklet_hi_schedule(&slot
->cover_tasklet
);
868 static void mmc_omap_cover_timer(unsigned long arg
)
870 struct mmc_omap_slot
*slot
= (struct mmc_omap_slot
*) arg
;
871 tasklet_schedule(&slot
->cover_tasklet
);
874 static void mmc_omap_cover_handler(unsigned long param
)
876 struct mmc_omap_slot
*slot
= (struct mmc_omap_slot
*)param
;
877 int cover_open
= mmc_omap_cover_is_open(slot
);
879 mmc_detect_change(slot
->mmc
, 0);
884 * If no card is inserted, we postpone polling until
885 * the cover has been closed.
887 if (slot
->mmc
->card
== NULL
|| !mmc_card_present(slot
->mmc
->card
))
890 mod_timer(&slot
->cover_timer
,
891 jiffies
+ msecs_to_jiffies(OMAP_MMC_COVER_POLL_DELAY
));
894 /* Prepare to transfer the next segment of a scatterlist */
896 mmc_omap_prepare_dma(struct mmc_omap_host
*host
, struct mmc_data
*data
)
898 int dma_ch
= host
->dma_ch
;
899 unsigned long data_addr
;
902 struct scatterlist
*sg
= &data
->sg
[host
->sg_idx
];
907 data_addr
= host
->phys_base
+ OMAP_MMC_REG(host
, DATA
);
909 count
= sg_dma_len(sg
);
911 if ((data
->blocks
== 1) && (count
> data
->blksz
))
914 host
->dma_len
= count
;
916 /* FIFO is 16x2 bytes on 15xx, and 32x2 bytes on 16xx and 24xx.
917 * Use 16 or 32 word frames when the blocksize is at least that large.
918 * Blocksize is usually 512 bytes; but not for some SD reads.
920 if (cpu_is_omap15xx() && frame
> 32)
927 if (!(data
->flags
& MMC_DATA_WRITE
)) {
928 buf
= 0x800f | ((frame
- 1) << 8);
930 if (cpu_class_is_omap1()) {
931 src_port
= OMAP_DMA_PORT_TIPB
;
932 dst_port
= OMAP_DMA_PORT_EMIFF
;
934 if (cpu_is_omap24xx())
935 sync_dev
= OMAP24XX_DMA_MMC1_RX
;
937 omap_set_dma_src_params(dma_ch
, src_port
,
938 OMAP_DMA_AMODE_CONSTANT
,
940 omap_set_dma_dest_params(dma_ch
, dst_port
,
941 OMAP_DMA_AMODE_POST_INC
,
942 sg_dma_address(sg
), 0, 0);
943 omap_set_dma_dest_data_pack(dma_ch
, 1);
944 omap_set_dma_dest_burst_mode(dma_ch
, OMAP_DMA_DATA_BURST_4
);
946 buf
= 0x0f80 | ((frame
- 1) << 0);
948 if (cpu_class_is_omap1()) {
949 src_port
= OMAP_DMA_PORT_EMIFF
;
950 dst_port
= OMAP_DMA_PORT_TIPB
;
952 if (cpu_is_omap24xx())
953 sync_dev
= OMAP24XX_DMA_MMC1_TX
;
955 omap_set_dma_dest_params(dma_ch
, dst_port
,
956 OMAP_DMA_AMODE_CONSTANT
,
958 omap_set_dma_src_params(dma_ch
, src_port
,
959 OMAP_DMA_AMODE_POST_INC
,
960 sg_dma_address(sg
), 0, 0);
961 omap_set_dma_src_data_pack(dma_ch
, 1);
962 omap_set_dma_src_burst_mode(dma_ch
, OMAP_DMA_DATA_BURST_4
);
965 /* Max limit for DMA frame count is 0xffff */
966 BUG_ON(count
> 0xffff);
968 OMAP_MMC_WRITE(host
, BUF
, buf
);
969 omap_set_dma_transfer_params(dma_ch
, OMAP_DMA_DATA_TYPE_S16
,
970 frame
, count
, OMAP_DMA_SYNC_FRAME
,
974 /* A scatterlist segment completed */
975 static void mmc_omap_dma_cb(int lch
, u16 ch_status
, void *data
)
977 struct mmc_omap_host
*host
= (struct mmc_omap_host
*) data
;
978 struct mmc_data
*mmcdat
= host
->data
;
980 if (unlikely(host
->dma_ch
< 0)) {
981 dev_err(mmc_dev(host
->mmc
),
982 "DMA callback while DMA not enabled\n");
985 /* FIXME: We really should do something to _handle_ the errors */
986 if (ch_status
& OMAP1_DMA_TOUT_IRQ
) {
987 dev_err(mmc_dev(host
->mmc
),"DMA timeout\n");
990 if (ch_status
& OMAP_DMA_DROP_IRQ
) {
991 dev_err(mmc_dev(host
->mmc
), "DMA sync error\n");
994 if (!(ch_status
& OMAP_DMA_BLOCK_IRQ
)) {
997 mmcdat
->bytes_xfered
+= host
->dma_len
;
999 if (host
->sg_idx
< host
->sg_len
) {
1000 mmc_omap_prepare_dma(host
, host
->data
);
1001 omap_start_dma(host
->dma_ch
);
1003 mmc_omap_dma_done(host
, host
->data
);
1006 static int mmc_omap_get_dma_channel(struct mmc_omap_host
*host
, struct mmc_data
*data
)
1008 const char *dma_dev_name
;
1009 int sync_dev
, dma_ch
, is_read
, r
;
1011 is_read
= !(data
->flags
& MMC_DATA_WRITE
);
1012 del_timer_sync(&host
->dma_timer
);
1013 if (host
->dma_ch
>= 0) {
1014 if (is_read
== host
->dma_is_read
)
1016 omap_free_dma(host
->dma_ch
);
1021 if (host
->id
== 0) {
1022 sync_dev
= OMAP_DMA_MMC_RX
;
1023 dma_dev_name
= "MMC1 read";
1025 sync_dev
= OMAP_DMA_MMC2_RX
;
1026 dma_dev_name
= "MMC2 read";
1029 if (host
->id
== 0) {
1030 sync_dev
= OMAP_DMA_MMC_TX
;
1031 dma_dev_name
= "MMC1 write";
1033 sync_dev
= OMAP_DMA_MMC2_TX
;
1034 dma_dev_name
= "MMC2 write";
1037 r
= omap_request_dma(sync_dev
, dma_dev_name
, mmc_omap_dma_cb
,
1040 dev_dbg(mmc_dev(host
->mmc
), "omap_request_dma() failed with %d\n", r
);
1043 host
->dma_ch
= dma_ch
;
1044 host
->dma_is_read
= is_read
;
1049 static inline void set_cmd_timeout(struct mmc_omap_host
*host
, struct mmc_request
*req
)
1053 reg
= OMAP_MMC_READ(host
, SDIO
);
1055 OMAP_MMC_WRITE(host
, SDIO
, reg
);
1056 /* Set maximum timeout */
1057 OMAP_MMC_WRITE(host
, CTO
, 0xff);
1060 static inline void set_data_timeout(struct mmc_omap_host
*host
, struct mmc_request
*req
)
1062 unsigned int timeout
, cycle_ns
;
1065 cycle_ns
= 1000000000 / host
->current_slot
->fclk_freq
;
1066 timeout
= req
->data
->timeout_ns
/ cycle_ns
;
1067 timeout
+= req
->data
->timeout_clks
;
1069 /* Check if we need to use timeout multiplier register */
1070 reg
= OMAP_MMC_READ(host
, SDIO
);
1071 if (timeout
> 0xffff) {
1076 OMAP_MMC_WRITE(host
, SDIO
, reg
);
1077 OMAP_MMC_WRITE(host
, DTO
, timeout
);
1081 mmc_omap_prepare_data(struct mmc_omap_host
*host
, struct mmc_request
*req
)
1083 struct mmc_data
*data
= req
->data
;
1084 int i
, use_dma
, block_size
;
1089 OMAP_MMC_WRITE(host
, BLEN
, 0);
1090 OMAP_MMC_WRITE(host
, NBLK
, 0);
1091 OMAP_MMC_WRITE(host
, BUF
, 0);
1092 host
->dma_in_use
= 0;
1093 set_cmd_timeout(host
, req
);
1097 block_size
= data
->blksz
;
1099 OMAP_MMC_WRITE(host
, NBLK
, data
->blocks
- 1);
1100 OMAP_MMC_WRITE(host
, BLEN
, block_size
- 1);
1101 set_data_timeout(host
, req
);
1103 /* cope with calling layer confusion; it issues "single
1104 * block" writes using multi-block scatterlists.
1106 sg_len
= (data
->blocks
== 1) ? 1 : data
->sg_len
;
1108 /* Only do DMA for entire blocks */
1109 use_dma
= host
->use_dma
;
1111 for (i
= 0; i
< sg_len
; i
++) {
1112 if ((data
->sg
[i
].length
% block_size
) != 0) {
1121 if (mmc_omap_get_dma_channel(host
, data
) == 0) {
1122 enum dma_data_direction dma_data_dir
;
1124 if (data
->flags
& MMC_DATA_WRITE
)
1125 dma_data_dir
= DMA_TO_DEVICE
;
1127 dma_data_dir
= DMA_FROM_DEVICE
;
1129 host
->sg_len
= dma_map_sg(mmc_dev(host
->mmc
), data
->sg
,
1130 sg_len
, dma_data_dir
);
1131 host
->total_bytes_left
= 0;
1132 mmc_omap_prepare_dma(host
, req
->data
);
1133 host
->brs_received
= 0;
1135 host
->dma_in_use
= 1;
1140 /* Revert to PIO? */
1142 OMAP_MMC_WRITE(host
, BUF
, 0x1f1f);
1143 host
->total_bytes_left
= data
->blocks
* block_size
;
1144 host
->sg_len
= sg_len
;
1145 mmc_omap_sg_to_buf(host
);
1146 host
->dma_in_use
= 0;
1150 static void mmc_omap_start_request(struct mmc_omap_host
*host
,
1151 struct mmc_request
*req
)
1153 BUG_ON(host
->mrq
!= NULL
);
1157 /* only touch fifo AFTER the controller readies it */
1158 mmc_omap_prepare_data(host
, req
);
1159 mmc_omap_start_command(host
, req
->cmd
);
1160 if (host
->dma_in_use
)
1161 omap_start_dma(host
->dma_ch
);
1164 static void mmc_omap_request(struct mmc_host
*mmc
, struct mmc_request
*req
)
1166 struct mmc_omap_slot
*slot
= mmc_priv(mmc
);
1167 struct mmc_omap_host
*host
= slot
->host
;
1168 unsigned long flags
;
1170 spin_lock_irqsave(&host
->slot_lock
, flags
);
1171 if (host
->mmc
!= NULL
) {
1172 BUG_ON(slot
->mrq
!= NULL
);
1174 spin_unlock_irqrestore(&host
->slot_lock
, flags
);
1178 spin_unlock_irqrestore(&host
->slot_lock
, flags
);
1179 mmc_omap_select_slot(slot
, 1);
1180 mmc_omap_start_request(host
, req
);
1183 static void mmc_omap_set_power(struct mmc_omap_slot
*slot
, int power_on
,
1186 struct mmc_omap_host
*host
;
1190 if (slot
->pdata
->set_power
!= NULL
)
1191 slot
->pdata
->set_power(mmc_dev(slot
->mmc
), slot
->id
, power_on
,
1194 if (cpu_is_omap24xx()) {
1198 w
= OMAP_MMC_READ(host
, CON
);
1199 OMAP_MMC_WRITE(host
, CON
, w
| (1 << 11));
1201 w
= OMAP_MMC_READ(host
, CON
);
1202 OMAP_MMC_WRITE(host
, CON
, w
& ~(1 << 11));
1207 static int mmc_omap_calc_divisor(struct mmc_host
*mmc
, struct mmc_ios
*ios
)
1209 struct mmc_omap_slot
*slot
= mmc_priv(mmc
);
1210 struct mmc_omap_host
*host
= slot
->host
;
1211 int func_clk_rate
= clk_get_rate(host
->fclk
);
1214 if (ios
->clock
== 0)
1217 dsor
= func_clk_rate
/ ios
->clock
;
1221 if (func_clk_rate
/ dsor
> ios
->clock
)
1227 slot
->fclk_freq
= func_clk_rate
/ dsor
;
1229 if (ios
->bus_width
== MMC_BUS_WIDTH_4
)
1235 static void mmc_omap_set_ios(struct mmc_host
*mmc
, struct mmc_ios
*ios
)
1237 struct mmc_omap_slot
*slot
= mmc_priv(mmc
);
1238 struct mmc_omap_host
*host
= slot
->host
;
1242 mmc_omap_select_slot(slot
, 0);
1244 dsor
= mmc_omap_calc_divisor(mmc
, ios
);
1246 if (ios
->vdd
!= slot
->vdd
)
1247 slot
->vdd
= ios
->vdd
;
1250 switch (ios
->power_mode
) {
1252 mmc_omap_set_power(slot
, 0, ios
->vdd
);
1255 /* Cannot touch dsor yet, just power up MMC */
1256 mmc_omap_set_power(slot
, 1, ios
->vdd
);
1259 mmc_omap_fclk_enable(host
, 1);
1265 if (slot
->bus_mode
!= ios
->bus_mode
) {
1266 if (slot
->pdata
->set_bus_mode
!= NULL
)
1267 slot
->pdata
->set_bus_mode(mmc_dev(mmc
), slot
->id
,
1269 slot
->bus_mode
= ios
->bus_mode
;
1272 /* On insanely high arm_per frequencies something sometimes
1273 * goes somehow out of sync, and the POW bit is not being set,
1274 * which results in the while loop below getting stuck.
1275 * Writing to the CON register twice seems to do the trick. */
1276 for (i
= 0; i
< 2; i
++)
1277 OMAP_MMC_WRITE(host
, CON
, dsor
);
1278 slot
->saved_con
= dsor
;
1279 if (ios
->power_mode
== MMC_POWER_ON
) {
1280 /* worst case at 400kHz, 80 cycles makes 200 microsecs */
1283 /* Send clock cycles, poll completion */
1284 OMAP_MMC_WRITE(host
, IE
, 0);
1285 OMAP_MMC_WRITE(host
, STAT
, 0xffff);
1286 OMAP_MMC_WRITE(host
, CMD
, 1 << 7);
1287 while (usecs
> 0 && (OMAP_MMC_READ(host
, STAT
) & 1) == 0) {
1291 OMAP_MMC_WRITE(host
, STAT
, 1);
1295 mmc_omap_release_slot(slot
, clk_enabled
);
1298 static const struct mmc_host_ops mmc_omap_ops
= {
1299 .request
= mmc_omap_request
,
1300 .set_ios
= mmc_omap_set_ios
,
1303 static int __init
mmc_omap_new_slot(struct mmc_omap_host
*host
, int id
)
1305 struct mmc_omap_slot
*slot
= NULL
;
1306 struct mmc_host
*mmc
;
1309 mmc
= mmc_alloc_host(sizeof(struct mmc_omap_slot
), host
->dev
);
1313 slot
= mmc_priv(mmc
);
1317 slot
->pdata
= &host
->pdata
->slots
[id
];
1319 host
->slots
[id
] = slot
;
1322 if (host
->pdata
->slots
[id
].wires
>= 4)
1323 mmc
->caps
|= MMC_CAP_4_BIT_DATA
;
1325 mmc
->ops
= &mmc_omap_ops
;
1326 mmc
->f_min
= 400000;
1328 if (cpu_class_is_omap2())
1329 mmc
->f_max
= 48000000;
1331 mmc
->f_max
= 24000000;
1332 if (host
->pdata
->max_freq
)
1333 mmc
->f_max
= min(host
->pdata
->max_freq
, mmc
->f_max
);
1334 mmc
->ocr_avail
= slot
->pdata
->ocr_mask
;
1336 /* Use scatterlist DMA to reduce per-transfer costs.
1337 * NOTE max_seg_size assumption that small blocks aren't
1338 * normally used (except e.g. for reading SD registers).
1341 mmc
->max_blk_size
= 2048; /* BLEN is 11 bits (+1) */
1342 mmc
->max_blk_count
= 2048; /* NBLK is 11 bits (+1) */
1343 mmc
->max_req_size
= mmc
->max_blk_size
* mmc
->max_blk_count
;
1344 mmc
->max_seg_size
= mmc
->max_req_size
;
1346 r
= mmc_add_host(mmc
);
1348 goto err_remove_host
;
1350 if (slot
->pdata
->name
!= NULL
) {
1351 r
= device_create_file(&mmc
->class_dev
,
1352 &dev_attr_slot_name
);
1354 goto err_remove_host
;
1357 if (slot
->pdata
->get_cover_state
!= NULL
) {
1358 r
= device_create_file(&mmc
->class_dev
,
1359 &dev_attr_cover_switch
);
1361 goto err_remove_slot_name
;
1363 setup_timer(&slot
->cover_timer
, mmc_omap_cover_timer
,
1364 (unsigned long)slot
);
1365 tasklet_init(&slot
->cover_tasklet
, mmc_omap_cover_handler
,
1366 (unsigned long)slot
);
1367 tasklet_schedule(&slot
->cover_tasklet
);
1372 err_remove_slot_name
:
1373 if (slot
->pdata
->name
!= NULL
)
1374 device_remove_file(&mmc
->class_dev
, &dev_attr_slot_name
);
1376 mmc_remove_host(mmc
);
1381 static void mmc_omap_remove_slot(struct mmc_omap_slot
*slot
)
1383 struct mmc_host
*mmc
= slot
->mmc
;
1385 if (slot
->pdata
->name
!= NULL
)
1386 device_remove_file(&mmc
->class_dev
, &dev_attr_slot_name
);
1387 if (slot
->pdata
->get_cover_state
!= NULL
)
1388 device_remove_file(&mmc
->class_dev
, &dev_attr_cover_switch
);
1390 tasklet_kill(&slot
->cover_tasklet
);
1391 del_timer_sync(&slot
->cover_timer
);
1392 flush_workqueue(mmc_omap_wq
);
1394 mmc_remove_host(mmc
);
1398 static int __init
mmc_omap_probe(struct platform_device
*pdev
)
1400 struct omap_mmc_platform_data
*pdata
= pdev
->dev
.platform_data
;
1401 struct mmc_omap_host
*host
= NULL
;
1402 struct resource
*res
;
1406 if (pdata
== NULL
) {
1407 dev_err(&pdev
->dev
, "platform data missing\n");
1410 if (pdata
->nr_slots
== 0) {
1411 dev_err(&pdev
->dev
, "no slots\n");
1415 res
= platform_get_resource(pdev
, IORESOURCE_MEM
, 0);
1416 irq
= platform_get_irq(pdev
, 0);
1417 if (res
== NULL
|| irq
< 0)
1420 res
= request_mem_region(res
->start
, resource_size(res
),
1425 host
= kzalloc(sizeof(struct mmc_omap_host
), GFP_KERNEL
);
1428 goto err_free_mem_region
;
1431 INIT_WORK(&host
->slot_release_work
, mmc_omap_slot_release_work
);
1432 INIT_WORK(&host
->send_stop_work
, mmc_omap_send_stop_work
);
1434 INIT_WORK(&host
->cmd_abort_work
, mmc_omap_abort_command
);
1435 setup_timer(&host
->cmd_abort_timer
, mmc_omap_cmd_timer
,
1436 (unsigned long) host
);
1438 spin_lock_init(&host
->clk_lock
);
1439 setup_timer(&host
->clk_timer
, mmc_omap_clk_timer
, (unsigned long) host
);
1441 spin_lock_init(&host
->dma_lock
);
1442 setup_timer(&host
->dma_timer
, mmc_omap_dma_timer
, (unsigned long) host
);
1443 spin_lock_init(&host
->slot_lock
);
1444 init_waitqueue_head(&host
->slot_wq
);
1446 host
->pdata
= pdata
;
1447 host
->dev
= &pdev
->dev
;
1448 platform_set_drvdata(pdev
, host
);
1450 host
->id
= pdev
->id
;
1451 host
->mem_res
= res
;
1455 host
->dev
->dma_mask
= &pdata
->dma_mask
;
1459 host
->phys_base
= host
->mem_res
->start
;
1460 host
->virt_base
= ioremap(res
->start
, resource_size(res
));
1461 if (!host
->virt_base
)
1464 host
->iclk
= clk_get(&pdev
->dev
, "ick");
1465 if (IS_ERR(host
->iclk
)) {
1466 ret
= PTR_ERR(host
->iclk
);
1467 goto err_free_mmc_host
;
1469 clk_enable(host
->iclk
);
1471 host
->fclk
= clk_get(&pdev
->dev
, "fck");
1472 if (IS_ERR(host
->fclk
)) {
1473 ret
= PTR_ERR(host
->fclk
);
1477 ret
= request_irq(host
->irq
, mmc_omap_irq
, 0, DRIVER_NAME
, host
);
1481 if (pdata
->init
!= NULL
) {
1482 ret
= pdata
->init(&pdev
->dev
);
1487 host
->nr_slots
= pdata
->nr_slots
;
1488 for (i
= 0; i
< pdata
->nr_slots
; i
++) {
1489 ret
= mmc_omap_new_slot(host
, i
);
1492 mmc_omap_remove_slot(host
->slots
[i
]);
1494 goto err_plat_cleanup
;
1498 host
->reg_shift
= (cpu_is_omap7xx() ? 1 : 2);
1504 pdata
->cleanup(&pdev
->dev
);
1506 free_irq(host
->irq
, host
);
1508 clk_put(host
->fclk
);
1510 clk_disable(host
->iclk
);
1511 clk_put(host
->iclk
);
1513 iounmap(host
->virt_base
);
1516 err_free_mem_region
:
1517 release_mem_region(res
->start
, resource_size(res
));
1521 static int mmc_omap_remove(struct platform_device
*pdev
)
1523 struct mmc_omap_host
*host
= platform_get_drvdata(pdev
);
1526 platform_set_drvdata(pdev
, NULL
);
1528 BUG_ON(host
== NULL
);
1530 for (i
= 0; i
< host
->nr_slots
; i
++)
1531 mmc_omap_remove_slot(host
->slots
[i
]);
1533 if (host
->pdata
->cleanup
)
1534 host
->pdata
->cleanup(&pdev
->dev
);
1536 mmc_omap_fclk_enable(host
, 0);
1537 free_irq(host
->irq
, host
);
1538 clk_put(host
->fclk
);
1539 clk_disable(host
->iclk
);
1540 clk_put(host
->iclk
);
1542 iounmap(host
->virt_base
);
1543 release_mem_region(pdev
->resource
[0].start
,
1544 pdev
->resource
[0].end
- pdev
->resource
[0].start
+ 1);
1552 static int mmc_omap_suspend(struct platform_device
*pdev
, pm_message_t mesg
)
1555 struct mmc_omap_host
*host
= platform_get_drvdata(pdev
);
1557 if (host
== NULL
|| host
->suspended
)
1560 for (i
= 0; i
< host
->nr_slots
; i
++) {
1561 struct mmc_omap_slot
*slot
;
1563 slot
= host
->slots
[i
];
1564 ret
= mmc_suspend_host(slot
->mmc
);
1567 slot
= host
->slots
[i
];
1568 mmc_resume_host(slot
->mmc
);
1573 host
->suspended
= 1;
1577 static int mmc_omap_resume(struct platform_device
*pdev
)
1580 struct mmc_omap_host
*host
= platform_get_drvdata(pdev
);
1582 if (host
== NULL
|| !host
->suspended
)
1585 for (i
= 0; i
< host
->nr_slots
; i
++) {
1586 struct mmc_omap_slot
*slot
;
1587 slot
= host
->slots
[i
];
1588 ret
= mmc_resume_host(slot
->mmc
);
1592 host
->suspended
= 0;
1597 #define mmc_omap_suspend NULL
1598 #define mmc_omap_resume NULL
1601 static struct platform_driver mmc_omap_driver
= {
1602 .remove
= mmc_omap_remove
,
1603 .suspend
= mmc_omap_suspend
,
1604 .resume
= mmc_omap_resume
,
1606 .name
= DRIVER_NAME
,
1607 .owner
= THIS_MODULE
,
1611 static int __init
mmc_omap_init(void)
1615 mmc_omap_wq
= alloc_workqueue("mmc_omap", 0, 0);
1619 ret
= platform_driver_probe(&mmc_omap_driver
, mmc_omap_probe
);
1621 destroy_workqueue(mmc_omap_wq
);
1625 static void __exit
mmc_omap_exit(void)
1627 platform_driver_unregister(&mmc_omap_driver
);
1628 destroy_workqueue(mmc_omap_wq
);
1631 module_init(mmc_omap_init
);
1632 module_exit(mmc_omap_exit
);
1634 MODULE_DESCRIPTION("OMAP Multimedia Card driver");
1635 MODULE_LICENSE("GPL");
1636 MODULE_ALIAS("platform:" DRIVER_NAME
);
1637 MODULE_AUTHOR("Juha Yrjölä");