x86/amd-iommu: Add per IOMMU reference counting
[linux/fpc-iii.git] / drivers / block / mg_disk.c
blobe0339aaa18154303316c8405d72c08af4b036ffd
1 /*
2 * drivers/block/mg_disk.c
4 * Support for the mGine m[g]flash IO mode.
5 * Based on legacy hd.c
7 * (c) 2008 mGine Co.,LTD
8 * (c) 2008 unsik Kim <donari75@gmail.com>
10 * This program is free software; you can redistribute it and/or modify
11 * it under the terms of the GNU General Public License version 2 as
12 * published by the Free Software Foundation.
15 #include <linux/kernel.h>
16 #include <linux/module.h>
17 #include <linux/fs.h>
18 #include <linux/blkdev.h>
19 #include <linux/hdreg.h>
20 #include <linux/ata.h>
21 #include <linux/interrupt.h>
22 #include <linux/delay.h>
23 #include <linux/platform_device.h>
24 #include <linux/gpio.h>
25 #include <linux/mg_disk.h>
27 #define MG_RES_SEC (CONFIG_MG_DISK_RES << 1)
29 /* name for block device */
30 #define MG_DISK_NAME "mgd"
32 #define MG_DISK_MAJ 0
33 #define MG_DISK_MAX_PART 16
34 #define MG_SECTOR_SIZE 512
35 #define MG_MAX_SECTS 256
37 /* Register offsets */
38 #define MG_BUFF_OFFSET 0x8000
39 #define MG_REG_OFFSET 0xC000
40 #define MG_REG_FEATURE (MG_REG_OFFSET + 2) /* write case */
41 #define MG_REG_ERROR (MG_REG_OFFSET + 2) /* read case */
42 #define MG_REG_SECT_CNT (MG_REG_OFFSET + 4)
43 #define MG_REG_SECT_NUM (MG_REG_OFFSET + 6)
44 #define MG_REG_CYL_LOW (MG_REG_OFFSET + 8)
45 #define MG_REG_CYL_HIGH (MG_REG_OFFSET + 0xA)
46 #define MG_REG_DRV_HEAD (MG_REG_OFFSET + 0xC)
47 #define MG_REG_COMMAND (MG_REG_OFFSET + 0xE) /* write case */
48 #define MG_REG_STATUS (MG_REG_OFFSET + 0xE) /* read case */
49 #define MG_REG_DRV_CTRL (MG_REG_OFFSET + 0x10)
50 #define MG_REG_BURST_CTRL (MG_REG_OFFSET + 0x12)
52 /* handy status */
53 #define MG_STAT_READY (ATA_DRDY | ATA_DSC)
54 #define MG_READY_OK(s) (((s) & (MG_STAT_READY | (ATA_BUSY | ATA_DF | \
55 ATA_ERR))) == MG_STAT_READY)
57 /* error code for others */
58 #define MG_ERR_NONE 0
59 #define MG_ERR_TIMEOUT 0x100
60 #define MG_ERR_INIT_STAT 0x101
61 #define MG_ERR_TRANSLATION 0x102
62 #define MG_ERR_CTRL_RST 0x103
63 #define MG_ERR_INV_STAT 0x104
64 #define MG_ERR_RSTOUT 0x105
66 #define MG_MAX_ERRORS 6 /* Max read/write errors */
68 /* command */
69 #define MG_CMD_RD 0x20
70 #define MG_CMD_WR 0x30
71 #define MG_CMD_SLEEP 0x99
72 #define MG_CMD_WAKEUP 0xC3
73 #define MG_CMD_ID 0xEC
74 #define MG_CMD_WR_CONF 0x3C
75 #define MG_CMD_RD_CONF 0x40
77 /* operation mode */
78 #define MG_OP_CASCADE (1 << 0)
79 #define MG_OP_CASCADE_SYNC_RD (1 << 1)
80 #define MG_OP_CASCADE_SYNC_WR (1 << 2)
81 #define MG_OP_INTERLEAVE (1 << 3)
83 /* synchronous */
84 #define MG_BURST_LAT_4 (3 << 4)
85 #define MG_BURST_LAT_5 (4 << 4)
86 #define MG_BURST_LAT_6 (5 << 4)
87 #define MG_BURST_LAT_7 (6 << 4)
88 #define MG_BURST_LAT_8 (7 << 4)
89 #define MG_BURST_LEN_4 (1 << 1)
90 #define MG_BURST_LEN_8 (2 << 1)
91 #define MG_BURST_LEN_16 (3 << 1)
92 #define MG_BURST_LEN_32 (4 << 1)
93 #define MG_BURST_LEN_CONT (0 << 1)
95 /* timeout value (unit: ms) */
96 #define MG_TMAX_CONF_TO_CMD 1
97 #define MG_TMAX_WAIT_RD_DRQ 10
98 #define MG_TMAX_WAIT_WR_DRQ 500
99 #define MG_TMAX_RST_TO_BUSY 10
100 #define MG_TMAX_HDRST_TO_RDY 500
101 #define MG_TMAX_SWRST_TO_RDY 500
102 #define MG_TMAX_RSTOUT 3000
104 #define MG_DEV_MASK (MG_BOOT_DEV | MG_STORAGE_DEV | MG_STORAGE_DEV_SKIP_RST)
106 /* main structure for mflash driver */
107 struct mg_host {
108 struct device *dev;
110 struct request_queue *breq;
111 struct request *req;
112 spinlock_t lock;
113 struct gendisk *gd;
115 struct timer_list timer;
116 void (*mg_do_intr) (struct mg_host *);
118 u16 id[ATA_ID_WORDS];
120 u16 cyls;
121 u16 heads;
122 u16 sectors;
123 u32 n_sectors;
124 u32 nres_sectors;
126 void __iomem *dev_base;
127 unsigned int irq;
128 unsigned int rst;
129 unsigned int rstout;
131 u32 major;
132 u32 error;
136 * Debugging macro and defines
138 #undef DO_MG_DEBUG
139 #ifdef DO_MG_DEBUG
140 # define MG_DBG(fmt, args...) \
141 printk(KERN_DEBUG "%s:%d "fmt, __func__, __LINE__, ##args)
142 #else /* CONFIG_MG_DEBUG */
143 # define MG_DBG(fmt, args...) do { } while (0)
144 #endif /* CONFIG_MG_DEBUG */
146 static void mg_request(struct request_queue *);
148 static bool mg_end_request(struct mg_host *host, int err, unsigned int nr_bytes)
150 if (__blk_end_request(host->req, err, nr_bytes))
151 return true;
153 host->req = NULL;
154 return false;
157 static bool mg_end_request_cur(struct mg_host *host, int err)
159 return mg_end_request(host, err, blk_rq_cur_bytes(host->req));
162 static void mg_dump_status(const char *msg, unsigned int stat,
163 struct mg_host *host)
165 char *name = MG_DISK_NAME;
167 if (host->req)
168 name = host->req->rq_disk->disk_name;
170 printk(KERN_ERR "%s: %s: status=0x%02x { ", name, msg, stat & 0xff);
171 if (stat & ATA_BUSY)
172 printk("Busy ");
173 if (stat & ATA_DRDY)
174 printk("DriveReady ");
175 if (stat & ATA_DF)
176 printk("WriteFault ");
177 if (stat & ATA_DSC)
178 printk("SeekComplete ");
179 if (stat & ATA_DRQ)
180 printk("DataRequest ");
181 if (stat & ATA_CORR)
182 printk("CorrectedError ");
183 if (stat & ATA_ERR)
184 printk("Error ");
185 printk("}\n");
186 if ((stat & ATA_ERR) == 0) {
187 host->error = 0;
188 } else {
189 host->error = inb((unsigned long)host->dev_base + MG_REG_ERROR);
190 printk(KERN_ERR "%s: %s: error=0x%02x { ", name, msg,
191 host->error & 0xff);
192 if (host->error & ATA_BBK)
193 printk("BadSector ");
194 if (host->error & ATA_UNC)
195 printk("UncorrectableError ");
196 if (host->error & ATA_IDNF)
197 printk("SectorIdNotFound ");
198 if (host->error & ATA_ABORTED)
199 printk("DriveStatusError ");
200 if (host->error & ATA_AMNF)
201 printk("AddrMarkNotFound ");
202 printk("}");
203 if (host->error & (ATA_BBK | ATA_UNC | ATA_IDNF | ATA_AMNF)) {
204 if (host->req)
205 printk(", sector=%u",
206 (unsigned int)blk_rq_pos(host->req));
208 printk("\n");
212 static unsigned int mg_wait(struct mg_host *host, u32 expect, u32 msec)
214 u8 status;
215 unsigned long expire, cur_jiffies;
216 struct mg_drv_data *prv_data = host->dev->platform_data;
218 host->error = MG_ERR_NONE;
219 expire = jiffies + msecs_to_jiffies(msec);
221 /* These 2 times dummy status read prevents reading invalid
222 * status. A very little time (3 times of mflash operating clk)
223 * is required for busy bit is set. Use dummy read instead of
224 * busy wait, because mflash's PLL is machine dependent.
226 if (prv_data->use_polling) {
227 status = inb((unsigned long)host->dev_base + MG_REG_STATUS);
228 status = inb((unsigned long)host->dev_base + MG_REG_STATUS);
231 status = inb((unsigned long)host->dev_base + MG_REG_STATUS);
233 do {
234 cur_jiffies = jiffies;
235 if (status & ATA_BUSY) {
236 if (expect == ATA_BUSY)
237 break;
238 } else {
239 /* Check the error condition! */
240 if (status & ATA_ERR) {
241 mg_dump_status("mg_wait", status, host);
242 break;
245 if (expect == MG_STAT_READY)
246 if (MG_READY_OK(status))
247 break;
249 if (expect == ATA_DRQ)
250 if (status & ATA_DRQ)
251 break;
253 if (!msec) {
254 mg_dump_status("not ready", status, host);
255 return MG_ERR_INV_STAT;
258 status = inb((unsigned long)host->dev_base + MG_REG_STATUS);
259 } while (time_before(cur_jiffies, expire));
261 if (time_after_eq(cur_jiffies, expire) && msec)
262 host->error = MG_ERR_TIMEOUT;
264 return host->error;
267 static unsigned int mg_wait_rstout(u32 rstout, u32 msec)
269 unsigned long expire;
271 expire = jiffies + msecs_to_jiffies(msec);
272 while (time_before(jiffies, expire)) {
273 if (gpio_get_value(rstout) == 1)
274 return MG_ERR_NONE;
275 msleep(10);
278 return MG_ERR_RSTOUT;
281 static void mg_unexpected_intr(struct mg_host *host)
283 u32 status = inb((unsigned long)host->dev_base + MG_REG_STATUS);
285 mg_dump_status("mg_unexpected_intr", status, host);
288 static irqreturn_t mg_irq(int irq, void *dev_id)
290 struct mg_host *host = dev_id;
291 void (*handler)(struct mg_host *) = host->mg_do_intr;
293 spin_lock(&host->lock);
295 host->mg_do_intr = NULL;
296 del_timer(&host->timer);
297 if (!handler)
298 handler = mg_unexpected_intr;
299 handler(host);
301 spin_unlock(&host->lock);
303 return IRQ_HANDLED;
306 /* local copy of ata_id_string() */
307 static void mg_id_string(const u16 *id, unsigned char *s,
308 unsigned int ofs, unsigned int len)
310 unsigned int c;
312 BUG_ON(len & 1);
314 while (len > 0) {
315 c = id[ofs] >> 8;
316 *s = c;
317 s++;
319 c = id[ofs] & 0xff;
320 *s = c;
321 s++;
323 ofs++;
324 len -= 2;
328 /* local copy of ata_id_c_string() */
329 static void mg_id_c_string(const u16 *id, unsigned char *s,
330 unsigned int ofs, unsigned int len)
332 unsigned char *p;
334 mg_id_string(id, s, ofs, len - 1);
336 p = s + strnlen(s, len - 1);
337 while (p > s && p[-1] == ' ')
338 p--;
339 *p = '\0';
342 static int mg_get_disk_id(struct mg_host *host)
344 u32 i;
345 s32 err;
346 const u16 *id = host->id;
347 struct mg_drv_data *prv_data = host->dev->platform_data;
348 char fwrev[ATA_ID_FW_REV_LEN + 1];
349 char model[ATA_ID_PROD_LEN + 1];
350 char serial[ATA_ID_SERNO_LEN + 1];
352 if (!prv_data->use_polling)
353 outb(ATA_NIEN, (unsigned long)host->dev_base + MG_REG_DRV_CTRL);
355 outb(MG_CMD_ID, (unsigned long)host->dev_base + MG_REG_COMMAND);
356 err = mg_wait(host, ATA_DRQ, MG_TMAX_WAIT_RD_DRQ);
357 if (err)
358 return err;
360 for (i = 0; i < (MG_SECTOR_SIZE >> 1); i++)
361 host->id[i] = le16_to_cpu(inw((unsigned long)host->dev_base +
362 MG_BUFF_OFFSET + i * 2));
364 outb(MG_CMD_RD_CONF, (unsigned long)host->dev_base + MG_REG_COMMAND);
365 err = mg_wait(host, MG_STAT_READY, MG_TMAX_CONF_TO_CMD);
366 if (err)
367 return err;
369 if ((id[ATA_ID_FIELD_VALID] & 1) == 0)
370 return MG_ERR_TRANSLATION;
372 host->n_sectors = ata_id_u32(id, ATA_ID_LBA_CAPACITY);
373 host->cyls = id[ATA_ID_CYLS];
374 host->heads = id[ATA_ID_HEADS];
375 host->sectors = id[ATA_ID_SECTORS];
377 if (MG_RES_SEC && host->heads && host->sectors) {
378 /* modify cyls, n_sectors */
379 host->cyls = (host->n_sectors - MG_RES_SEC) /
380 host->heads / host->sectors;
381 host->nres_sectors = host->n_sectors - host->cyls *
382 host->heads * host->sectors;
383 host->n_sectors -= host->nres_sectors;
386 mg_id_c_string(id, fwrev, ATA_ID_FW_REV, sizeof(fwrev));
387 mg_id_c_string(id, model, ATA_ID_PROD, sizeof(model));
388 mg_id_c_string(id, serial, ATA_ID_SERNO, sizeof(serial));
389 printk(KERN_INFO "mg_disk: model: %s\n", model);
390 printk(KERN_INFO "mg_disk: firm: %.8s\n", fwrev);
391 printk(KERN_INFO "mg_disk: serial: %s\n", serial);
392 printk(KERN_INFO "mg_disk: %d + reserved %d sectors\n",
393 host->n_sectors, host->nres_sectors);
395 if (!prv_data->use_polling)
396 outb(0, (unsigned long)host->dev_base + MG_REG_DRV_CTRL);
398 return err;
402 static int mg_disk_init(struct mg_host *host)
404 struct mg_drv_data *prv_data = host->dev->platform_data;
405 s32 err;
406 u8 init_status;
408 /* hdd rst low */
409 gpio_set_value(host->rst, 0);
410 err = mg_wait(host, ATA_BUSY, MG_TMAX_RST_TO_BUSY);
411 if (err)
412 return err;
414 /* hdd rst high */
415 gpio_set_value(host->rst, 1);
416 err = mg_wait(host, MG_STAT_READY, MG_TMAX_HDRST_TO_RDY);
417 if (err)
418 return err;
420 /* soft reset on */
421 outb(ATA_SRST | (prv_data->use_polling ? ATA_NIEN : 0),
422 (unsigned long)host->dev_base + MG_REG_DRV_CTRL);
423 err = mg_wait(host, ATA_BUSY, MG_TMAX_RST_TO_BUSY);
424 if (err)
425 return err;
427 /* soft reset off */
428 outb(prv_data->use_polling ? ATA_NIEN : 0,
429 (unsigned long)host->dev_base + MG_REG_DRV_CTRL);
430 err = mg_wait(host, MG_STAT_READY, MG_TMAX_SWRST_TO_RDY);
431 if (err)
432 return err;
434 init_status = inb((unsigned long)host->dev_base + MG_REG_STATUS) & 0xf;
436 if (init_status == 0xf)
437 return MG_ERR_INIT_STAT;
439 return err;
442 static void mg_bad_rw_intr(struct mg_host *host)
444 if (host->req)
445 if (++host->req->errors >= MG_MAX_ERRORS ||
446 host->error == MG_ERR_TIMEOUT)
447 mg_end_request_cur(host, -EIO);
450 static unsigned int mg_out(struct mg_host *host,
451 unsigned int sect_num,
452 unsigned int sect_cnt,
453 unsigned int cmd,
454 void (*intr_addr)(struct mg_host *))
456 struct mg_drv_data *prv_data = host->dev->platform_data;
458 if (mg_wait(host, MG_STAT_READY, MG_TMAX_CONF_TO_CMD))
459 return host->error;
461 if (!prv_data->use_polling) {
462 host->mg_do_intr = intr_addr;
463 mod_timer(&host->timer, jiffies + 3 * HZ);
465 if (MG_RES_SEC)
466 sect_num += MG_RES_SEC;
467 outb((u8)sect_cnt, (unsigned long)host->dev_base + MG_REG_SECT_CNT);
468 outb((u8)sect_num, (unsigned long)host->dev_base + MG_REG_SECT_NUM);
469 outb((u8)(sect_num >> 8), (unsigned long)host->dev_base +
470 MG_REG_CYL_LOW);
471 outb((u8)(sect_num >> 16), (unsigned long)host->dev_base +
472 MG_REG_CYL_HIGH);
473 outb((u8)((sect_num >> 24) | ATA_LBA | ATA_DEVICE_OBS),
474 (unsigned long)host->dev_base + MG_REG_DRV_HEAD);
475 outb(cmd, (unsigned long)host->dev_base + MG_REG_COMMAND);
476 return MG_ERR_NONE;
479 static void mg_read_one(struct mg_host *host, struct request *req)
481 u16 *buff = (u16 *)req->buffer;
482 u32 i;
484 for (i = 0; i < MG_SECTOR_SIZE >> 1; i++)
485 *buff++ = inw((unsigned long)host->dev_base + MG_BUFF_OFFSET +
486 (i << 1));
489 static void mg_read(struct request *req)
491 struct mg_host *host = req->rq_disk->private_data;
493 if (mg_out(host, blk_rq_pos(req), blk_rq_sectors(req),
494 MG_CMD_RD, NULL) != MG_ERR_NONE)
495 mg_bad_rw_intr(host);
497 MG_DBG("requested %d sects (from %ld), buffer=0x%p\n",
498 blk_rq_sectors(req), blk_rq_pos(req), req->buffer);
500 do {
501 if (mg_wait(host, ATA_DRQ,
502 MG_TMAX_WAIT_RD_DRQ) != MG_ERR_NONE) {
503 mg_bad_rw_intr(host);
504 return;
507 mg_read_one(host, req);
509 outb(MG_CMD_RD_CONF, (unsigned long)host->dev_base +
510 MG_REG_COMMAND);
511 } while (mg_end_request(host, 0, MG_SECTOR_SIZE));
514 static void mg_write_one(struct mg_host *host, struct request *req)
516 u16 *buff = (u16 *)req->buffer;
517 u32 i;
519 for (i = 0; i < MG_SECTOR_SIZE >> 1; i++)
520 outw(*buff++, (unsigned long)host->dev_base + MG_BUFF_OFFSET +
521 (i << 1));
524 static void mg_write(struct request *req)
526 struct mg_host *host = req->rq_disk->private_data;
527 unsigned int rem = blk_rq_sectors(req);
529 if (mg_out(host, blk_rq_pos(req), rem,
530 MG_CMD_WR, NULL) != MG_ERR_NONE) {
531 mg_bad_rw_intr(host);
532 return;
535 MG_DBG("requested %d sects (from %ld), buffer=0x%p\n",
536 rem, blk_rq_pos(req), req->buffer);
538 if (mg_wait(host, ATA_DRQ,
539 MG_TMAX_WAIT_WR_DRQ) != MG_ERR_NONE) {
540 mg_bad_rw_intr(host);
541 return;
544 do {
545 mg_write_one(host, req);
547 outb(MG_CMD_WR_CONF, (unsigned long)host->dev_base +
548 MG_REG_COMMAND);
550 rem--;
551 if (rem > 1 && mg_wait(host, ATA_DRQ,
552 MG_TMAX_WAIT_WR_DRQ) != MG_ERR_NONE) {
553 mg_bad_rw_intr(host);
554 return;
555 } else if (mg_wait(host, MG_STAT_READY,
556 MG_TMAX_WAIT_WR_DRQ) != MG_ERR_NONE) {
557 mg_bad_rw_intr(host);
558 return;
560 } while (mg_end_request(host, 0, MG_SECTOR_SIZE));
563 static void mg_read_intr(struct mg_host *host)
565 struct request *req = host->req;
566 u32 i;
568 /* check status */
569 do {
570 i = inb((unsigned long)host->dev_base + MG_REG_STATUS);
571 if (i & ATA_BUSY)
572 break;
573 if (!MG_READY_OK(i))
574 break;
575 if (i & ATA_DRQ)
576 goto ok_to_read;
577 } while (0);
578 mg_dump_status("mg_read_intr", i, host);
579 mg_bad_rw_intr(host);
580 mg_request(host->breq);
581 return;
583 ok_to_read:
584 mg_read_one(host, req);
586 MG_DBG("sector %ld, remaining=%ld, buffer=0x%p\n",
587 blk_rq_pos(req), blk_rq_sectors(req) - 1, req->buffer);
589 /* send read confirm */
590 outb(MG_CMD_RD_CONF, (unsigned long)host->dev_base + MG_REG_COMMAND);
592 if (mg_end_request(host, 0, MG_SECTOR_SIZE)) {
593 /* set handler if read remains */
594 host->mg_do_intr = mg_read_intr;
595 mod_timer(&host->timer, jiffies + 3 * HZ);
596 } else /* goto next request */
597 mg_request(host->breq);
600 static void mg_write_intr(struct mg_host *host)
602 struct request *req = host->req;
603 u32 i;
604 bool rem;
606 /* check status */
607 do {
608 i = inb((unsigned long)host->dev_base + MG_REG_STATUS);
609 if (i & ATA_BUSY)
610 break;
611 if (!MG_READY_OK(i))
612 break;
613 if ((blk_rq_sectors(req) <= 1) || (i & ATA_DRQ))
614 goto ok_to_write;
615 } while (0);
616 mg_dump_status("mg_write_intr", i, host);
617 mg_bad_rw_intr(host);
618 mg_request(host->breq);
619 return;
621 ok_to_write:
622 if ((rem = mg_end_request(host, 0, MG_SECTOR_SIZE))) {
623 /* write 1 sector and set handler if remains */
624 mg_write_one(host, req);
625 MG_DBG("sector %ld, remaining=%ld, buffer=0x%p\n",
626 blk_rq_pos(req), blk_rq_sectors(req), req->buffer);
627 host->mg_do_intr = mg_write_intr;
628 mod_timer(&host->timer, jiffies + 3 * HZ);
631 /* send write confirm */
632 outb(MG_CMD_WR_CONF, (unsigned long)host->dev_base + MG_REG_COMMAND);
634 if (!rem)
635 mg_request(host->breq);
638 void mg_times_out(unsigned long data)
640 struct mg_host *host = (struct mg_host *)data;
641 char *name;
643 spin_lock_irq(&host->lock);
645 if (!host->req)
646 goto out_unlock;
648 host->mg_do_intr = NULL;
650 name = host->req->rq_disk->disk_name;
651 printk(KERN_DEBUG "%s: timeout\n", name);
653 host->error = MG_ERR_TIMEOUT;
654 mg_bad_rw_intr(host);
656 out_unlock:
657 mg_request(host->breq);
658 spin_unlock_irq(&host->lock);
661 static void mg_request_poll(struct request_queue *q)
663 struct mg_host *host = q->queuedata;
665 while (1) {
666 if (!host->req) {
667 host->req = blk_fetch_request(q);
668 if (!host->req)
669 break;
672 if (unlikely(!blk_fs_request(host->req))) {
673 mg_end_request_cur(host, -EIO);
674 continue;
677 if (rq_data_dir(host->req) == READ)
678 mg_read(host->req);
679 else
680 mg_write(host->req);
684 static unsigned int mg_issue_req(struct request *req,
685 struct mg_host *host,
686 unsigned int sect_num,
687 unsigned int sect_cnt)
689 switch (rq_data_dir(req)) {
690 case READ:
691 if (mg_out(host, sect_num, sect_cnt, MG_CMD_RD, &mg_read_intr)
692 != MG_ERR_NONE) {
693 mg_bad_rw_intr(host);
694 return host->error;
696 break;
697 case WRITE:
698 /* TODO : handler */
699 outb(ATA_NIEN, (unsigned long)host->dev_base + MG_REG_DRV_CTRL);
700 if (mg_out(host, sect_num, sect_cnt, MG_CMD_WR, &mg_write_intr)
701 != MG_ERR_NONE) {
702 mg_bad_rw_intr(host);
703 return host->error;
705 del_timer(&host->timer);
706 mg_wait(host, ATA_DRQ, MG_TMAX_WAIT_WR_DRQ);
707 outb(0, (unsigned long)host->dev_base + MG_REG_DRV_CTRL);
708 if (host->error) {
709 mg_bad_rw_intr(host);
710 return host->error;
712 mg_write_one(host, req);
713 mod_timer(&host->timer, jiffies + 3 * HZ);
714 outb(MG_CMD_WR_CONF, (unsigned long)host->dev_base +
715 MG_REG_COMMAND);
716 break;
718 return MG_ERR_NONE;
721 /* This function also called from IRQ context */
722 static void mg_request(struct request_queue *q)
724 struct mg_host *host = q->queuedata;
725 struct request *req;
726 u32 sect_num, sect_cnt;
728 while (1) {
729 if (!host->req) {
730 host->req = blk_fetch_request(q);
731 if (!host->req)
732 break;
734 req = host->req;
736 /* check unwanted request call */
737 if (host->mg_do_intr)
738 return;
740 del_timer(&host->timer);
742 sect_num = blk_rq_pos(req);
743 /* deal whole segments */
744 sect_cnt = blk_rq_sectors(req);
746 /* sanity check */
747 if (sect_num >= get_capacity(req->rq_disk) ||
748 ((sect_num + sect_cnt) >
749 get_capacity(req->rq_disk))) {
750 printk(KERN_WARNING
751 "%s: bad access: sector=%d, count=%d\n",
752 req->rq_disk->disk_name,
753 sect_num, sect_cnt);
754 mg_end_request_cur(host, -EIO);
755 continue;
758 if (unlikely(!blk_fs_request(req))) {
759 mg_end_request_cur(host, -EIO);
760 continue;
763 if (!mg_issue_req(req, host, sect_num, sect_cnt))
764 return;
768 static int mg_getgeo(struct block_device *bdev, struct hd_geometry *geo)
770 struct mg_host *host = bdev->bd_disk->private_data;
772 geo->cylinders = (unsigned short)host->cyls;
773 geo->heads = (unsigned char)host->heads;
774 geo->sectors = (unsigned char)host->sectors;
775 return 0;
778 static const struct block_device_operations mg_disk_ops = {
779 .getgeo = mg_getgeo
782 static int mg_suspend(struct platform_device *plat_dev, pm_message_t state)
784 struct mg_drv_data *prv_data = plat_dev->dev.platform_data;
785 struct mg_host *host = prv_data->host;
787 if (mg_wait(host, MG_STAT_READY, MG_TMAX_CONF_TO_CMD))
788 return -EIO;
790 if (!prv_data->use_polling)
791 outb(ATA_NIEN, (unsigned long)host->dev_base + MG_REG_DRV_CTRL);
793 outb(MG_CMD_SLEEP, (unsigned long)host->dev_base + MG_REG_COMMAND);
794 /* wait until mflash deep sleep */
795 msleep(1);
797 if (mg_wait(host, MG_STAT_READY, MG_TMAX_CONF_TO_CMD)) {
798 if (!prv_data->use_polling)
799 outb(0, (unsigned long)host->dev_base + MG_REG_DRV_CTRL);
800 return -EIO;
803 return 0;
806 static int mg_resume(struct platform_device *plat_dev)
808 struct mg_drv_data *prv_data = plat_dev->dev.platform_data;
809 struct mg_host *host = prv_data->host;
811 if (mg_wait(host, MG_STAT_READY, MG_TMAX_CONF_TO_CMD))
812 return -EIO;
814 outb(MG_CMD_WAKEUP, (unsigned long)host->dev_base + MG_REG_COMMAND);
815 /* wait until mflash wakeup */
816 msleep(1);
818 if (mg_wait(host, MG_STAT_READY, MG_TMAX_CONF_TO_CMD))
819 return -EIO;
821 if (!prv_data->use_polling)
822 outb(0, (unsigned long)host->dev_base + MG_REG_DRV_CTRL);
824 return 0;
827 static int mg_probe(struct platform_device *plat_dev)
829 struct mg_host *host;
830 struct resource *rsc;
831 struct mg_drv_data *prv_data = plat_dev->dev.platform_data;
832 int err = 0;
834 if (!prv_data) {
835 printk(KERN_ERR "%s:%d fail (no driver_data)\n",
836 __func__, __LINE__);
837 err = -EINVAL;
838 goto probe_err;
841 /* alloc mg_host */
842 host = kzalloc(sizeof(struct mg_host), GFP_KERNEL);
843 if (!host) {
844 printk(KERN_ERR "%s:%d fail (no memory for mg_host)\n",
845 __func__, __LINE__);
846 err = -ENOMEM;
847 goto probe_err;
849 host->major = MG_DISK_MAJ;
851 /* link each other */
852 prv_data->host = host;
853 host->dev = &plat_dev->dev;
855 /* io remap */
856 rsc = platform_get_resource(plat_dev, IORESOURCE_MEM, 0);
857 if (!rsc) {
858 printk(KERN_ERR "%s:%d platform_get_resource fail\n",
859 __func__, __LINE__);
860 err = -EINVAL;
861 goto probe_err_2;
863 host->dev_base = ioremap(rsc->start , rsc->end + 1);
864 if (!host->dev_base) {
865 printk(KERN_ERR "%s:%d ioremap fail\n",
866 __func__, __LINE__);
867 err = -EIO;
868 goto probe_err_2;
870 MG_DBG("dev_base = 0x%x\n", (u32)host->dev_base);
872 /* get reset pin */
873 rsc = platform_get_resource_byname(plat_dev, IORESOURCE_IO,
874 MG_RST_PIN);
875 if (!rsc) {
876 printk(KERN_ERR "%s:%d get reset pin fail\n",
877 __func__, __LINE__);
878 err = -EIO;
879 goto probe_err_3;
881 host->rst = rsc->start;
883 /* init rst pin */
884 err = gpio_request(host->rst, MG_RST_PIN);
885 if (err)
886 goto probe_err_3;
887 gpio_direction_output(host->rst, 1);
889 /* reset out pin */
890 if (!(prv_data->dev_attr & MG_DEV_MASK))
891 goto probe_err_3a;
893 if (prv_data->dev_attr != MG_BOOT_DEV) {
894 rsc = platform_get_resource_byname(plat_dev, IORESOURCE_IO,
895 MG_RSTOUT_PIN);
896 if (!rsc) {
897 printk(KERN_ERR "%s:%d get reset-out pin fail\n",
898 __func__, __LINE__);
899 err = -EIO;
900 goto probe_err_3a;
902 host->rstout = rsc->start;
903 err = gpio_request(host->rstout, MG_RSTOUT_PIN);
904 if (err)
905 goto probe_err_3a;
906 gpio_direction_input(host->rstout);
909 /* disk reset */
910 if (prv_data->dev_attr == MG_STORAGE_DEV) {
911 /* If POR seq. not yet finised, wait */
912 err = mg_wait_rstout(host->rstout, MG_TMAX_RSTOUT);
913 if (err)
914 goto probe_err_3b;
915 err = mg_disk_init(host);
916 if (err) {
917 printk(KERN_ERR "%s:%d fail (err code : %d)\n",
918 __func__, __LINE__, err);
919 err = -EIO;
920 goto probe_err_3b;
924 /* get irq resource */
925 if (!prv_data->use_polling) {
926 host->irq = platform_get_irq(plat_dev, 0);
927 if (host->irq == -ENXIO) {
928 err = host->irq;
929 goto probe_err_3b;
931 err = request_irq(host->irq, mg_irq,
932 IRQF_DISABLED | IRQF_TRIGGER_RISING,
933 MG_DEV_NAME, host);
934 if (err) {
935 printk(KERN_ERR "%s:%d fail (request_irq err=%d)\n",
936 __func__, __LINE__, err);
937 goto probe_err_3b;
942 /* get disk id */
943 err = mg_get_disk_id(host);
944 if (err) {
945 printk(KERN_ERR "%s:%d fail (err code : %d)\n",
946 __func__, __LINE__, err);
947 err = -EIO;
948 goto probe_err_4;
951 err = register_blkdev(host->major, MG_DISK_NAME);
952 if (err < 0) {
953 printk(KERN_ERR "%s:%d register_blkdev fail (err code : %d)\n",
954 __func__, __LINE__, err);
955 goto probe_err_4;
957 if (!host->major)
958 host->major = err;
960 spin_lock_init(&host->lock);
962 if (prv_data->use_polling)
963 host->breq = blk_init_queue(mg_request_poll, &host->lock);
964 else
965 host->breq = blk_init_queue(mg_request, &host->lock);
967 if (!host->breq) {
968 err = -ENOMEM;
969 printk(KERN_ERR "%s:%d (blk_init_queue) fail\n",
970 __func__, __LINE__);
971 goto probe_err_5;
973 host->breq->queuedata = host;
975 /* mflash is random device, thanx for the noop */
976 elevator_exit(host->breq->elevator);
977 err = elevator_init(host->breq, "noop");
978 if (err) {
979 printk(KERN_ERR "%s:%d (elevator_init) fail\n",
980 __func__, __LINE__);
981 goto probe_err_6;
983 blk_queue_max_sectors(host->breq, MG_MAX_SECTS);
984 blk_queue_logical_block_size(host->breq, MG_SECTOR_SIZE);
986 init_timer(&host->timer);
987 host->timer.function = mg_times_out;
988 host->timer.data = (unsigned long)host;
990 host->gd = alloc_disk(MG_DISK_MAX_PART);
991 if (!host->gd) {
992 printk(KERN_ERR "%s:%d (alloc_disk) fail\n",
993 __func__, __LINE__);
994 err = -ENOMEM;
995 goto probe_err_7;
997 host->gd->major = host->major;
998 host->gd->first_minor = 0;
999 host->gd->fops = &mg_disk_ops;
1000 host->gd->queue = host->breq;
1001 host->gd->private_data = host;
1002 sprintf(host->gd->disk_name, MG_DISK_NAME"a");
1004 set_capacity(host->gd, host->n_sectors);
1006 add_disk(host->gd);
1008 return err;
1010 probe_err_7:
1011 del_timer_sync(&host->timer);
1012 probe_err_6:
1013 blk_cleanup_queue(host->breq);
1014 probe_err_5:
1015 unregister_blkdev(MG_DISK_MAJ, MG_DISK_NAME);
1016 probe_err_4:
1017 if (!prv_data->use_polling)
1018 free_irq(host->irq, host);
1019 probe_err_3b:
1020 gpio_free(host->rstout);
1021 probe_err_3a:
1022 gpio_free(host->rst);
1023 probe_err_3:
1024 iounmap(host->dev_base);
1025 probe_err_2:
1026 kfree(host);
1027 probe_err:
1028 return err;
1031 static int mg_remove(struct platform_device *plat_dev)
1033 struct mg_drv_data *prv_data = plat_dev->dev.platform_data;
1034 struct mg_host *host = prv_data->host;
1035 int err = 0;
1037 /* delete timer */
1038 del_timer_sync(&host->timer);
1040 /* remove disk */
1041 if (host->gd) {
1042 del_gendisk(host->gd);
1043 put_disk(host->gd);
1045 /* remove queue */
1046 if (host->breq)
1047 blk_cleanup_queue(host->breq);
1049 /* unregister blk device */
1050 unregister_blkdev(host->major, MG_DISK_NAME);
1052 /* free irq */
1053 if (!prv_data->use_polling)
1054 free_irq(host->irq, host);
1056 /* free reset-out pin */
1057 if (prv_data->dev_attr != MG_BOOT_DEV)
1058 gpio_free(host->rstout);
1060 /* free rst pin */
1061 if (host->rst)
1062 gpio_free(host->rst);
1064 /* unmap io */
1065 if (host->dev_base)
1066 iounmap(host->dev_base);
1068 /* free mg_host */
1069 kfree(host);
1071 return err;
1074 static struct platform_driver mg_disk_driver = {
1075 .probe = mg_probe,
1076 .remove = mg_remove,
1077 .suspend = mg_suspend,
1078 .resume = mg_resume,
1079 .driver = {
1080 .name = MG_DEV_NAME,
1081 .owner = THIS_MODULE,
1085 /****************************************************************************
1087 * Module stuff
1089 ****************************************************************************/
1091 static int __init mg_init(void)
1093 printk(KERN_INFO "mGine mflash driver, (c) 2008 mGine Co.\n");
1094 return platform_driver_register(&mg_disk_driver);
1097 static void __exit mg_exit(void)
1099 printk(KERN_INFO "mflash driver : bye bye\n");
1100 platform_driver_unregister(&mg_disk_driver);
1103 module_init(mg_init);
1104 module_exit(mg_exit);
1106 MODULE_LICENSE("GPL");
1107 MODULE_AUTHOR("unsik Kim <donari75@gmail.com>");
1108 MODULE_DESCRIPTION("mGine m[g]flash device driver");