2 * Driver for MMC and SSD cards for Cavium OCTEON SOCs.
4 * This file is subject to the terms and conditions of the GNU General Public
5 * License. See the file "COPYING" in the main directory of this archive
8 * Copyright (C) 2012-2017 Cavium Inc.
10 #include <linux/dma-mapping.h>
11 #include <linux/gpio/consumer.h>
12 #include <linux/interrupt.h>
13 #include <linux/mmc/mmc.h>
14 #include <linux/mmc/slot-gpio.h>
15 #include <linux/module.h>
16 #include <linux/of_platform.h>
17 #include <asm/octeon/octeon.h>
20 #define CVMX_MIO_BOOT_CTL CVMX_ADD_IO_SEG(0x00011800000000D0ull)
23 * The l2c* functions below are used for the EMMC-17978 workaround.
25 * Due to a bug in the design of the MMC bus hardware, the 2nd to last
26 * cache block of a DMA read must be locked into the L2 Cache.
27 * Otherwise, data corruption may occur.
29 static inline void *phys_to_ptr(u64 address
)
31 return (void *)(address
| (1ull << 63)); /* XKPHYS */
35 * Lock a single line into L2. The line is zeroed before locking
36 * to make sure no dram accesses are made.
38 static void l2c_lock_line(u64 addr
)
40 char *addr_ptr
= phys_to_ptr(addr
);
43 "cache 31, %[line]" /* Unlock the line */
44 ::[line
] "m" (*addr_ptr
));
47 /* Unlock a single line in the L2 cache. */
48 static void l2c_unlock_line(u64 addr
)
50 char *addr_ptr
= phys_to_ptr(addr
);
53 "cache 23, %[line]" /* Unlock the line */
54 ::[line
] "m" (*addr_ptr
));
57 /* Locks a memory region in the L2 cache. */
58 static void l2c_lock_mem_region(u64 start
, u64 len
)
62 /* Round start/end to cache line boundaries */
63 end
= ALIGN(start
+ len
- 1, CVMX_CACHE_LINE_SIZE
);
64 start
= ALIGN(start
, CVMX_CACHE_LINE_SIZE
);
66 while (start
<= end
) {
68 start
+= CVMX_CACHE_LINE_SIZE
;
73 /* Unlock a memory region in the L2 cache. */
74 static void l2c_unlock_mem_region(u64 start
, u64 len
)
78 /* Round start/end to cache line boundaries */
79 end
= ALIGN(start
+ len
- 1, CVMX_CACHE_LINE_SIZE
);
80 start
= ALIGN(start
, CVMX_CACHE_LINE_SIZE
);
82 while (start
<= end
) {
83 l2c_unlock_line(start
);
84 start
+= CVMX_CACHE_LINE_SIZE
;
88 static void octeon_mmc_acquire_bus(struct cvm_mmc_host
*host
)
90 if (!host
->has_ciu3
) {
91 down(&octeon_bootbus_sem
);
92 /* For CN70XX, switch the MMC controller onto the bus. */
93 if (OCTEON_IS_MODEL(OCTEON_CN70XX
))
94 writeq(0, (void __iomem
*)CVMX_MIO_BOOT_CTL
);
96 down(&host
->mmc_serializer
);
100 static void octeon_mmc_release_bus(struct cvm_mmc_host
*host
)
103 up(&octeon_bootbus_sem
);
105 up(&host
->mmc_serializer
);
108 static void octeon_mmc_int_enable(struct cvm_mmc_host
*host
, u64 val
)
110 writeq(val
, host
->base
+ MIO_EMM_INT(host
));
112 writeq(val
, host
->base
+ MIO_EMM_INT_EN(host
));
115 static void octeon_mmc_set_shared_power(struct cvm_mmc_host
*host
, int dir
)
118 if (!atomic_dec_return(&host
->shared_power_users
))
119 gpiod_set_value_cansleep(host
->global_pwr_gpiod
, 0);
121 if (atomic_inc_return(&host
->shared_power_users
) == 1)
122 gpiod_set_value_cansleep(host
->global_pwr_gpiod
, 1);
125 static void octeon_mmc_dmar_fixup(struct cvm_mmc_host
*host
,
126 struct mmc_command
*cmd
,
127 struct mmc_data
*data
,
130 if (cmd
->opcode
!= MMC_WRITE_MULTIPLE_BLOCK
)
132 if (data
->blksz
* data
->blocks
<= 1024)
135 host
->n_minus_one
= addr
+ (data
->blksz
* data
->blocks
) - 1024;
136 l2c_lock_mem_region(host
->n_minus_one
, 512);
139 static void octeon_mmc_dmar_fixup_done(struct cvm_mmc_host
*host
)
141 if (!host
->n_minus_one
)
143 l2c_unlock_mem_region(host
->n_minus_one
, 512);
144 host
->n_minus_one
= 0;
147 static int octeon_mmc_probe(struct platform_device
*pdev
)
149 struct device_node
*cn
, *node
= pdev
->dev
.of_node
;
150 struct cvm_mmc_host
*host
;
151 struct resource
*res
;
157 host
= devm_kzalloc(&pdev
->dev
, sizeof(*host
), GFP_KERNEL
);
161 spin_lock_init(&host
->irq_handler_lock
);
162 sema_init(&host
->mmc_serializer
, 1);
164 host
->dev
= &pdev
->dev
;
165 host
->acquire_bus
= octeon_mmc_acquire_bus
;
166 host
->release_bus
= octeon_mmc_release_bus
;
167 host
->int_enable
= octeon_mmc_int_enable
;
168 host
->set_shared_power
= octeon_mmc_set_shared_power
;
169 if (OCTEON_IS_MODEL(OCTEON_CN6XXX
) ||
170 OCTEON_IS_MODEL(OCTEON_CNF7XXX
)) {
171 host
->dmar_fixup
= octeon_mmc_dmar_fixup
;
172 host
->dmar_fixup_done
= octeon_mmc_dmar_fixup_done
;
175 host
->sys_freq
= octeon_get_io_clock_rate();
177 if (of_device_is_compatible(node
, "cavium,octeon-7890-mmc")) {
178 host
->big_dma_addr
= true;
179 host
->need_irq_handler_lock
= true;
180 host
->has_ciu3
= true;
183 * First seven are the EMM_INT bits 0..6, then two for
184 * the EMM_DMA_INT bits
186 for (i
= 0; i
< 9; i
++) {
187 mmc_irq
[i
] = platform_get_irq(pdev
, i
);
191 /* work around legacy u-boot device trees */
192 irq_set_irq_type(mmc_irq
[i
], IRQ_TYPE_EDGE_RISING
);
195 host
->big_dma_addr
= false;
196 host
->need_irq_handler_lock
= false;
197 host
->has_ciu3
= false;
198 /* First one is EMM second DMA */
199 for (i
= 0; i
< 2; i
++) {
200 mmc_irq
[i
] = platform_get_irq(pdev
, i
);
206 host
->last_slot
= -1;
208 res
= platform_get_resource(pdev
, IORESOURCE_MEM
, 0);
210 dev_err(&pdev
->dev
, "Platform resource[0] is missing\n");
213 base
= devm_ioremap_resource(&pdev
->dev
, res
);
215 return PTR_ERR(base
);
216 host
->base
= (void __iomem
*)base
;
219 res
= platform_get_resource(pdev
, IORESOURCE_MEM
, 1);
221 dev_err(&pdev
->dev
, "Platform resource[1] is missing\n");
224 base
= devm_ioremap_resource(&pdev
->dev
, res
);
226 return PTR_ERR(base
);
227 host
->dma_base
= (void __iomem
*)base
;
229 * To keep the register addresses shared we intentionaly use
230 * a negative offset here, first register used on Octeon therefore
231 * starts at 0x20 (MIO_EMM_DMA_CFG).
233 host
->reg_off_dma
= -0x20;
235 ret
= dma_set_mask(&pdev
->dev
, DMA_BIT_MASK(64));
240 * Clear out any pending interrupts that may be left over from
243 val
= readq(host
->base
+ MIO_EMM_INT(host
));
244 writeq(val
, host
->base
+ MIO_EMM_INT(host
));
246 if (host
->has_ciu3
) {
247 /* Only CMD_DONE, DMA_DONE, CMD_ERR, DMA_ERR */
248 for (i
= 1; i
<= 4; i
++) {
249 ret
= devm_request_irq(&pdev
->dev
, mmc_irq
[i
],
251 0, cvm_mmc_irq_names
[i
], host
);
253 dev_err(&pdev
->dev
, "Error: devm_request_irq %d\n",
259 ret
= devm_request_irq(&pdev
->dev
, mmc_irq
[0],
260 cvm_mmc_interrupt
, 0, KBUILD_MODNAME
,
263 dev_err(&pdev
->dev
, "Error: devm_request_irq %d\n",
269 host
->global_pwr_gpiod
= devm_gpiod_get_optional(&pdev
->dev
,
272 if (IS_ERR(host
->global_pwr_gpiod
)) {
273 dev_err(&pdev
->dev
, "Invalid power GPIO\n");
274 return PTR_ERR(host
->global_pwr_gpiod
);
277 platform_set_drvdata(pdev
, host
);
280 for_each_child_of_node(node
, cn
) {
282 of_platform_device_create(cn
, NULL
, &pdev
->dev
);
283 if (!host
->slot_pdev
[i
]) {
287 ret
= cvm_mmc_of_slot_probe(&host
->slot_pdev
[i
]->dev
, host
);
289 dev_err(&pdev
->dev
, "Error populating slots\n");
290 octeon_mmc_set_shared_power(host
, 0);
298 for (i
= 0; i
< CAVIUM_MAX_MMC
; i
++) {
300 cvm_mmc_of_slot_remove(host
->slot
[i
]);
301 if (host
->slot_pdev
[i
])
302 of_platform_device_destroy(&host
->slot_pdev
[i
]->dev
, NULL
);
307 static int octeon_mmc_remove(struct platform_device
*pdev
)
309 struct cvm_mmc_host
*host
= platform_get_drvdata(pdev
);
313 for (i
= 0; i
< CAVIUM_MAX_MMC
; i
++)
315 cvm_mmc_of_slot_remove(host
->slot
[i
]);
317 dma_cfg
= readq(host
->dma_base
+ MIO_EMM_DMA_CFG(host
));
318 dma_cfg
&= ~MIO_EMM_DMA_CFG_EN
;
319 writeq(dma_cfg
, host
->dma_base
+ MIO_EMM_DMA_CFG(host
));
321 octeon_mmc_set_shared_power(host
, 0);
325 static const struct of_device_id octeon_mmc_match
[] = {
327 .compatible
= "cavium,octeon-6130-mmc",
330 .compatible
= "cavium,octeon-7890-mmc",
334 MODULE_DEVICE_TABLE(of
, octeon_mmc_match
);
336 static struct platform_driver octeon_mmc_driver
= {
337 .probe
= octeon_mmc_probe
,
338 .remove
= octeon_mmc_remove
,
340 .name
= KBUILD_MODNAME
,
341 .of_match_table
= octeon_mmc_match
,
345 module_platform_driver(octeon_mmc_driver
);
347 MODULE_AUTHOR("Cavium Inc. <support@cavium.com>");
348 MODULE_DESCRIPTION("Low-level driver for Cavium OCTEON MMC/SSD card");
349 MODULE_LICENSE("GPL");