2 * This file is subject to the terms and conditions of the GNU General Public
3 * License. See the file "COPYING" in the main directory of this archive
6 * Copyright (C) 1996 David S. Miller (davem@davemloft.net)
7 * Copyright (C) 1997, 1998, 1999, 2000, 2001, 2002 Ralf Baechle (ralf@gnu.org)
8 * Copyright (C) 1999, 2000 Silicon Graphics, Inc.
10 #include <linux/hardirq.h>
11 #include <linux/init.h>
12 #include <linux/highmem.h>
13 #include <linux/kernel.h>
14 #include <linux/linkage.h>
15 #include <linux/sched.h>
16 #include <linux/smp.h>
18 #include <linux/module.h>
19 #include <linux/bitops.h>
21 #include <asm/bcache.h>
22 #include <asm/bootinfo.h>
23 #include <asm/cache.h>
24 #include <asm/cacheops.h>
26 #include <asm/cpu-features.h>
29 #include <asm/pgtable.h>
30 #include <asm/r4kcache.h>
31 #include <asm/sections.h>
32 #include <asm/mmu_context.h>
34 #include <asm/cacheflush.h> /* for run_uncached() */
35 #include <asm/traps.h>
36 #include <asm/dma-coherence.h>
39 * Special Variant of smp_call_function for use by cache functions:
42 * o collapses to normal function call on UP kernels
43 * o collapses to normal function call on systems with a single shared
45 * o doesn't disable interrupts on the local CPU
47 static inline void r4k_on_each_cpu(void (*func
) (void *info
), void *info
)
51 #if !defined(CONFIG_MIPS_MT_SMP) && !defined(CONFIG_MIPS_MT_SMTC)
52 smp_call_function(func
, info
, 1);
58 #if defined(CONFIG_MIPS_CMP)
59 #define cpu_has_safe_index_cacheops 0
61 #define cpu_has_safe_index_cacheops 1
67 static unsigned long icache_size __read_mostly
;
68 static unsigned long dcache_size __read_mostly
;
69 static unsigned long scache_size __read_mostly
;
72 * Dummy cache handling routines for machines without boardcaches
74 static void cache_noop(void) {}
76 static struct bcache_ops no_sc_ops
= {
77 .bc_enable
= (void *)cache_noop
,
78 .bc_disable
= (void *)cache_noop
,
79 .bc_wback_inv
= (void *)cache_noop
,
80 .bc_inv
= (void *)cache_noop
83 struct bcache_ops
*bcops
= &no_sc_ops
;
85 #define cpu_is_r4600_v1_x() ((read_c0_prid() & 0xfffffff0) == 0x00002010)
86 #define cpu_is_r4600_v2_x() ((read_c0_prid() & 0xfffffff0) == 0x00002020)
88 #define R4600_HIT_CACHEOP_WAR_IMPL \
90 if (R4600_V2_HIT_CACHEOP_WAR && cpu_is_r4600_v2_x()) \
91 *(volatile unsigned long *)CKSEG1; \
92 if (R4600_V1_HIT_CACHEOP_WAR) \
93 __asm__ __volatile__("nop;nop;nop;nop"); \
96 static void (*r4k_blast_dcache_page
)(unsigned long addr
);
98 static inline void r4k_blast_dcache_page_dc32(unsigned long addr
)
100 R4600_HIT_CACHEOP_WAR_IMPL
;
101 blast_dcache32_page(addr
);
104 static inline void r4k_blast_dcache_page_dc64(unsigned long addr
)
106 R4600_HIT_CACHEOP_WAR_IMPL
;
107 blast_dcache64_page(addr
);
110 static void r4k_blast_dcache_page_setup(void)
112 unsigned long dc_lsize
= cpu_dcache_line_size();
115 r4k_blast_dcache_page
= (void *)cache_noop
;
116 else if (dc_lsize
== 16)
117 r4k_blast_dcache_page
= blast_dcache16_page
;
118 else if (dc_lsize
== 32)
119 r4k_blast_dcache_page
= r4k_blast_dcache_page_dc32
;
120 else if (dc_lsize
== 64)
121 r4k_blast_dcache_page
= r4k_blast_dcache_page_dc64
;
124 static void (* r4k_blast_dcache_page_indexed
)(unsigned long addr
);
126 static void r4k_blast_dcache_page_indexed_setup(void)
128 unsigned long dc_lsize
= cpu_dcache_line_size();
131 r4k_blast_dcache_page_indexed
= (void *)cache_noop
;
132 else if (dc_lsize
== 16)
133 r4k_blast_dcache_page_indexed
= blast_dcache16_page_indexed
;
134 else if (dc_lsize
== 32)
135 r4k_blast_dcache_page_indexed
= blast_dcache32_page_indexed
;
136 else if (dc_lsize
== 64)
137 r4k_blast_dcache_page_indexed
= blast_dcache64_page_indexed
;
140 void (* r4k_blast_dcache
)(void);
141 EXPORT_SYMBOL(r4k_blast_dcache
);
143 static void r4k_blast_dcache_setup(void)
145 unsigned long dc_lsize
= cpu_dcache_line_size();
148 r4k_blast_dcache
= (void *)cache_noop
;
149 else if (dc_lsize
== 16)
150 r4k_blast_dcache
= blast_dcache16
;
151 else if (dc_lsize
== 32)
152 r4k_blast_dcache
= blast_dcache32
;
153 else if (dc_lsize
== 64)
154 r4k_blast_dcache
= blast_dcache64
;
157 /* force code alignment (used for TX49XX_ICACHE_INDEX_INV_WAR) */
158 #define JUMP_TO_ALIGN(order) \
159 __asm__ __volatile__( \
161 ".align\t" #order "\n\t" \
164 #define CACHE32_UNROLL32_ALIGN JUMP_TO_ALIGN(10) /* 32 * 32 = 1024 */
165 #define CACHE32_UNROLL32_ALIGN2 JUMP_TO_ALIGN(11)
167 static inline void blast_r4600_v1_icache32(void)
171 local_irq_save(flags
);
173 local_irq_restore(flags
);
176 static inline void tx49_blast_icache32(void)
178 unsigned long start
= INDEX_BASE
;
179 unsigned long end
= start
+ current_cpu_data
.icache
.waysize
;
180 unsigned long ws_inc
= 1UL << current_cpu_data
.icache
.waybit
;
181 unsigned long ws_end
= current_cpu_data
.icache
.ways
<<
182 current_cpu_data
.icache
.waybit
;
183 unsigned long ws
, addr
;
185 CACHE32_UNROLL32_ALIGN2
;
186 /* I'm in even chunk. blast odd chunks */
187 for (ws
= 0; ws
< ws_end
; ws
+= ws_inc
)
188 for (addr
= start
+ 0x400; addr
< end
; addr
+= 0x400 * 2)
189 cache32_unroll32(addr
|ws
, Index_Invalidate_I
);
190 CACHE32_UNROLL32_ALIGN
;
191 /* I'm in odd chunk. blast even chunks */
192 for (ws
= 0; ws
< ws_end
; ws
+= ws_inc
)
193 for (addr
= start
; addr
< end
; addr
+= 0x400 * 2)
194 cache32_unroll32(addr
|ws
, Index_Invalidate_I
);
197 static inline void blast_icache32_r4600_v1_page_indexed(unsigned long page
)
201 local_irq_save(flags
);
202 blast_icache32_page_indexed(page
);
203 local_irq_restore(flags
);
206 static inline void tx49_blast_icache32_page_indexed(unsigned long page
)
208 unsigned long indexmask
= current_cpu_data
.icache
.waysize
- 1;
209 unsigned long start
= INDEX_BASE
+ (page
& indexmask
);
210 unsigned long end
= start
+ PAGE_SIZE
;
211 unsigned long ws_inc
= 1UL << current_cpu_data
.icache
.waybit
;
212 unsigned long ws_end
= current_cpu_data
.icache
.ways
<<
213 current_cpu_data
.icache
.waybit
;
214 unsigned long ws
, addr
;
216 CACHE32_UNROLL32_ALIGN2
;
217 /* I'm in even chunk. blast odd chunks */
218 for (ws
= 0; ws
< ws_end
; ws
+= ws_inc
)
219 for (addr
= start
+ 0x400; addr
< end
; addr
+= 0x400 * 2)
220 cache32_unroll32(addr
|ws
, Index_Invalidate_I
);
221 CACHE32_UNROLL32_ALIGN
;
222 /* I'm in odd chunk. blast even chunks */
223 for (ws
= 0; ws
< ws_end
; ws
+= ws_inc
)
224 for (addr
= start
; addr
< end
; addr
+= 0x400 * 2)
225 cache32_unroll32(addr
|ws
, Index_Invalidate_I
);
228 static void (* r4k_blast_icache_page
)(unsigned long addr
);
230 static void r4k_blast_icache_page_setup(void)
232 unsigned long ic_lsize
= cpu_icache_line_size();
235 r4k_blast_icache_page
= (void *)cache_noop
;
236 else if (ic_lsize
== 16)
237 r4k_blast_icache_page
= blast_icache16_page
;
238 else if (ic_lsize
== 32)
239 r4k_blast_icache_page
= blast_icache32_page
;
240 else if (ic_lsize
== 64)
241 r4k_blast_icache_page
= blast_icache64_page
;
245 static void (* r4k_blast_icache_page_indexed
)(unsigned long addr
);
247 static void r4k_blast_icache_page_indexed_setup(void)
249 unsigned long ic_lsize
= cpu_icache_line_size();
252 r4k_blast_icache_page_indexed
= (void *)cache_noop
;
253 else if (ic_lsize
== 16)
254 r4k_blast_icache_page_indexed
= blast_icache16_page_indexed
;
255 else if (ic_lsize
== 32) {
256 if (R4600_V1_INDEX_ICACHEOP_WAR
&& cpu_is_r4600_v1_x())
257 r4k_blast_icache_page_indexed
=
258 blast_icache32_r4600_v1_page_indexed
;
259 else if (TX49XX_ICACHE_INDEX_INV_WAR
)
260 r4k_blast_icache_page_indexed
=
261 tx49_blast_icache32_page_indexed
;
263 r4k_blast_icache_page_indexed
=
264 blast_icache32_page_indexed
;
265 } else if (ic_lsize
== 64)
266 r4k_blast_icache_page_indexed
= blast_icache64_page_indexed
;
269 void (* r4k_blast_icache
)(void);
270 EXPORT_SYMBOL(r4k_blast_icache
);
272 static void r4k_blast_icache_setup(void)
274 unsigned long ic_lsize
= cpu_icache_line_size();
277 r4k_blast_icache
= (void *)cache_noop
;
278 else if (ic_lsize
== 16)
279 r4k_blast_icache
= blast_icache16
;
280 else if (ic_lsize
== 32) {
281 if (R4600_V1_INDEX_ICACHEOP_WAR
&& cpu_is_r4600_v1_x())
282 r4k_blast_icache
= blast_r4600_v1_icache32
;
283 else if (TX49XX_ICACHE_INDEX_INV_WAR
)
284 r4k_blast_icache
= tx49_blast_icache32
;
286 r4k_blast_icache
= blast_icache32
;
287 } else if (ic_lsize
== 64)
288 r4k_blast_icache
= blast_icache64
;
291 static void (* r4k_blast_scache_page
)(unsigned long addr
);
293 static void r4k_blast_scache_page_setup(void)
295 unsigned long sc_lsize
= cpu_scache_line_size();
297 if (scache_size
== 0)
298 r4k_blast_scache_page
= (void *)cache_noop
;
299 else if (sc_lsize
== 16)
300 r4k_blast_scache_page
= blast_scache16_page
;
301 else if (sc_lsize
== 32)
302 r4k_blast_scache_page
= blast_scache32_page
;
303 else if (sc_lsize
== 64)
304 r4k_blast_scache_page
= blast_scache64_page
;
305 else if (sc_lsize
== 128)
306 r4k_blast_scache_page
= blast_scache128_page
;
309 static void (* r4k_blast_scache_page_indexed
)(unsigned long addr
);
311 static void r4k_blast_scache_page_indexed_setup(void)
313 unsigned long sc_lsize
= cpu_scache_line_size();
315 if (scache_size
== 0)
316 r4k_blast_scache_page_indexed
= (void *)cache_noop
;
317 else if (sc_lsize
== 16)
318 r4k_blast_scache_page_indexed
= blast_scache16_page_indexed
;
319 else if (sc_lsize
== 32)
320 r4k_blast_scache_page_indexed
= blast_scache32_page_indexed
;
321 else if (sc_lsize
== 64)
322 r4k_blast_scache_page_indexed
= blast_scache64_page_indexed
;
323 else if (sc_lsize
== 128)
324 r4k_blast_scache_page_indexed
= blast_scache128_page_indexed
;
327 static void (* r4k_blast_scache
)(void);
329 static void r4k_blast_scache_setup(void)
331 unsigned long sc_lsize
= cpu_scache_line_size();
333 if (scache_size
== 0)
334 r4k_blast_scache
= (void *)cache_noop
;
335 else if (sc_lsize
== 16)
336 r4k_blast_scache
= blast_scache16
;
337 else if (sc_lsize
== 32)
338 r4k_blast_scache
= blast_scache32
;
339 else if (sc_lsize
== 64)
340 r4k_blast_scache
= blast_scache64
;
341 else if (sc_lsize
== 128)
342 r4k_blast_scache
= blast_scache128
;
345 static inline void local_r4k___flush_cache_all(void * args
)
347 #if defined(CONFIG_CPU_LOONGSON2)
354 switch (current_cpu_type()) {
366 static void r4k___flush_cache_all(void)
368 r4k_on_each_cpu(local_r4k___flush_cache_all
, NULL
);
371 static inline int has_valid_asid(const struct mm_struct
*mm
)
373 #if defined(CONFIG_MIPS_MT_SMP) || defined(CONFIG_MIPS_MT_SMTC)
376 for_each_online_cpu(i
)
377 if (cpu_context(i
, mm
))
382 return cpu_context(smp_processor_id(), mm
);
386 static void r4k__flush_cache_vmap(void)
391 static void r4k__flush_cache_vunmap(void)
396 static inline void local_r4k_flush_cache_range(void * args
)
398 struct vm_area_struct
*vma
= args
;
399 int exec
= vma
->vm_flags
& VM_EXEC
;
401 if (!(has_valid_asid(vma
->vm_mm
)))
409 static void r4k_flush_cache_range(struct vm_area_struct
*vma
,
410 unsigned long start
, unsigned long end
)
412 int exec
= vma
->vm_flags
& VM_EXEC
;
414 if (cpu_has_dc_aliases
|| (exec
&& !cpu_has_ic_fills_f_dc
))
415 r4k_on_each_cpu(local_r4k_flush_cache_range
, vma
);
418 static inline void local_r4k_flush_cache_mm(void * args
)
420 struct mm_struct
*mm
= args
;
422 if (!has_valid_asid(mm
))
426 * Kludge alert. For obscure reasons R4000SC and R4400SC go nuts if we
427 * only flush the primary caches but R10000 and R12000 behave sane ...
428 * R4000SC and R4400SC indexed S-cache ops also invalidate primary
429 * caches, so we can bail out early.
431 if (current_cpu_type() == CPU_R4000SC
||
432 current_cpu_type() == CPU_R4000MC
||
433 current_cpu_type() == CPU_R4400SC
||
434 current_cpu_type() == CPU_R4400MC
) {
442 static void r4k_flush_cache_mm(struct mm_struct
*mm
)
444 if (!cpu_has_dc_aliases
)
447 r4k_on_each_cpu(local_r4k_flush_cache_mm
, mm
);
450 struct flush_cache_page_args
{
451 struct vm_area_struct
*vma
;
456 static inline void local_r4k_flush_cache_page(void *args
)
458 struct flush_cache_page_args
*fcp_args
= args
;
459 struct vm_area_struct
*vma
= fcp_args
->vma
;
460 unsigned long addr
= fcp_args
->addr
;
461 struct page
*page
= pfn_to_page(fcp_args
->pfn
);
462 int exec
= vma
->vm_flags
& VM_EXEC
;
463 struct mm_struct
*mm
= vma
->vm_mm
;
464 int map_coherent
= 0;
472 * If ownes no valid ASID yet, cannot possibly have gotten
473 * this page into the cache.
475 if (!has_valid_asid(mm
))
479 pgdp
= pgd_offset(mm
, addr
);
480 pudp
= pud_offset(pgdp
, addr
);
481 pmdp
= pmd_offset(pudp
, addr
);
482 ptep
= pte_offset(pmdp
, addr
);
485 * If the page isn't marked valid, the page cannot possibly be
488 if (!(pte_present(*ptep
)))
491 if ((mm
== current
->active_mm
) && (pte_val(*ptep
) & _PAGE_VALID
))
495 * Use kmap_coherent or kmap_atomic to do flushes for
496 * another ASID than the current one.
498 map_coherent
= (cpu_has_dc_aliases
&&
499 page_mapped(page
) && !Page_dcache_dirty(page
));
501 vaddr
= kmap_coherent(page
, addr
);
503 vaddr
= kmap_atomic(page
);
504 addr
= (unsigned long)vaddr
;
507 if (cpu_has_dc_aliases
|| (exec
&& !cpu_has_ic_fills_f_dc
)) {
508 r4k_blast_dcache_page(addr
);
509 if (exec
&& !cpu_icache_snoops_remote_store
)
510 r4k_blast_scache_page(addr
);
513 if (vaddr
&& cpu_has_vtag_icache
&& mm
== current
->active_mm
) {
514 int cpu
= smp_processor_id();
516 if (cpu_context(cpu
, mm
) != 0)
517 drop_mmu_context(mm
, cpu
);
519 r4k_blast_icache_page(addr
);
526 kunmap_atomic(vaddr
);
530 static void r4k_flush_cache_page(struct vm_area_struct
*vma
,
531 unsigned long addr
, unsigned long pfn
)
533 struct flush_cache_page_args args
;
539 r4k_on_each_cpu(local_r4k_flush_cache_page
, &args
);
542 static inline void local_r4k_flush_data_cache_page(void * addr
)
544 r4k_blast_dcache_page((unsigned long) addr
);
547 static void r4k_flush_data_cache_page(unsigned long addr
)
550 local_r4k_flush_data_cache_page((void *)addr
);
552 r4k_on_each_cpu(local_r4k_flush_data_cache_page
, (void *) addr
);
555 struct flush_icache_range_args
{
560 static inline void local_r4k_flush_icache_range(unsigned long start
, unsigned long end
)
562 if (!cpu_has_ic_fills_f_dc
) {
563 if (end
- start
>= dcache_size
) {
566 R4600_HIT_CACHEOP_WAR_IMPL
;
567 protected_blast_dcache_range(start
, end
);
571 if (end
- start
> icache_size
)
574 protected_blast_icache_range(start
, end
);
577 static inline void local_r4k_flush_icache_range_ipi(void *args
)
579 struct flush_icache_range_args
*fir_args
= args
;
580 unsigned long start
= fir_args
->start
;
581 unsigned long end
= fir_args
->end
;
583 local_r4k_flush_icache_range(start
, end
);
586 static void r4k_flush_icache_range(unsigned long start
, unsigned long end
)
588 struct flush_icache_range_args args
;
593 r4k_on_each_cpu(local_r4k_flush_icache_range_ipi
, &args
);
594 instruction_hazard();
597 #ifdef CONFIG_DMA_NONCOHERENT
599 static void r4k_dma_cache_wback_inv(unsigned long addr
, unsigned long size
)
601 /* Catch bad driver code */
604 if (cpu_has_inclusive_pcaches
) {
605 if (size
>= scache_size
)
608 blast_scache_range(addr
, addr
+ size
);
614 * Either no secondary cache or the available caches don't have the
615 * subset property so we have to flush the primary caches
618 if (cpu_has_safe_index_cacheops
&& size
>= dcache_size
) {
621 R4600_HIT_CACHEOP_WAR_IMPL
;
622 blast_dcache_range(addr
, addr
+ size
);
625 bc_wback_inv(addr
, size
);
629 static void r4k_dma_cache_inv(unsigned long addr
, unsigned long size
)
631 /* Catch bad driver code */
634 if (cpu_has_inclusive_pcaches
) {
635 if (size
>= scache_size
)
639 * There is no clearly documented alignment requirement
640 * for the cache instruction on MIPS processors and
641 * some processors, among them the RM5200 and RM7000
642 * QED processors will throw an address error for cache
643 * hit ops with insufficient alignment. Solved by
644 * aligning the address to cache line size.
646 blast_inv_scache_range(addr
, addr
+ size
);
652 if (cpu_has_safe_index_cacheops
&& size
>= dcache_size
) {
655 R4600_HIT_CACHEOP_WAR_IMPL
;
656 blast_inv_dcache_range(addr
, addr
+ size
);
662 #endif /* CONFIG_DMA_NONCOHERENT */
665 * While we're protected against bad userland addresses we don't care
666 * very much about what happens in that case. Usually a segmentation
667 * fault will dump the process later on anyway ...
669 static void local_r4k_flush_cache_sigtramp(void * arg
)
671 unsigned long ic_lsize
= cpu_icache_line_size();
672 unsigned long dc_lsize
= cpu_dcache_line_size();
673 unsigned long sc_lsize
= cpu_scache_line_size();
674 unsigned long addr
= (unsigned long) arg
;
676 R4600_HIT_CACHEOP_WAR_IMPL
;
678 protected_writeback_dcache_line(addr
& ~(dc_lsize
- 1));
679 if (!cpu_icache_snoops_remote_store
&& scache_size
)
680 protected_writeback_scache_line(addr
& ~(sc_lsize
- 1));
682 protected_flush_icache_line(addr
& ~(ic_lsize
- 1));
683 if (MIPS4K_ICACHE_REFILL_WAR
) {
684 __asm__
__volatile__ (
699 : "i" (Hit_Invalidate_I
));
701 if (MIPS_CACHE_SYNC_WAR
)
702 __asm__
__volatile__ ("sync");
705 static void r4k_flush_cache_sigtramp(unsigned long addr
)
707 r4k_on_each_cpu(local_r4k_flush_cache_sigtramp
, (void *) addr
);
710 static void r4k_flush_icache_all(void)
712 if (cpu_has_vtag_icache
)
716 struct flush_kernel_vmap_range_args
{
721 static inline void local_r4k_flush_kernel_vmap_range(void *args
)
723 struct flush_kernel_vmap_range_args
*vmra
= args
;
724 unsigned long vaddr
= vmra
->vaddr
;
725 int size
= vmra
->size
;
728 * Aliases only affect the primary caches so don't bother with
729 * S-caches or T-caches.
731 if (cpu_has_safe_index_cacheops
&& size
>= dcache_size
)
734 R4600_HIT_CACHEOP_WAR_IMPL
;
735 blast_dcache_range(vaddr
, vaddr
+ size
);
739 static void r4k_flush_kernel_vmap_range(unsigned long vaddr
, int size
)
741 struct flush_kernel_vmap_range_args args
;
743 args
.vaddr
= (unsigned long) vaddr
;
746 r4k_on_each_cpu(local_r4k_flush_kernel_vmap_range
, &args
);
749 static inline void rm7k_erratum31(void)
751 const unsigned long ic_lsize
= 32;
754 /* RM7000 erratum #31. The icache is screwed at startup. */
758 for (addr
= INDEX_BASE
; addr
<= INDEX_BASE
+ 4096; addr
+= ic_lsize
) {
759 __asm__
__volatile__ (
763 "cache\t%1, 0(%0)\n\t"
764 "cache\t%1, 0x1000(%0)\n\t"
765 "cache\t%1, 0x2000(%0)\n\t"
766 "cache\t%1, 0x3000(%0)\n\t"
767 "cache\t%2, 0(%0)\n\t"
768 "cache\t%2, 0x1000(%0)\n\t"
769 "cache\t%2, 0x2000(%0)\n\t"
770 "cache\t%2, 0x3000(%0)\n\t"
771 "cache\t%1, 0(%0)\n\t"
772 "cache\t%1, 0x1000(%0)\n\t"
773 "cache\t%1, 0x2000(%0)\n\t"
774 "cache\t%1, 0x3000(%0)\n\t"
777 : "r" (addr
), "i" (Index_Store_Tag_I
), "i" (Fill
));
781 static inline void alias_74k_erratum(struct cpuinfo_mips
*c
)
784 * Early versions of the 74K do not update the cache tags on a
785 * vtag miss/ptag hit which can occur in the case of KSEG0/KUSEG
786 * aliases. In this case it is better to treat the cache as always
789 if ((c
->processor_id
& 0xff) <= PRID_REV_ENCODE_332(2, 4, 0))
790 c
->dcache
.flags
|= MIPS_CACHE_VTAG
;
791 if ((c
->processor_id
& 0xff) == PRID_REV_ENCODE_332(2, 4, 0))
792 write_c0_config6(read_c0_config6() | MIPS_CONF6_SYND
);
793 if (((c
->processor_id
& 0xff00) == PRID_IMP_1074K
) &&
794 ((c
->processor_id
& 0xff) <= PRID_REV_ENCODE_332(1, 1, 0))) {
795 c
->dcache
.flags
|= MIPS_CACHE_VTAG
;
796 write_c0_config6(read_c0_config6() | MIPS_CONF6_SYND
);
800 static char *way_string
[] = { NULL
, "direct mapped", "2-way",
801 "3-way", "4-way", "5-way", "6-way", "7-way", "8-way"
804 static void probe_pcache(void)
806 struct cpuinfo_mips
*c
= ¤t_cpu_data
;
807 unsigned int config
= read_c0_config();
808 unsigned int prid
= read_c0_prid();
809 unsigned long config1
;
812 switch (c
->cputype
) {
813 case CPU_R4600
: /* QED style two way caches? */
817 icache_size
= 1 << (12 + ((config
& CONF_IC
) >> 9));
818 c
->icache
.linesz
= 16 << ((config
& CONF_IB
) >> 5);
820 c
->icache
.waybit
= __ffs(icache_size
/2);
822 dcache_size
= 1 << (12 + ((config
& CONF_DC
) >> 6));
823 c
->dcache
.linesz
= 16 << ((config
& CONF_DB
) >> 4);
825 c
->dcache
.waybit
= __ffs(dcache_size
/2);
827 c
->options
|= MIPS_CPU_CACHE_CDEX_P
;
832 icache_size
= 1 << (12 + ((config
& CONF_IC
) >> 9));
833 c
->icache
.linesz
= 16 << ((config
& CONF_IB
) >> 5);
837 dcache_size
= 1 << (12 + ((config
& CONF_DC
) >> 6));
838 c
->dcache
.linesz
= 16 << ((config
& CONF_DB
) >> 4);
840 c
->dcache
.waybit
= 0;
842 c
->options
|= MIPS_CPU_CACHE_CDEX_P
| MIPS_CPU_PREFETCH
;
846 icache_size
= 1 << (12 + ((config
& CONF_IC
) >> 9));
847 c
->icache
.linesz
= 16 << ((config
& CONF_IB
) >> 5);
851 dcache_size
= 1 << (12 + ((config
& CONF_DC
) >> 6));
852 c
->dcache
.linesz
= 16 << ((config
& CONF_DB
) >> 4);
854 c
->dcache
.waybit
= 0;
856 c
->options
|= MIPS_CPU_CACHE_CDEX_P
;
857 c
->options
|= MIPS_CPU_PREFETCH
;
867 icache_size
= 1 << (12 + ((config
& CONF_IC
) >> 9));
868 c
->icache
.linesz
= 16 << ((config
& CONF_IB
) >> 5);
870 c
->icache
.waybit
= 0; /* doesn't matter */
872 dcache_size
= 1 << (12 + ((config
& CONF_DC
) >> 6));
873 c
->dcache
.linesz
= 16 << ((config
& CONF_DB
) >> 4);
875 c
->dcache
.waybit
= 0; /* does not matter */
877 c
->options
|= MIPS_CPU_CACHE_CDEX_P
;
883 icache_size
= 1 << (12 + ((config
& R10K_CONF_IC
) >> 29));
884 c
->icache
.linesz
= 64;
886 c
->icache
.waybit
= 0;
888 dcache_size
= 1 << (12 + ((config
& R10K_CONF_DC
) >> 26));
889 c
->dcache
.linesz
= 32;
891 c
->dcache
.waybit
= 0;
893 c
->options
|= MIPS_CPU_PREFETCH
;
897 write_c0_config(config
& ~VR41_CONF_P4K
);
899 /* Workaround for cache instruction bug of VR4131 */
900 if (c
->processor_id
== 0x0c80U
|| c
->processor_id
== 0x0c81U
||
901 c
->processor_id
== 0x0c82U
) {
902 config
|= 0x00400000U
;
903 if (c
->processor_id
== 0x0c80U
)
904 config
|= VR41_CONF_BP
;
905 write_c0_config(config
);
907 c
->options
|= MIPS_CPU_CACHE_CDEX_P
;
909 icache_size
= 1 << (10 + ((config
& CONF_IC
) >> 9));
910 c
->icache
.linesz
= 16 << ((config
& CONF_IB
) >> 5);
912 c
->icache
.waybit
= __ffs(icache_size
/2);
914 dcache_size
= 1 << (10 + ((config
& CONF_DC
) >> 6));
915 c
->dcache
.linesz
= 16 << ((config
& CONF_DB
) >> 4);
917 c
->dcache
.waybit
= __ffs(dcache_size
/2);
926 icache_size
= 1 << (10 + ((config
& CONF_IC
) >> 9));
927 c
->icache
.linesz
= 16 << ((config
& CONF_IB
) >> 5);
929 c
->icache
.waybit
= 0; /* doesn't matter */
931 dcache_size
= 1 << (10 + ((config
& CONF_DC
) >> 6));
932 c
->dcache
.linesz
= 16 << ((config
& CONF_DB
) >> 4);
934 c
->dcache
.waybit
= 0; /* does not matter */
936 c
->options
|= MIPS_CPU_CACHE_CDEX_P
;
942 icache_size
= 1 << (12 + ((config
& CONF_IC
) >> 9));
943 c
->icache
.linesz
= 16 << ((config
& CONF_IB
) >> 5);
945 c
->icache
.waybit
= __ffs(icache_size
/ c
->icache
.ways
);
947 dcache_size
= 1 << (12 + ((config
& CONF_DC
) >> 6));
948 c
->dcache
.linesz
= 16 << ((config
& CONF_DB
) >> 4);
950 c
->dcache
.waybit
= __ffs(dcache_size
/ c
->dcache
.ways
);
952 c
->options
|= MIPS_CPU_CACHE_CDEX_P
;
953 c
->options
|= MIPS_CPU_PREFETCH
;
957 icache_size
= 1 << (12 + ((config
& CONF_IC
) >> 9));
958 c
->icache
.linesz
= 16 << ((config
& CONF_IB
) >> 5);
963 c
->icache
.waybit
= 0;
965 dcache_size
= 1 << (12 + ((config
& CONF_DC
) >> 6));
966 c
->dcache
.linesz
= 16 << ((config
& CONF_DB
) >> 4);
971 c
->dcache
.waybit
= 0;
975 if (!(config
& MIPS_CONF_M
))
976 panic("Don't know how to probe P-caches on this cpu.");
979 * So we seem to be a MIPS32 or MIPS64 CPU
980 * So let's probe the I-cache ...
982 config1
= read_c0_config1();
984 if ((lsize
= ((config1
>> 19) & 7)))
985 c
->icache
.linesz
= 2 << lsize
;
987 c
->icache
.linesz
= lsize
;
988 c
->icache
.sets
= 32 << (((config1
>> 22) + 1) & 7);
989 c
->icache
.ways
= 1 + ((config1
>> 16) & 7);
991 icache_size
= c
->icache
.sets
*
994 c
->icache
.waybit
= __ffs(icache_size
/c
->icache
.ways
);
996 if (config
& 0x8) /* VI bit */
997 c
->icache
.flags
|= MIPS_CACHE_VTAG
;
1000 * Now probe the MIPS32 / MIPS64 data cache.
1002 c
->dcache
.flags
= 0;
1004 if ((lsize
= ((config1
>> 10) & 7)))
1005 c
->dcache
.linesz
= 2 << lsize
;
1007 c
->dcache
.linesz
= lsize
;
1008 c
->dcache
.sets
= 32 << (((config1
>> 13) + 1) & 7);
1009 c
->dcache
.ways
= 1 + ((config1
>> 7) & 7);
1011 dcache_size
= c
->dcache
.sets
*
1014 c
->dcache
.waybit
= __ffs(dcache_size
/c
->dcache
.ways
);
1016 c
->options
|= MIPS_CPU_PREFETCH
;
1021 * Processor configuration sanity check for the R4000SC erratum
1022 * #5. With page sizes larger than 32kB there is no possibility
1023 * to get a VCE exception anymore so we don't care about this
1024 * misconfiguration. The case is rather theoretical anyway;
1025 * presumably no vendor is shipping his hardware in the "bad"
1028 if ((prid
& 0xff00) == PRID_IMP_R4000
&& (prid
& 0xff) < 0x40 &&
1029 !(config
& CONF_SC
) && c
->icache
.linesz
!= 16 &&
1030 PAGE_SIZE
<= 0x8000)
1031 panic("Improper R4000SC processor configuration detected");
1033 /* compute a couple of other cache variables */
1034 c
->icache
.waysize
= icache_size
/ c
->icache
.ways
;
1035 c
->dcache
.waysize
= dcache_size
/ c
->dcache
.ways
;
1037 c
->icache
.sets
= c
->icache
.linesz
?
1038 icache_size
/ (c
->icache
.linesz
* c
->icache
.ways
) : 0;
1039 c
->dcache
.sets
= c
->dcache
.linesz
?
1040 dcache_size
/ (c
->dcache
.linesz
* c
->dcache
.ways
) : 0;
1043 * R10000 and R12000 P-caches are odd in a positive way. They're 32kB
1044 * 2-way virtually indexed so normally would suffer from aliases. So
1045 * normally they'd suffer from aliases but magic in the hardware deals
1046 * with that for us so we don't need to take care ourselves.
1048 switch (c
->cputype
) {
1054 c
->dcache
.flags
|= MIPS_CACHE_PINDEX
;
1068 if (c
->cputype
== CPU_74K
)
1069 alias_74k_erratum(c
);
1070 if ((read_c0_config7() & (1 << 16))) {
1071 /* effectively physically indexed dcache,
1072 thus no virtual aliases. */
1073 c
->dcache
.flags
|= MIPS_CACHE_PINDEX
;
1077 if (c
->dcache
.waysize
> PAGE_SIZE
)
1078 c
->dcache
.flags
|= MIPS_CACHE_ALIASES
;
1081 switch (c
->cputype
) {
1084 * Some older 20Kc chips doesn't have the 'VI' bit in
1085 * the config register.
1087 c
->icache
.flags
|= MIPS_CACHE_VTAG
;
1091 c
->icache
.flags
|= MIPS_CACHE_IC_F_DC
;
1095 #ifdef CONFIG_CPU_LOONGSON2
1097 * LOONGSON2 has 4 way icache, but when using indexed cache op,
1098 * one op will act on all 4 ways
1103 printk("Primary instruction cache %ldkB, %s, %s, linesize %d bytes.\n",
1105 c
->icache
.flags
& MIPS_CACHE_VTAG
? "VIVT" : "VIPT",
1106 way_string
[c
->icache
.ways
], c
->icache
.linesz
);
1108 printk("Primary data cache %ldkB, %s, %s, %s, linesize %d bytes\n",
1109 dcache_size
>> 10, way_string
[c
->dcache
.ways
],
1110 (c
->dcache
.flags
& MIPS_CACHE_PINDEX
) ? "PIPT" : "VIPT",
1111 (c
->dcache
.flags
& MIPS_CACHE_ALIASES
) ?
1112 "cache aliases" : "no aliases",
1117 * If you even _breathe_ on this function, look at the gcc output and make sure
1118 * it does not pop things on and off the stack for the cache sizing loop that
1119 * executes in KSEG1 space or else you will crash and burn badly. You have
1122 static int probe_scache(void)
1124 unsigned long flags
, addr
, begin
, end
, pow2
;
1125 unsigned int config
= read_c0_config();
1126 struct cpuinfo_mips
*c
= ¤t_cpu_data
;
1128 if (config
& CONF_SC
)
1131 begin
= (unsigned long) &_stext
;
1132 begin
&= ~((4 * 1024 * 1024) - 1);
1133 end
= begin
+ (4 * 1024 * 1024);
1136 * This is such a bitch, you'd think they would make it easy to do
1137 * this. Away you daemons of stupidity!
1139 local_irq_save(flags
);
1141 /* Fill each size-multiple cache line with a valid tag. */
1143 for (addr
= begin
; addr
< end
; addr
= (begin
+ pow2
)) {
1144 unsigned long *p
= (unsigned long *) addr
;
1145 __asm__
__volatile__("nop" : : "r" (*p
)); /* whee... */
1149 /* Load first line with zero (therefore invalid) tag. */
1152 __asm__
__volatile__("nop; nop; nop; nop;"); /* avoid the hazard */
1153 cache_op(Index_Store_Tag_I
, begin
);
1154 cache_op(Index_Store_Tag_D
, begin
);
1155 cache_op(Index_Store_Tag_SD
, begin
);
1157 /* Now search for the wrap around point. */
1158 pow2
= (128 * 1024);
1159 for (addr
= begin
+ (128 * 1024); addr
< end
; addr
= begin
+ pow2
) {
1160 cache_op(Index_Load_Tag_SD
, addr
);
1161 __asm__
__volatile__("nop; nop; nop; nop;"); /* hazard... */
1162 if (!read_c0_taglo())
1166 local_irq_restore(flags
);
1170 c
->scache
.linesz
= 16 << ((config
& R4K_CONF_SB
) >> 22);
1172 c
->dcache
.waybit
= 0; /* does not matter */
1177 #if defined(CONFIG_CPU_LOONGSON2)
1178 static void __init
loongson2_sc_init(void)
1180 struct cpuinfo_mips
*c
= ¤t_cpu_data
;
1182 scache_size
= 512*1024;
1183 c
->scache
.linesz
= 32;
1185 c
->scache
.waybit
= 0;
1186 c
->scache
.waysize
= scache_size
/ (c
->scache
.ways
);
1187 c
->scache
.sets
= scache_size
/ (c
->scache
.linesz
* c
->scache
.ways
);
1188 pr_info("Unified secondary cache %ldkB %s, linesize %d bytes.\n",
1189 scache_size
>> 10, way_string
[c
->scache
.ways
], c
->scache
.linesz
);
1191 c
->options
|= MIPS_CPU_INCLUSIVE_CACHES
;
1195 extern int r5k_sc_init(void);
1196 extern int rm7k_sc_init(void);
1197 extern int mips_sc_init(void);
1199 static void setup_scache(void)
1201 struct cpuinfo_mips
*c
= ¤t_cpu_data
;
1202 unsigned int config
= read_c0_config();
1206 * Do the probing thing on R4000SC and R4400SC processors. Other
1207 * processors don't have a S-cache that would be relevant to the
1208 * Linux memory management.
1210 switch (c
->cputype
) {
1215 sc_present
= run_uncached(probe_scache
);
1217 c
->options
|= MIPS_CPU_CACHE_CDEX_S
;
1223 scache_size
= 0x80000 << ((config
& R10K_CONF_SS
) >> 16);
1224 c
->scache
.linesz
= 64 << ((config
>> 13) & 1);
1226 c
->scache
.waybit
= 0;
1232 #ifdef CONFIG_R5000_CPU_SCACHE
1238 #ifdef CONFIG_RM7000_CPU_SCACHE
1243 #if defined(CONFIG_CPU_LOONGSON2)
1245 loongson2_sc_init();
1249 /* don't need to worry about L2, fully coherent */
1253 if (c
->isa_level
& (MIPS_CPU_ISA_M32R1
| MIPS_CPU_ISA_M32R2
|
1254 MIPS_CPU_ISA_M64R1
| MIPS_CPU_ISA_M64R2
)) {
1255 #ifdef CONFIG_MIPS_CPU_SCACHE
1256 if (mips_sc_init ()) {
1257 scache_size
= c
->scache
.ways
* c
->scache
.sets
* c
->scache
.linesz
;
1258 printk("MIPS secondary cache %ldkB, %s, linesize %d bytes.\n",
1260 way_string
[c
->scache
.ways
], c
->scache
.linesz
);
1263 if (!(c
->scache
.flags
& MIPS_CACHE_NOT_PRESENT
))
1264 panic("Dunno how to handle MIPS32 / MIPS64 second level cache");
1274 /* compute a couple of other cache variables */
1275 c
->scache
.waysize
= scache_size
/ c
->scache
.ways
;
1277 c
->scache
.sets
= scache_size
/ (c
->scache
.linesz
* c
->scache
.ways
);
1279 printk("Unified secondary cache %ldkB %s, linesize %d bytes.\n",
1280 scache_size
>> 10, way_string
[c
->scache
.ways
], c
->scache
.linesz
);
1282 c
->options
|= MIPS_CPU_INCLUSIVE_CACHES
;
1285 void au1x00_fixup_config_od(void)
1288 * c0_config.od (bit 19) was write only (and read as 0)
1289 * on the early revisions of Alchemy SOCs. It disables the bus
1290 * transaction overlapping and needs to be set to fix various errata.
1292 switch (read_c0_prid()) {
1293 case 0x00030100: /* Au1000 DA */
1294 case 0x00030201: /* Au1000 HA */
1295 case 0x00030202: /* Au1000 HB */
1296 case 0x01030200: /* Au1500 AB */
1298 * Au1100 errata actually keeps silence about this bit, so we set it
1299 * just in case for those revisions that require it to be set according
1300 * to the (now gone) cpu table.
1302 case 0x02030200: /* Au1100 AB */
1303 case 0x02030201: /* Au1100 BA */
1304 case 0x02030202: /* Au1100 BC */
1305 set_c0_config(1 << 19);
1310 /* CP0 hazard avoidance. */
1311 #define NXP_BARRIER() \
1312 __asm__ __volatile__( \
1313 ".set noreorder\n\t" \
1314 "nop; nop; nop; nop; nop; nop;\n\t" \
1317 static void nxp_pr4450_fixup_config(void)
1319 unsigned long config0
;
1321 config0
= read_c0_config();
1323 /* clear all three cache coherency fields */
1324 config0
&= ~(0x7 | (7 << 25) | (7 << 28));
1325 config0
|= (((_page_cachable_default
>> _CACHE_SHIFT
) << 0) |
1326 ((_page_cachable_default
>> _CACHE_SHIFT
) << 25) |
1327 ((_page_cachable_default
>> _CACHE_SHIFT
) << 28));
1328 write_c0_config(config0
);
1332 static int cca
= -1;
1334 static int __init
cca_setup(char *str
)
1336 get_option(&str
, &cca
);
1341 early_param("cca", cca_setup
);
1343 static void coherency_setup(void)
1345 if (cca
< 0 || cca
> 7)
1346 cca
= read_c0_config() & CONF_CM_CMASK
;
1347 _page_cachable_default
= cca
<< _CACHE_SHIFT
;
1349 pr_debug("Using cache attribute %d\n", cca
);
1350 change_c0_config(CONF_CM_CMASK
, cca
);
1353 * c0_status.cu=0 specifies that updates by the sc instruction use
1354 * the coherency mode specified by the TLB; 1 means cachable
1355 * coherent update on write will be used. Not all processors have
1356 * this bit and; some wire it to zero, others like Toshiba had the
1357 * silly idea of putting something else there ...
1359 switch (current_cpu_type()) {
1366 clear_c0_config(CONF_CU
);
1369 * We need to catch the early Alchemy SOCs with
1370 * the write-only co_config.od bit and set it back to one on:
1371 * Au1000 rev DA, HA, HB; Au1100 AB, BA, BC, Au1500 AB
1374 au1x00_fixup_config_od();
1377 case PRID_IMP_PR4450
:
1378 nxp_pr4450_fixup_config();
1383 static void r4k_cache_error_setup(void)
1385 extern char __weak except_vec2_generic
;
1386 extern char __weak except_vec2_sb1
;
1387 struct cpuinfo_mips
*c
= ¤t_cpu_data
;
1389 switch (c
->cputype
) {
1392 set_uncached_handler(0x100, &except_vec2_sb1
, 0x80);
1396 set_uncached_handler(0x100, &except_vec2_generic
, 0x80);
1401 void r4k_cache_init(void)
1403 extern void build_clear_page(void);
1404 extern void build_copy_page(void);
1405 struct cpuinfo_mips
*c
= ¤t_cpu_data
;
1410 r4k_blast_dcache_page_setup();
1411 r4k_blast_dcache_page_indexed_setup();
1412 r4k_blast_dcache_setup();
1413 r4k_blast_icache_page_setup();
1414 r4k_blast_icache_page_indexed_setup();
1415 r4k_blast_icache_setup();
1416 r4k_blast_scache_page_setup();
1417 r4k_blast_scache_page_indexed_setup();
1418 r4k_blast_scache_setup();
1421 * Some MIPS32 and MIPS64 processors have physically indexed caches.
1422 * This code supports virtually indexed processors and will be
1423 * unnecessarily inefficient on physically indexed processors.
1425 if (c
->dcache
.linesz
)
1426 shm_align_mask
= max_t( unsigned long,
1427 c
->dcache
.sets
* c
->dcache
.linesz
- 1,
1430 shm_align_mask
= PAGE_SIZE
-1;
1432 __flush_cache_vmap
= r4k__flush_cache_vmap
;
1433 __flush_cache_vunmap
= r4k__flush_cache_vunmap
;
1435 flush_cache_all
= cache_noop
;
1436 __flush_cache_all
= r4k___flush_cache_all
;
1437 flush_cache_mm
= r4k_flush_cache_mm
;
1438 flush_cache_page
= r4k_flush_cache_page
;
1439 flush_cache_range
= r4k_flush_cache_range
;
1441 __flush_kernel_vmap_range
= r4k_flush_kernel_vmap_range
;
1443 flush_cache_sigtramp
= r4k_flush_cache_sigtramp
;
1444 flush_icache_all
= r4k_flush_icache_all
;
1445 local_flush_data_cache_page
= local_r4k_flush_data_cache_page
;
1446 flush_data_cache_page
= r4k_flush_data_cache_page
;
1447 flush_icache_range
= r4k_flush_icache_range
;
1448 local_flush_icache_range
= local_r4k_flush_icache_range
;
1450 #if defined(CONFIG_DMA_NONCOHERENT)
1452 _dma_cache_wback_inv
= (void *)cache_noop
;
1453 _dma_cache_wback
= (void *)cache_noop
;
1454 _dma_cache_inv
= (void *)cache_noop
;
1456 _dma_cache_wback_inv
= r4k_dma_cache_wback_inv
;
1457 _dma_cache_wback
= r4k_dma_cache_wback_inv
;
1458 _dma_cache_inv
= r4k_dma_cache_inv
;
1466 * We want to run CMP kernels on core with and without coherent
1467 * caches. Therefore, do not use CONFIG_MIPS_CMP to decide whether
1468 * or not to flush caches.
1470 local_r4k___flush_cache_all(NULL
);
1473 board_cache_error_setup
= r4k_cache_error_setup
;