2 * This file is subject to the terms and conditions of the GNU General Public
3 * License. See the file "COPYING" in the main directory of this archive
6 * A small micro-assembler. It is intentionally kept simple, does only
7 * support a subset of instructions, and does not try to hide pipeline
8 * effects like branch delay slots.
10 * Copyright (C) 2004, 2005, 2006, 2008 Thiemo Seufer
11 * Copyright (C) 2005, 2007 Maciej W. Rozycki
12 * Copyright (C) 2006 Ralf Baechle (ralf@linux-mips.org)
15 #include <linux/kernel.h>
16 #include <linux/types.h>
17 #include <linux/init.h>
48 #define IMM_MASK 0xffff
50 #define JIMM_MASK 0x3ffffff
52 #define FUNC_MASK 0x3f
59 insn_addu
, insn_addiu
, insn_and
, insn_andi
, insn_beq
,
60 insn_beql
, insn_bgez
, insn_bgezl
, insn_bltz
, insn_bltzl
,
61 insn_bne
, insn_cache
, insn_daddu
, insn_daddiu
, insn_dmfc0
,
62 insn_dmtc0
, insn_dsll
, insn_dsll32
, insn_dsra
, insn_dsrl
,
63 insn_dsrl32
, insn_dsubu
, insn_eret
, insn_j
, insn_jal
, insn_jr
,
64 insn_ld
, insn_ll
, insn_lld
, insn_lui
, insn_lw
, insn_mfc0
,
65 insn_mtc0
, insn_ori
, insn_pref
, insn_rfe
, insn_sc
, insn_scd
,
66 insn_sd
, insn_sll
, insn_sra
, insn_srl
, insn_subu
, insn_sw
,
67 insn_tlbp
, insn_tlbwi
, insn_tlbwr
, insn_xor
, insn_xori
76 /* This macro sets the non-variable bits of an instruction. */
77 #define M(a, b, c, d, e, f) \
85 static struct insn insn_table
[] __cpuinitdata
= {
86 { insn_addiu
, M(addiu_op
, 0, 0, 0, 0, 0), RS
| RT
| SIMM
},
87 { insn_addu
, M(spec_op
, 0, 0, 0, 0, addu_op
), RS
| RT
| RD
},
88 { insn_and
, M(spec_op
, 0, 0, 0, 0, and_op
), RS
| RT
| RD
},
89 { insn_andi
, M(andi_op
, 0, 0, 0, 0, 0), RS
| RT
| UIMM
},
90 { insn_beq
, M(beq_op
, 0, 0, 0, 0, 0), RS
| RT
| BIMM
},
91 { insn_beql
, M(beql_op
, 0, 0, 0, 0, 0), RS
| RT
| BIMM
},
92 { insn_bgez
, M(bcond_op
, 0, bgez_op
, 0, 0, 0), RS
| BIMM
},
93 { insn_bgezl
, M(bcond_op
, 0, bgezl_op
, 0, 0, 0), RS
| BIMM
},
94 { insn_bltz
, M(bcond_op
, 0, bltz_op
, 0, 0, 0), RS
| BIMM
},
95 { insn_bltzl
, M(bcond_op
, 0, bltzl_op
, 0, 0, 0), RS
| BIMM
},
96 { insn_bne
, M(bne_op
, 0, 0, 0, 0, 0), RS
| RT
| BIMM
},
97 { insn_cache
, M(cache_op
, 0, 0, 0, 0, 0), RS
| RT
| SIMM
},
98 { insn_daddiu
, M(daddiu_op
, 0, 0, 0, 0, 0), RS
| RT
| SIMM
},
99 { insn_daddu
, M(spec_op
, 0, 0, 0, 0, daddu_op
), RS
| RT
| RD
},
100 { insn_dmfc0
, M(cop0_op
, dmfc_op
, 0, 0, 0, 0), RT
| RD
| SET
},
101 { insn_dmtc0
, M(cop0_op
, dmtc_op
, 0, 0, 0, 0), RT
| RD
| SET
},
102 { insn_dsll
, M(spec_op
, 0, 0, 0, 0, dsll_op
), RT
| RD
| RE
},
103 { insn_dsll32
, M(spec_op
, 0, 0, 0, 0, dsll32_op
), RT
| RD
| RE
},
104 { insn_dsra
, M(spec_op
, 0, 0, 0, 0, dsra_op
), RT
| RD
| RE
},
105 { insn_dsrl
, M(spec_op
, 0, 0, 0, 0, dsrl_op
), RT
| RD
| RE
},
106 { insn_dsrl32
, M(spec_op
, 0, 0, 0, 0, dsrl32_op
), RT
| RD
| RE
},
107 { insn_dsubu
, M(spec_op
, 0, 0, 0, 0, dsubu_op
), RS
| RT
| RD
},
108 { insn_eret
, M(cop0_op
, cop_op
, 0, 0, 0, eret_op
), 0 },
109 { insn_j
, M(j_op
, 0, 0, 0, 0, 0), JIMM
},
110 { insn_jal
, M(jal_op
, 0, 0, 0, 0, 0), JIMM
},
111 { insn_jr
, M(spec_op
, 0, 0, 0, 0, jr_op
), RS
},
112 { insn_ld
, M(ld_op
, 0, 0, 0, 0, 0), RS
| RT
| SIMM
},
113 { insn_ll
, M(ll_op
, 0, 0, 0, 0, 0), RS
| RT
| SIMM
},
114 { insn_lld
, M(lld_op
, 0, 0, 0, 0, 0), RS
| RT
| SIMM
},
115 { insn_lui
, M(lui_op
, 0, 0, 0, 0, 0), RT
| SIMM
},
116 { insn_lw
, M(lw_op
, 0, 0, 0, 0, 0), RS
| RT
| SIMM
},
117 { insn_mfc0
, M(cop0_op
, mfc_op
, 0, 0, 0, 0), RT
| RD
| SET
},
118 { insn_mtc0
, M(cop0_op
, mtc_op
, 0, 0, 0, 0), RT
| RD
| SET
},
119 { insn_ori
, M(ori_op
, 0, 0, 0, 0, 0), RS
| RT
| UIMM
},
120 { insn_pref
, M(pref_op
, 0, 0, 0, 0, 0), RS
| RT
| SIMM
},
121 { insn_rfe
, M(cop0_op
, cop_op
, 0, 0, 0, rfe_op
), 0 },
122 { insn_sc
, M(sc_op
, 0, 0, 0, 0, 0), RS
| RT
| SIMM
},
123 { insn_scd
, M(scd_op
, 0, 0, 0, 0, 0), RS
| RT
| SIMM
},
124 { insn_sd
, M(sd_op
, 0, 0, 0, 0, 0), RS
| RT
| SIMM
},
125 { insn_sll
, M(spec_op
, 0, 0, 0, 0, sll_op
), RT
| RD
| RE
},
126 { insn_sra
, M(spec_op
, 0, 0, 0, 0, sra_op
), RT
| RD
| RE
},
127 { insn_srl
, M(spec_op
, 0, 0, 0, 0, srl_op
), RT
| RD
| RE
},
128 { insn_subu
, M(spec_op
, 0, 0, 0, 0, subu_op
), RS
| RT
| RD
},
129 { insn_sw
, M(sw_op
, 0, 0, 0, 0, 0), RS
| RT
| SIMM
},
130 { insn_tlbp
, M(cop0_op
, cop_op
, 0, 0, 0, tlbp_op
), 0 },
131 { insn_tlbwi
, M(cop0_op
, cop_op
, 0, 0, 0, tlbwi_op
), 0 },
132 { insn_tlbwr
, M(cop0_op
, cop_op
, 0, 0, 0, tlbwr_op
), 0 },
133 { insn_xor
, M(spec_op
, 0, 0, 0, 0, xor_op
), RS
| RT
| RD
},
134 { insn_xori
, M(xori_op
, 0, 0, 0, 0, 0), RS
| RT
| UIMM
},
135 { insn_invalid
, 0, 0 }
140 static inline __cpuinit u32
build_rs(u32 arg
)
143 printk(KERN_WARNING
"Micro-assembler field overflow\n");
145 return (arg
& RS_MASK
) << RS_SH
;
148 static inline __cpuinit u32
build_rt(u32 arg
)
151 printk(KERN_WARNING
"Micro-assembler field overflow\n");
153 return (arg
& RT_MASK
) << RT_SH
;
156 static inline __cpuinit u32
build_rd(u32 arg
)
159 printk(KERN_WARNING
"Micro-assembler field overflow\n");
161 return (arg
& RD_MASK
) << RD_SH
;
164 static inline __cpuinit u32
build_re(u32 arg
)
167 printk(KERN_WARNING
"Micro-assembler field overflow\n");
169 return (arg
& RE_MASK
) << RE_SH
;
172 static inline __cpuinit u32
build_simm(s32 arg
)
174 if (arg
> 0x7fff || arg
< -0x8000)
175 printk(KERN_WARNING
"Micro-assembler field overflow\n");
180 static inline __cpuinit u32
build_uimm(u32 arg
)
183 printk(KERN_WARNING
"Micro-assembler field overflow\n");
185 return arg
& IMM_MASK
;
188 static inline __cpuinit u32
build_bimm(s32 arg
)
190 if (arg
> 0x1ffff || arg
< -0x20000)
191 printk(KERN_WARNING
"Micro-assembler field overflow\n");
194 printk(KERN_WARNING
"Invalid micro-assembler branch target\n");
196 return ((arg
< 0) ? (1 << 15) : 0) | ((arg
>> 2) & 0x7fff);
199 static inline __cpuinit u32
build_jimm(u32 arg
)
201 if (arg
& ~((JIMM_MASK
) << 2))
202 printk(KERN_WARNING
"Micro-assembler field overflow\n");
204 return (arg
>> 2) & JIMM_MASK
;
207 static inline __cpuinit u32
build_func(u32 arg
)
209 if (arg
& ~FUNC_MASK
)
210 printk(KERN_WARNING
"Micro-assembler field overflow\n");
212 return arg
& FUNC_MASK
;
215 static inline __cpuinit u32
build_set(u32 arg
)
218 printk(KERN_WARNING
"Micro-assembler field overflow\n");
220 return arg
& SET_MASK
;
224 * The order of opcode arguments is implicitly left to right,
225 * starting with RS and ending with FUNC or IMM.
227 static void __cpuinit
build_insn(u32
**buf
, enum opcode opc
, ...)
229 struct insn
*ip
= NULL
;
234 for (i
= 0; insn_table
[i
].opcode
!= insn_invalid
; i
++)
235 if (insn_table
[i
].opcode
== opc
) {
240 if (!ip
|| (opc
== insn_daddiu
&& r4k_daddiu_bug()))
241 panic("Unsupported Micro-assembler instruction %d", opc
);
246 op
|= build_rs(va_arg(ap
, u32
));
248 op
|= build_rt(va_arg(ap
, u32
));
250 op
|= build_rd(va_arg(ap
, u32
));
252 op
|= build_re(va_arg(ap
, u32
));
253 if (ip
->fields
& SIMM
)
254 op
|= build_simm(va_arg(ap
, s32
));
255 if (ip
->fields
& UIMM
)
256 op
|= build_uimm(va_arg(ap
, u32
));
257 if (ip
->fields
& BIMM
)
258 op
|= build_bimm(va_arg(ap
, s32
));
259 if (ip
->fields
& JIMM
)
260 op
|= build_jimm(va_arg(ap
, u32
));
261 if (ip
->fields
& FUNC
)
262 op
|= build_func(va_arg(ap
, u32
));
263 if (ip
->fields
& SET
)
264 op
|= build_set(va_arg(ap
, u32
));
271 #define I_u1u2u3(op) \
274 build_insn(buf, insn##op, a, b, c); \
277 #define I_u2u1u3(op) \
280 build_insn(buf, insn##op, b, a, c); \
283 #define I_u3u1u2(op) \
286 build_insn(buf, insn##op, b, c, a); \
289 #define I_u1u2s3(op) \
292 build_insn(buf, insn##op, a, b, c); \
295 #define I_u2s3u1(op) \
298 build_insn(buf, insn##op, c, a, b); \
301 #define I_u2u1s3(op) \
304 build_insn(buf, insn##op, b, a, c); \
310 build_insn(buf, insn##op, a, b); \
316 build_insn(buf, insn##op, a, b); \
322 build_insn(buf, insn##op, a); \
328 build_insn(buf, insn##op); \
382 void __cpuinit
uasm_build_label(struct uasm_label
**lab
, u32
*addr
, int lid
)
389 int __cpuinit
uasm_in_compat_space_p(long addr
)
391 /* Is this address in 32bit compat space? */
393 return (((addr
) & 0xffffffff00000000L
) == 0xffffffff00000000L
);
399 static int __cpuinit
uasm_rel_highest(long val
)
402 return ((((val
+ 0x800080008000L
) >> 48) & 0xffff) ^ 0x8000) - 0x8000;
408 static int __cpuinit
uasm_rel_higher(long val
)
411 return ((((val
+ 0x80008000L
) >> 32) & 0xffff) ^ 0x8000) - 0x8000;
417 int __cpuinit
uasm_rel_hi(long val
)
419 return ((((val
+ 0x8000L
) >> 16) & 0xffff) ^ 0x8000) - 0x8000;
422 int __cpuinit
uasm_rel_lo(long val
)
424 return ((val
& 0xffff) ^ 0x8000) - 0x8000;
427 void __cpuinit
UASM_i_LA_mostly(u32
**buf
, unsigned int rs
, long addr
)
429 if (!uasm_in_compat_space_p(addr
)) {
430 uasm_i_lui(buf
, rs
, uasm_rel_highest(addr
));
431 if (uasm_rel_higher(addr
))
432 uasm_i_daddiu(buf
, rs
, rs
, uasm_rel_higher(addr
));
433 if (uasm_rel_hi(addr
)) {
434 uasm_i_dsll(buf
, rs
, rs
, 16);
435 uasm_i_daddiu(buf
, rs
, rs
, uasm_rel_hi(addr
));
436 uasm_i_dsll(buf
, rs
, rs
, 16);
438 uasm_i_dsll32(buf
, rs
, rs
, 0);
440 uasm_i_lui(buf
, rs
, uasm_rel_hi(addr
));
443 void __cpuinit
UASM_i_LA(u32
**buf
, unsigned int rs
, long addr
)
445 UASM_i_LA_mostly(buf
, rs
, addr
);
446 if (uasm_rel_lo(addr
)) {
447 if (!uasm_in_compat_space_p(addr
))
448 uasm_i_daddiu(buf
, rs
, rs
, uasm_rel_lo(addr
));
450 uasm_i_addiu(buf
, rs
, rs
, uasm_rel_lo(addr
));
454 /* Handle relocations. */
456 uasm_r_mips_pc16(struct uasm_reloc
**rel
, u32
*addr
, int lid
)
459 (*rel
)->type
= R_MIPS_PC16
;
464 static inline void __cpuinit
465 __resolve_relocs(struct uasm_reloc
*rel
, struct uasm_label
*lab
)
467 long laddr
= (long)lab
->addr
;
468 long raddr
= (long)rel
->addr
;
472 *rel
->addr
|= build_bimm(laddr
- (raddr
+ 4));
476 panic("Unsupported Micro-assembler relocation %d",
482 uasm_resolve_relocs(struct uasm_reloc
*rel
, struct uasm_label
*lab
)
484 struct uasm_label
*l
;
486 for (; rel
->lab
!= UASM_LABEL_INVALID
; rel
++)
487 for (l
= lab
; l
->lab
!= UASM_LABEL_INVALID
; l
++)
488 if (rel
->lab
== l
->lab
)
489 __resolve_relocs(rel
, l
);
493 uasm_move_relocs(struct uasm_reloc
*rel
, u32
*first
, u32
*end
, long off
)
495 for (; rel
->lab
!= UASM_LABEL_INVALID
; rel
++)
496 if (rel
->addr
>= first
&& rel
->addr
< end
)
501 uasm_move_labels(struct uasm_label
*lab
, u32
*first
, u32
*end
, long off
)
503 for (; lab
->lab
!= UASM_LABEL_INVALID
; lab
++)
504 if (lab
->addr
>= first
&& lab
->addr
< end
)
509 uasm_copy_handler(struct uasm_reloc
*rel
, struct uasm_label
*lab
, u32
*first
,
510 u32
*end
, u32
*target
)
512 long off
= (long)(target
- first
);
514 memcpy(target
, first
, (end
- first
) * sizeof(u32
));
516 uasm_move_relocs(rel
, first
, end
, off
);
517 uasm_move_labels(lab
, first
, end
, off
);
520 int __cpuinit
uasm_insn_has_bdelay(struct uasm_reloc
*rel
, u32
*addr
)
522 for (; rel
->lab
!= UASM_LABEL_INVALID
; rel
++) {
523 if (rel
->addr
== addr
524 && (rel
->type
== R_MIPS_PC16
525 || rel
->type
== R_MIPS_26
))
532 /* Convenience functions for labeled branches. */
534 uasm_il_bltz(u32
**p
, struct uasm_reloc
**r
, unsigned int reg
, int lid
)
536 uasm_r_mips_pc16(r
, *p
, lid
);
537 uasm_i_bltz(p
, reg
, 0);
541 uasm_il_b(u32
**p
, struct uasm_reloc
**r
, int lid
)
543 uasm_r_mips_pc16(r
, *p
, lid
);
548 uasm_il_beqz(u32
**p
, struct uasm_reloc
**r
, unsigned int reg
, int lid
)
550 uasm_r_mips_pc16(r
, *p
, lid
);
551 uasm_i_beqz(p
, reg
, 0);
555 uasm_il_beqzl(u32
**p
, struct uasm_reloc
**r
, unsigned int reg
, int lid
)
557 uasm_r_mips_pc16(r
, *p
, lid
);
558 uasm_i_beqzl(p
, reg
, 0);
562 uasm_il_bne(u32
**p
, struct uasm_reloc
**r
, unsigned int reg1
,
563 unsigned int reg2
, int lid
)
565 uasm_r_mips_pc16(r
, *p
, lid
);
566 uasm_i_bne(p
, reg1
, reg2
, 0);
570 uasm_il_bnez(u32
**p
, struct uasm_reloc
**r
, unsigned int reg
, int lid
)
572 uasm_r_mips_pc16(r
, *p
, lid
);
573 uasm_i_bnez(p
, reg
, 0);
577 uasm_il_bgezl(u32
**p
, struct uasm_reloc
**r
, unsigned int reg
, int lid
)
579 uasm_r_mips_pc16(r
, *p
, lid
);
580 uasm_i_bgezl(p
, reg
, 0);
584 uasm_il_bgez(u32
**p
, struct uasm_reloc
**r
, unsigned int reg
, int lid
)
586 uasm_r_mips_pc16(r
, *p
, lid
);
587 uasm_i_bgez(p
, reg
, 0);