2 * Stack-less Just-In-Time compiler
4 * Copyright Zoltan Herczeg (hzmester@freemail.hu). All rights reserved.
6 * Redistribution and use in source and binary forms, with or without modification, are
7 * permitted provided that the following conditions are met:
9 * 1. Redistributions of source code must retain the above copyright notice, this list of
10 * conditions and the following disclaimer.
12 * 2. Redistributions in binary form must reproduce the above copyright notice, this list
13 * of conditions and the following disclaimer in the documentation and/or other materials
14 * provided with the distribution.
16 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDER(S) AND CONTRIBUTORS ``AS IS'' AND ANY
17 * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
18 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT
19 * SHALL THE COPYRIGHT HOLDER(S) OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
20 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
21 * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
22 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
23 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
24 * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27 SLJIT_API_FUNC_ATTRIBUTE
const char* sljit_get_platform_name(void)
29 return "ARM-64" SLJIT_CPUINFO
;
32 /* Length of an instruction word */
33 typedef sljit_u32 sljit_ins
;
37 #define TMP_REG1 (SLJIT_NUMBER_OF_REGISTERS + 2)
38 #define TMP_REG2 (SLJIT_NUMBER_OF_REGISTERS + 3)
39 #define TMP_LR (SLJIT_NUMBER_OF_REGISTERS + 4)
40 #define TMP_FP (SLJIT_NUMBER_OF_REGISTERS + 5)
42 #define TMP_FREG1 (SLJIT_NUMBER_OF_FLOAT_REGISTERS + 1)
43 #define TMP_FREG2 (SLJIT_NUMBER_OF_FLOAT_REGISTERS + 2)
45 /* r18 - platform register, currently not used */
46 static const sljit_u8 reg_map
[SLJIT_NUMBER_OF_REGISTERS
+ 8] = {
47 31, 0, 1, 2, 3, 4, 5, 6, 7, 11, 12, 13, 14, 15, 16, 17, 8, 28, 27, 26, 25, 24, 23, 22, 21, 20, 19, 31, 9, 10, 30, 29
50 static const sljit_u8 freg_map
[SLJIT_NUMBER_OF_FLOAT_REGISTERS
+ 3] = {
51 0, 0, 1, 2, 3, 4, 5, 6, 7, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 15, 14, 13, 12, 11, 10, 9, 8, 30, 31
54 #define W_OP ((sljit_ins)1 << 31)
55 #define RD(rd) ((sljit_ins)reg_map[rd])
56 #define RT(rt) ((sljit_ins)reg_map[rt])
57 #define RN(rn) ((sljit_ins)reg_map[rn] << 5)
58 #define RT2(rt2) ((sljit_ins)reg_map[rt2] << 10)
59 #define RM(rm) ((sljit_ins)reg_map[rm] << 16)
60 #define VD(vd) ((sljit_ins)freg_map[vd])
61 #define VT(vt) ((sljit_ins)freg_map[vt])
62 #define VT2(vt) ((sljit_ins)freg_map[vt] << 10)
63 #define VN(vn) ((sljit_ins)freg_map[vn] << 5)
64 #define VM(vm) ((sljit_ins)freg_map[vm] << 16)
66 /* --------------------------------------------------------------------- */
67 /* Instrucion forms */
68 /* --------------------------------------------------------------------- */
70 #define ADC 0x9a000000
71 #define ADD 0x8b000000
72 #define ADDE 0x8b200000
73 #define ADDI 0x91000000
74 #define AND 0x8a000000
75 #define ANDI 0x92000000
76 #define ASRV 0x9ac02800
78 #define B_CC 0x54000000
80 #define BLR 0xd63f0000
82 #define BRK 0xd4200000
83 #define CBZ 0xb4000000
84 #define CLZ 0xdac01000
85 #define CSEL 0x9a800000
86 #define CSINC 0x9a800400
87 #define EOR 0xca000000
88 #define EORI 0xd2000000
89 #define FABS 0x1e60c000
90 #define FADD 0x1e602800
91 #define FCMP 0x1e602000
92 #define FCVT 0x1e224000
93 #define FCVTZS 0x9e780000
94 #define FDIV 0x1e601800
95 #define FMOV 0x1e604000
96 #define FMUL 0x1e600800
97 #define FNEG 0x1e614000
98 #define FSUB 0x1e603800
99 #define LDRI 0xf9400000
100 #define LDRI_F64 0xfd400000
101 #define LDP 0xa9400000
102 #define LDP_F64 0x6d400000
103 #define LDP_POST 0xa8c00000
104 #define LDR_PRE 0xf8400c00
105 #define LSLV 0x9ac02000
106 #define LSRV 0x9ac02400
107 #define MADD 0x9b000000
108 #define MOVK 0xf2800000
109 #define MOVN 0x92800000
110 #define MOVZ 0xd2800000
111 #define NOP 0xd503201f
112 #define ORN 0xaa200000
113 #define ORR 0xaa000000
114 #define ORRI 0xb2000000
115 #define RET 0xd65f0000
116 #define SBC 0xda000000
117 #define SBFM 0x93000000
118 #define SCVTF 0x9e620000
119 #define SDIV 0x9ac00c00
120 #define SMADDL 0x9b200000
121 #define SMULH 0x9b403c00
122 #define STP 0xa9000000
123 #define STP_F64 0x6d000000
124 #define STP_PRE 0xa9800000
125 #define STRB 0x38206800
126 #define STRBI 0x39000000
127 #define STRI 0xf9000000
128 #define STRI_F64 0xfd000000
129 #define STR_FI 0x3d000000
130 #define STR_FR 0x3c206800
131 #define STUR_FI 0x3c000000
132 #define STURBI 0x38000000
133 #define SUB 0xcb000000
134 #define SUBI 0xd1000000
135 #define SUBS 0xeb000000
136 #define UBFM 0xd3000000
137 #define UDIV 0x9ac00800
138 #define UMULH 0x9bc03c00
140 static sljit_s32
push_inst(struct sljit_compiler
*compiler
, sljit_ins ins
)
142 sljit_ins
*ptr
= (sljit_ins
*)ensure_buf(compiler
, sizeof(sljit_ins
));
146 return SLJIT_SUCCESS
;
149 static SLJIT_INLINE sljit_s32
emit_imm64_const(struct sljit_compiler
*compiler
, sljit_s32 dst
, sljit_uw imm
)
151 FAIL_IF(push_inst(compiler
, MOVZ
| RD(dst
) | ((sljit_ins
)(imm
& 0xffff) << 5)));
152 FAIL_IF(push_inst(compiler
, MOVK
| RD(dst
) | (((sljit_ins
)(imm
>> 16) & 0xffff) << 5) | (1 << 21)));
153 FAIL_IF(push_inst(compiler
, MOVK
| RD(dst
) | (((sljit_ins
)(imm
>> 32) & 0xffff) << 5) | (2 << 21)));
154 return push_inst(compiler
, MOVK
| RD(dst
) | ((sljit_ins
)(imm
>> 48) << 5) | (3 << 21));
157 static SLJIT_INLINE sljit_sw
detect_jump_type(struct sljit_jump
*jump
, sljit_ins
*code_ptr
, sljit_ins
*code
, sljit_sw executable_offset
)
160 sljit_uw target_addr
;
162 if (jump
->flags
& SLJIT_REWRITABLE_JUMP
) {
163 jump
->flags
|= PATCH_ABS64
;
167 if (jump
->flags
& JUMP_ADDR
)
168 target_addr
= jump
->u
.target
;
170 SLJIT_ASSERT(jump
->flags
& JUMP_LABEL
);
171 target_addr
= (sljit_uw
)(code
+ jump
->u
.label
->size
) + (sljit_uw
)executable_offset
;
174 diff
= (sljit_sw
)target_addr
- (sljit_sw
)(code_ptr
+ 4) - executable_offset
;
176 if (jump
->flags
& IS_COND
) {
177 diff
+= SSIZE_OF(ins
);
178 if (diff
<= 0xfffff && diff
>= -0x100000) {
179 code_ptr
[-5] ^= (jump
->flags
& IS_CBZ
) ? (0x1 << 24) : 0x1;
180 jump
->addr
-= sizeof(sljit_ins
);
181 jump
->flags
|= PATCH_COND
;
184 diff
-= SSIZE_OF(ins
);
187 if (diff
<= 0x7ffffff && diff
>= -0x8000000) {
188 jump
->flags
|= PATCH_B
;
192 if (target_addr
< 0x100000000l
) {
193 if (jump
->flags
& IS_COND
)
194 code_ptr
[-5] -= (2 << 5);
195 code_ptr
[-2] = code_ptr
[0];
199 if (target_addr
< 0x1000000000000l
) {
200 if (jump
->flags
& IS_COND
)
201 code_ptr
[-5] -= (1 << 5);
202 jump
->flags
|= PATCH_ABS48
;
203 code_ptr
[-1] = code_ptr
[0];
207 jump
->flags
|= PATCH_ABS64
;
211 static SLJIT_INLINE sljit_sw
put_label_get_length(struct sljit_put_label
*put_label
, sljit_uw max_label
)
213 if (max_label
< 0x100000000l
) {
214 put_label
->flags
= 0;
218 if (max_label
< 0x1000000000000l
) {
219 put_label
->flags
= 1;
223 put_label
->flags
= 2;
227 SLJIT_API_FUNC_ATTRIBUTE
void* sljit_generate_code(struct sljit_compiler
*compiler
)
229 struct sljit_memory_fragment
*buf
;
236 sljit_sw executable_offset
;
240 struct sljit_label
*label
;
241 struct sljit_jump
*jump
;
242 struct sljit_const
*const_
;
243 struct sljit_put_label
*put_label
;
246 CHECK_PTR(check_sljit_generate_code(compiler
));
247 reverse_buf(compiler
);
249 code
= (sljit_ins
*)SLJIT_MALLOC_EXEC(compiler
->size
* sizeof(sljit_ins
), compiler
->exec_allocator_data
);
250 PTR_FAIL_WITH_EXEC_IF(code
);
256 executable_offset
= SLJIT_EXEC_OFFSET(code
);
258 label
= compiler
->labels
;
259 jump
= compiler
->jumps
;
260 const_
= compiler
->consts
;
261 put_label
= compiler
->put_labels
;
264 buf_ptr
= (sljit_ins
*)buf
->memory
;
265 buf_end
= buf_ptr
+ (buf
->used_size
>> 2);
267 *code_ptr
= *buf_ptr
++;
268 if (next_addr
== word_count
) {
269 SLJIT_ASSERT(!label
|| label
->size
>= word_count
);
270 SLJIT_ASSERT(!jump
|| jump
->addr
>= word_count
);
271 SLJIT_ASSERT(!const_
|| const_
->addr
>= word_count
);
272 SLJIT_ASSERT(!put_label
|| put_label
->addr
>= word_count
);
274 /* These structures are ordered by their address. */
275 if (label
&& label
->size
== word_count
) {
276 label
->addr
= (sljit_uw
)SLJIT_ADD_EXEC_OFFSET(code_ptr
, executable_offset
);
277 label
->size
= (sljit_uw
)(code_ptr
- code
);
280 if (jump
&& jump
->addr
== word_count
) {
281 jump
->addr
= (sljit_uw
)(code_ptr
- 4);
282 code_ptr
-= detect_jump_type(jump
, code_ptr
, code
, executable_offset
);
285 if (const_
&& const_
->addr
== word_count
) {
286 const_
->addr
= (sljit_uw
)code_ptr
;
287 const_
= const_
->next
;
289 if (put_label
&& put_label
->addr
== word_count
) {
290 SLJIT_ASSERT(put_label
->label
);
291 put_label
->addr
= (sljit_uw
)(code_ptr
- 3);
292 code_ptr
-= put_label_get_length(put_label
, (sljit_uw
)(SLJIT_ADD_EXEC_OFFSET(code
, executable_offset
) + put_label
->label
->size
));
293 put_label
= put_label
->next
;
295 next_addr
= compute_next_addr(label
, jump
, const_
, put_label
);
299 } while (buf_ptr
< buf_end
);
304 if (label
&& label
->size
== word_count
) {
305 label
->addr
= (sljit_uw
)SLJIT_ADD_EXEC_OFFSET(code_ptr
, executable_offset
);
306 label
->size
= (sljit_uw
)(code_ptr
- code
);
310 SLJIT_ASSERT(!label
);
312 SLJIT_ASSERT(!const_
);
313 SLJIT_ASSERT(!put_label
);
314 SLJIT_ASSERT(code_ptr
- code
<= (sljit_sw
)compiler
->size
);
316 jump
= compiler
->jumps
;
319 addr
= (sljit_sw
)((jump
->flags
& JUMP_LABEL
) ? jump
->u
.label
->addr
: jump
->u
.target
);
320 buf_ptr
= (sljit_ins
*)jump
->addr
;
322 if (jump
->flags
& PATCH_B
) {
323 addr
= (addr
- (sljit_sw
)SLJIT_ADD_EXEC_OFFSET(buf_ptr
, executable_offset
)) >> 2;
324 SLJIT_ASSERT(addr
<= 0x1ffffff && addr
>= -0x2000000);
325 buf_ptr
[0] = ((jump
->flags
& IS_BL
) ? BL
: B
) | (sljit_ins
)(addr
& 0x3ffffff);
326 if (jump
->flags
& IS_COND
)
327 buf_ptr
[-1] -= (4 << 5);
330 if (jump
->flags
& PATCH_COND
) {
331 addr
= (addr
- (sljit_sw
)SLJIT_ADD_EXEC_OFFSET(buf_ptr
, executable_offset
)) >> 2;
332 SLJIT_ASSERT(addr
<= 0x3ffff && addr
>= -0x40000);
333 buf_ptr
[0] = (buf_ptr
[0] & ~(sljit_ins
)0xffffe0) | (sljit_ins
)((addr
& 0x7ffff) << 5);
337 SLJIT_ASSERT((jump
->flags
& (PATCH_ABS48
| PATCH_ABS64
)) || (sljit_uw
)addr
<= (sljit_uw
)0xffffffff);
338 SLJIT_ASSERT((jump
->flags
& PATCH_ABS64
) || (sljit_uw
)addr
<= (sljit_uw
)0xffffffffffff);
340 dst
= buf_ptr
[0] & 0x1f;
341 buf_ptr
[0] = MOVZ
| dst
| (((sljit_ins
)addr
& 0xffff) << 5);
342 buf_ptr
[1] = MOVK
| dst
| (((sljit_ins
)(addr
>> 16) & 0xffff) << 5) | (1 << 21);
343 if (jump
->flags
& (PATCH_ABS48
| PATCH_ABS64
))
344 buf_ptr
[2] = MOVK
| dst
| (((sljit_ins
)(addr
>> 32) & 0xffff) << 5) | (2 << 21);
345 if (jump
->flags
& PATCH_ABS64
)
346 buf_ptr
[3] = MOVK
| dst
| ((sljit_ins
)(addr
>> 48) << 5) | (3 << 21);
351 put_label
= compiler
->put_labels
;
353 addr
= (sljit_sw
)put_label
->label
->addr
;
354 buf_ptr
= (sljit_ins
*)put_label
->addr
;
356 buf_ptr
[0] |= ((sljit_ins
)addr
& 0xffff) << 5;
357 buf_ptr
[1] |= ((sljit_ins
)(addr
>> 16) & 0xffff) << 5;
359 if (put_label
->flags
>= 1)
360 buf_ptr
[2] |= ((sljit_ins
)(addr
>> 32) & 0xffff) << 5;
362 if (put_label
->flags
>= 2)
363 buf_ptr
[3] |= (sljit_ins
)(addr
>> 48) << 5;
365 put_label
= put_label
->next
;
368 compiler
->error
= SLJIT_ERR_COMPILED
;
369 compiler
->executable_offset
= executable_offset
;
370 compiler
->executable_size
= (sljit_uw
)(code_ptr
- code
) * sizeof(sljit_ins
);
372 code
= (sljit_ins
*)SLJIT_ADD_EXEC_OFFSET(code
, executable_offset
);
373 code_ptr
= (sljit_ins
*)SLJIT_ADD_EXEC_OFFSET(code_ptr
, executable_offset
);
375 SLJIT_CACHE_FLUSH(code
, code_ptr
);
376 SLJIT_UPDATE_WX_FLAGS(code
, code_ptr
, 1);
380 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_has_cpu_feature(sljit_s32 feature_type
)
382 switch (feature_type
) {
384 #ifdef SLJIT_IS_FPU_AVAILABLE
385 return SLJIT_IS_FPU_AVAILABLE
;
387 /* Available by default. */
393 case SLJIT_HAS_PREFETCH
:
401 /* --------------------------------------------------------------------- */
402 /* Core code generator functions. */
403 /* --------------------------------------------------------------------- */
405 #define COUNT_TRAILING_ZERO(value, result) \
407 if (!(value & 0xffffffff)) { \
411 if (!(value & 0xffff)) { \
415 if (!(value & 0xff)) { \
419 if (!(value & 0xf)) { \
423 if (!(value & 0x3)) { \
427 if (!(value & 0x1)) { \
432 #define LOGICAL_IMM_CHECK (sljit_ins)0x100
434 static sljit_ins
logical_imm(sljit_sw imm
, sljit_u32 len
)
437 sljit_u32 ones
, right
;
441 if (len
& LOGICAL_IMM_CHECK
) {
442 len
&= ~LOGICAL_IMM_CHECK
;
443 if (len
== 32 && (imm
== 0 || imm
== -1))
445 if (len
== 16 && ((sljit_s32
)imm
== 0 || (sljit_s32
)imm
== -1))
449 SLJIT_ASSERT((len
== 32 && imm
!= 0 && imm
!= -1)
450 || (len
== 16 && (sljit_s32
)imm
!= 0 && (sljit_s32
)imm
!= -1));
452 uimm
= (sljit_uw
)imm
;
459 mask
= ((sljit_uw
)1 << len
) - 1;
460 if ((uimm
& mask
) != ((uimm
>> len
) & mask
))
474 uimm
&= ((sljit_uw
)1 << len
) - 1;
476 /* Unsigned right shift. */
477 COUNT_TRAILING_ZERO(uimm
, right
);
479 /* Signed shift. We also know that the highest bit is set. */
480 imm
= (sljit_sw
)~uimm
;
481 SLJIT_ASSERT(imm
< 0);
483 COUNT_TRAILING_ZERO(imm
, ones
);
491 ins
= (0x3f - ((len
<< 1) - 1)) << 10;
494 return ins
| ((len
- ones
- 1) << 10) | ((len
- ones
- right
) << 16);
496 return ins
| ((ones
- 1) << 10) | ((len
- right
) << 16);
499 #undef COUNT_TRAILING_ZERO
501 static sljit_s32
load_immediate(struct sljit_compiler
*compiler
, sljit_s32 dst
, sljit_sw simm
)
503 sljit_uw imm
= (sljit_uw
)simm
;
504 sljit_u32 i
, zeros
, ones
, first
;
507 /* Handling simple immediates first. */
509 return push_inst(compiler
, MOVZ
| RD(dst
) | ((sljit_ins
)imm
<< 5));
511 if (simm
< 0 && simm
>= -0x10000)
512 return push_inst(compiler
, MOVN
| RD(dst
) | (((sljit_ins
)~imm
& 0xffff) << 5));
514 if (imm
<= 0xffffffffl
) {
515 if ((imm
& 0xffff) == 0)
516 return push_inst(compiler
, MOVZ
| RD(dst
) | ((sljit_ins
)(imm
>> 16) << 5) | (1 << 21));
517 if ((imm
& 0xffff0000l
) == 0xffff0000)
518 return push_inst(compiler
, (MOVN
^ W_OP
) | RD(dst
) | (((sljit_ins
)~imm
& 0xffff) << 5));
519 if ((imm
& 0xffff) == 0xffff)
520 return push_inst(compiler
, (MOVN
^ W_OP
) | RD(dst
) | (((sljit_ins
)~imm
& 0xffff0000u
) >> (16 - 5)) | (1 << 21));
522 bitmask
= logical_imm(simm
, 16);
524 return push_inst(compiler
, (ORRI
^ W_OP
) | RD(dst
) | RN(TMP_ZERO
) | bitmask
);
526 FAIL_IF(push_inst(compiler
, MOVZ
| RD(dst
) | (((sljit_ins
)imm
& 0xffff) << 5)));
527 return push_inst(compiler
, MOVK
| RD(dst
) | (((sljit_ins
)imm
& 0xffff0000u
) >> (16 - 5)) | (1 << 21));
530 bitmask
= logical_imm(simm
, 32);
532 return push_inst(compiler
, ORRI
| RD(dst
) | RN(TMP_ZERO
) | bitmask
);
534 if (simm
< 0 && simm
>= -0x100000000l
) {
535 if ((imm
& 0xffff) == 0xffff)
536 return push_inst(compiler
, MOVN
| RD(dst
) | (((sljit_ins
)~imm
& 0xffff0000u
) >> (16 - 5)) | (1 << 21));
538 FAIL_IF(push_inst(compiler
, MOVN
| RD(dst
) | (((sljit_ins
)~imm
& 0xffff) << 5)));
539 return push_inst(compiler
, MOVK
| RD(dst
) | (((sljit_ins
)imm
& 0xffff0000u
) >> (16 - 5)) | (1 << 21));
542 /* A large amount of number can be constructed from ORR and MOVx, but computing them is costly. */
546 for (i
= 4; i
> 0; i
--) {
547 if ((simm
& 0xffff) == 0)
549 if ((simm
& 0xffff) == 0xffff)
554 simm
= (sljit_sw
)imm
;
558 for (i
= 0; i
< 4; i
++) {
559 if (!(simm
& 0xffff)) {
565 FAIL_IF(push_inst(compiler
, MOVN
| RD(dst
) | (((sljit_ins
)simm
& 0xffff) << 5) | (i
<< 21)));
568 FAIL_IF(push_inst(compiler
, MOVK
| RD(dst
) | (((sljit_ins
)~simm
& 0xffff) << 5) | (i
<< 21)));
571 return SLJIT_SUCCESS
;
574 for (i
= 0; i
< 4; i
++) {
575 if (!(simm
& 0xffff)) {
581 FAIL_IF(push_inst(compiler
, MOVZ
| RD(dst
) | (((sljit_ins
)simm
& 0xffff) << 5) | (i
<< 21)));
584 FAIL_IF(push_inst(compiler
, MOVK
| RD(dst
) | (((sljit_ins
)simm
& 0xffff) << 5) | (i
<< 21)));
587 return SLJIT_SUCCESS
;
590 #define ARG1_IMM 0x0010000
591 #define ARG2_IMM 0x0020000
592 #define INT_OP 0x0040000
593 #define SET_FLAGS 0x0080000
594 #define UNUSED_RETURN 0x0100000
596 #define CHECK_FLAGS(flag_bits) \
597 if (flags & SET_FLAGS) { \
598 inv_bits |= flag_bits; \
599 if (flags & UNUSED_RETURN) \
603 static sljit_s32
emit_op_imm(struct sljit_compiler
*compiler
, sljit_s32 flags
, sljit_s32 dst
, sljit_sw arg1
, sljit_sw arg2
)
605 /* dst must be register, TMP_REG1
606 arg1 must be register, TMP_REG1, imm
607 arg2 must be register, TMP_REG2, imm */
608 sljit_ins inv_bits
= (flags
& INT_OP
) ? W_OP
: 0;
610 sljit_s32 op
= (flags
& 0xffff);
614 if (SLJIT_UNLIKELY((flags
& (ARG1_IMM
| ARG2_IMM
)) == (ARG1_IMM
| ARG2_IMM
))) {
615 /* Both are immediates. */
617 if (arg1
== 0 && op
!= SLJIT_ADD
&& op
!= SLJIT_SUB
)
620 FAIL_IF(load_immediate(compiler
, TMP_REG1
, arg1
));
625 if (flags
& (ARG1_IMM
| ARG2_IMM
)) {
626 reg
= (sljit_s32
)((flags
& ARG2_IMM
) ? arg1
: arg2
);
627 imm
= (flags
& ARG2_IMM
) ? arg2
: arg1
;
634 /* No form with immediate operand (except imm 0, which
635 is represented by a ZERO register). */
638 SLJIT_ASSERT(!(flags
& SET_FLAGS
) && (flags
& ARG2_IMM
) && arg1
== TMP_REG1
);
639 return load_immediate(compiler
, dst
, imm
);
641 SLJIT_ASSERT(flags
& ARG2_IMM
);
642 FAIL_IF(load_immediate(compiler
, dst
, (flags
& INT_OP
) ? (~imm
& 0xffffffff) : ~imm
));
645 compiler
->status_flags_state
= SLJIT_CURRENT_FLAGS_SUB
;
646 if (flags
& ARG1_IMM
)
652 compiler
->status_flags_state
= SLJIT_CURRENT_FLAGS_ADD
;
655 CHECK_FLAGS(1 << 29);
656 return push_inst(compiler
, ((op
== SLJIT_ADD
? ADDI
: SUBI
) ^ inv_bits
) | RD(dst
) | RN(reg
));
658 if (imm
> 0 && imm
<= 0xfff) {
659 CHECK_FLAGS(1 << 29);
660 return push_inst(compiler
, (ADDI
^ inv_bits
) | RD(dst
) | RN(reg
) | ((sljit_ins
)imm
<< 10));
663 if (nimm
> 0 && nimm
<= 0xfff) {
664 CHECK_FLAGS(1 << 29);
665 return push_inst(compiler
, (SUBI
^ inv_bits
) | RD(dst
) | RN(reg
) | ((sljit_ins
)nimm
<< 10));
667 if (imm
> 0 && imm
<= 0xffffff && !(imm
& 0xfff)) {
668 CHECK_FLAGS(1 << 29);
669 return push_inst(compiler
, (ADDI
^ inv_bits
) | RD(dst
) | RN(reg
) | (((sljit_ins
)imm
>> 12) << 10) | (1 << 22));
671 if (nimm
> 0 && nimm
<= 0xffffff && !(nimm
& 0xfff)) {
672 CHECK_FLAGS(1 << 29);
673 return push_inst(compiler
, (SUBI
^ inv_bits
) | RD(dst
) | RN(reg
) | (((sljit_ins
)nimm
>> 12) << 10) | (1 << 22));
675 if (imm
> 0 && imm
<= 0xffffff && !(flags
& SET_FLAGS
)) {
676 FAIL_IF(push_inst(compiler
, (ADDI
^ inv_bits
) | RD(dst
) | RN(reg
) | (((sljit_ins
)imm
>> 12) << 10) | (1 << 22)));
677 return push_inst(compiler
, (ADDI
^ inv_bits
) | RD(dst
) | RN(dst
) | (((sljit_ins
)imm
& 0xfff) << 10));
679 if (nimm
> 0 && nimm
<= 0xffffff && !(flags
& SET_FLAGS
)) {
680 FAIL_IF(push_inst(compiler
, (SUBI
^ inv_bits
) | RD(dst
) | RN(reg
) | (((sljit_ins
)nimm
>> 12) << 10) | (1 << 22)));
681 return push_inst(compiler
, (SUBI
^ inv_bits
) | RD(dst
) | RN(dst
) | (((sljit_ins
)nimm
& 0xfff) << 10));
685 inst_bits
= logical_imm(imm
, LOGICAL_IMM_CHECK
| ((flags
& INT_OP
) ? 16 : 32));
688 CHECK_FLAGS(3 << 29);
689 return push_inst(compiler
, (ANDI
^ inv_bits
) | RD(dst
) | RN(reg
) | inst_bits
);
692 inst_bits
= logical_imm(imm
, LOGICAL_IMM_CHECK
| ((flags
& INT_OP
) ? 16 : 32));
699 FAIL_IF(push_inst(compiler
, (inst_bits
^ inv_bits
) | RD(dst
) | RN(reg
)));
702 if (flags
& ARG1_IMM
)
704 if (flags
& INT_OP
) {
706 FAIL_IF(push_inst(compiler
, (UBFM
^ inv_bits
) | RD(dst
) | RN(arg1
)
707 | (((sljit_ins
)-imm
& 0x1f) << 16) | ((31 - (sljit_ins
)imm
) << 10)));
711 FAIL_IF(push_inst(compiler
, (UBFM
^ inv_bits
) | RD(dst
) | RN(arg1
) | (1 << 22)
712 | (((sljit_ins
)-imm
& 0x3f) << 16) | ((63 - (sljit_ins
)imm
) << 10)));
717 if (flags
& ARG1_IMM
)
719 if (op
== SLJIT_ASHR
)
721 if (flags
& INT_OP
) {
723 FAIL_IF(push_inst(compiler
, (UBFM
^ inv_bits
) | RD(dst
) | RN(arg1
)
724 | ((sljit_ins
)imm
<< 16) | (31 << 10)));
728 FAIL_IF(push_inst(compiler
, (UBFM
^ inv_bits
) | RD(dst
) | RN(arg1
)
729 | (1 << 22) | ((sljit_ins
)imm
<< 16) | (63 << 10)));
737 if (flags
& ARG2_IMM
) {
741 FAIL_IF(load_immediate(compiler
, TMP_REG2
, arg2
));
749 FAIL_IF(load_immediate(compiler
, TMP_REG1
, arg1
));
755 /* Both arguments are registers. */
759 SLJIT_ASSERT(!(flags
& SET_FLAGS
) && arg1
== TMP_REG1
);
761 return SLJIT_SUCCESS
;
762 return push_inst(compiler
, ORR
| RD(dst
) | RN(TMP_ZERO
) | RM(arg2
));
764 SLJIT_ASSERT(!(flags
& SET_FLAGS
) && arg1
== TMP_REG1
);
765 return push_inst(compiler
, (UBFM
^ W_OP
) | RD(dst
) | RN(arg2
) | (7 << 10));
767 SLJIT_ASSERT(!(flags
& SET_FLAGS
) && arg1
== TMP_REG1
);
768 if (!(flags
& INT_OP
))
770 return push_inst(compiler
, (SBFM
^ inv_bits
) | RD(dst
) | RN(arg2
) | (7 << 10));
772 SLJIT_ASSERT(!(flags
& SET_FLAGS
) && arg1
== TMP_REG1
);
773 return push_inst(compiler
, (UBFM
^ W_OP
) | RD(dst
) | RN(arg2
) | (15 << 10));
775 SLJIT_ASSERT(!(flags
& SET_FLAGS
) && arg1
== TMP_REG1
);
776 if (!(flags
& INT_OP
))
778 return push_inst(compiler
, (SBFM
^ inv_bits
) | RD(dst
) | RN(arg2
) | (15 << 10));
780 SLJIT_ASSERT(!(flags
& SET_FLAGS
) && arg1
== TMP_REG1
);
782 return SLJIT_SUCCESS
;
785 SLJIT_ASSERT(!(flags
& SET_FLAGS
) && arg1
== TMP_REG1
);
786 return push_inst(compiler
, (ORR
^ W_OP
) | RD(dst
) | RN(TMP_ZERO
) | RM(arg2
));
788 SLJIT_ASSERT(!(flags
& SET_FLAGS
) && arg1
== TMP_REG1
);
789 return push_inst(compiler
, SBFM
| (1 << 22) | RD(dst
) | RN(arg2
) | (31 << 10));
791 SLJIT_ASSERT(arg1
== TMP_REG1
);
792 FAIL_IF(push_inst(compiler
, (ORN
^ inv_bits
) | RD(dst
) | RN(TMP_ZERO
) | RM(arg2
)));
793 break; /* Set flags. */
795 SLJIT_ASSERT(arg1
== TMP_REG1
);
796 return push_inst(compiler
, (CLZ
^ inv_bits
) | RD(dst
) | RN(arg2
));
798 compiler
->status_flags_state
= SLJIT_CURRENT_FLAGS_ADD
;
799 CHECK_FLAGS(1 << 29);
800 return push_inst(compiler
, (ADD
^ inv_bits
) | RD(dst
) | RN(arg1
) | RM(arg2
));
802 compiler
->status_flags_state
= SLJIT_CURRENT_FLAGS_ADD
;
803 CHECK_FLAGS(1 << 29);
804 return push_inst(compiler
, (ADC
^ inv_bits
) | RD(dst
) | RN(arg1
) | RM(arg2
));
806 compiler
->status_flags_state
= SLJIT_CURRENT_FLAGS_SUB
;
807 CHECK_FLAGS(1 << 29);
808 return push_inst(compiler
, (SUB
^ inv_bits
) | RD(dst
) | RN(arg1
) | RM(arg2
));
810 compiler
->status_flags_state
= SLJIT_CURRENT_FLAGS_SUB
;
811 CHECK_FLAGS(1 << 29);
812 return push_inst(compiler
, (SBC
^ inv_bits
) | RD(dst
) | RN(arg1
) | RM(arg2
));
814 compiler
->status_flags_state
= 0;
815 if (!(flags
& SET_FLAGS
))
816 return push_inst(compiler
, (MADD
^ inv_bits
) | RD(dst
) | RN(arg1
) | RM(arg2
) | RT2(TMP_ZERO
));
817 if (flags
& INT_OP
) {
818 FAIL_IF(push_inst(compiler
, SMADDL
| RD(dst
) | RN(arg1
) | RM(arg2
) | (31 << 10)));
819 FAIL_IF(push_inst(compiler
, ADD
| RD(TMP_LR
) | RN(TMP_ZERO
) | RM(dst
) | (2 << 22) | (31 << 10)));
820 return push_inst(compiler
, SUBS
| RD(TMP_ZERO
) | RN(TMP_LR
) | RM(dst
) | (2 << 22) | (63 << 10));
822 FAIL_IF(push_inst(compiler
, SMULH
| RD(TMP_LR
) | RN(arg1
) | RM(arg2
)));
823 FAIL_IF(push_inst(compiler
, MADD
| RD(dst
) | RN(arg1
) | RM(arg2
) | RT2(TMP_ZERO
)));
824 return push_inst(compiler
, SUBS
| RD(TMP_ZERO
) | RN(TMP_LR
) | RM(dst
) | (2 << 22) | (63 << 10));
826 CHECK_FLAGS(3 << 29);
827 return push_inst(compiler
, (AND
^ inv_bits
) | RD(dst
) | RN(arg1
) | RM(arg2
));
829 FAIL_IF(push_inst(compiler
, (ORR
^ inv_bits
) | RD(dst
) | RN(arg1
) | RM(arg2
)));
830 break; /* Set flags. */
832 FAIL_IF(push_inst(compiler
, (EOR
^ inv_bits
) | RD(dst
) | RN(arg1
) | RM(arg2
)));
833 break; /* Set flags. */
835 FAIL_IF(push_inst(compiler
, (LSLV
^ inv_bits
) | RD(dst
) | RN(arg1
) | RM(arg2
)));
836 break; /* Set flags. */
838 FAIL_IF(push_inst(compiler
, (LSRV
^ inv_bits
) | RD(dst
) | RN(arg1
) | RM(arg2
)));
839 break; /* Set flags. */
841 FAIL_IF(push_inst(compiler
, (ASRV
^ inv_bits
) | RD(dst
) | RN(arg1
) | RM(arg2
)));
842 break; /* Set flags. */
845 return SLJIT_SUCCESS
;
849 if (flags
& SET_FLAGS
)
850 return push_inst(compiler
, (SUBS
^ inv_bits
) | RD(TMP_ZERO
) | RN(dst
) | RM(TMP_ZERO
));
851 return SLJIT_SUCCESS
;
857 #define BYTE_SIZE 0x0
858 #define HALF_SIZE 0x1
860 #define WORD_SIZE 0x3
862 #define MEM_SIZE_SHIFT(flags) ((sljit_ins)(flags) & 0x3)
864 static sljit_s32
emit_op_mem(struct sljit_compiler
*compiler
, sljit_s32 flags
, sljit_s32 reg
,
865 sljit_s32 arg
, sljit_sw argw
, sljit_s32 tmp_reg
)
867 sljit_u32 shift
= MEM_SIZE_SHIFT(flags
);
868 sljit_u32 type
= (shift
<< 30);
870 if (!(flags
& STORE
))
871 type
|= (flags
& SIGNED
) ? 0x00800000 : 0x00400000;
873 SLJIT_ASSERT(arg
& SLJIT_MEM
);
875 if (SLJIT_UNLIKELY(arg
& OFFS_REG_MASK
)) {
878 if (argw
== 0 || argw
== shift
)
879 return push_inst(compiler
, STRB
| type
| RT(reg
)
880 | RN(arg
& REG_MASK
) | RM(OFFS_REG(arg
)) | (argw
? (1 << 12) : 0));
882 FAIL_IF(push_inst(compiler
, ADD
| RD(tmp_reg
) | RN(arg
& REG_MASK
) | RM(OFFS_REG(arg
)) | ((sljit_ins
)argw
<< 10)));
883 return push_inst(compiler
, STRBI
| type
| RT(reg
) | RN(tmp_reg
));
889 FAIL_IF(load_immediate(compiler
, tmp_reg
, argw
& ~(0xfff << shift
)));
891 argw
= (argw
>> shift
) & 0xfff;
893 return push_inst(compiler
, STRBI
| type
| RT(reg
) | RN(tmp_reg
) | ((sljit_ins
)argw
<< 10));
896 if (argw
>= 0 && (argw
& ((1 << shift
) - 1)) == 0) {
897 if ((argw
>> shift
) <= 0xfff)
898 return push_inst(compiler
, STRBI
| type
| RT(reg
) | RN(arg
) | ((sljit_ins
)argw
<< (10 - shift
)));
900 if (argw
<= 0xffffff) {
901 FAIL_IF(push_inst(compiler
, ADDI
| (1 << 22) | RD(tmp_reg
) | RN(arg
) | (((sljit_ins
)argw
>> 12) << 10)));
903 argw
= ((argw
& 0xfff) >> shift
);
904 return push_inst(compiler
, STRBI
| type
| RT(reg
) | RN(tmp_reg
) | ((sljit_ins
)argw
<< 10));
908 if (argw
<= 255 && argw
>= -256)
909 return push_inst(compiler
, STURBI
| type
| RT(reg
) | RN(arg
) | (((sljit_ins
)argw
& 0x1ff) << 12));
911 FAIL_IF(load_immediate(compiler
, tmp_reg
, argw
));
913 return push_inst(compiler
, STRB
| type
| RT(reg
) | RN(arg
) | RM(tmp_reg
));
916 /* --------------------------------------------------------------------- */
918 /* --------------------------------------------------------------------- */
920 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_enter(struct sljit_compiler
*compiler
,
921 sljit_s32 options
, sljit_s32 arg_types
, sljit_s32 scratches
, sljit_s32 saveds
,
922 sljit_s32 fscratches
, sljit_s32 fsaveds
, sljit_s32 local_size
)
924 sljit_s32 prev
, fprev
, saved_regs_size
, i
, tmp
;
925 sljit_s32 saved_arg_count
= SLJIT_KEPT_SAVEDS_COUNT(options
);
929 CHECK(check_sljit_emit_enter(compiler
, options
, arg_types
, scratches
, saveds
, fscratches
, fsaveds
, local_size
));
930 set_emit_enter(compiler
, options
, arg_types
, scratches
, saveds
, fscratches
, fsaveds
, local_size
);
932 saved_regs_size
= GET_SAVED_REGISTERS_SIZE(scratches
, saveds
- saved_arg_count
, 2);
933 saved_regs_size
+= GET_SAVED_FLOAT_REGISTERS_SIZE(fscratches
, fsaveds
, SSIZE_OF(f64
));
935 local_size
= (local_size
+ saved_regs_size
+ 0xf) & ~0xf;
936 compiler
->local_size
= local_size
;
938 if (local_size
<= 512) {
939 FAIL_IF(push_inst(compiler
, STP_PRE
| RT(TMP_FP
) | RT2(TMP_LR
)
940 | RN(SLJIT_SP
) | (sljit_ins
)((-(local_size
>> 3) & 0x7f) << 15)));
941 offs
= (sljit_ins
)(local_size
- 2 * SSIZE_OF(sw
)) << (15 - 3);
944 saved_regs_size
= ((saved_regs_size
- 2 * SSIZE_OF(sw
)) + 0xf) & ~0xf;
946 FAIL_IF(push_inst(compiler
, SUBI
| RD(SLJIT_SP
) | RN(SLJIT_SP
) | ((sljit_ins
)saved_regs_size
<< 10)));
947 offs
= (sljit_ins
)(saved_regs_size
- 2 * SSIZE_OF(sw
)) << (15 - 3);
948 local_size
-= saved_regs_size
;
949 SLJIT_ASSERT(local_size
> 0);
954 tmp
= SLJIT_S0
- saveds
;
955 for (i
= SLJIT_S0
- saved_arg_count
; i
> tmp
; i
--) {
960 FAIL_IF(push_inst(compiler
, STP
| RT(prev
) | RT2(i
) | RN(SLJIT_SP
) | offs
));
961 offs
-= (sljit_ins
)2 << 15;
965 for (i
= scratches
; i
>= SLJIT_FIRST_SAVED_REG
; i
--) {
970 FAIL_IF(push_inst(compiler
, STP
| RT(prev
) | RT2(i
) | RN(SLJIT_SP
) | offs
));
971 offs
-= (sljit_ins
)2 << 15;
977 tmp
= SLJIT_FS0
- fsaveds
;
978 for (i
= SLJIT_FS0
; i
> tmp
; i
--) {
983 FAIL_IF(push_inst(compiler
, STP_F64
| VT(fprev
) | VT2(i
) | RN(SLJIT_SP
) | offs
));
984 offs
-= (sljit_ins
)2 << 15;
988 for (i
= fscratches
; i
>= SLJIT_FIRST_SAVED_FLOAT_REG
; i
--) {
993 FAIL_IF(push_inst(compiler
, STP_F64
| VT(fprev
) | VT2(i
) | RN(SLJIT_SP
) | offs
));
994 offs
-= (sljit_ins
)2 << 15;
999 FAIL_IF(push_inst(compiler
, STRI_F64
| VT(fprev
) | RN(SLJIT_SP
) | (offs
>> 5) | (1 << 10)));
1002 FAIL_IF(push_inst(compiler
, STRI
| RT(prev
) | RN(SLJIT_SP
) | (offs
>> 5) | ((fprev
== -1) ? (1 << 10) : 0)));
1006 if (local_size
> 4096)
1007 FAIL_IF(push_inst(compiler
, SUBI
| RD(SLJIT_SP
) | RN(SLJIT_SP
) | (1 << 10) | (1 << 22)));
1010 if (!(options
& SLJIT_ENTER_REG_ARG
)) {
1011 arg_types
>>= SLJIT_ARG_SHIFT
;
1012 saved_arg_count
= 0;
1016 if ((arg_types
& SLJIT_ARG_MASK
) < SLJIT_ARG_TYPE_F64
) {
1017 if (!(arg_types
& SLJIT_ARG_TYPE_SCRATCH_REG
)) {
1018 FAIL_IF(push_inst(compiler
, ORR
| RD(SLJIT_S0
- saved_arg_count
) | RN(TMP_ZERO
) | RM(tmp
)));
1023 arg_types
>>= SLJIT_ARG_SHIFT
;
1028 if (local_size
> 4096) {
1029 if (local_size
< 4 * 4096) {
1030 /* No need for a loop. */
1032 if (local_size
>= 2 * 4096) {
1033 if (local_size
>= 3 * 4096) {
1034 FAIL_IF(push_inst(compiler
, LDRI
| RT(TMP_ZERO
) | RN(SLJIT_SP
)));
1035 FAIL_IF(push_inst(compiler
, SUBI
| RD(SLJIT_SP
) | RN(SLJIT_SP
) | (1 << 10) | (1 << 22)));
1038 FAIL_IF(push_inst(compiler
, LDRI
| RT(TMP_ZERO
) | RN(SLJIT_SP
)));
1039 FAIL_IF(push_inst(compiler
, SUBI
| RD(SLJIT_SP
) | RN(SLJIT_SP
) | (1 << 10) | (1 << 22)));
1043 FAIL_IF(push_inst(compiler
, MOVZ
| RD(TMP_REG1
) | ((((sljit_ins
)local_size
>> 12) - 1) << 5)));
1044 FAIL_IF(push_inst(compiler
, LDRI
| RT(TMP_ZERO
) | RN(SLJIT_SP
)));
1045 FAIL_IF(push_inst(compiler
, SUBI
| RD(SLJIT_SP
) | RN(SLJIT_SP
) | (1 << 10) | (1 << 22)));
1046 FAIL_IF(push_inst(compiler
, SUBI
| (1 << 29) | RD(TMP_REG1
) | RN(TMP_REG1
) | (1 << 10)));
1047 FAIL_IF(push_inst(compiler
, B_CC
| ((((sljit_ins
) -3) & 0x7ffff) << 5) | 0x1 /* not-equal */));
1050 local_size
&= 0xfff;
1053 FAIL_IF(push_inst(compiler
, LDRI
| RT(TMP_ZERO
) | RN(SLJIT_SP
)));
1055 FAIL_IF(push_inst(compiler
, STP
| RT(TMP_FP
) | RT2(TMP_LR
) | RN(SLJIT_SP
)));
1058 if (local_size
> 0) {
1059 if (local_size
<= 512)
1060 FAIL_IF(push_inst(compiler
, STP_PRE
| RT(TMP_FP
) | RT2(TMP_LR
)
1061 | RN(SLJIT_SP
) | (sljit_ins
)((-(local_size
>> 3) & 0x7f) << 15)));
1063 if (local_size
>= 4096)
1064 local_size
= (1 << (22 - 10));
1066 FAIL_IF(push_inst(compiler
, SUBI
| RD(SLJIT_SP
) | RN(SLJIT_SP
) | ((sljit_ins
)local_size
<< 10)));
1067 FAIL_IF(push_inst(compiler
, STP
| RT(TMP_FP
) | RT2(TMP_LR
) | RN(SLJIT_SP
)));
1073 /* The local_size does not include saved registers size. */
1074 if (local_size
!= 0) {
1075 if (local_size
> 0xfff) {
1076 FAIL_IF(push_inst(compiler
, SUBI
| RD(SLJIT_SP
) | RN(SLJIT_SP
) | (((sljit_ins
)local_size
>> 12) << 10) | (1 << 22)));
1077 local_size
&= 0xfff;
1080 if (local_size
> 512 || local_size
== 0) {
1081 if (local_size
!= 0)
1082 FAIL_IF(push_inst(compiler
, SUBI
| RD(SLJIT_SP
) | RN(SLJIT_SP
) | ((sljit_ins
)local_size
<< 10)));
1084 FAIL_IF(push_inst(compiler
, STP
| RT(TMP_FP
) | RT2(TMP_LR
) | RN(SLJIT_SP
)));
1086 FAIL_IF(push_inst(compiler
, STP_PRE
| RT(TMP_FP
) | RT2(TMP_LR
)
1087 | RN(SLJIT_SP
) | (sljit_ins
)((-(local_size
>> 3) & 0x7f) << 15)));
1092 return push_inst(compiler
, ADDI
| RD(TMP_FP
) | RN(SLJIT_SP
) | (0 << 10));
1095 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_set_context(struct sljit_compiler
*compiler
,
1096 sljit_s32 options
, sljit_s32 arg_types
, sljit_s32 scratches
, sljit_s32 saveds
,
1097 sljit_s32 fscratches
, sljit_s32 fsaveds
, sljit_s32 local_size
)
1099 sljit_s32 saved_regs_size
;
1102 CHECK(check_sljit_set_context(compiler
, options
, arg_types
, scratches
, saveds
, fscratches
, fsaveds
, local_size
));
1103 set_set_context(compiler
, options
, arg_types
, scratches
, saveds
, fscratches
, fsaveds
, local_size
);
1105 saved_regs_size
= GET_SAVED_REGISTERS_SIZE(scratches
, saveds
- SLJIT_KEPT_SAVEDS_COUNT(options
), 2);
1106 saved_regs_size
+= GET_SAVED_FLOAT_REGISTERS_SIZE(fscratches
, fsaveds
, SSIZE_OF(f64
));
1108 compiler
->local_size
= (local_size
+ saved_regs_size
+ 0xf) & ~0xf;
1109 return SLJIT_SUCCESS
;
1112 static sljit_s32
emit_stack_frame_release(struct sljit_compiler
*compiler
)
1114 sljit_s32 local_size
, prev
, fprev
, i
, tmp
;
1117 local_size
= compiler
->local_size
;
1119 if (local_size
> 512 && local_size
<= 512 + 496) {
1120 FAIL_IF(push_inst(compiler
, LDP_POST
| RT(TMP_FP
) | RT2(TMP_LR
)
1121 | RN(SLJIT_SP
) | ((sljit_ins
)(local_size
- 512) << (15 - 3))));
1124 FAIL_IF(push_inst(compiler
, LDP
| RT(TMP_FP
) | RT2(TMP_LR
) | RN(SLJIT_SP
)));
1126 if (local_size
> 512) {
1128 if (local_size
> 0xfff) {
1129 FAIL_IF(push_inst(compiler
, ADDI
| RD(SLJIT_SP
) | RN(SLJIT_SP
)
1130 | (((sljit_ins
)local_size
>> 12) << 10) | (1 << 22)));
1131 local_size
&= 0xfff;
1134 FAIL_IF(push_inst(compiler
, ADDI
| RD(SLJIT_SP
) | RN(SLJIT_SP
) | ((sljit_ins
)local_size
<< 10)));
1138 offs
= (sljit_ins
)(local_size
- 2 * SSIZE_OF(sw
)) << (15 - 3);
1141 tmp
= SLJIT_S0
- compiler
->saveds
;
1142 for (i
= SLJIT_S0
- SLJIT_KEPT_SAVEDS_COUNT(compiler
->options
); i
> tmp
; i
--) {
1147 FAIL_IF(push_inst(compiler
, LDP
| RT(prev
) | RT2(i
) | RN(SLJIT_SP
) | offs
));
1148 offs
-= (sljit_ins
)2 << 15;
1152 for (i
= compiler
->scratches
; i
>= SLJIT_FIRST_SAVED_REG
; i
--) {
1157 FAIL_IF(push_inst(compiler
, LDP
| RT(prev
) | RT2(i
) | RN(SLJIT_SP
) | offs
));
1158 offs
-= (sljit_ins
)2 << 15;
1164 tmp
= SLJIT_FS0
- compiler
->fsaveds
;
1165 for (i
= SLJIT_FS0
; i
> tmp
; i
--) {
1170 FAIL_IF(push_inst(compiler
, LDP_F64
| VT(fprev
) | VT2(i
) | RN(SLJIT_SP
) | offs
));
1171 offs
-= (sljit_ins
)2 << 15;
1175 for (i
= compiler
->fscratches
; i
>= SLJIT_FIRST_SAVED_FLOAT_REG
; i
--) {
1180 FAIL_IF(push_inst(compiler
, LDP_F64
| VT(fprev
) | VT2(i
) | RN(SLJIT_SP
) | offs
));
1181 offs
-= (sljit_ins
)2 << 15;
1186 FAIL_IF(push_inst(compiler
, LDRI_F64
| VT(fprev
) | RN(SLJIT_SP
) | (offs
>> 5) | (1 << 10)));
1189 FAIL_IF(push_inst(compiler
, LDRI
| RT(prev
) | RN(SLJIT_SP
) | (offs
>> 5) | ((fprev
== -1) ? (1 << 10) : 0)));
1191 /* This and the next call/jump instruction can be executed parallelly. */
1192 return push_inst(compiler
, ADDI
| RD(SLJIT_SP
) | RN(SLJIT_SP
) | (sljit_ins
)(local_size
<< 10));
1195 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_return_void(struct sljit_compiler
*compiler
)
1198 CHECK(check_sljit_emit_return_void(compiler
));
1200 FAIL_IF(emit_stack_frame_release(compiler
));
1202 return push_inst(compiler
, RET
| RN(TMP_LR
));
1205 /* --------------------------------------------------------------------- */
1207 /* --------------------------------------------------------------------- */
1209 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_op0(struct sljit_compiler
*compiler
, sljit_s32 op
)
1211 sljit_ins inv_bits
= (op
& SLJIT_32
) ? W_OP
: 0;
1214 CHECK(check_sljit_emit_op0(compiler
, op
));
1216 op
= GET_OPCODE(op
);
1218 case SLJIT_BREAKPOINT
:
1219 return push_inst(compiler
, BRK
);
1221 return push_inst(compiler
, NOP
);
1224 FAIL_IF(push_inst(compiler
, ORR
| RD(TMP_REG1
) | RN(TMP_ZERO
) | RM(SLJIT_R0
)));
1225 FAIL_IF(push_inst(compiler
, MADD
| RD(SLJIT_R0
) | RN(SLJIT_R0
) | RM(SLJIT_R1
) | RT2(TMP_ZERO
)));
1226 return push_inst(compiler
, (op
== SLJIT_LMUL_UW
? UMULH
: SMULH
) | RD(SLJIT_R1
) | RN(TMP_REG1
) | RM(SLJIT_R1
));
1227 case SLJIT_DIVMOD_UW
:
1228 case SLJIT_DIVMOD_SW
:
1229 FAIL_IF(push_inst(compiler
, (ORR
^ inv_bits
) | RD(TMP_REG1
) | RN(TMP_ZERO
) | RM(SLJIT_R0
)));
1230 FAIL_IF(push_inst(compiler
, ((op
== SLJIT_DIVMOD_UW
? UDIV
: SDIV
) ^ inv_bits
) | RD(SLJIT_R0
) | RN(SLJIT_R0
) | RM(SLJIT_R1
)));
1231 FAIL_IF(push_inst(compiler
, (MADD
^ inv_bits
) | RD(SLJIT_R1
) | RN(SLJIT_R0
) | RM(SLJIT_R1
) | RT2(TMP_ZERO
)));
1232 return push_inst(compiler
, (SUB
^ inv_bits
) | RD(SLJIT_R1
) | RN(TMP_REG1
) | RM(SLJIT_R1
));
1235 return push_inst(compiler
, ((op
== SLJIT_DIV_UW
? UDIV
: SDIV
) ^ inv_bits
) | RD(SLJIT_R0
) | RN(SLJIT_R0
) | RM(SLJIT_R1
));
1237 case SLJIT_SKIP_FRAMES_BEFORE_RETURN
:
1238 return SLJIT_SUCCESS
;
1241 return SLJIT_SUCCESS
;
1244 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_op1(struct sljit_compiler
*compiler
, sljit_s32 op
,
1245 sljit_s32 dst
, sljit_sw dstw
,
1246 sljit_s32 src
, sljit_sw srcw
)
1248 sljit_s32 dst_r
, flags
, mem_flags
;
1249 sljit_s32 op_flags
= GET_ALL_FLAGS(op
);
1252 CHECK(check_sljit_emit_op1(compiler
, op
, dst
, dstw
, src
, srcw
));
1253 ADJUST_LOCAL_OFFSET(dst
, dstw
);
1254 ADJUST_LOCAL_OFFSET(src
, srcw
);
1256 dst_r
= FAST_IS_REG(dst
) ? dst
: TMP_REG1
;
1258 op
= GET_OPCODE(op
);
1259 if (op
>= SLJIT_MOV
&& op
<= SLJIT_MOV_P
) {
1260 /* Both operands are registers. */
1261 if (dst_r
!= TMP_REG1
&& FAST_IS_REG(src
))
1262 return emit_op_imm(compiler
, op
| ((op_flags
& SLJIT_32
) ? INT_OP
: 0), dst_r
, TMP_REG1
, src
);
1267 mem_flags
= WORD_SIZE
;
1270 mem_flags
= BYTE_SIZE
;
1271 if (src
& SLJIT_IMM
)
1272 srcw
= (sljit_u8
)srcw
;
1275 mem_flags
= BYTE_SIZE
| SIGNED
;
1276 if (src
& SLJIT_IMM
)
1277 srcw
= (sljit_s8
)srcw
;
1280 mem_flags
= HALF_SIZE
;
1281 if (src
& SLJIT_IMM
)
1282 srcw
= (sljit_u16
)srcw
;
1285 mem_flags
= HALF_SIZE
| SIGNED
;
1286 if (src
& SLJIT_IMM
)
1287 srcw
= (sljit_s16
)srcw
;
1290 mem_flags
= INT_SIZE
;
1291 if (src
& SLJIT_IMM
)
1292 srcw
= (sljit_u32
)srcw
;
1296 mem_flags
= INT_SIZE
| SIGNED
;
1297 if (src
& SLJIT_IMM
)
1298 srcw
= (sljit_s32
)srcw
;
1301 SLJIT_UNREACHABLE();
1306 if (src
& SLJIT_IMM
)
1307 FAIL_IF(emit_op_imm(compiler
, SLJIT_MOV
| ARG2_IMM
, dst_r
, TMP_REG1
, srcw
));
1308 else if (!(src
& SLJIT_MEM
))
1311 FAIL_IF(emit_op_mem(compiler
, mem_flags
, dst_r
, src
, srcw
, TMP_REG1
));
1313 if (dst
& SLJIT_MEM
)
1314 return emit_op_mem(compiler
, mem_flags
| STORE
, dst_r
, dst
, dstw
, TMP_REG2
);
1315 return SLJIT_SUCCESS
;
1318 flags
= HAS_FLAGS(op_flags
) ? SET_FLAGS
: 0;
1319 mem_flags
= WORD_SIZE
;
1321 if (op_flags
& SLJIT_32
) {
1323 mem_flags
= INT_SIZE
;
1326 if (src
& SLJIT_MEM
) {
1327 FAIL_IF(emit_op_mem(compiler
, mem_flags
, TMP_REG2
, src
, srcw
, TMP_REG2
));
1331 emit_op_imm(compiler
, flags
| op
, dst_r
, TMP_REG1
, src
);
1333 if (SLJIT_UNLIKELY(dst
& SLJIT_MEM
))
1334 return emit_op_mem(compiler
, mem_flags
| STORE
, dst_r
, dst
, dstw
, TMP_REG2
);
1335 return SLJIT_SUCCESS
;
1338 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_op2(struct sljit_compiler
*compiler
, sljit_s32 op
,
1339 sljit_s32 dst
, sljit_sw dstw
,
1340 sljit_s32 src1
, sljit_sw src1w
,
1341 sljit_s32 src2
, sljit_sw src2w
)
1343 sljit_s32 dst_r
, flags
, mem_flags
;
1346 CHECK(check_sljit_emit_op2(compiler
, op
, 0, dst
, dstw
, src1
, src1w
, src2
, src2w
));
1347 ADJUST_LOCAL_OFFSET(dst
, dstw
);
1348 ADJUST_LOCAL_OFFSET(src1
, src1w
);
1349 ADJUST_LOCAL_OFFSET(src2
, src2w
);
1351 dst_r
= FAST_IS_REG(dst
) ? dst
: TMP_REG1
;
1352 flags
= HAS_FLAGS(op
) ? SET_FLAGS
: 0;
1353 mem_flags
= WORD_SIZE
;
1355 if (op
& SLJIT_32
) {
1357 mem_flags
= INT_SIZE
;
1360 if (dst
== TMP_REG1
)
1361 flags
|= UNUSED_RETURN
;
1363 if (src1
& SLJIT_MEM
) {
1364 FAIL_IF(emit_op_mem(compiler
, mem_flags
, TMP_REG1
, src1
, src1w
, TMP_REG1
));
1368 if (src2
& SLJIT_MEM
) {
1369 FAIL_IF(emit_op_mem(compiler
, mem_flags
, TMP_REG2
, src2
, src2w
, TMP_REG2
));
1373 if (src1
& SLJIT_IMM
)
1378 if (src2
& SLJIT_IMM
)
1383 emit_op_imm(compiler
, flags
| GET_OPCODE(op
), dst_r
, src1w
, src2w
);
1385 if (dst
& SLJIT_MEM
)
1386 return emit_op_mem(compiler
, mem_flags
| STORE
, dst_r
, dst
, dstw
, TMP_REG2
);
1387 return SLJIT_SUCCESS
;
1390 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_op2u(struct sljit_compiler
*compiler
, sljit_s32 op
,
1391 sljit_s32 src1
, sljit_sw src1w
,
1392 sljit_s32 src2
, sljit_sw src2w
)
1395 CHECK(check_sljit_emit_op2(compiler
, op
, 1, 0, 0, src1
, src1w
, src2
, src2w
));
1397 SLJIT_SKIP_CHECKS(compiler
);
1398 return sljit_emit_op2(compiler
, op
, TMP_REG1
, 0, src1
, src1w
, src2
, src2w
);
1401 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_op_src(struct sljit_compiler
*compiler
, sljit_s32 op
,
1402 sljit_s32 src
, sljit_sw srcw
)
1405 CHECK(check_sljit_emit_op_src(compiler
, op
, src
, srcw
));
1406 ADJUST_LOCAL_OFFSET(src
, srcw
);
1409 case SLJIT_FAST_RETURN
:
1410 if (FAST_IS_REG(src
))
1411 FAIL_IF(push_inst(compiler
, ORR
| RD(TMP_LR
) | RN(TMP_ZERO
) | RM(src
)));
1413 FAIL_IF(emit_op_mem(compiler
, WORD_SIZE
, TMP_LR
, src
, srcw
, TMP_REG1
));
1415 return push_inst(compiler
, RET
| RN(TMP_LR
));
1416 case SLJIT_SKIP_FRAMES_BEFORE_FAST_RETURN
:
1417 return SLJIT_SUCCESS
;
1418 case SLJIT_PREFETCH_L1
:
1419 case SLJIT_PREFETCH_L2
:
1420 case SLJIT_PREFETCH_L3
:
1421 case SLJIT_PREFETCH_ONCE
:
1422 SLJIT_ASSERT(reg_map
[1] == 0 && reg_map
[3] == 2 && reg_map
[5] == 4);
1424 /* The reg_map[op] should provide the appropriate constant. */
1425 if (op
== SLJIT_PREFETCH_L1
)
1427 else if (op
== SLJIT_PREFETCH_L2
)
1429 else if (op
== SLJIT_PREFETCH_L3
)
1434 /* Signed word sized load is the prefetch instruction. */
1435 return emit_op_mem(compiler
, WORD_SIZE
| SIGNED
, op
, src
, srcw
, TMP_REG1
);
1438 return SLJIT_SUCCESS
;
1441 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_get_register_index(sljit_s32 reg
)
1443 CHECK_REG_INDEX(check_sljit_get_register_index(reg
));
1444 return reg_map
[reg
];
1447 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_get_float_register_index(sljit_s32 reg
)
1449 CHECK_REG_INDEX(check_sljit_get_float_register_index(reg
));
1450 return freg_map
[reg
];
1453 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_op_custom(struct sljit_compiler
*compiler
,
1454 void *instruction
, sljit_u32 size
)
1456 SLJIT_UNUSED_ARG(size
);
1458 CHECK(check_sljit_emit_op_custom(compiler
, instruction
, size
));
1460 return push_inst(compiler
, *(sljit_ins
*)instruction
);
1463 /* --------------------------------------------------------------------- */
1464 /* Floating point operators */
1465 /* --------------------------------------------------------------------- */
1467 static sljit_s32
emit_fop_mem(struct sljit_compiler
*compiler
, sljit_s32 flags
, sljit_s32 reg
, sljit_s32 arg
, sljit_sw argw
)
1469 sljit_u32 shift
= MEM_SIZE_SHIFT(flags
);
1470 sljit_ins type
= (shift
<< 30);
1472 SLJIT_ASSERT(arg
& SLJIT_MEM
);
1474 if (!(flags
& STORE
))
1477 if (arg
& OFFS_REG_MASK
) {
1479 if (argw
== 0 || argw
== shift
)
1480 return push_inst(compiler
, STR_FR
| type
| VT(reg
)
1481 | RN(arg
& REG_MASK
) | RM(OFFS_REG(arg
)) | (argw
? (1 << 12) : 0));
1483 FAIL_IF(push_inst(compiler
, ADD
| RD(TMP_REG1
) | RN(arg
& REG_MASK
) | RM(OFFS_REG(arg
)) | ((sljit_ins
)argw
<< 10)));
1484 return push_inst(compiler
, STR_FI
| type
| VT(reg
) | RN(TMP_REG1
));
1490 FAIL_IF(load_immediate(compiler
, TMP_REG1
, argw
& ~(0xfff << shift
)));
1492 argw
= (argw
>> shift
) & 0xfff;
1494 return push_inst(compiler
, STR_FI
| type
| VT(reg
) | RN(TMP_REG1
) | ((sljit_ins
)argw
<< 10));
1497 if (argw
>= 0 && (argw
& ((1 << shift
) - 1)) == 0) {
1498 if ((argw
>> shift
) <= 0xfff)
1499 return push_inst(compiler
, STR_FI
| type
| VT(reg
) | RN(arg
) | ((sljit_ins
)argw
<< (10 - shift
)));
1501 if (argw
<= 0xffffff) {
1502 FAIL_IF(push_inst(compiler
, ADDI
| (1 << 22) | RD(TMP_REG1
) | RN(arg
) | (((sljit_ins
)argw
>> 12) << 10)));
1504 argw
= ((argw
& 0xfff) >> shift
);
1505 return push_inst(compiler
, STR_FI
| type
| VT(reg
) | RN(TMP_REG1
) | ((sljit_ins
)argw
<< 10));
1509 if (argw
<= 255 && argw
>= -256)
1510 return push_inst(compiler
, STUR_FI
| type
| VT(reg
) | RN(arg
) | (((sljit_ins
)argw
& 0x1ff) << 12));
1512 FAIL_IF(load_immediate(compiler
, TMP_REG1
, argw
));
1513 return push_inst(compiler
, STR_FR
| type
| VT(reg
) | RN(arg
) | RM(TMP_REG1
));
1516 static SLJIT_INLINE sljit_s32
sljit_emit_fop1_conv_sw_from_f64(struct sljit_compiler
*compiler
, sljit_s32 op
,
1517 sljit_s32 dst
, sljit_sw dstw
,
1518 sljit_s32 src
, sljit_sw srcw
)
1520 sljit_s32 dst_r
= FAST_IS_REG(dst
) ? dst
: TMP_REG1
;
1521 sljit_ins inv_bits
= (op
& SLJIT_32
) ? (1 << 22) : 0;
1523 if (GET_OPCODE(op
) == SLJIT_CONV_S32_FROM_F64
)
1526 if (src
& SLJIT_MEM
) {
1527 emit_fop_mem(compiler
, (op
& SLJIT_32
) ? INT_SIZE
: WORD_SIZE
, TMP_FREG1
, src
, srcw
);
1531 FAIL_IF(push_inst(compiler
, (FCVTZS
^ inv_bits
) | RD(dst_r
) | VN(src
)));
1533 if (dst
& SLJIT_MEM
)
1534 return emit_op_mem(compiler
, ((GET_OPCODE(op
) == SLJIT_CONV_S32_FROM_F64
) ? INT_SIZE
: WORD_SIZE
) | STORE
, TMP_REG1
, dst
, dstw
, TMP_REG2
);
1535 return SLJIT_SUCCESS
;
1538 static SLJIT_INLINE sljit_s32
sljit_emit_fop1_conv_f64_from_sw(struct sljit_compiler
*compiler
, sljit_s32 op
,
1539 sljit_s32 dst
, sljit_sw dstw
,
1540 sljit_s32 src
, sljit_sw srcw
)
1542 sljit_s32 dst_r
= FAST_IS_REG(dst
) ? dst
: TMP_FREG1
;
1543 sljit_ins inv_bits
= (op
& SLJIT_32
) ? (1 << 22) : 0;
1545 if (GET_OPCODE(op
) == SLJIT_CONV_F64_FROM_S32
)
1548 if (src
& SLJIT_MEM
) {
1549 emit_op_mem(compiler
, ((GET_OPCODE(op
) == SLJIT_CONV_F64_FROM_S32
) ? INT_SIZE
: WORD_SIZE
), TMP_REG1
, src
, srcw
, TMP_REG1
);
1551 } else if (src
& SLJIT_IMM
) {
1552 if (GET_OPCODE(op
) == SLJIT_CONV_F64_FROM_S32
)
1553 srcw
= (sljit_s32
)srcw
;
1555 FAIL_IF(load_immediate(compiler
, TMP_REG1
, srcw
));
1559 FAIL_IF(push_inst(compiler
, (SCVTF
^ inv_bits
) | VD(dst_r
) | RN(src
)));
1561 if (dst
& SLJIT_MEM
)
1562 return emit_fop_mem(compiler
, ((op
& SLJIT_32
) ? INT_SIZE
: WORD_SIZE
) | STORE
, TMP_FREG1
, dst
, dstw
);
1563 return SLJIT_SUCCESS
;
1566 static SLJIT_INLINE sljit_s32
sljit_emit_fop1_cmp(struct sljit_compiler
*compiler
, sljit_s32 op
,
1567 sljit_s32 src1
, sljit_sw src1w
,
1568 sljit_s32 src2
, sljit_sw src2w
)
1570 sljit_s32 mem_flags
= (op
& SLJIT_32
) ? INT_SIZE
: WORD_SIZE
;
1571 sljit_ins inv_bits
= (op
& SLJIT_32
) ? (1 << 22) : 0;
1573 if (src1
& SLJIT_MEM
) {
1574 emit_fop_mem(compiler
, mem_flags
, TMP_FREG1
, src1
, src1w
);
1578 if (src2
& SLJIT_MEM
) {
1579 emit_fop_mem(compiler
, mem_flags
, TMP_FREG2
, src2
, src2w
);
1583 return push_inst(compiler
, (FCMP
^ inv_bits
) | VN(src1
) | VM(src2
));
1586 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_fop1(struct sljit_compiler
*compiler
, sljit_s32 op
,
1587 sljit_s32 dst
, sljit_sw dstw
,
1588 sljit_s32 src
, sljit_sw srcw
)
1590 sljit_s32 dst_r
, mem_flags
= (op
& SLJIT_32
) ? INT_SIZE
: WORD_SIZE
;
1595 SLJIT_COMPILE_ASSERT((INT_SIZE
^ 0x1) == WORD_SIZE
, must_be_one_bit_difference
);
1596 SELECT_FOP1_OPERATION_WITH_CHECKS(compiler
, op
, dst
, dstw
, src
, srcw
);
1598 inv_bits
= (op
& SLJIT_32
) ? (1 << 22) : 0;
1599 dst_r
= FAST_IS_REG(dst
) ? dst
: TMP_FREG1
;
1601 if (src
& SLJIT_MEM
) {
1602 emit_fop_mem(compiler
, (GET_OPCODE(op
) == SLJIT_CONV_F64_FROM_F32
) ? (mem_flags
^ 0x1) : mem_flags
, dst_r
, src
, srcw
);
1606 switch (GET_OPCODE(op
)) {
1609 if (dst_r
!= TMP_FREG1
)
1610 FAIL_IF(push_inst(compiler
, (FMOV
^ inv_bits
) | VD(dst_r
) | VN(src
)));
1616 FAIL_IF(push_inst(compiler
, (FNEG
^ inv_bits
) | VD(dst_r
) | VN(src
)));
1619 FAIL_IF(push_inst(compiler
, (FABS
^ inv_bits
) | VD(dst_r
) | VN(src
)));
1621 case SLJIT_CONV_F64_FROM_F32
:
1622 FAIL_IF(push_inst(compiler
, FCVT
| (sljit_ins
)((op
& SLJIT_32
) ? (1 << 22) : (1 << 15)) | VD(dst_r
) | VN(src
)));
1626 if (dst
& SLJIT_MEM
)
1627 return emit_fop_mem(compiler
, mem_flags
| STORE
, dst_r
, dst
, dstw
);
1628 return SLJIT_SUCCESS
;
1631 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_fop2(struct sljit_compiler
*compiler
, sljit_s32 op
,
1632 sljit_s32 dst
, sljit_sw dstw
,
1633 sljit_s32 src1
, sljit_sw src1w
,
1634 sljit_s32 src2
, sljit_sw src2w
)
1636 sljit_s32 dst_r
, mem_flags
= (op
& SLJIT_32
) ? INT_SIZE
: WORD_SIZE
;
1637 sljit_ins inv_bits
= (op
& SLJIT_32
) ? (1 << 22) : 0;
1640 CHECK(check_sljit_emit_fop2(compiler
, op
, dst
, dstw
, src1
, src1w
, src2
, src2w
));
1641 ADJUST_LOCAL_OFFSET(dst
, dstw
);
1642 ADJUST_LOCAL_OFFSET(src1
, src1w
);
1643 ADJUST_LOCAL_OFFSET(src2
, src2w
);
1645 dst_r
= FAST_IS_REG(dst
) ? dst
: TMP_FREG1
;
1646 if (src1
& SLJIT_MEM
) {
1647 emit_fop_mem(compiler
, mem_flags
, TMP_FREG1
, src1
, src1w
);
1650 if (src2
& SLJIT_MEM
) {
1651 emit_fop_mem(compiler
, mem_flags
, TMP_FREG2
, src2
, src2w
);
1655 switch (GET_OPCODE(op
)) {
1657 FAIL_IF(push_inst(compiler
, (FADD
^ inv_bits
) | VD(dst_r
) | VN(src1
) | VM(src2
)));
1660 FAIL_IF(push_inst(compiler
, (FSUB
^ inv_bits
) | VD(dst_r
) | VN(src1
) | VM(src2
)));
1663 FAIL_IF(push_inst(compiler
, (FMUL
^ inv_bits
) | VD(dst_r
) | VN(src1
) | VM(src2
)));
1666 FAIL_IF(push_inst(compiler
, (FDIV
^ inv_bits
) | VD(dst_r
) | VN(src1
) | VM(src2
)));
1670 if (!(dst
& SLJIT_MEM
))
1671 return SLJIT_SUCCESS
;
1672 return emit_fop_mem(compiler
, mem_flags
| STORE
, TMP_FREG1
, dst
, dstw
);
1675 /* --------------------------------------------------------------------- */
1676 /* Other instructions */
1677 /* --------------------------------------------------------------------- */
1679 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_fast_enter(struct sljit_compiler
*compiler
, sljit_s32 dst
, sljit_sw dstw
)
1682 CHECK(check_sljit_emit_fast_enter(compiler
, dst
, dstw
));
1683 ADJUST_LOCAL_OFFSET(dst
, dstw
);
1685 if (FAST_IS_REG(dst
))
1686 return push_inst(compiler
, ORR
| RD(dst
) | RN(TMP_ZERO
) | RM(TMP_LR
));
1689 return emit_op_mem(compiler
, WORD_SIZE
| STORE
, TMP_LR
, dst
, dstw
, TMP_REG1
);
1692 /* --------------------------------------------------------------------- */
1693 /* Conditional instructions */
1694 /* --------------------------------------------------------------------- */
1696 static sljit_ins
get_cc(struct sljit_compiler
*compiler
, sljit_s32 type
)
1701 case SLJIT_ORDERED_EQUAL
:
1702 case SLJIT_UNORDERED_OR_EQUAL
: /* Not supported. */
1705 case SLJIT_NOT_EQUAL
:
1706 case SLJIT_F_NOT_EQUAL
:
1707 case SLJIT_UNORDERED_OR_NOT_EQUAL
:
1708 case SLJIT_ORDERED_NOT_EQUAL
: /* Not supported. */
1712 if (compiler
->status_flags_state
& SLJIT_CURRENT_FLAGS_ADD
)
1719 case SLJIT_NOT_CARRY
:
1720 if (compiler
->status_flags_state
& SLJIT_CURRENT_FLAGS_ADD
)
1724 case SLJIT_GREATER_EQUAL
:
1728 case SLJIT_UNORDERED_OR_GREATER
:
1731 case SLJIT_LESS_EQUAL
:
1732 case SLJIT_F_LESS_EQUAL
:
1733 case SLJIT_ORDERED_LESS_EQUAL
:
1736 case SLJIT_SIG_LESS
:
1737 case SLJIT_UNORDERED_OR_LESS
:
1740 case SLJIT_SIG_GREATER_EQUAL
:
1741 case SLJIT_F_GREATER_EQUAL
:
1742 case SLJIT_ORDERED_GREATER_EQUAL
:
1745 case SLJIT_SIG_GREATER
:
1746 case SLJIT_F_GREATER
:
1747 case SLJIT_ORDERED_GREATER
:
1750 case SLJIT_SIG_LESS_EQUAL
:
1751 case SLJIT_UNORDERED_OR_LESS_EQUAL
:
1754 case SLJIT_OVERFLOW
:
1755 if (!(compiler
->status_flags_state
& (SLJIT_CURRENT_FLAGS_ADD
| SLJIT_CURRENT_FLAGS_SUB
)))
1759 case SLJIT_UNORDERED
:
1762 case SLJIT_NOT_OVERFLOW
:
1763 if (!(compiler
->status_flags_state
& (SLJIT_CURRENT_FLAGS_ADD
| SLJIT_CURRENT_FLAGS_SUB
)))
1771 case SLJIT_ORDERED_LESS
:
1774 case SLJIT_UNORDERED_OR_GREATER_EQUAL
:
1778 SLJIT_UNREACHABLE();
1783 SLJIT_API_FUNC_ATTRIBUTE
struct sljit_label
* sljit_emit_label(struct sljit_compiler
*compiler
)
1785 struct sljit_label
*label
;
1788 CHECK_PTR(check_sljit_emit_label(compiler
));
1790 if (compiler
->last_label
&& compiler
->last_label
->size
== compiler
->size
)
1791 return compiler
->last_label
;
1793 label
= (struct sljit_label
*)ensure_abuf(compiler
, sizeof(struct sljit_label
));
1794 PTR_FAIL_IF(!label
);
1795 set_label(label
, compiler
);
1799 SLJIT_API_FUNC_ATTRIBUTE
struct sljit_jump
* sljit_emit_jump(struct sljit_compiler
*compiler
, sljit_s32 type
)
1801 struct sljit_jump
*jump
;
1804 CHECK_PTR(check_sljit_emit_jump(compiler
, type
));
1806 jump
= (struct sljit_jump
*)ensure_abuf(compiler
, sizeof(struct sljit_jump
));
1808 set_jump(jump
, compiler
, type
& SLJIT_REWRITABLE_JUMP
);
1811 if (type
< SLJIT_JUMP
) {
1812 jump
->flags
|= IS_COND
;
1813 PTR_FAIL_IF(push_inst(compiler
, B_CC
| (6 << 5) | get_cc(compiler
, type
)));
1815 else if (type
>= SLJIT_FAST_CALL
)
1816 jump
->flags
|= IS_BL
;
1818 PTR_FAIL_IF(emit_imm64_const(compiler
, TMP_REG1
, 0));
1819 jump
->addr
= compiler
->size
;
1820 PTR_FAIL_IF(push_inst(compiler
, ((type
>= SLJIT_FAST_CALL
) ? BLR
: BR
) | RN(TMP_REG1
)));
1825 SLJIT_API_FUNC_ATTRIBUTE
struct sljit_jump
* sljit_emit_call(struct sljit_compiler
*compiler
, sljit_s32 type
,
1826 sljit_s32 arg_types
)
1828 SLJIT_UNUSED_ARG(arg_types
);
1830 CHECK_PTR(check_sljit_emit_call(compiler
, type
, arg_types
));
1832 if (type
& SLJIT_CALL_RETURN
) {
1833 PTR_FAIL_IF(emit_stack_frame_release(compiler
));
1834 type
= SLJIT_JUMP
| (type
& SLJIT_REWRITABLE_JUMP
);
1837 SLJIT_SKIP_CHECKS(compiler
);
1838 return sljit_emit_jump(compiler
, type
);
1841 static SLJIT_INLINE
struct sljit_jump
* emit_cmp_to0(struct sljit_compiler
*compiler
, sljit_s32 type
,
1842 sljit_s32 src
, sljit_sw srcw
)
1844 struct sljit_jump
*jump
;
1845 sljit_ins inv_bits
= (type
& SLJIT_32
) ? W_OP
: 0;
1847 SLJIT_ASSERT((type
& 0xff) == SLJIT_EQUAL
|| (type
& 0xff) == SLJIT_NOT_EQUAL
);
1848 ADJUST_LOCAL_OFFSET(src
, srcw
);
1850 jump
= (struct sljit_jump
*)ensure_abuf(compiler
, sizeof(struct sljit_jump
));
1852 set_jump(jump
, compiler
, type
& SLJIT_REWRITABLE_JUMP
);
1853 jump
->flags
|= IS_CBZ
| IS_COND
;
1855 if (src
& SLJIT_MEM
) {
1856 PTR_FAIL_IF(emit_op_mem(compiler
, inv_bits
? INT_SIZE
: WORD_SIZE
, TMP_REG1
, src
, srcw
, TMP_REG1
));
1859 else if (src
& SLJIT_IMM
) {
1860 PTR_FAIL_IF(load_immediate(compiler
, TMP_REG1
, srcw
));
1864 SLJIT_ASSERT(FAST_IS_REG(src
));
1866 if ((type
& 0xff) == SLJIT_EQUAL
)
1867 inv_bits
|= 1 << 24;
1869 PTR_FAIL_IF(push_inst(compiler
, (CBZ
^ inv_bits
) | (6 << 5) | RT(src
)));
1870 PTR_FAIL_IF(emit_imm64_const(compiler
, TMP_REG1
, 0));
1871 jump
->addr
= compiler
->size
;
1872 PTR_FAIL_IF(push_inst(compiler
, BR
| RN(TMP_REG1
)));
1876 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_ijump(struct sljit_compiler
*compiler
, sljit_s32 type
, sljit_s32 src
, sljit_sw srcw
)
1878 struct sljit_jump
*jump
;
1881 CHECK(check_sljit_emit_ijump(compiler
, type
, src
, srcw
));
1882 ADJUST_LOCAL_OFFSET(src
, srcw
);
1884 if (!(src
& SLJIT_IMM
)) {
1885 if (src
& SLJIT_MEM
) {
1886 FAIL_IF(emit_op_mem(compiler
, WORD_SIZE
, TMP_REG1
, src
, srcw
, TMP_REG1
));
1889 return push_inst(compiler
, ((type
>= SLJIT_FAST_CALL
) ? BLR
: BR
) | RN(src
));
1892 /* These jumps are converted to jump/call instructions when possible. */
1893 jump
= (struct sljit_jump
*)ensure_abuf(compiler
, sizeof(struct sljit_jump
));
1895 set_jump(jump
, compiler
, JUMP_ADDR
| ((type
>= SLJIT_FAST_CALL
) ? IS_BL
: 0));
1896 jump
->u
.target
= (sljit_uw
)srcw
;
1898 FAIL_IF(emit_imm64_const(compiler
, TMP_REG1
, 0));
1899 jump
->addr
= compiler
->size
;
1900 return push_inst(compiler
, ((type
>= SLJIT_FAST_CALL
) ? BLR
: BR
) | RN(TMP_REG1
));
1903 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_icall(struct sljit_compiler
*compiler
, sljit_s32 type
,
1904 sljit_s32 arg_types
,
1905 sljit_s32 src
, sljit_sw srcw
)
1907 SLJIT_UNUSED_ARG(arg_types
);
1909 CHECK(check_sljit_emit_icall(compiler
, type
, arg_types
, src
, srcw
));
1910 ADJUST_LOCAL_OFFSET(src
, srcw
);
1912 if (src
& SLJIT_MEM
) {
1913 FAIL_IF(emit_op_mem(compiler
, WORD_SIZE
, TMP_REG1
, src
, srcw
, TMP_REG1
));
1917 if (type
& SLJIT_CALL_RETURN
) {
1918 if (src
>= SLJIT_FIRST_SAVED_REG
&& src
<= SLJIT_S0
) {
1919 FAIL_IF(push_inst(compiler
, ORR
| RD(TMP_REG1
) | RN(TMP_ZERO
) | RM(src
)));
1923 FAIL_IF(emit_stack_frame_release(compiler
));
1927 SLJIT_SKIP_CHECKS(compiler
);
1928 return sljit_emit_ijump(compiler
, type
, src
, srcw
);
1931 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_op_flags(struct sljit_compiler
*compiler
, sljit_s32 op
,
1932 sljit_s32 dst
, sljit_sw dstw
,
1935 sljit_s32 dst_r
, src_r
, flags
, mem_flags
;
1939 CHECK(check_sljit_emit_op_flags(compiler
, op
, dst
, dstw
, type
));
1940 ADJUST_LOCAL_OFFSET(dst
, dstw
);
1942 cc
= get_cc(compiler
, type
);
1943 dst_r
= FAST_IS_REG(dst
) ? dst
: TMP_REG1
;
1945 if (GET_OPCODE(op
) < SLJIT_ADD
) {
1946 FAIL_IF(push_inst(compiler
, CSINC
| (cc
<< 12) | RD(dst_r
) | RN(TMP_ZERO
) | RM(TMP_ZERO
)));
1948 if (dst_r
== TMP_REG1
) {
1949 mem_flags
= (GET_OPCODE(op
) == SLJIT_MOV
? WORD_SIZE
: INT_SIZE
) | STORE
;
1950 return emit_op_mem(compiler
, mem_flags
, TMP_REG1
, dst
, dstw
, TMP_REG2
);
1953 return SLJIT_SUCCESS
;
1956 flags
= HAS_FLAGS(op
) ? SET_FLAGS
: 0;
1957 mem_flags
= WORD_SIZE
;
1959 if (op
& SLJIT_32
) {
1961 mem_flags
= INT_SIZE
;
1966 if (dst
& SLJIT_MEM
) {
1967 FAIL_IF(emit_op_mem(compiler
, mem_flags
, TMP_REG1
, dst
, dstw
, TMP_REG1
));
1971 FAIL_IF(push_inst(compiler
, CSINC
| (cc
<< 12) | RD(TMP_REG2
) | RN(TMP_ZERO
) | RM(TMP_ZERO
)));
1972 emit_op_imm(compiler
, flags
| GET_OPCODE(op
), dst_r
, src_r
, TMP_REG2
);
1974 if (dst
& SLJIT_MEM
)
1975 return emit_op_mem(compiler
, mem_flags
| STORE
, TMP_REG1
, dst
, dstw
, TMP_REG2
);
1976 return SLJIT_SUCCESS
;
1979 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_cmov(struct sljit_compiler
*compiler
, sljit_s32 type
,
1981 sljit_s32 src
, sljit_sw srcw
)
1983 sljit_ins inv_bits
= (dst_reg
& SLJIT_32
) ? W_OP
: 0;
1987 CHECK(check_sljit_emit_cmov(compiler
, type
, dst_reg
, src
, srcw
));
1989 if (SLJIT_UNLIKELY(src
& SLJIT_IMM
)) {
1990 if (dst_reg
& SLJIT_32
)
1991 srcw
= (sljit_s32
)srcw
;
1992 FAIL_IF(load_immediate(compiler
, TMP_REG1
, srcw
));
1997 cc
= get_cc(compiler
, type
);
1998 dst_reg
&= ~SLJIT_32
;
2000 return push_inst(compiler
, (CSEL
^ inv_bits
) | (cc
<< 12) | RD(dst_reg
) | RN(dst_reg
) | RM(src
));
2003 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_mem(struct sljit_compiler
*compiler
, sljit_s32 type
,
2005 sljit_s32 mem
, sljit_sw memw
)
2007 sljit_u32 sign
= 0, inst
;
2010 CHECK(check_sljit_emit_mem(compiler
, type
, reg
, mem
, memw
));
2012 if (type
& SLJIT_MEM_UNALIGNED
)
2013 return sljit_emit_mem_unaligned(compiler
, type
, reg
, mem
, memw
);
2015 if ((mem
& OFFS_REG_MASK
) || (memw
> 255 || memw
< -256))
2016 return SLJIT_ERR_UNSUPPORTED
;
2018 if (type
& SLJIT_MEM_SUPP
)
2019 return SLJIT_SUCCESS
;
2021 switch (type
& 0xff) {
2024 inst
= STURBI
| (MEM_SIZE_SHIFT(WORD_SIZE
) << 30) | 0x400;
2030 inst
= STURBI
| (MEM_SIZE_SHIFT(BYTE_SIZE
) << 30) | 0x400;
2036 inst
= STURBI
| (MEM_SIZE_SHIFT(HALF_SIZE
) << 30) | 0x400;
2043 inst
= STURBI
| (MEM_SIZE_SHIFT(INT_SIZE
) << 30) | 0x400;
2046 SLJIT_UNREACHABLE();
2047 inst
= STURBI
| (MEM_SIZE_SHIFT(WORD_SIZE
) << 30) | 0x400;
2051 if (!(type
& SLJIT_MEM_STORE
))
2052 inst
|= sign
? 0x00800000 : 0x00400000;
2054 if (type
& SLJIT_MEM_PRE
)
2057 return push_inst(compiler
, inst
| RT(reg
) | RN(mem
& REG_MASK
) | (sljit_ins
)((memw
& 0x1ff) << 12));
2060 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_emit_fmem(struct sljit_compiler
*compiler
, sljit_s32 type
,
2062 sljit_s32 mem
, sljit_sw memw
)
2067 CHECK(check_sljit_emit_fmem(compiler
, type
, freg
, mem
, memw
));
2069 if (type
& SLJIT_MEM_UNALIGNED
)
2070 return sljit_emit_fmem_unaligned(compiler
, type
, freg
, mem
, memw
);
2072 if ((mem
& OFFS_REG_MASK
) || (memw
> 255 || memw
< -256))
2073 return SLJIT_ERR_UNSUPPORTED
;
2075 if (type
& SLJIT_MEM_SUPP
)
2076 return SLJIT_SUCCESS
;
2078 inst
= STUR_FI
| 0x80000400;
2080 if (!(type
& SLJIT_32
))
2083 if (!(type
& SLJIT_MEM_STORE
))
2086 if (type
& SLJIT_MEM_PRE
)
2089 return push_inst(compiler
, inst
| VT(freg
) | RN(mem
& REG_MASK
) | (sljit_ins
)((memw
& 0x1ff) << 12));
2092 SLJIT_API_FUNC_ATTRIBUTE sljit_s32
sljit_get_local_base(struct sljit_compiler
*compiler
, sljit_s32 dst
, sljit_sw dstw
, sljit_sw offset
)
2098 CHECK(check_sljit_get_local_base(compiler
, dst
, dstw
, offset
));
2099 ADJUST_LOCAL_OFFSET(SLJIT_MEM1(SLJIT_SP
), offset
);
2101 dst_reg
= FAST_IS_REG(dst
) ? dst
: TMP_REG1
;
2103 /* Not all instruction forms support accessing SP register. */
2104 if (offset
<= 0xffffff && offset
>= -0xffffff) {
2111 if (offset
<= 0xfff)
2112 FAIL_IF(push_inst(compiler
, ins
| RD(dst_reg
) | RN(SLJIT_SP
) | (sljit_ins
)(offset
<< 10)));
2114 FAIL_IF(push_inst(compiler
, ins
| RD(dst_reg
) | RN(SLJIT_SP
) | (sljit_ins
)((offset
& 0xfff000) >> (12 - 10)) | (1 << 22)));
2118 FAIL_IF(push_inst(compiler
, ins
| RD(dst_reg
) | RN(dst_reg
) | (sljit_ins
)(offset
<< 10)));
2122 FAIL_IF(load_immediate (compiler
, dst_reg
, offset
));
2123 /* Add extended register form. */
2124 FAIL_IF(push_inst(compiler
, ADDE
| (0x3 << 13) | RD(dst_reg
) | RN(SLJIT_SP
) | RM(dst_reg
)));
2127 if (SLJIT_UNLIKELY(dst
& SLJIT_MEM
))
2128 return emit_op_mem(compiler
, WORD_SIZE
| STORE
, dst_reg
, dst
, dstw
, TMP_REG1
);
2129 return SLJIT_SUCCESS
;
2132 SLJIT_API_FUNC_ATTRIBUTE
struct sljit_const
* sljit_emit_const(struct sljit_compiler
*compiler
, sljit_s32 dst
, sljit_sw dstw
, sljit_sw init_value
)
2134 struct sljit_const
*const_
;
2138 CHECK_PTR(check_sljit_emit_const(compiler
, dst
, dstw
, init_value
));
2139 ADJUST_LOCAL_OFFSET(dst
, dstw
);
2141 const_
= (struct sljit_const
*)ensure_abuf(compiler
, sizeof(struct sljit_const
));
2142 PTR_FAIL_IF(!const_
);
2143 set_const(const_
, compiler
);
2145 dst_r
= FAST_IS_REG(dst
) ? dst
: TMP_REG1
;
2146 PTR_FAIL_IF(emit_imm64_const(compiler
, dst_r
, (sljit_uw
)init_value
));
2148 if (dst
& SLJIT_MEM
)
2149 PTR_FAIL_IF(emit_op_mem(compiler
, WORD_SIZE
| STORE
, dst_r
, dst
, dstw
, TMP_REG2
));
2153 SLJIT_API_FUNC_ATTRIBUTE
struct sljit_put_label
* sljit_emit_put_label(struct sljit_compiler
*compiler
, sljit_s32 dst
, sljit_sw dstw
)
2155 struct sljit_put_label
*put_label
;
2159 CHECK_PTR(check_sljit_emit_put_label(compiler
, dst
, dstw
));
2160 ADJUST_LOCAL_OFFSET(dst
, dstw
);
2162 dst_r
= FAST_IS_REG(dst
) ? dst
: TMP_REG1
;
2163 PTR_FAIL_IF(emit_imm64_const(compiler
, dst_r
, 0));
2165 put_label
= (struct sljit_put_label
*)ensure_abuf(compiler
, sizeof(struct sljit_put_label
));
2166 PTR_FAIL_IF(!put_label
);
2167 set_put_label(put_label
, compiler
, 1);
2169 if (dst
& SLJIT_MEM
)
2170 PTR_FAIL_IF(emit_op_mem(compiler
, WORD_SIZE
| STORE
, dst_r
, dst
, dstw
, TMP_REG2
));
2175 SLJIT_API_FUNC_ATTRIBUTE
void sljit_set_jump_addr(sljit_uw addr
, sljit_uw new_target
, sljit_sw executable_offset
)
2177 sljit_ins
* inst
= (sljit_ins
*)addr
;
2179 SLJIT_UNUSED_ARG(executable_offset
);
2181 SLJIT_UPDATE_WX_FLAGS(inst
, inst
+ 4, 0);
2183 dst
= inst
[0] & 0x1f;
2184 SLJIT_ASSERT((inst
[0] & 0xffe00000) == MOVZ
&& (inst
[1] & 0xffe00000) == (MOVK
| (1 << 21)));
2185 inst
[0] = MOVZ
| dst
| (((sljit_u32
)new_target
& 0xffff) << 5);
2186 inst
[1] = MOVK
| dst
| (((sljit_u32
)(new_target
>> 16) & 0xffff) << 5) | (1 << 21);
2187 inst
[2] = MOVK
| dst
| (((sljit_u32
)(new_target
>> 32) & 0xffff) << 5) | (2 << 21);
2188 inst
[3] = MOVK
| dst
| ((sljit_u32
)(new_target
>> 48) << 5) | (3 << 21);
2190 SLJIT_UPDATE_WX_FLAGS(inst
, inst
+ 4, 1);
2191 inst
= (sljit_ins
*)SLJIT_ADD_EXEC_OFFSET(inst
, executable_offset
);
2192 SLJIT_CACHE_FLUSH(inst
, inst
+ 4);
2195 SLJIT_API_FUNC_ATTRIBUTE
void sljit_set_const(sljit_uw addr
, sljit_sw new_constant
, sljit_sw executable_offset
)
2197 sljit_set_jump_addr(addr
, (sljit_uw
)new_constant
, executable_offset
);