2 * Copyright (C) 2024 Mikulas Patocka
4 * This file is part of Ajla.
6 * Ajla is free software: you can redistribute it and/or modify it under the
7 * terms of the GNU General Public License as published by the Free Software
8 * Foundation, either version 3 of the License, or (at your option) any later
11 * Ajla is distributed in the hope that it will be useful, but WITHOUT ANY
12 * WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR
13 * A PARTICULAR PURPOSE. See the GNU General Public License for more details.
15 * You should have received a copy of the GNU General Public License along with
16 * Ajla. If not, see <https://www.gnu.org/licenses/>.
19 #define OP_SIZE_NATIVE OP_SIZE_8
20 #define OP_SIZE_ADDRESS OP_SIZE_NATIVE
22 #define JMP_LIMIT JMP_LONG
24 #define UNALIGNED_TRAP (!cpu_test_feature(CPU_FEATURE_unaligned))
26 #define ALU_WRITES_FLAGS(alu, im) 0
27 #define ALU1_WRITES_FLAGS(alu) 0
28 #define ROT_WRITES_FLAGS(alu) 0
29 #define COND_IS_LOGICAL(cond) 0
31 #define ARCH_PARTIAL_ALU(size) 0
32 #define ARCH_IS_3ADDRESS 1
33 #define ARCH_HAS_FLAGS 0
34 #define ARCH_PREFERS_SX(size) 0
35 #define ARCH_HAS_BWX 1
36 #define ARCH_HAS_MUL 1
37 #define ARCH_HAS_DIV 1
38 #define ARCH_HAS_ANDN 1
39 #define ARCH_HAS_SHIFTED_ADD(bits) 0
40 #define ARCH_HAS_BTX(btx, size, cnst) (((btx) == BTX_BTR || (btx) == BTX_BTEXT) && (cnst))
41 #define ARCH_SHIFT_SIZE OP_SIZE_4
42 #define ARCH_NEEDS_BARRIER 0
44 #define i_size(size) OP_SIZE_NATIVE
45 #define i_size_rot(size) maximum(size, OP_SIZE_4)
68 #define R_RESERVED 0x15
114 #define R_UPCALL R_S1
115 #define R_TIMESTAMP R_S2
117 #define R_SCRATCH_1 R_A0
118 #define R_SCRATCH_2 R_A1
119 #define R_SCRATCH_3 R_A2
120 #define R_SCRATCH_4 R_A3
121 #define R_SCRATCH_NA_1 R_A4
122 #define R_SCRATCH_NA_2 R_A5
123 #define R_SCRATCH_NA_3 R_A6
125 #define R_SAVED_1 R_S3
126 #define R_SAVED_2 R_S4
135 #define R_OFFSET_IMM R_T0
136 #define R_CONST_IMM R_T1
137 #define R_CMP_RESULT R_T2
139 #define FR_SCRATCH_1 R_FA0
140 #define FR_SCRATCH_2 R_FA1
142 #define SUPPORTED_FP 0x6
144 #define FRAME_SIZE 0x30
146 static bool reg_is_fp(unsigned reg)
148 return reg >= 0x20 && reg < 0x40;
151 static bool attr_w gen_load_constant(struct codegen_context *ctx, unsigned reg, uint64_t c)
153 uint64_t c0 = c & 0x0000000000000fffULL;
154 uint64_t c1 = c & 0x00000000fffff000ULL;
155 uint64_t c2 = c & 0x000fffff00000000ULL;
156 uint64_t c3 = c & 0xfff0000000000000ULL;
157 uint64_t top_bits = 0;
158 if (!(c0 | c1 | c2)) {
159 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
165 if (c0 & 0x800ULL && c1 == 0xfffff000ULL) {
166 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
169 gen_eight(c0 | 0xfffffffffffff000ULL);
170 top_bits = 0xffffffff00000000ULL;
172 bool have_reg = false;
174 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
177 gen_eight((uint64_t)(int32_t)c1);
178 top_bits = (uint64_t)(int32_t)c1 & 0xffffffff00000000ULL;
181 if (!have_reg || c0) {
183 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
188 gen_insn(INSN_ALU, OP_SIZE_NATIVE, ALU_OR, 0);
196 if (top_bits != (c2 | c3)) {
198 if (c2 & 0x0008000000000000ULL)
199 c2x |= 0xfff0000000000000ULL;
200 if (top_bits != c2x) {
201 gen_insn(INSN_MOV_MASK, OP_SIZE_NATIVE, MOV_MASK_32_64, 0);
205 gen_eight(c2x >> 32);
207 top_bits = c2x & 0xfff0000000000000ULL;
208 if (top_bits != c3) {
209 gen_insn(INSN_MOV_MASK, OP_SIZE_NATIVE, MOV_MASK_52_64, 0);
219 static bool attr_w gen_address(struct codegen_context *ctx, unsigned base, int64_t imm, unsigned purpose, unsigned size)
221 ctx->base_reg = base;
222 ctx->offset_imm = imm;
223 ctx->offset_reg = false;
225 case IMM_PURPOSE_LDR_OFFSET:
226 case IMM_PURPOSE_LDR_SX_OFFSET:
227 case IMM_PURPOSE_STR_OFFSET:
228 case IMM_PURPOSE_VLDR_VSTR_OFFSET:
229 case IMM_PURPOSE_MVI_CLI_OFFSET:
230 if (likely(imm >= -0x800) && likely(imm < 0x800)) {
233 if (imm >= -0x8000 && imm < 0x8000 && !(imm & 3)) {
234 if (size == OP_SIZE_NATIVE)
236 if (purpose == IMM_PURPOSE_LDR_SX_OFFSET && size == OP_SIZE_4)
238 if (purpose == IMM_PURPOSE_STR_OFFSET && size == OP_SIZE_4)
243 internal(file_line, "gen_address: invalid purpose %u (imm %"PRIxMAX", size %u)", purpose, (uintmax_t)imm, size);
245 g(gen_load_constant(ctx, R_OFFSET_IMM, imm));
246 ctx->offset_reg = true;
250 static bool is_direct_const(int64_t imm, unsigned purpose, unsigned size)
253 case IMM_PURPOSE_STORE_VALUE:
257 case IMM_PURPOSE_ADD:
258 case IMM_PURPOSE_CMP:
259 if (likely(imm >= -0x800) && likely(imm < 0x800))
262 case IMM_PURPOSE_SUB:
263 if (likely(imm > -0x800) && likely(imm <= 0x800))
266 case IMM_PURPOSE_AND:
268 case IMM_PURPOSE_XOR:
269 if (likely(imm >= 0) && likely(imm < 0x1000))
272 case IMM_PURPOSE_ANDN:
274 case IMM_PURPOSE_TEST:
276 case IMM_PURPOSE_JMP_2REGS:
278 case IMM_PURPOSE_MUL:
280 case IMM_PURPOSE_BITWISE:
283 internal(file_line, "is_direct_const: invalid purpose %u (imm %"PRIxMAX", size %u)", purpose, (uintmax_t)imm, size);
288 static bool attr_w gen_imm(struct codegen_context *ctx, int64_t imm, unsigned purpose, unsigned size)
290 if (is_direct_const(imm, purpose, size)) {
291 ctx->const_imm = imm;
292 ctx->const_reg = false;
294 g(gen_load_constant(ctx, R_CONST_IMM, imm));
295 ctx->const_reg = true;
300 static bool attr_w gen_entry(struct codegen_context *ctx)
302 g(gen_imm(ctx, -FRAME_SIZE, IMM_PURPOSE_ADD, OP_SIZE_NATIVE));
303 gen_insn(INSN_ALU, OP_SIZE_NATIVE, ALU_ADD, 0);
308 g(gen_address(ctx, R_SP, FRAME_SIZE - 0x08, IMM_PURPOSE_STR_OFFSET, OP_SIZE_NATIVE));
309 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
310 gen_address_offset();
313 g(gen_address(ctx, R_SP, FRAME_SIZE - 0x10, IMM_PURPOSE_STR_OFFSET, OP_SIZE_NATIVE));
314 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
315 gen_address_offset();
318 g(gen_address(ctx, R_SP, FRAME_SIZE - 0x18, IMM_PURPOSE_STR_OFFSET, OP_SIZE_NATIVE));
319 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
320 gen_address_offset();
323 g(gen_address(ctx, R_SP, FRAME_SIZE - 0x20, IMM_PURPOSE_STR_OFFSET, OP_SIZE_NATIVE));
324 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
325 gen_address_offset();
326 gen_one(R_TIMESTAMP);
328 g(gen_address(ctx, R_SP, FRAME_SIZE - 0x28, IMM_PURPOSE_STR_OFFSET, OP_SIZE_NATIVE));
329 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
330 gen_address_offset();
333 g(gen_address(ctx, R_SP, FRAME_SIZE - 0x30, IMM_PURPOSE_STR_OFFSET, OP_SIZE_NATIVE));
334 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
335 gen_address_offset();
338 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
342 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
346 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
347 gen_one(R_TIMESTAMP);
350 gen_insn(INSN_JMP_INDIRECT, 0, 0, 0);
356 static bool attr_w gen_escape_arg(struct codegen_context *ctx, ip_t ip, uint32_t escape_label)
358 g(gen_load_constant(ctx, R_RET1, ip));
360 gen_insn(INSN_JMP, 0, 0, 0);
361 gen_four(escape_label);
366 static bool attr_w gen_escape(struct codegen_context *ctx)
368 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
372 g(gen_address(ctx, R_SP, FRAME_SIZE - 0x08, IMM_PURPOSE_LDR_OFFSET, OP_SIZE_NATIVE));
373 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
375 gen_address_offset();
377 g(gen_address(ctx, R_SP, FRAME_SIZE - 0x10, IMM_PURPOSE_LDR_OFFSET, OP_SIZE_NATIVE));
378 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
380 gen_address_offset();
382 g(gen_address(ctx, R_SP, FRAME_SIZE - 0x18, IMM_PURPOSE_LDR_OFFSET, OP_SIZE_NATIVE));
383 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
385 gen_address_offset();
387 g(gen_address(ctx, R_SP, FRAME_SIZE - 0x20, IMM_PURPOSE_LDR_OFFSET, OP_SIZE_NATIVE));
388 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
389 gen_one(R_TIMESTAMP);
390 gen_address_offset();
392 g(gen_address(ctx, R_SP, FRAME_SIZE - 0x28, IMM_PURPOSE_LDR_OFFSET, OP_SIZE_NATIVE));
393 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
395 gen_address_offset();
397 g(gen_address(ctx, R_SP, FRAME_SIZE - 0x30, IMM_PURPOSE_LDR_OFFSET, OP_SIZE_NATIVE));
398 gen_insn(INSN_MOV, OP_SIZE_NATIVE, 0, 0);
400 gen_address_offset();
402 g(gen_imm(ctx, FRAME_SIZE, IMM_PURPOSE_ADD, OP_SIZE_NATIVE));
403 gen_insn(INSN_ALU, OP_SIZE_NATIVE, ALU_ADD, 0);
408 gen_insn(INSN_RET, 0, 0, 0);
413 static bool attr_w gen_upcall_argument(struct codegen_context attr_unused *ctx, unsigned attr_unused arg)
418 static bool attr_w gen_upcall(struct codegen_context *ctx, unsigned offset, unsigned attr_unused n_args)
420 g(gen_address(ctx, R_UPCALL, offset, IMM_PURPOSE_LDR_OFFSET, OP_SIZE_ADDRESS));
421 gen_insn(INSN_MOV, OP_SIZE_ADDRESS, 0, 0);
422 gen_one(R_SCRATCH_NA_1);
423 gen_address_offset();
425 gen_insn(INSN_CALL_INDIRECT, OP_SIZE_ADDRESS, 0, 0);
426 gen_one(R_SCRATCH_NA_1);
431 static bool attr_w gen_cmp_test_jmp(struct codegen_context *ctx, unsigned insn, unsigned op_size, unsigned reg1, unsigned reg2, unsigned cond, uint32_t label);
433 static bool attr_w gen_timestamp_test(struct codegen_context *ctx, uint32_t escape_label)
435 g(gen_address(ctx, R_UPCALL, offsetof(struct cg_upcall_vector_s, ts), IMM_PURPOSE_LDR_SX_OFFSET, OP_SIZE_4));
436 gen_insn(INSN_MOVSX, OP_SIZE_4, 0, 0);
437 gen_one(R_SCRATCH_1);
438 gen_address_offset();
440 g(gen_cmp_test_jmp(ctx, INSN_CMP, OP_SIZE_NATIVE, R_SCRATCH_1, R_TIMESTAMP, COND_NE, escape_label));