1/* $NetBSD: sljitNativeARM_64.c,v 1.5 2024/04/02 22:29:57 riastradh Exp $ */ 2 3/* 4 * Stack-less Just-In-Time compiler 5 * 6 * Copyright Zoltan Herczeg (hzmester@freemail.hu). All rights reserved. 7 * 8 * Redistribution and use in source and binary forms, with or without modification, are 9 * permitted provided that the following conditions are met: 10 * 11 * 1. Redistributions of source code must retain the above copyright notice, this list of 12 * conditions and the following disclaimer. 13 * 14 * 2. Redistributions in binary form must reproduce the above copyright notice, this list 15 * of conditions and the following disclaimer in the documentation and/or other materials 16 * provided with the distribution. 17 * 18 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDER(S) AND CONTRIBUTORS ``AS IS'' AND ANY 19 * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 20 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT 21 * SHALL THE COPYRIGHT HOLDER(S) OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, 22 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED 23 * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR 24 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 25 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN 26 * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 27 */ 28 29SLJIT_API_FUNC_ATTRIBUTE const char* sljit_get_platform_name(void) 30{ 31 return "ARM-64" SLJIT_CPUINFO; 32} 33 34/* Length of an instruction word */ 35typedef sljit_u32 sljit_ins; 36 37#define TMP_ZERO (0) 38 39#define TMP_REG1 (SLJIT_NUMBER_OF_REGISTERS + 2) 40#define TMP_REG2 (SLJIT_NUMBER_OF_REGISTERS + 3) 41#define TMP_REG3 (SLJIT_NUMBER_OF_REGISTERS + 4) 42#define TMP_LR (SLJIT_NUMBER_OF_REGISTERS + 5) 43#define TMP_SP (SLJIT_NUMBER_OF_REGISTERS + 6) 44 45#define TMP_FREG1 (0) 46#define TMP_FREG2 (SLJIT_NUMBER_OF_FLOAT_REGISTERS + 1) 47 48static const sljit_u8 reg_map[SLJIT_NUMBER_OF_REGISTERS + 8] = { 49 31, 0, 1, 2, 3, 4, 5, 6, 7, 12, 13, 14, 15, 16, 17, 8, 28, 27, 26, 25, 24, 23, 22, 21, 20, 19, 29, 9, 10, 11, 30, 31 50}; 51 52#define W_OP (1 << 31) 53#define RD(rd) (reg_map[rd]) 54#define RT(rt) (reg_map[rt]) 55#define RN(rn) (reg_map[rn] << 5) 56#define RT2(rt2) (reg_map[rt2] << 10) 57#define RM(rm) (reg_map[rm] << 16) 58#define VD(vd) (vd) 59#define VT(vt) (vt) 60#define VN(vn) ((vn) << 5) 61#define VM(vm) ((vm) << 16) 62 63/* --------------------------------------------------------------------- */ 64/* Instrucion forms */ 65/* --------------------------------------------------------------------- */ 66 67#define ADC 0x9a000000 68#define ADD 0x8b000000 69#define ADDI 0x91000000 70#define AND 0x8a000000 71#define ANDI 0x92000000 72#define ASRV 0x9ac02800 73#define B 0x14000000 74#define B_CC 0x54000000 75#define BL 0x94000000 76#define BLR 0xd63f0000 77#define BR 0xd61f0000 78#define BRK 0xd4200000 79#define CBZ 0xb4000000 80#define CLZ 0xdac01000 81#define CSINC 0x9a800400 82#define EOR 0xca000000 83#define EORI 0xd2000000 84#define FABS 0x1e60c000 85#define FADD 0x1e602800 86#define FCMP 0x1e602000 87#define FCVT 0x1e224000 88#define FCVTZS 0x9e780000 89#define FDIV 0x1e601800 90#define FMOV 0x1e604000 91#define FMUL 0x1e600800 92#define FNEG 0x1e614000 93#define FSUB 0x1e603800 94#define LDRI 0xf9400000 95#define LDP 0xa9400000 96#define LDP_PST 0xa8c00000 97#define LSLV 0x9ac02000 98#define LSRV 0x9ac02400 99#define MADD 0x9b000000 100#define MOVK 0xf2800000 101#define MOVN 0x92800000 102#define MOVZ 0xd2800000 103#define NOP 0xd503201f 104#define ORN 0xaa200000 105#define ORR 0xaa000000 106#define ORRI 0xb2000000 107#define RET 0xd65f0000 108#define SBC 0xda000000 109#define SBFM 0x93000000 110#define SCVTF 0x9e620000 111#define SDIV 0x9ac00c00 112#define SMADDL 0x9b200000 113#define SMULH 0x9b403c00 114#define STP 0xa9000000 115#define STP_PRE 0xa9800000 116#define STRI 0xf9000000 117#define STR_FI 0x3d000000 118#define STR_FR 0x3c206800 119#define STUR_FI 0x3c000000 120#define SUB 0xcb000000 121#define SUBI 0xd1000000 122#define SUBS 0xeb000000 123#define UBFM 0xd3000000 124#define UDIV 0x9ac00800 125#define UMULH 0x9bc03c00 126 127/* dest_reg is the absolute name of the register 128 Useful for reordering instructions in the delay slot. */ 129static sljit_s32 push_inst(struct sljit_compiler *compiler, sljit_ins ins) 130{ 131 sljit_ins *ptr = (sljit_ins*)ensure_buf(compiler, sizeof(sljit_ins)); 132 FAIL_IF(!ptr); 133 *ptr = ins; 134 compiler->size++; 135 return SLJIT_SUCCESS; 136} 137 138static SLJIT_INLINE sljit_s32 emit_imm64_const(struct sljit_compiler *compiler, sljit_s32 dst, sljit_uw imm) 139{ 140 FAIL_IF(push_inst(compiler, MOVZ | RD(dst) | ((imm & 0xffff) << 5))); 141 FAIL_IF(push_inst(compiler, MOVK | RD(dst) | (((imm >> 16) & 0xffff) << 5) | (1 << 21))); 142 FAIL_IF(push_inst(compiler, MOVK | RD(dst) | (((imm >> 32) & 0xffff) << 5) | (2 << 21))); 143 return push_inst(compiler, MOVK | RD(dst) | ((imm >> 48) << 5) | (3 << 21)); 144} 145 146static SLJIT_INLINE void modify_imm64_const(sljit_ins* inst, sljit_uw new_imm) 147{ 148 sljit_s32 dst = inst[0] & 0x1f; 149 SLJIT_ASSERT((inst[0] & 0xffe00000) == MOVZ && (inst[1] & 0xffe00000) == (MOVK | (1 << 21))); 150 inst[0] = MOVZ | dst | ((new_imm & 0xffff) << 5); 151 inst[1] = MOVK | dst | (((new_imm >> 16) & 0xffff) << 5) | (1 << 21); 152 inst[2] = MOVK | dst | (((new_imm >> 32) & 0xffff) << 5) | (2 << 21); 153 inst[3] = MOVK | dst | ((new_imm >> 48) << 5) | (3 << 21); 154} 155 156static SLJIT_INLINE sljit_s32 detect_jump_type(struct sljit_jump *jump, sljit_ins *code_ptr, sljit_ins *code, sljit_sw executable_offset) 157{ 158 sljit_sw diff; 159 sljit_uw target_addr; 160 161 if (jump->flags & SLJIT_REWRITABLE_JUMP) { 162 jump->flags |= PATCH_ABS64; 163 return 0; 164 } 165 166 if (jump->flags & JUMP_ADDR) 167 target_addr = jump->u.target; 168 else { 169 SLJIT_ASSERT(jump->flags & JUMP_LABEL); 170 target_addr = (sljit_uw)(code + jump->u.label->size) + (sljit_uw)executable_offset; 171 } 172 173 diff = (sljit_sw)target_addr - (sljit_sw)(code_ptr + 4) - executable_offset; 174 175 if (jump->flags & IS_COND) { 176 diff += sizeof(sljit_ins); 177 if (diff <= 0xfffff && diff >= -0x100000) { 178 code_ptr[-5] ^= (jump->flags & IS_CBZ) ? (0x1 << 24) : 0x1; 179 jump->addr -= sizeof(sljit_ins); 180 jump->flags |= PATCH_COND; 181 return 5; 182 } 183 diff -= sizeof(sljit_ins); 184 } 185 186 if (diff <= 0x7ffffff && diff >= -0x8000000) { 187 jump->flags |= PATCH_B; 188 return 4; 189 } 190 191 if (target_addr <= 0xffffffffl) { 192 if (jump->flags & IS_COND) 193 code_ptr[-5] -= (2 << 5); 194 code_ptr[-2] = code_ptr[0]; 195 return 2; 196 } 197 if (target_addr <= 0xffffffffffffl) { 198 if (jump->flags & IS_COND) 199 code_ptr[-5] -= (1 << 5); 200 jump->flags |= PATCH_ABS48; 201 code_ptr[-1] = code_ptr[0]; 202 return 1; 203 } 204 205 jump->flags |= PATCH_ABS64; 206 return 0; 207} 208 209SLJIT_API_FUNC_ATTRIBUTE void* sljit_generate_code(struct sljit_compiler *compiler) 210{ 211 struct sljit_memory_fragment *buf; 212 sljit_ins *code; 213 sljit_ins *code_ptr; 214 sljit_ins *buf_ptr; 215 sljit_ins *buf_end; 216 sljit_uw word_count; 217 sljit_sw executable_offset; 218 sljit_uw addr; 219 sljit_s32 dst; 220 221 struct sljit_label *label; 222 struct sljit_jump *jump; 223 struct sljit_const *const_; 224 225 CHECK_ERROR_PTR(); 226 CHECK_PTR(check_sljit_generate_code(compiler)); 227 reverse_buf(compiler); 228 229 code = (sljit_ins*)SLJIT_MALLOC_EXEC(compiler->size * sizeof(sljit_ins)); 230 PTR_FAIL_WITH_EXEC_IF(code); 231 buf = compiler->buf; 232 233 code_ptr = code; 234 word_count = 0; 235 executable_offset = SLJIT_EXEC_OFFSET(code); 236 237 label = compiler->labels; 238 jump = compiler->jumps; 239 const_ = compiler->consts; 240 241 do { 242 buf_ptr = (sljit_ins*)buf->memory; 243 buf_end = buf_ptr + (buf->used_size >> 2); 244 do { 245 *code_ptr = *buf_ptr++; 246 /* These structures are ordered by their address. */ 247 SLJIT_ASSERT(!label || label->size >= word_count); 248 SLJIT_ASSERT(!jump || jump->addr >= word_count); 249 SLJIT_ASSERT(!const_ || const_->addr >= word_count); 250 if (label && label->size == word_count) { 251 label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset); 252 label->size = code_ptr - code; 253 label = label->next; 254 } 255 if (jump && jump->addr == word_count) { 256 jump->addr = (sljit_uw)(code_ptr - 4); 257 code_ptr -= detect_jump_type(jump, code_ptr, code, executable_offset); 258 jump = jump->next; 259 } 260 if (const_ && const_->addr == word_count) { 261 const_->addr = (sljit_uw)code_ptr; 262 const_ = const_->next; 263 } 264 code_ptr ++; 265 word_count ++; 266 } while (buf_ptr < buf_end); 267 268 buf = buf->next; 269 } while (buf); 270 271 if (label && label->size == word_count) { 272 label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset); 273 label->size = code_ptr - code; 274 label = label->next; 275 } 276 277 SLJIT_ASSERT(!label); 278 SLJIT_ASSERT(!jump); 279 SLJIT_ASSERT(!const_); 280 SLJIT_ASSERT(code_ptr - code <= (sljit_sw)compiler->size); 281 282 jump = compiler->jumps; 283 while (jump) { 284 do { 285 addr = (jump->flags & JUMP_LABEL) ? jump->u.label->addr : jump->u.target; 286 buf_ptr = (sljit_ins *)jump->addr; 287 288 if (jump->flags & PATCH_B) { 289 addr = (sljit_sw)(addr - (sljit_uw)SLJIT_ADD_EXEC_OFFSET(buf_ptr, executable_offset)) >> 2; 290 SLJIT_ASSERT((sljit_sw)addr <= 0x1ffffff && (sljit_sw)addr >= -0x2000000); 291 buf_ptr[0] = ((jump->flags & IS_BL) ? BL : B) | (addr & 0x3ffffff); 292 if (jump->flags & IS_COND) 293 buf_ptr[-1] -= (4 << 5); 294 break; 295 } 296 if (jump->flags & PATCH_COND) { 297 addr = (sljit_sw)(addr - (sljit_uw)SLJIT_ADD_EXEC_OFFSET(buf_ptr, executable_offset)) >> 2; 298 SLJIT_ASSERT((sljit_sw)addr <= 0x3ffff && (sljit_sw)addr >= -0x40000); 299 buf_ptr[0] = (buf_ptr[0] & ~0xffffe0) | ((addr & 0x7ffff) << 5); 300 break; 301 } 302 303 SLJIT_ASSERT((jump->flags & (PATCH_ABS48 | PATCH_ABS64)) || addr <= 0xffffffffl); 304 SLJIT_ASSERT((jump->flags & PATCH_ABS64) || addr <= 0xffffffffffffl); 305 306 dst = buf_ptr[0] & 0x1f; 307 buf_ptr[0] = MOVZ | dst | ((addr & 0xffff) << 5); 308 buf_ptr[1] = MOVK | dst | (((addr >> 16) & 0xffff) << 5) | (1 << 21); 309 if (jump->flags & (PATCH_ABS48 | PATCH_ABS64)) 310 buf_ptr[2] = MOVK | dst | (((addr >> 32) & 0xffff) << 5) | (2 << 21); 311 if (jump->flags & PATCH_ABS64) 312 buf_ptr[3] = MOVK | dst | (((addr >> 48) & 0xffff) << 5) | (3 << 21); 313 } while (0); 314 jump = jump->next; 315 } 316 317 compiler->error = SLJIT_ERR_COMPILED; 318 compiler->executable_offset = executable_offset; 319 compiler->executable_size = (code_ptr - code) * sizeof(sljit_ins); 320 321 code = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(code, executable_offset); 322 code_ptr = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset); 323 324 SLJIT_CACHE_FLUSH(code, code_ptr); 325 return code; 326} 327 328/* --------------------------------------------------------------------- */ 329/* Core code generator functions. */ 330/* --------------------------------------------------------------------- */ 331 332#define COUNT_TRAILING_ZERO(value, result) \ 333 result = 0; \ 334 if (!(value & 0xffffffff)) { \ 335 result += 32; \ 336 value >>= 32; \ 337 } \ 338 if (!(value & 0xffff)) { \ 339 result += 16; \ 340 value >>= 16; \ 341 } \ 342 if (!(value & 0xff)) { \ 343 result += 8; \ 344 value >>= 8; \ 345 } \ 346 if (!(value & 0xf)) { \ 347 result += 4; \ 348 value >>= 4; \ 349 } \ 350 if (!(value & 0x3)) { \ 351 result += 2; \ 352 value >>= 2; \ 353 } \ 354 if (!(value & 0x1)) { \ 355 result += 1; \ 356 value >>= 1; \ 357 } 358 359#define LOGICAL_IMM_CHECK 0x100 360 361static sljit_ins logical_imm(sljit_sw imm, sljit_s32 len) 362{ 363 sljit_s32 negated, ones, right; 364 sljit_uw mask, uimm; 365 sljit_ins ins; 366 367 if (len & LOGICAL_IMM_CHECK) { 368 len &= ~LOGICAL_IMM_CHECK; 369 if (len == 32 && (imm == 0 || imm == -1)) 370 return 0; 371 if (len == 16 && ((sljit_s32)imm == 0 || (sljit_s32)imm == -1)) 372 return 0; 373 } 374 375 SLJIT_ASSERT((len == 32 && imm != 0 && imm != -1) 376 || (len == 16 && (sljit_s32)imm != 0 && (sljit_s32)imm != -1)); 377 uimm = (sljit_uw)imm; 378 while (1) { 379 if (len <= 0) { 380 SLJIT_UNREACHABLE(); 381 return 0; 382 } 383 mask = ((sljit_uw)1 << len) - 1; 384 if ((uimm & mask) != ((uimm >> len) & mask)) 385 break; 386 len >>= 1; 387 } 388 389 len <<= 1; 390 391 negated = 0; 392 if (uimm & 0x1) { 393 negated = 1; 394 uimm = ~uimm; 395 } 396 397 if (len < 64) 398 uimm &= ((sljit_uw)1 << len) - 1; 399 400 /* Unsigned right shift. */ 401 COUNT_TRAILING_ZERO(uimm, right); 402 403 /* Signed shift. We also know that the highest bit is set. */ 404 imm = (sljit_sw)~uimm; 405 SLJIT_ASSERT(imm < 0); 406 407 COUNT_TRAILING_ZERO(imm, ones); 408 409 if (~imm) 410 return 0; 411 412 if (len == 64) 413 ins = 1 << 22; 414 else 415 ins = (0x3f - ((len << 1) - 1)) << 10; 416 417 if (negated) 418 return ins | ((len - ones - 1) << 10) | ((len - ones - right) << 16); 419 420 return ins | ((ones - 1) << 10) | ((len - right) << 16); 421} 422 423#undef COUNT_TRAILING_ZERO 424 425static sljit_s32 load_immediate(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw simm) 426{ 427 sljit_uw imm = (sljit_uw)simm; 428 sljit_s32 i, zeros, ones, first; 429 sljit_ins bitmask; 430 431 if (imm <= 0xffff) 432 return push_inst(compiler, MOVZ | RD(dst) | (imm << 5)); 433 434 if (simm >= -0x10000 && simm < 0) 435 return push_inst(compiler, MOVN | RD(dst) | ((~imm & 0xffff) << 5)); 436 437 if (imm <= 0xffffffffl) { 438 if ((imm & 0xffff0000l) == 0xffff0000) 439 return push_inst(compiler, (MOVN ^ W_OP) | RD(dst) | ((~imm & 0xffff) << 5)); 440 if ((imm & 0xffff) == 0xffff) 441 return push_inst(compiler, (MOVN ^ W_OP) | RD(dst) | ((~imm & 0xffff0000l) >> (16 - 5)) | (1 << 21)); 442 bitmask = logical_imm(simm, 16); 443 if (bitmask != 0) 444 return push_inst(compiler, (ORRI ^ W_OP) | RD(dst) | RN(TMP_ZERO) | bitmask); 445 } 446 else { 447 bitmask = logical_imm(simm, 32); 448 if (bitmask != 0) 449 return push_inst(compiler, ORRI | RD(dst) | RN(TMP_ZERO) | bitmask); 450 } 451 452 if (imm <= 0xffffffffl) { 453 FAIL_IF(push_inst(compiler, MOVZ | RD(dst) | ((imm & 0xffff) << 5))); 454 return push_inst(compiler, MOVK | RD(dst) | ((imm & 0xffff0000l) >> (16 - 5)) | (1 << 21)); 455 } 456 457 if (simm >= -0x100000000l && simm < 0) { 458 FAIL_IF(push_inst(compiler, MOVN | RD(dst) | ((~imm & 0xffff) << 5))); 459 return push_inst(compiler, MOVK | RD(dst) | ((imm & 0xffff0000l) >> (16 - 5)) | (1 << 21)); 460 } 461 462 /* A large amount of number can be constructed from ORR and MOVx, 463 but computing them is costly. We don't */ 464 465 zeros = 0; 466 ones = 0; 467 for (i = 4; i > 0; i--) { 468 if ((simm & 0xffff) == 0) 469 zeros++; 470 if ((simm & 0xffff) == 0xffff) 471 ones++; 472 simm >>= 16; 473 } 474 475 simm = (sljit_sw)imm; 476 first = 1; 477 if (ones > zeros) { 478 simm = ~simm; 479 for (i = 0; i < 4; i++) { 480 if (!(simm & 0xffff)) { 481 simm >>= 16; 482 continue; 483 } 484 if (first) { 485 first = 0; 486 FAIL_IF(push_inst(compiler, MOVN | RD(dst) | ((simm & 0xffff) << 5) | (i << 21))); 487 } 488 else 489 FAIL_IF(push_inst(compiler, MOVK | RD(dst) | ((~simm & 0xffff) << 5) | (i << 21))); 490 simm >>= 16; 491 } 492 return SLJIT_SUCCESS; 493 } 494 495 for (i = 0; i < 4; i++) { 496 if (!(simm & 0xffff)) { 497 simm >>= 16; 498 continue; 499 } 500 if (first) { 501 first = 0; 502 FAIL_IF(push_inst(compiler, MOVZ | RD(dst) | ((simm & 0xffff) << 5) | (i << 21))); 503 } 504 else 505 FAIL_IF(push_inst(compiler, MOVK | RD(dst) | ((simm & 0xffff) << 5) | (i << 21))); 506 simm >>= 16; 507 } 508 return SLJIT_SUCCESS; 509} 510 511#define ARG1_IMM 0x0010000 512#define ARG2_IMM 0x0020000 513#define INT_OP 0x0040000 514#define SET_FLAGS 0x0080000 515#define UNUSED_RETURN 0x0100000 516#define SLOW_DEST 0x0200000 517#define SLOW_SRC1 0x0400000 518#define SLOW_SRC2 0x0800000 519 520#define CHECK_FLAGS(flag_bits) \ 521 if (flags & SET_FLAGS) { \ 522 inv_bits |= flag_bits; \ 523 if (flags & UNUSED_RETURN) \ 524 dst = TMP_ZERO; \ 525 } 526 527static sljit_s32 emit_op_imm(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 dst, sljit_sw arg1, sljit_sw arg2) 528{ 529 /* dst must be register, TMP_REG1 530 arg1 must be register, TMP_REG1, imm 531 arg2 must be register, TMP_REG2, imm */ 532 sljit_ins inv_bits = (flags & INT_OP) ? (1 << 31) : 0; 533 sljit_ins inst_bits; 534 sljit_s32 op = (flags & 0xffff); 535 sljit_s32 reg; 536 sljit_sw imm, nimm; 537 538 if (SLJIT_UNLIKELY((flags & (ARG1_IMM | ARG2_IMM)) == (ARG1_IMM | ARG2_IMM))) { 539 /* Both are immediates. */ 540 flags &= ~ARG1_IMM; 541 if (arg1 == 0 && op != SLJIT_ADD && op != SLJIT_SUB) 542 arg1 = TMP_ZERO; 543 else { 544 FAIL_IF(load_immediate(compiler, TMP_REG1, arg1)); 545 arg1 = TMP_REG1; 546 } 547 } 548 549 if (flags & (ARG1_IMM | ARG2_IMM)) { 550 reg = (flags & ARG2_IMM) ? arg1 : arg2; 551 imm = (flags & ARG2_IMM) ? arg2 : arg1; 552 553 switch (op) { 554 case SLJIT_MUL: 555 case SLJIT_NEG: 556 case SLJIT_CLZ: 557 case SLJIT_ADDC: 558 case SLJIT_SUBC: 559 /* No form with immediate operand (except imm 0, which 560 is represented by a ZERO register). */ 561 break; 562 case SLJIT_MOV: 563 SLJIT_ASSERT(!(flags & SET_FLAGS) && (flags & ARG2_IMM) && arg1 == TMP_REG1); 564 return load_immediate(compiler, dst, imm); 565 case SLJIT_NOT: 566 SLJIT_ASSERT(flags & ARG2_IMM); 567 FAIL_IF(load_immediate(compiler, dst, (flags & INT_OP) ? (~imm & 0xffffffff) : ~imm)); 568 goto set_flags; 569 case SLJIT_SUB: 570 if (flags & ARG1_IMM) 571 break; 572 imm = -imm; 573 /* Fall through. */ 574 case SLJIT_ADD: 575 if (imm == 0) { 576 CHECK_FLAGS(1 << 29); 577 return push_inst(compiler, ((op == SLJIT_ADD ? ADDI : SUBI) ^ inv_bits) | RD(dst) | RN(reg)); 578 } 579 if (imm > 0 && imm <= 0xfff) { 580 CHECK_FLAGS(1 << 29); 581 return push_inst(compiler, (ADDI ^ inv_bits) | RD(dst) | RN(reg) | (imm << 10)); 582 } 583 nimm = -imm; 584 if (nimm > 0 && nimm <= 0xfff) { 585 CHECK_FLAGS(1 << 29); 586 return push_inst(compiler, (SUBI ^ inv_bits) | RD(dst) | RN(reg) | (nimm << 10)); 587 } 588 if (imm > 0 && imm <= 0xffffff && !(imm & 0xfff)) { 589 CHECK_FLAGS(1 << 29); 590 return push_inst(compiler, (ADDI ^ inv_bits) | RD(dst) | RN(reg) | ((imm >> 12) << 10) | (1 << 22)); 591 } 592 if (nimm > 0 && nimm <= 0xffffff && !(nimm & 0xfff)) { 593 CHECK_FLAGS(1 << 29); 594 return push_inst(compiler, (SUBI ^ inv_bits) | RD(dst) | RN(reg) | ((nimm >> 12) << 10) | (1 << 22)); 595 } 596 if (imm > 0 && imm <= 0xffffff && !(flags & SET_FLAGS)) { 597 FAIL_IF(push_inst(compiler, (ADDI ^ inv_bits) | RD(dst) | RN(reg) | ((imm >> 12) << 10) | (1 << 22))); 598 return push_inst(compiler, (ADDI ^ inv_bits) | RD(dst) | RN(dst) | ((imm & 0xfff) << 10)); 599 } 600 if (nimm > 0 && nimm <= 0xffffff && !(flags & SET_FLAGS)) { 601 FAIL_IF(push_inst(compiler, (SUBI ^ inv_bits) | RD(dst) | RN(reg) | ((nimm >> 12) << 10) | (1 << 22))); 602 return push_inst(compiler, (SUBI ^ inv_bits) | RD(dst) | RN(dst) | ((nimm & 0xfff) << 10)); 603 } 604 break; 605 case SLJIT_AND: 606 inst_bits = logical_imm(imm, LOGICAL_IMM_CHECK | ((flags & INT_OP) ? 16 : 32)); 607 if (!inst_bits) 608 break; 609 CHECK_FLAGS(3 << 29); 610 return push_inst(compiler, (ANDI ^ inv_bits) | RD(dst) | RN(reg) | inst_bits); 611 case SLJIT_OR: 612 case SLJIT_XOR: 613 inst_bits = logical_imm(imm, LOGICAL_IMM_CHECK | ((flags & INT_OP) ? 16 : 32)); 614 if (!inst_bits) 615 break; 616 if (op == SLJIT_OR) 617 inst_bits |= ORRI; 618 else 619 inst_bits |= EORI; 620 FAIL_IF(push_inst(compiler, (inst_bits ^ inv_bits) | RD(dst) | RN(reg))); 621 goto set_flags; 622 case SLJIT_SHL: 623 if (flags & ARG1_IMM) 624 break; 625 if (flags & INT_OP) { 626 imm &= 0x1f; 627 FAIL_IF(push_inst(compiler, (UBFM ^ inv_bits) | RD(dst) | RN(arg1) | ((-imm & 0x1f) << 16) | ((31 - imm) << 10))); 628 } 629 else { 630 imm &= 0x3f; 631 FAIL_IF(push_inst(compiler, (UBFM ^ inv_bits) | RD(dst) | RN(arg1) | (1 << 22) | ((-imm & 0x3f) << 16) | ((63 - imm) << 10))); 632 } 633 goto set_flags; 634 case SLJIT_LSHR: 635 case SLJIT_ASHR: 636 if (flags & ARG1_IMM) 637 break; 638 if (op == SLJIT_ASHR) 639 inv_bits |= 1 << 30; 640 if (flags & INT_OP) { 641 imm &= 0x1f; 642 FAIL_IF(push_inst(compiler, (UBFM ^ inv_bits) | RD(dst) | RN(arg1) | (imm << 16) | (31 << 10))); 643 } 644 else { 645 imm &= 0x3f; 646 FAIL_IF(push_inst(compiler, (UBFM ^ inv_bits) | RD(dst) | RN(arg1) | (1 << 22) | (imm << 16) | (63 << 10))); 647 } 648 goto set_flags; 649 default: 650 SLJIT_UNREACHABLE(); 651 break; 652 } 653 654 if (flags & ARG2_IMM) { 655 if (arg2 == 0) 656 arg2 = TMP_ZERO; 657 else { 658 FAIL_IF(load_immediate(compiler, TMP_REG2, arg2)); 659 arg2 = TMP_REG2; 660 } 661 } 662 else { 663 if (arg1 == 0) 664 arg1 = TMP_ZERO; 665 else { 666 FAIL_IF(load_immediate(compiler, TMP_REG1, arg1)); 667 arg1 = TMP_REG1; 668 } 669 } 670 } 671 672 /* Both arguments are registers. */ 673 switch (op) { 674 case SLJIT_MOV: 675 case SLJIT_MOV_P: 676 case SLJIT_MOVU: 677 case SLJIT_MOVU_P: 678 SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1); 679 if (dst == arg2) 680 return SLJIT_SUCCESS; 681 return push_inst(compiler, ORR | RD(dst) | RN(TMP_ZERO) | RM(arg2)); 682 case SLJIT_MOV_U8: 683 case SLJIT_MOVU_U8: 684 SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1); 685 return push_inst(compiler, (UBFM ^ (1 << 31)) | RD(dst) | RN(arg2) | (7 << 10)); 686 case SLJIT_MOV_S8: 687 case SLJIT_MOVU_S8: 688 SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1); 689 if (!(flags & INT_OP)) 690 inv_bits |= 1 << 22; 691 return push_inst(compiler, (SBFM ^ inv_bits) | RD(dst) | RN(arg2) | (7 << 10)); 692 case SLJIT_MOV_U16: 693 case SLJIT_MOVU_U16: 694 SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1); 695 return push_inst(compiler, (UBFM ^ (1 << 31)) | RD(dst) | RN(arg2) | (15 << 10)); 696 case SLJIT_MOV_S16: 697 case SLJIT_MOVU_S16: 698 SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1); 699 if (!(flags & INT_OP)) 700 inv_bits |= 1 << 22; 701 return push_inst(compiler, (SBFM ^ inv_bits) | RD(dst) | RN(arg2) | (15 << 10)); 702 case SLJIT_MOV_U32: 703 case SLJIT_MOVU_U32: 704 SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1); 705 if ((flags & INT_OP) && dst == arg2) 706 return SLJIT_SUCCESS; 707 return push_inst(compiler, (ORR ^ (1 << 31)) | RD(dst) | RN(TMP_ZERO) | RM(arg2)); 708 case SLJIT_MOV_S32: 709 case SLJIT_MOVU_S32: 710 SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1); 711 if ((flags & INT_OP) && dst == arg2) 712 return SLJIT_SUCCESS; 713 return push_inst(compiler, SBFM | (1 << 22) | RD(dst) | RN(arg2) | (31 << 10)); 714 case SLJIT_NOT: 715 SLJIT_ASSERT(arg1 == TMP_REG1); 716 FAIL_IF(push_inst(compiler, (ORN ^ inv_bits) | RD(dst) | RN(TMP_ZERO) | RM(arg2))); 717 goto set_flags; 718 case SLJIT_NEG: 719 SLJIT_ASSERT(arg1 == TMP_REG1); 720 if (flags & SET_FLAGS) 721 inv_bits |= 1 << 29; 722 return push_inst(compiler, (SUB ^ inv_bits) | RD(dst) | RN(TMP_ZERO) | RM(arg2)); 723 case SLJIT_CLZ: 724 SLJIT_ASSERT(arg1 == TMP_REG1); 725 FAIL_IF(push_inst(compiler, (CLZ ^ inv_bits) | RD(dst) | RN(arg2))); 726 goto set_flags; 727 case SLJIT_ADD: 728 CHECK_FLAGS(1 << 29); 729 return push_inst(compiler, (ADD ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)); 730 case SLJIT_ADDC: 731 CHECK_FLAGS(1 << 29); 732 return push_inst(compiler, (ADC ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)); 733 case SLJIT_SUB: 734 CHECK_FLAGS(1 << 29); 735 return push_inst(compiler, (SUB ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)); 736 case SLJIT_SUBC: 737 CHECK_FLAGS(1 << 29); 738 return push_inst(compiler, (SBC ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)); 739 case SLJIT_MUL: 740 if (!(flags & SET_FLAGS)) 741 return push_inst(compiler, (MADD ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2) | RT2(TMP_ZERO)); 742 if (flags & INT_OP) { 743 FAIL_IF(push_inst(compiler, SMADDL | RD(dst) | RN(arg1) | RM(arg2) | (31 << 10))); 744 FAIL_IF(push_inst(compiler, ADD | RD(TMP_LR) | RN(TMP_ZERO) | RM(dst) | (2 << 22) | (31 << 10))); 745 return push_inst(compiler, SUBS | RD(TMP_ZERO) | RN(TMP_LR) | RM(dst) | (2 << 22) | (63 << 10)); 746 } 747 FAIL_IF(push_inst(compiler, SMULH | RD(TMP_LR) | RN(arg1) | RM(arg2))); 748 FAIL_IF(push_inst(compiler, MADD | RD(dst) | RN(arg1) | RM(arg2) | RT2(TMP_ZERO))); 749 return push_inst(compiler, SUBS | RD(TMP_ZERO) | RN(TMP_LR) | RM(dst) | (2 << 22) | (63 << 10)); 750 case SLJIT_AND: 751 CHECK_FLAGS(3 << 29); 752 return push_inst(compiler, (AND ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)); 753 case SLJIT_OR: 754 FAIL_IF(push_inst(compiler, (ORR ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2))); 755 goto set_flags; 756 case SLJIT_XOR: 757 FAIL_IF(push_inst(compiler, (EOR ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2))); 758 goto set_flags; 759 case SLJIT_SHL: 760 FAIL_IF(push_inst(compiler, (LSLV ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2))); 761 goto set_flags; 762 case SLJIT_LSHR: 763 FAIL_IF(push_inst(compiler, (LSRV ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2))); 764 goto set_flags; 765 case SLJIT_ASHR: 766 FAIL_IF(push_inst(compiler, (ASRV ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2))); 767 goto set_flags; 768 } 769 770 SLJIT_UNREACHABLE(); 771 return SLJIT_SUCCESS; 772 773set_flags: 774 if (flags & SET_FLAGS) 775 return push_inst(compiler, (SUBS ^ inv_bits) | RD(TMP_ZERO) | RN(dst) | RM(TMP_ZERO)); 776 return SLJIT_SUCCESS; 777} 778 779#define STORE 0x01 780#define SIGNED 0x02 781 782#define UPDATE 0x04 783#define ARG_TEST 0x08 784 785#define BYTE_SIZE 0x000 786#define HALF_SIZE 0x100 787#define INT_SIZE 0x200 788#define WORD_SIZE 0x300 789 790#define MEM_SIZE_SHIFT(flags) ((flags) >> 8) 791 792static const sljit_ins sljit_mem_imm[4] = { 793/* u l */ 0x39400000 /* ldrb [reg,imm] */, 794/* u s */ 0x39000000 /* strb [reg,imm] */, 795/* s l */ 0x39800000 /* ldrsb [reg,imm] */, 796/* s s */ 0x39000000 /* strb [reg,imm] */, 797}; 798 799static const sljit_ins sljit_mem_simm[4] = { 800/* u l */ 0x38400000 /* ldurb [reg,imm] */, 801/* u s */ 0x38000000 /* sturb [reg,imm] */, 802/* s l */ 0x38800000 /* ldursb [reg,imm] */, 803/* s s */ 0x38000000 /* sturb [reg,imm] */, 804}; 805 806static const sljit_ins sljit_mem_pre_simm[4] = { 807/* u l */ 0x38400c00 /* ldrb [reg,imm]! */, 808/* u s */ 0x38000c00 /* strb [reg,imm]! */, 809/* s l */ 0x38800c00 /* ldrsb [reg,imm]! */, 810/* s s */ 0x38000c00 /* strb [reg,imm]! */, 811}; 812 813static const sljit_ins sljit_mem_reg[4] = { 814/* u l */ 0x38606800 /* ldrb [reg,reg] */, 815/* u s */ 0x38206800 /* strb [reg,reg] */, 816/* s l */ 0x38a06800 /* ldrsb [reg,reg] */, 817/* s s */ 0x38206800 /* strb [reg,reg] */, 818}; 819 820/* Helper function. Dst should be reg + value, using at most 1 instruction, flags does not set. */ 821static sljit_s32 emit_set_delta(struct sljit_compiler *compiler, sljit_s32 dst, sljit_s32 reg, sljit_sw value) 822{ 823 if (value >= 0) { 824 if (value <= 0xfff) 825 return push_inst(compiler, ADDI | RD(dst) | RN(reg) | (value << 10)); 826 if (value <= 0xffffff && !(value & 0xfff)) 827 return push_inst(compiler, ADDI | (1 << 22) | RD(dst) | RN(reg) | (value >> 2)); 828 } 829 else { 830 value = -value; 831 if (value <= 0xfff) 832 return push_inst(compiler, SUBI | RD(dst) | RN(reg) | (value << 10)); 833 if (value <= 0xffffff && !(value & 0xfff)) 834 return push_inst(compiler, SUBI | (1 << 22) | RD(dst) | RN(reg) | (value >> 2)); 835 } 836 return SLJIT_ERR_UNSUPPORTED; 837} 838 839/* Can perform an operation using at most 1 instruction. */ 840static sljit_s32 getput_arg_fast(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg, sljit_s32 arg, sljit_sw argw) 841{ 842 sljit_u32 shift = MEM_SIZE_SHIFT(flags); 843 844 SLJIT_ASSERT(arg & SLJIT_MEM); 845 846 if (SLJIT_UNLIKELY(flags & UPDATE)) { 847 if ((arg & REG_MASK) && !(arg & OFFS_REG_MASK) && argw <= 255 && argw >= -256) { 848 if (SLJIT_UNLIKELY(flags & ARG_TEST)) 849 return 1; 850 851 arg &= REG_MASK; 852 argw &= 0x1ff; 853 FAIL_IF(push_inst(compiler, sljit_mem_pre_simm[flags & 0x3] 854 | (shift << 30) | RT(reg) | RN(arg) | (argw << 12))); 855 return -1; 856 } 857 return 0; 858 } 859 860 if (SLJIT_UNLIKELY(arg & OFFS_REG_MASK)) { 861 argw &= 0x3; 862 if (argw && argw != shift) 863 return 0; 864 865 if (SLJIT_UNLIKELY(flags & ARG_TEST)) 866 return 1; 867 868 FAIL_IF(push_inst(compiler, sljit_mem_reg[flags & 0x3] | (shift << 30) | RT(reg) 869 | RN(arg & REG_MASK) | RM(OFFS_REG(arg)) | (argw ? (1 << 12) : 0))); 870 return -1; 871 } 872 873 arg &= REG_MASK; 874 if (argw >= 0 && (argw >> shift) <= 0xfff && (argw & ((1 << shift) - 1)) == 0) { 875 if (SLJIT_UNLIKELY(flags & ARG_TEST)) 876 return 1; 877 878 FAIL_IF(push_inst(compiler, sljit_mem_imm[flags & 0x3] | (shift << 30) 879 | RT(reg) | RN(arg) | (argw << (10 - shift)))); 880 return -1; 881 } 882 883 if (argw > 255 || argw < -256) 884 return 0; 885 886 if (SLJIT_UNLIKELY(flags & ARG_TEST)) 887 return 1; 888 889 FAIL_IF(push_inst(compiler, sljit_mem_simm[flags & 0x3] | (shift << 30) 890 | RT(reg) | RN(arg) | ((argw & 0x1ff) << 12))); 891 return -1; 892} 893 894/* see getput_arg below. 895 Note: can_cache is called only for binary operators. Those 896 operators always uses word arguments without write back. */ 897static sljit_s32 can_cache(sljit_s32 arg, sljit_sw argw, sljit_s32 next_arg, sljit_sw next_argw) 898{ 899 sljit_sw diff; 900 if ((arg & OFFS_REG_MASK) || !(next_arg & SLJIT_MEM)) 901 return 0; 902 903 if (!(arg & REG_MASK)) { 904 diff = argw - next_argw; 905 if (diff <= 0xfff && diff >= -0xfff) 906 return 1; 907 return 0; 908 } 909 910 if (argw == next_argw) 911 return 1; 912 913 diff = argw - next_argw; 914 if (arg == next_arg && diff <= 0xfff && diff >= -0xfff) 915 return 1; 916 917 return 0; 918} 919 920/* Emit the necessary instructions. See can_cache above. */ 921static sljit_s32 getput_arg(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg, 922 sljit_s32 arg, sljit_sw argw, sljit_s32 next_arg, sljit_sw next_argw) 923{ 924 sljit_u32 shift = MEM_SIZE_SHIFT(flags); 925 sljit_s32 tmp_r, other_r; 926 sljit_sw diff; 927 928 SLJIT_ASSERT(arg & SLJIT_MEM); 929 if (!(next_arg & SLJIT_MEM)) { 930 next_arg = 0; 931 next_argw = 0; 932 } 933 934 tmp_r = (flags & STORE) ? TMP_REG3 : reg; 935 936 if (SLJIT_UNLIKELY((flags & UPDATE) && (arg & REG_MASK))) { 937 /* Update only applies if a base register exists. */ 938 other_r = OFFS_REG(arg); 939 if (!other_r) { 940 other_r = arg & REG_MASK; 941 SLJIT_ASSERT(other_r != reg); 942 943 if (argw >= 0 && argw <= 0xffffff) { 944 if ((argw & 0xfff) != 0) 945 FAIL_IF(push_inst(compiler, ADDI | RD(other_r) | RN(other_r) | ((argw & 0xfff) << 10))); 946 if (argw >> 12) 947 FAIL_IF(push_inst(compiler, ADDI | (1 << 22) | RD(other_r) | RN(other_r) | ((argw >> 12) << 10))); 948 return push_inst(compiler, sljit_mem_imm[flags & 0x3] | (shift << 30) | RT(reg) | RN(other_r)); 949 } 950 else if (argw < 0 && argw >= -0xffffff) { 951 argw = -argw; 952 if ((argw & 0xfff) != 0) 953 FAIL_IF(push_inst(compiler, SUBI | RD(other_r) | RN(other_r) | ((argw & 0xfff) << 10))); 954 if (argw >> 12) 955 FAIL_IF(push_inst(compiler, SUBI | (1 << 22) | RD(other_r) | RN(other_r) | ((argw >> 12) << 10))); 956 return push_inst(compiler, sljit_mem_imm[flags & 0x3] | (shift << 30) | RT(reg) | RN(other_r)); 957 } 958 959 if (compiler->cache_arg == SLJIT_MEM) { 960 if (argw == compiler->cache_argw) { 961 other_r = TMP_REG3; 962 argw = 0; 963 } 964 else if (emit_set_delta(compiler, TMP_REG3, TMP_REG3, argw - compiler->cache_argw) != SLJIT_ERR_UNSUPPORTED) { 965 FAIL_IF(compiler->error); 966 compiler->cache_argw = argw; 967 other_r = TMP_REG3; 968 argw = 0; 969 } 970 } 971 972 if (argw) { 973 FAIL_IF(load_immediate(compiler, TMP_REG3, argw)); 974 compiler->cache_arg = SLJIT_MEM; 975 compiler->cache_argw = argw; 976 other_r = TMP_REG3; 977 argw = 0; 978 } 979 } 980 981 /* No caching here. */ 982 arg &= REG_MASK; 983 FAIL_IF(push_inst(compiler, sljit_mem_reg[flags & 0x3] | (shift << 30) | RT(reg) | RN(arg) | RM(other_r))); 984 return push_inst(compiler, ADD | RD(arg) | RN(arg) | RM(other_r)); 985 } 986 987 if (arg & OFFS_REG_MASK) { 988 other_r = OFFS_REG(arg); 989 arg &= REG_MASK; 990 FAIL_IF(push_inst(compiler, ADD | RD(tmp_r) | RN(arg) | RM(other_r) | ((argw & 0x3) << 10))); 991 return push_inst(compiler, sljit_mem_imm[flags & 0x3] | (shift << 30) | RT(reg) | RN(tmp_r)); 992 } 993 994 if (compiler->cache_arg == arg) { 995 diff = argw - compiler->cache_argw; 996 if (diff <= 255 && diff >= -256) 997 return push_inst(compiler, sljit_mem_simm[flags & 0x3] | (shift << 30) 998 | RT(reg) | RN(TMP_REG3) | ((diff & 0x1ff) << 12)); 999 if (emit_set_delta(compiler, TMP_REG3, TMP_REG3, diff) != SLJIT_ERR_UNSUPPORTED) { 1000 FAIL_IF(compiler->error); 1001 return push_inst(compiler, sljit_mem_imm[flags & 0x3] | (shift << 30) | RT(reg) | RN(arg)); 1002 } 1003 } 1004 1005 if (argw >= 0 && argw <= 0xffffff && (argw & ((1 << shift) - 1)) == 0) { 1006 FAIL_IF(push_inst(compiler, ADDI | (1 << 22) | RD(tmp_r) | RN(arg & REG_MASK) | ((argw >> 12) << 10))); 1007 return push_inst(compiler, sljit_mem_imm[flags & 0x3] | (shift << 30) 1008 | RT(reg) | RN(tmp_r) | ((argw & 0xfff) << (10 - shift))); 1009 } 1010 1011 diff = argw - next_argw; 1012 next_arg = (arg & REG_MASK) && (arg == next_arg) && diff <= 0xfff && diff >= -0xfff && diff != 0; 1013 arg &= REG_MASK; 1014 1015 if (arg && compiler->cache_arg == SLJIT_MEM) { 1016 if (compiler->cache_argw == argw) 1017 return push_inst(compiler, sljit_mem_reg[flags & 0x3] | (shift << 30) | RT(reg) | RN(arg) | RM(TMP_REG3)); 1018 if (emit_set_delta(compiler, TMP_REG3, TMP_REG3, argw - compiler->cache_argw) != SLJIT_ERR_UNSUPPORTED) { 1019 FAIL_IF(compiler->error); 1020 compiler->cache_argw = argw; 1021 return push_inst(compiler, sljit_mem_reg[flags & 0x3] | (shift << 30) | RT(reg) | RN(arg) | RM(TMP_REG3)); 1022 } 1023 } 1024 1025 compiler->cache_argw = argw; 1026 if (next_arg && emit_set_delta(compiler, TMP_REG3, arg, argw) != SLJIT_ERR_UNSUPPORTED) { 1027 FAIL_IF(compiler->error); 1028 compiler->cache_arg = SLJIT_MEM | arg; 1029 arg = 0; 1030 } 1031 else { 1032 FAIL_IF(load_immediate(compiler, TMP_REG3, argw)); 1033 compiler->cache_arg = SLJIT_MEM; 1034 1035 if (next_arg) { 1036 FAIL_IF(push_inst(compiler, ADD | RD(TMP_REG3) | RN(TMP_REG3) | RM(arg))); 1037 compiler->cache_arg = SLJIT_MEM | arg; 1038 arg = 0; 1039 } 1040 } 1041 1042 if (arg) 1043 return push_inst(compiler, sljit_mem_reg[flags & 0x3] | (shift << 30) | RT(reg) | RN(arg) | RM(TMP_REG3)); 1044 return push_inst(compiler, sljit_mem_imm[flags & 0x3] | (shift << 30) | RT(reg) | RN(TMP_REG3)); 1045} 1046 1047static SLJIT_INLINE sljit_s32 emit_op_mem(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg, sljit_s32 arg, sljit_sw argw) 1048{ 1049 if (getput_arg_fast(compiler, flags, reg, arg, argw)) 1050 return compiler->error; 1051 compiler->cache_arg = 0; 1052 compiler->cache_argw = 0; 1053 return getput_arg(compiler, flags, reg, arg, argw, 0, 0); 1054} 1055 1056static SLJIT_INLINE sljit_s32 emit_op_mem2(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg, sljit_s32 arg1, sljit_sw arg1w, sljit_s32 arg2, sljit_sw arg2w) 1057{ 1058 if (getput_arg_fast(compiler, flags, reg, arg1, arg1w)) 1059 return compiler->error; 1060 return getput_arg(compiler, flags, reg, arg1, arg1w, arg2, arg2w); 1061} 1062 1063/* --------------------------------------------------------------------- */ 1064/* Entry, exit */ 1065/* --------------------------------------------------------------------- */ 1066 1067SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_enter(struct sljit_compiler *compiler, 1068 sljit_s32 options, sljit_s32 args, sljit_s32 scratches, sljit_s32 saveds, 1069 sljit_s32 fscratches, sljit_s32 fsaveds, sljit_s32 local_size) 1070{ 1071 sljit_s32 i, tmp, offs, prev, saved_regs_size; 1072 1073 CHECK_ERROR(); 1074 CHECK(check_sljit_emit_enter(compiler, options, args, scratches, saveds, fscratches, fsaveds, local_size)); 1075 set_emit_enter(compiler, options, args, scratches, saveds, fscratches, fsaveds, local_size); 1076 1077 saved_regs_size = GET_SAVED_REGISTERS_SIZE(scratches, saveds, 0); 1078 local_size += saved_regs_size + SLJIT_LOCALS_OFFSET; 1079 local_size = (local_size + 15) & ~0xf; 1080 compiler->local_size = local_size; 1081 1082 SLJIT_ASSERT(local_size >= 0); 1083 if ((size_t)local_size <= (63 * sizeof(sljit_sw))) { 1084 FAIL_IF(push_inst(compiler, STP_PRE | 29 | RT2(TMP_LR) 1085 | RN(TMP_SP) | ((-(local_size >> 3) & 0x7f) << 15))); 1086 FAIL_IF(push_inst(compiler, ADDI | RD(SLJIT_SP) | RN(TMP_SP) | (0 << 10))); 1087 offs = (local_size - saved_regs_size) << (15 - 3); 1088 } else { 1089 offs = 0 << 15; 1090 if (saved_regs_size & 0x8) { 1091 offs = 1 << 15; 1092 saved_regs_size += sizeof(sljit_sw); 1093 } 1094 local_size -= saved_regs_size + SLJIT_LOCALS_OFFSET; 1095 if (saved_regs_size > 0) 1096 FAIL_IF(push_inst(compiler, SUBI | RD(TMP_SP) | RN(TMP_SP) | (saved_regs_size << 10))); 1097 } 1098 1099 tmp = saveds < SLJIT_NUMBER_OF_SAVED_REGISTERS ? (SLJIT_S0 + 1 - saveds) : SLJIT_FIRST_SAVED_REG; 1100 prev = -1; 1101 for (i = SLJIT_S0; i >= tmp; i--) { 1102 if (prev == -1) { 1103 if (!(offs & (1 << 15))) { 1104 prev = i; 1105 continue; 1106 } 1107 FAIL_IF(push_inst(compiler, STRI | RT(i) | RN(TMP_SP) | (offs >> 5))); 1108 offs += 1 << 15; 1109 continue; 1110 } 1111 FAIL_IF(push_inst(compiler, STP | RT(prev) | RT2(i) | RN(TMP_SP) | offs)); 1112 offs += 2 << 15; 1113 prev = -1; 1114 } 1115 1116 for (i = scratches; i >= SLJIT_FIRST_SAVED_REG; i--) { 1117 if (prev == -1) { 1118 if (!(offs & (1 << 15))) { 1119 prev = i; 1120 continue; 1121 } 1122 FAIL_IF(push_inst(compiler, STRI | RT(i) | RN(TMP_SP) | (offs >> 5))); 1123 offs += 1 << 15; 1124 continue; 1125 } 1126 FAIL_IF(push_inst(compiler, STP | RT(prev) | RT2(i) | RN(TMP_SP) | offs)); 1127 offs += 2 << 15; 1128 prev = -1; 1129 } 1130 1131 SLJIT_ASSERT(prev == -1); 1132 1133 SLJIT_ASSERT(compiler->local_size >= 0); 1134 if ((size_t)compiler->local_size > (63 * sizeof(sljit_sw))) { 1135 /* The local_size is already adjusted by the saved registers. */ 1136 if (local_size > 0xfff) { 1137 FAIL_IF(push_inst(compiler, SUBI | RD(TMP_SP) | RN(TMP_SP) | ((local_size >> 12) << 10) | (1 << 22))); 1138 local_size &= 0xfff; 1139 } 1140 if (local_size) 1141 FAIL_IF(push_inst(compiler, SUBI | RD(TMP_SP) | RN(TMP_SP) | (local_size << 10))); 1142 FAIL_IF(push_inst(compiler, STP_PRE | 29 | RT2(TMP_LR) 1143 | RN(TMP_SP) | ((-(16 >> 3) & 0x7f) << 15))); 1144 FAIL_IF(push_inst(compiler, ADDI | RD(SLJIT_SP) | RN(TMP_SP) | (0 << 10))); 1145 } 1146 1147 if (args >= 1) 1148 FAIL_IF(push_inst(compiler, ORR | RD(SLJIT_S0) | RN(TMP_ZERO) | RM(SLJIT_R0))); 1149 if (args >= 2) 1150 FAIL_IF(push_inst(compiler, ORR | RD(SLJIT_S1) | RN(TMP_ZERO) | RM(SLJIT_R1))); 1151 if (args >= 3) 1152 FAIL_IF(push_inst(compiler, ORR | RD(SLJIT_S2) | RN(TMP_ZERO) | RM(SLJIT_R2))); 1153 1154 return SLJIT_SUCCESS; 1155} 1156 1157SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_set_context(struct sljit_compiler *compiler, 1158 sljit_s32 options, sljit_s32 args, sljit_s32 scratches, sljit_s32 saveds, 1159 sljit_s32 fscratches, sljit_s32 fsaveds, sljit_s32 local_size) 1160{ 1161 CHECK_ERROR(); 1162 CHECK(check_sljit_set_context(compiler, options, args, scratches, saveds, fscratches, fsaveds, local_size)); 1163 set_set_context(compiler, options, args, scratches, saveds, fscratches, fsaveds, local_size); 1164 1165 local_size += GET_SAVED_REGISTERS_SIZE(scratches, saveds, 0) + SLJIT_LOCALS_OFFSET; 1166 local_size = (local_size + 15) & ~0xf; 1167 compiler->local_size = local_size; 1168 return SLJIT_SUCCESS; 1169} 1170 1171SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_return(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 src, sljit_sw srcw) 1172{ 1173 sljit_s32 local_size; 1174 sljit_s32 i, tmp, offs, prev, saved_regs_size; 1175 1176 CHECK_ERROR(); 1177 CHECK(check_sljit_emit_return(compiler, op, src, srcw)); 1178 1179 FAIL_IF(emit_mov_before_return(compiler, op, src, srcw)); 1180 1181 local_size = compiler->local_size; 1182 1183 saved_regs_size = GET_SAVED_REGISTERS_SIZE(compiler->scratches, compiler->saveds, 0); 1184 SLJIT_ASSERT(local_size >= 0); 1185 if ((size_t)local_size <= (63 * sizeof(sljit_sw))) 1186 offs = (local_size - saved_regs_size) << (15 - 3); 1187 else { 1188 FAIL_IF(push_inst(compiler, LDP_PST | 29 | RT2(TMP_LR) 1189 | RN(TMP_SP) | (((16 >> 3) & 0x7f) << 15))); 1190 offs = 0 << 15; 1191 if (saved_regs_size & 0x8) { 1192 offs = 1 << 15; 1193 saved_regs_size += sizeof(sljit_sw); 1194 } 1195 local_size -= saved_regs_size + SLJIT_LOCALS_OFFSET; 1196 if (local_size > 0xfff) { 1197 FAIL_IF(push_inst(compiler, ADDI | RD(TMP_SP) | RN(TMP_SP) | ((local_size >> 12) << 10) | (1 << 22))); 1198 local_size &= 0xfff; 1199 } 1200 if (local_size) 1201 FAIL_IF(push_inst(compiler, ADDI | RD(TMP_SP) | RN(TMP_SP) | (local_size << 10))); 1202 } 1203 1204 tmp = compiler->saveds < SLJIT_NUMBER_OF_SAVED_REGISTERS ? (SLJIT_S0 + 1 - compiler->saveds) : SLJIT_FIRST_SAVED_REG; 1205 prev = -1; 1206 for (i = SLJIT_S0; i >= tmp; i--) { 1207 if (prev == -1) { 1208 if (!(offs & (1 << 15))) { 1209 prev = i; 1210 continue; 1211 } 1212 FAIL_IF(push_inst(compiler, LDRI | RT(i) | RN(TMP_SP) | (offs >> 5))); 1213 offs += 1 << 15; 1214 continue; 1215 } 1216 FAIL_IF(push_inst(compiler, LDP | RT(prev) | RT2(i) | RN(TMP_SP) | offs)); 1217 offs += 2 << 15; 1218 prev = -1; 1219 } 1220 1221 for (i = compiler->scratches; i >= SLJIT_FIRST_SAVED_REG; i--) { 1222 if (prev == -1) { 1223 if (!(offs & (1 << 15))) { 1224 prev = i; 1225 continue; 1226 } 1227 FAIL_IF(push_inst(compiler, LDRI | RT(i) | RN(TMP_SP) | (offs >> 5))); 1228 offs += 1 << 15; 1229 continue; 1230 } 1231 FAIL_IF(push_inst(compiler, LDP | RT(prev) | RT2(i) | RN(TMP_SP) | offs)); 1232 offs += 2 << 15; 1233 prev = -1; 1234 } 1235 1236 SLJIT_ASSERT(prev == -1); 1237 1238 SLJIT_ASSERT(compiler->local_size >= 0); 1239 if ((size_t)compiler->local_size <= (63 * sizeof(sljit_sw))) { 1240 FAIL_IF(push_inst(compiler, LDP_PST | 29 | RT2(TMP_LR) 1241 | RN(TMP_SP) | (((local_size >> 3) & 0x7f) << 15))); 1242 } else if (saved_regs_size > 0) { 1243 FAIL_IF(push_inst(compiler, ADDI | RD(TMP_SP) | RN(TMP_SP) | (saved_regs_size << 10))); 1244 } 1245 1246 FAIL_IF(push_inst(compiler, RET | RN(TMP_LR))); 1247 return SLJIT_SUCCESS; 1248} 1249 1250/* --------------------------------------------------------------------- */ 1251/* Operators */ 1252/* --------------------------------------------------------------------- */ 1253 1254SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op0(struct sljit_compiler *compiler, sljit_s32 op) 1255{ 1256 sljit_ins inv_bits = (op & SLJIT_I32_OP) ? (1 << 31) : 0; 1257 1258 CHECK_ERROR(); 1259 CHECK(check_sljit_emit_op0(compiler, op)); 1260 1261 op = GET_OPCODE(op); 1262 switch (op) { 1263 case SLJIT_BREAKPOINT: 1264 return push_inst(compiler, BRK); 1265 case SLJIT_NOP: 1266 return push_inst(compiler, NOP); 1267 case SLJIT_LMUL_UW: 1268 case SLJIT_LMUL_SW: 1269 FAIL_IF(push_inst(compiler, ORR | RD(TMP_REG1) | RN(TMP_ZERO) | RM(SLJIT_R0))); 1270 FAIL_IF(push_inst(compiler, MADD | RD(SLJIT_R0) | RN(SLJIT_R0) | RM(SLJIT_R1) | RT2(TMP_ZERO))); 1271 return push_inst(compiler, (op == SLJIT_LMUL_UW ? UMULH : SMULH) | RD(SLJIT_R1) | RN(TMP_REG1) | RM(SLJIT_R1)); 1272 case SLJIT_DIVMOD_UW: 1273 case SLJIT_DIVMOD_SW: 1274 FAIL_IF(push_inst(compiler, (ORR ^ inv_bits) | RD(TMP_REG1) | RN(TMP_ZERO) | RM(SLJIT_R0))); 1275 FAIL_IF(push_inst(compiler, ((op == SLJIT_DIVMOD_UW ? UDIV : SDIV) ^ inv_bits) | RD(SLJIT_R0) | RN(SLJIT_R0) | RM(SLJIT_R1))); 1276 FAIL_IF(push_inst(compiler, (MADD ^ inv_bits) | RD(SLJIT_R1) | RN(SLJIT_R0) | RM(SLJIT_R1) | RT2(TMP_ZERO))); 1277 return push_inst(compiler, (SUB ^ inv_bits) | RD(SLJIT_R1) | RN(TMP_REG1) | RM(SLJIT_R1)); 1278 case SLJIT_DIV_UW: 1279 case SLJIT_DIV_SW: 1280 return push_inst(compiler, ((op == SLJIT_DIV_UW ? UDIV : SDIV) ^ inv_bits) | RD(SLJIT_R0) | RN(SLJIT_R0) | RM(SLJIT_R1)); 1281 } 1282 1283 return SLJIT_SUCCESS; 1284} 1285 1286SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op1(struct sljit_compiler *compiler, sljit_s32 op, 1287 sljit_s32 dst, sljit_sw dstw, 1288 sljit_s32 src, sljit_sw srcw) 1289{ 1290 sljit_s32 dst_r, flags, mem_flags; 1291 sljit_s32 op_flags = GET_ALL_FLAGS(op); 1292 1293 CHECK_ERROR(); 1294 CHECK(check_sljit_emit_op1(compiler, op, dst, dstw, src, srcw)); 1295 ADJUST_LOCAL_OFFSET(dst, dstw); 1296 ADJUST_LOCAL_OFFSET(src, srcw); 1297 1298 compiler->cache_arg = 0; 1299 compiler->cache_argw = 0; 1300 1301 dst_r = SLOW_IS_REG(dst) ? dst : TMP_REG1; 1302 1303 op = GET_OPCODE(op); 1304 if (op >= SLJIT_MOV && op <= SLJIT_MOVU_P) { 1305 switch (op) { 1306 case SLJIT_MOV: 1307 case SLJIT_MOV_P: 1308 flags = WORD_SIZE; 1309 break; 1310 case SLJIT_MOV_U8: 1311 flags = BYTE_SIZE; 1312 if (src & SLJIT_IMM) 1313 srcw = (sljit_u8)srcw; 1314 break; 1315 case SLJIT_MOV_S8: 1316 flags = BYTE_SIZE | SIGNED; 1317 if (src & SLJIT_IMM) 1318 srcw = (sljit_s8)srcw; 1319 break; 1320 case SLJIT_MOV_U16: 1321 flags = HALF_SIZE; 1322 if (src & SLJIT_IMM) 1323 srcw = (sljit_u16)srcw; 1324 break; 1325 case SLJIT_MOV_S16: 1326 flags = HALF_SIZE | SIGNED; 1327 if (src & SLJIT_IMM) 1328 srcw = (sljit_s16)srcw; 1329 break; 1330 case SLJIT_MOV_U32: 1331 flags = INT_SIZE; 1332 if (src & SLJIT_IMM) 1333 srcw = (sljit_u32)srcw; 1334 break; 1335 case SLJIT_MOV_S32: 1336 flags = INT_SIZE | SIGNED; 1337 if (src & SLJIT_IMM) 1338 srcw = (sljit_s32)srcw; 1339 break; 1340 case SLJIT_MOVU: 1341 case SLJIT_MOVU_P: 1342 flags = WORD_SIZE | UPDATE; 1343 break; 1344 case SLJIT_MOVU_U8: 1345 flags = BYTE_SIZE | UPDATE; 1346 if (src & SLJIT_IMM) 1347 srcw = (sljit_u8)srcw; 1348 break; 1349 case SLJIT_MOVU_S8: 1350 flags = BYTE_SIZE | SIGNED | UPDATE; 1351 if (src & SLJIT_IMM) 1352 srcw = (sljit_s8)srcw; 1353 break; 1354 case SLJIT_MOVU_U16: 1355 flags = HALF_SIZE | UPDATE; 1356 if (src & SLJIT_IMM) 1357 srcw = (sljit_u16)srcw; 1358 break; 1359 case SLJIT_MOVU_S16: 1360 flags = HALF_SIZE | SIGNED | UPDATE; 1361 if (src & SLJIT_IMM) 1362 srcw = (sljit_s16)srcw; 1363 break; 1364 case SLJIT_MOVU_U32: 1365 flags = INT_SIZE | UPDATE; 1366 if (src & SLJIT_IMM) 1367 srcw = (sljit_u32)srcw; 1368 break; 1369 case SLJIT_MOVU_S32: 1370 flags = INT_SIZE | SIGNED | UPDATE; 1371 if (src & SLJIT_IMM) 1372 srcw = (sljit_s32)srcw; 1373 break; 1374 default: 1375 SLJIT_UNREACHABLE(); 1376 flags = 0; 1377 break; 1378 } 1379 1380 if (src & SLJIT_IMM) 1381 FAIL_IF(emit_op_imm(compiler, SLJIT_MOV | ARG2_IMM, dst_r, TMP_REG1, srcw)); 1382 else if (src & SLJIT_MEM) { 1383 if (getput_arg_fast(compiler, flags, dst_r, src, srcw)) 1384 FAIL_IF(compiler->error); 1385 else 1386 FAIL_IF(getput_arg(compiler, flags, dst_r, src, srcw, dst, dstw)); 1387 } else { 1388 if (dst_r != TMP_REG1) 1389 return emit_op_imm(compiler, op | ((op_flags & SLJIT_I32_OP) ? INT_OP : 0), dst_r, TMP_REG1, src); 1390 dst_r = src; 1391 } 1392 1393 if (dst & SLJIT_MEM) { 1394 if (getput_arg_fast(compiler, flags | STORE, dst_r, dst, dstw)) 1395 return compiler->error; 1396 else 1397 return getput_arg(compiler, flags | STORE, dst_r, dst, dstw, 0, 0); 1398 } 1399 return SLJIT_SUCCESS; 1400 } 1401 1402 flags = HAS_FLAGS(op_flags) ? SET_FLAGS : 0; 1403 mem_flags = WORD_SIZE; 1404 if (op_flags & SLJIT_I32_OP) { 1405 flags |= INT_OP; 1406 mem_flags = INT_SIZE; 1407 } 1408 1409 if (dst == SLJIT_UNUSED) 1410 flags |= UNUSED_RETURN; 1411 1412 if (src & SLJIT_MEM) { 1413 if (getput_arg_fast(compiler, mem_flags, TMP_REG2, src, srcw)) 1414 FAIL_IF(compiler->error); 1415 else 1416 FAIL_IF(getput_arg(compiler, mem_flags, TMP_REG2, src, srcw, dst, dstw)); 1417 src = TMP_REG2; 1418 } 1419 1420 if (src & SLJIT_IMM) { 1421 flags |= ARG2_IMM; 1422 if (op_flags & SLJIT_I32_OP) 1423 srcw = (sljit_s32)srcw; 1424 } else 1425 srcw = src; 1426 1427 emit_op_imm(compiler, flags | op, dst_r, TMP_REG1, srcw); 1428 1429 if (dst & SLJIT_MEM) { 1430 if (getput_arg_fast(compiler, mem_flags | STORE, dst_r, dst, dstw)) 1431 return compiler->error; 1432 else 1433 return getput_arg(compiler, mem_flags | STORE, dst_r, dst, dstw, 0, 0); 1434 } 1435 return SLJIT_SUCCESS; 1436} 1437 1438SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op2(struct sljit_compiler *compiler, sljit_s32 op, 1439 sljit_s32 dst, sljit_sw dstw, 1440 sljit_s32 src1, sljit_sw src1w, 1441 sljit_s32 src2, sljit_sw src2w) 1442{ 1443 sljit_s32 dst_r, flags, mem_flags; 1444 1445 CHECK_ERROR(); 1446 CHECK(check_sljit_emit_op2(compiler, op, dst, dstw, src1, src1w, src2, src2w)); 1447 ADJUST_LOCAL_OFFSET(dst, dstw); 1448 ADJUST_LOCAL_OFFSET(src1, src1w); 1449 ADJUST_LOCAL_OFFSET(src2, src2w); 1450 1451 compiler->cache_arg = 0; 1452 compiler->cache_argw = 0; 1453 1454 dst_r = SLOW_IS_REG(dst) ? dst : TMP_REG1; 1455 flags = HAS_FLAGS(op) ? SET_FLAGS : 0; 1456 mem_flags = WORD_SIZE; 1457 if (op & SLJIT_I32_OP) { 1458 flags |= INT_OP; 1459 mem_flags = INT_SIZE; 1460 } 1461 1462 if (dst == SLJIT_UNUSED) 1463 flags |= UNUSED_RETURN; 1464 1465 if ((dst & SLJIT_MEM) && !getput_arg_fast(compiler, mem_flags | STORE | ARG_TEST, TMP_REG1, dst, dstw)) 1466 flags |= SLOW_DEST; 1467 1468 if (src1 & SLJIT_MEM) { 1469 if (getput_arg_fast(compiler, mem_flags, TMP_REG1, src1, src1w)) 1470 FAIL_IF(compiler->error); 1471 else 1472 flags |= SLOW_SRC1; 1473 } 1474 if (src2 & SLJIT_MEM) { 1475 if (getput_arg_fast(compiler, mem_flags, TMP_REG2, src2, src2w)) 1476 FAIL_IF(compiler->error); 1477 else 1478 flags |= SLOW_SRC2; 1479 } 1480 1481 if ((flags & (SLOW_SRC1 | SLOW_SRC2)) == (SLOW_SRC1 | SLOW_SRC2)) { 1482 if (!can_cache(src1, src1w, src2, src2w) && can_cache(src1, src1w, dst, dstw)) { 1483 FAIL_IF(getput_arg(compiler, mem_flags, TMP_REG2, src2, src2w, src1, src1w)); 1484 FAIL_IF(getput_arg(compiler, mem_flags, TMP_REG1, src1, src1w, dst, dstw)); 1485 } 1486 else { 1487 FAIL_IF(getput_arg(compiler, mem_flags, TMP_REG1, src1, src1w, src2, src2w)); 1488 FAIL_IF(getput_arg(compiler, mem_flags, TMP_REG2, src2, src2w, dst, dstw)); 1489 } 1490 } 1491 else if (flags & SLOW_SRC1) 1492 FAIL_IF(getput_arg(compiler, mem_flags, TMP_REG1, src1, src1w, dst, dstw)); 1493 else if (flags & SLOW_SRC2) 1494 FAIL_IF(getput_arg(compiler, mem_flags, TMP_REG2, src2, src2w, dst, dstw)); 1495 1496 if (src1 & SLJIT_MEM) 1497 src1 = TMP_REG1; 1498 if (src2 & SLJIT_MEM) 1499 src2 = TMP_REG2; 1500 1501 if (src1 & SLJIT_IMM) 1502 flags |= ARG1_IMM; 1503 else 1504 src1w = src1; 1505 if (src2 & SLJIT_IMM) 1506 flags |= ARG2_IMM; 1507 else 1508 src2w = src2; 1509 1510 emit_op_imm(compiler, flags | GET_OPCODE(op), dst_r, src1w, src2w); 1511 1512 if (dst & SLJIT_MEM) { 1513 if (!(flags & SLOW_DEST)) { 1514 getput_arg_fast(compiler, mem_flags | STORE, dst_r, dst, dstw); 1515 return compiler->error; 1516 } 1517 return getput_arg(compiler, mem_flags | STORE, TMP_REG1, dst, dstw, 0, 0); 1518 } 1519 1520 return SLJIT_SUCCESS; 1521} 1522 1523SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_get_register_index(sljit_s32 reg) 1524{ 1525 CHECK_REG_INDEX(check_sljit_get_register_index(reg)); 1526 return reg_map[reg]; 1527} 1528 1529SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_get_float_register_index(sljit_s32 reg) 1530{ 1531 CHECK_REG_INDEX(check_sljit_get_float_register_index(reg)); 1532 return reg; 1533} 1534 1535SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op_custom(struct sljit_compiler *compiler, 1536 void *instruction, sljit_s32 size) 1537{ 1538 CHECK_ERROR(); 1539 CHECK(check_sljit_emit_op_custom(compiler, instruction, size)); 1540 1541 return push_inst(compiler, *(sljit_ins*)instruction); 1542} 1543 1544/* --------------------------------------------------------------------- */ 1545/* Floating point operators */ 1546/* --------------------------------------------------------------------- */ 1547 1548SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_is_fpu_available(void) 1549{ 1550#ifdef SLJIT_IS_FPU_AVAILABLE 1551 return SLJIT_IS_FPU_AVAILABLE; 1552#else 1553 /* Available by default. */ 1554 return 1; 1555#endif 1556} 1557 1558static sljit_s32 emit_fop_mem(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg, sljit_s32 arg, sljit_sw argw) 1559{ 1560 sljit_u32 shift = MEM_SIZE_SHIFT(flags); 1561 sljit_ins ins_bits = (shift << 30); 1562 sljit_s32 other_r; 1563 sljit_sw diff; 1564 1565 SLJIT_ASSERT(arg & SLJIT_MEM); 1566 1567 if (!(flags & STORE)) 1568 ins_bits |= 1 << 22; 1569 1570 if (arg & OFFS_REG_MASK) { 1571 argw &= 3; 1572 if (!argw || argw == shift) 1573 return push_inst(compiler, STR_FR | ins_bits | VT(reg) 1574 | RN(arg & REG_MASK) | RM(OFFS_REG(arg)) | (argw ? (1 << 12) : 0)); 1575 other_r = OFFS_REG(arg); 1576 arg &= REG_MASK; 1577 FAIL_IF(push_inst(compiler, ADD | RD(TMP_REG1) | RN(arg) | RM(other_r) | (argw << 10))); 1578 arg = TMP_REG1; 1579 argw = 0; 1580 } 1581 1582 arg &= REG_MASK; 1583 if (arg && argw >= 0 && ((argw >> shift) <= 0xfff) && (argw & ((1 << shift) - 1)) == 0) 1584 return push_inst(compiler, STR_FI | ins_bits | VT(reg) | RN(arg) | (argw << (10 - shift))); 1585 1586 if (arg && argw <= 255 && argw >= -256) 1587 return push_inst(compiler, STUR_FI | ins_bits | VT(reg) | RN(arg) | ((argw & 0x1ff) << 12)); 1588 1589 /* Slow cases */ 1590 if (compiler->cache_arg == SLJIT_MEM && argw != compiler->cache_argw) { 1591 diff = argw - compiler->cache_argw; 1592 if (!arg && diff <= 255 && diff >= -256) 1593 return push_inst(compiler, STUR_FI | ins_bits | VT(reg) | RN(TMP_REG3) | ((diff & 0x1ff) << 12)); 1594 if (emit_set_delta(compiler, TMP_REG3, TMP_REG3, argw - compiler->cache_argw) != SLJIT_ERR_UNSUPPORTED) { 1595 FAIL_IF(compiler->error); 1596 compiler->cache_argw = argw; 1597 } 1598 } 1599 1600 if (compiler->cache_arg != SLJIT_MEM || argw != compiler->cache_argw) { 1601 compiler->cache_arg = SLJIT_MEM; 1602 compiler->cache_argw = argw; 1603 FAIL_IF(load_immediate(compiler, TMP_REG3, argw)); 1604 } 1605 1606 if (arg & REG_MASK) 1607 return push_inst(compiler, STR_FR | ins_bits | VT(reg) | RN(arg) | RM(TMP_REG3)); 1608 return push_inst(compiler, STR_FI | ins_bits | VT(reg) | RN(TMP_REG3)); 1609} 1610 1611static SLJIT_INLINE sljit_s32 sljit_emit_fop1_conv_sw_from_f64(struct sljit_compiler *compiler, sljit_s32 op, 1612 sljit_s32 dst, sljit_sw dstw, 1613 sljit_s32 src, sljit_sw srcw) 1614{ 1615 sljit_s32 dst_r = SLOW_IS_REG(dst) ? dst : TMP_REG1; 1616 sljit_ins inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0; 1617 1618 if (GET_OPCODE(op) == SLJIT_CONV_S32_FROM_F64) 1619 inv_bits |= (1 << 31); 1620 1621 if (src & SLJIT_MEM) { 1622 emit_fop_mem(compiler, (op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE, TMP_FREG1, src, srcw); 1623 src = TMP_FREG1; 1624 } 1625 1626 FAIL_IF(push_inst(compiler, (FCVTZS ^ inv_bits) | RD(dst_r) | VN(src))); 1627 1628 if (dst_r == TMP_REG1 && dst != SLJIT_UNUSED) 1629 return emit_op_mem(compiler, ((GET_OPCODE(op) == SLJIT_CONV_S32_FROM_F64) ? INT_SIZE : WORD_SIZE) | STORE, TMP_REG1, dst, dstw); 1630 return SLJIT_SUCCESS; 1631} 1632 1633static SLJIT_INLINE sljit_s32 sljit_emit_fop1_conv_f64_from_sw(struct sljit_compiler *compiler, sljit_s32 op, 1634 sljit_s32 dst, sljit_sw dstw, 1635 sljit_s32 src, sljit_sw srcw) 1636{ 1637 sljit_s32 dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG1; 1638 sljit_ins inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0; 1639 1640 if (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_S32) 1641 inv_bits |= (1 << 31); 1642 1643 if (src & SLJIT_MEM) { 1644 emit_op_mem(compiler, ((GET_OPCODE(op) == SLJIT_CONV_F64_FROM_S32) ? INT_SIZE : WORD_SIZE), TMP_REG1, src, srcw); 1645 src = TMP_REG1; 1646 } else if (src & SLJIT_IMM) { 1647#if (defined SLJIT_CONFIG_X86_64 && SLJIT_CONFIG_X86_64) 1648 if (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_S32) 1649 srcw = (sljit_s32)srcw; 1650#endif 1651 FAIL_IF(load_immediate(compiler, TMP_REG1, srcw)); 1652 src = TMP_REG1; 1653 } 1654 1655 FAIL_IF(push_inst(compiler, (SCVTF ^ inv_bits) | VD(dst_r) | RN(src))); 1656 1657 if (dst & SLJIT_MEM) 1658 return emit_fop_mem(compiler, ((op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE) | STORE, TMP_FREG1, dst, dstw); 1659 return SLJIT_SUCCESS; 1660} 1661 1662static SLJIT_INLINE sljit_s32 sljit_emit_fop1_cmp(struct sljit_compiler *compiler, sljit_s32 op, 1663 sljit_s32 src1, sljit_sw src1w, 1664 sljit_s32 src2, sljit_sw src2w) 1665{ 1666 sljit_s32 mem_flags = (op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE; 1667 sljit_ins inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0; 1668 1669 if (src1 & SLJIT_MEM) { 1670 emit_fop_mem(compiler, mem_flags, TMP_FREG1, src1, src1w); 1671 src1 = TMP_FREG1; 1672 } 1673 1674 if (src2 & SLJIT_MEM) { 1675 emit_fop_mem(compiler, mem_flags, TMP_FREG2, src2, src2w); 1676 src2 = TMP_FREG2; 1677 } 1678 1679 return push_inst(compiler, (FCMP ^ inv_bits) | VN(src1) | VM(src2)); 1680} 1681 1682SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fop1(struct sljit_compiler *compiler, sljit_s32 op, 1683 sljit_s32 dst, sljit_sw dstw, 1684 sljit_s32 src, sljit_sw srcw) 1685{ 1686 sljit_s32 dst_r, mem_flags = (op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE; 1687 sljit_ins inv_bits; 1688 1689 CHECK_ERROR(); 1690 compiler->cache_arg = 0; 1691 compiler->cache_argw = 0; 1692 1693 SLJIT_COMPILE_ASSERT((INT_SIZE ^ 0x100) == WORD_SIZE, must_be_one_bit_difference); 1694 SELECT_FOP1_OPERATION_WITH_CHECKS(compiler, op, dst, dstw, src, srcw); 1695 1696 inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0; 1697 dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG1; 1698 1699 if (src & SLJIT_MEM) { 1700 emit_fop_mem(compiler, (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_F32) ? (mem_flags ^ 0x100) : mem_flags, dst_r, src, srcw); 1701 src = dst_r; 1702 } 1703 1704 switch (GET_OPCODE(op)) { 1705 case SLJIT_MOV_F64: 1706 if (src != dst_r) { 1707 if (dst_r != TMP_FREG1) 1708 FAIL_IF(push_inst(compiler, (FMOV ^ inv_bits) | VD(dst_r) | VN(src))); 1709 else 1710 dst_r = src; 1711 } 1712 break; 1713 case SLJIT_NEG_F64: 1714 FAIL_IF(push_inst(compiler, (FNEG ^ inv_bits) | VD(dst_r) | VN(src))); 1715 break; 1716 case SLJIT_ABS_F64: 1717 FAIL_IF(push_inst(compiler, (FABS ^ inv_bits) | VD(dst_r) | VN(src))); 1718 break; 1719 case SLJIT_CONV_F64_FROM_F32: 1720 FAIL_IF(push_inst(compiler, FCVT | ((op & SLJIT_F32_OP) ? (1 << 22) : (1 << 15)) | VD(dst_r) | VN(src))); 1721 break; 1722 } 1723 1724 if (dst & SLJIT_MEM) 1725 return emit_fop_mem(compiler, mem_flags | STORE, dst_r, dst, dstw); 1726 return SLJIT_SUCCESS; 1727} 1728 1729SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fop2(struct sljit_compiler *compiler, sljit_s32 op, 1730 sljit_s32 dst, sljit_sw dstw, 1731 sljit_s32 src1, sljit_sw src1w, 1732 sljit_s32 src2, sljit_sw src2w) 1733{ 1734 sljit_s32 dst_r, mem_flags = (op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE; 1735 sljit_ins inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0; 1736 1737 CHECK_ERROR(); 1738 CHECK(check_sljit_emit_fop2(compiler, op, dst, dstw, src1, src1w, src2, src2w)); 1739 ADJUST_LOCAL_OFFSET(dst, dstw); 1740 ADJUST_LOCAL_OFFSET(src1, src1w); 1741 ADJUST_LOCAL_OFFSET(src2, src2w); 1742 1743 compiler->cache_arg = 0; 1744 compiler->cache_argw = 0; 1745 1746 dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG1; 1747 if (src1 & SLJIT_MEM) { 1748 emit_fop_mem(compiler, mem_flags, TMP_FREG1, src1, src1w); 1749 src1 = TMP_FREG1; 1750 } 1751 if (src2 & SLJIT_MEM) { 1752 emit_fop_mem(compiler, mem_flags, TMP_FREG2, src2, src2w); 1753 src2 = TMP_FREG2; 1754 } 1755 1756 switch (GET_OPCODE(op)) { 1757 case SLJIT_ADD_F64: 1758 FAIL_IF(push_inst(compiler, (FADD ^ inv_bits) | VD(dst_r) | VN(src1) | VM(src2))); 1759 break; 1760 case SLJIT_SUB_F64: 1761 FAIL_IF(push_inst(compiler, (FSUB ^ inv_bits) | VD(dst_r) | VN(src1) | VM(src2))); 1762 break; 1763 case SLJIT_MUL_F64: 1764 FAIL_IF(push_inst(compiler, (FMUL ^ inv_bits) | VD(dst_r) | VN(src1) | VM(src2))); 1765 break; 1766 case SLJIT_DIV_F64: 1767 FAIL_IF(push_inst(compiler, (FDIV ^ inv_bits) | VD(dst_r) | VN(src1) | VM(src2))); 1768 break; 1769 } 1770 1771 if (!(dst & SLJIT_MEM)) 1772 return SLJIT_SUCCESS; 1773 return emit_fop_mem(compiler, mem_flags | STORE, TMP_FREG1, dst, dstw); 1774} 1775 1776/* --------------------------------------------------------------------- */ 1777/* Other instructions */ 1778/* --------------------------------------------------------------------- */ 1779 1780SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fast_enter(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw) 1781{ 1782 CHECK_ERROR(); 1783 CHECK(check_sljit_emit_fast_enter(compiler, dst, dstw)); 1784 ADJUST_LOCAL_OFFSET(dst, dstw); 1785 1786 /* For UNUSED dst. Uncommon, but possible. */ 1787 if (dst == SLJIT_UNUSED) 1788 return SLJIT_SUCCESS; 1789 1790 if (FAST_IS_REG(dst)) 1791 return push_inst(compiler, ORR | RD(dst) | RN(TMP_ZERO) | RM(TMP_LR)); 1792 1793 /* Memory. */ 1794 return emit_op_mem(compiler, WORD_SIZE | STORE, TMP_LR, dst, dstw); 1795} 1796 1797SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fast_return(struct sljit_compiler *compiler, sljit_s32 src, sljit_sw srcw) 1798{ 1799 CHECK_ERROR(); 1800 CHECK(check_sljit_emit_fast_return(compiler, src, srcw)); 1801 ADJUST_LOCAL_OFFSET(src, srcw); 1802 1803 if (FAST_IS_REG(src)) 1804 FAIL_IF(push_inst(compiler, ORR | RD(TMP_LR) | RN(TMP_ZERO) | RM(src))); 1805 else if (src & SLJIT_MEM) 1806 FAIL_IF(emit_op_mem(compiler, WORD_SIZE, TMP_LR, src, srcw)); 1807 else if (src & SLJIT_IMM) 1808 FAIL_IF(load_immediate(compiler, TMP_LR, srcw)); 1809 1810 return push_inst(compiler, RET | RN(TMP_LR)); 1811} 1812 1813/* --------------------------------------------------------------------- */ 1814/* Conditional instructions */ 1815/* --------------------------------------------------------------------- */ 1816 1817static sljit_uw get_cc(sljit_s32 type) 1818{ 1819 switch (type) { 1820 case SLJIT_EQUAL: 1821 case SLJIT_MUL_NOT_OVERFLOW: 1822 case SLJIT_EQUAL_F64: 1823 return 0x1; 1824 1825 case SLJIT_NOT_EQUAL: 1826 case SLJIT_MUL_OVERFLOW: 1827 case SLJIT_NOT_EQUAL_F64: 1828 return 0x0; 1829 1830 case SLJIT_LESS: 1831 case SLJIT_LESS_F64: 1832 return 0x2; 1833 1834 case SLJIT_GREATER_EQUAL: 1835 case SLJIT_GREATER_EQUAL_F64: 1836 return 0x3; 1837 1838 case SLJIT_GREATER: 1839 case SLJIT_GREATER_F64: 1840 return 0x9; 1841 1842 case SLJIT_LESS_EQUAL: 1843 case SLJIT_LESS_EQUAL_F64: 1844 return 0x8; 1845 1846 case SLJIT_SIG_LESS: 1847 return 0xa; 1848 1849 case SLJIT_SIG_GREATER_EQUAL: 1850 return 0xb; 1851 1852 case SLJIT_SIG_GREATER: 1853 return 0xd; 1854 1855 case SLJIT_SIG_LESS_EQUAL: 1856 return 0xc; 1857 1858 case SLJIT_OVERFLOW: 1859 case SLJIT_UNORDERED_F64: 1860 return 0x7; 1861 1862 case SLJIT_NOT_OVERFLOW: 1863 case SLJIT_ORDERED_F64: 1864 return 0x6; 1865 1866 default: 1867 SLJIT_UNREACHABLE(); 1868 return 0xe; 1869 } 1870} 1871 1872SLJIT_API_FUNC_ATTRIBUTE struct sljit_label* sljit_emit_label(struct sljit_compiler *compiler) 1873{ 1874 struct sljit_label *label; 1875 1876 CHECK_ERROR_PTR(); 1877 CHECK_PTR(check_sljit_emit_label(compiler)); 1878 1879 if (compiler->last_label && compiler->last_label->size == compiler->size) 1880 return compiler->last_label; 1881 1882 label = (struct sljit_label*)ensure_abuf(compiler, sizeof(struct sljit_label)); 1883 PTR_FAIL_IF(!label); 1884 set_label(label, compiler); 1885 return label; 1886} 1887 1888SLJIT_API_FUNC_ATTRIBUTE struct sljit_jump* sljit_emit_jump(struct sljit_compiler *compiler, sljit_s32 type) 1889{ 1890 struct sljit_jump *jump; 1891 1892 CHECK_ERROR_PTR(); 1893 CHECK_PTR(check_sljit_emit_jump(compiler, type)); 1894 1895 jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump)); 1896 PTR_FAIL_IF(!jump); 1897 set_jump(jump, compiler, type & SLJIT_REWRITABLE_JUMP); 1898 type &= 0xff; 1899 1900 if (type < SLJIT_JUMP) { 1901 jump->flags |= IS_COND; 1902 PTR_FAIL_IF(push_inst(compiler, B_CC | (6 << 5) | get_cc(type))); 1903 } 1904 else if (type >= SLJIT_FAST_CALL) 1905 jump->flags |= IS_BL; 1906 1907 PTR_FAIL_IF(emit_imm64_const(compiler, TMP_REG1, 0)); 1908 jump->addr = compiler->size; 1909 PTR_FAIL_IF(push_inst(compiler, ((type >= SLJIT_FAST_CALL) ? BLR : BR) | RN(TMP_REG1))); 1910 1911 return jump; 1912} 1913 1914static SLJIT_INLINE struct sljit_jump* emit_cmp_to0(struct sljit_compiler *compiler, sljit_s32 type, 1915 sljit_s32 src, sljit_sw srcw) 1916{ 1917 struct sljit_jump *jump; 1918 sljit_ins inv_bits = (type & SLJIT_I32_OP) ? (1 << 31) : 0; 1919 1920 SLJIT_ASSERT((type & 0xff) == SLJIT_EQUAL || (type & 0xff) == SLJIT_NOT_EQUAL); 1921 ADJUST_LOCAL_OFFSET(src, srcw); 1922 1923 jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump)); 1924 PTR_FAIL_IF(!jump); 1925 set_jump(jump, compiler, type & SLJIT_REWRITABLE_JUMP); 1926 jump->flags |= IS_CBZ | IS_COND; 1927 1928 if (src & SLJIT_MEM) { 1929 PTR_FAIL_IF(emit_op_mem(compiler, inv_bits ? INT_SIZE : WORD_SIZE, TMP_REG1, src, srcw)); 1930 src = TMP_REG1; 1931 } 1932 else if (src & SLJIT_IMM) { 1933 PTR_FAIL_IF(load_immediate(compiler, TMP_REG1, srcw)); 1934 src = TMP_REG1; 1935 } 1936 SLJIT_ASSERT(FAST_IS_REG(src)); 1937 1938 if ((type & 0xff) == SLJIT_EQUAL) 1939 inv_bits |= 1 << 24; 1940 1941 PTR_FAIL_IF(push_inst(compiler, (CBZ ^ inv_bits) | (6 << 5) | RT(src))); 1942 PTR_FAIL_IF(emit_imm64_const(compiler, TMP_REG1, 0)); 1943 jump->addr = compiler->size; 1944 PTR_FAIL_IF(push_inst(compiler, BR | RN(TMP_REG1))); 1945 return jump; 1946} 1947 1948SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_ijump(struct sljit_compiler *compiler, sljit_s32 type, sljit_s32 src, sljit_sw srcw) 1949{ 1950 struct sljit_jump *jump; 1951 1952 CHECK_ERROR(); 1953 CHECK(check_sljit_emit_ijump(compiler, type, src, srcw)); 1954 ADJUST_LOCAL_OFFSET(src, srcw); 1955 1956 /* In ARM, we don't need to touch the arguments. */ 1957 if (!(src & SLJIT_IMM)) { 1958 if (src & SLJIT_MEM) { 1959 FAIL_IF(emit_op_mem(compiler, WORD_SIZE, TMP_REG1, src, srcw)); 1960 src = TMP_REG1; 1961 } 1962 return push_inst(compiler, ((type >= SLJIT_FAST_CALL) ? BLR : BR) | RN(src)); 1963 } 1964 1965 jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump)); 1966 FAIL_IF(!jump); 1967 set_jump(jump, compiler, JUMP_ADDR | ((type >= SLJIT_FAST_CALL) ? IS_BL : 0)); 1968 jump->u.target = srcw; 1969 1970 FAIL_IF(emit_imm64_const(compiler, TMP_REG1, 0)); 1971 jump->addr = compiler->size; 1972 return push_inst(compiler, ((type >= SLJIT_FAST_CALL) ? BLR : BR) | RN(TMP_REG1)); 1973} 1974 1975SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op_flags(struct sljit_compiler *compiler, sljit_s32 op, 1976 sljit_s32 dst, sljit_sw dstw, 1977 sljit_s32 src, sljit_sw srcw, 1978 sljit_s32 type) 1979{ 1980 sljit_s32 dst_r, flags, mem_flags; 1981 sljit_ins cc; 1982 1983 CHECK_ERROR(); 1984 CHECK(check_sljit_emit_op_flags(compiler, op, dst, dstw, src, srcw, type)); 1985 ADJUST_LOCAL_OFFSET(dst, dstw); 1986 ADJUST_LOCAL_OFFSET(src, srcw); 1987 1988 if (dst == SLJIT_UNUSED) 1989 return SLJIT_SUCCESS; 1990 1991 cc = get_cc(type & 0xff); 1992 dst_r = FAST_IS_REG(dst) ? dst : TMP_REG1; 1993 1994 if (GET_OPCODE(op) < SLJIT_ADD) { 1995 FAIL_IF(push_inst(compiler, CSINC | (cc << 12) | RD(dst_r) | RN(TMP_ZERO) | RM(TMP_ZERO))); 1996 if (dst_r != TMP_REG1) 1997 return SLJIT_SUCCESS; 1998 return emit_op_mem(compiler, (GET_OPCODE(op) == SLJIT_MOV ? WORD_SIZE : INT_SIZE) | STORE, TMP_REG1, dst, dstw); 1999 } 2000 2001 compiler->cache_arg = 0; 2002 compiler->cache_argw = 0; 2003 flags = HAS_FLAGS(op) ? SET_FLAGS : 0; 2004 mem_flags = WORD_SIZE; 2005 if (op & SLJIT_I32_OP) { 2006 flags |= INT_OP; 2007 mem_flags = INT_SIZE; 2008 } 2009 2010 if (src & SLJIT_MEM) { 2011 FAIL_IF(emit_op_mem2(compiler, mem_flags, TMP_REG1, src, srcw, dst, dstw)); 2012 src = TMP_REG1; 2013 srcw = 0; 2014 } else if (src & SLJIT_IMM) 2015 flags |= ARG1_IMM; 2016 2017 FAIL_IF(push_inst(compiler, CSINC | (cc << 12) | RD(TMP_REG2) | RN(TMP_ZERO) | RM(TMP_ZERO))); 2018 emit_op_imm(compiler, flags | GET_OPCODE(op), dst_r, src, TMP_REG2); 2019 2020 if (dst_r != TMP_REG1) 2021 return SLJIT_SUCCESS; 2022 return emit_op_mem2(compiler, mem_flags | STORE, TMP_REG1, dst, dstw, 0, 0); 2023} 2024 2025SLJIT_API_FUNC_ATTRIBUTE struct sljit_const* sljit_emit_const(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw, sljit_sw init_value) 2026{ 2027 struct sljit_const *const_; 2028 sljit_s32 dst_r; 2029 2030 CHECK_ERROR_PTR(); 2031 CHECK_PTR(check_sljit_emit_const(compiler, dst, dstw, init_value)); 2032 ADJUST_LOCAL_OFFSET(dst, dstw); 2033 2034 const_ = (struct sljit_const*)ensure_abuf(compiler, sizeof(struct sljit_const)); 2035 PTR_FAIL_IF(!const_); 2036 set_const(const_, compiler); 2037 2038 dst_r = SLOW_IS_REG(dst) ? dst : TMP_REG1; 2039 PTR_FAIL_IF(emit_imm64_const(compiler, dst_r, init_value)); 2040 2041 if (dst & SLJIT_MEM) 2042 PTR_FAIL_IF(emit_op_mem(compiler, WORD_SIZE | STORE, dst_r, dst, dstw)); 2043 return const_; 2044} 2045 2046SLJIT_API_FUNC_ATTRIBUTE void sljit_set_jump_addr(sljit_uw addr, sljit_uw new_target, sljit_sw executable_offset) 2047{ 2048 sljit_ins* inst = (sljit_ins*)addr; 2049 modify_imm64_const(inst, new_target); 2050 inst = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(inst, executable_offset); 2051 SLJIT_CACHE_FLUSH(inst, inst + 4); 2052} 2053 2054SLJIT_API_FUNC_ATTRIBUTE void sljit_set_const(sljit_uw addr, sljit_sw new_constant, sljit_sw executable_offset) 2055{ 2056 sljit_ins* inst = (sljit_ins*)addr; 2057 modify_imm64_const(inst, new_constant); 2058 inst = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(inst, executable_offset); 2059 SLJIT_CACHE_FLUSH(inst, inst + 4); 2060} 2061