1/*	$NetBSD: sljitNativeARM_64.c,v 1.5 2024/04/02 22:29:57 riastradh Exp $	*/
2
3/*
4 *    Stack-less Just-In-Time compiler
5 *
6 *    Copyright Zoltan Herczeg (hzmester@freemail.hu). All rights reserved.
7 *
8 * Redistribution and use in source and binary forms, with or without modification, are
9 * permitted provided that the following conditions are met:
10 *
11 *   1. Redistributions of source code must retain the above copyright notice, this list of
12 *      conditions and the following disclaimer.
13 *
14 *   2. Redistributions in binary form must reproduce the above copyright notice, this list
15 *      of conditions and the following disclaimer in the documentation and/or other materials
16 *      provided with the distribution.
17 *
18 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDER(S) AND CONTRIBUTORS ``AS IS'' AND ANY
19 * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
20 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT
21 * SHALL THE COPYRIGHT HOLDER(S) OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
22 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
23 * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
24 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
25 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
26 * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27 */
28
29SLJIT_API_FUNC_ATTRIBUTE const char* sljit_get_platform_name(void)
30{
31	return "ARM-64" SLJIT_CPUINFO;
32}
33
34/* Length of an instruction word */
35typedef sljit_u32 sljit_ins;
36
37#define TMP_ZERO	(0)
38
39#define TMP_REG1	(SLJIT_NUMBER_OF_REGISTERS + 2)
40#define TMP_REG2	(SLJIT_NUMBER_OF_REGISTERS + 3)
41#define TMP_REG3	(SLJIT_NUMBER_OF_REGISTERS + 4)
42#define TMP_LR		(SLJIT_NUMBER_OF_REGISTERS + 5)
43#define TMP_SP		(SLJIT_NUMBER_OF_REGISTERS + 6)
44
45#define TMP_FREG1	(0)
46#define TMP_FREG2	(SLJIT_NUMBER_OF_FLOAT_REGISTERS + 1)
47
48static const sljit_u8 reg_map[SLJIT_NUMBER_OF_REGISTERS + 8] = {
49  31, 0, 1, 2, 3, 4, 5, 6, 7, 12, 13, 14, 15, 16, 17, 8, 28, 27, 26, 25, 24, 23, 22, 21, 20, 19, 29, 9, 10, 11, 30, 31
50};
51
52#define W_OP (1 << 31)
53#define RD(rd) (reg_map[rd])
54#define RT(rt) (reg_map[rt])
55#define RN(rn) (reg_map[rn] << 5)
56#define RT2(rt2) (reg_map[rt2] << 10)
57#define RM(rm) (reg_map[rm] << 16)
58#define VD(vd) (vd)
59#define VT(vt) (vt)
60#define VN(vn) ((vn) << 5)
61#define VM(vm) ((vm) << 16)
62
63/* --------------------------------------------------------------------- */
64/*  Instrucion forms                                                     */
65/* --------------------------------------------------------------------- */
66
67#define ADC 0x9a000000
68#define ADD 0x8b000000
69#define ADDI 0x91000000
70#define AND 0x8a000000
71#define ANDI 0x92000000
72#define ASRV 0x9ac02800
73#define B 0x14000000
74#define B_CC 0x54000000
75#define BL 0x94000000
76#define BLR 0xd63f0000
77#define BR 0xd61f0000
78#define BRK 0xd4200000
79#define CBZ 0xb4000000
80#define CLZ 0xdac01000
81#define CSINC 0x9a800400
82#define EOR 0xca000000
83#define EORI 0xd2000000
84#define FABS 0x1e60c000
85#define FADD 0x1e602800
86#define FCMP 0x1e602000
87#define FCVT 0x1e224000
88#define FCVTZS 0x9e780000
89#define FDIV 0x1e601800
90#define FMOV 0x1e604000
91#define FMUL 0x1e600800
92#define FNEG 0x1e614000
93#define FSUB 0x1e603800
94#define LDRI 0xf9400000
95#define LDP 0xa9400000
96#define LDP_PST 0xa8c00000
97#define LSLV 0x9ac02000
98#define LSRV 0x9ac02400
99#define MADD 0x9b000000
100#define MOVK 0xf2800000
101#define MOVN 0x92800000
102#define MOVZ 0xd2800000
103#define NOP 0xd503201f
104#define ORN 0xaa200000
105#define ORR 0xaa000000
106#define ORRI 0xb2000000
107#define RET 0xd65f0000
108#define SBC 0xda000000
109#define SBFM 0x93000000
110#define SCVTF 0x9e620000
111#define SDIV 0x9ac00c00
112#define SMADDL 0x9b200000
113#define SMULH 0x9b403c00
114#define STP 0xa9000000
115#define STP_PRE 0xa9800000
116#define STRI 0xf9000000
117#define STR_FI 0x3d000000
118#define STR_FR 0x3c206800
119#define STUR_FI 0x3c000000
120#define SUB 0xcb000000
121#define SUBI 0xd1000000
122#define SUBS 0xeb000000
123#define UBFM 0xd3000000
124#define UDIV 0x9ac00800
125#define UMULH 0x9bc03c00
126
127/* dest_reg is the absolute name of the register
128   Useful for reordering instructions in the delay slot. */
129static sljit_s32 push_inst(struct sljit_compiler *compiler, sljit_ins ins)
130{
131	sljit_ins *ptr = (sljit_ins*)ensure_buf(compiler, sizeof(sljit_ins));
132	FAIL_IF(!ptr);
133	*ptr = ins;
134	compiler->size++;
135	return SLJIT_SUCCESS;
136}
137
138static SLJIT_INLINE sljit_s32 emit_imm64_const(struct sljit_compiler *compiler, sljit_s32 dst, sljit_uw imm)
139{
140	FAIL_IF(push_inst(compiler, MOVZ | RD(dst) | ((imm & 0xffff) << 5)));
141	FAIL_IF(push_inst(compiler, MOVK | RD(dst) | (((imm >> 16) & 0xffff) << 5) | (1 << 21)));
142	FAIL_IF(push_inst(compiler, MOVK | RD(dst) | (((imm >> 32) & 0xffff) << 5) | (2 << 21)));
143	return push_inst(compiler, MOVK | RD(dst) | ((imm >> 48) << 5) | (3 << 21));
144}
145
146static SLJIT_INLINE void modify_imm64_const(sljit_ins* inst, sljit_uw new_imm)
147{
148	sljit_s32 dst = inst[0] & 0x1f;
149	SLJIT_ASSERT((inst[0] & 0xffe00000) == MOVZ && (inst[1] & 0xffe00000) == (MOVK | (1 << 21)));
150	inst[0] = MOVZ | dst | ((new_imm & 0xffff) << 5);
151	inst[1] = MOVK | dst | (((new_imm >> 16) & 0xffff) << 5) | (1 << 21);
152	inst[2] = MOVK | dst | (((new_imm >> 32) & 0xffff) << 5) | (2 << 21);
153	inst[3] = MOVK | dst | ((new_imm >> 48) << 5) | (3 << 21);
154}
155
156static SLJIT_INLINE sljit_s32 detect_jump_type(struct sljit_jump *jump, sljit_ins *code_ptr, sljit_ins *code, sljit_sw executable_offset)
157{
158	sljit_sw diff;
159	sljit_uw target_addr;
160
161	if (jump->flags & SLJIT_REWRITABLE_JUMP) {
162		jump->flags |= PATCH_ABS64;
163		return 0;
164	}
165
166	if (jump->flags & JUMP_ADDR)
167		target_addr = jump->u.target;
168	else {
169		SLJIT_ASSERT(jump->flags & JUMP_LABEL);
170		target_addr = (sljit_uw)(code + jump->u.label->size) + (sljit_uw)executable_offset;
171	}
172
173	diff = (sljit_sw)target_addr - (sljit_sw)(code_ptr + 4) - executable_offset;
174
175	if (jump->flags & IS_COND) {
176		diff += sizeof(sljit_ins);
177		if (diff <= 0xfffff && diff >= -0x100000) {
178			code_ptr[-5] ^= (jump->flags & IS_CBZ) ? (0x1 << 24) : 0x1;
179			jump->addr -= sizeof(sljit_ins);
180			jump->flags |= PATCH_COND;
181			return 5;
182		}
183		diff -= sizeof(sljit_ins);
184	}
185
186	if (diff <= 0x7ffffff && diff >= -0x8000000) {
187		jump->flags |= PATCH_B;
188		return 4;
189	}
190
191	if (target_addr <= 0xffffffffl) {
192		if (jump->flags & IS_COND)
193			code_ptr[-5] -= (2 << 5);
194		code_ptr[-2] = code_ptr[0];
195		return 2;
196	}
197	if (target_addr <= 0xffffffffffffl) {
198		if (jump->flags & IS_COND)
199			code_ptr[-5] -= (1 << 5);
200		jump->flags |= PATCH_ABS48;
201		code_ptr[-1] = code_ptr[0];
202		return 1;
203	}
204
205	jump->flags |= PATCH_ABS64;
206	return 0;
207}
208
209SLJIT_API_FUNC_ATTRIBUTE void* sljit_generate_code(struct sljit_compiler *compiler)
210{
211	struct sljit_memory_fragment *buf;
212	sljit_ins *code;
213	sljit_ins *code_ptr;
214	sljit_ins *buf_ptr;
215	sljit_ins *buf_end;
216	sljit_uw word_count;
217	sljit_sw executable_offset;
218	sljit_uw addr;
219	sljit_s32 dst;
220
221	struct sljit_label *label;
222	struct sljit_jump *jump;
223	struct sljit_const *const_;
224
225	CHECK_ERROR_PTR();
226	CHECK_PTR(check_sljit_generate_code(compiler));
227	reverse_buf(compiler);
228
229	code = (sljit_ins*)SLJIT_MALLOC_EXEC(compiler->size * sizeof(sljit_ins));
230	PTR_FAIL_WITH_EXEC_IF(code);
231	buf = compiler->buf;
232
233	code_ptr = code;
234	word_count = 0;
235	executable_offset = SLJIT_EXEC_OFFSET(code);
236
237	label = compiler->labels;
238	jump = compiler->jumps;
239	const_ = compiler->consts;
240
241	do {
242		buf_ptr = (sljit_ins*)buf->memory;
243		buf_end = buf_ptr + (buf->used_size >> 2);
244		do {
245			*code_ptr = *buf_ptr++;
246			/* These structures are ordered by their address. */
247			SLJIT_ASSERT(!label || label->size >= word_count);
248			SLJIT_ASSERT(!jump || jump->addr >= word_count);
249			SLJIT_ASSERT(!const_ || const_->addr >= word_count);
250			if (label && label->size == word_count) {
251				label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset);
252				label->size = code_ptr - code;
253				label = label->next;
254			}
255			if (jump && jump->addr == word_count) {
256					jump->addr = (sljit_uw)(code_ptr - 4);
257					code_ptr -= detect_jump_type(jump, code_ptr, code, executable_offset);
258					jump = jump->next;
259			}
260			if (const_ && const_->addr == word_count) {
261				const_->addr = (sljit_uw)code_ptr;
262				const_ = const_->next;
263			}
264			code_ptr ++;
265			word_count ++;
266		} while (buf_ptr < buf_end);
267
268		buf = buf->next;
269	} while (buf);
270
271	if (label && label->size == word_count) {
272		label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset);
273		label->size = code_ptr - code;
274		label = label->next;
275	}
276
277	SLJIT_ASSERT(!label);
278	SLJIT_ASSERT(!jump);
279	SLJIT_ASSERT(!const_);
280	SLJIT_ASSERT(code_ptr - code <= (sljit_sw)compiler->size);
281
282	jump = compiler->jumps;
283	while (jump) {
284		do {
285			addr = (jump->flags & JUMP_LABEL) ? jump->u.label->addr : jump->u.target;
286			buf_ptr = (sljit_ins *)jump->addr;
287
288			if (jump->flags & PATCH_B) {
289				addr = (sljit_sw)(addr - (sljit_uw)SLJIT_ADD_EXEC_OFFSET(buf_ptr, executable_offset)) >> 2;
290				SLJIT_ASSERT((sljit_sw)addr <= 0x1ffffff && (sljit_sw)addr >= -0x2000000);
291				buf_ptr[0] = ((jump->flags & IS_BL) ? BL : B) | (addr & 0x3ffffff);
292				if (jump->flags & IS_COND)
293					buf_ptr[-1] -= (4 << 5);
294				break;
295			}
296			if (jump->flags & PATCH_COND) {
297				addr = (sljit_sw)(addr - (sljit_uw)SLJIT_ADD_EXEC_OFFSET(buf_ptr, executable_offset)) >> 2;
298				SLJIT_ASSERT((sljit_sw)addr <= 0x3ffff && (sljit_sw)addr >= -0x40000);
299				buf_ptr[0] = (buf_ptr[0] & ~0xffffe0) | ((addr & 0x7ffff) << 5);
300				break;
301			}
302
303			SLJIT_ASSERT((jump->flags & (PATCH_ABS48 | PATCH_ABS64)) || addr <= 0xffffffffl);
304			SLJIT_ASSERT((jump->flags & PATCH_ABS64) || addr <= 0xffffffffffffl);
305
306			dst = buf_ptr[0] & 0x1f;
307			buf_ptr[0] = MOVZ | dst | ((addr & 0xffff) << 5);
308			buf_ptr[1] = MOVK | dst | (((addr >> 16) & 0xffff) << 5) | (1 << 21);
309			if (jump->flags & (PATCH_ABS48 | PATCH_ABS64))
310				buf_ptr[2] = MOVK | dst | (((addr >> 32) & 0xffff) << 5) | (2 << 21);
311			if (jump->flags & PATCH_ABS64)
312				buf_ptr[3] = MOVK | dst | (((addr >> 48) & 0xffff) << 5) | (3 << 21);
313		} while (0);
314		jump = jump->next;
315	}
316
317	compiler->error = SLJIT_ERR_COMPILED;
318	compiler->executable_offset = executable_offset;
319	compiler->executable_size = (code_ptr - code) * sizeof(sljit_ins);
320
321	code = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(code, executable_offset);
322	code_ptr = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset);
323
324	SLJIT_CACHE_FLUSH(code, code_ptr);
325	return code;
326}
327
328/* --------------------------------------------------------------------- */
329/*  Core code generator functions.                                       */
330/* --------------------------------------------------------------------- */
331
332#define COUNT_TRAILING_ZERO(value, result) \
333	result = 0; \
334	if (!(value & 0xffffffff)) { \
335		result += 32; \
336		value >>= 32; \
337	} \
338	if (!(value & 0xffff)) { \
339		result += 16; \
340		value >>= 16; \
341	} \
342	if (!(value & 0xff)) { \
343		result += 8; \
344		value >>= 8; \
345	} \
346	if (!(value & 0xf)) { \
347		result += 4; \
348		value >>= 4; \
349	} \
350	if (!(value & 0x3)) { \
351		result += 2; \
352		value >>= 2; \
353	} \
354	if (!(value & 0x1)) { \
355		result += 1; \
356		value >>= 1; \
357	}
358
359#define LOGICAL_IMM_CHECK 0x100
360
361static sljit_ins logical_imm(sljit_sw imm, sljit_s32 len)
362{
363	sljit_s32 negated, ones, right;
364	sljit_uw mask, uimm;
365	sljit_ins ins;
366
367	if (len & LOGICAL_IMM_CHECK) {
368		len &= ~LOGICAL_IMM_CHECK;
369		if (len == 32 && (imm == 0 || imm == -1))
370			return 0;
371		if (len == 16 && ((sljit_s32)imm == 0 || (sljit_s32)imm == -1))
372			return 0;
373	}
374
375	SLJIT_ASSERT((len == 32 && imm != 0 && imm != -1)
376		|| (len == 16 && (sljit_s32)imm != 0 && (sljit_s32)imm != -1));
377	uimm = (sljit_uw)imm;
378	while (1) {
379		if (len <= 0) {
380			SLJIT_UNREACHABLE();
381			return 0;
382		}
383		mask = ((sljit_uw)1 << len) - 1;
384		if ((uimm & mask) != ((uimm >> len) & mask))
385			break;
386		len >>= 1;
387	}
388
389	len <<= 1;
390
391	negated = 0;
392	if (uimm & 0x1) {
393		negated = 1;
394		uimm = ~uimm;
395	}
396
397	if (len < 64)
398		uimm &= ((sljit_uw)1 << len) - 1;
399
400	/* Unsigned right shift. */
401	COUNT_TRAILING_ZERO(uimm, right);
402
403	/* Signed shift. We also know that the highest bit is set. */
404	imm = (sljit_sw)~uimm;
405	SLJIT_ASSERT(imm < 0);
406
407	COUNT_TRAILING_ZERO(imm, ones);
408
409	if (~imm)
410		return 0;
411
412	if (len == 64)
413		ins = 1 << 22;
414	else
415		ins = (0x3f - ((len << 1) - 1)) << 10;
416
417	if (negated)
418		return ins | ((len - ones - 1) << 10) | ((len - ones - right) << 16);
419
420	return ins | ((ones - 1) << 10) | ((len - right) << 16);
421}
422
423#undef COUNT_TRAILING_ZERO
424
425static sljit_s32 load_immediate(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw simm)
426{
427	sljit_uw imm = (sljit_uw)simm;
428	sljit_s32 i, zeros, ones, first;
429	sljit_ins bitmask;
430
431	if (imm <= 0xffff)
432		return push_inst(compiler, MOVZ | RD(dst) | (imm << 5));
433
434	if (simm >= -0x10000 && simm < 0)
435		return push_inst(compiler, MOVN | RD(dst) | ((~imm & 0xffff) << 5));
436
437	if (imm <= 0xffffffffl) {
438		if ((imm & 0xffff0000l) == 0xffff0000)
439			return push_inst(compiler, (MOVN ^ W_OP) | RD(dst) | ((~imm & 0xffff) << 5));
440		if ((imm & 0xffff) == 0xffff)
441			return push_inst(compiler, (MOVN ^ W_OP) | RD(dst) | ((~imm & 0xffff0000l) >> (16 - 5)) | (1 << 21));
442		bitmask = logical_imm(simm, 16);
443		if (bitmask != 0)
444			return push_inst(compiler, (ORRI ^ W_OP) | RD(dst) | RN(TMP_ZERO) | bitmask);
445	}
446	else {
447		bitmask = logical_imm(simm, 32);
448		if (bitmask != 0)
449			return push_inst(compiler, ORRI | RD(dst) | RN(TMP_ZERO) | bitmask);
450	}
451
452	if (imm <= 0xffffffffl) {
453		FAIL_IF(push_inst(compiler, MOVZ | RD(dst) | ((imm & 0xffff) << 5)));
454		return push_inst(compiler, MOVK | RD(dst) | ((imm & 0xffff0000l) >> (16 - 5)) | (1 << 21));
455	}
456
457	if (simm >= -0x100000000l && simm < 0) {
458		FAIL_IF(push_inst(compiler, MOVN | RD(dst) | ((~imm & 0xffff) << 5)));
459		return push_inst(compiler, MOVK | RD(dst) | ((imm & 0xffff0000l) >> (16 - 5)) | (1 << 21));
460	}
461
462	/* A large amount of number can be constructed from ORR and MOVx,
463	but computing them is costly. We don't  */
464
465	zeros = 0;
466	ones = 0;
467	for (i = 4; i > 0; i--) {
468		if ((simm & 0xffff) == 0)
469			zeros++;
470		if ((simm & 0xffff) == 0xffff)
471			ones++;
472		simm >>= 16;
473	}
474
475	simm = (sljit_sw)imm;
476	first = 1;
477	if (ones > zeros) {
478		simm = ~simm;
479		for (i = 0; i < 4; i++) {
480			if (!(simm & 0xffff)) {
481				simm >>= 16;
482				continue;
483			}
484			if (first) {
485				first = 0;
486				FAIL_IF(push_inst(compiler, MOVN | RD(dst) | ((simm & 0xffff) << 5) | (i << 21)));
487			}
488			else
489				FAIL_IF(push_inst(compiler, MOVK | RD(dst) | ((~simm & 0xffff) << 5) | (i << 21)));
490			simm >>= 16;
491		}
492		return SLJIT_SUCCESS;
493	}
494
495	for (i = 0; i < 4; i++) {
496		if (!(simm & 0xffff)) {
497			simm >>= 16;
498			continue;
499		}
500		if (first) {
501			first = 0;
502			FAIL_IF(push_inst(compiler, MOVZ | RD(dst) | ((simm & 0xffff) << 5) | (i << 21)));
503		}
504		else
505			FAIL_IF(push_inst(compiler, MOVK | RD(dst) | ((simm & 0xffff) << 5) | (i << 21)));
506		simm >>= 16;
507	}
508	return SLJIT_SUCCESS;
509}
510
511#define ARG1_IMM	0x0010000
512#define ARG2_IMM	0x0020000
513#define INT_OP		0x0040000
514#define SET_FLAGS	0x0080000
515#define UNUSED_RETURN	0x0100000
516#define SLOW_DEST	0x0200000
517#define SLOW_SRC1	0x0400000
518#define SLOW_SRC2	0x0800000
519
520#define CHECK_FLAGS(flag_bits) \
521	if (flags & SET_FLAGS) { \
522		inv_bits |= flag_bits; \
523		if (flags & UNUSED_RETURN) \
524			dst = TMP_ZERO; \
525	}
526
527static sljit_s32 emit_op_imm(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 dst, sljit_sw arg1, sljit_sw arg2)
528{
529	/* dst must be register, TMP_REG1
530	   arg1 must be register, TMP_REG1, imm
531	   arg2 must be register, TMP_REG2, imm */
532	sljit_ins inv_bits = (flags & INT_OP) ? (1 << 31) : 0;
533	sljit_ins inst_bits;
534	sljit_s32 op = (flags & 0xffff);
535	sljit_s32 reg;
536	sljit_sw imm, nimm;
537
538	if (SLJIT_UNLIKELY((flags & (ARG1_IMM | ARG2_IMM)) == (ARG1_IMM | ARG2_IMM))) {
539		/* Both are immediates. */
540		flags &= ~ARG1_IMM;
541		if (arg1 == 0 && op != SLJIT_ADD && op != SLJIT_SUB)
542			arg1 = TMP_ZERO;
543		else {
544			FAIL_IF(load_immediate(compiler, TMP_REG1, arg1));
545			arg1 = TMP_REG1;
546		}
547	}
548
549	if (flags & (ARG1_IMM | ARG2_IMM)) {
550		reg = (flags & ARG2_IMM) ? arg1 : arg2;
551		imm = (flags & ARG2_IMM) ? arg2 : arg1;
552
553		switch (op) {
554		case SLJIT_MUL:
555		case SLJIT_NEG:
556		case SLJIT_CLZ:
557		case SLJIT_ADDC:
558		case SLJIT_SUBC:
559			/* No form with immediate operand (except imm 0, which
560			is represented by a ZERO register). */
561			break;
562		case SLJIT_MOV:
563			SLJIT_ASSERT(!(flags & SET_FLAGS) && (flags & ARG2_IMM) && arg1 == TMP_REG1);
564			return load_immediate(compiler, dst, imm);
565		case SLJIT_NOT:
566			SLJIT_ASSERT(flags & ARG2_IMM);
567			FAIL_IF(load_immediate(compiler, dst, (flags & INT_OP) ? (~imm & 0xffffffff) : ~imm));
568			goto set_flags;
569		case SLJIT_SUB:
570			if (flags & ARG1_IMM)
571				break;
572			imm = -imm;
573			/* Fall through. */
574		case SLJIT_ADD:
575			if (imm == 0) {
576				CHECK_FLAGS(1 << 29);
577				return push_inst(compiler, ((op == SLJIT_ADD ? ADDI : SUBI) ^ inv_bits) | RD(dst) | RN(reg));
578			}
579			if (imm > 0 && imm <= 0xfff) {
580				CHECK_FLAGS(1 << 29);
581				return push_inst(compiler, (ADDI ^ inv_bits) | RD(dst) | RN(reg) | (imm << 10));
582			}
583			nimm = -imm;
584			if (nimm > 0 && nimm <= 0xfff) {
585				CHECK_FLAGS(1 << 29);
586				return push_inst(compiler, (SUBI ^ inv_bits) | RD(dst) | RN(reg) | (nimm << 10));
587			}
588			if (imm > 0 && imm <= 0xffffff && !(imm & 0xfff)) {
589				CHECK_FLAGS(1 << 29);
590				return push_inst(compiler, (ADDI ^ inv_bits) | RD(dst) | RN(reg) | ((imm >> 12) << 10) | (1 << 22));
591			}
592			if (nimm > 0 && nimm <= 0xffffff && !(nimm & 0xfff)) {
593				CHECK_FLAGS(1 << 29);
594				return push_inst(compiler, (SUBI ^ inv_bits) | RD(dst) | RN(reg) | ((nimm >> 12) << 10) | (1 << 22));
595			}
596			if (imm > 0 && imm <= 0xffffff && !(flags & SET_FLAGS)) {
597				FAIL_IF(push_inst(compiler, (ADDI ^ inv_bits) | RD(dst) | RN(reg) | ((imm >> 12) << 10) | (1 << 22)));
598				return push_inst(compiler, (ADDI ^ inv_bits) | RD(dst) | RN(dst) | ((imm & 0xfff) << 10));
599			}
600			if (nimm > 0 && nimm <= 0xffffff && !(flags & SET_FLAGS)) {
601				FAIL_IF(push_inst(compiler, (SUBI ^ inv_bits) | RD(dst) | RN(reg) | ((nimm >> 12) << 10) | (1 << 22)));
602				return push_inst(compiler, (SUBI ^ inv_bits) | RD(dst) | RN(dst) | ((nimm & 0xfff) << 10));
603			}
604			break;
605		case SLJIT_AND:
606			inst_bits = logical_imm(imm, LOGICAL_IMM_CHECK | ((flags & INT_OP) ? 16 : 32));
607			if (!inst_bits)
608				break;
609			CHECK_FLAGS(3 << 29);
610			return push_inst(compiler, (ANDI ^ inv_bits) | RD(dst) | RN(reg) | inst_bits);
611		case SLJIT_OR:
612		case SLJIT_XOR:
613			inst_bits = logical_imm(imm, LOGICAL_IMM_CHECK | ((flags & INT_OP) ? 16 : 32));
614			if (!inst_bits)
615				break;
616			if (op == SLJIT_OR)
617				inst_bits |= ORRI;
618			else
619				inst_bits |= EORI;
620			FAIL_IF(push_inst(compiler, (inst_bits ^ inv_bits) | RD(dst) | RN(reg)));
621			goto set_flags;
622		case SLJIT_SHL:
623			if (flags & ARG1_IMM)
624				break;
625			if (flags & INT_OP) {
626				imm &= 0x1f;
627				FAIL_IF(push_inst(compiler, (UBFM ^ inv_bits) | RD(dst) | RN(arg1) | ((-imm & 0x1f) << 16) | ((31 - imm) << 10)));
628			}
629			else {
630				imm &= 0x3f;
631				FAIL_IF(push_inst(compiler, (UBFM ^ inv_bits) | RD(dst) | RN(arg1) | (1 << 22) | ((-imm & 0x3f) << 16) | ((63 - imm) << 10)));
632			}
633			goto set_flags;
634		case SLJIT_LSHR:
635		case SLJIT_ASHR:
636			if (flags & ARG1_IMM)
637				break;
638			if (op == SLJIT_ASHR)
639				inv_bits |= 1 << 30;
640			if (flags & INT_OP) {
641				imm &= 0x1f;
642				FAIL_IF(push_inst(compiler, (UBFM ^ inv_bits) | RD(dst) | RN(arg1) | (imm << 16) | (31 << 10)));
643			}
644			else {
645				imm &= 0x3f;
646				FAIL_IF(push_inst(compiler, (UBFM ^ inv_bits) | RD(dst) | RN(arg1) | (1 << 22) | (imm << 16) | (63 << 10)));
647			}
648			goto set_flags;
649		default:
650			SLJIT_UNREACHABLE();
651			break;
652		}
653
654		if (flags & ARG2_IMM) {
655			if (arg2 == 0)
656				arg2 = TMP_ZERO;
657			else {
658				FAIL_IF(load_immediate(compiler, TMP_REG2, arg2));
659				arg2 = TMP_REG2;
660			}
661		}
662		else {
663			if (arg1 == 0)
664				arg1 = TMP_ZERO;
665			else {
666				FAIL_IF(load_immediate(compiler, TMP_REG1, arg1));
667				arg1 = TMP_REG1;
668			}
669		}
670	}
671
672	/* Both arguments are registers. */
673	switch (op) {
674	case SLJIT_MOV:
675	case SLJIT_MOV_P:
676	case SLJIT_MOVU:
677	case SLJIT_MOVU_P:
678		SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1);
679		if (dst == arg2)
680			return SLJIT_SUCCESS;
681		return push_inst(compiler, ORR | RD(dst) | RN(TMP_ZERO) | RM(arg2));
682	case SLJIT_MOV_U8:
683	case SLJIT_MOVU_U8:
684		SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1);
685		return push_inst(compiler, (UBFM ^ (1 << 31)) | RD(dst) | RN(arg2) | (7 << 10));
686	case SLJIT_MOV_S8:
687	case SLJIT_MOVU_S8:
688		SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1);
689		if (!(flags & INT_OP))
690			inv_bits |= 1 << 22;
691		return push_inst(compiler, (SBFM ^ inv_bits) | RD(dst) | RN(arg2) | (7 << 10));
692	case SLJIT_MOV_U16:
693	case SLJIT_MOVU_U16:
694		SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1);
695		return push_inst(compiler, (UBFM ^ (1 << 31)) | RD(dst) | RN(arg2) | (15 << 10));
696	case SLJIT_MOV_S16:
697	case SLJIT_MOVU_S16:
698		SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1);
699		if (!(flags & INT_OP))
700			inv_bits |= 1 << 22;
701		return push_inst(compiler, (SBFM ^ inv_bits) | RD(dst) | RN(arg2) | (15 << 10));
702	case SLJIT_MOV_U32:
703	case SLJIT_MOVU_U32:
704		SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1);
705		if ((flags & INT_OP) && dst == arg2)
706			return SLJIT_SUCCESS;
707		return push_inst(compiler, (ORR ^ (1 << 31)) | RD(dst) | RN(TMP_ZERO) | RM(arg2));
708	case SLJIT_MOV_S32:
709	case SLJIT_MOVU_S32:
710		SLJIT_ASSERT(!(flags & SET_FLAGS) && arg1 == TMP_REG1);
711		if ((flags & INT_OP) && dst == arg2)
712			return SLJIT_SUCCESS;
713		return push_inst(compiler, SBFM | (1 << 22) | RD(dst) | RN(arg2) | (31 << 10));
714	case SLJIT_NOT:
715		SLJIT_ASSERT(arg1 == TMP_REG1);
716		FAIL_IF(push_inst(compiler, (ORN ^ inv_bits) | RD(dst) | RN(TMP_ZERO) | RM(arg2)));
717		goto set_flags;
718	case SLJIT_NEG:
719		SLJIT_ASSERT(arg1 == TMP_REG1);
720		if (flags & SET_FLAGS)
721			inv_bits |= 1 << 29;
722		return push_inst(compiler, (SUB ^ inv_bits) | RD(dst) | RN(TMP_ZERO) | RM(arg2));
723	case SLJIT_CLZ:
724		SLJIT_ASSERT(arg1 == TMP_REG1);
725		FAIL_IF(push_inst(compiler, (CLZ ^ inv_bits) | RD(dst) | RN(arg2)));
726		goto set_flags;
727	case SLJIT_ADD:
728		CHECK_FLAGS(1 << 29);
729		return push_inst(compiler, (ADD ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2));
730	case SLJIT_ADDC:
731		CHECK_FLAGS(1 << 29);
732		return push_inst(compiler, (ADC ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2));
733	case SLJIT_SUB:
734		CHECK_FLAGS(1 << 29);
735		return push_inst(compiler, (SUB ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2));
736	case SLJIT_SUBC:
737		CHECK_FLAGS(1 << 29);
738		return push_inst(compiler, (SBC ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2));
739	case SLJIT_MUL:
740		if (!(flags & SET_FLAGS))
741			return push_inst(compiler, (MADD ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2) | RT2(TMP_ZERO));
742		if (flags & INT_OP) {
743			FAIL_IF(push_inst(compiler, SMADDL | RD(dst) | RN(arg1) | RM(arg2) | (31 << 10)));
744			FAIL_IF(push_inst(compiler, ADD | RD(TMP_LR) | RN(TMP_ZERO) | RM(dst) | (2 << 22) | (31 << 10)));
745			return push_inst(compiler, SUBS | RD(TMP_ZERO) | RN(TMP_LR) | RM(dst) | (2 << 22) | (63 << 10));
746		}
747		FAIL_IF(push_inst(compiler, SMULH | RD(TMP_LR) | RN(arg1) | RM(arg2)));
748		FAIL_IF(push_inst(compiler, MADD | RD(dst) | RN(arg1) | RM(arg2) | RT2(TMP_ZERO)));
749		return push_inst(compiler, SUBS | RD(TMP_ZERO) | RN(TMP_LR) | RM(dst) | (2 << 22) | (63 << 10));
750	case SLJIT_AND:
751		CHECK_FLAGS(3 << 29);
752		return push_inst(compiler, (AND ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2));
753	case SLJIT_OR:
754		FAIL_IF(push_inst(compiler, (ORR ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)));
755		goto set_flags;
756	case SLJIT_XOR:
757		FAIL_IF(push_inst(compiler, (EOR ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)));
758		goto set_flags;
759	case SLJIT_SHL:
760		FAIL_IF(push_inst(compiler, (LSLV ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)));
761		goto set_flags;
762	case SLJIT_LSHR:
763		FAIL_IF(push_inst(compiler, (LSRV ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)));
764		goto set_flags;
765	case SLJIT_ASHR:
766		FAIL_IF(push_inst(compiler, (ASRV ^ inv_bits) | RD(dst) | RN(arg1) | RM(arg2)));
767		goto set_flags;
768	}
769
770	SLJIT_UNREACHABLE();
771	return SLJIT_SUCCESS;
772
773set_flags:
774	if (flags & SET_FLAGS)
775		return push_inst(compiler, (SUBS ^ inv_bits) | RD(TMP_ZERO) | RN(dst) | RM(TMP_ZERO));
776	return SLJIT_SUCCESS;
777}
778
779#define STORE		0x01
780#define SIGNED		0x02
781
782#define UPDATE		0x04
783#define ARG_TEST	0x08
784
785#define BYTE_SIZE	0x000
786#define HALF_SIZE	0x100
787#define INT_SIZE	0x200
788#define WORD_SIZE	0x300
789
790#define MEM_SIZE_SHIFT(flags) ((flags) >> 8)
791
792static const sljit_ins sljit_mem_imm[4] = {
793/* u l */ 0x39400000 /* ldrb [reg,imm] */,
794/* u s */ 0x39000000 /* strb [reg,imm] */,
795/* s l */ 0x39800000 /* ldrsb [reg,imm] */,
796/* s s */ 0x39000000 /* strb [reg,imm] */,
797};
798
799static const sljit_ins sljit_mem_simm[4] = {
800/* u l */ 0x38400000 /* ldurb [reg,imm] */,
801/* u s */ 0x38000000 /* sturb [reg,imm] */,
802/* s l */ 0x38800000 /* ldursb [reg,imm] */,
803/* s s */ 0x38000000 /* sturb [reg,imm] */,
804};
805
806static const sljit_ins sljit_mem_pre_simm[4] = {
807/* u l */ 0x38400c00 /* ldrb [reg,imm]! */,
808/* u s */ 0x38000c00 /* strb [reg,imm]! */,
809/* s l */ 0x38800c00 /* ldrsb [reg,imm]! */,
810/* s s */ 0x38000c00 /* strb [reg,imm]! */,
811};
812
813static const sljit_ins sljit_mem_reg[4] = {
814/* u l */ 0x38606800 /* ldrb [reg,reg] */,
815/* u s */ 0x38206800 /* strb [reg,reg] */,
816/* s l */ 0x38a06800 /* ldrsb [reg,reg] */,
817/* s s */ 0x38206800 /* strb [reg,reg] */,
818};
819
820/* Helper function. Dst should be reg + value, using at most 1 instruction, flags does not set. */
821static sljit_s32 emit_set_delta(struct sljit_compiler *compiler, sljit_s32 dst, sljit_s32 reg, sljit_sw value)
822{
823	if (value >= 0) {
824		if (value <= 0xfff)
825			return push_inst(compiler, ADDI | RD(dst) | RN(reg) | (value << 10));
826		if (value <= 0xffffff && !(value & 0xfff))
827			return push_inst(compiler, ADDI | (1 << 22) | RD(dst) | RN(reg) | (value >> 2));
828	}
829	else {
830		value = -value;
831		if (value <= 0xfff)
832			return push_inst(compiler, SUBI | RD(dst) | RN(reg) | (value << 10));
833		if (value <= 0xffffff && !(value & 0xfff))
834			return push_inst(compiler, SUBI | (1 << 22) | RD(dst) | RN(reg) | (value >> 2));
835	}
836	return SLJIT_ERR_UNSUPPORTED;
837}
838
839/* Can perform an operation using at most 1 instruction. */
840static sljit_s32 getput_arg_fast(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg, sljit_s32 arg, sljit_sw argw)
841{
842	sljit_u32 shift = MEM_SIZE_SHIFT(flags);
843
844	SLJIT_ASSERT(arg & SLJIT_MEM);
845
846	if (SLJIT_UNLIKELY(flags & UPDATE)) {
847		if ((arg & REG_MASK) && !(arg & OFFS_REG_MASK) && argw <= 255 && argw >= -256) {
848			if (SLJIT_UNLIKELY(flags & ARG_TEST))
849				return 1;
850
851			arg &= REG_MASK;
852			argw &= 0x1ff;
853			FAIL_IF(push_inst(compiler, sljit_mem_pre_simm[flags & 0x3]
854				| (shift << 30) | RT(reg) | RN(arg) | (argw << 12)));
855			return -1;
856		}
857		return 0;
858	}
859
860	if (SLJIT_UNLIKELY(arg & OFFS_REG_MASK)) {
861		argw &= 0x3;
862		if (argw && argw != shift)
863			return 0;
864
865		if (SLJIT_UNLIKELY(flags & ARG_TEST))
866			return 1;
867
868		FAIL_IF(push_inst(compiler, sljit_mem_reg[flags & 0x3] | (shift << 30) | RT(reg)
869			| RN(arg & REG_MASK) | RM(OFFS_REG(arg)) | (argw ? (1 << 12) : 0)));
870		return -1;
871	}
872
873	arg &= REG_MASK;
874	if (argw >= 0 && (argw >> shift) <= 0xfff && (argw & ((1 << shift) - 1)) == 0) {
875		if (SLJIT_UNLIKELY(flags & ARG_TEST))
876			return 1;
877
878		FAIL_IF(push_inst(compiler, sljit_mem_imm[flags & 0x3] | (shift << 30)
879			| RT(reg) | RN(arg) | (argw << (10 - shift))));
880		return -1;
881	}
882
883	if (argw > 255 || argw < -256)
884		return 0;
885
886	if (SLJIT_UNLIKELY(flags & ARG_TEST))
887		return 1;
888
889	FAIL_IF(push_inst(compiler, sljit_mem_simm[flags & 0x3] | (shift << 30)
890		| RT(reg) | RN(arg) | ((argw & 0x1ff) << 12)));
891	return -1;
892}
893
894/* see getput_arg below.
895   Note: can_cache is called only for binary operators. Those
896   operators always uses word arguments without write back. */
897static sljit_s32 can_cache(sljit_s32 arg, sljit_sw argw, sljit_s32 next_arg, sljit_sw next_argw)
898{
899	sljit_sw diff;
900	if ((arg & OFFS_REG_MASK) || !(next_arg & SLJIT_MEM))
901		return 0;
902
903	if (!(arg & REG_MASK)) {
904		diff = argw - next_argw;
905		if (diff <= 0xfff && diff >= -0xfff)
906			return 1;
907		return 0;
908	}
909
910	if (argw == next_argw)
911		return 1;
912
913	diff = argw - next_argw;
914	if (arg == next_arg && diff <= 0xfff && diff >= -0xfff)
915		return 1;
916
917	return 0;
918}
919
920/* Emit the necessary instructions. See can_cache above. */
921static sljit_s32 getput_arg(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg,
922	sljit_s32 arg, sljit_sw argw, sljit_s32 next_arg, sljit_sw next_argw)
923{
924	sljit_u32 shift = MEM_SIZE_SHIFT(flags);
925	sljit_s32 tmp_r, other_r;
926	sljit_sw diff;
927
928	SLJIT_ASSERT(arg & SLJIT_MEM);
929	if (!(next_arg & SLJIT_MEM)) {
930		next_arg = 0;
931		next_argw = 0;
932	}
933
934	tmp_r = (flags & STORE) ? TMP_REG3 : reg;
935
936	if (SLJIT_UNLIKELY((flags & UPDATE) && (arg & REG_MASK))) {
937		/* Update only applies if a base register exists. */
938		other_r = OFFS_REG(arg);
939		if (!other_r) {
940			other_r = arg & REG_MASK;
941			SLJIT_ASSERT(other_r != reg);
942
943			if (argw >= 0 && argw <= 0xffffff) {
944				if ((argw & 0xfff) != 0)
945					FAIL_IF(push_inst(compiler, ADDI | RD(other_r) | RN(other_r) | ((argw & 0xfff) << 10)));
946				if (argw >> 12)
947					FAIL_IF(push_inst(compiler, ADDI | (1 << 22) | RD(other_r) | RN(other_r) | ((argw >> 12) << 10)));
948				return push_inst(compiler, sljit_mem_imm[flags & 0x3] | (shift << 30) | RT(reg) | RN(other_r));
949			}
950			else if (argw < 0 && argw >= -0xffffff) {
951				argw = -argw;
952				if ((argw & 0xfff) != 0)
953					FAIL_IF(push_inst(compiler, SUBI | RD(other_r) | RN(other_r) | ((argw & 0xfff) << 10)));
954				if (argw >> 12)
955					FAIL_IF(push_inst(compiler, SUBI | (1 << 22) | RD(other_r) | RN(other_r) | ((argw >> 12) << 10)));
956				return push_inst(compiler, sljit_mem_imm[flags & 0x3] | (shift << 30) | RT(reg) | RN(other_r));
957			}
958
959			if (compiler->cache_arg == SLJIT_MEM) {
960				if (argw == compiler->cache_argw) {
961					other_r = TMP_REG3;
962					argw = 0;
963				}
964				else if (emit_set_delta(compiler, TMP_REG3, TMP_REG3, argw - compiler->cache_argw) != SLJIT_ERR_UNSUPPORTED) {
965					FAIL_IF(compiler->error);
966					compiler->cache_argw = argw;
967					other_r = TMP_REG3;
968					argw = 0;
969				}
970			}
971
972			if (argw) {
973				FAIL_IF(load_immediate(compiler, TMP_REG3, argw));
974				compiler->cache_arg = SLJIT_MEM;
975				compiler->cache_argw = argw;
976				other_r = TMP_REG3;
977				argw = 0;
978			}
979		}
980
981		/* No caching here. */
982		arg &= REG_MASK;
983		FAIL_IF(push_inst(compiler, sljit_mem_reg[flags & 0x3] | (shift << 30) | RT(reg) | RN(arg) | RM(other_r)));
984		return push_inst(compiler, ADD | RD(arg) | RN(arg) | RM(other_r));
985	}
986
987	if (arg & OFFS_REG_MASK) {
988		other_r = OFFS_REG(arg);
989		arg &= REG_MASK;
990		FAIL_IF(push_inst(compiler, ADD | RD(tmp_r) | RN(arg) | RM(other_r) | ((argw & 0x3) << 10)));
991		return push_inst(compiler, sljit_mem_imm[flags & 0x3] | (shift << 30) | RT(reg) | RN(tmp_r));
992	}
993
994	if (compiler->cache_arg == arg) {
995		diff = argw - compiler->cache_argw;
996		if (diff <= 255 && diff >= -256)
997			return push_inst(compiler, sljit_mem_simm[flags & 0x3] | (shift << 30)
998				| RT(reg) | RN(TMP_REG3) | ((diff & 0x1ff) << 12));
999		if (emit_set_delta(compiler, TMP_REG3, TMP_REG3, diff) != SLJIT_ERR_UNSUPPORTED) {
1000			FAIL_IF(compiler->error);
1001			return push_inst(compiler, sljit_mem_imm[flags & 0x3] | (shift << 30) | RT(reg) | RN(arg));
1002		}
1003	}
1004
1005	if (argw >= 0 && argw <= 0xffffff && (argw & ((1 << shift) - 1)) == 0) {
1006		FAIL_IF(push_inst(compiler, ADDI | (1 << 22) | RD(tmp_r) | RN(arg & REG_MASK) | ((argw >> 12) << 10)));
1007		return push_inst(compiler, sljit_mem_imm[flags & 0x3] | (shift << 30)
1008			| RT(reg) | RN(tmp_r) | ((argw & 0xfff) << (10 - shift)));
1009	}
1010
1011	diff = argw - next_argw;
1012	next_arg = (arg & REG_MASK) && (arg == next_arg) && diff <= 0xfff && diff >= -0xfff && diff != 0;
1013	arg &= REG_MASK;
1014
1015	if (arg && compiler->cache_arg == SLJIT_MEM) {
1016		if (compiler->cache_argw == argw)
1017			return push_inst(compiler, sljit_mem_reg[flags & 0x3] | (shift << 30) | RT(reg) | RN(arg) | RM(TMP_REG3));
1018		if (emit_set_delta(compiler, TMP_REG3, TMP_REG3, argw - compiler->cache_argw) != SLJIT_ERR_UNSUPPORTED) {
1019			FAIL_IF(compiler->error);
1020			compiler->cache_argw = argw;
1021			return push_inst(compiler, sljit_mem_reg[flags & 0x3] | (shift << 30) | RT(reg) | RN(arg) | RM(TMP_REG3));
1022		}
1023	}
1024
1025	compiler->cache_argw = argw;
1026	if (next_arg && emit_set_delta(compiler, TMP_REG3, arg, argw) != SLJIT_ERR_UNSUPPORTED) {
1027		FAIL_IF(compiler->error);
1028		compiler->cache_arg = SLJIT_MEM | arg;
1029		arg = 0;
1030	}
1031	else {
1032		FAIL_IF(load_immediate(compiler, TMP_REG3, argw));
1033		compiler->cache_arg = SLJIT_MEM;
1034
1035		if (next_arg) {
1036			FAIL_IF(push_inst(compiler, ADD | RD(TMP_REG3) | RN(TMP_REG3) | RM(arg)));
1037			compiler->cache_arg = SLJIT_MEM | arg;
1038			arg = 0;
1039		}
1040	}
1041
1042	if (arg)
1043		return push_inst(compiler, sljit_mem_reg[flags & 0x3] | (shift << 30) | RT(reg) | RN(arg) | RM(TMP_REG3));
1044	return push_inst(compiler, sljit_mem_imm[flags & 0x3] | (shift << 30) | RT(reg) | RN(TMP_REG3));
1045}
1046
1047static SLJIT_INLINE sljit_s32 emit_op_mem(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg, sljit_s32 arg, sljit_sw argw)
1048{
1049	if (getput_arg_fast(compiler, flags, reg, arg, argw))
1050		return compiler->error;
1051	compiler->cache_arg = 0;
1052	compiler->cache_argw = 0;
1053	return getput_arg(compiler, flags, reg, arg, argw, 0, 0);
1054}
1055
1056static SLJIT_INLINE sljit_s32 emit_op_mem2(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg, sljit_s32 arg1, sljit_sw arg1w, sljit_s32 arg2, sljit_sw arg2w)
1057{
1058	if (getput_arg_fast(compiler, flags, reg, arg1, arg1w))
1059		return compiler->error;
1060	return getput_arg(compiler, flags, reg, arg1, arg1w, arg2, arg2w);
1061}
1062
1063/* --------------------------------------------------------------------- */
1064/*  Entry, exit                                                          */
1065/* --------------------------------------------------------------------- */
1066
1067SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_enter(struct sljit_compiler *compiler,
1068	sljit_s32 options, sljit_s32 args, sljit_s32 scratches, sljit_s32 saveds,
1069	sljit_s32 fscratches, sljit_s32 fsaveds, sljit_s32 local_size)
1070{
1071	sljit_s32 i, tmp, offs, prev, saved_regs_size;
1072
1073	CHECK_ERROR();
1074	CHECK(check_sljit_emit_enter(compiler, options, args, scratches, saveds, fscratches, fsaveds, local_size));
1075	set_emit_enter(compiler, options, args, scratches, saveds, fscratches, fsaveds, local_size);
1076
1077	saved_regs_size = GET_SAVED_REGISTERS_SIZE(scratches, saveds, 0);
1078	local_size += saved_regs_size + SLJIT_LOCALS_OFFSET;
1079	local_size = (local_size + 15) & ~0xf;
1080	compiler->local_size = local_size;
1081
1082	SLJIT_ASSERT(local_size >= 0);
1083	if ((size_t)local_size <= (63 * sizeof(sljit_sw))) {
1084		FAIL_IF(push_inst(compiler, STP_PRE | 29 | RT2(TMP_LR)
1085			| RN(TMP_SP) | ((-(local_size >> 3) & 0x7f) << 15)));
1086		FAIL_IF(push_inst(compiler, ADDI | RD(SLJIT_SP) | RN(TMP_SP) | (0 << 10)));
1087		offs = (local_size - saved_regs_size) << (15 - 3);
1088	} else {
1089		offs = 0 << 15;
1090		if (saved_regs_size & 0x8) {
1091			offs = 1 << 15;
1092			saved_regs_size += sizeof(sljit_sw);
1093		}
1094		local_size -= saved_regs_size + SLJIT_LOCALS_OFFSET;
1095		if (saved_regs_size > 0)
1096			FAIL_IF(push_inst(compiler, SUBI | RD(TMP_SP) | RN(TMP_SP) | (saved_regs_size << 10)));
1097	}
1098
1099	tmp = saveds < SLJIT_NUMBER_OF_SAVED_REGISTERS ? (SLJIT_S0 + 1 - saveds) : SLJIT_FIRST_SAVED_REG;
1100	prev = -1;
1101	for (i = SLJIT_S0; i >= tmp; i--) {
1102		if (prev == -1) {
1103			if (!(offs & (1 << 15))) {
1104				prev = i;
1105				continue;
1106			}
1107			FAIL_IF(push_inst(compiler, STRI | RT(i) | RN(TMP_SP) | (offs >> 5)));
1108			offs += 1 << 15;
1109			continue;
1110		}
1111		FAIL_IF(push_inst(compiler, STP | RT(prev) | RT2(i) | RN(TMP_SP) | offs));
1112		offs += 2 << 15;
1113		prev = -1;
1114	}
1115
1116	for (i = scratches; i >= SLJIT_FIRST_SAVED_REG; i--) {
1117		if (prev == -1) {
1118			if (!(offs & (1 << 15))) {
1119				prev = i;
1120				continue;
1121			}
1122			FAIL_IF(push_inst(compiler, STRI | RT(i) | RN(TMP_SP) | (offs >> 5)));
1123			offs += 1 << 15;
1124			continue;
1125		}
1126		FAIL_IF(push_inst(compiler, STP | RT(prev) | RT2(i) | RN(TMP_SP) | offs));
1127		offs += 2 << 15;
1128		prev = -1;
1129	}
1130
1131	SLJIT_ASSERT(prev == -1);
1132
1133	SLJIT_ASSERT(compiler->local_size >= 0);
1134	if ((size_t)compiler->local_size > (63 * sizeof(sljit_sw))) {
1135		/* The local_size is already adjusted by the saved registers. */
1136		if (local_size > 0xfff) {
1137			FAIL_IF(push_inst(compiler, SUBI | RD(TMP_SP) | RN(TMP_SP) | ((local_size >> 12) << 10) | (1 << 22)));
1138			local_size &= 0xfff;
1139		}
1140		if (local_size)
1141			FAIL_IF(push_inst(compiler, SUBI | RD(TMP_SP) | RN(TMP_SP) | (local_size << 10)));
1142		FAIL_IF(push_inst(compiler, STP_PRE | 29 | RT2(TMP_LR)
1143			| RN(TMP_SP) | ((-(16 >> 3) & 0x7f) << 15)));
1144		FAIL_IF(push_inst(compiler, ADDI | RD(SLJIT_SP) | RN(TMP_SP) | (0 << 10)));
1145	}
1146
1147	if (args >= 1)
1148		FAIL_IF(push_inst(compiler, ORR | RD(SLJIT_S0) | RN(TMP_ZERO) | RM(SLJIT_R0)));
1149	if (args >= 2)
1150		FAIL_IF(push_inst(compiler, ORR | RD(SLJIT_S1) | RN(TMP_ZERO) | RM(SLJIT_R1)));
1151	if (args >= 3)
1152		FAIL_IF(push_inst(compiler, ORR | RD(SLJIT_S2) | RN(TMP_ZERO) | RM(SLJIT_R2)));
1153
1154	return SLJIT_SUCCESS;
1155}
1156
1157SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_set_context(struct sljit_compiler *compiler,
1158	sljit_s32 options, sljit_s32 args, sljit_s32 scratches, sljit_s32 saveds,
1159	sljit_s32 fscratches, sljit_s32 fsaveds, sljit_s32 local_size)
1160{
1161	CHECK_ERROR();
1162	CHECK(check_sljit_set_context(compiler, options, args, scratches, saveds, fscratches, fsaveds, local_size));
1163	set_set_context(compiler, options, args, scratches, saveds, fscratches, fsaveds, local_size);
1164
1165	local_size += GET_SAVED_REGISTERS_SIZE(scratches, saveds, 0) + SLJIT_LOCALS_OFFSET;
1166	local_size = (local_size + 15) & ~0xf;
1167	compiler->local_size = local_size;
1168	return SLJIT_SUCCESS;
1169}
1170
1171SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_return(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 src, sljit_sw srcw)
1172{
1173	sljit_s32 local_size;
1174	sljit_s32 i, tmp, offs, prev, saved_regs_size;
1175
1176	CHECK_ERROR();
1177	CHECK(check_sljit_emit_return(compiler, op, src, srcw));
1178
1179	FAIL_IF(emit_mov_before_return(compiler, op, src, srcw));
1180
1181	local_size = compiler->local_size;
1182
1183	saved_regs_size = GET_SAVED_REGISTERS_SIZE(compiler->scratches, compiler->saveds, 0);
1184	SLJIT_ASSERT(local_size >= 0);
1185	if ((size_t)local_size <= (63 * sizeof(sljit_sw)))
1186		offs = (local_size - saved_regs_size) << (15 - 3);
1187	else {
1188		FAIL_IF(push_inst(compiler, LDP_PST | 29 | RT2(TMP_LR)
1189			| RN(TMP_SP) | (((16 >> 3) & 0x7f) << 15)));
1190		offs = 0 << 15;
1191		if (saved_regs_size & 0x8) {
1192			offs = 1 << 15;
1193			saved_regs_size += sizeof(sljit_sw);
1194		}
1195		local_size -= saved_regs_size + SLJIT_LOCALS_OFFSET;
1196		if (local_size > 0xfff) {
1197			FAIL_IF(push_inst(compiler, ADDI | RD(TMP_SP) | RN(TMP_SP) | ((local_size >> 12) << 10) | (1 << 22)));
1198			local_size &= 0xfff;
1199		}
1200		if (local_size)
1201			FAIL_IF(push_inst(compiler, ADDI | RD(TMP_SP) | RN(TMP_SP) | (local_size << 10)));
1202	}
1203
1204	tmp = compiler->saveds < SLJIT_NUMBER_OF_SAVED_REGISTERS ? (SLJIT_S0 + 1 - compiler->saveds) : SLJIT_FIRST_SAVED_REG;
1205	prev = -1;
1206	for (i = SLJIT_S0; i >= tmp; i--) {
1207		if (prev == -1) {
1208			if (!(offs & (1 << 15))) {
1209				prev = i;
1210				continue;
1211			}
1212			FAIL_IF(push_inst(compiler, LDRI | RT(i) | RN(TMP_SP) | (offs >> 5)));
1213			offs += 1 << 15;
1214			continue;
1215		}
1216		FAIL_IF(push_inst(compiler, LDP | RT(prev) | RT2(i) | RN(TMP_SP) | offs));
1217		offs += 2 << 15;
1218		prev = -1;
1219	}
1220
1221	for (i = compiler->scratches; i >= SLJIT_FIRST_SAVED_REG; i--) {
1222		if (prev == -1) {
1223			if (!(offs & (1 << 15))) {
1224				prev = i;
1225				continue;
1226			}
1227			FAIL_IF(push_inst(compiler, LDRI | RT(i) | RN(TMP_SP) | (offs >> 5)));
1228			offs += 1 << 15;
1229			continue;
1230		}
1231		FAIL_IF(push_inst(compiler, LDP | RT(prev) | RT2(i) | RN(TMP_SP) | offs));
1232		offs += 2 << 15;
1233		prev = -1;
1234	}
1235
1236	SLJIT_ASSERT(prev == -1);
1237
1238	SLJIT_ASSERT(compiler->local_size >= 0);
1239	if ((size_t)compiler->local_size <= (63 * sizeof(sljit_sw))) {
1240		FAIL_IF(push_inst(compiler, LDP_PST | 29 | RT2(TMP_LR)
1241			| RN(TMP_SP) | (((local_size >> 3) & 0x7f) << 15)));
1242	} else if (saved_regs_size > 0) {
1243		FAIL_IF(push_inst(compiler, ADDI | RD(TMP_SP) | RN(TMP_SP) | (saved_regs_size << 10)));
1244	}
1245
1246	FAIL_IF(push_inst(compiler, RET | RN(TMP_LR)));
1247	return SLJIT_SUCCESS;
1248}
1249
1250/* --------------------------------------------------------------------- */
1251/*  Operators                                                            */
1252/* --------------------------------------------------------------------- */
1253
1254SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op0(struct sljit_compiler *compiler, sljit_s32 op)
1255{
1256	sljit_ins inv_bits = (op & SLJIT_I32_OP) ? (1 << 31) : 0;
1257
1258	CHECK_ERROR();
1259	CHECK(check_sljit_emit_op0(compiler, op));
1260
1261	op = GET_OPCODE(op);
1262	switch (op) {
1263	case SLJIT_BREAKPOINT:
1264		return push_inst(compiler, BRK);
1265	case SLJIT_NOP:
1266		return push_inst(compiler, NOP);
1267	case SLJIT_LMUL_UW:
1268	case SLJIT_LMUL_SW:
1269		FAIL_IF(push_inst(compiler, ORR | RD(TMP_REG1) | RN(TMP_ZERO) | RM(SLJIT_R0)));
1270		FAIL_IF(push_inst(compiler, MADD | RD(SLJIT_R0) | RN(SLJIT_R0) | RM(SLJIT_R1) | RT2(TMP_ZERO)));
1271		return push_inst(compiler, (op == SLJIT_LMUL_UW ? UMULH : SMULH) | RD(SLJIT_R1) | RN(TMP_REG1) | RM(SLJIT_R1));
1272	case SLJIT_DIVMOD_UW:
1273	case SLJIT_DIVMOD_SW:
1274		FAIL_IF(push_inst(compiler, (ORR ^ inv_bits) | RD(TMP_REG1) | RN(TMP_ZERO) | RM(SLJIT_R0)));
1275		FAIL_IF(push_inst(compiler, ((op == SLJIT_DIVMOD_UW ? UDIV : SDIV) ^ inv_bits) | RD(SLJIT_R0) | RN(SLJIT_R0) | RM(SLJIT_R1)));
1276		FAIL_IF(push_inst(compiler, (MADD ^ inv_bits) | RD(SLJIT_R1) | RN(SLJIT_R0) | RM(SLJIT_R1) | RT2(TMP_ZERO)));
1277		return push_inst(compiler, (SUB ^ inv_bits) | RD(SLJIT_R1) | RN(TMP_REG1) | RM(SLJIT_R1));
1278	case SLJIT_DIV_UW:
1279	case SLJIT_DIV_SW:
1280		return push_inst(compiler, ((op == SLJIT_DIV_UW ? UDIV : SDIV) ^ inv_bits) | RD(SLJIT_R0) | RN(SLJIT_R0) | RM(SLJIT_R1));
1281	}
1282
1283	return SLJIT_SUCCESS;
1284}
1285
1286SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op1(struct sljit_compiler *compiler, sljit_s32 op,
1287	sljit_s32 dst, sljit_sw dstw,
1288	sljit_s32 src, sljit_sw srcw)
1289{
1290	sljit_s32 dst_r, flags, mem_flags;
1291	sljit_s32 op_flags = GET_ALL_FLAGS(op);
1292
1293	CHECK_ERROR();
1294	CHECK(check_sljit_emit_op1(compiler, op, dst, dstw, src, srcw));
1295	ADJUST_LOCAL_OFFSET(dst, dstw);
1296	ADJUST_LOCAL_OFFSET(src, srcw);
1297
1298	compiler->cache_arg = 0;
1299	compiler->cache_argw = 0;
1300
1301	dst_r = SLOW_IS_REG(dst) ? dst : TMP_REG1;
1302
1303	op = GET_OPCODE(op);
1304	if (op >= SLJIT_MOV && op <= SLJIT_MOVU_P) {
1305		switch (op) {
1306		case SLJIT_MOV:
1307		case SLJIT_MOV_P:
1308			flags = WORD_SIZE;
1309			break;
1310		case SLJIT_MOV_U8:
1311			flags = BYTE_SIZE;
1312			if (src & SLJIT_IMM)
1313				srcw = (sljit_u8)srcw;
1314			break;
1315		case SLJIT_MOV_S8:
1316			flags = BYTE_SIZE | SIGNED;
1317			if (src & SLJIT_IMM)
1318				srcw = (sljit_s8)srcw;
1319			break;
1320		case SLJIT_MOV_U16:
1321			flags = HALF_SIZE;
1322			if (src & SLJIT_IMM)
1323				srcw = (sljit_u16)srcw;
1324			break;
1325		case SLJIT_MOV_S16:
1326			flags = HALF_SIZE | SIGNED;
1327			if (src & SLJIT_IMM)
1328				srcw = (sljit_s16)srcw;
1329			break;
1330		case SLJIT_MOV_U32:
1331			flags = INT_SIZE;
1332			if (src & SLJIT_IMM)
1333				srcw = (sljit_u32)srcw;
1334			break;
1335		case SLJIT_MOV_S32:
1336			flags = INT_SIZE | SIGNED;
1337			if (src & SLJIT_IMM)
1338				srcw = (sljit_s32)srcw;
1339			break;
1340		case SLJIT_MOVU:
1341		case SLJIT_MOVU_P:
1342			flags = WORD_SIZE | UPDATE;
1343			break;
1344		case SLJIT_MOVU_U8:
1345			flags = BYTE_SIZE | UPDATE;
1346			if (src & SLJIT_IMM)
1347				srcw = (sljit_u8)srcw;
1348			break;
1349		case SLJIT_MOVU_S8:
1350			flags = BYTE_SIZE | SIGNED | UPDATE;
1351			if (src & SLJIT_IMM)
1352				srcw = (sljit_s8)srcw;
1353			break;
1354		case SLJIT_MOVU_U16:
1355			flags = HALF_SIZE | UPDATE;
1356			if (src & SLJIT_IMM)
1357				srcw = (sljit_u16)srcw;
1358			break;
1359		case SLJIT_MOVU_S16:
1360			flags = HALF_SIZE | SIGNED | UPDATE;
1361			if (src & SLJIT_IMM)
1362				srcw = (sljit_s16)srcw;
1363			break;
1364		case SLJIT_MOVU_U32:
1365			flags = INT_SIZE | UPDATE;
1366			if (src & SLJIT_IMM)
1367				srcw = (sljit_u32)srcw;
1368			break;
1369		case SLJIT_MOVU_S32:
1370			flags = INT_SIZE | SIGNED | UPDATE;
1371			if (src & SLJIT_IMM)
1372				srcw = (sljit_s32)srcw;
1373			break;
1374		default:
1375			SLJIT_UNREACHABLE();
1376			flags = 0;
1377			break;
1378		}
1379
1380		if (src & SLJIT_IMM)
1381			FAIL_IF(emit_op_imm(compiler, SLJIT_MOV | ARG2_IMM, dst_r, TMP_REG1, srcw));
1382		else if (src & SLJIT_MEM) {
1383			if (getput_arg_fast(compiler, flags, dst_r, src, srcw))
1384				FAIL_IF(compiler->error);
1385			else
1386				FAIL_IF(getput_arg(compiler, flags, dst_r, src, srcw, dst, dstw));
1387		} else {
1388			if (dst_r != TMP_REG1)
1389				return emit_op_imm(compiler, op | ((op_flags & SLJIT_I32_OP) ? INT_OP : 0), dst_r, TMP_REG1, src);
1390			dst_r = src;
1391		}
1392
1393		if (dst & SLJIT_MEM) {
1394			if (getput_arg_fast(compiler, flags | STORE, dst_r, dst, dstw))
1395				return compiler->error;
1396			else
1397				return getput_arg(compiler, flags | STORE, dst_r, dst, dstw, 0, 0);
1398		}
1399		return SLJIT_SUCCESS;
1400	}
1401
1402	flags = HAS_FLAGS(op_flags) ? SET_FLAGS : 0;
1403	mem_flags = WORD_SIZE;
1404	if (op_flags & SLJIT_I32_OP) {
1405		flags |= INT_OP;
1406		mem_flags = INT_SIZE;
1407	}
1408
1409	if (dst == SLJIT_UNUSED)
1410		flags |= UNUSED_RETURN;
1411
1412	if (src & SLJIT_MEM) {
1413		if (getput_arg_fast(compiler, mem_flags, TMP_REG2, src, srcw))
1414			FAIL_IF(compiler->error);
1415		else
1416			FAIL_IF(getput_arg(compiler, mem_flags, TMP_REG2, src, srcw, dst, dstw));
1417		src = TMP_REG2;
1418	}
1419
1420	if (src & SLJIT_IMM) {
1421		flags |= ARG2_IMM;
1422		if (op_flags & SLJIT_I32_OP)
1423			srcw = (sljit_s32)srcw;
1424	} else
1425		srcw = src;
1426
1427	emit_op_imm(compiler, flags | op, dst_r, TMP_REG1, srcw);
1428
1429	if (dst & SLJIT_MEM) {
1430		if (getput_arg_fast(compiler, mem_flags | STORE, dst_r, dst, dstw))
1431			return compiler->error;
1432		else
1433			return getput_arg(compiler, mem_flags | STORE, dst_r, dst, dstw, 0, 0);
1434	}
1435	return SLJIT_SUCCESS;
1436}
1437
1438SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op2(struct sljit_compiler *compiler, sljit_s32 op,
1439	sljit_s32 dst, sljit_sw dstw,
1440	sljit_s32 src1, sljit_sw src1w,
1441	sljit_s32 src2, sljit_sw src2w)
1442{
1443	sljit_s32 dst_r, flags, mem_flags;
1444
1445	CHECK_ERROR();
1446	CHECK(check_sljit_emit_op2(compiler, op, dst, dstw, src1, src1w, src2, src2w));
1447	ADJUST_LOCAL_OFFSET(dst, dstw);
1448	ADJUST_LOCAL_OFFSET(src1, src1w);
1449	ADJUST_LOCAL_OFFSET(src2, src2w);
1450
1451	compiler->cache_arg = 0;
1452	compiler->cache_argw = 0;
1453
1454	dst_r = SLOW_IS_REG(dst) ? dst : TMP_REG1;
1455	flags = HAS_FLAGS(op) ? SET_FLAGS : 0;
1456	mem_flags = WORD_SIZE;
1457	if (op & SLJIT_I32_OP) {
1458		flags |= INT_OP;
1459		mem_flags = INT_SIZE;
1460	}
1461
1462	if (dst == SLJIT_UNUSED)
1463		flags |= UNUSED_RETURN;
1464
1465	if ((dst & SLJIT_MEM) && !getput_arg_fast(compiler, mem_flags | STORE | ARG_TEST, TMP_REG1, dst, dstw))
1466		flags |= SLOW_DEST;
1467
1468	if (src1 & SLJIT_MEM) {
1469		if (getput_arg_fast(compiler, mem_flags, TMP_REG1, src1, src1w))
1470			FAIL_IF(compiler->error);
1471		else
1472			flags |= SLOW_SRC1;
1473	}
1474	if (src2 & SLJIT_MEM) {
1475		if (getput_arg_fast(compiler, mem_flags, TMP_REG2, src2, src2w))
1476			FAIL_IF(compiler->error);
1477		else
1478			flags |= SLOW_SRC2;
1479	}
1480
1481	if ((flags & (SLOW_SRC1 | SLOW_SRC2)) == (SLOW_SRC1 | SLOW_SRC2)) {
1482		if (!can_cache(src1, src1w, src2, src2w) && can_cache(src1, src1w, dst, dstw)) {
1483			FAIL_IF(getput_arg(compiler, mem_flags, TMP_REG2, src2, src2w, src1, src1w));
1484			FAIL_IF(getput_arg(compiler, mem_flags, TMP_REG1, src1, src1w, dst, dstw));
1485		}
1486		else {
1487			FAIL_IF(getput_arg(compiler, mem_flags, TMP_REG1, src1, src1w, src2, src2w));
1488			FAIL_IF(getput_arg(compiler, mem_flags, TMP_REG2, src2, src2w, dst, dstw));
1489		}
1490	}
1491	else if (flags & SLOW_SRC1)
1492		FAIL_IF(getput_arg(compiler, mem_flags, TMP_REG1, src1, src1w, dst, dstw));
1493	else if (flags & SLOW_SRC2)
1494		FAIL_IF(getput_arg(compiler, mem_flags, TMP_REG2, src2, src2w, dst, dstw));
1495
1496	if (src1 & SLJIT_MEM)
1497		src1 = TMP_REG1;
1498	if (src2 & SLJIT_MEM)
1499		src2 = TMP_REG2;
1500
1501	if (src1 & SLJIT_IMM)
1502		flags |= ARG1_IMM;
1503	else
1504		src1w = src1;
1505	if (src2 & SLJIT_IMM)
1506		flags |= ARG2_IMM;
1507	else
1508		src2w = src2;
1509
1510	emit_op_imm(compiler, flags | GET_OPCODE(op), dst_r, src1w, src2w);
1511
1512	if (dst & SLJIT_MEM) {
1513		if (!(flags & SLOW_DEST)) {
1514			getput_arg_fast(compiler, mem_flags | STORE, dst_r, dst, dstw);
1515			return compiler->error;
1516		}
1517		return getput_arg(compiler, mem_flags | STORE, TMP_REG1, dst, dstw, 0, 0);
1518	}
1519
1520	return SLJIT_SUCCESS;
1521}
1522
1523SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_get_register_index(sljit_s32 reg)
1524{
1525	CHECK_REG_INDEX(check_sljit_get_register_index(reg));
1526	return reg_map[reg];
1527}
1528
1529SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_get_float_register_index(sljit_s32 reg)
1530{
1531	CHECK_REG_INDEX(check_sljit_get_float_register_index(reg));
1532	return reg;
1533}
1534
1535SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op_custom(struct sljit_compiler *compiler,
1536	void *instruction, sljit_s32 size)
1537{
1538	CHECK_ERROR();
1539	CHECK(check_sljit_emit_op_custom(compiler, instruction, size));
1540
1541	return push_inst(compiler, *(sljit_ins*)instruction);
1542}
1543
1544/* --------------------------------------------------------------------- */
1545/*  Floating point operators                                             */
1546/* --------------------------------------------------------------------- */
1547
1548SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_is_fpu_available(void)
1549{
1550#ifdef SLJIT_IS_FPU_AVAILABLE
1551	return SLJIT_IS_FPU_AVAILABLE;
1552#else
1553	/* Available by default. */
1554	return 1;
1555#endif
1556}
1557
1558static sljit_s32 emit_fop_mem(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg, sljit_s32 arg, sljit_sw argw)
1559{
1560	sljit_u32 shift = MEM_SIZE_SHIFT(flags);
1561	sljit_ins ins_bits = (shift << 30);
1562	sljit_s32 other_r;
1563	sljit_sw diff;
1564
1565	SLJIT_ASSERT(arg & SLJIT_MEM);
1566
1567	if (!(flags & STORE))
1568		ins_bits |= 1 << 22;
1569
1570	if (arg & OFFS_REG_MASK) {
1571		argw &= 3;
1572		if (!argw || argw == shift)
1573			return push_inst(compiler, STR_FR | ins_bits | VT(reg)
1574				| RN(arg & REG_MASK) | RM(OFFS_REG(arg)) | (argw ? (1 << 12) : 0));
1575		other_r = OFFS_REG(arg);
1576		arg &= REG_MASK;
1577		FAIL_IF(push_inst(compiler, ADD | RD(TMP_REG1) | RN(arg) | RM(other_r) | (argw << 10)));
1578		arg = TMP_REG1;
1579		argw = 0;
1580	}
1581
1582	arg &= REG_MASK;
1583	if (arg && argw >= 0 && ((argw >> shift) <= 0xfff) && (argw & ((1 << shift) - 1)) == 0)
1584		return push_inst(compiler, STR_FI | ins_bits | VT(reg) | RN(arg) | (argw << (10 - shift)));
1585
1586	if (arg && argw <= 255 && argw >= -256)
1587		return push_inst(compiler, STUR_FI | ins_bits | VT(reg) | RN(arg) | ((argw & 0x1ff) << 12));
1588
1589	/* Slow cases */
1590	if (compiler->cache_arg == SLJIT_MEM && argw != compiler->cache_argw) {
1591		diff = argw - compiler->cache_argw;
1592		if (!arg && diff <= 255 && diff >= -256)
1593			return push_inst(compiler, STUR_FI | ins_bits | VT(reg) | RN(TMP_REG3) | ((diff & 0x1ff) << 12));
1594		if (emit_set_delta(compiler, TMP_REG3, TMP_REG3, argw - compiler->cache_argw) != SLJIT_ERR_UNSUPPORTED) {
1595			FAIL_IF(compiler->error);
1596			compiler->cache_argw = argw;
1597		}
1598	}
1599
1600	if (compiler->cache_arg != SLJIT_MEM || argw != compiler->cache_argw) {
1601		compiler->cache_arg = SLJIT_MEM;
1602		compiler->cache_argw = argw;
1603		FAIL_IF(load_immediate(compiler, TMP_REG3, argw));
1604	}
1605
1606	if (arg & REG_MASK)
1607		return push_inst(compiler, STR_FR | ins_bits | VT(reg) | RN(arg) | RM(TMP_REG3));
1608	return push_inst(compiler, STR_FI | ins_bits | VT(reg) | RN(TMP_REG3));
1609}
1610
1611static SLJIT_INLINE sljit_s32 sljit_emit_fop1_conv_sw_from_f64(struct sljit_compiler *compiler, sljit_s32 op,
1612	sljit_s32 dst, sljit_sw dstw,
1613	sljit_s32 src, sljit_sw srcw)
1614{
1615	sljit_s32 dst_r = SLOW_IS_REG(dst) ? dst : TMP_REG1;
1616	sljit_ins inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0;
1617
1618	if (GET_OPCODE(op) == SLJIT_CONV_S32_FROM_F64)
1619		inv_bits |= (1 << 31);
1620
1621	if (src & SLJIT_MEM) {
1622		emit_fop_mem(compiler, (op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE, TMP_FREG1, src, srcw);
1623		src = TMP_FREG1;
1624	}
1625
1626	FAIL_IF(push_inst(compiler, (FCVTZS ^ inv_bits) | RD(dst_r) | VN(src)));
1627
1628	if (dst_r == TMP_REG1 && dst != SLJIT_UNUSED)
1629		return emit_op_mem(compiler, ((GET_OPCODE(op) == SLJIT_CONV_S32_FROM_F64) ? INT_SIZE : WORD_SIZE) | STORE, TMP_REG1, dst, dstw);
1630	return SLJIT_SUCCESS;
1631}
1632
1633static SLJIT_INLINE sljit_s32 sljit_emit_fop1_conv_f64_from_sw(struct sljit_compiler *compiler, sljit_s32 op,
1634	sljit_s32 dst, sljit_sw dstw,
1635	sljit_s32 src, sljit_sw srcw)
1636{
1637	sljit_s32 dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG1;
1638	sljit_ins inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0;
1639
1640	if (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_S32)
1641		inv_bits |= (1 << 31);
1642
1643	if (src & SLJIT_MEM) {
1644		emit_op_mem(compiler, ((GET_OPCODE(op) == SLJIT_CONV_F64_FROM_S32) ? INT_SIZE : WORD_SIZE), TMP_REG1, src, srcw);
1645		src = TMP_REG1;
1646	} else if (src & SLJIT_IMM) {
1647#if (defined SLJIT_CONFIG_X86_64 && SLJIT_CONFIG_X86_64)
1648		if (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_S32)
1649			srcw = (sljit_s32)srcw;
1650#endif
1651		FAIL_IF(load_immediate(compiler, TMP_REG1, srcw));
1652		src = TMP_REG1;
1653	}
1654
1655	FAIL_IF(push_inst(compiler, (SCVTF ^ inv_bits) | VD(dst_r) | RN(src)));
1656
1657	if (dst & SLJIT_MEM)
1658		return emit_fop_mem(compiler, ((op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE) | STORE, TMP_FREG1, dst, dstw);
1659	return SLJIT_SUCCESS;
1660}
1661
1662static SLJIT_INLINE sljit_s32 sljit_emit_fop1_cmp(struct sljit_compiler *compiler, sljit_s32 op,
1663	sljit_s32 src1, sljit_sw src1w,
1664	sljit_s32 src2, sljit_sw src2w)
1665{
1666	sljit_s32 mem_flags = (op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE;
1667	sljit_ins inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0;
1668
1669	if (src1 & SLJIT_MEM) {
1670		emit_fop_mem(compiler, mem_flags, TMP_FREG1, src1, src1w);
1671		src1 = TMP_FREG1;
1672	}
1673
1674	if (src2 & SLJIT_MEM) {
1675		emit_fop_mem(compiler, mem_flags, TMP_FREG2, src2, src2w);
1676		src2 = TMP_FREG2;
1677	}
1678
1679	return push_inst(compiler, (FCMP ^ inv_bits) | VN(src1) | VM(src2));
1680}
1681
1682SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fop1(struct sljit_compiler *compiler, sljit_s32 op,
1683	sljit_s32 dst, sljit_sw dstw,
1684	sljit_s32 src, sljit_sw srcw)
1685{
1686	sljit_s32 dst_r, mem_flags = (op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE;
1687	sljit_ins inv_bits;
1688
1689	CHECK_ERROR();
1690	compiler->cache_arg = 0;
1691	compiler->cache_argw = 0;
1692
1693	SLJIT_COMPILE_ASSERT((INT_SIZE ^ 0x100) == WORD_SIZE, must_be_one_bit_difference);
1694	SELECT_FOP1_OPERATION_WITH_CHECKS(compiler, op, dst, dstw, src, srcw);
1695
1696	inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0;
1697	dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG1;
1698
1699	if (src & SLJIT_MEM) {
1700		emit_fop_mem(compiler, (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_F32) ? (mem_flags ^ 0x100) : mem_flags, dst_r, src, srcw);
1701		src = dst_r;
1702	}
1703
1704	switch (GET_OPCODE(op)) {
1705	case SLJIT_MOV_F64:
1706		if (src != dst_r) {
1707			if (dst_r != TMP_FREG1)
1708				FAIL_IF(push_inst(compiler, (FMOV ^ inv_bits) | VD(dst_r) | VN(src)));
1709			else
1710				dst_r = src;
1711		}
1712		break;
1713	case SLJIT_NEG_F64:
1714		FAIL_IF(push_inst(compiler, (FNEG ^ inv_bits) | VD(dst_r) | VN(src)));
1715		break;
1716	case SLJIT_ABS_F64:
1717		FAIL_IF(push_inst(compiler, (FABS ^ inv_bits) | VD(dst_r) | VN(src)));
1718		break;
1719	case SLJIT_CONV_F64_FROM_F32:
1720		FAIL_IF(push_inst(compiler, FCVT | ((op & SLJIT_F32_OP) ? (1 << 22) : (1 << 15)) | VD(dst_r) | VN(src)));
1721		break;
1722	}
1723
1724	if (dst & SLJIT_MEM)
1725		return emit_fop_mem(compiler, mem_flags | STORE, dst_r, dst, dstw);
1726	return SLJIT_SUCCESS;
1727}
1728
1729SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fop2(struct sljit_compiler *compiler, sljit_s32 op,
1730	sljit_s32 dst, sljit_sw dstw,
1731	sljit_s32 src1, sljit_sw src1w,
1732	sljit_s32 src2, sljit_sw src2w)
1733{
1734	sljit_s32 dst_r, mem_flags = (op & SLJIT_F32_OP) ? INT_SIZE : WORD_SIZE;
1735	sljit_ins inv_bits = (op & SLJIT_F32_OP) ? (1 << 22) : 0;
1736
1737	CHECK_ERROR();
1738	CHECK(check_sljit_emit_fop2(compiler, op, dst, dstw, src1, src1w, src2, src2w));
1739	ADJUST_LOCAL_OFFSET(dst, dstw);
1740	ADJUST_LOCAL_OFFSET(src1, src1w);
1741	ADJUST_LOCAL_OFFSET(src2, src2w);
1742
1743	compiler->cache_arg = 0;
1744	compiler->cache_argw = 0;
1745
1746	dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG1;
1747	if (src1 & SLJIT_MEM) {
1748		emit_fop_mem(compiler, mem_flags, TMP_FREG1, src1, src1w);
1749		src1 = TMP_FREG1;
1750	}
1751	if (src2 & SLJIT_MEM) {
1752		emit_fop_mem(compiler, mem_flags, TMP_FREG2, src2, src2w);
1753		src2 = TMP_FREG2;
1754	}
1755
1756	switch (GET_OPCODE(op)) {
1757	case SLJIT_ADD_F64:
1758		FAIL_IF(push_inst(compiler, (FADD ^ inv_bits) | VD(dst_r) | VN(src1) | VM(src2)));
1759		break;
1760	case SLJIT_SUB_F64:
1761		FAIL_IF(push_inst(compiler, (FSUB ^ inv_bits) | VD(dst_r) | VN(src1) | VM(src2)));
1762		break;
1763	case SLJIT_MUL_F64:
1764		FAIL_IF(push_inst(compiler, (FMUL ^ inv_bits) | VD(dst_r) | VN(src1) | VM(src2)));
1765		break;
1766	case SLJIT_DIV_F64:
1767		FAIL_IF(push_inst(compiler, (FDIV ^ inv_bits) | VD(dst_r) | VN(src1) | VM(src2)));
1768		break;
1769	}
1770
1771	if (!(dst & SLJIT_MEM))
1772		return SLJIT_SUCCESS;
1773	return emit_fop_mem(compiler, mem_flags | STORE, TMP_FREG1, dst, dstw);
1774}
1775
1776/* --------------------------------------------------------------------- */
1777/*  Other instructions                                                   */
1778/* --------------------------------------------------------------------- */
1779
1780SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fast_enter(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw)
1781{
1782	CHECK_ERROR();
1783	CHECK(check_sljit_emit_fast_enter(compiler, dst, dstw));
1784	ADJUST_LOCAL_OFFSET(dst, dstw);
1785
1786	/* For UNUSED dst. Uncommon, but possible. */
1787	if (dst == SLJIT_UNUSED)
1788		return SLJIT_SUCCESS;
1789
1790	if (FAST_IS_REG(dst))
1791		return push_inst(compiler, ORR | RD(dst) | RN(TMP_ZERO) | RM(TMP_LR));
1792
1793	/* Memory. */
1794	return emit_op_mem(compiler, WORD_SIZE | STORE, TMP_LR, dst, dstw);
1795}
1796
1797SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fast_return(struct sljit_compiler *compiler, sljit_s32 src, sljit_sw srcw)
1798{
1799	CHECK_ERROR();
1800	CHECK(check_sljit_emit_fast_return(compiler, src, srcw));
1801	ADJUST_LOCAL_OFFSET(src, srcw);
1802
1803	if (FAST_IS_REG(src))
1804		FAIL_IF(push_inst(compiler, ORR | RD(TMP_LR) | RN(TMP_ZERO) | RM(src)));
1805	else if (src & SLJIT_MEM)
1806		FAIL_IF(emit_op_mem(compiler, WORD_SIZE, TMP_LR, src, srcw));
1807	else if (src & SLJIT_IMM)
1808		FAIL_IF(load_immediate(compiler, TMP_LR, srcw));
1809
1810	return push_inst(compiler, RET | RN(TMP_LR));
1811}
1812
1813/* --------------------------------------------------------------------- */
1814/*  Conditional instructions                                             */
1815/* --------------------------------------------------------------------- */
1816
1817static sljit_uw get_cc(sljit_s32 type)
1818{
1819	switch (type) {
1820	case SLJIT_EQUAL:
1821	case SLJIT_MUL_NOT_OVERFLOW:
1822	case SLJIT_EQUAL_F64:
1823		return 0x1;
1824
1825	case SLJIT_NOT_EQUAL:
1826	case SLJIT_MUL_OVERFLOW:
1827	case SLJIT_NOT_EQUAL_F64:
1828		return 0x0;
1829
1830	case SLJIT_LESS:
1831	case SLJIT_LESS_F64:
1832		return 0x2;
1833
1834	case SLJIT_GREATER_EQUAL:
1835	case SLJIT_GREATER_EQUAL_F64:
1836		return 0x3;
1837
1838	case SLJIT_GREATER:
1839	case SLJIT_GREATER_F64:
1840		return 0x9;
1841
1842	case SLJIT_LESS_EQUAL:
1843	case SLJIT_LESS_EQUAL_F64:
1844		return 0x8;
1845
1846	case SLJIT_SIG_LESS:
1847		return 0xa;
1848
1849	case SLJIT_SIG_GREATER_EQUAL:
1850		return 0xb;
1851
1852	case SLJIT_SIG_GREATER:
1853		return 0xd;
1854
1855	case SLJIT_SIG_LESS_EQUAL:
1856		return 0xc;
1857
1858	case SLJIT_OVERFLOW:
1859	case SLJIT_UNORDERED_F64:
1860		return 0x7;
1861
1862	case SLJIT_NOT_OVERFLOW:
1863	case SLJIT_ORDERED_F64:
1864		return 0x6;
1865
1866	default:
1867		SLJIT_UNREACHABLE();
1868		return 0xe;
1869	}
1870}
1871
1872SLJIT_API_FUNC_ATTRIBUTE struct sljit_label* sljit_emit_label(struct sljit_compiler *compiler)
1873{
1874	struct sljit_label *label;
1875
1876	CHECK_ERROR_PTR();
1877	CHECK_PTR(check_sljit_emit_label(compiler));
1878
1879	if (compiler->last_label && compiler->last_label->size == compiler->size)
1880		return compiler->last_label;
1881
1882	label = (struct sljit_label*)ensure_abuf(compiler, sizeof(struct sljit_label));
1883	PTR_FAIL_IF(!label);
1884	set_label(label, compiler);
1885	return label;
1886}
1887
1888SLJIT_API_FUNC_ATTRIBUTE struct sljit_jump* sljit_emit_jump(struct sljit_compiler *compiler, sljit_s32 type)
1889{
1890	struct sljit_jump *jump;
1891
1892	CHECK_ERROR_PTR();
1893	CHECK_PTR(check_sljit_emit_jump(compiler, type));
1894
1895	jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump));
1896	PTR_FAIL_IF(!jump);
1897	set_jump(jump, compiler, type & SLJIT_REWRITABLE_JUMP);
1898	type &= 0xff;
1899
1900	if (type < SLJIT_JUMP) {
1901		jump->flags |= IS_COND;
1902		PTR_FAIL_IF(push_inst(compiler, B_CC | (6 << 5) | get_cc(type)));
1903	}
1904	else if (type >= SLJIT_FAST_CALL)
1905		jump->flags |= IS_BL;
1906
1907	PTR_FAIL_IF(emit_imm64_const(compiler, TMP_REG1, 0));
1908	jump->addr = compiler->size;
1909	PTR_FAIL_IF(push_inst(compiler, ((type >= SLJIT_FAST_CALL) ? BLR : BR) | RN(TMP_REG1)));
1910
1911	return jump;
1912}
1913
1914static SLJIT_INLINE struct sljit_jump* emit_cmp_to0(struct sljit_compiler *compiler, sljit_s32 type,
1915	sljit_s32 src, sljit_sw srcw)
1916{
1917	struct sljit_jump *jump;
1918	sljit_ins inv_bits = (type & SLJIT_I32_OP) ? (1 << 31) : 0;
1919
1920	SLJIT_ASSERT((type & 0xff) == SLJIT_EQUAL || (type & 0xff) == SLJIT_NOT_EQUAL);
1921	ADJUST_LOCAL_OFFSET(src, srcw);
1922
1923	jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump));
1924	PTR_FAIL_IF(!jump);
1925	set_jump(jump, compiler, type & SLJIT_REWRITABLE_JUMP);
1926	jump->flags |= IS_CBZ | IS_COND;
1927
1928	if (src & SLJIT_MEM) {
1929		PTR_FAIL_IF(emit_op_mem(compiler, inv_bits ? INT_SIZE : WORD_SIZE, TMP_REG1, src, srcw));
1930		src = TMP_REG1;
1931	}
1932	else if (src & SLJIT_IMM) {
1933		PTR_FAIL_IF(load_immediate(compiler, TMP_REG1, srcw));
1934		src = TMP_REG1;
1935	}
1936	SLJIT_ASSERT(FAST_IS_REG(src));
1937
1938	if ((type & 0xff) == SLJIT_EQUAL)
1939		inv_bits |= 1 << 24;
1940
1941	PTR_FAIL_IF(push_inst(compiler, (CBZ ^ inv_bits) | (6 << 5) | RT(src)));
1942	PTR_FAIL_IF(emit_imm64_const(compiler, TMP_REG1, 0));
1943	jump->addr = compiler->size;
1944	PTR_FAIL_IF(push_inst(compiler, BR | RN(TMP_REG1)));
1945	return jump;
1946}
1947
1948SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_ijump(struct sljit_compiler *compiler, sljit_s32 type, sljit_s32 src, sljit_sw srcw)
1949{
1950	struct sljit_jump *jump;
1951
1952	CHECK_ERROR();
1953	CHECK(check_sljit_emit_ijump(compiler, type, src, srcw));
1954	ADJUST_LOCAL_OFFSET(src, srcw);
1955
1956	/* In ARM, we don't need to touch the arguments. */
1957	if (!(src & SLJIT_IMM)) {
1958		if (src & SLJIT_MEM) {
1959			FAIL_IF(emit_op_mem(compiler, WORD_SIZE, TMP_REG1, src, srcw));
1960			src = TMP_REG1;
1961		}
1962		return push_inst(compiler, ((type >= SLJIT_FAST_CALL) ? BLR : BR) | RN(src));
1963	}
1964
1965	jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump));
1966	FAIL_IF(!jump);
1967	set_jump(jump, compiler, JUMP_ADDR | ((type >= SLJIT_FAST_CALL) ? IS_BL : 0));
1968	jump->u.target = srcw;
1969
1970	FAIL_IF(emit_imm64_const(compiler, TMP_REG1, 0));
1971	jump->addr = compiler->size;
1972	return push_inst(compiler, ((type >= SLJIT_FAST_CALL) ? BLR : BR) | RN(TMP_REG1));
1973}
1974
1975SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op_flags(struct sljit_compiler *compiler, sljit_s32 op,
1976	sljit_s32 dst, sljit_sw dstw,
1977	sljit_s32 src, sljit_sw srcw,
1978	sljit_s32 type)
1979{
1980	sljit_s32 dst_r, flags, mem_flags;
1981	sljit_ins cc;
1982
1983	CHECK_ERROR();
1984	CHECK(check_sljit_emit_op_flags(compiler, op, dst, dstw, src, srcw, type));
1985	ADJUST_LOCAL_OFFSET(dst, dstw);
1986	ADJUST_LOCAL_OFFSET(src, srcw);
1987
1988	if (dst == SLJIT_UNUSED)
1989		return SLJIT_SUCCESS;
1990
1991	cc = get_cc(type & 0xff);
1992	dst_r = FAST_IS_REG(dst) ? dst : TMP_REG1;
1993
1994	if (GET_OPCODE(op) < SLJIT_ADD) {
1995		FAIL_IF(push_inst(compiler, CSINC | (cc << 12) | RD(dst_r) | RN(TMP_ZERO) | RM(TMP_ZERO)));
1996		if (dst_r != TMP_REG1)
1997			return SLJIT_SUCCESS;
1998		return emit_op_mem(compiler, (GET_OPCODE(op) == SLJIT_MOV ? WORD_SIZE : INT_SIZE) | STORE, TMP_REG1, dst, dstw);
1999	}
2000
2001	compiler->cache_arg = 0;
2002	compiler->cache_argw = 0;
2003	flags = HAS_FLAGS(op) ? SET_FLAGS : 0;
2004	mem_flags = WORD_SIZE;
2005	if (op & SLJIT_I32_OP) {
2006		flags |= INT_OP;
2007		mem_flags = INT_SIZE;
2008	}
2009
2010	if (src & SLJIT_MEM) {
2011		FAIL_IF(emit_op_mem2(compiler, mem_flags, TMP_REG1, src, srcw, dst, dstw));
2012		src = TMP_REG1;
2013		srcw = 0;
2014	} else if (src & SLJIT_IMM)
2015		flags |= ARG1_IMM;
2016
2017	FAIL_IF(push_inst(compiler, CSINC | (cc << 12) | RD(TMP_REG2) | RN(TMP_ZERO) | RM(TMP_ZERO)));
2018	emit_op_imm(compiler, flags | GET_OPCODE(op), dst_r, src, TMP_REG2);
2019
2020	if (dst_r != TMP_REG1)
2021		return SLJIT_SUCCESS;
2022	return emit_op_mem2(compiler, mem_flags | STORE, TMP_REG1, dst, dstw, 0, 0);
2023}
2024
2025SLJIT_API_FUNC_ATTRIBUTE struct sljit_const* sljit_emit_const(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw, sljit_sw init_value)
2026{
2027	struct sljit_const *const_;
2028	sljit_s32 dst_r;
2029
2030	CHECK_ERROR_PTR();
2031	CHECK_PTR(check_sljit_emit_const(compiler, dst, dstw, init_value));
2032	ADJUST_LOCAL_OFFSET(dst, dstw);
2033
2034	const_ = (struct sljit_const*)ensure_abuf(compiler, sizeof(struct sljit_const));
2035	PTR_FAIL_IF(!const_);
2036	set_const(const_, compiler);
2037
2038	dst_r = SLOW_IS_REG(dst) ? dst : TMP_REG1;
2039	PTR_FAIL_IF(emit_imm64_const(compiler, dst_r, init_value));
2040
2041	if (dst & SLJIT_MEM)
2042		PTR_FAIL_IF(emit_op_mem(compiler, WORD_SIZE | STORE, dst_r, dst, dstw));
2043	return const_;
2044}
2045
2046SLJIT_API_FUNC_ATTRIBUTE void sljit_set_jump_addr(sljit_uw addr, sljit_uw new_target, sljit_sw executable_offset)
2047{
2048	sljit_ins* inst = (sljit_ins*)addr;
2049	modify_imm64_const(inst, new_target);
2050	inst = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(inst, executable_offset);
2051	SLJIT_CACHE_FLUSH(inst, inst + 4);
2052}
2053
2054SLJIT_API_FUNC_ATTRIBUTE void sljit_set_const(sljit_uw addr, sljit_sw new_constant, sljit_sw executable_offset)
2055{
2056	sljit_ins* inst = (sljit_ins*)addr;
2057	modify_imm64_const(inst, new_constant);
2058	inst = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(inst, executable_offset);
2059	SLJIT_CACHE_FLUSH(inst, inst + 4);
2060}
2061