Lines Matching refs:emit

68 static inline void emit(const u32 insn, struct jit_ctx *ctx)  in emit()  function
82 emit(A64_MOVZ(1, reg, tmp & 0xffff, shift), ctx); in emit_a64_mov_i64()
87 emit(A64_MOVK(1, reg, tmp & 0xffff, shift), ctx); in emit_a64_mov_i64()
101 emit(A64_MOVN(is64, reg, (u16)~lo, 0), ctx); in emit_a64_mov_i()
103 emit(A64_MOVN(is64, reg, (u16)~hi, 16), ctx); in emit_a64_mov_i()
104 emit(A64_MOVK(is64, reg, lo, 0), ctx); in emit_a64_mov_i()
107 emit(A64_MOVZ(is64, reg, lo, 0), ctx); in emit_a64_mov_i()
109 emit(A64_MOVK(is64, reg, hi, 16), ctx); in emit_a64_mov_i()
186 emit(A64_PUSH(A64_FP, A64_LR, A64_SP), ctx); in build_prologue()
187 emit(A64_MOV(1, A64_FP, A64_SP), ctx); in build_prologue()
190 emit(A64_PUSH(r6, r7, A64_SP), ctx); in build_prologue()
191 emit(A64_PUSH(r8, r9, A64_SP), ctx); in build_prologue()
193 emit(A64_PUSH(tmp1, tmp2, A64_SP), ctx); in build_prologue()
196 emit(A64_PUSH(fp, A64_R(26), A64_SP), ctx); in build_prologue()
199 emit(A64_MOV(1, fp, A64_SP), ctx); in build_prologue()
202 emit(A64_SUB_I(1, A64_SP, A64_SP, STACK_SIZE), ctx); in build_prologue()
221 emit(A64_ADD_I(1, A64_SP, A64_SP, STACK_SIZE), ctx); in build_epilogue()
224 emit(A64_POP(fp, A64_R(26), A64_SP), ctx); in build_epilogue()
228 emit(A64_POP(tmp1, tmp2, A64_SP), ctx); in build_epilogue()
229 emit(A64_POP(r8, r9, A64_SP), ctx); in build_epilogue()
230 emit(A64_POP(r6, r7, A64_SP), ctx); in build_epilogue()
233 emit(A64_POP(A64_FP, A64_LR, A64_SP), ctx); in build_epilogue()
236 emit(A64_MOV(1, A64_R(0), r0), ctx); in build_epilogue()
238 emit(A64_RET(A64_LR), ctx); in build_epilogue()
276 emit(A64_MOV(is64, dst, src), ctx); in build_insn()
281 emit(A64_ADD(is64, dst, dst, src), ctx); in build_insn()
285 emit(A64_SUB(is64, dst, dst, src), ctx); in build_insn()
289 emit(A64_AND(is64, dst, dst, src), ctx); in build_insn()
293 emit(A64_ORR(is64, dst, dst, src), ctx); in build_insn()
297 emit(A64_EOR(is64, dst, dst, src), ctx); in build_insn()
301 emit(A64_MUL(is64, dst, dst, src), ctx); in build_insn()
313 emit(A64_CBNZ(is64, src, jmp_offset), ctx); in build_insn()
314 emit(A64_MOVZ(1, r0, 0, 0), ctx); in build_insn()
317 emit(A64_B(jmp_offset), ctx); in build_insn()
321 emit(A64_UDIV(is64, dst, dst, src), ctx); in build_insn()
325 emit(A64_UDIV(is64, tmp, dst, src), ctx); in build_insn()
326 emit(A64_MUL(is64, tmp, tmp, src), ctx); in build_insn()
327 emit(A64_SUB(is64, dst, dst, tmp), ctx); in build_insn()
334 emit(A64_LSLV(is64, dst, dst, src), ctx); in build_insn()
338 emit(A64_LSRV(is64, dst, dst, src), ctx); in build_insn()
342 emit(A64_ASRV(is64, dst, dst, src), ctx); in build_insn()
347 emit(A64_NEG(is64, dst, dst), ctx); in build_insn()
361 emit(A64_REV16(is64, dst, dst), ctx); in build_insn()
363 emit(A64_UXTH(is64, dst, dst), ctx); in build_insn()
366 emit(A64_REV32(is64, dst, dst), ctx); in build_insn()
370 emit(A64_REV64(dst, dst), ctx); in build_insn()
378 emit(A64_UXTH(is64, dst, dst), ctx); in build_insn()
382 emit(A64_UXTW(is64, dst, dst), ctx); in build_insn()
399 emit(A64_ADD(is64, dst, dst, tmp), ctx); in build_insn()
405 emit(A64_SUB(is64, dst, dst, tmp), ctx); in build_insn()
411 emit(A64_AND(is64, dst, dst, tmp), ctx); in build_insn()
417 emit(A64_ORR(is64, dst, dst, tmp), ctx); in build_insn()
423 emit(A64_EOR(is64, dst, dst, tmp), ctx); in build_insn()
429 emit(A64_MUL(is64, dst, dst, tmp), ctx); in build_insn()
435 emit(A64_UDIV(is64, dst, dst, tmp), ctx); in build_insn()
441 emit(A64_UDIV(is64, tmp, dst, tmp2), ctx); in build_insn()
442 emit(A64_MUL(is64, tmp, tmp, tmp2), ctx); in build_insn()
443 emit(A64_SUB(is64, dst, dst, tmp), ctx); in build_insn()
447 emit(A64_LSL(is64, dst, dst, imm), ctx); in build_insn()
451 emit(A64_LSR(is64, dst, dst, imm), ctx); in build_insn()
455 emit(A64_ASR(is64, dst, dst, imm), ctx); in build_insn()
462 emit(A64_B(jmp_offset), ctx); in build_insn()
471 emit(A64_CMP(1, dst, src), ctx); in build_insn()
497 emit(A64_B_(jmp_cond, jmp_offset), ctx); in build_insn()
500 emit(A64_TST(1, dst, src), ctx); in build_insn()
511 emit(A64_CMP(1, dst, tmp), ctx); in build_insn()
516 emit(A64_TST(1, dst, tmp), ctx); in build_insn()
526 emit(A64_PUSH(A64_FP, A64_LR, A64_SP), ctx); in build_insn()
527 emit(A64_MOV(1, A64_FP, A64_SP), ctx); in build_insn()
528 emit(A64_BLR(tmp), ctx); in build_insn()
529 emit(A64_MOV(1, r0, A64_R(0)), ctx); in build_insn()
530 emit(A64_POP(A64_FP, A64_LR, A64_SP), ctx); in build_insn()
541 emit(A64_B(jmp_offset), ctx); in build_insn()
574 emit(A64_LDR32(dst, src, tmp), ctx); in build_insn()
577 emit(A64_LDRH(dst, src, tmp), ctx); in build_insn()
580 emit(A64_LDRB(dst, src, tmp), ctx); in build_insn()
583 emit(A64_LDR64(dst, src, tmp), ctx); in build_insn()
599 emit(A64_STR32(tmp, dst, tmp2), ctx); in build_insn()
602 emit(A64_STRH(tmp, dst, tmp2), ctx); in build_insn()
605 emit(A64_STRB(tmp, dst, tmp2), ctx); in build_insn()
608 emit(A64_STR64(tmp, dst, tmp2), ctx); in build_insn()
622 emit(A64_STR32(src, dst, tmp), ctx); in build_insn()
625 emit(A64_STRH(src, dst, tmp), ctx); in build_insn()
628 emit(A64_STRB(src, dst, tmp), ctx); in build_insn()
631 emit(A64_STR64(src, dst, tmp), ctx); in build_insn()
660 emit(A64_MOV(1, r1, r6), ctx); in build_insn()
663 emit(A64_ADD(0, r2, r2, src), ctx); in build_insn()
678 emit(A64_SUB_I(1, r4, fp, STACK_SIZE), ctx); in build_insn()
680 emit(A64_PUSH(A64_FP, A64_LR, A64_SP), ctx); in build_insn()
681 emit(A64_MOV(1, A64_FP, A64_SP), ctx); in build_insn()
682 emit(A64_BLR(r5), ctx); in build_insn()
683 emit(A64_MOV(1, r0, A64_R(0)), ctx); in build_insn()
684 emit(A64_POP(A64_FP, A64_LR, A64_SP), ctx); in build_insn()
688 emit(A64_CBZ(1, r0, jmp_offset), ctx); in build_insn()
689 emit(A64_MOV(1, r5, r0), ctx); in build_insn()
692 emit(A64_LDR32(r0, r5, A64_ZR), ctx); in build_insn()
694 emit(A64_REV32(0, r0, r0), ctx); in build_insn()
698 emit(A64_LDRH(r0, r5, A64_ZR), ctx); in build_insn()
700 emit(A64_REV16(0, r0, r0), ctx); in build_insn()
704 emit(A64_LDRB(r0, r5, A64_ZR), ctx); in build_insn()