imm16 11 arch/arm/include/asm/opcodes-virt.h #define __HVC(imm16) __inst_arm_thumb32( \ imm16 12 arch/arm/include/asm/opcodes-virt.h 0xE1400070 | (((imm16) & 0xFFF0) << 4) | ((imm16) & 0x000F), \ imm16 13 arch/arm/include/asm/opcodes-virt.h 0xF7E08000 | (((imm16) & 0xF000) << 4) | ((imm16) & 0x0FFF) \ imm16 130 arch/arm64/net/bpf_jit.h #define A64_MOVEW(sf, Rd, imm16, shift, type) \ imm16 131 arch/arm64/net/bpf_jit.h aarch64_insn_gen_movewide(Rd, imm16, shift, \ imm16 136 arch/arm64/net/bpf_jit.h #define A64_MOVN(sf, Rd, imm16, shift) A64_MOVEW(sf, Rd, imm16, shift, INVERSE) imm16 137 arch/arm64/net/bpf_jit.h #define A64_MOVZ(sf, Rd, imm16, shift) A64_MOVEW(sf, Rd, imm16, shift, ZERO) imm16 138 arch/arm64/net/bpf_jit.h #define A64_MOVK(sf, Rd, imm16, shift) A64_MOVEW(sf, Rd, imm16, shift, KEEP) imm16 71 arch/nios2/kernel/misaligned.c s16 imm16; imm16 92 arch/nios2/kernel/misaligned.c imm16 = (isn >> 6) & 0xffff; imm16 93 arch/nios2/kernel/misaligned.c addr = get_reg_val(fp, a) + imm16;