imm16              11 arch/arm/include/asm/opcodes-virt.h #define __HVC(imm16) __inst_arm_thumb32(				\
imm16              12 arch/arm/include/asm/opcodes-virt.h 	0xE1400070 | (((imm16) & 0xFFF0) << 4) | ((imm16) & 0x000F),	\
imm16              13 arch/arm/include/asm/opcodes-virt.h 	0xF7E08000 | (((imm16) & 0xF000) << 4) | ((imm16) & 0x0FFF)	\
imm16             130 arch/arm64/net/bpf_jit.h #define A64_MOVEW(sf, Rd, imm16, shift, type) \
imm16             131 arch/arm64/net/bpf_jit.h 	aarch64_insn_gen_movewide(Rd, imm16, shift, \
imm16             136 arch/arm64/net/bpf_jit.h #define A64_MOVN(sf, Rd, imm16, shift) A64_MOVEW(sf, Rd, imm16, shift, INVERSE)
imm16             137 arch/arm64/net/bpf_jit.h #define A64_MOVZ(sf, Rd, imm16, shift) A64_MOVEW(sf, Rd, imm16, shift, ZERO)
imm16             138 arch/arm64/net/bpf_jit.h #define A64_MOVK(sf, Rd, imm16, shift) A64_MOVEW(sf, Rd, imm16, shift, KEEP)
imm16              71 arch/nios2/kernel/misaligned.c 	s16 imm16;
imm16              92 arch/nios2/kernel/misaligned.c 		imm16 = (isn >> 6) & 0xffff;
imm16              93 arch/nios2/kernel/misaligned.c 		addr = get_reg_val(fp, a) + imm16;