imm32             311 arch/x86/net/bpf_jit_comp.c 			   u32 dst_reg, const u32 imm32)
imm32             321 arch/x86/net/bpf_jit_comp.c 	if (sign_propagate && (s32)imm32 < 0) {
imm32             326 arch/x86/net/bpf_jit_comp.c 		EMIT3_off32(b1, b2, add_1reg(b3, dst_reg), imm32);
imm32             334 arch/x86/net/bpf_jit_comp.c 	if (imm32 == 0) {
imm32             346 arch/x86/net/bpf_jit_comp.c 	EMIT1_off32(add_1reg(0xB8, dst_reg), imm32);
imm32             409 arch/x86/net/bpf_jit_comp.c 		const s32 imm32 = insn->imm;
imm32             503 arch/x86/net/bpf_jit_comp.c 			if (is_imm8(imm32))
imm32             504 arch/x86/net/bpf_jit_comp.c 				EMIT3(0x83, add_1reg(b3, dst_reg), imm32);
imm32             506 arch/x86/net/bpf_jit_comp.c 				EMIT1_off32(b2, imm32);
imm32             508 arch/x86/net/bpf_jit_comp.c 				EMIT2_off32(0x81, add_1reg(b3, dst_reg), imm32);
imm32             514 arch/x86/net/bpf_jit_comp.c 				       dst_reg, imm32);
imm32             540 arch/x86/net/bpf_jit_comp.c 				EMIT3_off32(0x49, 0xC7, 0xC3, imm32);
imm32             590 arch/x86/net/bpf_jit_comp.c 				emit_mov_imm32(&prog, is64, BPF_REG_0, imm32);
imm32             626 arch/x86/net/bpf_jit_comp.c 			if (imm32 == 1)
imm32             629 arch/x86/net/bpf_jit_comp.c 				EMIT3(0xC1, add_1reg(b3, dst_reg), imm32);
imm32             675 arch/x86/net/bpf_jit_comp.c 			switch (imm32) {
imm32             707 arch/x86/net/bpf_jit_comp.c 			switch (imm32) {
imm32             758 arch/x86/net/bpf_jit_comp.c 			EMIT(imm32, bpf_size_to_x86_bytes(BPF_SIZE(insn->code)));
imm32             841 arch/x86/net/bpf_jit_comp.c 			func = (u8 *) __bpf_call_base + imm32;
imm32             843 arch/x86/net/bpf_jit_comp.c 			if (!imm32 || !is_simm32(jmp_offset)) {
imm32             845 arch/x86/net/bpf_jit_comp.c 				       imm32, func, image);
imm32             901 arch/x86/net/bpf_jit_comp.c 			EMIT2_off32(0xF7, add_1reg(0xC0, dst_reg), imm32);
imm32             930 arch/x86/net/bpf_jit_comp.c 			if (is_imm8(imm32))
imm32             931 arch/x86/net/bpf_jit_comp.c 				EMIT3(0x83, add_1reg(0xF8, dst_reg), imm32);
imm32             933 arch/x86/net/bpf_jit_comp.c 				EMIT2_off32(0x81, add_1reg(0xF8, dst_reg), imm32);
imm32            1476 arch/x86/net/bpf_jit_comp32.c 		const s32 imm32 = insn->imm;
imm32            1498 arch/x86/net/bpf_jit_comp32.c 				if (imm32 == 1) {
imm32            1508 arch/x86/net/bpf_jit_comp32.c 				emit_ia32_mov_i64(is64, dst, imm32,
imm32            1549 arch/x86/net/bpf_jit_comp32.c 						  imm32, dstk, &prog,
imm32            1564 arch/x86/net/bpf_jit_comp32.c 					    imm32);
imm32            1584 arch/x86/net/bpf_jit_comp32.c 					    imm32);
imm32            1607 arch/x86/net/bpf_jit_comp32.c 					    imm32);
imm32            1625 arch/x86/net/bpf_jit_comp32.c 			if (unlikely(imm32 > 31))
imm32            1628 arch/x86/net/bpf_jit_comp32.c 			EMIT2_off32(0xC7, add_1reg(0xC0, IA32_ECX), imm32);
imm32            1636 arch/x86/net/bpf_jit_comp32.c 			if (unlikely(imm32 > 63))
imm32            1638 arch/x86/net/bpf_jit_comp32.c 			emit_ia32_lsh_i64(dst, imm32, dstk, &prog);
imm32            1642 arch/x86/net/bpf_jit_comp32.c 			if (unlikely(imm32 > 63))
imm32            1644 arch/x86/net/bpf_jit_comp32.c 			emit_ia32_rsh_i64(dst, imm32, dstk, &prog);
imm32            1660 arch/x86/net/bpf_jit_comp32.c 			if (unlikely(imm32 > 63))
imm32            1662 arch/x86/net/bpf_jit_comp32.c 			emit_ia32_arsh_i64(dst, imm32, dstk, &prog);
imm32            1683 arch/x86/net/bpf_jit_comp32.c 				emit_ia32_mul_i64(dst, imm32, dstk, &prog);
imm32            1689 arch/x86/net/bpf_jit_comp32.c 			emit_ia32_to_le_r64(dst, imm32, dstk, &prog,
imm32            1694 arch/x86/net/bpf_jit_comp32.c 			emit_ia32_to_be_r64(dst, imm32, dstk, &prog,
imm32            1699 arch/x86/net/bpf_jit_comp32.c 			s32 hi, lo = imm32;
imm32            1736 arch/x86/net/bpf_jit_comp32.c 			EMIT(imm32, bpf_size_to_x86_bytes(BPF_SIZE(code)));
imm32            1741 arch/x86/net/bpf_jit_comp32.c 				hi = imm32 & (1<<31) ? (u32)~0 : 0;
imm32            1891 arch/x86/net/bpf_jit_comp32.c 			func = (u8 *) __bpf_call_base + imm32;
imm32            1894 arch/x86/net/bpf_jit_comp32.c 			if (!imm32 || !is_simm32(jmp_offset)) {
imm32            1896 arch/x86/net/bpf_jit_comp32.c 				       imm32, func, image);
imm32            2086 arch/x86/net/bpf_jit_comp32.c 			EMIT2_off32(0xC7, add_1reg(0xC0, sreg_lo), imm32);
imm32            2091 arch/x86/net/bpf_jit_comp32.c 				hi = imm32 & (1 << 31) ? (u32)~0 : 0;
imm32            2135 arch/x86/net/bpf_jit_comp32.c 			EMIT2_off32(0xC7, add_1reg(0xC0, IA32_ECX), imm32);
imm32            2137 arch/x86/net/bpf_jit_comp32.c 				hi = imm32 & (1 << 31) ? (u32)~0 : 0;
imm32            2181 arch/x86/net/bpf_jit_comp32.c 			EMIT2_off32(0xC7, add_1reg(0xC0, IA32_ECX), imm32);
imm32            2182 arch/x86/net/bpf_jit_comp32.c 			hi = imm32 & (1 << 31) ? (u32)~0 : 0;