imm32 311 arch/x86/net/bpf_jit_comp.c u32 dst_reg, const u32 imm32) imm32 321 arch/x86/net/bpf_jit_comp.c if (sign_propagate && (s32)imm32 < 0) { imm32 326 arch/x86/net/bpf_jit_comp.c EMIT3_off32(b1, b2, add_1reg(b3, dst_reg), imm32); imm32 334 arch/x86/net/bpf_jit_comp.c if (imm32 == 0) { imm32 346 arch/x86/net/bpf_jit_comp.c EMIT1_off32(add_1reg(0xB8, dst_reg), imm32); imm32 409 arch/x86/net/bpf_jit_comp.c const s32 imm32 = insn->imm; imm32 503 arch/x86/net/bpf_jit_comp.c if (is_imm8(imm32)) imm32 504 arch/x86/net/bpf_jit_comp.c EMIT3(0x83, add_1reg(b3, dst_reg), imm32); imm32 506 arch/x86/net/bpf_jit_comp.c EMIT1_off32(b2, imm32); imm32 508 arch/x86/net/bpf_jit_comp.c EMIT2_off32(0x81, add_1reg(b3, dst_reg), imm32); imm32 514 arch/x86/net/bpf_jit_comp.c dst_reg, imm32); imm32 540 arch/x86/net/bpf_jit_comp.c EMIT3_off32(0x49, 0xC7, 0xC3, imm32); imm32 590 arch/x86/net/bpf_jit_comp.c emit_mov_imm32(&prog, is64, BPF_REG_0, imm32); imm32 626 arch/x86/net/bpf_jit_comp.c if (imm32 == 1) imm32 629 arch/x86/net/bpf_jit_comp.c EMIT3(0xC1, add_1reg(b3, dst_reg), imm32); imm32 675 arch/x86/net/bpf_jit_comp.c switch (imm32) { imm32 707 arch/x86/net/bpf_jit_comp.c switch (imm32) { imm32 758 arch/x86/net/bpf_jit_comp.c EMIT(imm32, bpf_size_to_x86_bytes(BPF_SIZE(insn->code))); imm32 841 arch/x86/net/bpf_jit_comp.c func = (u8 *) __bpf_call_base + imm32; imm32 843 arch/x86/net/bpf_jit_comp.c if (!imm32 || !is_simm32(jmp_offset)) { imm32 845 arch/x86/net/bpf_jit_comp.c imm32, func, image); imm32 901 arch/x86/net/bpf_jit_comp.c EMIT2_off32(0xF7, add_1reg(0xC0, dst_reg), imm32); imm32 930 arch/x86/net/bpf_jit_comp.c if (is_imm8(imm32)) imm32 931 arch/x86/net/bpf_jit_comp.c EMIT3(0x83, add_1reg(0xF8, dst_reg), imm32); imm32 933 arch/x86/net/bpf_jit_comp.c EMIT2_off32(0x81, add_1reg(0xF8, dst_reg), imm32); imm32 1476 arch/x86/net/bpf_jit_comp32.c const s32 imm32 = insn->imm; imm32 1498 arch/x86/net/bpf_jit_comp32.c if (imm32 == 1) { imm32 1508 arch/x86/net/bpf_jit_comp32.c emit_ia32_mov_i64(is64, dst, imm32, imm32 1549 arch/x86/net/bpf_jit_comp32.c imm32, dstk, &prog, imm32 1564 arch/x86/net/bpf_jit_comp32.c imm32); imm32 1584 arch/x86/net/bpf_jit_comp32.c imm32); imm32 1607 arch/x86/net/bpf_jit_comp32.c imm32); imm32 1625 arch/x86/net/bpf_jit_comp32.c if (unlikely(imm32 > 31)) imm32 1628 arch/x86/net/bpf_jit_comp32.c EMIT2_off32(0xC7, add_1reg(0xC0, IA32_ECX), imm32); imm32 1636 arch/x86/net/bpf_jit_comp32.c if (unlikely(imm32 > 63)) imm32 1638 arch/x86/net/bpf_jit_comp32.c emit_ia32_lsh_i64(dst, imm32, dstk, &prog); imm32 1642 arch/x86/net/bpf_jit_comp32.c if (unlikely(imm32 > 63)) imm32 1644 arch/x86/net/bpf_jit_comp32.c emit_ia32_rsh_i64(dst, imm32, dstk, &prog); imm32 1660 arch/x86/net/bpf_jit_comp32.c if (unlikely(imm32 > 63)) imm32 1662 arch/x86/net/bpf_jit_comp32.c emit_ia32_arsh_i64(dst, imm32, dstk, &prog); imm32 1683 arch/x86/net/bpf_jit_comp32.c emit_ia32_mul_i64(dst, imm32, dstk, &prog); imm32 1689 arch/x86/net/bpf_jit_comp32.c emit_ia32_to_le_r64(dst, imm32, dstk, &prog, imm32 1694 arch/x86/net/bpf_jit_comp32.c emit_ia32_to_be_r64(dst, imm32, dstk, &prog, imm32 1699 arch/x86/net/bpf_jit_comp32.c s32 hi, lo = imm32; imm32 1736 arch/x86/net/bpf_jit_comp32.c EMIT(imm32, bpf_size_to_x86_bytes(BPF_SIZE(code))); imm32 1741 arch/x86/net/bpf_jit_comp32.c hi = imm32 & (1<<31) ? (u32)~0 : 0; imm32 1891 arch/x86/net/bpf_jit_comp32.c func = (u8 *) __bpf_call_base + imm32; imm32 1894 arch/x86/net/bpf_jit_comp32.c if (!imm32 || !is_simm32(jmp_offset)) { imm32 1896 arch/x86/net/bpf_jit_comp32.c imm32, func, image); imm32 2086 arch/x86/net/bpf_jit_comp32.c EMIT2_off32(0xC7, add_1reg(0xC0, sreg_lo), imm32); imm32 2091 arch/x86/net/bpf_jit_comp32.c hi = imm32 & (1 << 31) ? (u32)~0 : 0; imm32 2135 arch/x86/net/bpf_jit_comp32.c EMIT2_off32(0xC7, add_1reg(0xC0, IA32_ECX), imm32); imm32 2137 arch/x86/net/bpf_jit_comp32.c hi = imm32 & (1 << 31) ? (u32)~0 : 0; imm32 2181 arch/x86/net/bpf_jit_comp32.c EMIT2_off32(0xC7, add_1reg(0xC0, IA32_ECX), imm32); imm32 2182 arch/x86/net/bpf_jit_comp32.c hi = imm32 & (1 << 31) ? (u32)~0 : 0;