imm64 671 arch/arm64/net/bpf_jit_comp.c u64 imm64; imm64 673 arch/arm64/net/bpf_jit_comp.c imm64 = (u64)insn1.imm << 32 | (u32)imm; imm64 674 arch/arm64/net/bpf_jit_comp.c emit_a64_mov_i64(dst, imm64, ctx); imm64 311 arch/powerpc/net/bpf_jit_comp64.c u64 imm64; imm64 741 arch/powerpc/net/bpf_jit_comp64.c imm64 = ((u64)(u32) insn[i].imm) | imm64 745 arch/powerpc/net/bpf_jit_comp64.c PPC_LI64(dst_reg, imm64); imm64 1261 arch/riscv/net/bpf_jit_comp.c u64 imm64; imm64 1263 arch/riscv/net/bpf_jit_comp.c imm64 = (u64)insn1.imm << 32 | (u32)imm; imm64 1264 arch/riscv/net/bpf_jit_comp.c emit_imm(rd, imm64, ctx); imm64 550 arch/s390/net/bpf_jit_comp.c u64 imm64; imm64 552 arch/s390/net/bpf_jit_comp.c imm64 = (u64)(u32) insn[0].imm | ((u64)(u32) insn[1].imm) << 32; imm64 555 arch/s390/net/bpf_jit_comp.c EMIT_CONST_U64(imm64)); imm64 1246 arch/sparc/net/bpf_jit_comp_64.c u64 imm64; imm64 1248 arch/sparc/net/bpf_jit_comp_64.c imm64 = (u64)insn1.imm << 32 | (u32)imm; imm64 1249 arch/sparc/net/bpf_jit_comp_64.c emit_loadimm64(imm64, dst, ctx);