imm64             671 arch/arm64/net/bpf_jit_comp.c 		u64 imm64;
imm64             673 arch/arm64/net/bpf_jit_comp.c 		imm64 = (u64)insn1.imm << 32 | (u32)imm;
imm64             674 arch/arm64/net/bpf_jit_comp.c 		emit_a64_mov_i64(dst, imm64, ctx);
imm64             311 arch/powerpc/net/bpf_jit_comp64.c 		u64 imm64;
imm64             741 arch/powerpc/net/bpf_jit_comp64.c 			imm64 = ((u64)(u32) insn[i].imm) |
imm64             745 arch/powerpc/net/bpf_jit_comp64.c 			PPC_LI64(dst_reg, imm64);
imm64            1261 arch/riscv/net/bpf_jit_comp.c 		u64 imm64;
imm64            1263 arch/riscv/net/bpf_jit_comp.c 		imm64 = (u64)insn1.imm << 32 | (u32)imm;
imm64            1264 arch/riscv/net/bpf_jit_comp.c 		emit_imm(rd, imm64, ctx);
imm64             550 arch/s390/net/bpf_jit_comp.c 		u64 imm64;
imm64             552 arch/s390/net/bpf_jit_comp.c 		imm64 = (u64)(u32) insn[0].imm | ((u64)(u32) insn[1].imm) << 32;
imm64             555 arch/s390/net/bpf_jit_comp.c 			      EMIT_CONST_U64(imm64));
imm64            1246 arch/sparc/net/bpf_jit_comp_64.c 		u64 imm64;
imm64            1248 arch/sparc/net/bpf_jit_comp_64.c 		imm64 = (u64)insn1.imm << 32 | (u32)imm;
imm64            1249 arch/sparc/net/bpf_jit_comp_64.c 		emit_loadimm64(imm64, dst, ctx);