/linux-4.4.14/arch/x86/purgatory/ |
D | entry64.S | 21 lgdt gdt(%rip) 32 leaq stack_init(%rip), %rsp 34 leaq new_cs_exit(%rip), %rax 40 movq rax(%rip), %rax 41 movq rbx(%rip), %rbx 42 movq rcx(%rip), %rcx 43 movq rdx(%rip), %rdx 44 movq rsi(%rip), %rsi 45 movq rdi(%rip), %rdi 46 movq rsp(%rip), %rsp [all …]
|
D | setup-x86_64.S | 20 lgdt gdt(%rip) 31 leaq lstack_end(%rip), %rsp
|
/linux-4.4.14/arch/x86/kernel/ |
D | head_64.S | 75 leaq _text(%rip), %rbp 87 leaq _text(%rip), %rax 94 addq %rbp, early_level4_pgt + (L4_START_KERNEL*8)(%rip) 96 addq %rbp, level3_kernel_pgt + (510*8)(%rip) 97 addq %rbp, level3_kernel_pgt + (511*8)(%rip) 99 addq %rbp, level2_fixmap_pgt + (506*8)(%rip) 107 leaq _text(%rip), %rdi 108 leaq early_level4_pgt(%rip), %rbx 130 leaq (_end - 1)(%rip), %rcx 149 leaq level2_kernel_pgt(%rip), %rdi [all …]
|
D | mcount_64.S | 160 movq function_trace_op(%rip), %rdx 197 movq function_trace_op(%rip), %rdx
|
D | kexec-bzimage64.c | 475 regs64.rip = kernel_load_addr + 0x200; in bzImage64_load()
|
/linux-4.4.14/arch/x86/platform/efi/ |
D | efi_stub_64.S | 43 movq %r15, efi_scratch(%rip) 44 movq %r14, efi_scratch+8(%rip) 50 movq efi_scratch+8(%rip), %r14 51 movq efi_scratch(%rip), %r15 55 cmpb $0, efi_scratch+24(%rip) 57 movq %r15, efi_scratch(%rip) # r15 60 movq %r15, efi_scratch+8(%rip) # prev_cr3 61 movq efi_scratch+16(%rip), %r15 # EFI pgt 67 cmpb $0, efi_scratch+24(%rip) 69 movq efi_scratch+8(%rip), %r15 [all …]
|
D | efi_thunk_64.S | 34 movq %rsp, efi_saved_sp(%rip) 35 movq efi_scratch+25(%rip), %rsp 41 subq phys_base(%rip), %rax 50 leaq efi_exit32(%rip), %rbx 54 leaq __efi64_thunk(%rip), %rbx 58 movq efi_saved_sp(%rip), %rsp 86 leaq 1f(%rip), %rbx 87 movq %rbx, func_rt_ptr(%rip) 91 leaq efi_enter32(%rip), %rax 119 movq func_rt_ptr(%rip), %rax
|
/linux-4.4.14/arch/x86/boot/compressed/ |
D | head_64.S | 245 movq %rcx, efi64_config(%rip) /* Handle */ 246 movq %rdx, efi64_config+8(%rip) /* EFI System table pointer */ 248 leaq efi64_config(%rip), %rax 249 movq %rax, efi_config(%rip) 258 addq %rbp, efi64_config+88(%rip) 265 leaq startup_32(%rip), %rax 277 movq efi_config(%rip), %rax 280 movq efi_config(%rip), %rdi 320 leaq startup_32(%rip) /* - $startup_32 */, %rbp 347 leaq (_bss-8)(%rip), %rsi [all …]
|
D | efi_thunk_64.S | 30 leaq efi_exit32(%rip), %rax 32 leaq efi_gdt64(%rip), %rax 55 sgdt save_gdt(%rip) 57 leaq 1f(%rip), %rbx 58 movq %rbx, func_rt_ptr(%rip) 65 leaq efi32_boot_gdt(%rip), %rax 69 leaq efi_enter32(%rip), %rax 75 lgdt save_gdt(%rip) 102 movq func_rt_ptr(%rip), %rax
|
/linux-4.4.14/arch/x86/power/ |
D | hibernate_asm_64.S | 47 movq %rax, restore_jump_address(%rip) 50 movq %rax, restore_cr3(%rip) 58 movq temp_level4_pgt(%rip), %rax 62 movq mmu_cr4_features(%rip), %rax 71 movq restore_jump_address(%rip), %rax 72 movq restore_cr3(%rip), %rbx 75 movq restore_pblist(%rip), %rdx 76 movq relocated_restore_code(%rip), %rcx 114 movq mmu_cr4_features(%rip), %rax 148 movq %rax, in_suspend(%rip)
|
/linux-4.4.14/arch/x86/crypto/ |
D | aesni-intel_avx-x86_64.S | 392 vpshufb SHUF_MASK(%rip), reg_i, reg_i 400 vpshufb SHUF_MASK(%rip), \CTR, \CTR 406 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0 408 vpshufb SHUF_MASK(%rip), reg_i, reg_i # perform a 16Byte swap 458 vpshufb SHUF_MASK(%rip), reg_i, reg_i # prepare ciphertext for GHASH computations 486 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0 488 vpshufb SHUF_MASK(%rip), \XMM1, \XMM1 # perform a 16Byte swap 490 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0 492 vpshufb SHUF_MASK(%rip), \XMM2, \XMM2 # perform a 16Byte swap 494 vpaddd ONE(%rip), \CTR, \CTR # INCR Y0 [all …]
|
D | crc32-pclmul_asm.S | 124 movdqa .Lconstant_R2R1(%rip), CONSTANT 173 movdqa .Lconstant_R4R3(%rip), CONSTANT 220 movdqa .Lconstant_R5(%rip), CONSTANT 221 movdqa .Lconstant_mask32(%rip), %xmm3 233 movdqa .Lconstant_RUpoly(%rip), CONSTANT
|
D | aes_ctrby8_avx-x86_64.S | 173 vpaddq var_ddq_add(%rip), xcounter, var_xdata 174 vptest ddq_low_msk(%rip), var_xdata 176 vpaddq ddq_high_add_1(%rip), var_xdata, var_xdata 177 vpaddq ddq_high_add_1(%rip), xcounter, xcounter 187 vpaddq var_ddq_add(%rip), xcounter, xcounter 188 vptest ddq_low_msk(%rip), xcounter 190 vpaddq ddq_high_add_1(%rip), xcounter, xcounter 443 vmovdqa byteswap_const(%rip), xbyteswap
|
D | aesni-intel_asm.S | 228 MOVADQ SHUF_MASK(%rip), %xmm14 330 MOVADQ ONE(%rip), \TMP1 453 MOVADQ SHUF_MASK(%rip), %xmm14 670 movdqa SHUF_MASK(%rip), %xmm15 676 paddd ONE(%rip), \XMM0 # INCR CNT 680 paddd ONE(%rip), \XMM0 # INCR CNT 682 paddd ONE(%rip), \XMM0 # INCR CNT 684 paddd ONE(%rip), \XMM0 # INCR CNT 878 movdqa SHUF_MASK(%rip), %xmm15 884 paddd ONE(%rip), \XMM0 # INCR CNT [all …]
|
D | poly1305-avx2-x86_64.S | 163 vpand ANMASK(%rip),hc0,hc0 176 vpand ANMASK(%rip),hc1,hc1 189 vpand ANMASK(%rip),hc2,hc2 202 vpand ANMASK(%rip),hc3,hc3 215 vpor ORMASK(%rip),hc4,hc4
|
D | poly1305-sse2-x86_64.S | 79 movdqa ANMASK(%rip),mask 360 pand ANMASK(%rip),hc0 368 pand ANMASK(%rip),hc1 376 pand ANMASK(%rip),hc2 384 pand ANMASK(%rip),hc3 392 por ORMASK(%rip),hc4
|
D | sha1_ni_asm.S | 110 pand UPPER_WORD_MASK(%rip), E0 113 movdqa PSHUFFLE_BYTE_FLIP_MASK(%rip), SHUF_MASK
|
D | sha256_ni_asm.S | 120 movdqa PSHUFFLE_BYTE_FLIP_MASK(%rip), SHUF_MASK 121 lea K256(%rip), SHA256CONSTANTS
|
D | sha256-avx2-asm.S | 564 vmovdqa PSHUFFLE_BYTE_FLIP_MASK(%rip), BYTE_FLIP_MASK 565 vmovdqa _SHUF_00BA(%rip), SHUF_00BA 566 vmovdqa _SHUF_DC00(%rip), SHUF_DC00 571 lea K256(%rip), TBL 680 lea K256(%rip), TBL 706 vmovdqa PSHUFFLE_BYTE_FLIP_MASK(%rip), BYTE_FLIP_MASK 707 vmovdqa _SHUF_00BA(%rip), SHUF_00BA 708 vmovdqa _SHUF_DC00(%rip), SHUF_DC00
|
D | chacha20-ssse3-x86_64.S | 45 movdqa ROT8(%rip),%xmm4 46 movdqa ROT16(%rip),%xmm5 195 movdqa CTRINC(%rip),%xmm1 196 movdqa ROT8(%rip),%xmm2 197 movdqa ROT16(%rip),%xmm3
|
D | sha256-ssse3-asm.S | 384 movdqa PSHUFFLE_BYTE_FLIP_MASK(%rip), BYTE_FLIP_MASK 385 movdqa _SHUF_00BA(%rip), SHUF_00BA 386 movdqa _SHUF_DC00(%rip), SHUF_DC00 389 lea K256(%rip), TBL
|
D | sha256-avx-asm.S | 378 vmovdqa PSHUFFLE_BYTE_FLIP_MASK(%rip), BYTE_FLIP_MASK 379 vmovdqa _SHUF_00BA(%rip), SHUF_00BA 380 vmovdqa _SHUF_DC00(%rip), SHUF_DC00 382 lea K256(%rip), TBL
|
D | chacha20-avx2-x86_64.S | 70 vmovdqa CTRINC(%rip),%ymm1 71 vmovdqa ROT8(%rip),%ymm2 72 vmovdqa ROT16(%rip),%ymm3
|
D | sha1_avx2_x86_64_asm.S | 644 lea K_XMM_AR(%rip), K_BASE 658 xmm_mov BSWAP_SHUFB_CTL(%rip), YMM_SHUFB_BSWAP
|
D | crc32c-pcl-intel-asm_64.S | 220 lea (K_table-8)(%rip), bufp # first entry is for idx 1
|
D | sha1_ssse3_asm.S | 91 lea K_XMM_AR(%rip), K_BASE 92 xmm_mov BSWAP_SHUFB_CTL(%rip), XMM_SHUFB_BSWAP
|
D | sha512-ssse3-asm.S | 97 #define K_t(i) 8*i+K512(%rip) 314 movdqa XMM_QWORD_BSWAP(%rip), %xmm1
|
D | sha512-avx-asm.S | 100 #define K_t(i) 8*i+K512(%rip) 315 vmovdqa XMM_QWORD_BSWAP(%rip), %xmm1
|
D | sha512-avx2-asm.S | 238 vpand MASK_YMM_LO(%rip), YTMP0, YTMP0 # YTMP0 = W[-16] + W[-7] + s0 {DC00} 601 vmovdqa PSHUFFLE_BYTE_FLIP_MASK(%rip), BYTE_FLIP_MASK 604 lea K512(%rip), TBL
|
/linux-4.4.14/arch/x86/kvm/ |
D | trace.h | 498 TP_PROTO(__u64 rip, __u64 vmcb, __u64 nested_rip, __u32 int_ctl, 500 TP_ARGS(rip, vmcb, nested_rip, int_ctl, event_inj, npt), 503 __field( __u64, rip ) 512 __entry->rip = rip; 522 __entry->rip, __entry->vmcb, __entry->nested_rip, 553 TP_PROTO(__u64 rip, __u32 exit_code, 556 TP_ARGS(rip, exit_code, exit_info1, exit_info2, 560 __field( __u64, rip ) 570 __entry->rip = rip; 580 __entry->rip, [all …]
|
D | svm.c | 550 unsigned long rip, old_rip = kvm_rip_read(&svm->vcpu); in svm_queue_exception() local 560 rip = kvm_rip_read(&svm->vcpu); in svm_queue_exception() 561 svm->int3_rip = rip + svm->vmcb->save.cs.base; in svm_queue_exception() 562 svm->int3_injected = rip - old_rip; in svm_queue_exception() 1077 save->rip = 0x0000fff0; in init_vmcb() 1078 svm->vcpu.arch.regs[VCPU_REGS_RIP] = save->rip; in init_vmcb() 1676 svm->vmcb->save.cs.base + svm->vmcb->save.rip; in db_interception() 1689 kvm_run->debug.arch.pc = svm->vmcb->save.cs.base + svm->vmcb->save.rip; in bp_interception() 2006 trace_kvm_nested_intr_vmexit(svm->vmcb->save.rip); in nested_svm_intr() 2263 nested_vmcb->save.rip = vmcb->save.rip; in nested_svm_vmexit() [all …]
|
D | x86.c | 5938 unsigned long rip = kvm_rip_read(vcpu); in emulator_fix_hypercall() local 5942 return emulator_write_emulated(ctxt, rip, instruction, 3, NULL); in emulator_fix_hypercall() 6600 unsigned long rip = kvm_rip_read(vcpu); in vcpu_enter_guest() local 6601 profile_hit(KVM_PROFILING, (void *)rip); in vcpu_enter_guest() 6869 regs->rip = kvm_rip_read(vcpu); in kvm_arch_vcpu_ioctl_get_regs() 6899 kvm_rip_write(vcpu, regs->rip); in kvm_arch_vcpu_ioctl_set_regs() 7175 fpu->last_ip = fxsave->rip; in kvm_arch_vcpu_ioctl_get_fpu() 7192 fxsave->rip = fpu->last_ip; in kvm_arch_vcpu_ioctl_set_fpu()
|
D | vmx.c | 2245 unsigned long rip; in skip_emulated_instruction() local 2247 rip = kvm_rip_read(vcpu); in skip_emulated_instruction() 2248 rip += vmcs_read32(VM_EXIT_INSTRUCTION_LEN); in skip_emulated_instruction() 2249 kvm_rip_write(vcpu, rip); in skip_emulated_instruction() 5226 unsigned long cr2, rip, dr6; in handle_exception() local 5319 rip = kvm_rip_read(vcpu); in handle_exception() 5320 kvm_run->debug.arch.pc = vmcs_readl(GUEST_CS_BASE) + rip; in handle_exception()
|
D | lapic.c | 1058 run->tpr_access.rip = kvm_rip_read(vcpu); in __report_tpr_access()
|
/linux-4.4.14/arch/x86/crypto/sha-mb/ |
D | sha1_mb_mgr_flush_avx2.S | 136 cmovne one(%rip), idx 139 cmovne two(%rip), idx 142 cmovne three(%rip), idx 145 cmovne four(%rip), idx 148 cmovne five(%rip), idx 151 cmovne six(%rip), idx 154 cmovne seven(%rip), idx 192 vpand clear_low_nibble(%rip), %xmm2, %xmm2
|
D | sha1_x8_avx2.S | 327 vmovdqu PSHUFFLE_BYTE_FLIP_MASK(%rip), F 369 vmovdqu K00_19(%rip), K 388 vmovdqu K20_39(%rip), K 396 vmovdqu K40_59(%rip), K 404 vmovdqu K60_79(%rip), K
|
D | sha1_mb_mgr_submit_avx2.S | 165 vpand clear_low_nibble(%rip), %xmm2, %xmm2
|
/linux-4.4.14/arch/ia64/mm/ |
D | extable.c | 17 u64 rip = (u64) &r->addr + r->addr; in cmp_ex() local 20 if (lip > rip) in cmp_ex() 22 if (lip < rip) in cmp_ex()
|
/linux-4.4.14/tools/lib/traceevent/ |
D | plugin_kvm.c | 39 static const char *disassemble(unsigned char *insn, int len, uint64_t rip, in disassemble() argument 56 ud_set_pc(&ud, rip); in disassemble() 69 static const char *disassemble(unsigned char *insn, int len, uint64_t rip, in disassemble() argument 297 unsigned long long rip, csbase, len, flags, failed; in kvm_emulate_insn_handler() local 302 if (pevent_get_field_val(s, event, "rip", record, &rip, 1) < 0) in kvm_emulate_insn_handler() 321 disasm = disassemble(insn, len, rip, in kvm_emulate_insn_handler() 327 trace_seq_printf(s, "%llx:%llx: %s%s", csbase, rip, disasm, in kvm_emulate_insn_handler()
|
/linux-4.4.14/fs/xfs/ |
D | xfs_mount.c | 631 struct xfs_inode *rip; in xfs_mountfs() local 860 error = xfs_iget(mp, NULL, sbp->sb_rootino, 0, XFS_ILOCK_EXCL, &rip); in xfs_mountfs() 866 ASSERT(rip != NULL); in xfs_mountfs() 868 if (unlikely(!S_ISDIR(rip->i_d.di_mode))) { in xfs_mountfs() 870 (unsigned long long)rip->i_ino); in xfs_mountfs() 871 xfs_iunlock(rip, XFS_ILOCK_EXCL); in xfs_mountfs() 877 mp->m_rootip = rip; /* save it */ in xfs_mountfs() 879 xfs_iunlock(rip, XFS_ILOCK_EXCL); in xfs_mountfs() 974 IRELE(rip); in xfs_mountfs()
|
/linux-4.4.14/arch/x86/include/uapi/asm/ |
D | ptrace.h | 70 unsigned long rip; member
|
D | sigcontext.h | 154 __u64 rip; member 341 __u64 rip; member
|
D | kvm.h | 118 __u64 rip, rflags; member
|
/linux-4.4.14/arch/x86/include/asm/ |
D | user_64.h | 56 __u64 rip; member
|
D | kexec.h | 207 uint64_t rip; member
|
D | svm.h | 160 u64 rip; member
|
D | paravirt.h | 897 #define PARA_INDIRECT(addr) *addr(%rip)
|
/linux-4.4.14/arch/x86/realmode/rm/ |
D | trampoline_64.S | 125 jmpq *tr_start(%rip)
|
/linux-4.4.14/arch/x86/include/asm/xen/ |
D | interface_64.h | 83 uint64_t rax, r11, rcx, flags, rip, cs, rflags, rsp, ss; member
|
/linux-4.4.14/arch/x86/kernel/fpu/ |
D | regset.c | 217 env->fip = fxsave->rip; in convert_from_fxsr() 255 fxsave->rip = env->fip; in convert_to_fxsr()
|
D | xstate.c | 136 fx->rip = 0; in fpstate_sanitize_xstate()
|
/linux-4.4.14/arch/x86/include/asm/fpu/ |
D | types.h | 40 u64 rip; /* Instruction Pointer */ member
|
/linux-4.4.14/tools/perf/util/ |
D | map.c | 426 u64 map__rip_2objdump(struct map *map, u64 rip) in map__rip_2objdump() argument 429 return rip; in map__rip_2objdump() 432 return rip - map->pgoff; in map__rip_2objdump() 434 return map->unmap_ip(map, rip) - map->reloc; in map__rip_2objdump()
|
D | map.h | 106 u64 map__rip_2objdump(struct map *map, u64 rip);
|
/linux-4.4.14/arch/x86/kernel/acpi/ |
D | wakeup_64.S | 65 movq $.Lresume_point, saved_rip(%rip)
|
/linux-4.4.14/Documentation/prctl/ |
D | seccomp_filter.txt | 207 The rip value seen by the tracer will be the syscall entry address; 209 rip or rsp. (Do not rely on other changes terminating the process. 215 0xFFFFFFFFFF600000. (For SECCOMP_RET_TRACE, use rip. For
|
/linux-4.4.14/tools/testing/selftests/x86/ |
D | ptrace_syscall.c | 26 # define user_ip rip
|
/linux-4.4.14/drivers/net/ethernet/alteon/ |
D | acenic.c | 1930 struct ring_info *rip; in ace_rx_int() local 1956 rip = &ap->skb->rx_std_skbuff[skbidx]; in ace_rx_int() 1962 rip = &ap->skb->rx_jumbo_skbuff[skbidx]; in ace_rx_int() 1968 rip = &ap->skb->rx_mini_skbuff[skbidx]; in ace_rx_int() 1980 skb = rip->skb; in ace_rx_int() 1981 rip->skb = NULL; in ace_rx_int() 1983 dma_unmap_addr(rip, mapping), in ace_rx_int()
|
/linux-4.4.14/arch/x86/entry/ |
D | entry_64.S | 992 idtentry machine_check has_error_code=0 paranoid=1 do_sym=*machine_check_vector(%rip) 1086 leaq native_irq_return_iret(%rip), %rcx
|
/linux-4.4.14/Documentation/ |
D | static-keys.txt | 214 ffffffff810441f0: 8b 05 8a 52 d8 00 mov 0xd8528a(%rip),%eax # ffffffff81dc948…
|
D | devices.txt | 756 /dev/inet/rip -> /dev/inet/udp
|
/linux-4.4.14/include/uapi/linux/ |
D | kvm.h | 262 __u64 rip; member
|
/linux-4.4.14/include/linux/ |
D | mm.h | 2191 void print_vma_addr(char *prefix, unsigned long rip);
|
/linux-4.4.14/Documentation/networking/ |
D | filter.txt | 688 movq _f2(%rip), %rax
|
/linux-4.4.14/Documentation/virtual/kvm/ |
D | api.txt | 308 __u64 rip, rflags; 3177 __u64 rip;
|
/linux-4.4.14/drivers/scsi/aic7xxx/ |
D | aic7xxx.seq | 1365 * Command phase. Set up the DMA registers and let 'er rip.
|
D | aic79xx.seq | 865 * Command phase. Set up the DMA registers and let 'er rip.
|