MSR_DR 138 arch/powerpc/include/asm/reg.h #define __MSR (MSR_ME | MSR_RI | MSR_IR | MSR_DR | MSR_ISF |MSR_HV) MSR_DR 151 arch/powerpc/include/asm/reg.h #define MSR_KERNEL (MSR_ME|MSR_RI|MSR_IR|MSR_DR) MSR_DR 46 arch/powerpc/include/asm/reg_booke.h #define MSR_KERNEL (MSR_ME|MSR_RI|MSR_IR|MSR_DR|MSR_CE) MSR_DR 57 arch/powerpc/kernel/btext.c if (!(mfmsr() & MSR_DR)) MSR_DR 65 arch/powerpc/kernel/btext.c if (!(mfmsr() & MSR_DR)) MSR_DR 52 arch/powerpc/kernel/head_32.h li r10,MSR_KERNEL & ~(MSR_IR|MSR_DR) /* can take exceptions */ MSR_DR 82 arch/powerpc/kernel/head_32.h LOAD_REG_IMMEDIATE(r10, MSR_KERNEL & ~(MSR_IR|MSR_DR)) /* can take exceptions */ MSR_DR 267 arch/powerpc/kernel/kprobes.c if (!(regs->msr & MSR_IR) || !(regs->msr & MSR_DR)) MSR_DR 192 arch/powerpc/kernel/paca.c new_paca->kernel_msr = MSR_KERNEL & ~(MSR_IR | MSR_DR); MSR_DR 1323 arch/powerpc/kernel/process.c {MSR_DR, "DR"}, MSR_DR 457 arch/powerpc/kvm/book3s.c int relocated = (kvmppc_get_msr(vcpu) & (data ? MSR_DR : MSR_IR)); MSR_DR 471 arch/powerpc/kvm/book3s.c if ((kvmppc_get_msr(vcpu) & (MSR_IR | MSR_DR)) == MSR_DR && MSR_DR 361 arch/powerpc/kvm/book3s_32_mmu.c if (msr & (MSR_DR|MSR_IR)) { MSR_DR 370 arch/powerpc/kvm/book3s_32_mmu.c switch (msr & (MSR_DR|MSR_IR)) { MSR_DR 377 arch/powerpc/kvm/book3s_32_mmu.c case MSR_DR: MSR_DR 380 arch/powerpc/kvm/book3s_32_mmu.c case MSR_DR|MSR_IR: MSR_DR 597 arch/powerpc/kvm/book3s_64_mmu.c if (msr & (MSR_DR|MSR_IR)) { MSR_DR 610 arch/powerpc/kvm/book3s_64_mmu.c switch (msr & (MSR_DR|MSR_IR)) { MSR_DR 617 arch/powerpc/kvm/book3s_64_mmu.c case MSR_DR: MSR_DR 620 arch/powerpc/kvm/book3s_64_mmu.c case MSR_DR|MSR_IR: MSR_DR 350 arch/powerpc/kvm/book3s_64_mmu_hv.c int virtmode = vcpu->arch.shregs.msr & (data ? MSR_DR : MSR_IR); MSR_DR 570 arch/powerpc/kvm/book3s_hv_builtin.c return !(mfmsr() & MSR_DR); MSR_DR 1276 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (data && (vcpu->arch.shregs.msr & MSR_DR)) { MSR_DR 69 arch/powerpc/kvm/book3s_pr.c return (msr & (MSR_IR|MSR_DR)) == MSR_DR; MSR_DR 78 arch/powerpc/kvm/book3s_pr.c if ((msr & (MSR_IR|MSR_DR)) != MSR_DR) MSR_DR 201 arch/powerpc/kvm/book3s_pr.c smsr |= MSR_ME | MSR_RI | MSR_IR | MSR_DR | MSR_PR | MSR_EE; MSR_DR 495 arch/powerpc/kvm/book3s_pr.c if ((kvmppc_get_msr(vcpu) & (MSR_PR|MSR_IR|MSR_DR)) != MSR_DR 496 arch/powerpc/kvm/book3s_pr.c (old_msr & (MSR_PR|MSR_IR|MSR_DR))) { MSR_DR 504 arch/powerpc/kvm/book3s_pr.c if (msr & MSR_DR) MSR_DR 676 arch/powerpc/kvm/book3s_pr.c bool dr = (kvmppc_get_msr(vcpu) & MSR_DR) ? true : false; MSR_DR 698 arch/powerpc/kvm/book3s_pr.c switch (kvmppc_get_msr(vcpu) & (MSR_DR|MSR_IR)) { MSR_DR 702 arch/powerpc/kvm/book3s_pr.c case MSR_DR: MSR_DR 711 arch/powerpc/kvm/book3s_pr.c if ((kvmppc_get_msr(vcpu) & (MSR_DR|MSR_IR)) == MSR_DR) MSR_DR 441 arch/powerpc/mm/mem.c msr = msr0 & ~MSR_DR; MSR_DR 27 arch/powerpc/platforms/82xx/pq2.c mtmsr(mfmsr() & ~(MSR_ME | MSR_EE | MSR_IR | MSR_DR)); MSR_DR 389 arch/powerpc/platforms/powernv/idle.c WARN_ON_ONCE(mfmsr() & (MSR_IR|MSR_DR)); MSR_DR 700 arch/powerpc/platforms/powernv/idle.c WARN_ON_ONCE(mfmsr() & (MSR_IR|MSR_DR)); MSR_DR 239 arch/powerpc/platforms/powernv/npu-dma.c pnv_npu2_map_lpar_dev(gpdev, 0, MSR_DR | MSR_PR | MSR_HV); MSR_DR 100 arch/powerpc/platforms/powernv/opal-call.c bool mmu = (msr & (MSR_IR|MSR_DR)); MSR_DR 1296 arch/powerpc/platforms/powernv/pci-ioda.c pnv_npu2_map_lpar(pe, MSR_DR | MSR_PR | MSR_HV); MSR_DR 476 arch/powerpc/platforms/pseries/ras.c (MSR_LE|MSR_RI|MSR_DR|MSR_IR|MSR_ME|MSR_PR| MSR_DR 477 arch/powerpc/platforms/pseries/ras.c MSR_ILE|MSR_HV|MSR_SF)) == (MSR_DR|MSR_SF)) { MSR_DR 696 arch/powerpc/platforms/pseries/ras.c mtmsr(mfmsr() | MSR_IR | MSR_DR); MSR_DR 208 drivers/misc/cxl/cxl.h #define CXL_PSL_SR_An_R MSR_DR /* Relocate, GA1: 1 */ MSR_DR 204 drivers/vfio/pci/vfio_pci_nvlink2.c kvm->arch.lpid, MSR_DR | MSR_PR))