magic_page_pa 765 arch/powerpc/include/asm/kvm_host.h unsigned long magic_page_pa; /* phys addr to map the magic page to */ magic_page_pa 429 arch/powerpc/kvm/book3s.c ulong mp_pa = vcpu->arch.magic_page_pa & KVM_PAM; magic_page_pa 310 arch/powerpc/kvm/book3s_32_mmu.c pte->raddr = vcpu->arch.magic_page_pa | (pte->raddr & 0xfff); magic_page_pa 229 arch/powerpc/kvm/book3s_64_mmu.c gpte->raddr = vcpu->arch.magic_page_pa | (gpte->raddr & 0xfff); magic_page_pa 501 arch/powerpc/kvm/book3s_pr.c if (!(msr & MSR_PR) && vcpu->arch.magic_page_pa) { magic_page_pa 507 arch/powerpc/kvm/book3s_pr.c kvmppc_mmu_map_segment(vcpu, a->magic_page_pa); magic_page_pa 519 arch/powerpc/kvm/book3s_pr.c if (vcpu->arch.magic_page_pa && magic_page_pa 522 arch/powerpc/kvm/book3s_pr.c kvmppc_mmu_pte_flush(vcpu, (uint32_t)vcpu->arch.magic_page_pa, magic_page_pa 654 arch/powerpc/kvm/book3s_pr.c ulong mp_pa = vcpu->arch.magic_page_pa; magic_page_pa 1973 arch/powerpc/kvm/booke.c pte->raddr = (vcpu->arch.magic_page_pa & PAGE_MASK) | magic_page_pa 201 arch/powerpc/kvm/powerpc.c vcpu->arch.magic_page_pa = param1 & ~0xfffULL; magic_page_pa 209 arch/powerpc/kvm/powerpc.c if ((vcpu->arch.magic_page_pa & 0xf000) != magic_page_pa 216 arch/powerpc/kvm/powerpc.c shared |= vcpu->arch.magic_page_pa & 0xf000; magic_page_pa 326 arch/powerpc/kvm/powerpc.c ulong mp_pa = vcpu->arch.magic_page_pa & KVM_PAM & PAGE_MASK; magic_page_pa 369 arch/powerpc/kvm/powerpc.c ulong mp_pa = vcpu->arch.magic_page_pa & KVM_PAM & PAGE_MASK;