magic_page_pa     765 arch/powerpc/include/asm/kvm_host.h 	unsigned long magic_page_pa; /* phys addr to map the magic page to */
magic_page_pa     429 arch/powerpc/kvm/book3s.c 	ulong mp_pa = vcpu->arch.magic_page_pa & KVM_PAM;
magic_page_pa     310 arch/powerpc/kvm/book3s_32_mmu.c 		pte->raddr = vcpu->arch.magic_page_pa | (pte->raddr & 0xfff);
magic_page_pa     229 arch/powerpc/kvm/book3s_64_mmu.c 		gpte->raddr = vcpu->arch.magic_page_pa | (gpte->raddr & 0xfff);
magic_page_pa     501 arch/powerpc/kvm/book3s_pr.c 		if (!(msr & MSR_PR) && vcpu->arch.magic_page_pa) {
magic_page_pa     507 arch/powerpc/kvm/book3s_pr.c 				kvmppc_mmu_map_segment(vcpu, a->magic_page_pa);
magic_page_pa     519 arch/powerpc/kvm/book3s_pr.c 	if (vcpu->arch.magic_page_pa &&
magic_page_pa     522 arch/powerpc/kvm/book3s_pr.c 		kvmppc_mmu_pte_flush(vcpu, (uint32_t)vcpu->arch.magic_page_pa,
magic_page_pa     654 arch/powerpc/kvm/book3s_pr.c 	ulong mp_pa = vcpu->arch.magic_page_pa;
magic_page_pa    1973 arch/powerpc/kvm/booke.c 		pte->raddr = (vcpu->arch.magic_page_pa & PAGE_MASK) |
magic_page_pa     201 arch/powerpc/kvm/powerpc.c 		vcpu->arch.magic_page_pa = param1 & ~0xfffULL;
magic_page_pa     209 arch/powerpc/kvm/powerpc.c 		if ((vcpu->arch.magic_page_pa & 0xf000) !=
magic_page_pa     216 arch/powerpc/kvm/powerpc.c 			shared |= vcpu->arch.magic_page_pa & 0xf000;
magic_page_pa     326 arch/powerpc/kvm/powerpc.c 	ulong mp_pa = vcpu->arch.magic_page_pa & KVM_PAM & PAGE_MASK;
magic_page_pa     369 arch/powerpc/kvm/powerpc.c 	ulong mp_pa = vcpu->arch.magic_page_pa & KVM_PAM & PAGE_MASK;