gpa_mm 186 arch/mips/include/asm/kvm_host.h struct mm_struct gpa_mm; gpa_mm 323 arch/mips/kvm/entry.c UASM_i_LW(&p, A0, offsetof(struct kvm, arch.gpa_mm.pgd), S0); gpa_mm 366 arch/mips/kvm/entry.c offsetof(struct kvm, arch.gpa_mm.context.asid)); gpa_mm 146 arch/mips/kvm/mips.c kvm->arch.gpa_mm.pgd = kvm_pgd_alloc(); gpa_mm 147 arch/mips/kvm/mips.c if (!kvm->arch.gpa_mm.pgd) gpa_mm 176 arch/mips/kvm/mips.c pgd_free(NULL, kvm->arch.gpa_mm.pgd); gpa_mm 177 arch/mips/kvm/mmu.c return kvm_mips_walk_pgd(kvm->arch.gpa_mm.pgd, cache, addr); gpa_mm 301 arch/mips/kvm/mmu.c return kvm_mips_flush_gpa_pgd(kvm->arch.gpa_mm.pgd, gpa_mm 425 arch/mips/kvm/mmu.c return kvm_mips_mkclean_pgd(kvm->arch.gpa_mm.pgd, gpa_mm 463 arch/mips/kvm/mmu.c return kvm_mips_mkold_pgd(kvm->arch.gpa_mm.pgd, gpa_mm 46 arch/mips/kvm/tlb.c struct mm_struct *gpa_mm = &vcpu->kvm->arch.gpa_mm; gpa_mm 51 arch/mips/kvm/tlb.c return cpu_asid(smp_processor_id(), gpa_mm); gpa_mm 2406 arch/mips/kvm/vz.c struct mm_struct *gpa_mm = &kvm->arch.gpa_mm; gpa_mm 2458 arch/mips/kvm/vz.c get_new_mmu_context(gpa_mm); gpa_mm 2460 arch/mips/kvm/vz.c check_mmu_context(gpa_mm);