gpa_mm            186 arch/mips/include/asm/kvm_host.h 	struct mm_struct gpa_mm;
gpa_mm            323 arch/mips/kvm/entry.c 	UASM_i_LW(&p, A0, offsetof(struct kvm, arch.gpa_mm.pgd), S0);
gpa_mm            366 arch/mips/kvm/entry.c 		     offsetof(struct kvm, arch.gpa_mm.context.asid));
gpa_mm            146 arch/mips/kvm/mips.c 	kvm->arch.gpa_mm.pgd = kvm_pgd_alloc();
gpa_mm            147 arch/mips/kvm/mips.c 	if (!kvm->arch.gpa_mm.pgd)
gpa_mm            176 arch/mips/kvm/mips.c 	pgd_free(NULL, kvm->arch.gpa_mm.pgd);
gpa_mm            177 arch/mips/kvm/mmu.c 	return kvm_mips_walk_pgd(kvm->arch.gpa_mm.pgd, cache, addr);
gpa_mm            301 arch/mips/kvm/mmu.c 	return kvm_mips_flush_gpa_pgd(kvm->arch.gpa_mm.pgd,
gpa_mm            425 arch/mips/kvm/mmu.c 	return kvm_mips_mkclean_pgd(kvm->arch.gpa_mm.pgd,
gpa_mm            463 arch/mips/kvm/mmu.c 	return kvm_mips_mkold_pgd(kvm->arch.gpa_mm.pgd,
gpa_mm             46 arch/mips/kvm/tlb.c 	struct mm_struct *gpa_mm = &vcpu->kvm->arch.gpa_mm;
gpa_mm             51 arch/mips/kvm/tlb.c 		return cpu_asid(smp_processor_id(), gpa_mm);
gpa_mm           2406 arch/mips/kvm/vz.c 	struct mm_struct *gpa_mm = &kvm->arch.gpa_mm;
gpa_mm           2458 arch/mips/kvm/vz.c 			get_new_mmu_context(gpa_mm);
gpa_mm           2460 arch/mips/kvm/vz.c 			check_mmu_context(gpa_mm);