map_writable     3326 arch/x86/kvm/mmu.c 			int map_writable, int level, kvm_pfn_t pfn,
map_writable     3363 arch/x86/kvm/mmu.c 			   map_writable);
map_writable     3654 arch/x86/kvm/mmu.c 	bool map_writable, write = error_code & PFERR_WRITE_MASK;
map_writable     3678 arch/x86/kvm/mmu.c 	if (try_async_pf(vcpu, prefault, gfn, gpa, &pfn, write, &map_writable))
map_writable     3692 arch/x86/kvm/mmu.c 	r = __direct_map(vcpu, gpa, write, map_writable, level, pfn,
map_writable     4288 arch/x86/kvm/mmu.c 	bool map_writable;
map_writable     4318 arch/x86/kvm/mmu.c 	if (try_async_pf(vcpu, prefault, gfn, gpa, &pfn, write, &map_writable))
map_writable     4332 arch/x86/kvm/mmu.c 	r = __direct_map(vcpu, gpa, write, map_writable, level, pfn,
map_writable      617 arch/x86/kvm/paging_tmpl.h 			 kvm_pfn_t pfn, bool map_writable, bool prefault,
map_writable      705 arch/x86/kvm/paging_tmpl.h 			   it.level, base_gfn, pfn, prefault, map_writable);
map_writable      778 arch/x86/kvm/paging_tmpl.h 	bool map_writable, is_self_change_mapping;
map_writable      834 arch/x86/kvm/paging_tmpl.h 			 &map_writable))
map_writable      871 arch/x86/kvm/paging_tmpl.h 			 level, pfn, map_writable, prefault, lpage_disallowed);