map_writable 3326 arch/x86/kvm/mmu.c int map_writable, int level, kvm_pfn_t pfn, map_writable 3363 arch/x86/kvm/mmu.c map_writable); map_writable 3654 arch/x86/kvm/mmu.c bool map_writable, write = error_code & PFERR_WRITE_MASK; map_writable 3678 arch/x86/kvm/mmu.c if (try_async_pf(vcpu, prefault, gfn, gpa, &pfn, write, &map_writable)) map_writable 3692 arch/x86/kvm/mmu.c r = __direct_map(vcpu, gpa, write, map_writable, level, pfn, map_writable 4288 arch/x86/kvm/mmu.c bool map_writable; map_writable 4318 arch/x86/kvm/mmu.c if (try_async_pf(vcpu, prefault, gfn, gpa, &pfn, write, &map_writable)) map_writable 4332 arch/x86/kvm/mmu.c r = __direct_map(vcpu, gpa, write, map_writable, level, pfn, map_writable 617 arch/x86/kvm/paging_tmpl.h kvm_pfn_t pfn, bool map_writable, bool prefault, map_writable 705 arch/x86/kvm/paging_tmpl.h it.level, base_gfn, pfn, prefault, map_writable); map_writable 778 arch/x86/kvm/paging_tmpl.h bool map_writable, is_self_change_mapping; map_writable 834 arch/x86/kvm/paging_tmpl.h &map_writable)) map_writable 871 arch/x86/kvm/paging_tmpl.h level, pfn, map_writable, prefault, lpage_disallowed);