Searched refs:write_fault (Results 1 - 5 of 5) sorted by relevance

/linux-4.1.27/arch/x86/kvm/
H A Dpaging_tmpl.h210 int write_fault) update_accessed_dirty_bits()
231 if (level == walker->level && write_fault && update_accessed_dirty_bits()
279 const int write_fault = access & PFERR_WRITE_MASK; walk_addr_generic() local
385 if (!write_fault) walk_addr_generic()
397 ret = FNAME(update_accessed_dirty_bits)(vcpu, mmu, walker, write_fault); walk_addr_generic()
411 errcode |= write_fault | user_fault; walk_addr_generic()
562 int write_fault, int hlevel, fetch()
634 mmu_set_spte(vcpu, it.sptep, gw->pte_access, write_fault, &emulate, fetch()
704 int write_fault = error_code & PFERR_WRITE_MASK; page_fault() local
768 if (try_async_pf(vcpu, prefault, walker.gfn, addr, &pfn, write_fault, page_fault()
780 if (write_fault && !(walker.pte_access & ACC_WRITE_MASK) && page_fault()
804 r = FNAME(fetch)(vcpu, addr, &walker, write_fault, page_fault()
207 update_accessed_dirty_bits(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, struct guest_walker *walker, int write_fault) update_accessed_dirty_bits() argument
560 fetch(struct kvm_vcpu *vcpu, gva_t addr, struct guest_walker *gw, int write_fault, int hlevel, pfn_t pfn, bool map_writable, bool prefault) fetch() argument
H A Dmmu.c2594 unsigned pte_access, int write_fault, int *emulate, mmu_set_spte()
2601 pgprintk("%s: spte %llx write_fault %d gfn %llx\n", __func__, mmu_set_spte()
2602 *sptep, write_fault, gfn); mmu_set_spte()
2628 if (write_fault) mmu_set_spte()
2593 mmu_set_spte(struct kvm_vcpu *vcpu, u64 *sptep, unsigned pte_access, int write_fault, int *emulate, int level, gfn_t gfn, pfn_t pfn, bool speculative, bool host_writable) mmu_set_spte() argument
/linux-4.1.27/virt/kvm/
H A Dkvm_main.c1217 bool write_fault, bool *writable, pfn_t *pfn) hva_to_pfn_fast()
1230 if (!(write_fault || writable)) hva_to_pfn_fast()
1249 static int hva_to_pfn_slow(unsigned long addr, bool *async, bool write_fault, hva_to_pfn_slow() argument
1258 *writable = write_fault; hva_to_pfn_slow()
1263 addr, write_fault, page); hva_to_pfn_slow()
1267 write_fault, 0, page, hva_to_pfn_slow()
1273 if (unlikely(!write_fault) && writable) { hva_to_pfn_slow()
1289 static bool vma_is_valid(struct vm_area_struct *vma, bool write_fault) vma_is_valid() argument
1294 if (write_fault && (unlikely(!(vma->vm_flags & VM_WRITE)))) vma_is_valid()
1306 * @write_fault: whether we should get a writable host page
1307 * @writable: whether it allows to map a writable host page for !@write_fault
1310 * 1): @write_fault = true
1311 * 2): @write_fault = false && @writable, @writable will tell the caller
1315 bool write_fault, bool *writable) hva_to_pfn()
1324 if (hva_to_pfn_fast(addr, atomic, async, write_fault, writable, &pfn)) hva_to_pfn()
1330 npages = hva_to_pfn_slow(addr, async, write_fault, writable, &pfn); hva_to_pfn()
1350 if (async && vma_is_valid(vma, write_fault)) hva_to_pfn()
1361 bool *async, bool write_fault, bool *writable) __gfn_to_pfn_memslot()
1363 unsigned long addr = __gfn_to_hva_many(slot, gfn, NULL, write_fault); __gfn_to_pfn_memslot()
1377 return hva_to_pfn(addr, atomic, async, write_fault, __gfn_to_pfn_memslot()
1382 bool write_fault, bool *writable) __gfn_to_pfn()
1391 return __gfn_to_pfn_memslot(slot, gfn, atomic, async, write_fault, __gfn_to_pfn()
1402 bool write_fault, bool *writable) gfn_to_pfn_async()
1404 return __gfn_to_pfn(kvm, gfn, false, async, write_fault, writable); gfn_to_pfn_async()
1414 pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, gfn_to_pfn_prot() argument
1417 return __gfn_to_pfn(kvm, gfn, false, NULL, write_fault, writable); gfn_to_pfn_prot()
1216 hva_to_pfn_fast(unsigned long addr, bool atomic, bool *async, bool write_fault, bool *writable, pfn_t *pfn) hva_to_pfn_fast() argument
1314 hva_to_pfn(unsigned long addr, bool atomic, bool *async, bool write_fault, bool *writable) hva_to_pfn() argument
1360 __gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn, bool atomic, bool *async, bool write_fault, bool *writable) __gfn_to_pfn_memslot() argument
1381 __gfn_to_pfn(struct kvm *kvm, gfn_t gfn, bool atomic, bool *async, bool write_fault, bool *writable) __gfn_to_pfn() argument
1401 gfn_to_pfn_async(struct kvm *kvm, gfn_t gfn, bool *async, bool write_fault, bool *writable) gfn_to_pfn_async() argument
/linux-4.1.27/arch/arm/kvm/
H A Dmmu.c1219 bool write_fault, writable, hugetlb = false, force_pte = false; user_mem_abort() local
1231 write_fault = kvm_is_write_fault(vcpu); user_mem_abort()
1232 if (fault_status == FSC_PERM && !write_fault) { user_mem_abort()
1283 pfn = gfn_to_pfn_prot(kvm, gfn, write_fault, &writable); user_mem_abort()
1303 if (!write_fault) user_mem_abort()
1402 bool is_iabt, write_fault, writable; kvm_handle_guest_abort() local
1428 write_fault = kvm_is_write_fault(vcpu); kvm_handle_guest_abort()
1429 if (kvm_is_error_hva(hva) || (write_fault && !writable)) { kvm_handle_guest_abort()
/linux-4.1.27/include/linux/
H A Dkvm_host.h553 bool write_fault, bool *writable);
555 pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault,

Completed in 185 milliseconds