Lines Matching refs:gfn_t
107 struct kvm_memory_slot *memslot, gfn_t gfn);
768 gfn_t base_gfn; in __kvm_set_memory_region()
1048 gfn_t offset; in kvm_get_dirty_log_protect()
1089 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) in gfn_to_memslot()
1095 int kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn) in kvm_is_visible_gfn()
1107 unsigned long kvm_host_page_size(struct kvm *kvm, gfn_t gfn) in kvm_host_page_size()
1136 static unsigned long __gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, in __gfn_to_hva_many()
1137 gfn_t *nr_pages, bool write) in __gfn_to_hva_many()
1151 static unsigned long gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, in gfn_to_hva_many()
1152 gfn_t *nr_pages) in gfn_to_hva_many()
1158 gfn_t gfn) in gfn_to_hva_memslot()
1164 unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn) in gfn_to_hva()
1175 gfn_t gfn, bool *writable) in gfn_to_hva_memslot_prot()
1185 unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable) in gfn_to_hva_prot()
1360 __gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn, bool atomic, in __gfn_to_pfn_memslot()
1381 static pfn_t __gfn_to_pfn(struct kvm *kvm, gfn_t gfn, bool atomic, bool *async, in __gfn_to_pfn()
1395 pfn_t gfn_to_pfn_atomic(struct kvm *kvm, gfn_t gfn) in gfn_to_pfn_atomic()
1401 pfn_t gfn_to_pfn_async(struct kvm *kvm, gfn_t gfn, bool *async, in gfn_to_pfn_async()
1408 pfn_t gfn_to_pfn(struct kvm *kvm, gfn_t gfn) in gfn_to_pfn()
1414 pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, in gfn_to_pfn_prot()
1421 pfn_t gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn) in gfn_to_pfn_memslot()
1426 pfn_t gfn_to_pfn_memslot_atomic(struct kvm_memory_slot *slot, gfn_t gfn) in gfn_to_pfn_memslot_atomic()
1432 int gfn_to_page_many_atomic(struct kvm *kvm, gfn_t gfn, struct page **pages, in gfn_to_page_many_atomic()
1436 gfn_t entry; in gfn_to_page_many_atomic()
1462 struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn) in gfn_to_page()
1534 int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset, in kvm_read_guest_page()
1552 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_read_guest()
1575 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_read_guest_atomic()
1590 int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, const void *data, in kvm_write_guest_page()
1610 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_write_guest()
1633 gfn_t start_gfn = gpa >> PAGE_SHIFT; in kvm_gfn_to_hva_cache_init()
1634 gfn_t end_gfn = (gpa + len - 1) >> PAGE_SHIFT; in kvm_gfn_to_hva_cache_init()
1635 gfn_t nr_pages_needed = end_gfn - start_gfn + 1; in kvm_gfn_to_hva_cache_init()
1636 gfn_t nr_pages_avail; in kvm_gfn_to_hva_cache_init()
1716 int kvm_clear_guest_page(struct kvm *kvm, gfn_t gfn, int offset, int len) in kvm_clear_guest_page()
1726 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_clear_guest()
1745 gfn_t gfn) in mark_page_dirty_in_slot()
1754 void mark_page_dirty(struct kvm *kvm, gfn_t gfn) in mark_page_dirty()