gfn_t 919 arch/mips/include/asm/kvm_host.h bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn); gfn_t 920 arch/mips/include/asm/kvm_host.h int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn); gfn_t 299 arch/mips/kvm/mmu.c bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn) gfn_t 423 arch/mips/kvm/mmu.c int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn) gfn_t 443 arch/mips/kvm/mmu.c gfn_t gfn_offset, unsigned long mask) gfn_t 445 arch/mips/kvm/mmu.c gfn_t base_gfn = slot->base_gfn + gfn_offset; gfn_t 446 arch/mips/kvm/mmu.c gfn_t start = base_gfn + __ffs(mask); gfn_t 447 arch/mips/kvm/mmu.c gfn_t end = base_gfn + __fls(mask); gfn_t 460 arch/mips/kvm/mmu.c static int kvm_mips_mkold_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t 461 arch/mips/kvm/mmu.c gfn_t end_gfn) gfn_t 471 arch/mips/kvm/mmu.c int (*handler)(struct kvm *kvm, gfn_t gfn, gfn_t 486 arch/mips/kvm/mmu.c gfn_t gfn, gfn_end; gfn_t 508 arch/mips/kvm/mmu.c static int kvm_unmap_hva_handler(struct kvm *kvm, gfn_t gfn, gfn_t gfn_end, gfn_t 523 arch/mips/kvm/mmu.c static int kvm_set_spte_handler(struct kvm *kvm, gfn_t gfn, gfn_t gfn_end, gfn_t 565 arch/mips/kvm/mmu.c static int kvm_age_hva_handler(struct kvm *kvm, gfn_t gfn, gfn_t gfn_end, gfn_t 571 arch/mips/kvm/mmu.c static int kvm_test_age_hva_handler(struct kvm *kvm, gfn_t gfn, gfn_t gfn_end, gfn_t 615 arch/mips/kvm/mmu.c gfn_t gfn = gpa >> PAGE_SHIFT; gfn_t 692 arch/mips/kvm/mmu.c gfn_t gfn = gpa >> PAGE_SHIFT; gfn_t 430 arch/powerpc/kvm/book3s.c gfn_t gfn = gpa >> PAGE_SHIFT; gfn_t 782 arch/powerpc/kvm/book3s_64_mmu_hv.c gfn_t gfn, gfn_end; gfn_t 403 arch/powerpc/kvm/book3s_pr.c gfn_t gfn, gfn_end; gfn_t 565 arch/powerpc/kvm/book3s_xive_native.c gfn_t gfn; gfn_t 1239 arch/powerpc/kvm/booke.c gfn_t gfn; gfn_t 1296 arch/powerpc/kvm/booke.c gfn_t gfn; gfn_t 323 arch/powerpc/kvm/e500_mmu_host.c u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, gfn_t 560 arch/powerpc/kvm/e500_mmu_host.c u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, gfn_t 612 arch/powerpc/kvm/e500_mmu_host.c gfn_t gfn = gpaddr >> PAGE_SHIFT; gfn_t 579 arch/s390/kvm/kvm-s390.c gfn_t cur_gfn, last_gfn; gfn_t 1916 arch/s390/kvm/kvm-s390.c static int gfn_to_memslot_approx(struct kvm_memslots *slots, gfn_t gfn) gfn_t 120 arch/x86/include/asm/kvm_host.h static inline gfn_t gfn_to_index(gfn_t gfn, gfn_t base_gfn, int level) gfn_t 327 arch/x86/include/asm/kvm_host.h gfn_t gfn; gfn_t 331 arch/x86/include/asm/kvm_host.h gfn_t *gfns; gfn_t 719 arch/x86/include/asm/kvm_host.h gfn_t mmio_gfn; gfn_t 736 arch/x86/include/asm/kvm_host.h gfn_t gfns[roundup_pow_of_two(ASYNC_PF_PER_VCPU)]; gfn_t 1107 arch/x86/include/asm/kvm_host.h u64 (*get_mt_mask)(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio); gfn_t 1162 arch/x86/include/asm/kvm_host.h gfn_t offset, unsigned long mask); gfn_t 1225 arch/x86/include/asm/kvm_host.h gfn_t gfn; gfn_t 1278 arch/x86/include/asm/kvm_host.h gfn_t gfn_offset, unsigned long mask); gfn_t 1406 arch/x86/include/asm/kvm_host.h gfn_t gfn, void *data, int offset, int len, gfn_t 1432 arch/x86/include/asm/kvm_host.h int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn); gfn_t 1590 arch/x86/include/asm/kvm_host.h extern bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn); gfn_t 58 arch/x86/include/asm/kvm_page_track.h struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 61 arch/x86/include/asm/kvm_page_track.h struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 63 arch/x86/include/asm/kvm_page_track.h bool kvm_page_track_is_active(struct kvm_vcpu *vcpu, gfn_t gfn, gfn_t 569 arch/x86/kvm/hyperv.c gfn_t msg_page_gfn; gfn_t 473 arch/x86/kvm/mmu.c static gfn_t get_mmio_spte_gfn(u64 spte) gfn_t 488 arch/x86/kvm/mmu.c static bool set_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, gfn_t 638 arch/x86/kvm/mmu.c static gfn_t pse36_gfn_delta(u32 gpte) gfn_t 1155 arch/x86/kvm/mmu.c static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) gfn_t 1163 arch/x86/kvm/mmu.c static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn) gfn_t 1181 arch/x86/kvm/mmu.c static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn, gfn_t 1192 arch/x86/kvm/mmu.c gfn_t gfn, int count) gfn_t 1204 arch/x86/kvm/mmu.c void kvm_mmu_gfn_disallow_lpage(struct kvm_memory_slot *slot, gfn_t gfn) gfn_t 1209 arch/x86/kvm/mmu.c void kvm_mmu_gfn_allow_lpage(struct kvm_memory_slot *slot, gfn_t gfn) gfn_t 1218 arch/x86/kvm/mmu.c gfn_t gfn; gfn_t 1248 arch/x86/kvm/mmu.c gfn_t gfn; gfn_t 1268 arch/x86/kvm/mmu.c static bool __mmu_gfn_lpage_is_disallowed(gfn_t gfn, int level, gfn_t 1281 arch/x86/kvm/mmu.c static bool mmu_gfn_lpage_is_disallowed(struct kvm_vcpu *vcpu, gfn_t gfn, gfn_t 1290 arch/x86/kvm/mmu.c static int host_mapping_level(struct kvm_vcpu *vcpu, gfn_t gfn) gfn_t 1319 arch/x86/kvm/mmu.c gfn_to_memslot_dirty_bitmap(struct kvm_vcpu *vcpu, gfn_t gfn, gfn_t 1331 arch/x86/kvm/mmu.c static int mapping_level(struct kvm_vcpu *vcpu, gfn_t large_gfn, gfn_t 1469 arch/x86/kvm/mmu.c static struct kvm_rmap_head *__gfn_to_rmap(gfn_t gfn, int level, gfn_t 1478 arch/x86/kvm/mmu.c static struct kvm_rmap_head *gfn_to_rmap(struct kvm *kvm, gfn_t gfn, gfn_t 1497 arch/x86/kvm/mmu.c static int rmap_add(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) gfn_t 1511 arch/x86/kvm/mmu.c gfn_t gfn; gfn_t 1753 arch/x86/kvm/mmu.c gfn_t gfn_offset, unsigned long mask) gfn_t 1779 arch/x86/kvm/mmu.c gfn_t gfn_offset, unsigned long mask) gfn_t 1806 arch/x86/kvm/mmu.c gfn_t gfn_offset, unsigned long mask) gfn_t 1870 arch/x86/kvm/mmu.c struct kvm_memory_slot *slot, gfn_t gfn, int level, gfn_t 1877 arch/x86/kvm/mmu.c struct kvm_memory_slot *slot, gfn_t gfn, int level, gfn_t 1925 arch/x86/kvm/mmu.c gfn_t start_gfn; gfn_t 1926 arch/x86/kvm/mmu.c gfn_t end_gfn; gfn_t 1931 arch/x86/kvm/mmu.c gfn_t gfn; gfn_t 1952 arch/x86/kvm/mmu.c int end_level, gfn_t start_gfn, gfn_t end_gfn) gfn_t 1997 arch/x86/kvm/mmu.c gfn_t gfn, gfn_t 2011 arch/x86/kvm/mmu.c gfn_t gfn_start, gfn_end; gfn_t 2042 arch/x86/kvm/mmu.c gfn_t gfn, int level, gfn_t 2059 arch/x86/kvm/mmu.c struct kvm_memory_slot *slot, gfn_t gfn, int level, gfn_t 2074 arch/x86/kvm/mmu.c struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 2088 arch/x86/kvm/mmu.c static void rmap_recycle(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) gfn_t 2151 arch/x86/kvm/mmu.c static unsigned kvm_page_table_hashfn(gfn_t gfn) gfn_t 2414 arch/x86/kvm/mmu.c static bool kvm_sync_pages(struct kvm_vcpu *vcpu, gfn_t gfn, gfn_t 2551 arch/x86/kvm/mmu.c gfn_t gfn, gfn_t 2932 arch/x86/kvm/mmu.c int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn) gfn_t 2963 arch/x86/kvm/mmu.c static bool mmu_need_write_protect(struct kvm_vcpu *vcpu, gfn_t gfn, gfn_t 3051 arch/x86/kvm/mmu.c gfn_t gfn, kvm_pfn_t pfn, bool speculative, gfn_t 3154 arch/x86/kvm/mmu.c int write_fault, int level, gfn_t gfn, kvm_pfn_t pfn, gfn_t 3219 arch/x86/kvm/mmu.c static kvm_pfn_t pte_prefetch_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, gfn_t 3239 arch/x86/kvm/mmu.c gfn_t gfn; gfn_t 3303 arch/x86/kvm/mmu.c gfn_t gfn, kvm_pfn_t *pfnp, int *levelp) gfn_t 3332 arch/x86/kvm/mmu.c gfn_t gfn = gpa >> PAGE_SHIFT; gfn_t 3333 arch/x86/kvm/mmu.c gfn_t base_gfn = gfn; gfn_t 3374 arch/x86/kvm/mmu.c static int kvm_handle_bad_page(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_pfn_t pfn) gfn_t 3393 arch/x86/kvm/mmu.c gfn_t gfn, kvm_pfn_t *pfnp, gfn_t 3431 arch/x86/kvm/mmu.c static bool handle_abnormal_pfn(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, gfn_t 3488 arch/x86/kvm/mmu.c gfn_t gfn; gfn_t 3641 arch/x86/kvm/mmu.c static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn, gfn_t 3647 arch/x86/kvm/mmu.c gfn_t gfn, bool prefault) gfn_t 3765 arch/x86/kvm/mmu.c static int mmu_check_root(struct kvm_vcpu *vcpu, gfn_t root_gfn) gfn_t 3822 arch/x86/kvm/mmu.c gfn_t root_gfn, root_cr3; gfn_t 4100 arch/x86/kvm/mmu.c gfn_t gfn = get_mmio_spte_gfn(spte); gfn_t 4122 arch/x86/kvm/mmu.c u32 error_code, gfn_t gfn) gfn_t 4161 arch/x86/kvm/mmu.c gfn_t gfn = gpa >> PAGE_SHIFT; gfn_t 4182 arch/x86/kvm/mmu.c gfn_t gfn) gfn_t 4195 arch/x86/kvm/mmu.c static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn, gfn_t 4269 arch/x86/kvm/mmu.c check_hugepage_cache_consistency(struct kvm_vcpu *vcpu, gfn_t gfn, int level) gfn_t 4285 arch/x86/kvm/mmu.c gfn_t gfn = gpa >> PAGE_SHIFT; gfn_t 4463 arch/x86/kvm/mmu.c static bool sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, gfn_t 5435 arch/x86/kvm/mmu.c gfn_t gfn = gpa >> PAGE_SHIFT; gfn_t 5693 arch/x86/kvm/mmu.c gfn_t start_gfn, gfn_t end_gfn, bool lock_flush_tlb) gfn_t 5947 arch/x86/kvm/mmu.c void kvm_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end) gfn_t 5957 arch/x86/kvm/mmu.c gfn_t start, end; gfn_t 206 arch/x86/kvm/mmu.h void kvm_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end); gfn_t 208 arch/x86/kvm/mmu.h void kvm_mmu_gfn_disallow_lpage(struct kvm_memory_slot *slot, gfn_t gfn); gfn_t 209 arch/x86/kvm/mmu.h void kvm_mmu_gfn_allow_lpage(struct kvm_memory_slot *slot, gfn_t gfn); gfn_t 96 arch/x86/kvm/mmu_audit.c gfn_t gfn; gfn_t 133 arch/x86/kvm/mmu_audit.c gfn_t gfn; gfn_t 205 arch/x86/kvm/mmutrace.h TP_PROTO(u64 *sptep, gfn_t gfn, unsigned access, unsigned int gen), gfn_t 210 arch/x86/kvm/mmutrace.h __field(gfn_t, gfn) gfn_t 228 arch/x86/kvm/mmutrace.h TP_PROTO(u64 addr, gfn_t gfn, unsigned access), gfn_t 233 arch/x86/kvm/mmutrace.h __field(gfn_t, gfn) gfn_t 331 arch/x86/kvm/mmutrace.h TP_PROTO(int level, gfn_t gfn, u64 *sptep), gfn_t 313 arch/x86/kvm/mtrr.c gfn_t start, end; gfn_t 619 arch/x86/kvm/mtrr.c u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn) gfn_t 695 arch/x86/kvm/mtrr.c bool kvm_mtrr_check_gfn_range_consistency(struct kvm_vcpu *vcpu, gfn_t gfn, gfn_t 63 arch/x86/kvm/page_track.c static void update_gfn_track(struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 91 arch/x86/kvm/page_track.c struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 126 arch/x86/kvm/page_track.c struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 145 arch/x86/kvm/page_track.c bool kvm_page_track_is_active(struct kvm_vcpu *vcpu, gfn_t gfn, gfn_t 87 arch/x86/kvm/paging_tmpl.h gfn_t table_gfn[PT_MAX_FULL_LEVELS]; gfn_t 95 arch/x86/kvm/paging_tmpl.h gfn_t gfn; gfn_t 99 arch/x86/kvm/paging_tmpl.h static gfn_t gpte_to_gfn_lvl(pt_element_t gpte, int lvl) gfn_t 228 arch/x86/kvm/paging_tmpl.h gfn_t table_gfn; gfn_t 303 arch/x86/kvm/paging_tmpl.h gfn_t table_gfn; gfn_t 316 arch/x86/kvm/paging_tmpl.h gfn_t gfn; gfn_t 348 arch/x86/kvm/paging_tmpl.h gfn_t real_gfn; gfn_t 520 arch/x86/kvm/paging_tmpl.h gfn_t gfn; gfn_t 624 arch/x86/kvm/paging_tmpl.h gfn_t gfn, base_gfn; gfn_t 646 arch/x86/kvm/paging_tmpl.h gfn_t table_gfn; gfn_t 737 arch/x86/kvm/paging_tmpl.h gfn_t mask = ~(KVM_PAGES_PER_HPAGE(walker->level) - 1); gfn_t 745 arch/x86/kvm/paging_tmpl.h gfn_t gfn = walker->gfn ^ walker->table_gfn[level - 1]; gfn_t 1023 arch/x86/kvm/paging_tmpl.h gfn_t gfn; gfn_t 5938 arch/x86/kvm/svm.c static u64 svm_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio) gfn_t 3388 arch/x86/kvm/vmx/nested.c gfn_t gfn; gfn_t 3428 arch/x86/kvm/vmx/vmx.c gfn_t fn; gfn_t 6863 arch/x86/kvm/vmx/vmx.c static u64 vmx_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio) gfn_t 7313 arch/x86/kvm/vmx/vmx.c gfn_t offset, unsigned long mask) gfn_t 664 arch/x86/kvm/x86.c gfn_t ngfn, void *data, int offset, int len, gfn_t 668 arch/x86/kvm/x86.c gfn_t real_gfn; gfn_t 682 arch/x86/kvm/x86.c static int kvm_read_nested_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, gfn_t 700 arch/x86/kvm/x86.c gfn_t pdpt_gfn = cr3 >> PAGE_SHIFT; gfn_t 738 arch/x86/kvm/x86.c gfn_t gfn; gfn_t 10082 arch/x86/kvm/x86.c static inline u32 kvm_async_pf_hash_fn(gfn_t gfn) gfn_t 10092 arch/x86/kvm/x86.c static void kvm_add_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) gfn_t 10102 arch/x86/kvm/x86.c static u32 kvm_async_pf_gfn_slot(struct kvm_vcpu *vcpu, gfn_t gfn) gfn_t 10115 arch/x86/kvm/x86.c bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) gfn_t 10120 arch/x86/kvm/x86.c static void kvm_del_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) gfn_t 187 arch/x86/kvm/x86.h gva_t gva, gfn_t gfn, unsigned access) gfn_t 282 arch/x86/kvm/x86.h u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn); gfn_t 286 arch/x86/kvm/x86.h bool kvm_mtrr_check_gfn_range_consistency(struct kvm_vcpu *vcpu, gfn_t gfn, gfn_t 87 drivers/gpu/drm/i915/gvt/kvmgt.c gfn_t gfn; gfn_t 105 drivers/gpu/drm/i915/gvt/kvmgt.c gfn_t gfn; gfn_t 238 drivers/gpu/drm/i915/gvt/kvmgt.c static struct gvt_dma *__gvt_cache_find_gfn(struct intel_vgpu *vgpu, gfn_t gfn) gfn_t 256 drivers/gpu/drm/i915/gvt/kvmgt.c static int __gvt_cache_add(struct intel_vgpu *vgpu, gfn_t gfn, gfn_t 359 drivers/gpu/drm/i915/gvt/kvmgt.c __kvmgt_protect_table_find(struct kvmgt_guest_info *info, gfn_t gfn) gfn_t 374 drivers/gpu/drm/i915/gvt/kvmgt.c gfn_t gfn) gfn_t 382 drivers/gpu/drm/i915/gvt/kvmgt.c static void kvmgt_protect_table_add(struct kvmgt_guest_info *info, gfn_t gfn) gfn_t 398 drivers/gpu/drm/i915/gvt/kvmgt.c gfn_t gfn) gfn_t 1719 drivers/gpu/drm/i915/gvt/kvmgt.c gfn_t gfn; gfn_t 344 include/linux/kvm_host.h gfn_t base_gfn; gfn_t 702 include/linux/kvm_host.h int gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 705 include/linux/kvm_host.h struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn); gfn_t 706 include/linux/kvm_host.h unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn); gfn_t 707 include/linux/kvm_host.h unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable); gfn_t 708 include/linux/kvm_host.h unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn); gfn_t 709 include/linux/kvm_host.h unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 715 include/linux/kvm_host.h kvm_pfn_t gfn_to_pfn_atomic(struct kvm *kvm, gfn_t gfn); gfn_t 716 include/linux/kvm_host.h kvm_pfn_t gfn_to_pfn(struct kvm *kvm, gfn_t gfn); gfn_t 717 include/linux/kvm_host.h kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, gfn_t 719 include/linux/kvm_host.h kvm_pfn_t gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn); gfn_t 720 include/linux/kvm_host.h kvm_pfn_t gfn_to_pfn_memslot_atomic(struct kvm_memory_slot *slot, gfn_t gfn); gfn_t 721 include/linux/kvm_host.h kvm_pfn_t __gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 732 include/linux/kvm_host.h int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset, gfn_t 739 include/linux/kvm_host.h int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, const void *data, gfn_t 750 include/linux/kvm_host.h int kvm_clear_guest_page(struct kvm *kvm, gfn_t gfn, int offset, int len); gfn_t 752 include/linux/kvm_host.h struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn); gfn_t 753 include/linux/kvm_host.h bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn); gfn_t 754 include/linux/kvm_host.h unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn); gfn_t 755 include/linux/kvm_host.h void mark_page_dirty(struct kvm *kvm, gfn_t gfn); gfn_t 758 include/linux/kvm_host.h struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn); gfn_t 759 include/linux/kvm_host.h kvm_pfn_t kvm_vcpu_gfn_to_pfn_atomic(struct kvm_vcpu *vcpu, gfn_t gfn); gfn_t 760 include/linux/kvm_host.h kvm_pfn_t kvm_vcpu_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn); gfn_t 762 include/linux/kvm_host.h int kvm_map_gfn(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map, gfn_t 764 include/linux/kvm_host.h struct page *kvm_vcpu_gfn_to_page(struct kvm_vcpu *vcpu, gfn_t gfn); gfn_t 768 include/linux/kvm_host.h unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn); gfn_t 769 include/linux/kvm_host.h unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable); gfn_t 770 include/linux/kvm_host.h int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, int offset, gfn_t 776 include/linux/kvm_host.h int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, const void *data, gfn_t 780 include/linux/kvm_host.h void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn); gfn_t 818 include/linux/kvm_host.h gfn_t gfn_offset, gfn_t 1011 include/linux/kvm_host.h search_memslots(struct kvm_memslots *slots, gfn_t gfn) gfn_t 1040 include/linux/kvm_host.h __gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn) gfn_t 1046 include/linux/kvm_host.h __gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn) gfn_t 1051 include/linux/kvm_host.h static inline int memslot_id(struct kvm *kvm, gfn_t gfn) gfn_t 1056 include/linux/kvm_host.h static inline gfn_t gfn_t 1059 include/linux/kvm_host.h gfn_t gfn_offset = (hva - slot->userspace_addr) >> PAGE_SHIFT; gfn_t 1064 include/linux/kvm_host.h static inline gpa_t gfn_to_gpa(gfn_t gfn) gfn_t 1069 include/linux/kvm_host.h static inline gfn_t gpa_to_gfn(gpa_t gpa) gfn_t 1071 include/linux/kvm_host.h return (gfn_t)(gpa >> PAGE_SHIFT); gfn_t 54 include/linux/kvm_types.h gfn_t gfn; gfn_t 1379 virt/kvm/arm/mmu.c gfn_t gfn = *ipap >> PAGE_SHIFT; gfn_t 1566 virt/kvm/arm/mmu.c gfn_t gfn_offset, unsigned long mask) gfn_t 1584 virt/kvm/arm/mmu.c gfn_t gfn_offset, unsigned long mask) gfn_t 1677 virt/kvm/arm/mmu.c gfn_t gfn = fault_ipa >> PAGE_SHIFT; gfn_t 1915 virt/kvm/arm/mmu.c gfn_t gfn; gfn_t 2030 virt/kvm/arm/mmu.c gfn_t gpa; gfn_t 904 virt/kvm/arm/vgic/vgic-its.c gfn_t gfn; gfn_t 147 virt/kvm/kvm_main.c static void mark_page_dirty_in_slot(struct kvm_memory_slot *memslot, gfn_t gfn); gfn_t 995 virt/kvm/kvm_main.c gfn_t base_gfn; gfn_t 1272 virt/kvm/kvm_main.c gfn_t offset; gfn_t 1307 virt/kvm/kvm_main.c gfn_t offset; gfn_t 1380 virt/kvm/kvm_main.c struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) gfn_t 1386 virt/kvm/kvm_main.c struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn) gfn_t 1391 virt/kvm/kvm_main.c bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn) gfn_t 1403 virt/kvm/kvm_main.c unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn) gfn_t 1432 virt/kvm/kvm_main.c static unsigned long __gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 1433 virt/kvm/kvm_main.c gfn_t *nr_pages, bool write) gfn_t 1447 virt/kvm/kvm_main.c static unsigned long gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 1448 virt/kvm/kvm_main.c gfn_t *nr_pages) gfn_t 1454 virt/kvm/kvm_main.c gfn_t gfn) gfn_t 1460 virt/kvm/kvm_main.c unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn) gfn_t 1466 virt/kvm/kvm_main.c unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn) gfn_t 1481 virt/kvm/kvm_main.c gfn_t gfn, bool *writable) gfn_t 1491 virt/kvm/kvm_main.c unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable) gfn_t 1498 virt/kvm/kvm_main.c unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable) gfn_t 1705 virt/kvm/kvm_main.c kvm_pfn_t __gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 1734 virt/kvm/kvm_main.c kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, gfn_t 1742 virt/kvm/kvm_main.c kvm_pfn_t gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn) gfn_t 1748 virt/kvm/kvm_main.c kvm_pfn_t gfn_to_pfn_memslot_atomic(struct kvm_memory_slot *slot, gfn_t gfn) gfn_t 1754 virt/kvm/kvm_main.c kvm_pfn_t gfn_to_pfn_atomic(struct kvm *kvm, gfn_t gfn) gfn_t 1760 virt/kvm/kvm_main.c kvm_pfn_t kvm_vcpu_gfn_to_pfn_atomic(struct kvm_vcpu *vcpu, gfn_t gfn) gfn_t 1766 virt/kvm/kvm_main.c kvm_pfn_t gfn_to_pfn(struct kvm *kvm, gfn_t gfn) gfn_t 1772 virt/kvm/kvm_main.c kvm_pfn_t kvm_vcpu_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn) gfn_t 1778 virt/kvm/kvm_main.c int gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 1782 virt/kvm/kvm_main.c gfn_t entry = 0; gfn_t 1808 virt/kvm/kvm_main.c struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn) gfn_t 1832 virt/kvm/kvm_main.c static void kvm_cache_gfn_to_pfn(struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 1843 virt/kvm/kvm_main.c static int __kvm_map_gfn(struct kvm_memslots *slots, gfn_t gfn, gfn_t 1898 virt/kvm/kvm_main.c int kvm_map_gfn(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map, gfn_t 1906 virt/kvm/kvm_main.c int kvm_vcpu_map(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map) gfn_t 1965 virt/kvm/kvm_main.c struct page *kvm_vcpu_gfn_to_page(struct kvm_vcpu *vcpu, gfn_t gfn) gfn_t 2037 virt/kvm/kvm_main.c static int __kvm_read_guest_page(struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 2052 virt/kvm/kvm_main.c int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset, gfn_t 2061 virt/kvm/kvm_main.c int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, gfn_t 2072 virt/kvm/kvm_main.c gfn_t gfn = gpa >> PAGE_SHIFT; gfn_t 2092 virt/kvm/kvm_main.c gfn_t gfn = gpa >> PAGE_SHIFT; gfn_t 2110 virt/kvm/kvm_main.c static int __kvm_read_guest_atomic(struct kvm_memory_slot *slot, gfn_t gfn, gfn_t 2130 virt/kvm/kvm_main.c gfn_t gfn = gpa >> PAGE_SHIFT; gfn_t 2141 virt/kvm/kvm_main.c gfn_t gfn = gpa >> PAGE_SHIFT; gfn_t 2149 virt/kvm/kvm_main.c static int __kvm_write_guest_page(struct kvm_memory_slot *memslot, gfn_t gfn, gfn_t 2165 virt/kvm/kvm_main.c int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, gfn_t 2174 virt/kvm/kvm_main.c int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, gfn_t 2186 virt/kvm/kvm_main.c gfn_t gfn = gpa >> PAGE_SHIFT; gfn_t 2207 virt/kvm/kvm_main.c gfn_t gfn = gpa >> PAGE_SHIFT; gfn_t 2230 virt/kvm/kvm_main.c gfn_t start_gfn = gpa >> PAGE_SHIFT; gfn_t 2231 virt/kvm/kvm_main.c gfn_t end_gfn = (gpa + len - 1) >> PAGE_SHIFT; gfn_t 2232 virt/kvm/kvm_main.c gfn_t nr_pages_needed = end_gfn - start_gfn + 1; gfn_t 2233 virt/kvm/kvm_main.c gfn_t nr_pages_avail; gfn_t 2331 virt/kvm/kvm_main.c int kvm_clear_guest_page(struct kvm *kvm, gfn_t gfn, int offset, int len) gfn_t 2341 virt/kvm/kvm_main.c gfn_t gfn = gpa >> PAGE_SHIFT; gfn_t 2359 virt/kvm/kvm_main.c gfn_t gfn) gfn_t 2368 virt/kvm/kvm_main.c void mark_page_dirty(struct kvm *kvm, gfn_t gfn) gfn_t 2377 virt/kvm/kvm_main.c void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn)