gfn_t             919 arch/mips/include/asm/kvm_host.h bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn);
gfn_t             920 arch/mips/include/asm/kvm_host.h int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn);
gfn_t             299 arch/mips/kvm/mmu.c bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn)
gfn_t             423 arch/mips/kvm/mmu.c int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn)
gfn_t             443 arch/mips/kvm/mmu.c 		gfn_t gfn_offset, unsigned long mask)
gfn_t             445 arch/mips/kvm/mmu.c 	gfn_t base_gfn = slot->base_gfn + gfn_offset;
gfn_t             446 arch/mips/kvm/mmu.c 	gfn_t start = base_gfn +  __ffs(mask);
gfn_t             447 arch/mips/kvm/mmu.c 	gfn_t end = base_gfn + __fls(mask);
gfn_t             460 arch/mips/kvm/mmu.c static int kvm_mips_mkold_gpa_pt(struct kvm *kvm, gfn_t start_gfn,
gfn_t             461 arch/mips/kvm/mmu.c 				 gfn_t end_gfn)
gfn_t             471 arch/mips/kvm/mmu.c 			     int (*handler)(struct kvm *kvm, gfn_t gfn,
gfn_t             486 arch/mips/kvm/mmu.c 		gfn_t gfn, gfn_end;
gfn_t             508 arch/mips/kvm/mmu.c static int kvm_unmap_hva_handler(struct kvm *kvm, gfn_t gfn, gfn_t gfn_end,
gfn_t             523 arch/mips/kvm/mmu.c static int kvm_set_spte_handler(struct kvm *kvm, gfn_t gfn, gfn_t gfn_end,
gfn_t             565 arch/mips/kvm/mmu.c static int kvm_age_hva_handler(struct kvm *kvm, gfn_t gfn, gfn_t gfn_end,
gfn_t             571 arch/mips/kvm/mmu.c static int kvm_test_age_hva_handler(struct kvm *kvm, gfn_t gfn, gfn_t gfn_end,
gfn_t             615 arch/mips/kvm/mmu.c 	gfn_t gfn = gpa >> PAGE_SHIFT;
gfn_t             692 arch/mips/kvm/mmu.c 	gfn_t gfn = gpa >> PAGE_SHIFT;
gfn_t             430 arch/powerpc/kvm/book3s.c 	gfn_t gfn = gpa >> PAGE_SHIFT;
gfn_t             782 arch/powerpc/kvm/book3s_64_mmu_hv.c 		gfn_t gfn, gfn_end;
gfn_t             403 arch/powerpc/kvm/book3s_pr.c 		gfn_t gfn, gfn_end;
gfn_t             565 arch/powerpc/kvm/book3s_xive_native.c 	gfn_t gfn;
gfn_t            1239 arch/powerpc/kvm/booke.c 		gfn_t gfn;
gfn_t            1296 arch/powerpc/kvm/booke.c 		gfn_t gfn;
gfn_t             323 arch/powerpc/kvm/e500_mmu_host.c 	u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe,
gfn_t             560 arch/powerpc/kvm/e500_mmu_host.c 		u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe,
gfn_t             612 arch/powerpc/kvm/e500_mmu_host.c 		gfn_t gfn = gpaddr >> PAGE_SHIFT;
gfn_t             579 arch/s390/kvm/kvm-s390.c 	gfn_t cur_gfn, last_gfn;
gfn_t            1916 arch/s390/kvm/kvm-s390.c static int gfn_to_memslot_approx(struct kvm_memslots *slots, gfn_t gfn)
gfn_t             120 arch/x86/include/asm/kvm_host.h static inline gfn_t gfn_to_index(gfn_t gfn, gfn_t base_gfn, int level)
gfn_t             327 arch/x86/include/asm/kvm_host.h 	gfn_t gfn;
gfn_t             331 arch/x86/include/asm/kvm_host.h 	gfn_t *gfns;
gfn_t             719 arch/x86/include/asm/kvm_host.h 	gfn_t mmio_gfn;
gfn_t             736 arch/x86/include/asm/kvm_host.h 		gfn_t gfns[roundup_pow_of_two(ASYNC_PF_PER_VCPU)];
gfn_t            1107 arch/x86/include/asm/kvm_host.h 	u64 (*get_mt_mask)(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio);
gfn_t            1162 arch/x86/include/asm/kvm_host.h 					   gfn_t offset, unsigned long mask);
gfn_t            1225 arch/x86/include/asm/kvm_host.h 	gfn_t gfn;
gfn_t            1278 arch/x86/include/asm/kvm_host.h 				   gfn_t gfn_offset, unsigned long mask);
gfn_t            1406 arch/x86/include/asm/kvm_host.h 			    gfn_t gfn, void *data, int offset, int len,
gfn_t            1432 arch/x86/include/asm/kvm_host.h int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn);
gfn_t            1590 arch/x86/include/asm/kvm_host.h extern bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn);
gfn_t              58 arch/x86/include/asm/kvm_page_track.h 				  struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t              61 arch/x86/include/asm/kvm_page_track.h 				     struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t              63 arch/x86/include/asm/kvm_page_track.h bool kvm_page_track_is_active(struct kvm_vcpu *vcpu, gfn_t gfn,
gfn_t             569 arch/x86/kvm/hyperv.c 	gfn_t msg_page_gfn;
gfn_t             473 arch/x86/kvm/mmu.c static gfn_t get_mmio_spte_gfn(u64 spte)
gfn_t             488 arch/x86/kvm/mmu.c static bool set_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn,
gfn_t             638 arch/x86/kvm/mmu.c static gfn_t pse36_gfn_delta(u32 gpte)
gfn_t            1155 arch/x86/kvm/mmu.c static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index)
gfn_t            1163 arch/x86/kvm/mmu.c static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn)
gfn_t            1181 arch/x86/kvm/mmu.c static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn,
gfn_t            1192 arch/x86/kvm/mmu.c 					    gfn_t gfn, int count)
gfn_t            1204 arch/x86/kvm/mmu.c void kvm_mmu_gfn_disallow_lpage(struct kvm_memory_slot *slot, gfn_t gfn)
gfn_t            1209 arch/x86/kvm/mmu.c void kvm_mmu_gfn_allow_lpage(struct kvm_memory_slot *slot, gfn_t gfn)
gfn_t            1218 arch/x86/kvm/mmu.c 	gfn_t gfn;
gfn_t            1248 arch/x86/kvm/mmu.c 	gfn_t gfn;
gfn_t            1268 arch/x86/kvm/mmu.c static bool __mmu_gfn_lpage_is_disallowed(gfn_t gfn, int level,
gfn_t            1281 arch/x86/kvm/mmu.c static bool mmu_gfn_lpage_is_disallowed(struct kvm_vcpu *vcpu, gfn_t gfn,
gfn_t            1290 arch/x86/kvm/mmu.c static int host_mapping_level(struct kvm_vcpu *vcpu, gfn_t gfn)
gfn_t            1319 arch/x86/kvm/mmu.c gfn_to_memslot_dirty_bitmap(struct kvm_vcpu *vcpu, gfn_t gfn,
gfn_t            1331 arch/x86/kvm/mmu.c static int mapping_level(struct kvm_vcpu *vcpu, gfn_t large_gfn,
gfn_t            1469 arch/x86/kvm/mmu.c static struct kvm_rmap_head *__gfn_to_rmap(gfn_t gfn, int level,
gfn_t            1478 arch/x86/kvm/mmu.c static struct kvm_rmap_head *gfn_to_rmap(struct kvm *kvm, gfn_t gfn,
gfn_t            1497 arch/x86/kvm/mmu.c static int rmap_add(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn)
gfn_t            1511 arch/x86/kvm/mmu.c 	gfn_t gfn;
gfn_t            1753 arch/x86/kvm/mmu.c 				     gfn_t gfn_offset, unsigned long mask)
gfn_t            1779 arch/x86/kvm/mmu.c 				     gfn_t gfn_offset, unsigned long mask)
gfn_t            1806 arch/x86/kvm/mmu.c 				gfn_t gfn_offset, unsigned long mask)
gfn_t            1870 arch/x86/kvm/mmu.c 			   struct kvm_memory_slot *slot, gfn_t gfn, int level,
gfn_t            1877 arch/x86/kvm/mmu.c 			     struct kvm_memory_slot *slot, gfn_t gfn, int level,
gfn_t            1925 arch/x86/kvm/mmu.c 	gfn_t start_gfn;
gfn_t            1926 arch/x86/kvm/mmu.c 	gfn_t end_gfn;
gfn_t            1931 arch/x86/kvm/mmu.c 	gfn_t gfn;
gfn_t            1952 arch/x86/kvm/mmu.c 		    int end_level, gfn_t start_gfn, gfn_t end_gfn)
gfn_t            1997 arch/x86/kvm/mmu.c 					       gfn_t gfn,
gfn_t            2011 arch/x86/kvm/mmu.c 			gfn_t gfn_start, gfn_end;
gfn_t            2042 arch/x86/kvm/mmu.c 					 gfn_t gfn, int level,
gfn_t            2059 arch/x86/kvm/mmu.c 			 struct kvm_memory_slot *slot, gfn_t gfn, int level,
gfn_t            2074 arch/x86/kvm/mmu.c 			      struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t            2088 arch/x86/kvm/mmu.c static void rmap_recycle(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn)
gfn_t            2151 arch/x86/kvm/mmu.c static unsigned kvm_page_table_hashfn(gfn_t gfn)
gfn_t            2414 arch/x86/kvm/mmu.c static bool kvm_sync_pages(struct kvm_vcpu *vcpu, gfn_t gfn,
gfn_t            2551 arch/x86/kvm/mmu.c 					     gfn_t gfn,
gfn_t            2932 arch/x86/kvm/mmu.c int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn)
gfn_t            2963 arch/x86/kvm/mmu.c static bool mmu_need_write_protect(struct kvm_vcpu *vcpu, gfn_t gfn,
gfn_t            3051 arch/x86/kvm/mmu.c 		    gfn_t gfn, kvm_pfn_t pfn, bool speculative,
gfn_t            3154 arch/x86/kvm/mmu.c 			int write_fault, int level, gfn_t gfn, kvm_pfn_t pfn,
gfn_t            3219 arch/x86/kvm/mmu.c static kvm_pfn_t pte_prefetch_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn,
gfn_t            3239 arch/x86/kvm/mmu.c 	gfn_t gfn;
gfn_t            3303 arch/x86/kvm/mmu.c 				       gfn_t gfn, kvm_pfn_t *pfnp, int *levelp)
gfn_t            3332 arch/x86/kvm/mmu.c 	gfn_t gfn = gpa >> PAGE_SHIFT;
gfn_t            3333 arch/x86/kvm/mmu.c 	gfn_t base_gfn = gfn;
gfn_t            3374 arch/x86/kvm/mmu.c static int kvm_handle_bad_page(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_pfn_t pfn)
gfn_t            3393 arch/x86/kvm/mmu.c 					gfn_t gfn, kvm_pfn_t *pfnp,
gfn_t            3431 arch/x86/kvm/mmu.c static bool handle_abnormal_pfn(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn,
gfn_t            3488 arch/x86/kvm/mmu.c 	gfn_t gfn;
gfn_t            3641 arch/x86/kvm/mmu.c static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn,
gfn_t            3647 arch/x86/kvm/mmu.c 			 gfn_t gfn, bool prefault)
gfn_t            3765 arch/x86/kvm/mmu.c static int mmu_check_root(struct kvm_vcpu *vcpu, gfn_t root_gfn)
gfn_t            3822 arch/x86/kvm/mmu.c 	gfn_t root_gfn, root_cr3;
gfn_t            4100 arch/x86/kvm/mmu.c 		gfn_t gfn = get_mmio_spte_gfn(spte);
gfn_t            4122 arch/x86/kvm/mmu.c 					 u32 error_code, gfn_t gfn)
gfn_t            4161 arch/x86/kvm/mmu.c 	gfn_t gfn = gpa >> PAGE_SHIFT;
gfn_t            4182 arch/x86/kvm/mmu.c 				   gfn_t gfn)
gfn_t            4195 arch/x86/kvm/mmu.c static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn,
gfn_t            4269 arch/x86/kvm/mmu.c check_hugepage_cache_consistency(struct kvm_vcpu *vcpu, gfn_t gfn, int level)
gfn_t            4285 arch/x86/kvm/mmu.c 	gfn_t gfn = gpa >> PAGE_SHIFT;
gfn_t            4463 arch/x86/kvm/mmu.c static bool sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn,
gfn_t            5435 arch/x86/kvm/mmu.c 	gfn_t gfn = gpa >> PAGE_SHIFT;
gfn_t            5693 arch/x86/kvm/mmu.c 			gfn_t start_gfn, gfn_t end_gfn, bool lock_flush_tlb)
gfn_t            5947 arch/x86/kvm/mmu.c void kvm_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end)
gfn_t            5957 arch/x86/kvm/mmu.c 			gfn_t start, end;
gfn_t             206 arch/x86/kvm/mmu.h void kvm_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end);
gfn_t             208 arch/x86/kvm/mmu.h void kvm_mmu_gfn_disallow_lpage(struct kvm_memory_slot *slot, gfn_t gfn);
gfn_t             209 arch/x86/kvm/mmu.h void kvm_mmu_gfn_allow_lpage(struct kvm_memory_slot *slot, gfn_t gfn);
gfn_t              96 arch/x86/kvm/mmu_audit.c 	gfn_t gfn;
gfn_t             133 arch/x86/kvm/mmu_audit.c 	gfn_t gfn;
gfn_t             205 arch/x86/kvm/mmutrace.h 	TP_PROTO(u64 *sptep, gfn_t gfn, unsigned access, unsigned int gen),
gfn_t             210 arch/x86/kvm/mmutrace.h 		__field(gfn_t, gfn)
gfn_t             228 arch/x86/kvm/mmutrace.h 	TP_PROTO(u64 addr, gfn_t gfn, unsigned access),
gfn_t             233 arch/x86/kvm/mmutrace.h 		__field(gfn_t, gfn)
gfn_t             331 arch/x86/kvm/mmutrace.h 	TP_PROTO(int level, gfn_t gfn, u64 *sptep),
gfn_t             313 arch/x86/kvm/mtrr.c 	gfn_t start, end;
gfn_t             619 arch/x86/kvm/mtrr.c u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn)
gfn_t             695 arch/x86/kvm/mtrr.c bool kvm_mtrr_check_gfn_range_consistency(struct kvm_vcpu *vcpu, gfn_t gfn,
gfn_t              63 arch/x86/kvm/page_track.c static void update_gfn_track(struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t              91 arch/x86/kvm/page_track.c 				  struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t             126 arch/x86/kvm/page_track.c 				     struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t             145 arch/x86/kvm/page_track.c bool kvm_page_track_is_active(struct kvm_vcpu *vcpu, gfn_t gfn,
gfn_t              87 arch/x86/kvm/paging_tmpl.h 	gfn_t table_gfn[PT_MAX_FULL_LEVELS];
gfn_t              95 arch/x86/kvm/paging_tmpl.h 	gfn_t gfn;
gfn_t              99 arch/x86/kvm/paging_tmpl.h static gfn_t gpte_to_gfn_lvl(pt_element_t gpte, int lvl)
gfn_t             228 arch/x86/kvm/paging_tmpl.h 	gfn_t table_gfn;
gfn_t             303 arch/x86/kvm/paging_tmpl.h 	gfn_t table_gfn;
gfn_t             316 arch/x86/kvm/paging_tmpl.h 	gfn_t gfn;
gfn_t             348 arch/x86/kvm/paging_tmpl.h 		gfn_t real_gfn;
gfn_t             520 arch/x86/kvm/paging_tmpl.h 	gfn_t gfn;
gfn_t             624 arch/x86/kvm/paging_tmpl.h 	gfn_t gfn, base_gfn;
gfn_t             646 arch/x86/kvm/paging_tmpl.h 		gfn_t table_gfn;
gfn_t             737 arch/x86/kvm/paging_tmpl.h 	gfn_t mask = ~(KVM_PAGES_PER_HPAGE(walker->level) - 1);
gfn_t             745 arch/x86/kvm/paging_tmpl.h 		gfn_t gfn = walker->gfn ^ walker->table_gfn[level - 1];
gfn_t            1023 arch/x86/kvm/paging_tmpl.h 		gfn_t gfn;
gfn_t            5938 arch/x86/kvm/svm.c static u64 svm_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio)
gfn_t            3388 arch/x86/kvm/vmx/nested.c 	gfn_t gfn;
gfn_t            3428 arch/x86/kvm/vmx/vmx.c 	gfn_t fn;
gfn_t            6863 arch/x86/kvm/vmx/vmx.c static u64 vmx_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio)
gfn_t            7313 arch/x86/kvm/vmx/vmx.c 					   gfn_t offset, unsigned long mask)
gfn_t             664 arch/x86/kvm/x86.c 			    gfn_t ngfn, void *data, int offset, int len,
gfn_t             668 arch/x86/kvm/x86.c 	gfn_t real_gfn;
gfn_t             682 arch/x86/kvm/x86.c static int kvm_read_nested_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn,
gfn_t             700 arch/x86/kvm/x86.c 	gfn_t pdpt_gfn = cr3 >> PAGE_SHIFT;
gfn_t             738 arch/x86/kvm/x86.c 	gfn_t gfn;
gfn_t            10082 arch/x86/kvm/x86.c static inline u32 kvm_async_pf_hash_fn(gfn_t gfn)
gfn_t            10092 arch/x86/kvm/x86.c static void kvm_add_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn)
gfn_t            10102 arch/x86/kvm/x86.c static u32 kvm_async_pf_gfn_slot(struct kvm_vcpu *vcpu, gfn_t gfn)
gfn_t            10115 arch/x86/kvm/x86.c bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn)
gfn_t            10120 arch/x86/kvm/x86.c static void kvm_del_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn)
gfn_t             187 arch/x86/kvm/x86.h 					gva_t gva, gfn_t gfn, unsigned access)
gfn_t             282 arch/x86/kvm/x86.h u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn);
gfn_t             286 arch/x86/kvm/x86.h bool kvm_mtrr_check_gfn_range_consistency(struct kvm_vcpu *vcpu, gfn_t gfn,
gfn_t              87 drivers/gpu/drm/i915/gvt/kvmgt.c 	gfn_t gfn;
gfn_t             105 drivers/gpu/drm/i915/gvt/kvmgt.c 	gfn_t gfn;
gfn_t             238 drivers/gpu/drm/i915/gvt/kvmgt.c static struct gvt_dma *__gvt_cache_find_gfn(struct intel_vgpu *vgpu, gfn_t gfn)
gfn_t             256 drivers/gpu/drm/i915/gvt/kvmgt.c static int __gvt_cache_add(struct intel_vgpu *vgpu, gfn_t gfn,
gfn_t             359 drivers/gpu/drm/i915/gvt/kvmgt.c __kvmgt_protect_table_find(struct kvmgt_guest_info *info, gfn_t gfn)
gfn_t             374 drivers/gpu/drm/i915/gvt/kvmgt.c 				gfn_t gfn)
gfn_t             382 drivers/gpu/drm/i915/gvt/kvmgt.c static void kvmgt_protect_table_add(struct kvmgt_guest_info *info, gfn_t gfn)
gfn_t             398 drivers/gpu/drm/i915/gvt/kvmgt.c 				gfn_t gfn)
gfn_t            1719 drivers/gpu/drm/i915/gvt/kvmgt.c 	gfn_t gfn;
gfn_t             344 include/linux/kvm_host.h 	gfn_t base_gfn;
gfn_t             702 include/linux/kvm_host.h int gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t             705 include/linux/kvm_host.h struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn);
gfn_t             706 include/linux/kvm_host.h unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn);
gfn_t             707 include/linux/kvm_host.h unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable);
gfn_t             708 include/linux/kvm_host.h unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
gfn_t             709 include/linux/kvm_host.h unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t             715 include/linux/kvm_host.h kvm_pfn_t gfn_to_pfn_atomic(struct kvm *kvm, gfn_t gfn);
gfn_t             716 include/linux/kvm_host.h kvm_pfn_t gfn_to_pfn(struct kvm *kvm, gfn_t gfn);
gfn_t             717 include/linux/kvm_host.h kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault,
gfn_t             719 include/linux/kvm_host.h kvm_pfn_t gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn);
gfn_t             720 include/linux/kvm_host.h kvm_pfn_t gfn_to_pfn_memslot_atomic(struct kvm_memory_slot *slot, gfn_t gfn);
gfn_t             721 include/linux/kvm_host.h kvm_pfn_t __gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t             732 include/linux/kvm_host.h int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset,
gfn_t             739 include/linux/kvm_host.h int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn, const void *data,
gfn_t             750 include/linux/kvm_host.h int kvm_clear_guest_page(struct kvm *kvm, gfn_t gfn, int offset, int len);
gfn_t             752 include/linux/kvm_host.h struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn);
gfn_t             753 include/linux/kvm_host.h bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn);
gfn_t             754 include/linux/kvm_host.h unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn);
gfn_t             755 include/linux/kvm_host.h void mark_page_dirty(struct kvm *kvm, gfn_t gfn);
gfn_t             758 include/linux/kvm_host.h struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn);
gfn_t             759 include/linux/kvm_host.h kvm_pfn_t kvm_vcpu_gfn_to_pfn_atomic(struct kvm_vcpu *vcpu, gfn_t gfn);
gfn_t             760 include/linux/kvm_host.h kvm_pfn_t kvm_vcpu_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn);
gfn_t             762 include/linux/kvm_host.h int kvm_map_gfn(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map,
gfn_t             764 include/linux/kvm_host.h struct page *kvm_vcpu_gfn_to_page(struct kvm_vcpu *vcpu, gfn_t gfn);
gfn_t             768 include/linux/kvm_host.h unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn);
gfn_t             769 include/linux/kvm_host.h unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable);
gfn_t             770 include/linux/kvm_host.h int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, int offset,
gfn_t             776 include/linux/kvm_host.h int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, const void *data,
gfn_t             780 include/linux/kvm_host.h void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn);
gfn_t             818 include/linux/kvm_host.h 					gfn_t gfn_offset,
gfn_t            1011 include/linux/kvm_host.h search_memslots(struct kvm_memslots *slots, gfn_t gfn)
gfn_t            1040 include/linux/kvm_host.h __gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn)
gfn_t            1046 include/linux/kvm_host.h __gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn)
gfn_t            1051 include/linux/kvm_host.h static inline int memslot_id(struct kvm *kvm, gfn_t gfn)
gfn_t            1056 include/linux/kvm_host.h static inline gfn_t
gfn_t            1059 include/linux/kvm_host.h 	gfn_t gfn_offset = (hva - slot->userspace_addr) >> PAGE_SHIFT;
gfn_t            1064 include/linux/kvm_host.h static inline gpa_t gfn_to_gpa(gfn_t gfn)
gfn_t            1069 include/linux/kvm_host.h static inline gfn_t gpa_to_gfn(gpa_t gpa)
gfn_t            1071 include/linux/kvm_host.h 	return (gfn_t)(gpa >> PAGE_SHIFT);
gfn_t              54 include/linux/kvm_types.h 	gfn_t gfn;
gfn_t            1379 virt/kvm/arm/mmu.c 	gfn_t gfn = *ipap >> PAGE_SHIFT;
gfn_t            1566 virt/kvm/arm/mmu.c 		gfn_t gfn_offset, unsigned long mask)
gfn_t            1584 virt/kvm/arm/mmu.c 		gfn_t gfn_offset, unsigned long mask)
gfn_t            1677 virt/kvm/arm/mmu.c 	gfn_t gfn = fault_ipa >> PAGE_SHIFT;
gfn_t            1915 virt/kvm/arm/mmu.c 	gfn_t gfn;
gfn_t            2030 virt/kvm/arm/mmu.c 		gfn_t gpa;
gfn_t             904 virt/kvm/arm/vgic/vgic-its.c 	gfn_t gfn;
gfn_t             147 virt/kvm/kvm_main.c static void mark_page_dirty_in_slot(struct kvm_memory_slot *memslot, gfn_t gfn);
gfn_t             995 virt/kvm/kvm_main.c 	gfn_t base_gfn;
gfn_t            1272 virt/kvm/kvm_main.c 			gfn_t offset;
gfn_t            1307 virt/kvm/kvm_main.c 	gfn_t offset;
gfn_t            1380 virt/kvm/kvm_main.c struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn)
gfn_t            1386 virt/kvm/kvm_main.c struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn)
gfn_t            1391 virt/kvm/kvm_main.c bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn)
gfn_t            1403 virt/kvm/kvm_main.c unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn)
gfn_t            1432 virt/kvm/kvm_main.c static unsigned long __gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t            1433 virt/kvm/kvm_main.c 				       gfn_t *nr_pages, bool write)
gfn_t            1447 virt/kvm/kvm_main.c static unsigned long gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t            1448 virt/kvm/kvm_main.c 				     gfn_t *nr_pages)
gfn_t            1454 virt/kvm/kvm_main.c 					gfn_t gfn)
gfn_t            1460 virt/kvm/kvm_main.c unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn)
gfn_t            1466 virt/kvm/kvm_main.c unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn)
gfn_t            1481 virt/kvm/kvm_main.c 				      gfn_t gfn, bool *writable)
gfn_t            1491 virt/kvm/kvm_main.c unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable)
gfn_t            1498 virt/kvm/kvm_main.c unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable)
gfn_t            1705 virt/kvm/kvm_main.c kvm_pfn_t __gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t            1734 virt/kvm/kvm_main.c kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault,
gfn_t            1742 virt/kvm/kvm_main.c kvm_pfn_t gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn)
gfn_t            1748 virt/kvm/kvm_main.c kvm_pfn_t gfn_to_pfn_memslot_atomic(struct kvm_memory_slot *slot, gfn_t gfn)
gfn_t            1754 virt/kvm/kvm_main.c kvm_pfn_t gfn_to_pfn_atomic(struct kvm *kvm, gfn_t gfn)
gfn_t            1760 virt/kvm/kvm_main.c kvm_pfn_t kvm_vcpu_gfn_to_pfn_atomic(struct kvm_vcpu *vcpu, gfn_t gfn)
gfn_t            1766 virt/kvm/kvm_main.c kvm_pfn_t gfn_to_pfn(struct kvm *kvm, gfn_t gfn)
gfn_t            1772 virt/kvm/kvm_main.c kvm_pfn_t kvm_vcpu_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn)
gfn_t            1778 virt/kvm/kvm_main.c int gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t            1782 virt/kvm/kvm_main.c 	gfn_t entry = 0;
gfn_t            1808 virt/kvm/kvm_main.c struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn)
gfn_t            1832 virt/kvm/kvm_main.c static void kvm_cache_gfn_to_pfn(struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t            1843 virt/kvm/kvm_main.c static int __kvm_map_gfn(struct kvm_memslots *slots, gfn_t gfn,
gfn_t            1898 virt/kvm/kvm_main.c int kvm_map_gfn(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map,
gfn_t            1906 virt/kvm/kvm_main.c int kvm_vcpu_map(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map)
gfn_t            1965 virt/kvm/kvm_main.c struct page *kvm_vcpu_gfn_to_page(struct kvm_vcpu *vcpu, gfn_t gfn)
gfn_t            2037 virt/kvm/kvm_main.c static int __kvm_read_guest_page(struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t            2052 virt/kvm/kvm_main.c int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset,
gfn_t            2061 virt/kvm/kvm_main.c int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data,
gfn_t            2072 virt/kvm/kvm_main.c 	gfn_t gfn = gpa >> PAGE_SHIFT;
gfn_t            2092 virt/kvm/kvm_main.c 	gfn_t gfn = gpa >> PAGE_SHIFT;
gfn_t            2110 virt/kvm/kvm_main.c static int __kvm_read_guest_atomic(struct kvm_memory_slot *slot, gfn_t gfn,
gfn_t            2130 virt/kvm/kvm_main.c 	gfn_t gfn = gpa >> PAGE_SHIFT;
gfn_t            2141 virt/kvm/kvm_main.c 	gfn_t gfn = gpa >> PAGE_SHIFT;
gfn_t            2149 virt/kvm/kvm_main.c static int __kvm_write_guest_page(struct kvm_memory_slot *memslot, gfn_t gfn,
gfn_t            2165 virt/kvm/kvm_main.c int kvm_write_guest_page(struct kvm *kvm, gfn_t gfn,
gfn_t            2174 virt/kvm/kvm_main.c int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn,
gfn_t            2186 virt/kvm/kvm_main.c 	gfn_t gfn = gpa >> PAGE_SHIFT;
gfn_t            2207 virt/kvm/kvm_main.c 	gfn_t gfn = gpa >> PAGE_SHIFT;
gfn_t            2230 virt/kvm/kvm_main.c 	gfn_t start_gfn = gpa >> PAGE_SHIFT;
gfn_t            2231 virt/kvm/kvm_main.c 	gfn_t end_gfn = (gpa + len - 1) >> PAGE_SHIFT;
gfn_t            2232 virt/kvm/kvm_main.c 	gfn_t nr_pages_needed = end_gfn - start_gfn + 1;
gfn_t            2233 virt/kvm/kvm_main.c 	gfn_t nr_pages_avail;
gfn_t            2331 virt/kvm/kvm_main.c int kvm_clear_guest_page(struct kvm *kvm, gfn_t gfn, int offset, int len)
gfn_t            2341 virt/kvm/kvm_main.c 	gfn_t gfn = gpa >> PAGE_SHIFT;
gfn_t            2359 virt/kvm/kvm_main.c 				    gfn_t gfn)
gfn_t            2368 virt/kvm/kvm_main.c void mark_page_dirty(struct kvm *kvm, gfn_t gfn)
gfn_t            2377 virt/kvm/kvm_main.c void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn)