/linux-4.1.27/include/linux/ |
D | kvm_host.h | 285 gfn_t base_gfn; 538 int gfn_to_page_many_atomic(struct kvm *kvm, gfn_t gfn, struct page **pages, 541 struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn); 542 unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn); 543 unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable); 544 unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn); 545 unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn, 551 pfn_t gfn_to_pfn_atomic(struct kvm *kvm, gfn_t gfn); 552 pfn_t gfn_to_pfn_async(struct kvm *kvm, gfn_t gfn, bool *async, 554 pfn_t gfn_to_pfn(struct kvm *kvm, gfn_t gfn); [all …]
|
D | kvm_types.h | 49 typedef u64 gfn_t; typedef
|
/linux-4.1.27/arch/x86/kvm/ |
D | iommu.c | 44 gfn_t base_gfn, unsigned long npages); 46 static pfn_t kvm_pin_pages(struct kvm_memory_slot *slot, gfn_t gfn, in kvm_pin_pages() 49 gfn_t end_gfn; in kvm_pin_pages() 75 gfn_t gfn, end_gfn; in kvm_iommu_map_pages() 272 gfn_t base_gfn, unsigned long npages) in kvm_iommu_put_pages() 275 gfn_t end_gfn, gfn; in kvm_iommu_put_pages()
|
D | mmu.c | 249 static gfn_t get_mmio_spte_gfn(u64 spte) in get_mmio_spte_gfn() 261 static bool set_mmio_spte(struct kvm *kvm, u64 *sptep, gfn_t gfn, in set_mmio_spte() 333 static gfn_t pse36_gfn_delta(u32 gpte) in pse36_gfn_delta() 754 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_gfn() 762 static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn) in kvm_mmu_page_set_gfn() 774 static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn, in lpage_info_slot() 784 static void account_shadowed(struct kvm *kvm, gfn_t gfn) in account_shadowed() 799 static void unaccount_shadowed(struct kvm *kvm, gfn_t gfn) in unaccount_shadowed() 816 gfn_t gfn, in has_wrprotected_page() 831 static int host_mapping_level(struct kvm *kvm, gfn_t gfn) in host_mapping_level() [all …]
|
D | paging_tmpl.h | 97 gfn_t table_gfn[PT_MAX_FULL_LEVELS]; 105 gfn_t gfn; 109 static gfn_t gpte_to_gfn_lvl(pt_element_t gpte, int lvl) in gpte_to_gfn_lvl() 215 gfn_t table_gfn; in FNAME() 275 gfn_t table_gfn; in FNAME() 284 gfn_t gfn; in FNAME() 308 gfn_t real_gfn; in FNAME() 467 gfn_t gfn; in FNAME() 590 gfn_t table_gfn; in FNAME() 616 gfn_t direct_gfn; in FNAME() [all …]
|
D | mmutrace.h | 202 TP_PROTO(u64 *sptep, gfn_t gfn, unsigned access, unsigned int gen), 207 __field(gfn_t, gfn) 225 TP_PROTO(u64 addr, gfn_t gfn, unsigned access), 230 __field(gfn_t, gfn)
|
D | mmu_audit.c | 99 gfn_t gfn; in audit_mappings() 134 gfn_t gfn; in inspect_spte_has_rmap()
|
D | x86.h | 86 gva_t gva, gfn_t gfn, unsigned access) in vcpu_cache_mmio_info()
|
D | x86.c | 482 gfn_t ngfn, void *data, int offset, int len, in kvm_read_guest_page_mmu() 486 gfn_t real_gfn; in kvm_read_guest_page_mmu() 500 static int kvm_read_nested_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, in kvm_read_nested_guest_page() 512 gfn_t pdpt_gfn = cr3 >> PAGE_SHIFT; in load_pdptrs() 550 gfn_t gfn; in pdptrs_changed() 7833 static inline u32 kvm_async_pf_hash_fn(gfn_t gfn) in kvm_async_pf_hash_fn() 7843 static void kvm_add_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_add_async_pf_gfn() 7853 static u32 kvm_async_pf_gfn_slot(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_async_pf_gfn_slot() 7866 bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_find_async_pf_gfn() 7871 static void kvm_del_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_del_async_pf_gfn()
|
D | vmx.c | 4043 gfn_t fn; in init_rmode_tss() 8505 static u64 vmx_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio) in vmx_get_mt_mask() 10163 gfn_t offset, unsigned long mask) in vmx_enable_log_dirty_pt_masked()
|
D | svm.c | 4078 static u64 svm_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio) in svm_get_mt_mask()
|
/linux-4.1.27/arch/x86/include/asm/ |
D | kvm_host.h | 77 static inline gfn_t gfn_to_index(gfn_t gfn, gfn_t base_gfn, int level) in gfn_to_index() 222 gfn_t gfn; 227 gfn_t *gfns; 493 gfn_t mmio_gfn; 511 gfn_t gfns[roundup_pow_of_two(ASYNC_PF_PER_VCPU)]; 783 u64 (*get_mt_mask)(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio); 838 gfn_t offset, unsigned long mask); 843 gfn_t gfn; 883 gfn_t gfn_offset, unsigned long mask); 893 u8 kvm_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn); [all …]
|
/linux-4.1.27/virt/kvm/ |
D | kvm_main.c | 107 struct kvm_memory_slot *memslot, gfn_t gfn); 768 gfn_t base_gfn; in __kvm_set_memory_region() 1048 gfn_t offset; in kvm_get_dirty_log_protect() 1089 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) in gfn_to_memslot() 1095 int kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn) in kvm_is_visible_gfn() 1107 unsigned long kvm_host_page_size(struct kvm *kvm, gfn_t gfn) in kvm_host_page_size() 1136 static unsigned long __gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, in __gfn_to_hva_many() 1137 gfn_t *nr_pages, bool write) in __gfn_to_hva_many() 1151 static unsigned long gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, in gfn_to_hva_many() 1152 gfn_t *nr_pages) in gfn_to_hva_many() [all …]
|
/linux-4.1.27/arch/mips/kvm/ |
D | tlb.c | 41 pfn_t (*kvm_mips_gfn_to_pfn)(struct kvm *kvm, gfn_t gfn); 144 static int kvm_mips_map_page(struct kvm *kvm, gfn_t gfn) in kvm_mips_map_page() 171 gfn_t gfn; in kvm_mips_translate_guest_kseg0_to_hpa() 264 gfn_t gfn; in kvm_mips_handle_kseg0_tlb_fault()
|
D | emulate.c | 1538 gfn_t gfn; in kvm_mips_sync_icache()
|
/linux-4.1.27/arch/powerpc/kvm/ |
D | e500_mmu_host.c | 325 u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, in kvmppc_e500_shadow_map() 562 u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, in kvmppc_e500_tlb1_map() 614 gfn_t gfn = gpaddr >> PAGE_SHIFT; in kvmppc_mmu_map()
|
D | book3s_64_mmu_hv.c | 680 gfn_t gfn, gfn_end; in kvm_handle_hva_range() 695 gfn_t gfn_offset = gfn - memslot->base_gfn; in kvm_handle_hva_range()
|
D | book3s.c | 371 gfn_t gfn = gpa >> PAGE_SHIFT; in kvmppc_gpa_to_pfn()
|
D | booke.c | 1228 gfn_t gfn; in kvmppc_handle_exit() 1285 gfn_t gfn; in kvmppc_handle_exit()
|
D | book3s_pr.c | 262 gfn_t gfn, gfn_end; in do_kvm_unmap_hva()
|
/linux-4.1.27/arch/arm/kvm/ |
D | mmu.c | 1003 gfn_t gfn = *ipap >> PAGE_SHIFT; in transparent_hugepage_adjust() 1185 gfn_t gfn_offset, unsigned long mask) in kvm_mmu_write_protect_pt_masked() 1203 gfn_t gfn_offset, unsigned long mask) in kvm_arch_mmu_enable_log_dirty_pt_masked() 1221 gfn_t gfn = fault_ipa >> PAGE_SHIFT; in user_mem_abort() 1403 gfn_t gfn; in kvm_handle_guest_abort() 1481 gfn_t gfn, gfn_end; in handle_hva_to_gpa()
|
/linux-4.1.27/arch/mips/include/asm/ |
D | kvm_host.h | 103 extern pfn_t(*kvm_mips_gfn_to_pfn) (struct kvm *kvm, gfn_t gfn);
|
/linux-4.1.27/arch/s390/kvm/ |
D | kvm-s390.c | 214 gfn_t cur_gfn, last_gfn; in kvm_s390_sync_dirty_log()
|