hva               270 arch/arm/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte);
hva               275 arch/arm/include/asm/kvm_host.h int kvm_test_age_hva(struct kvm *kvm, unsigned long hva);
hva               431 arch/arm64/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte);
hva               433 arch/arm64/include/asm/kvm_host.h int kvm_test_age_hva(struct kvm *kvm, unsigned long hva);
hva               943 arch/mips/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte);
hva               945 arch/mips/include/asm/kvm_host.h int kvm_test_age_hva(struct kvm *kvm, unsigned long hva);
hva               554 arch/mips/kvm/mmu.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte)
hva               556 arch/mips/kvm/mmu.c 	unsigned long end = hva + PAGE_SIZE;
hva               559 arch/mips/kvm/mmu.c 	ret = handle_hva_to_gpa(kvm, hva, end, &kvm_set_spte_handler, &pte);
hva               587 arch/mips/kvm/mmu.c int kvm_test_age_hva(struct kvm *kvm, unsigned long hva)
hva               589 arch/mips/kvm/mmu.c 	return handle_hva_to_gpa(kvm, hva, hva, kvm_test_age_hva_handler, NULL);
hva                63 arch/powerpc/include/asm/kvm_host.h extern int kvm_test_age_hva(struct kvm *kvm, unsigned long hva);
hva                64 arch/powerpc/include/asm/kvm_host.h extern int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte);
hva               292 arch/powerpc/include/asm/kvm_ppc.h 	int (*test_age_hva)(struct kvm *kvm, unsigned long hva);
hva               293 arch/powerpc/include/asm/kvm_ppc.h 	void (*set_spte_hva)(struct kvm *kvm, unsigned long hva, pte_t pte);
hva               880 arch/powerpc/kvm/book3s.c int kvm_test_age_hva(struct kvm *kvm, unsigned long hva)
hva               882 arch/powerpc/kvm/book3s.c 	return kvm->arch.kvm_ops->test_age_hva(kvm, hva);
hva               885 arch/powerpc/kvm/book3s.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte)
hva               887 arch/powerpc/kvm/book3s.c 	kvm->arch.kvm_ops->set_spte_hva(kvm, hva, pte);
hva                16 arch/powerpc/kvm/book3s.h extern int kvm_test_age_hva_hv(struct kvm *kvm, unsigned long hva);
hva                17 arch/powerpc/kvm/book3s.h extern void kvm_set_spte_hva_hv(struct kvm *kvm, unsigned long hva, pte_t pte);
hva               500 arch/powerpc/kvm/book3s_64_mmu_hv.c 	unsigned long gpa, gfn, hva, pfn;
hva               591 arch/powerpc/kvm/book3s_64_mmu_hv.c 	hva = gfn_to_hva_memslot(memslot, gfn);
hva               592 arch/powerpc/kvm/book3s_64_mmu_hv.c 	npages = get_user_pages_fast(hva, 1, writing ? FOLL_WRITE : 0, pages);
hva               596 arch/powerpc/kvm/book3s_64_mmu_hv.c 		vma = find_vma(current->mm, hva);
hva               597 arch/powerpc/kvm/book3s_64_mmu_hv.c 		if (vma && vma->vm_start <= hva && hva + psize <= vma->vm_end &&
hva               600 arch/powerpc/kvm/book3s_64_mmu_hv.c 				((hva - vma->vm_start) >> PAGE_SHIFT);
hva               625 arch/powerpc/kvm/book3s_64_mmu_hv.c 						   hva, NULL, NULL);
hva               805 arch/powerpc/kvm/book3s_64_mmu_hv.c static int kvm_handle_hva(struct kvm *kvm, unsigned long hva,
hva               808 arch/powerpc/kvm/book3s_64_mmu_hv.c 	return kvm_handle_hva_range(kvm, hva, hva + 1, handler);
hva              1023 arch/powerpc/kvm/book3s_64_mmu_hv.c int kvm_test_age_hva_hv(struct kvm *kvm, unsigned long hva)
hva              1028 arch/powerpc/kvm/book3s_64_mmu_hv.c 	return kvm_handle_hva(kvm, hva, handler);
hva              1031 arch/powerpc/kvm/book3s_64_mmu_hv.c void kvm_set_spte_hva_hv(struct kvm *kvm, unsigned long hva, pte_t pte)
hva              1036 arch/powerpc/kvm/book3s_64_mmu_hv.c 	kvm_handle_hva(kvm, hva, handler);
hva              1177 arch/powerpc/kvm/book3s_64_mmu_hv.c 	unsigned long hva, offset;
hva              1184 arch/powerpc/kvm/book3s_64_mmu_hv.c 	hva = gfn_to_hva_memslot(memslot, gfn);
hva              1185 arch/powerpc/kvm/book3s_64_mmu_hv.c 	npages = get_user_pages_fast(hva, 1, FOLL_WRITE, pages);
hva               775 arch/powerpc/kvm/book3s_64_mmu_radix.c 	unsigned long hva, gfn = gpa >> PAGE_SHIFT;
hva               793 arch/powerpc/kvm/book3s_64_mmu_radix.c 	hva = gfn_to_hva_memslot(memslot, gfn);
hva               794 arch/powerpc/kvm/book3s_64_mmu_radix.c 	if (!kvm_ro && __get_user_pages_fast(hva, 1, 1, &page) == 1) {
hva               817 arch/powerpc/kvm/book3s_64_mmu_radix.c 	ptep = __find_linux_pte(vcpu->arch.pgdir, hva, NULL, &shift);
hva               837 arch/powerpc/kvm/book3s_64_mmu_radix.c 	    (hva & (PUD_SIZE - PAGE_SIZE))) {
hva               841 arch/powerpc/kvm/book3s_64_mmu_radix.c 		   (hva & (PMD_SIZE - PAGE_SIZE))) {
hva               852 arch/powerpc/kvm/book3s_64_mmu_radix.c 			pte = __pte(pte_val(pte) | (hva & rpnmask));
hva              4580 arch/powerpc/kvm/book3s_hv.c 	unsigned long hva;
hva              4617 arch/powerpc/kvm/book3s_hv.c 	hva = memslot->userspace_addr;
hva              4619 arch/powerpc/kvm/book3s_hv.c 	vma = find_vma(current->mm, hva);
hva              4620 arch/powerpc/kvm/book3s_hv.c 	if (!vma || vma->vm_start > hva || (vma->vm_flags & VM_IO))
hva               202 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	unsigned long slot_fn, hva;
hva               250 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	hva = __gfn_to_hva_memslot(memslot, gfn);
hva               261 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	ptep = __find_linux_pte(pgdir, hva, NULL, &hpage_shift);
hva               286 arch/powerpc/kvm/book3s_hv_rm_mmu.c 			pa |= hva & (host_pte_size - 1);
hva               897 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	unsigned long gfn, hva, pa, psize = PAGE_SHIFT;
hva               908 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	hva = __gfn_to_hva_memslot(memslot, gfn);
hva               911 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	ptep = __find_linux_pte(vcpu->arch.pgdir, hva, NULL, &shift);
hva               922 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	pa |= hva & (psize - 1);
hva               437 arch/powerpc/kvm/book3s_pr.c static int kvm_test_age_hva_pr(struct kvm *kvm, unsigned long hva)
hva               443 arch/powerpc/kvm/book3s_pr.c static void kvm_set_spte_hva_pr(struct kvm *kvm, unsigned long hva, pte_t pte)
hva               446 arch/powerpc/kvm/book3s_pr.c 	do_kvm_unmap_hva(kvm, hva, hva + PAGE_SIZE);
hva               329 arch/powerpc/kvm/e500_mmu_host.c 	unsigned long hva;
hva               354 arch/powerpc/kvm/e500_mmu_host.c 	hva = gfn_to_hva_memslot(slot, gfn);
hva               360 arch/powerpc/kvm/e500_mmu_host.c 		vma = find_vma(current->mm, hva);
hva               361 arch/powerpc/kvm/e500_mmu_host.c 		if (vma && hva >= vma->vm_start &&
hva               379 arch/powerpc/kvm/e500_mmu_host.c 			pfn = start + ((hva - vma->vm_start) >> PAGE_SHIFT);
hva               424 arch/powerpc/kvm/e500_mmu_host.c 		} else if (vma && hva >= vma->vm_start &&
hva               477 arch/powerpc/kvm/e500_mmu_host.c 	ptep = find_linux_pte(pgdir, hva, NULL, NULL);
hva               724 arch/powerpc/kvm/e500_mmu_host.c static int kvm_unmap_hva(struct kvm *kvm, unsigned long hva)
hva               726 arch/powerpc/kvm/e500_mmu_host.c 	trace_kvm_unmap_hva(hva);
hva               751 arch/powerpc/kvm/e500_mmu_host.c int kvm_test_age_hva(struct kvm *kvm, unsigned long hva)
hva               757 arch/powerpc/kvm/e500_mmu_host.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte)
hva               760 arch/powerpc/kvm/e500_mmu_host.c 	kvm_unmap_hva(kvm, hva);
hva                73 arch/powerpc/kvm/trace_booke.h 	TP_PROTO(unsigned long hva),
hva                74 arch/powerpc/kvm/trace_booke.h 	TP_ARGS(hva),
hva                77 arch/powerpc/kvm/trace_booke.h 		__field(	unsigned long,	hva		)
hva                81 arch/powerpc/kvm/trace_booke.h 		__entry->hva		= hva;
hva                84 arch/powerpc/kvm/trace_booke.h 	TP_printk("unmap hva 0x%lx\n", __entry->hva)
hva              1166 arch/s390/include/asm/pgtable.h int get_pgste(struct mm_struct *mm, unsigned long hva, unsigned long *pgstep);
hva              1167 arch/s390/include/asm/pgtable.h int pgste_perform_essa(struct mm_struct *mm, unsigned long hva, int orc,
hva              1793 arch/s390/kvm/kvm-s390.c 	uint64_t hva;
hva              1814 arch/s390/kvm/kvm-s390.c 		hva = gfn_to_hva(kvm, args->start_gfn + i);
hva              1815 arch/s390/kvm/kvm-s390.c 		if (kvm_is_error_hva(hva)) {
hva              1820 arch/s390/kvm/kvm-s390.c 		r = get_guest_storage_key(current->mm, hva, &keys[i]);
hva              1841 arch/s390/kvm/kvm-s390.c 	uint64_t hva;
hva              1873 arch/s390/kvm/kvm-s390.c 		hva = gfn_to_hva(kvm, args->start_gfn + i);
hva              1874 arch/s390/kvm/kvm-s390.c 		if (kvm_is_error_hva(hva)) {
hva              1885 arch/s390/kvm/kvm-s390.c 		r = set_guest_storage_key(current->mm, hva, keys[i], 0);
hva              1887 arch/s390/kvm/kvm-s390.c 			r = fixup_user_fault(current, current->mm, hva,
hva              1949 arch/s390/kvm/kvm-s390.c 	unsigned long pgstev, hva, cur_gfn = args->start_gfn;
hva              1953 arch/s390/kvm/kvm-s390.c 		hva = gfn_to_hva(kvm, cur_gfn);
hva              1958 arch/s390/kvm/kvm-s390.c 		if (kvm_is_error_hva(hva))
hva              1960 arch/s390/kvm/kvm-s390.c 		if (get_pgste(kvm->mm, hva, &pgstev) < 0)
hva              1997 arch/s390/kvm/kvm-s390.c 	unsigned long mem_end, cur_gfn, next_gfn, hva, pgstev;
hva              2011 arch/s390/kvm/kvm-s390.c 		hva = gfn_to_hva(kvm, cur_gfn);
hva              2012 arch/s390/kvm/kvm-s390.c 		if (kvm_is_error_hva(hva))
hva              2017 arch/s390/kvm/kvm-s390.c 		if (get_pgste(kvm->mm, hva, &pgstev) < 0)
hva              2111 arch/s390/kvm/kvm-s390.c 	unsigned long hva, mask, pgstev, i;
hva              2142 arch/s390/kvm/kvm-s390.c 		hva = gfn_to_hva(kvm, args->start_gfn + i);
hva              2143 arch/s390/kvm/kvm-s390.c 		if (kvm_is_error_hva(hva)) {
hva              2151 arch/s390/kvm/kvm-s390.c 		set_pgste_bits(kvm->mm, hva, mask, pgstev);
hva              3667 arch/s390/kvm/kvm-s390.c 	hva_t hva;
hva              3685 arch/s390/kvm/kvm-s390.c 	hva = gfn_to_hva(vcpu->kvm, gpa_to_gfn(current->thread.gmap_addr));
hva              3686 arch/s390/kvm/kvm-s390.c 	hva += current->thread.gmap_addr & ~PAGE_MASK;
hva              3690 arch/s390/kvm/kvm-s390.c 	rc = kvm_setup_async_pf(vcpu, current->thread.gmap_addr, hva, &arch);
hva              1125 arch/s390/kvm/priv.c 	unsigned long gfn, hva, res, pgstev, ptev;
hva              1135 arch/s390/kvm/priv.c 	hva = gfn_to_hva(vcpu->kvm, gfn);
hva              1138 arch/s390/kvm/priv.c 	if (kvm_is_error_hva(hva))
hva              1141 arch/s390/kvm/priv.c 	nappended = pgste_perform_essa(vcpu->kvm->mm, hva, orc, &ptev, &pgstev);
hva              1438 arch/s390/kvm/priv.c 	unsigned long hva, gpa;
hva              1475 arch/s390/kvm/priv.c 	hva = gfn_to_hva_prot(vcpu->kvm, gpa_to_gfn(gpa), &writable);
hva              1476 arch/s390/kvm/priv.c 	if (kvm_is_error_hva(hva)) {
hva               970 arch/s390/mm/pgtable.c int pgste_perform_essa(struct mm_struct *mm, unsigned long hva, int orc,
hva               982 arch/s390/mm/pgtable.c 	ptep = get_locked_pte(mm, hva, &ptl);
hva              1071 arch/s390/mm/pgtable.c int set_pgste_bits(struct mm_struct *mm, unsigned long hva,
hva              1078 arch/s390/mm/pgtable.c 	ptep = get_locked_pte(mm, hva, &ptl);
hva              1100 arch/s390/mm/pgtable.c int get_pgste(struct mm_struct *mm, unsigned long hva, unsigned long *pgstep)
hva              1105 arch/s390/mm/pgtable.c 	ptep = get_locked_pte(mm, hva, &ptl);
hva              1558 arch/x86/include/asm/kvm_host.h int kvm_test_age_hva(struct kvm *kvm, unsigned long hva);
hva              1559 arch/x86/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte);
hva              2037 arch/x86/kvm/mmu.c static int kvm_handle_hva(struct kvm *kvm, unsigned long hva,
hva              2045 arch/x86/kvm/mmu.c 	return kvm_handle_hva_range(kvm, hva, hva + 1, data, handler);
hva              2053 arch/x86/kvm/mmu.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte)
hva              2055 arch/x86/kvm/mmu.c 	return kvm_handle_hva(kvm, hva, (unsigned long)&pte, kvm_set_pte_rmapp);
hva              2107 arch/x86/kvm/mmu.c int kvm_test_age_hva(struct kvm *kvm, unsigned long hva)
hva              2109 arch/x86/kvm/mmu.c 	return kvm_handle_hva(kvm, hva, 0, kvm_test_age_rmapp);
hva              3372 arch/x86/kvm/svm.c 	nested_vmcb = map.hva;
hva              3665 arch/x86/kvm/svm.c 	nested_vmcb = map.hva;
hva              3763 arch/x86/kvm/svm.c 	nested_vmcb = map.hva;
hva              3789 arch/x86/kvm/svm.c 	nested_vmcb = map.hva;
hva              6310 arch/x86/kvm/svm.c 		nested_vmcb = map.hva;
hva               577 arch/x86/kvm/vmx/nested.c 	msr_bitmap_l1 = (unsigned long *)map->hva;
hva               673 arch/x86/kvm/vmx/nested.c 	memcpy(shadow, map.hva, VMCS12_SIZE);
hva              1842 arch/x86/kvm/vmx/nested.c 		vmx->nested.hv_evmcs = vmx->nested.hv_evmcs_map.hva;
hva              2750 arch/x86/kvm/vmx/nested.c 	shadow = map.hva;
hva              2997 arch/x86/kvm/vmx/nested.c 				(struct pi_desc *)(((void *)map->hva) +
hva              3422 arch/x86/kvm/vmx/nested.c 		vapic_page = vmx->nested.virtual_apic_map.hva;
hva              4848 arch/x86/kvm/vmx/nested.c 		new_vmcs12 = map.hva;
hva              3754 arch/x86/kvm/vmx/vmx.c 	vapic_page = vmx->nested.virtual_apic_map.hva;
hva              2660 arch/x86/kvm/x86.c 	st = map.hva +
hva              3558 arch/x86/kvm/x86.c 	st = map.hva +
hva              5858 arch/x86/kvm/x86.c 	kaddr = map.hva + offset_in_page(gpa);
hva              9625 arch/x86/kvm/x86.c 	unsigned long hva;
hva              9642 arch/x86/kvm/x86.c 		hva = vm_mmap(NULL, 0, size, PROT_READ | PROT_WRITE,
hva              9644 arch/x86/kvm/x86.c 		if (IS_ERR((void *)hva))
hva              9645 arch/x86/kvm/x86.c 			return PTR_ERR((void *)hva);
hva              9650 arch/x86/kvm/x86.c 		hva = 0;
hva              9660 arch/x86/kvm/x86.c 		m.userspace_addr = hva;
hva               276 drivers/media/platform/sti/hva/hva-debugfs.c 	struct hva_dev *hva = s->private;
hva               278 drivers/media/platform/sti/hva/hva-debugfs.c 	seq_printf(s, "[%s]\n", hva->v4l2_dev.name);
hva               279 drivers/media/platform/sti/hva/hva-debugfs.c 	seq_printf(s, "registered as /dev/video%d\n", hva->vdev->num);
hva               286 drivers/media/platform/sti/hva/hva-debugfs.c 	struct hva_dev *hva = s->private;
hva               290 drivers/media/platform/sti/hva/hva-debugfs.c 		   hva->nb_of_encoders);
hva               292 drivers/media/platform/sti/hva/hva-debugfs.c 	while (hva->encoders[i]) {
hva               293 drivers/media/platform/sti/hva/hva-debugfs.c 		seq_printf(s, "|- %s: %4.4s => %4.4s\n", hva->encoders[i]->name,
hva               294 drivers/media/platform/sti/hva/hva-debugfs.c 			   (char *)&hva->encoders[i]->pixelformat,
hva               295 drivers/media/platform/sti/hva/hva-debugfs.c 			   (char *)&hva->encoders[i]->streamformat);
hva               304 drivers/media/platform/sti/hva/hva-debugfs.c 	struct hva_dev *hva = s->private;
hva               305 drivers/media/platform/sti/hva/hva-debugfs.c 	struct hva_ctx *last_ctx = &hva->dbg.last_ctx;
hva               321 drivers/media/platform/sti/hva/hva-debugfs.c 	struct hva_dev *hva = s->private;
hva               323 drivers/media/platform/sti/hva/hva-debugfs.c 	hva_hw_dump_regs(hva, s);
hva               329 drivers/media/platform/sti/hva/hva-debugfs.c 	debugfs_create_file(#name, 0444, hva->dbg.debugfs_entry, hva, \
hva               337 drivers/media/platform/sti/hva/hva-debugfs.c void hva_debugfs_create(struct hva_dev *hva)
hva               339 drivers/media/platform/sti/hva/hva-debugfs.c 	hva->dbg.debugfs_entry = debugfs_create_dir(HVA_NAME, NULL);
hva               340 drivers/media/platform/sti/hva/hva-debugfs.c 	if (!hva->dbg.debugfs_entry)
hva               358 drivers/media/platform/sti/hva/hva-debugfs.c 	hva_debugfs_remove(hva);
hva               361 drivers/media/platform/sti/hva/hva-debugfs.c void hva_debugfs_remove(struct hva_dev *hva)
hva               363 drivers/media/platform/sti/hva/hva-debugfs.c 	debugfs_remove_recursive(hva->dbg.debugfs_entry);
hva               364 drivers/media/platform/sti/hva/hva-debugfs.c 	hva->dbg.debugfs_entry = NULL;
hva               387 drivers/media/platform/sti/hva/hva-debugfs.c 	struct hva_dev *hva = ctx->hva_dev;
hva               394 drivers/media/platform/sti/hva/hva-debugfs.c 	snprintf(name, sizeof(name), "%d", hva->instance_id);
hva               397 drivers/media/platform/sti/hva/hva-debugfs.c 						     hva->dbg.debugfs_entry,
hva               403 drivers/media/platform/sti/hva/hva-debugfs.c 	struct hva_dev *hva = ctx->hva_dev;
hva               407 drivers/media/platform/sti/hva/hva-debugfs.c 		memcpy(&hva->dbg.last_ctx, ctx, sizeof(*ctx));
hva               590 drivers/media/platform/sti/hva/hva-h264.c 	struct hva_dev *hva = ctx_to_hdev(pctx);
hva               601 drivers/media/platform/sti/hva/hva-h264.c 	u32 addr_esram = hva->esram_addr;
hva               891 drivers/media/platform/sti/hva/hva-h264.c 	struct hva_dev *hva = ctx_to_hdev(pctx);
hva               903 drivers/media/platform/sti/hva/hva-h264.c 	if (hva->esram_size < size) {
hva               905 drivers/media/platform/sti/hva/hva-h264.c 			pctx->name, hva->esram_size, size);
hva               103 drivers/media/platform/sti/hva/hva-hw.c 	struct hva_dev *hva = data;
hva               106 drivers/media/platform/sti/hva/hva-hw.c 	hva->sts_reg = readl_relaxed(hva->regs + HVA_HIF_FIFO_STS);
hva               107 drivers/media/platform/sti/hva/hva-hw.c 	hva->sfl_reg = readl_relaxed(hva->regs + HVA_HIF_REG_SFL);
hva               110 drivers/media/platform/sti/hva/hva-hw.c 	writel_relaxed(0x1, hva->regs + HVA_HIF_REG_IT_ACK);
hva               117 drivers/media/platform/sti/hva/hva-hw.c 	struct hva_dev *hva = arg;
hva               118 drivers/media/platform/sti/hva/hva-hw.c 	struct device *dev = hva_to_dev(hva);
hva               119 drivers/media/platform/sti/hva/hva-hw.c 	u32 status = hva->sts_reg & 0xFF;
hva               124 drivers/media/platform/sti/hva/hva-hw.c 		HVA_PREFIX, __func__, hva->sts_reg & 0xFF, hva->sfl_reg & 0xF);
hva               130 drivers/media/platform/sti/hva/hva-hw.c 	ctx_id = (hva->sts_reg & 0xFF00) >> 8;
hva               138 drivers/media/platform/sti/hva/hva-hw.c 	ctx = hva->instances[ctx_id];
hva               195 drivers/media/platform/sti/hva/hva-hw.c 	complete(&hva->interrupt);
hva               202 drivers/media/platform/sti/hva/hva-hw.c 	struct hva_dev *hva = data;
hva               205 drivers/media/platform/sti/hva/hva-hw.c 	hva->sts_reg = readl_relaxed(hva->regs + HVA_HIF_FIFO_STS);
hva               206 drivers/media/platform/sti/hva/hva-hw.c 	hva->sfl_reg = readl_relaxed(hva->regs + HVA_HIF_REG_SFL);
hva               209 drivers/media/platform/sti/hva/hva-hw.c 	hva->lmi_err_reg = readl_relaxed(hva->regs + HVA_HIF_REG_LMI_ERR);
hva               210 drivers/media/platform/sti/hva/hva-hw.c 	hva->emi_err_reg = readl_relaxed(hva->regs + HVA_HIF_REG_EMI_ERR);
hva               211 drivers/media/platform/sti/hva/hva-hw.c 	hva->hec_mif_err_reg = readl_relaxed(hva->regs +
hva               215 drivers/media/platform/sti/hva/hva-hw.c 	writel_relaxed(0x1, hva->regs + HVA_HIF_REG_IT_ACK);
hva               222 drivers/media/platform/sti/hva/hva-hw.c 	struct hva_dev *hva = arg;
hva               223 drivers/media/platform/sti/hva/hva-hw.c 	struct device *dev = hva_to_dev(hva);
hva               228 drivers/media/platform/sti/hva/hva-hw.c 		HVA_PREFIX, hva->sts_reg & 0xFF, hva->sfl_reg & 0xF);
hva               234 drivers/media/platform/sti/hva/hva-hw.c 	ctx_id = (hva->sts_reg & 0xFF00) >> 8;
hva               241 drivers/media/platform/sti/hva/hva-hw.c 	ctx = hva->instances[ctx_id];
hva               245 drivers/media/platform/sti/hva/hva-hw.c 	if (hva->lmi_err_reg) {
hva               247 drivers/media/platform/sti/hva/hva-hw.c 			ctx->name, hva->lmi_err_reg);
hva               251 drivers/media/platform/sti/hva/hva-hw.c 	if (hva->emi_err_reg) {
hva               253 drivers/media/platform/sti/hva/hva-hw.c 			ctx->name, hva->emi_err_reg);
hva               257 drivers/media/platform/sti/hva/hva-hw.c 	if (hva->hec_mif_err_reg) {
hva               259 drivers/media/platform/sti/hva/hva-hw.c 			ctx->name, hva->hec_mif_err_reg);
hva               263 drivers/media/platform/sti/hva/hva-hw.c 	complete(&hva->interrupt);
hva               268 drivers/media/platform/sti/hva/hva-hw.c static unsigned long int hva_hw_get_ip_version(struct hva_dev *hva)
hva               270 drivers/media/platform/sti/hva/hva-hw.c 	struct device *dev = hva_to_dev(hva);
hva               275 drivers/media/platform/sti/hva/hva-hw.c 		mutex_unlock(&hva->protect_mutex);
hva               279 drivers/media/platform/sti/hva/hva-hw.c 	version = readl_relaxed(hva->regs + HVA_HIF_REG_VERSION) &
hva               299 drivers/media/platform/sti/hva/hva-hw.c int hva_hw_probe(struct platform_device *pdev, struct hva_dev *hva)
hva               306 drivers/media/platform/sti/hva/hva-hw.c 	WARN_ON(!hva);
hva               310 drivers/media/platform/sti/hva/hva-hw.c 	hva->regs = devm_ioremap_resource(dev, regs);
hva               311 drivers/media/platform/sti/hva/hva-hw.c 	if (IS_ERR(hva->regs)) {
hva               313 drivers/media/platform/sti/hva/hva-hw.c 		return PTR_ERR(hva->regs);
hva               322 drivers/media/platform/sti/hva/hva-hw.c 	hva->esram_addr = esram->start;
hva               323 drivers/media/platform/sti/hva/hva-hw.c 	hva->esram_size = resource_size(esram);
hva               326 drivers/media/platform/sti/hva/hva-hw.c 		 HVA_PREFIX, hva->esram_addr, hva->esram_size);
hva               329 drivers/media/platform/sti/hva/hva-hw.c 	hva->clk = devm_clk_get(dev, "clk_hva");
hva               330 drivers/media/platform/sti/hva/hva-hw.c 	if (IS_ERR(hva->clk)) {
hva               332 drivers/media/platform/sti/hva/hva-hw.c 		return PTR_ERR(hva->clk);
hva               335 drivers/media/platform/sti/hva/hva-hw.c 	ret = clk_prepare(hva->clk);
hva               338 drivers/media/platform/sti/hva/hva-hw.c 		hva->clk = ERR_PTR(-EINVAL);
hva               346 drivers/media/platform/sti/hva/hva-hw.c 	hva->irq_its = ret;
hva               348 drivers/media/platform/sti/hva/hva-hw.c 	ret = devm_request_threaded_irq(dev, hva->irq_its, hva_hw_its_interrupt,
hva               351 drivers/media/platform/sti/hva/hva-hw.c 					"hva_its_irq", hva);
hva               354 drivers/media/platform/sti/hva/hva-hw.c 			HVA_PREFIX, hva->irq_its);
hva               357 drivers/media/platform/sti/hva/hva-hw.c 	disable_irq(hva->irq_its);
hva               363 drivers/media/platform/sti/hva/hva-hw.c 	hva->irq_err = ret;
hva               365 drivers/media/platform/sti/hva/hva-hw.c 	ret = devm_request_threaded_irq(dev, hva->irq_err, hva_hw_err_interrupt,
hva               368 drivers/media/platform/sti/hva/hva-hw.c 					"hva_err_irq", hva);
hva               371 drivers/media/platform/sti/hva/hva-hw.c 			HVA_PREFIX, hva->irq_err);
hva               374 drivers/media/platform/sti/hva/hva-hw.c 	disable_irq(hva->irq_err);
hva               377 drivers/media/platform/sti/hva/hva-hw.c 	mutex_init(&hva->protect_mutex);
hva               380 drivers/media/platform/sti/hva/hva-hw.c 	init_completion(&hva->interrupt);
hva               395 drivers/media/platform/sti/hva/hva-hw.c 	hva->ip_version = hva_hw_get_ip_version(hva);
hva               397 drivers/media/platform/sti/hva/hva-hw.c 	if (hva->ip_version == HVA_VERSION_UNKNOWN) {
hva               403 drivers/media/platform/sti/hva/hva-hw.c 		 hva->ip_version);
hva               410 drivers/media/platform/sti/hva/hva-hw.c 	if (hva->clk)
hva               411 drivers/media/platform/sti/hva/hva-hw.c 		clk_unprepare(hva->clk);
hva               416 drivers/media/platform/sti/hva/hva-hw.c void hva_hw_remove(struct hva_dev *hva)
hva               418 drivers/media/platform/sti/hva/hva-hw.c 	struct device *dev = hva_to_dev(hva);
hva               420 drivers/media/platform/sti/hva/hva-hw.c 	disable_irq(hva->irq_its);
hva               421 drivers/media/platform/sti/hva/hva-hw.c 	disable_irq(hva->irq_err);
hva               429 drivers/media/platform/sti/hva/hva-hw.c 	struct hva_dev *hva = dev_get_drvdata(dev);
hva               431 drivers/media/platform/sti/hva/hva-hw.c 	clk_disable_unprepare(hva->clk);
hva               438 drivers/media/platform/sti/hva/hva-hw.c 	struct hva_dev *hva = dev_get_drvdata(dev);
hva               440 drivers/media/platform/sti/hva/hva-hw.c 	if (clk_prepare_enable(hva->clk)) {
hva               441 drivers/media/platform/sti/hva/hva-hw.c 		dev_err(hva->dev, "%s     failed to prepare hva clk\n",
hva               446 drivers/media/platform/sti/hva/hva-hw.c 	if (clk_set_rate(hva->clk, CLK_RATE)) {
hva               458 drivers/media/platform/sti/hva/hva-hw.c 	struct hva_dev *hva = ctx_to_hdev(ctx);
hva               459 drivers/media/platform/sti/hva/hva-hw.c 	struct device *dev = hva_to_dev(hva);
hva               464 drivers/media/platform/sti/hva/hva-hw.c 	mutex_lock(&hva->protect_mutex);
hva               467 drivers/media/platform/sti/hva/hva-hw.c 	enable_irq(hva->irq_its);
hva               468 drivers/media/platform/sti/hva/hva-hw.c 	enable_irq(hva->irq_err);
hva               477 drivers/media/platform/sti/hva/hva-hw.c 	reg = readl_relaxed(hva->regs + HVA_HIF_REG_CLK_GATING);
hva               488 drivers/media/platform/sti/hva/hva-hw.c 	writel_relaxed(reg, hva->regs + HVA_HIF_REG_CLK_GATING);
hva               494 drivers/media/platform/sti/hva/hva-hw.c 	writel_relaxed(BSM_CFG_VAL1, hva->regs + HVA_HIF_REG_BSM);
hva               497 drivers/media/platform/sti/hva/hva-hw.c 	writel_relaxed(MIF_CFG_VAL3, hva->regs + HVA_HIF_REG_MIF_CFG);
hva               498 drivers/media/platform/sti/hva/hva-hw.c 	writel_relaxed(HEC_MIF_CFG_VAL, hva->regs + HVA_HIF_REG_HEC_MIF_CFG);
hva               508 drivers/media/platform/sti/hva/hva-hw.c 	writel_relaxed(cmd + (client_id << 8), hva->regs + HVA_HIF_FIFO_CMD);
hva               509 drivers/media/platform/sti/hva/hva-hw.c 	writel_relaxed(task->paddr, hva->regs + HVA_HIF_FIFO_CMD);
hva               511 drivers/media/platform/sti/hva/hva-hw.c 	if (!wait_for_completion_timeout(&hva->interrupt,
hva               526 drivers/media/platform/sti/hva/hva-hw.c 	disable_irq(hva->irq_its);
hva               527 drivers/media/platform/sti/hva/hva-hw.c 	disable_irq(hva->irq_err);
hva               532 drivers/media/platform/sti/hva/hva-hw.c 		writel_relaxed(reg, hva->regs + HVA_HIF_REG_CLK_GATING);
hva               539 drivers/media/platform/sti/hva/hva-hw.c 	mutex_unlock(&hva->protect_mutex);
hva               546 drivers/media/platform/sti/hva/hva-hw.c 			     #reg, readl_relaxed(hva->regs + reg))
hva               548 drivers/media/platform/sti/hva/hva-hw.c void hva_hw_dump_regs(struct hva_dev *hva, struct seq_file *s)
hva               550 drivers/media/platform/sti/hva/hva-hw.c 	struct device *dev = hva_to_dev(hva);
hva               552 drivers/media/platform/sti/hva/hva-hw.c 	mutex_lock(&hva->protect_mutex);
hva               556 drivers/media/platform/sti/hva/hva-hw.c 		mutex_unlock(&hva->protect_mutex);
hva               560 drivers/media/platform/sti/hva/hva-hw.c 	seq_printf(s, "Registers:\nReg @ = 0x%p\n", hva->regs);
hva               580 drivers/media/platform/sti/hva/hva-hw.c 	mutex_unlock(&hva->protect_mutex);
hva                35 drivers/media/platform/sti/hva/hva-hw.h int hva_hw_probe(struct platform_device *pdev, struct hva_dev *hva);
hva                36 drivers/media/platform/sti/hva/hva-hw.h void hva_hw_remove(struct hva_dev *hva);
hva                42 drivers/media/platform/sti/hva/hva-hw.h void hva_hw_dump_regs(struct hva_dev *hva, struct seq_file *s);
hva               124 drivers/media/platform/sti/hva/hva-v4l2.c 	struct hva_dev *hva = ctx_to_hdev(ctx);
hva               128 drivers/media/platform/sti/hva/hva-v4l2.c 	for (i = 0; i < hva->nb_of_encoders; i++) {
hva               129 drivers/media/platform/sti/hva/hva-v4l2.c 		enc = hva->encoders[i];
hva               154 drivers/media/platform/sti/hva/hva-v4l2.c static void register_formats(struct hva_dev *hva)
hva               158 drivers/media/platform/sti/hva/hva-v4l2.c 	for (i = 0; i < hva->nb_of_encoders; i++) {
hva               159 drivers/media/platform/sti/hva/hva-v4l2.c 		register_format(hva->encoders[i]->pixelformat,
hva               160 drivers/media/platform/sti/hva/hva-v4l2.c 				hva->pixelformats,
hva               161 drivers/media/platform/sti/hva/hva-v4l2.c 				&hva->nb_of_pixelformats);
hva               163 drivers/media/platform/sti/hva/hva-v4l2.c 		register_format(hva->encoders[i]->streamformat,
hva               164 drivers/media/platform/sti/hva/hva-v4l2.c 				hva->streamformats,
hva               165 drivers/media/platform/sti/hva/hva-v4l2.c 				&hva->nb_of_streamformats);
hva               169 drivers/media/platform/sti/hva/hva-v4l2.c static void register_encoders(struct hva_dev *hva)
hva               171 drivers/media/platform/sti/hva/hva-v4l2.c 	struct device *dev = hva_to_dev(hva);
hva               175 drivers/media/platform/sti/hva/hva-v4l2.c 		if (hva->nb_of_encoders >= HVA_MAX_ENCODERS) {
hva               183 drivers/media/platform/sti/hva/hva-v4l2.c 		hva->encoders[hva->nb_of_encoders++] = hva_encoders[i];
hva               192 drivers/media/platform/sti/hva/hva-v4l2.c 	struct hva_dev *hva = ctx_to_hdev(ctx);
hva               211 drivers/media/platform/sti/hva/hva-v4l2.c 		 hva->instance_id, (char *)&streamformat);
hva               258 drivers/media/platform/sti/hva/hva-v4l2.c 	struct hva_dev *hva = ctx_to_hdev(ctx);
hva               261 drivers/media/platform/sti/hva/hva-v4l2.c 	strscpy(cap->card, hva->vdev->name, sizeof(cap->card));
hva               263 drivers/media/platform/sti/hva/hva-v4l2.c 		 hva->pdev->name);
hva               272 drivers/media/platform/sti/hva/hva-v4l2.c 	struct hva_dev *hva = ctx_to_hdev(ctx);
hva               274 drivers/media/platform/sti/hva/hva-v4l2.c 	if (unlikely(f->index >= hva->nb_of_streamformats))
hva               277 drivers/media/platform/sti/hva/hva-v4l2.c 	f->pixelformat = hva->streamformats[f->index];
hva               286 drivers/media/platform/sti/hva/hva-v4l2.c 	struct hva_dev *hva = ctx_to_hdev(ctx);
hva               288 drivers/media/platform/sti/hva/hva-v4l2.c 	if (unlikely(f->index >= hva->nb_of_pixelformats))
hva               291 drivers/media/platform/sti/hva/hva-v4l2.c 	f->pixelformat = hva->pixelformats[f->index];
hva               858 drivers/media/platform/sti/hva/hva-v4l2.c 	struct hva_dev *hva = ctx_to_hdev(ctx);
hva               860 drivers/media/platform/sti/hva/hva-v4l2.c 	queue_work(hva->work_queue, &ctx->run_work);
hva               997 drivers/media/platform/sti/hva/hva-v4l2.c 	struct hva_dev *hva = ctx_to_hdev(ctx);
hva              1018 drivers/media/platform/sti/hva/hva-v4l2.c 		if (!hva->instances[i]) {
hva              1019 drivers/media/platform/sti/hva/hva-v4l2.c 			hva->instances[i] = ctx;
hva              1033 drivers/media/platform/sti/hva/hva-v4l2.c 	hva->nb_of_instances++;
hva              1047 drivers/media/platform/sti/hva/hva-v4l2.c 	hva->instances[ctx->id] = NULL;
hva              1048 drivers/media/platform/sti/hva/hva-v4l2.c 	hva->nb_of_instances--;
hva              1068 drivers/media/platform/sti/hva/hva-v4l2.c 	struct hva_dev *hva = ctx_to_hdev(ctx);
hva              1106 drivers/media/platform/sti/hva/hva-v4l2.c 		hva->instances[ctx->id] = NULL;
hva              1107 drivers/media/platform/sti/hva/hva-v4l2.c 		hva->nb_of_instances--;
hva              1165 drivers/media/platform/sti/hva/hva-v4l2.c 	struct hva_dev *hva = video_drvdata(file);
hva              1166 drivers/media/platform/sti/hva/hva-v4l2.c 	struct device *dev = hva_to_dev(hva);
hva              1175 drivers/media/platform/sti/hva/hva-v4l2.c 	ctx->hva_dev = hva;
hva              1193 drivers/media/platform/sti/hva/hva-v4l2.c 	ctx->fh.m2m_ctx = v4l2_m2m_ctx_init(hva->m2m_dev, ctx,
hva              1204 drivers/media/platform/sti/hva/hva-v4l2.c 	mutex_lock(&hva->lock);
hva              1205 drivers/media/platform/sti/hva/hva-v4l2.c 	hva->instance_id++;
hva              1207 drivers/media/platform/sti/hva/hva-v4l2.c 		 hva->instance_id);
hva              1208 drivers/media/platform/sti/hva/hva-v4l2.c 	mutex_unlock(&hva->lock);
hva              1234 drivers/media/platform/sti/hva/hva-v4l2.c 	struct hva_dev *hva = ctx_to_hdev(ctx);
hva              1244 drivers/media/platform/sti/hva/hva-v4l2.c 		hva->instances[ctx->id] = NULL;
hva              1245 drivers/media/platform/sti/hva/hva-v4l2.c 		hva->nb_of_instances--;
hva              1283 drivers/media/platform/sti/hva/hva-v4l2.c static int hva_register_device(struct hva_dev *hva)
hva              1289 drivers/media/platform/sti/hva/hva-v4l2.c 	if (!hva)
hva              1291 drivers/media/platform/sti/hva/hva-v4l2.c 	dev = hva_to_dev(hva);
hva              1293 drivers/media/platform/sti/hva/hva-v4l2.c 	hva->m2m_dev = v4l2_m2m_init(&hva_m2m_ops);
hva              1294 drivers/media/platform/sti/hva/hva-v4l2.c 	if (IS_ERR(hva->m2m_dev)) {
hva              1297 drivers/media/platform/sti/hva/hva-v4l2.c 		ret = PTR_ERR(hva->m2m_dev);
hva              1312 drivers/media/platform/sti/hva/hva-v4l2.c 	vdev->lock = &hva->lock;
hva              1315 drivers/media/platform/sti/hva/hva-v4l2.c 	vdev->v4l2_dev = &hva->v4l2_dev;
hva              1317 drivers/media/platform/sti/hva/hva-v4l2.c 		 hva->ip_version);
hva              1326 drivers/media/platform/sti/hva/hva-v4l2.c 	hva->vdev = vdev;
hva              1327 drivers/media/platform/sti/hva/hva-v4l2.c 	video_set_drvdata(vdev, hva);
hva              1333 drivers/media/platform/sti/hva/hva-v4l2.c 	v4l2_m2m_release(hva->m2m_dev);
hva              1338 drivers/media/platform/sti/hva/hva-v4l2.c static void hva_unregister_device(struct hva_dev *hva)
hva              1340 drivers/media/platform/sti/hva/hva-v4l2.c 	if (!hva)
hva              1343 drivers/media/platform/sti/hva/hva-v4l2.c 	if (hva->m2m_dev)
hva              1344 drivers/media/platform/sti/hva/hva-v4l2.c 		v4l2_m2m_release(hva->m2m_dev);
hva              1346 drivers/media/platform/sti/hva/hva-v4l2.c 	video_unregister_device(hva->vdev);
hva              1351 drivers/media/platform/sti/hva/hva-v4l2.c 	struct hva_dev *hva;
hva              1355 drivers/media/platform/sti/hva/hva-v4l2.c 	hva = devm_kzalloc(dev, sizeof(*hva), GFP_KERNEL);
hva              1356 drivers/media/platform/sti/hva/hva-v4l2.c 	if (!hva) {
hva              1365 drivers/media/platform/sti/hva/hva-v4l2.c 	hva->dev = dev;
hva              1366 drivers/media/platform/sti/hva/hva-v4l2.c 	hva->pdev = pdev;
hva              1367 drivers/media/platform/sti/hva/hva-v4l2.c 	platform_set_drvdata(pdev, hva);
hva              1369 drivers/media/platform/sti/hva/hva-v4l2.c 	mutex_init(&hva->lock);
hva              1372 drivers/media/platform/sti/hva/hva-v4l2.c 	ret = hva_hw_probe(pdev, hva);
hva              1377 drivers/media/platform/sti/hva/hva-v4l2.c 	register_encoders(hva);
hva              1380 drivers/media/platform/sti/hva/hva-v4l2.c 	register_formats(hva);
hva              1383 drivers/media/platform/sti/hva/hva-v4l2.c 	ret = v4l2_device_register(dev, &hva->v4l2_dev);
hva              1391 drivers/media/platform/sti/hva/hva-v4l2.c 	hva_debugfs_create(hva);
hva              1394 drivers/media/platform/sti/hva/hva-v4l2.c 	hva->work_queue = create_workqueue(HVA_NAME);
hva              1395 drivers/media/platform/sti/hva/hva-v4l2.c 	if (!hva->work_queue) {
hva              1403 drivers/media/platform/sti/hva/hva-v4l2.c 	ret = hva_register_device(hva);
hva              1408 drivers/media/platform/sti/hva/hva-v4l2.c 		 HVA_NAME, hva->vdev->num);
hva              1413 drivers/media/platform/sti/hva/hva-v4l2.c 	destroy_workqueue(hva->work_queue);
hva              1416 drivers/media/platform/sti/hva/hva-v4l2.c 	hva_debugfs_remove(hva);
hva              1418 drivers/media/platform/sti/hva/hva-v4l2.c 	v4l2_device_unregister(&hva->v4l2_dev);
hva              1420 drivers/media/platform/sti/hva/hva-v4l2.c 	hva_hw_remove(hva);
hva              1427 drivers/media/platform/sti/hva/hva-v4l2.c 	struct hva_dev *hva = platform_get_drvdata(pdev);
hva              1428 drivers/media/platform/sti/hva/hva-v4l2.c 	struct device *dev = hva_to_dev(hva);
hva              1430 drivers/media/platform/sti/hva/hva-v4l2.c 	hva_unregister_device(hva);
hva              1432 drivers/media/platform/sti/hva/hva-v4l2.c 	destroy_workqueue(hva->work_queue);
hva              1434 drivers/media/platform/sti/hva/hva-v4l2.c 	hva_hw_remove(hva);
hva              1437 drivers/media/platform/sti/hva/hva-v4l2.c 	hva_debugfs_remove(hva);
hva              1440 drivers/media/platform/sti/hva/hva-v4l2.c 	v4l2_device_unregister(&hva->v4l2_dev);
hva               401 drivers/media/platform/sti/hva/hva.h void hva_debugfs_create(struct hva_dev *hva);
hva               402 drivers/media/platform/sti/hva/hva.h void hva_debugfs_remove(struct hva_dev *hva);
hva              1870 drivers/vhost/vhost.c static int log_write_hva(struct vhost_virtqueue *vq, u64 hva, u64 len)
hva              1884 drivers/vhost/vhost.c 			if (u->userspace_addr > hva - 1 + len ||
hva              1885 drivers/vhost/vhost.c 			    u->userspace_addr - 1 + u->size < hva)
hva              1887 drivers/vhost/vhost.c 			start = max(u->userspace_addr, hva);
hva              1889 drivers/vhost/vhost.c 				  hva - 1 + len);
hva              1904 drivers/vhost/vhost.c 		hva += min;
hva               216 include/linux/kvm_host.h 		       unsigned long hva, struct kvm_arch_async_pf *arch);
hva               239 include/linux/kvm_host.h 	void *hva;
hva               250 include/linux/kvm_host.h 	return !!map->hva;
hva              1057 include/linux/kvm_host.h hva_to_gfn_memslot(unsigned long hva, struct kvm_memory_slot *slot)
hva              1059 include/linux/kvm_host.h 	gfn_t gfn_offset = (hva - slot->userspace_addr) >> PAGE_SHIFT;
hva              1087 include/linux/kvm_host.h 	unsigned long hva = gfn_to_hva(kvm, gpa_to_gfn(gpa));
hva              1089 include/linux/kvm_host.h 	return kvm_is_error_hva(hva);
hva                47 include/linux/kvm_types.h 	unsigned long hva;
hva               263 include/trace/events/kvm.h 		__field(	u64,	hva		)
hva               272 include/trace/events/kvm.h 		__entry->hva		= ((gfn - slot->base_gfn) <<
hva               278 include/trace/events/kvm.h 		  __entry->hva, __entry->gfn, __entry->level,
hva                80 tools/testing/selftests/kvm/include/kvm_util.h int kvm_memcmp_hva_gva(void *hva, struct kvm_vm *vm, const vm_vaddr_t gva,
hva               110 tools/testing/selftests/kvm/include/kvm_util.h vm_paddr_t addr_hva2gpa(struct kvm_vm *vm, void *hva);
hva               506 tools/testing/selftests/kvm/lib/kvm_util.c int kvm_memcmp_hva_gva(void *hva, struct kvm_vm *vm, vm_vaddr_t gva, size_t len)
hva               515 tools/testing/selftests/kvm/lib/kvm_util.c 		uintptr_t ptr1 = (uintptr_t)hva + offset;
hva              1061 tools/testing/selftests/kvm/lib/kvm_util.c vm_paddr_t addr_hva2gpa(struct kvm_vm *vm, void *hva)
hva              1066 tools/testing/selftests/kvm/lib/kvm_util.c 		if ((hva >= region->host_mem)
hva              1067 tools/testing/selftests/kvm/lib/kvm_util.c 			&& (hva <= (region->host_mem
hva              1071 tools/testing/selftests/kvm/lib/kvm_util.c 				+ (hva - (uintptr_t) region->host_mem));
hva              1075 tools/testing/selftests/kvm/lib/kvm_util.c 		"hva: %p", hva);
hva               926 virt/kvm/arm/mmu.c 	hva_t hva = memslot->userspace_addr;
hva               929 virt/kvm/arm/mmu.c 	hva_t reg_end = hva + size;
hva               944 virt/kvm/arm/mmu.c 		struct vm_area_struct *vma = find_vma(current->mm, hva);
hva               953 virt/kvm/arm/mmu.c 		vm_start = max(hva, vma->vm_start);
hva               960 virt/kvm/arm/mmu.c 		hva = vm_end;
hva               961 virt/kvm/arm/mmu.c 	} while (hva < reg_end);
hva              1613 virt/kvm/arm/mmu.c 					       unsigned long hva,
hva              1665 virt/kvm/arm/mmu.c 	return (hva & ~(map_size - 1)) >= uaddr_start &&
hva              1666 virt/kvm/arm/mmu.c 	       (hva & ~(map_size - 1)) + map_size <= uaddr_end;
hva              1670 virt/kvm/arm/mmu.c 			  struct kvm_memory_slot *memslot, unsigned long hva,
hva              1697 virt/kvm/arm/mmu.c 	vma = find_vma_intersection(current->mm, hva, hva + 1);
hva              1699 virt/kvm/arm/mmu.c 		kvm_err("Failed to find VMA for hva 0x%lx\n", hva);
hva              1707 virt/kvm/arm/mmu.c 	    !fault_supports_stage2_huge_mapping(memslot, hva, vma_pagesize)) {
hva              1744 virt/kvm/arm/mmu.c 		kvm_send_hwpoison_signal(hva, vma);
hva              1785 virt/kvm/arm/mmu.c 		if (fault_supports_stage2_huge_mapping(memslot, hva, PMD_SIZE) &&
hva              1913 virt/kvm/arm/mmu.c 	unsigned long hva;
hva              1955 virt/kvm/arm/mmu.c 	hva = gfn_to_hva_memslot_prot(memslot, gfn, &writable);
hva              1957 virt/kvm/arm/mmu.c 	if (kvm_is_error_hva(hva) || (write_fault && !writable)) {
hva              2000 virt/kvm/arm/mmu.c 	ret = user_mem_abort(vcpu, fault_ipa, memslot, hva, fault_status);
hva              2079 virt/kvm/arm/mmu.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte)
hva              2081 virt/kvm/arm/mmu.c 	unsigned long end = hva + PAGE_SIZE;
hva              2088 virt/kvm/arm/mmu.c 	trace_kvm_set_spte_hva(hva);
hva              2096 virt/kvm/arm/mmu.c 	handle_hva_to_gpa(kvm, hva, end, &kvm_set_spte_handler, &stage2_pte);
hva              2145 virt/kvm/arm/mmu.c int kvm_test_age_hva(struct kvm *kvm, unsigned long hva)
hva              2149 virt/kvm/arm/mmu.c 	trace_kvm_test_age_hva(hva);
hva              2150 virt/kvm/arm/mmu.c 	return handle_hva_to_gpa(kvm, hva, hva + PAGE_SIZE,
hva              2281 virt/kvm/arm/mmu.c 	hva_t hva = mem->userspace_addr;
hva              2282 virt/kvm/arm/mmu.c 	hva_t reg_end = hva + mem->memory_size;
hva              2312 virt/kvm/arm/mmu.c 		struct vm_area_struct *vma = find_vma(current->mm, hva);
hva              2330 virt/kvm/arm/mmu.c 		vm_start = max(hva, vma->vm_start);
hva              2353 virt/kvm/arm/mmu.c 		hva = vm_end;
hva              2354 virt/kvm/arm/mmu.c 	} while (hva < reg_end);
hva               157 virt/kvm/arm/trace.h 	TP_PROTO(unsigned long hva),
hva               158 virt/kvm/arm/trace.h 	TP_ARGS(hva),
hva               161 virt/kvm/arm/trace.h 		__field(	unsigned long,	hva		)
hva               165 virt/kvm/arm/trace.h 		__entry->hva		= hva;
hva               168 virt/kvm/arm/trace.h 	TP_printk("mmu notifier set pte hva: %#08lx", __entry->hva)
hva               190 virt/kvm/arm/trace.h 	TP_PROTO(unsigned long hva),
hva               191 virt/kvm/arm/trace.h 	TP_ARGS(hva),
hva               194 virt/kvm/arm/trace.h 		__field(	unsigned long,	hva		)
hva               198 virt/kvm/arm/trace.h 		__entry->hva		= hva;
hva               201 virt/kvm/arm/trace.h 	TP_printk("mmu notifier test age hva: %#08lx", __entry->hva)
hva               169 virt/kvm/async_pf.c 		       unsigned long hva, struct kvm_arch_async_pf *arch)
hva               189 virt/kvm/async_pf.c 	work->addr = hva;
hva              1483 virt/kvm/kvm_main.c 	unsigned long hva = __gfn_to_hva_many(slot, gfn, NULL, false);
hva              1485 virt/kvm/kvm_main.c 	if (!kvm_is_error_hva(hva) && writable)
hva              1488 virt/kvm/kvm_main.c 	return hva;
hva              1849 virt/kvm/kvm_main.c 	void *hva = NULL;
hva              1876 virt/kvm/kvm_main.c 			hva = kmap_atomic(page);
hva              1878 virt/kvm/kvm_main.c 			hva = kmap(page);
hva              1881 virt/kvm/kvm_main.c 		hva = memremap(pfn_to_hpa(pfn), PAGE_SIZE, MEMREMAP_WB);
hva              1887 virt/kvm/kvm_main.c 	if (!hva)
hva              1891 virt/kvm/kvm_main.c 	map->hva = hva;
hva              1921 virt/kvm/kvm_main.c 	if (!map->hva)
hva              1926 virt/kvm/kvm_main.c 			kunmap_atomic(map->hva);
hva              1932 virt/kvm/kvm_main.c 		memunmap(map->hva);
hva              1945 virt/kvm/kvm_main.c 	map->hva = NULL;
hva              2239 virt/kvm/kvm_main.c 	ghc->hva = KVM_HVA_ERR_BAD;
hva              2247 virt/kvm/kvm_main.c 		ghc->hva = gfn_to_hva_many(ghc->memslot, start_gfn,
hva              2249 virt/kvm/kvm_main.c 		if (kvm_is_error_hva(ghc->hva))
hva              2256 virt/kvm/kvm_main.c 		ghc->hva += offset;
hva              2284 virt/kvm/kvm_main.c 	if (kvm_is_error_hva(ghc->hva))
hva              2290 virt/kvm/kvm_main.c 	r = __copy_to_user((void __user *)ghc->hva + offset, data, len);
hva              2317 virt/kvm/kvm_main.c 	if (kvm_is_error_hva(ghc->hva))
hva              2323 virt/kvm/kvm_main.c 	r = __copy_from_user(data, (void __user *)ghc->hva, len);