hva 270 arch/arm/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte); hva 275 arch/arm/include/asm/kvm_host.h int kvm_test_age_hva(struct kvm *kvm, unsigned long hva); hva 431 arch/arm64/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte); hva 433 arch/arm64/include/asm/kvm_host.h int kvm_test_age_hva(struct kvm *kvm, unsigned long hva); hva 943 arch/mips/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte); hva 945 arch/mips/include/asm/kvm_host.h int kvm_test_age_hva(struct kvm *kvm, unsigned long hva); hva 554 arch/mips/kvm/mmu.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte) hva 556 arch/mips/kvm/mmu.c unsigned long end = hva + PAGE_SIZE; hva 559 arch/mips/kvm/mmu.c ret = handle_hva_to_gpa(kvm, hva, end, &kvm_set_spte_handler, &pte); hva 587 arch/mips/kvm/mmu.c int kvm_test_age_hva(struct kvm *kvm, unsigned long hva) hva 589 arch/mips/kvm/mmu.c return handle_hva_to_gpa(kvm, hva, hva, kvm_test_age_hva_handler, NULL); hva 63 arch/powerpc/include/asm/kvm_host.h extern int kvm_test_age_hva(struct kvm *kvm, unsigned long hva); hva 64 arch/powerpc/include/asm/kvm_host.h extern int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte); hva 292 arch/powerpc/include/asm/kvm_ppc.h int (*test_age_hva)(struct kvm *kvm, unsigned long hva); hva 293 arch/powerpc/include/asm/kvm_ppc.h void (*set_spte_hva)(struct kvm *kvm, unsigned long hva, pte_t pte); hva 880 arch/powerpc/kvm/book3s.c int kvm_test_age_hva(struct kvm *kvm, unsigned long hva) hva 882 arch/powerpc/kvm/book3s.c return kvm->arch.kvm_ops->test_age_hva(kvm, hva); hva 885 arch/powerpc/kvm/book3s.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte) hva 887 arch/powerpc/kvm/book3s.c kvm->arch.kvm_ops->set_spte_hva(kvm, hva, pte); hva 16 arch/powerpc/kvm/book3s.h extern int kvm_test_age_hva_hv(struct kvm *kvm, unsigned long hva); hva 17 arch/powerpc/kvm/book3s.h extern void kvm_set_spte_hva_hv(struct kvm *kvm, unsigned long hva, pte_t pte); hva 500 arch/powerpc/kvm/book3s_64_mmu_hv.c unsigned long gpa, gfn, hva, pfn; hva 591 arch/powerpc/kvm/book3s_64_mmu_hv.c hva = gfn_to_hva_memslot(memslot, gfn); hva 592 arch/powerpc/kvm/book3s_64_mmu_hv.c npages = get_user_pages_fast(hva, 1, writing ? FOLL_WRITE : 0, pages); hva 596 arch/powerpc/kvm/book3s_64_mmu_hv.c vma = find_vma(current->mm, hva); hva 597 arch/powerpc/kvm/book3s_64_mmu_hv.c if (vma && vma->vm_start <= hva && hva + psize <= vma->vm_end && hva 600 arch/powerpc/kvm/book3s_64_mmu_hv.c ((hva - vma->vm_start) >> PAGE_SHIFT); hva 625 arch/powerpc/kvm/book3s_64_mmu_hv.c hva, NULL, NULL); hva 805 arch/powerpc/kvm/book3s_64_mmu_hv.c static int kvm_handle_hva(struct kvm *kvm, unsigned long hva, hva 808 arch/powerpc/kvm/book3s_64_mmu_hv.c return kvm_handle_hva_range(kvm, hva, hva + 1, handler); hva 1023 arch/powerpc/kvm/book3s_64_mmu_hv.c int kvm_test_age_hva_hv(struct kvm *kvm, unsigned long hva) hva 1028 arch/powerpc/kvm/book3s_64_mmu_hv.c return kvm_handle_hva(kvm, hva, handler); hva 1031 arch/powerpc/kvm/book3s_64_mmu_hv.c void kvm_set_spte_hva_hv(struct kvm *kvm, unsigned long hva, pte_t pte) hva 1036 arch/powerpc/kvm/book3s_64_mmu_hv.c kvm_handle_hva(kvm, hva, handler); hva 1177 arch/powerpc/kvm/book3s_64_mmu_hv.c unsigned long hva, offset; hva 1184 arch/powerpc/kvm/book3s_64_mmu_hv.c hva = gfn_to_hva_memslot(memslot, gfn); hva 1185 arch/powerpc/kvm/book3s_64_mmu_hv.c npages = get_user_pages_fast(hva, 1, FOLL_WRITE, pages); hva 775 arch/powerpc/kvm/book3s_64_mmu_radix.c unsigned long hva, gfn = gpa >> PAGE_SHIFT; hva 793 arch/powerpc/kvm/book3s_64_mmu_radix.c hva = gfn_to_hva_memslot(memslot, gfn); hva 794 arch/powerpc/kvm/book3s_64_mmu_radix.c if (!kvm_ro && __get_user_pages_fast(hva, 1, 1, &page) == 1) { hva 817 arch/powerpc/kvm/book3s_64_mmu_radix.c ptep = __find_linux_pte(vcpu->arch.pgdir, hva, NULL, &shift); hva 837 arch/powerpc/kvm/book3s_64_mmu_radix.c (hva & (PUD_SIZE - PAGE_SIZE))) { hva 841 arch/powerpc/kvm/book3s_64_mmu_radix.c (hva & (PMD_SIZE - PAGE_SIZE))) { hva 852 arch/powerpc/kvm/book3s_64_mmu_radix.c pte = __pte(pte_val(pte) | (hva & rpnmask)); hva 4580 arch/powerpc/kvm/book3s_hv.c unsigned long hva; hva 4617 arch/powerpc/kvm/book3s_hv.c hva = memslot->userspace_addr; hva 4619 arch/powerpc/kvm/book3s_hv.c vma = find_vma(current->mm, hva); hva 4620 arch/powerpc/kvm/book3s_hv.c if (!vma || vma->vm_start > hva || (vma->vm_flags & VM_IO)) hva 202 arch/powerpc/kvm/book3s_hv_rm_mmu.c unsigned long slot_fn, hva; hva 250 arch/powerpc/kvm/book3s_hv_rm_mmu.c hva = __gfn_to_hva_memslot(memslot, gfn); hva 261 arch/powerpc/kvm/book3s_hv_rm_mmu.c ptep = __find_linux_pte(pgdir, hva, NULL, &hpage_shift); hva 286 arch/powerpc/kvm/book3s_hv_rm_mmu.c pa |= hva & (host_pte_size - 1); hva 897 arch/powerpc/kvm/book3s_hv_rm_mmu.c unsigned long gfn, hva, pa, psize = PAGE_SHIFT; hva 908 arch/powerpc/kvm/book3s_hv_rm_mmu.c hva = __gfn_to_hva_memslot(memslot, gfn); hva 911 arch/powerpc/kvm/book3s_hv_rm_mmu.c ptep = __find_linux_pte(vcpu->arch.pgdir, hva, NULL, &shift); hva 922 arch/powerpc/kvm/book3s_hv_rm_mmu.c pa |= hva & (psize - 1); hva 437 arch/powerpc/kvm/book3s_pr.c static int kvm_test_age_hva_pr(struct kvm *kvm, unsigned long hva) hva 443 arch/powerpc/kvm/book3s_pr.c static void kvm_set_spte_hva_pr(struct kvm *kvm, unsigned long hva, pte_t pte) hva 446 arch/powerpc/kvm/book3s_pr.c do_kvm_unmap_hva(kvm, hva, hva + PAGE_SIZE); hva 329 arch/powerpc/kvm/e500_mmu_host.c unsigned long hva; hva 354 arch/powerpc/kvm/e500_mmu_host.c hva = gfn_to_hva_memslot(slot, gfn); hva 360 arch/powerpc/kvm/e500_mmu_host.c vma = find_vma(current->mm, hva); hva 361 arch/powerpc/kvm/e500_mmu_host.c if (vma && hva >= vma->vm_start && hva 379 arch/powerpc/kvm/e500_mmu_host.c pfn = start + ((hva - vma->vm_start) >> PAGE_SHIFT); hva 424 arch/powerpc/kvm/e500_mmu_host.c } else if (vma && hva >= vma->vm_start && hva 477 arch/powerpc/kvm/e500_mmu_host.c ptep = find_linux_pte(pgdir, hva, NULL, NULL); hva 724 arch/powerpc/kvm/e500_mmu_host.c static int kvm_unmap_hva(struct kvm *kvm, unsigned long hva) hva 726 arch/powerpc/kvm/e500_mmu_host.c trace_kvm_unmap_hva(hva); hva 751 arch/powerpc/kvm/e500_mmu_host.c int kvm_test_age_hva(struct kvm *kvm, unsigned long hva) hva 757 arch/powerpc/kvm/e500_mmu_host.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte) hva 760 arch/powerpc/kvm/e500_mmu_host.c kvm_unmap_hva(kvm, hva); hva 73 arch/powerpc/kvm/trace_booke.h TP_PROTO(unsigned long hva), hva 74 arch/powerpc/kvm/trace_booke.h TP_ARGS(hva), hva 77 arch/powerpc/kvm/trace_booke.h __field( unsigned long, hva ) hva 81 arch/powerpc/kvm/trace_booke.h __entry->hva = hva; hva 84 arch/powerpc/kvm/trace_booke.h TP_printk("unmap hva 0x%lx\n", __entry->hva) hva 1166 arch/s390/include/asm/pgtable.h int get_pgste(struct mm_struct *mm, unsigned long hva, unsigned long *pgstep); hva 1167 arch/s390/include/asm/pgtable.h int pgste_perform_essa(struct mm_struct *mm, unsigned long hva, int orc, hva 1793 arch/s390/kvm/kvm-s390.c uint64_t hva; hva 1814 arch/s390/kvm/kvm-s390.c hva = gfn_to_hva(kvm, args->start_gfn + i); hva 1815 arch/s390/kvm/kvm-s390.c if (kvm_is_error_hva(hva)) { hva 1820 arch/s390/kvm/kvm-s390.c r = get_guest_storage_key(current->mm, hva, &keys[i]); hva 1841 arch/s390/kvm/kvm-s390.c uint64_t hva; hva 1873 arch/s390/kvm/kvm-s390.c hva = gfn_to_hva(kvm, args->start_gfn + i); hva 1874 arch/s390/kvm/kvm-s390.c if (kvm_is_error_hva(hva)) { hva 1885 arch/s390/kvm/kvm-s390.c r = set_guest_storage_key(current->mm, hva, keys[i], 0); hva 1887 arch/s390/kvm/kvm-s390.c r = fixup_user_fault(current, current->mm, hva, hva 1949 arch/s390/kvm/kvm-s390.c unsigned long pgstev, hva, cur_gfn = args->start_gfn; hva 1953 arch/s390/kvm/kvm-s390.c hva = gfn_to_hva(kvm, cur_gfn); hva 1958 arch/s390/kvm/kvm-s390.c if (kvm_is_error_hva(hva)) hva 1960 arch/s390/kvm/kvm-s390.c if (get_pgste(kvm->mm, hva, &pgstev) < 0) hva 1997 arch/s390/kvm/kvm-s390.c unsigned long mem_end, cur_gfn, next_gfn, hva, pgstev; hva 2011 arch/s390/kvm/kvm-s390.c hva = gfn_to_hva(kvm, cur_gfn); hva 2012 arch/s390/kvm/kvm-s390.c if (kvm_is_error_hva(hva)) hva 2017 arch/s390/kvm/kvm-s390.c if (get_pgste(kvm->mm, hva, &pgstev) < 0) hva 2111 arch/s390/kvm/kvm-s390.c unsigned long hva, mask, pgstev, i; hva 2142 arch/s390/kvm/kvm-s390.c hva = gfn_to_hva(kvm, args->start_gfn + i); hva 2143 arch/s390/kvm/kvm-s390.c if (kvm_is_error_hva(hva)) { hva 2151 arch/s390/kvm/kvm-s390.c set_pgste_bits(kvm->mm, hva, mask, pgstev); hva 3667 arch/s390/kvm/kvm-s390.c hva_t hva; hva 3685 arch/s390/kvm/kvm-s390.c hva = gfn_to_hva(vcpu->kvm, gpa_to_gfn(current->thread.gmap_addr)); hva 3686 arch/s390/kvm/kvm-s390.c hva += current->thread.gmap_addr & ~PAGE_MASK; hva 3690 arch/s390/kvm/kvm-s390.c rc = kvm_setup_async_pf(vcpu, current->thread.gmap_addr, hva, &arch); hva 1125 arch/s390/kvm/priv.c unsigned long gfn, hva, res, pgstev, ptev; hva 1135 arch/s390/kvm/priv.c hva = gfn_to_hva(vcpu->kvm, gfn); hva 1138 arch/s390/kvm/priv.c if (kvm_is_error_hva(hva)) hva 1141 arch/s390/kvm/priv.c nappended = pgste_perform_essa(vcpu->kvm->mm, hva, orc, &ptev, &pgstev); hva 1438 arch/s390/kvm/priv.c unsigned long hva, gpa; hva 1475 arch/s390/kvm/priv.c hva = gfn_to_hva_prot(vcpu->kvm, gpa_to_gfn(gpa), &writable); hva 1476 arch/s390/kvm/priv.c if (kvm_is_error_hva(hva)) { hva 970 arch/s390/mm/pgtable.c int pgste_perform_essa(struct mm_struct *mm, unsigned long hva, int orc, hva 982 arch/s390/mm/pgtable.c ptep = get_locked_pte(mm, hva, &ptl); hva 1071 arch/s390/mm/pgtable.c int set_pgste_bits(struct mm_struct *mm, unsigned long hva, hva 1078 arch/s390/mm/pgtable.c ptep = get_locked_pte(mm, hva, &ptl); hva 1100 arch/s390/mm/pgtable.c int get_pgste(struct mm_struct *mm, unsigned long hva, unsigned long *pgstep) hva 1105 arch/s390/mm/pgtable.c ptep = get_locked_pte(mm, hva, &ptl); hva 1558 arch/x86/include/asm/kvm_host.h int kvm_test_age_hva(struct kvm *kvm, unsigned long hva); hva 1559 arch/x86/include/asm/kvm_host.h int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte); hva 2037 arch/x86/kvm/mmu.c static int kvm_handle_hva(struct kvm *kvm, unsigned long hva, hva 2045 arch/x86/kvm/mmu.c return kvm_handle_hva_range(kvm, hva, hva + 1, data, handler); hva 2053 arch/x86/kvm/mmu.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte) hva 2055 arch/x86/kvm/mmu.c return kvm_handle_hva(kvm, hva, (unsigned long)&pte, kvm_set_pte_rmapp); hva 2107 arch/x86/kvm/mmu.c int kvm_test_age_hva(struct kvm *kvm, unsigned long hva) hva 2109 arch/x86/kvm/mmu.c return kvm_handle_hva(kvm, hva, 0, kvm_test_age_rmapp); hva 3372 arch/x86/kvm/svm.c nested_vmcb = map.hva; hva 3665 arch/x86/kvm/svm.c nested_vmcb = map.hva; hva 3763 arch/x86/kvm/svm.c nested_vmcb = map.hva; hva 3789 arch/x86/kvm/svm.c nested_vmcb = map.hva; hva 6310 arch/x86/kvm/svm.c nested_vmcb = map.hva; hva 577 arch/x86/kvm/vmx/nested.c msr_bitmap_l1 = (unsigned long *)map->hva; hva 673 arch/x86/kvm/vmx/nested.c memcpy(shadow, map.hva, VMCS12_SIZE); hva 1842 arch/x86/kvm/vmx/nested.c vmx->nested.hv_evmcs = vmx->nested.hv_evmcs_map.hva; hva 2750 arch/x86/kvm/vmx/nested.c shadow = map.hva; hva 2997 arch/x86/kvm/vmx/nested.c (struct pi_desc *)(((void *)map->hva) + hva 3422 arch/x86/kvm/vmx/nested.c vapic_page = vmx->nested.virtual_apic_map.hva; hva 4848 arch/x86/kvm/vmx/nested.c new_vmcs12 = map.hva; hva 3754 arch/x86/kvm/vmx/vmx.c vapic_page = vmx->nested.virtual_apic_map.hva; hva 2660 arch/x86/kvm/x86.c st = map.hva + hva 3558 arch/x86/kvm/x86.c st = map.hva + hva 5858 arch/x86/kvm/x86.c kaddr = map.hva + offset_in_page(gpa); hva 9625 arch/x86/kvm/x86.c unsigned long hva; hva 9642 arch/x86/kvm/x86.c hva = vm_mmap(NULL, 0, size, PROT_READ | PROT_WRITE, hva 9644 arch/x86/kvm/x86.c if (IS_ERR((void *)hva)) hva 9645 arch/x86/kvm/x86.c return PTR_ERR((void *)hva); hva 9650 arch/x86/kvm/x86.c hva = 0; hva 9660 arch/x86/kvm/x86.c m.userspace_addr = hva; hva 276 drivers/media/platform/sti/hva/hva-debugfs.c struct hva_dev *hva = s->private; hva 278 drivers/media/platform/sti/hva/hva-debugfs.c seq_printf(s, "[%s]\n", hva->v4l2_dev.name); hva 279 drivers/media/platform/sti/hva/hva-debugfs.c seq_printf(s, "registered as /dev/video%d\n", hva->vdev->num); hva 286 drivers/media/platform/sti/hva/hva-debugfs.c struct hva_dev *hva = s->private; hva 290 drivers/media/platform/sti/hva/hva-debugfs.c hva->nb_of_encoders); hva 292 drivers/media/platform/sti/hva/hva-debugfs.c while (hva->encoders[i]) { hva 293 drivers/media/platform/sti/hva/hva-debugfs.c seq_printf(s, "|- %s: %4.4s => %4.4s\n", hva->encoders[i]->name, hva 294 drivers/media/platform/sti/hva/hva-debugfs.c (char *)&hva->encoders[i]->pixelformat, hva 295 drivers/media/platform/sti/hva/hva-debugfs.c (char *)&hva->encoders[i]->streamformat); hva 304 drivers/media/platform/sti/hva/hva-debugfs.c struct hva_dev *hva = s->private; hva 305 drivers/media/platform/sti/hva/hva-debugfs.c struct hva_ctx *last_ctx = &hva->dbg.last_ctx; hva 321 drivers/media/platform/sti/hva/hva-debugfs.c struct hva_dev *hva = s->private; hva 323 drivers/media/platform/sti/hva/hva-debugfs.c hva_hw_dump_regs(hva, s); hva 329 drivers/media/platform/sti/hva/hva-debugfs.c debugfs_create_file(#name, 0444, hva->dbg.debugfs_entry, hva, \ hva 337 drivers/media/platform/sti/hva/hva-debugfs.c void hva_debugfs_create(struct hva_dev *hva) hva 339 drivers/media/platform/sti/hva/hva-debugfs.c hva->dbg.debugfs_entry = debugfs_create_dir(HVA_NAME, NULL); hva 340 drivers/media/platform/sti/hva/hva-debugfs.c if (!hva->dbg.debugfs_entry) hva 358 drivers/media/platform/sti/hva/hva-debugfs.c hva_debugfs_remove(hva); hva 361 drivers/media/platform/sti/hva/hva-debugfs.c void hva_debugfs_remove(struct hva_dev *hva) hva 363 drivers/media/platform/sti/hva/hva-debugfs.c debugfs_remove_recursive(hva->dbg.debugfs_entry); hva 364 drivers/media/platform/sti/hva/hva-debugfs.c hva->dbg.debugfs_entry = NULL; hva 387 drivers/media/platform/sti/hva/hva-debugfs.c struct hva_dev *hva = ctx->hva_dev; hva 394 drivers/media/platform/sti/hva/hva-debugfs.c snprintf(name, sizeof(name), "%d", hva->instance_id); hva 397 drivers/media/platform/sti/hva/hva-debugfs.c hva->dbg.debugfs_entry, hva 403 drivers/media/platform/sti/hva/hva-debugfs.c struct hva_dev *hva = ctx->hva_dev; hva 407 drivers/media/platform/sti/hva/hva-debugfs.c memcpy(&hva->dbg.last_ctx, ctx, sizeof(*ctx)); hva 590 drivers/media/platform/sti/hva/hva-h264.c struct hva_dev *hva = ctx_to_hdev(pctx); hva 601 drivers/media/platform/sti/hva/hva-h264.c u32 addr_esram = hva->esram_addr; hva 891 drivers/media/platform/sti/hva/hva-h264.c struct hva_dev *hva = ctx_to_hdev(pctx); hva 903 drivers/media/platform/sti/hva/hva-h264.c if (hva->esram_size < size) { hva 905 drivers/media/platform/sti/hva/hva-h264.c pctx->name, hva->esram_size, size); hva 103 drivers/media/platform/sti/hva/hva-hw.c struct hva_dev *hva = data; hva 106 drivers/media/platform/sti/hva/hva-hw.c hva->sts_reg = readl_relaxed(hva->regs + HVA_HIF_FIFO_STS); hva 107 drivers/media/platform/sti/hva/hva-hw.c hva->sfl_reg = readl_relaxed(hva->regs + HVA_HIF_REG_SFL); hva 110 drivers/media/platform/sti/hva/hva-hw.c writel_relaxed(0x1, hva->regs + HVA_HIF_REG_IT_ACK); hva 117 drivers/media/platform/sti/hva/hva-hw.c struct hva_dev *hva = arg; hva 118 drivers/media/platform/sti/hva/hva-hw.c struct device *dev = hva_to_dev(hva); hva 119 drivers/media/platform/sti/hva/hva-hw.c u32 status = hva->sts_reg & 0xFF; hva 124 drivers/media/platform/sti/hva/hva-hw.c HVA_PREFIX, __func__, hva->sts_reg & 0xFF, hva->sfl_reg & 0xF); hva 130 drivers/media/platform/sti/hva/hva-hw.c ctx_id = (hva->sts_reg & 0xFF00) >> 8; hva 138 drivers/media/platform/sti/hva/hva-hw.c ctx = hva->instances[ctx_id]; hva 195 drivers/media/platform/sti/hva/hva-hw.c complete(&hva->interrupt); hva 202 drivers/media/platform/sti/hva/hva-hw.c struct hva_dev *hva = data; hva 205 drivers/media/platform/sti/hva/hva-hw.c hva->sts_reg = readl_relaxed(hva->regs + HVA_HIF_FIFO_STS); hva 206 drivers/media/platform/sti/hva/hva-hw.c hva->sfl_reg = readl_relaxed(hva->regs + HVA_HIF_REG_SFL); hva 209 drivers/media/platform/sti/hva/hva-hw.c hva->lmi_err_reg = readl_relaxed(hva->regs + HVA_HIF_REG_LMI_ERR); hva 210 drivers/media/platform/sti/hva/hva-hw.c hva->emi_err_reg = readl_relaxed(hva->regs + HVA_HIF_REG_EMI_ERR); hva 211 drivers/media/platform/sti/hva/hva-hw.c hva->hec_mif_err_reg = readl_relaxed(hva->regs + hva 215 drivers/media/platform/sti/hva/hva-hw.c writel_relaxed(0x1, hva->regs + HVA_HIF_REG_IT_ACK); hva 222 drivers/media/platform/sti/hva/hva-hw.c struct hva_dev *hva = arg; hva 223 drivers/media/platform/sti/hva/hva-hw.c struct device *dev = hva_to_dev(hva); hva 228 drivers/media/platform/sti/hva/hva-hw.c HVA_PREFIX, hva->sts_reg & 0xFF, hva->sfl_reg & 0xF); hva 234 drivers/media/platform/sti/hva/hva-hw.c ctx_id = (hva->sts_reg & 0xFF00) >> 8; hva 241 drivers/media/platform/sti/hva/hva-hw.c ctx = hva->instances[ctx_id]; hva 245 drivers/media/platform/sti/hva/hva-hw.c if (hva->lmi_err_reg) { hva 247 drivers/media/platform/sti/hva/hva-hw.c ctx->name, hva->lmi_err_reg); hva 251 drivers/media/platform/sti/hva/hva-hw.c if (hva->emi_err_reg) { hva 253 drivers/media/platform/sti/hva/hva-hw.c ctx->name, hva->emi_err_reg); hva 257 drivers/media/platform/sti/hva/hva-hw.c if (hva->hec_mif_err_reg) { hva 259 drivers/media/platform/sti/hva/hva-hw.c ctx->name, hva->hec_mif_err_reg); hva 263 drivers/media/platform/sti/hva/hva-hw.c complete(&hva->interrupt); hva 268 drivers/media/platform/sti/hva/hva-hw.c static unsigned long int hva_hw_get_ip_version(struct hva_dev *hva) hva 270 drivers/media/platform/sti/hva/hva-hw.c struct device *dev = hva_to_dev(hva); hva 275 drivers/media/platform/sti/hva/hva-hw.c mutex_unlock(&hva->protect_mutex); hva 279 drivers/media/platform/sti/hva/hva-hw.c version = readl_relaxed(hva->regs + HVA_HIF_REG_VERSION) & hva 299 drivers/media/platform/sti/hva/hva-hw.c int hva_hw_probe(struct platform_device *pdev, struct hva_dev *hva) hva 306 drivers/media/platform/sti/hva/hva-hw.c WARN_ON(!hva); hva 310 drivers/media/platform/sti/hva/hva-hw.c hva->regs = devm_ioremap_resource(dev, regs); hva 311 drivers/media/platform/sti/hva/hva-hw.c if (IS_ERR(hva->regs)) { hva 313 drivers/media/platform/sti/hva/hva-hw.c return PTR_ERR(hva->regs); hva 322 drivers/media/platform/sti/hva/hva-hw.c hva->esram_addr = esram->start; hva 323 drivers/media/platform/sti/hva/hva-hw.c hva->esram_size = resource_size(esram); hva 326 drivers/media/platform/sti/hva/hva-hw.c HVA_PREFIX, hva->esram_addr, hva->esram_size); hva 329 drivers/media/platform/sti/hva/hva-hw.c hva->clk = devm_clk_get(dev, "clk_hva"); hva 330 drivers/media/platform/sti/hva/hva-hw.c if (IS_ERR(hva->clk)) { hva 332 drivers/media/platform/sti/hva/hva-hw.c return PTR_ERR(hva->clk); hva 335 drivers/media/platform/sti/hva/hva-hw.c ret = clk_prepare(hva->clk); hva 338 drivers/media/platform/sti/hva/hva-hw.c hva->clk = ERR_PTR(-EINVAL); hva 346 drivers/media/platform/sti/hva/hva-hw.c hva->irq_its = ret; hva 348 drivers/media/platform/sti/hva/hva-hw.c ret = devm_request_threaded_irq(dev, hva->irq_its, hva_hw_its_interrupt, hva 351 drivers/media/platform/sti/hva/hva-hw.c "hva_its_irq", hva); hva 354 drivers/media/platform/sti/hva/hva-hw.c HVA_PREFIX, hva->irq_its); hva 357 drivers/media/platform/sti/hva/hva-hw.c disable_irq(hva->irq_its); hva 363 drivers/media/platform/sti/hva/hva-hw.c hva->irq_err = ret; hva 365 drivers/media/platform/sti/hva/hva-hw.c ret = devm_request_threaded_irq(dev, hva->irq_err, hva_hw_err_interrupt, hva 368 drivers/media/platform/sti/hva/hva-hw.c "hva_err_irq", hva); hva 371 drivers/media/platform/sti/hva/hva-hw.c HVA_PREFIX, hva->irq_err); hva 374 drivers/media/platform/sti/hva/hva-hw.c disable_irq(hva->irq_err); hva 377 drivers/media/platform/sti/hva/hva-hw.c mutex_init(&hva->protect_mutex); hva 380 drivers/media/platform/sti/hva/hva-hw.c init_completion(&hva->interrupt); hva 395 drivers/media/platform/sti/hva/hva-hw.c hva->ip_version = hva_hw_get_ip_version(hva); hva 397 drivers/media/platform/sti/hva/hva-hw.c if (hva->ip_version == HVA_VERSION_UNKNOWN) { hva 403 drivers/media/platform/sti/hva/hva-hw.c hva->ip_version); hva 410 drivers/media/platform/sti/hva/hva-hw.c if (hva->clk) hva 411 drivers/media/platform/sti/hva/hva-hw.c clk_unprepare(hva->clk); hva 416 drivers/media/platform/sti/hva/hva-hw.c void hva_hw_remove(struct hva_dev *hva) hva 418 drivers/media/platform/sti/hva/hva-hw.c struct device *dev = hva_to_dev(hva); hva 420 drivers/media/platform/sti/hva/hva-hw.c disable_irq(hva->irq_its); hva 421 drivers/media/platform/sti/hva/hva-hw.c disable_irq(hva->irq_err); hva 429 drivers/media/platform/sti/hva/hva-hw.c struct hva_dev *hva = dev_get_drvdata(dev); hva 431 drivers/media/platform/sti/hva/hva-hw.c clk_disable_unprepare(hva->clk); hva 438 drivers/media/platform/sti/hva/hva-hw.c struct hva_dev *hva = dev_get_drvdata(dev); hva 440 drivers/media/platform/sti/hva/hva-hw.c if (clk_prepare_enable(hva->clk)) { hva 441 drivers/media/platform/sti/hva/hva-hw.c dev_err(hva->dev, "%s failed to prepare hva clk\n", hva 446 drivers/media/platform/sti/hva/hva-hw.c if (clk_set_rate(hva->clk, CLK_RATE)) { hva 458 drivers/media/platform/sti/hva/hva-hw.c struct hva_dev *hva = ctx_to_hdev(ctx); hva 459 drivers/media/platform/sti/hva/hva-hw.c struct device *dev = hva_to_dev(hva); hva 464 drivers/media/platform/sti/hva/hva-hw.c mutex_lock(&hva->protect_mutex); hva 467 drivers/media/platform/sti/hva/hva-hw.c enable_irq(hva->irq_its); hva 468 drivers/media/platform/sti/hva/hva-hw.c enable_irq(hva->irq_err); hva 477 drivers/media/platform/sti/hva/hva-hw.c reg = readl_relaxed(hva->regs + HVA_HIF_REG_CLK_GATING); hva 488 drivers/media/platform/sti/hva/hva-hw.c writel_relaxed(reg, hva->regs + HVA_HIF_REG_CLK_GATING); hva 494 drivers/media/platform/sti/hva/hva-hw.c writel_relaxed(BSM_CFG_VAL1, hva->regs + HVA_HIF_REG_BSM); hva 497 drivers/media/platform/sti/hva/hva-hw.c writel_relaxed(MIF_CFG_VAL3, hva->regs + HVA_HIF_REG_MIF_CFG); hva 498 drivers/media/platform/sti/hva/hva-hw.c writel_relaxed(HEC_MIF_CFG_VAL, hva->regs + HVA_HIF_REG_HEC_MIF_CFG); hva 508 drivers/media/platform/sti/hva/hva-hw.c writel_relaxed(cmd + (client_id << 8), hva->regs + HVA_HIF_FIFO_CMD); hva 509 drivers/media/platform/sti/hva/hva-hw.c writel_relaxed(task->paddr, hva->regs + HVA_HIF_FIFO_CMD); hva 511 drivers/media/platform/sti/hva/hva-hw.c if (!wait_for_completion_timeout(&hva->interrupt, hva 526 drivers/media/platform/sti/hva/hva-hw.c disable_irq(hva->irq_its); hva 527 drivers/media/platform/sti/hva/hva-hw.c disable_irq(hva->irq_err); hva 532 drivers/media/platform/sti/hva/hva-hw.c writel_relaxed(reg, hva->regs + HVA_HIF_REG_CLK_GATING); hva 539 drivers/media/platform/sti/hva/hva-hw.c mutex_unlock(&hva->protect_mutex); hva 546 drivers/media/platform/sti/hva/hva-hw.c #reg, readl_relaxed(hva->regs + reg)) hva 548 drivers/media/platform/sti/hva/hva-hw.c void hva_hw_dump_regs(struct hva_dev *hva, struct seq_file *s) hva 550 drivers/media/platform/sti/hva/hva-hw.c struct device *dev = hva_to_dev(hva); hva 552 drivers/media/platform/sti/hva/hva-hw.c mutex_lock(&hva->protect_mutex); hva 556 drivers/media/platform/sti/hva/hva-hw.c mutex_unlock(&hva->protect_mutex); hva 560 drivers/media/platform/sti/hva/hva-hw.c seq_printf(s, "Registers:\nReg @ = 0x%p\n", hva->regs); hva 580 drivers/media/platform/sti/hva/hva-hw.c mutex_unlock(&hva->protect_mutex); hva 35 drivers/media/platform/sti/hva/hva-hw.h int hva_hw_probe(struct platform_device *pdev, struct hva_dev *hva); hva 36 drivers/media/platform/sti/hva/hva-hw.h void hva_hw_remove(struct hva_dev *hva); hva 42 drivers/media/platform/sti/hva/hva-hw.h void hva_hw_dump_regs(struct hva_dev *hva, struct seq_file *s); hva 124 drivers/media/platform/sti/hva/hva-v4l2.c struct hva_dev *hva = ctx_to_hdev(ctx); hva 128 drivers/media/platform/sti/hva/hva-v4l2.c for (i = 0; i < hva->nb_of_encoders; i++) { hva 129 drivers/media/platform/sti/hva/hva-v4l2.c enc = hva->encoders[i]; hva 154 drivers/media/platform/sti/hva/hva-v4l2.c static void register_formats(struct hva_dev *hva) hva 158 drivers/media/platform/sti/hva/hva-v4l2.c for (i = 0; i < hva->nb_of_encoders; i++) { hva 159 drivers/media/platform/sti/hva/hva-v4l2.c register_format(hva->encoders[i]->pixelformat, hva 160 drivers/media/platform/sti/hva/hva-v4l2.c hva->pixelformats, hva 161 drivers/media/platform/sti/hva/hva-v4l2.c &hva->nb_of_pixelformats); hva 163 drivers/media/platform/sti/hva/hva-v4l2.c register_format(hva->encoders[i]->streamformat, hva 164 drivers/media/platform/sti/hva/hva-v4l2.c hva->streamformats, hva 165 drivers/media/platform/sti/hva/hva-v4l2.c &hva->nb_of_streamformats); hva 169 drivers/media/platform/sti/hva/hva-v4l2.c static void register_encoders(struct hva_dev *hva) hva 171 drivers/media/platform/sti/hva/hva-v4l2.c struct device *dev = hva_to_dev(hva); hva 175 drivers/media/platform/sti/hva/hva-v4l2.c if (hva->nb_of_encoders >= HVA_MAX_ENCODERS) { hva 183 drivers/media/platform/sti/hva/hva-v4l2.c hva->encoders[hva->nb_of_encoders++] = hva_encoders[i]; hva 192 drivers/media/platform/sti/hva/hva-v4l2.c struct hva_dev *hva = ctx_to_hdev(ctx); hva 211 drivers/media/platform/sti/hva/hva-v4l2.c hva->instance_id, (char *)&streamformat); hva 258 drivers/media/platform/sti/hva/hva-v4l2.c struct hva_dev *hva = ctx_to_hdev(ctx); hva 261 drivers/media/platform/sti/hva/hva-v4l2.c strscpy(cap->card, hva->vdev->name, sizeof(cap->card)); hva 263 drivers/media/platform/sti/hva/hva-v4l2.c hva->pdev->name); hva 272 drivers/media/platform/sti/hva/hva-v4l2.c struct hva_dev *hva = ctx_to_hdev(ctx); hva 274 drivers/media/platform/sti/hva/hva-v4l2.c if (unlikely(f->index >= hva->nb_of_streamformats)) hva 277 drivers/media/platform/sti/hva/hva-v4l2.c f->pixelformat = hva->streamformats[f->index]; hva 286 drivers/media/platform/sti/hva/hva-v4l2.c struct hva_dev *hva = ctx_to_hdev(ctx); hva 288 drivers/media/platform/sti/hva/hva-v4l2.c if (unlikely(f->index >= hva->nb_of_pixelformats)) hva 291 drivers/media/platform/sti/hva/hva-v4l2.c f->pixelformat = hva->pixelformats[f->index]; hva 858 drivers/media/platform/sti/hva/hva-v4l2.c struct hva_dev *hva = ctx_to_hdev(ctx); hva 860 drivers/media/platform/sti/hva/hva-v4l2.c queue_work(hva->work_queue, &ctx->run_work); hva 997 drivers/media/platform/sti/hva/hva-v4l2.c struct hva_dev *hva = ctx_to_hdev(ctx); hva 1018 drivers/media/platform/sti/hva/hva-v4l2.c if (!hva->instances[i]) { hva 1019 drivers/media/platform/sti/hva/hva-v4l2.c hva->instances[i] = ctx; hva 1033 drivers/media/platform/sti/hva/hva-v4l2.c hva->nb_of_instances++; hva 1047 drivers/media/platform/sti/hva/hva-v4l2.c hva->instances[ctx->id] = NULL; hva 1048 drivers/media/platform/sti/hva/hva-v4l2.c hva->nb_of_instances--; hva 1068 drivers/media/platform/sti/hva/hva-v4l2.c struct hva_dev *hva = ctx_to_hdev(ctx); hva 1106 drivers/media/platform/sti/hva/hva-v4l2.c hva->instances[ctx->id] = NULL; hva 1107 drivers/media/platform/sti/hva/hva-v4l2.c hva->nb_of_instances--; hva 1165 drivers/media/platform/sti/hva/hva-v4l2.c struct hva_dev *hva = video_drvdata(file); hva 1166 drivers/media/platform/sti/hva/hva-v4l2.c struct device *dev = hva_to_dev(hva); hva 1175 drivers/media/platform/sti/hva/hva-v4l2.c ctx->hva_dev = hva; hva 1193 drivers/media/platform/sti/hva/hva-v4l2.c ctx->fh.m2m_ctx = v4l2_m2m_ctx_init(hva->m2m_dev, ctx, hva 1204 drivers/media/platform/sti/hva/hva-v4l2.c mutex_lock(&hva->lock); hva 1205 drivers/media/platform/sti/hva/hva-v4l2.c hva->instance_id++; hva 1207 drivers/media/platform/sti/hva/hva-v4l2.c hva->instance_id); hva 1208 drivers/media/platform/sti/hva/hva-v4l2.c mutex_unlock(&hva->lock); hva 1234 drivers/media/platform/sti/hva/hva-v4l2.c struct hva_dev *hva = ctx_to_hdev(ctx); hva 1244 drivers/media/platform/sti/hva/hva-v4l2.c hva->instances[ctx->id] = NULL; hva 1245 drivers/media/platform/sti/hva/hva-v4l2.c hva->nb_of_instances--; hva 1283 drivers/media/platform/sti/hva/hva-v4l2.c static int hva_register_device(struct hva_dev *hva) hva 1289 drivers/media/platform/sti/hva/hva-v4l2.c if (!hva) hva 1291 drivers/media/platform/sti/hva/hva-v4l2.c dev = hva_to_dev(hva); hva 1293 drivers/media/platform/sti/hva/hva-v4l2.c hva->m2m_dev = v4l2_m2m_init(&hva_m2m_ops); hva 1294 drivers/media/platform/sti/hva/hva-v4l2.c if (IS_ERR(hva->m2m_dev)) { hva 1297 drivers/media/platform/sti/hva/hva-v4l2.c ret = PTR_ERR(hva->m2m_dev); hva 1312 drivers/media/platform/sti/hva/hva-v4l2.c vdev->lock = &hva->lock; hva 1315 drivers/media/platform/sti/hva/hva-v4l2.c vdev->v4l2_dev = &hva->v4l2_dev; hva 1317 drivers/media/platform/sti/hva/hva-v4l2.c hva->ip_version); hva 1326 drivers/media/platform/sti/hva/hva-v4l2.c hva->vdev = vdev; hva 1327 drivers/media/platform/sti/hva/hva-v4l2.c video_set_drvdata(vdev, hva); hva 1333 drivers/media/platform/sti/hva/hva-v4l2.c v4l2_m2m_release(hva->m2m_dev); hva 1338 drivers/media/platform/sti/hva/hva-v4l2.c static void hva_unregister_device(struct hva_dev *hva) hva 1340 drivers/media/platform/sti/hva/hva-v4l2.c if (!hva) hva 1343 drivers/media/platform/sti/hva/hva-v4l2.c if (hva->m2m_dev) hva 1344 drivers/media/platform/sti/hva/hva-v4l2.c v4l2_m2m_release(hva->m2m_dev); hva 1346 drivers/media/platform/sti/hva/hva-v4l2.c video_unregister_device(hva->vdev); hva 1351 drivers/media/platform/sti/hva/hva-v4l2.c struct hva_dev *hva; hva 1355 drivers/media/platform/sti/hva/hva-v4l2.c hva = devm_kzalloc(dev, sizeof(*hva), GFP_KERNEL); hva 1356 drivers/media/platform/sti/hva/hva-v4l2.c if (!hva) { hva 1365 drivers/media/platform/sti/hva/hva-v4l2.c hva->dev = dev; hva 1366 drivers/media/platform/sti/hva/hva-v4l2.c hva->pdev = pdev; hva 1367 drivers/media/platform/sti/hva/hva-v4l2.c platform_set_drvdata(pdev, hva); hva 1369 drivers/media/platform/sti/hva/hva-v4l2.c mutex_init(&hva->lock); hva 1372 drivers/media/platform/sti/hva/hva-v4l2.c ret = hva_hw_probe(pdev, hva); hva 1377 drivers/media/platform/sti/hva/hva-v4l2.c register_encoders(hva); hva 1380 drivers/media/platform/sti/hva/hva-v4l2.c register_formats(hva); hva 1383 drivers/media/platform/sti/hva/hva-v4l2.c ret = v4l2_device_register(dev, &hva->v4l2_dev); hva 1391 drivers/media/platform/sti/hva/hva-v4l2.c hva_debugfs_create(hva); hva 1394 drivers/media/platform/sti/hva/hva-v4l2.c hva->work_queue = create_workqueue(HVA_NAME); hva 1395 drivers/media/platform/sti/hva/hva-v4l2.c if (!hva->work_queue) { hva 1403 drivers/media/platform/sti/hva/hva-v4l2.c ret = hva_register_device(hva); hva 1408 drivers/media/platform/sti/hva/hva-v4l2.c HVA_NAME, hva->vdev->num); hva 1413 drivers/media/platform/sti/hva/hva-v4l2.c destroy_workqueue(hva->work_queue); hva 1416 drivers/media/platform/sti/hva/hva-v4l2.c hva_debugfs_remove(hva); hva 1418 drivers/media/platform/sti/hva/hva-v4l2.c v4l2_device_unregister(&hva->v4l2_dev); hva 1420 drivers/media/platform/sti/hva/hva-v4l2.c hva_hw_remove(hva); hva 1427 drivers/media/platform/sti/hva/hva-v4l2.c struct hva_dev *hva = platform_get_drvdata(pdev); hva 1428 drivers/media/platform/sti/hva/hva-v4l2.c struct device *dev = hva_to_dev(hva); hva 1430 drivers/media/platform/sti/hva/hva-v4l2.c hva_unregister_device(hva); hva 1432 drivers/media/platform/sti/hva/hva-v4l2.c destroy_workqueue(hva->work_queue); hva 1434 drivers/media/platform/sti/hva/hva-v4l2.c hva_hw_remove(hva); hva 1437 drivers/media/platform/sti/hva/hva-v4l2.c hva_debugfs_remove(hva); hva 1440 drivers/media/platform/sti/hva/hva-v4l2.c v4l2_device_unregister(&hva->v4l2_dev); hva 401 drivers/media/platform/sti/hva/hva.h void hva_debugfs_create(struct hva_dev *hva); hva 402 drivers/media/platform/sti/hva/hva.h void hva_debugfs_remove(struct hva_dev *hva); hva 1870 drivers/vhost/vhost.c static int log_write_hva(struct vhost_virtqueue *vq, u64 hva, u64 len) hva 1884 drivers/vhost/vhost.c if (u->userspace_addr > hva - 1 + len || hva 1885 drivers/vhost/vhost.c u->userspace_addr - 1 + u->size < hva) hva 1887 drivers/vhost/vhost.c start = max(u->userspace_addr, hva); hva 1889 drivers/vhost/vhost.c hva - 1 + len); hva 1904 drivers/vhost/vhost.c hva += min; hva 216 include/linux/kvm_host.h unsigned long hva, struct kvm_arch_async_pf *arch); hva 239 include/linux/kvm_host.h void *hva; hva 250 include/linux/kvm_host.h return !!map->hva; hva 1057 include/linux/kvm_host.h hva_to_gfn_memslot(unsigned long hva, struct kvm_memory_slot *slot) hva 1059 include/linux/kvm_host.h gfn_t gfn_offset = (hva - slot->userspace_addr) >> PAGE_SHIFT; hva 1087 include/linux/kvm_host.h unsigned long hva = gfn_to_hva(kvm, gpa_to_gfn(gpa)); hva 1089 include/linux/kvm_host.h return kvm_is_error_hva(hva); hva 47 include/linux/kvm_types.h unsigned long hva; hva 263 include/trace/events/kvm.h __field( u64, hva ) hva 272 include/trace/events/kvm.h __entry->hva = ((gfn - slot->base_gfn) << hva 278 include/trace/events/kvm.h __entry->hva, __entry->gfn, __entry->level, hva 80 tools/testing/selftests/kvm/include/kvm_util.h int kvm_memcmp_hva_gva(void *hva, struct kvm_vm *vm, const vm_vaddr_t gva, hva 110 tools/testing/selftests/kvm/include/kvm_util.h vm_paddr_t addr_hva2gpa(struct kvm_vm *vm, void *hva); hva 506 tools/testing/selftests/kvm/lib/kvm_util.c int kvm_memcmp_hva_gva(void *hva, struct kvm_vm *vm, vm_vaddr_t gva, size_t len) hva 515 tools/testing/selftests/kvm/lib/kvm_util.c uintptr_t ptr1 = (uintptr_t)hva + offset; hva 1061 tools/testing/selftests/kvm/lib/kvm_util.c vm_paddr_t addr_hva2gpa(struct kvm_vm *vm, void *hva) hva 1066 tools/testing/selftests/kvm/lib/kvm_util.c if ((hva >= region->host_mem) hva 1067 tools/testing/selftests/kvm/lib/kvm_util.c && (hva <= (region->host_mem hva 1071 tools/testing/selftests/kvm/lib/kvm_util.c + (hva - (uintptr_t) region->host_mem)); hva 1075 tools/testing/selftests/kvm/lib/kvm_util.c "hva: %p", hva); hva 926 virt/kvm/arm/mmu.c hva_t hva = memslot->userspace_addr; hva 929 virt/kvm/arm/mmu.c hva_t reg_end = hva + size; hva 944 virt/kvm/arm/mmu.c struct vm_area_struct *vma = find_vma(current->mm, hva); hva 953 virt/kvm/arm/mmu.c vm_start = max(hva, vma->vm_start); hva 960 virt/kvm/arm/mmu.c hva = vm_end; hva 961 virt/kvm/arm/mmu.c } while (hva < reg_end); hva 1613 virt/kvm/arm/mmu.c unsigned long hva, hva 1665 virt/kvm/arm/mmu.c return (hva & ~(map_size - 1)) >= uaddr_start && hva 1666 virt/kvm/arm/mmu.c (hva & ~(map_size - 1)) + map_size <= uaddr_end; hva 1670 virt/kvm/arm/mmu.c struct kvm_memory_slot *memslot, unsigned long hva, hva 1697 virt/kvm/arm/mmu.c vma = find_vma_intersection(current->mm, hva, hva + 1); hva 1699 virt/kvm/arm/mmu.c kvm_err("Failed to find VMA for hva 0x%lx\n", hva); hva 1707 virt/kvm/arm/mmu.c !fault_supports_stage2_huge_mapping(memslot, hva, vma_pagesize)) { hva 1744 virt/kvm/arm/mmu.c kvm_send_hwpoison_signal(hva, vma); hva 1785 virt/kvm/arm/mmu.c if (fault_supports_stage2_huge_mapping(memslot, hva, PMD_SIZE) && hva 1913 virt/kvm/arm/mmu.c unsigned long hva; hva 1955 virt/kvm/arm/mmu.c hva = gfn_to_hva_memslot_prot(memslot, gfn, &writable); hva 1957 virt/kvm/arm/mmu.c if (kvm_is_error_hva(hva) || (write_fault && !writable)) { hva 2000 virt/kvm/arm/mmu.c ret = user_mem_abort(vcpu, fault_ipa, memslot, hva, fault_status); hva 2079 virt/kvm/arm/mmu.c int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte) hva 2081 virt/kvm/arm/mmu.c unsigned long end = hva + PAGE_SIZE; hva 2088 virt/kvm/arm/mmu.c trace_kvm_set_spte_hva(hva); hva 2096 virt/kvm/arm/mmu.c handle_hva_to_gpa(kvm, hva, end, &kvm_set_spte_handler, &stage2_pte); hva 2145 virt/kvm/arm/mmu.c int kvm_test_age_hva(struct kvm *kvm, unsigned long hva) hva 2149 virt/kvm/arm/mmu.c trace_kvm_test_age_hva(hva); hva 2150 virt/kvm/arm/mmu.c return handle_hva_to_gpa(kvm, hva, hva + PAGE_SIZE, hva 2281 virt/kvm/arm/mmu.c hva_t hva = mem->userspace_addr; hva 2282 virt/kvm/arm/mmu.c hva_t reg_end = hva + mem->memory_size; hva 2312 virt/kvm/arm/mmu.c struct vm_area_struct *vma = find_vma(current->mm, hva); hva 2330 virt/kvm/arm/mmu.c vm_start = max(hva, vma->vm_start); hva 2353 virt/kvm/arm/mmu.c hva = vm_end; hva 2354 virt/kvm/arm/mmu.c } while (hva < reg_end); hva 157 virt/kvm/arm/trace.h TP_PROTO(unsigned long hva), hva 158 virt/kvm/arm/trace.h TP_ARGS(hva), hva 161 virt/kvm/arm/trace.h __field( unsigned long, hva ) hva 165 virt/kvm/arm/trace.h __entry->hva = hva; hva 168 virt/kvm/arm/trace.h TP_printk("mmu notifier set pte hva: %#08lx", __entry->hva) hva 190 virt/kvm/arm/trace.h TP_PROTO(unsigned long hva), hva 191 virt/kvm/arm/trace.h TP_ARGS(hva), hva 194 virt/kvm/arm/trace.h __field( unsigned long, hva ) hva 198 virt/kvm/arm/trace.h __entry->hva = hva; hva 201 virt/kvm/arm/trace.h TP_printk("mmu notifier test age hva: %#08lx", __entry->hva) hva 169 virt/kvm/async_pf.c unsigned long hva, struct kvm_arch_async_pf *arch) hva 189 virt/kvm/async_pf.c work->addr = hva; hva 1483 virt/kvm/kvm_main.c unsigned long hva = __gfn_to_hva_many(slot, gfn, NULL, false); hva 1485 virt/kvm/kvm_main.c if (!kvm_is_error_hva(hva) && writable) hva 1488 virt/kvm/kvm_main.c return hva; hva 1849 virt/kvm/kvm_main.c void *hva = NULL; hva 1876 virt/kvm/kvm_main.c hva = kmap_atomic(page); hva 1878 virt/kvm/kvm_main.c hva = kmap(page); hva 1881 virt/kvm/kvm_main.c hva = memremap(pfn_to_hpa(pfn), PAGE_SIZE, MEMREMAP_WB); hva 1887 virt/kvm/kvm_main.c if (!hva) hva 1891 virt/kvm/kvm_main.c map->hva = hva; hva 1921 virt/kvm/kvm_main.c if (!map->hva) hva 1926 virt/kvm/kvm_main.c kunmap_atomic(map->hva); hva 1932 virt/kvm/kvm_main.c memunmap(map->hva); hva 1945 virt/kvm/kvm_main.c map->hva = NULL; hva 2239 virt/kvm/kvm_main.c ghc->hva = KVM_HVA_ERR_BAD; hva 2247 virt/kvm/kvm_main.c ghc->hva = gfn_to_hva_many(ghc->memslot, start_gfn, hva 2249 virt/kvm/kvm_main.c if (kvm_is_error_hva(ghc->hva)) hva 2256 virt/kvm/kvm_main.c ghc->hva += offset; hva 2284 virt/kvm/kvm_main.c if (kvm_is_error_hva(ghc->hva)) hva 2290 virt/kvm/kvm_main.c r = __copy_to_user((void __user *)ghc->hva + offset, data, len); hva 2317 virt/kvm/kvm_main.c if (kvm_is_error_hva(ghc->hva)) hva 2323 virt/kvm/kvm_main.c r = __copy_from_user(data, (void __user *)ghc->hva, len);