bo_va 42 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h struct amdgpu_bo_va *bo_va; bo_va 89 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (entry->bo_va->base.vm == avm) bo_va 398 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c bo_va_entry->bo_va = amdgpu_vm_bo_add(adev, vm, bo); bo_va 399 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (!bo_va_entry->bo_va) { bo_va 425 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_vm_bo_rmv(adev, bo_va_entry->bo_va); bo_va 438 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_vm_bo_rmv(adev, entry->bo_va); bo_va 629 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if ((vm && vm != entry->bo_va->base.vm) || bo_va 651 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if ((vm && vm != entry->bo_va->base.vm) || bo_va 656 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_vm_get_pd_bo(entry->bo_va->base.vm, &ctx->list, bo_va 710 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_bo_va *bo_va = entry->bo_va; bo_va 711 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_vm *vm = bo_va->base.vm; bo_va 713 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_vm_bo_unmap(adev, bo_va, entry->va); bo_va 715 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_vm_clear_freed(adev, vm, &bo_va->last_pt_update); bo_va 717 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_sync_fence(NULL, sync, bo_va->last_pt_update, false); bo_va 727 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_bo_va *bo_va = entry->bo_va; bo_va 730 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = amdgpu_vm_bo_update(adev, bo_va, false); bo_va 736 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c return amdgpu_sync_fence(NULL, sync, bo_va->last_pt_update, false); bo_va 746 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = amdgpu_vm_bo_map(adev, entry->bo_va, entry->va, 0, bo_va 747 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_bo_size(entry->bo_va->base.bo), bo_va 1389 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (entry->bo_va->base.vm == vm && !entry->is_mapped) { bo_va 1468 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (entry->bo_va->base.vm == vm && entry->is_mapped) { bo_va 36 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.h struct amdgpu_bo_va *bo_va; bo_va 692 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c e->bo_va = amdgpu_vm_bo_find(vm, bo); bo_va 791 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c struct amdgpu_bo_va *bo_va; bo_va 876 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c bo_va = fpriv->csa_va; bo_va 877 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c BUG_ON(!bo_va); bo_va 878 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c r = amdgpu_vm_bo_update(adev, bo_va, false); bo_va 882 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c f = bo_va->last_pt_update; bo_va 896 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c bo_va = e->bo_va; bo_va 897 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (bo_va == NULL) bo_va 900 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c r = amdgpu_vm_bo_update(adev, bo_va, false); bo_va 904 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c f = bo_va->last_pt_update; bo_va 1724 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (!mapping || !mapping->bo_va || !mapping->bo_va->base.bo) bo_va 1727 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c *bo = mapping->bo_va->base.bo; bo_va 66 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c struct amdgpu_bo *bo, struct amdgpu_bo_va **bo_va, bo_va 89 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c *bo_va = amdgpu_vm_bo_add(adev, vm, bo); bo_va 90 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c if (!*bo_va) { bo_va 96 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c r = amdgpu_vm_bo_map(adev, *bo_va, csa_addr, 0, size, bo_va 102 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c amdgpu_vm_bo_rmv(adev, *bo_va); bo_va 35 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.h struct amdgpu_bo *bo, struct amdgpu_bo_va **bo_va, bo_va 128 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c struct amdgpu_bo_va *bo_va; bo_va 144 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c bo_va = amdgpu_vm_bo_find(vm, abo); bo_va 145 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c if (!bo_va) { bo_va 146 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c bo_va = amdgpu_vm_bo_add(adev, vm, abo); bo_va 148 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c ++bo_va->ref_count; bo_va 166 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c struct amdgpu_bo_va *bo_va; bo_va 184 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c bo_va = amdgpu_vm_bo_find(vm, bo); bo_va 185 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c if (bo_va && --bo_va->ref_count == 0) { bo_va 186 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c amdgpu_vm_bo_rmv(adev, bo_va); bo_va 511 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c struct amdgpu_bo_va *bo_va, bo_va 525 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c r = amdgpu_vm_bo_update(adev, bo_va, false); bo_va 551 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c struct amdgpu_bo_va *bo_va; bo_va 621 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c bo_va = amdgpu_vm_bo_find(&fpriv->vm, abo); bo_va 622 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c if (!bo_va) { bo_va 627 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c bo_va = fpriv->prt_va; bo_va 629 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c bo_va = NULL; bo_va 635 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c r = amdgpu_vm_bo_map(adev, bo_va, args->va_address, bo_va 640 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c r = amdgpu_vm_bo_unmap(adev, bo_va, args->va_address); bo_va 650 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c r = amdgpu_vm_bo_replace_map(adev, bo_va, args->va_address, bo_va 658 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c amdgpu_gem_va_update_vm(adev, &fpriv->vm, bo_va, bo_va 49 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h struct amdgpu_bo_va *bo_va; bo_va 244 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h TP_PROTO(struct amdgpu_bo_va *bo_va, bo_va 246 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h TP_ARGS(bo_va, mapping), bo_va 256 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __entry->bo = bo_va ? bo_va->base.bo : NULL; bo_va 268 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h TP_PROTO(struct amdgpu_bo_va *bo_va, bo_va 270 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h TP_ARGS(bo_va, mapping), bo_va 280 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __entry->bo = bo_va ? bo_va->base.bo : NULL; bo_va 1682 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo_va *bo_va, bo_va 1685 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo *bo = bo_va->base.bo; bo_va 1686 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_vm *vm = bo_va->base.vm; bo_va 1722 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c last_update = &bo_va->last_pt_update; bo_va 1724 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (!clear && bo_va->base.moved) { bo_va 1725 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c bo_va->base.moved = false; bo_va 1726 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_splice_init(&bo_va->valids, &bo_va->invalids); bo_va 1728 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c } else if (bo_va->cleared != clear) { bo_va 1729 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_splice_init(&bo_va->valids, &bo_va->invalids); bo_va 1732 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_for_each_entry(mapping, &bo_va->invalids, list) { bo_va 1754 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_vm_bo_evicted(&bo_va->base); bo_va 1756 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_vm_bo_idle(&bo_va->base); bo_va 1758 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_vm_bo_done(&bo_va->base); bo_va 1761 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_splice_init(&bo_va->invalids, &bo_va->valids); bo_va 1762 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c bo_va->cleared = clear; bo_va 1765 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_for_each_entry(mapping, &bo_va->valids, list) bo_va 1984 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo_va *bo_va, *tmp; bo_va 1989 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_for_each_entry_safe(bo_va, tmp, &vm->moved, base.vm_status) { bo_va 1991 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c r = amdgpu_vm_bo_update(adev, bo_va, false); bo_va 1998 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c bo_va = list_first_entry(&vm->invalidated, struct amdgpu_bo_va, bo_va 2000 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c resv = bo_va->base.bo->tbo.base.resv; bo_va 2010 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c r = amdgpu_vm_bo_update(adev, bo_va, clear); bo_va 2042 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo_va *bo_va; bo_va 2044 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c bo_va = kzalloc(sizeof(struct amdgpu_bo_va), GFP_KERNEL); bo_va 2045 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (bo_va == NULL) { bo_va 2048 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_vm_bo_base_init(&bo_va->base, vm, bo); bo_va 2050 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c bo_va->ref_count = 1; bo_va 2051 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c INIT_LIST_HEAD(&bo_va->valids); bo_va 2052 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c INIT_LIST_HEAD(&bo_va->invalids); bo_va 2056 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c bo_va->is_xgmi = true; bo_va 2064 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c return bo_va; bo_va 2078 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo_va *bo_va, bo_va 2081 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_vm *vm = bo_va->base.vm; bo_va 2082 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo *bo = bo_va->base.bo; bo_va 2084 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c mapping->bo_va = bo_va; bo_va 2085 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_add(&mapping->list, &bo_va->invalids); bo_va 2092 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c !bo_va->base.moved) { bo_va 2093 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_move(&bo_va->base.vm_status, &vm->moved); bo_va 2095 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c trace_amdgpu_vm_bo_map(bo_va, mapping); bo_va 2116 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo_va *bo_va, bo_va 2121 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo *bo = bo_va->base.bo; bo_va 2122 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_vm *vm = bo_va->base.vm; bo_va 2157 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_vm_bo_insert_map(adev, bo_va, mapping); bo_va 2181 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo_va *bo_va, bo_va 2186 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo *bo = bo_va->base.bo; bo_va 2206 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c r = amdgpu_vm_bo_clear_mappings(adev, bo_va->base.vm, saddr, size); bo_va 2220 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_vm_bo_insert_map(adev, bo_va, mapping); bo_va 2240 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo_va *bo_va, bo_va 2244 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_vm *vm = bo_va->base.vm; bo_va 2249 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_for_each_entry(mapping, &bo_va->valids, list) { bo_va 2254 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (&mapping->list == &bo_va->valids) { bo_va 2257 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_for_each_entry(mapping, &bo_va->invalids, list) { bo_va 2262 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (&mapping->list == &bo_va->invalids) bo_va 2268 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c mapping->bo_va = NULL; bo_va 2269 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c trace_amdgpu_vm_bo_unmap(bo_va, mapping); bo_va 2275 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c bo_va->last_pt_update); bo_va 2327 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c before->bo_va = tmp->bo_va; bo_va 2328 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_add(&before->list, &tmp->bo_va->invalids); bo_va 2338 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c after->bo_va = tmp->bo_va; bo_va 2339 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_add(&after->list, &tmp->bo_va->invalids); bo_va 2358 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c tmp->bo_va = NULL; bo_va 2419 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (mapping->bo_va && mapping->bo_va->base.bo) { bo_va 2422 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c bo = mapping->bo_va->base.bo; bo_va 2443 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo_va *bo_va) bo_va 2446 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_bo *bo = bo_va->base.bo; bo_va 2447 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct amdgpu_vm *vm = bo_va->base.vm; bo_va 2454 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c for (base = &bo_va->base.bo->vm_bo; *base; bo_va 2456 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (*base != &bo_va->base) bo_va 2459 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c *base = bo_va->base.next; bo_va 2465 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_del(&bo_va->base.vm_status); bo_va 2468 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_for_each_entry_safe(mapping, next, &bo_va->valids, list) { bo_va 2471 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c mapping->bo_va = NULL; bo_va 2472 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c trace_amdgpu_vm_bo_unmap(bo_va, mapping); bo_va 2475 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c list_for_each_entry_safe(mapping, next, &bo_va->invalids, list) { bo_va 2479 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c bo_va->last_pt_update); bo_va 2482 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c dma_fence_put(bo_va->last_pt_update); bo_va 2484 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (bo && bo_va->is_xgmi) { bo_va 2491 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c kfree(bo_va); bo_va 368 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h struct amdgpu_bo_va *bo_va, bo_va 379 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h struct amdgpu_bo_va *bo_va, bo_va 383 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h struct amdgpu_bo_va *bo_va, bo_va 387 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h struct amdgpu_bo_va *bo_va, bo_va 396 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h struct amdgpu_bo_va *bo_va); bo_va 88 drivers/gpu/drm/lima/lima_vm.c struct lima_bo_va *bo_va, *ret = NULL; bo_va 90 drivers/gpu/drm/lima/lima_vm.c list_for_each_entry(bo_va, &bo->va, list) { bo_va 91 drivers/gpu/drm/lima/lima_vm.c if (bo_va->vm == vm) { bo_va 92 drivers/gpu/drm/lima/lima_vm.c ret = bo_va; bo_va 102 drivers/gpu/drm/lima/lima_vm.c struct lima_bo_va *bo_va; bo_va 107 drivers/gpu/drm/lima/lima_vm.c bo_va = lima_vm_bo_find(vm, bo); bo_va 108 drivers/gpu/drm/lima/lima_vm.c if (bo_va) { bo_va 109 drivers/gpu/drm/lima/lima_vm.c bo_va->ref_count++; bo_va 120 drivers/gpu/drm/lima/lima_vm.c bo_va = kzalloc(sizeof(*bo_va), GFP_KERNEL); bo_va 121 drivers/gpu/drm/lima/lima_vm.c if (!bo_va) { bo_va 126 drivers/gpu/drm/lima/lima_vm.c bo_va->vm = vm; bo_va 127 drivers/gpu/drm/lima/lima_vm.c bo_va->ref_count = 1; bo_va 131 drivers/gpu/drm/lima/lima_vm.c err = drm_mm_insert_node(&vm->mm, &bo_va->node, bo->gem.size); bo_va 135 drivers/gpu/drm/lima/lima_vm.c err = lima_vm_map_page_table(vm, bo->pages_dma_addr, bo_va->node.start, bo_va 136 drivers/gpu/drm/lima/lima_vm.c bo_va->node.start + bo_va->node.size - 1); bo_va 142 drivers/gpu/drm/lima/lima_vm.c list_add_tail(&bo_va->list, &bo->va); bo_va 148 drivers/gpu/drm/lima/lima_vm.c drm_mm_remove_node(&bo_va->node); bo_va 151 drivers/gpu/drm/lima/lima_vm.c kfree(bo_va); bo_va 159 drivers/gpu/drm/lima/lima_vm.c struct lima_bo_va *bo_va; bo_va 163 drivers/gpu/drm/lima/lima_vm.c bo_va = lima_vm_bo_find(vm, bo); bo_va 164 drivers/gpu/drm/lima/lima_vm.c if (--bo_va->ref_count > 0) { bo_va 171 drivers/gpu/drm/lima/lima_vm.c lima_vm_unmap_page_table(vm, bo_va->node.start, bo_va 172 drivers/gpu/drm/lima/lima_vm.c bo_va->node.start + bo_va->node.size - 1); bo_va 174 drivers/gpu/drm/lima/lima_vm.c drm_mm_remove_node(&bo_va->node); bo_va 178 drivers/gpu/drm/lima/lima_vm.c list_del(&bo_va->list); bo_va 182 drivers/gpu/drm/lima/lima_vm.c kfree(bo_va); bo_va 187 drivers/gpu/drm/lima/lima_vm.c struct lima_bo_va *bo_va; bo_va 192 drivers/gpu/drm/lima/lima_vm.c bo_va = lima_vm_bo_find(vm, bo); bo_va 193 drivers/gpu/drm/lima/lima_vm.c ret = bo_va->node.start; bo_va 2853 drivers/gpu/drm/radeon/radeon.h struct radeon_bo_va *bo_va, bo_va 2863 drivers/gpu/drm/radeon/radeon.h struct radeon_bo_va *bo_va, bo_va 2867 drivers/gpu/drm/radeon/radeon.h struct radeon_bo_va *bo_va); bo_va 501 drivers/gpu/drm/radeon/radeon_cs.c struct radeon_bo_va *bo_va; bo_va 526 drivers/gpu/drm/radeon/radeon_cs.c bo_va = radeon_vm_bo_find(vm, bo); bo_va 527 drivers/gpu/drm/radeon/radeon_cs.c if (bo_va == NULL) { bo_va 532 drivers/gpu/drm/radeon/radeon_cs.c r = radeon_vm_bo_update(rdev, bo_va, &bo->tbo.mem); bo_va 536 drivers/gpu/drm/radeon/radeon_cs.c radeon_sync_fence(&p->ib.sync, bo_va->last_pt_update); bo_va 154 drivers/gpu/drm/radeon/radeon_gem.c struct radeon_bo_va *bo_va; bo_va 167 drivers/gpu/drm/radeon/radeon_gem.c bo_va = radeon_vm_bo_find(vm, rbo); bo_va 168 drivers/gpu/drm/radeon/radeon_gem.c if (!bo_va) { bo_va 169 drivers/gpu/drm/radeon/radeon_gem.c bo_va = radeon_vm_bo_add(rdev, vm, rbo); bo_va 171 drivers/gpu/drm/radeon/radeon_gem.c ++bo_va->ref_count; bo_va 185 drivers/gpu/drm/radeon/radeon_gem.c struct radeon_bo_va *bo_va; bo_va 199 drivers/gpu/drm/radeon/radeon_gem.c bo_va = radeon_vm_bo_find(vm, rbo); bo_va 200 drivers/gpu/drm/radeon/radeon_gem.c if (bo_va) { bo_va 201 drivers/gpu/drm/radeon/radeon_gem.c if (--bo_va->ref_count == 0) { bo_va 202 drivers/gpu/drm/radeon/radeon_gem.c radeon_vm_bo_rmv(rdev, bo_va); bo_va 550 drivers/gpu/drm/radeon/radeon_gem.c struct radeon_bo_va *bo_va) bo_va 561 drivers/gpu/drm/radeon/radeon_gem.c tv.bo = &bo_va->bo->tbo; bo_va 565 drivers/gpu/drm/radeon/radeon_gem.c vm_bos = radeon_vm_get_bos(rdev, bo_va->vm, &list); bo_va 581 drivers/gpu/drm/radeon/radeon_gem.c mutex_lock(&bo_va->vm->mutex); bo_va 582 drivers/gpu/drm/radeon/radeon_gem.c r = radeon_vm_clear_freed(rdev, bo_va->vm); bo_va 586 drivers/gpu/drm/radeon/radeon_gem.c if (bo_va->it.start) bo_va 587 drivers/gpu/drm/radeon/radeon_gem.c r = radeon_vm_bo_update(rdev, bo_va, &bo_va->bo->tbo.mem); bo_va 590 drivers/gpu/drm/radeon/radeon_gem.c mutex_unlock(&bo_va->vm->mutex); bo_va 610 drivers/gpu/drm/radeon/radeon_gem.c struct radeon_bo_va *bo_va; bo_va 673 drivers/gpu/drm/radeon/radeon_gem.c bo_va = radeon_vm_bo_find(&fpriv->vm, rbo); bo_va 674 drivers/gpu/drm/radeon/radeon_gem.c if (!bo_va) { bo_va 683 drivers/gpu/drm/radeon/radeon_gem.c if (bo_va->it.start) { bo_va 685 drivers/gpu/drm/radeon/radeon_gem.c args->offset = bo_va->it.start * RADEON_GPU_PAGE_SIZE; bo_va 689 drivers/gpu/drm/radeon/radeon_gem.c r = radeon_vm_bo_set_addr(rdev, bo_va, args->offset, args->flags); bo_va 692 drivers/gpu/drm/radeon/radeon_gem.c r = radeon_vm_bo_set_addr(rdev, bo_va, 0, 0); bo_va 698 drivers/gpu/drm/radeon/radeon_gem.c radeon_gem_va_update_vm(rdev, bo_va); bo_va 66 drivers/gpu/drm/radeon/radeon_trace.h TP_PROTO(struct radeon_bo_va *bo_va), bo_va 67 drivers/gpu/drm/radeon/radeon_trace.h TP_ARGS(bo_va), bo_va 75 drivers/gpu/drm/radeon/radeon_trace.h __entry->soffset = bo_va->it.start; bo_va 76 drivers/gpu/drm/radeon/radeon_trace.h __entry->eoffset = bo_va->it.last + 1; bo_va 77 drivers/gpu/drm/radeon/radeon_trace.h __entry->flags = bo_va->flags; bo_va 296 drivers/gpu/drm/radeon/radeon_vm.c struct radeon_bo_va *bo_va; bo_va 298 drivers/gpu/drm/radeon/radeon_vm.c list_for_each_entry(bo_va, &bo->va, bo_list) { bo_va 299 drivers/gpu/drm/radeon/radeon_vm.c if (bo_va->vm == vm) { bo_va 300 drivers/gpu/drm/radeon/radeon_vm.c return bo_va; bo_va 323 drivers/gpu/drm/radeon/radeon_vm.c struct radeon_bo_va *bo_va; bo_va 325 drivers/gpu/drm/radeon/radeon_vm.c bo_va = kzalloc(sizeof(struct radeon_bo_va), GFP_KERNEL); bo_va 326 drivers/gpu/drm/radeon/radeon_vm.c if (bo_va == NULL) { bo_va 329 drivers/gpu/drm/radeon/radeon_vm.c bo_va->vm = vm; bo_va 330 drivers/gpu/drm/radeon/radeon_vm.c bo_va->bo = bo; bo_va 331 drivers/gpu/drm/radeon/radeon_vm.c bo_va->it.start = 0; bo_va 332 drivers/gpu/drm/radeon/radeon_vm.c bo_va->it.last = 0; bo_va 333 drivers/gpu/drm/radeon/radeon_vm.c bo_va->flags = 0; bo_va 334 drivers/gpu/drm/radeon/radeon_vm.c bo_va->ref_count = 1; bo_va 335 drivers/gpu/drm/radeon/radeon_vm.c INIT_LIST_HEAD(&bo_va->bo_list); bo_va 336 drivers/gpu/drm/radeon/radeon_vm.c INIT_LIST_HEAD(&bo_va->vm_status); bo_va 339 drivers/gpu/drm/radeon/radeon_vm.c list_add_tail(&bo_va->bo_list, &bo->va); bo_va 342 drivers/gpu/drm/radeon/radeon_vm.c return bo_va; bo_va 447 drivers/gpu/drm/radeon/radeon_vm.c struct radeon_bo_va *bo_va, bo_va 451 drivers/gpu/drm/radeon/radeon_vm.c uint64_t size = radeon_bo_size(bo_va->bo); bo_va 452 drivers/gpu/drm/radeon/radeon_vm.c struct radeon_vm *vm = bo_va->vm; bo_va 483 drivers/gpu/drm/radeon/radeon_vm.c if (it && it != &bo_va->it) { bo_va 488 drivers/gpu/drm/radeon/radeon_vm.c "(bo %p 0x%010lx 0x%010lx)\n", bo_va->bo, bo_va 496 drivers/gpu/drm/radeon/radeon_vm.c if (bo_va->it.start || bo_va->it.last) { bo_va 505 drivers/gpu/drm/radeon/radeon_vm.c tmp->it.start = bo_va->it.start; bo_va 506 drivers/gpu/drm/radeon/radeon_vm.c tmp->it.last = bo_va->it.last; bo_va 508 drivers/gpu/drm/radeon/radeon_vm.c tmp->bo = radeon_bo_ref(bo_va->bo); bo_va 510 drivers/gpu/drm/radeon/radeon_vm.c interval_tree_remove(&bo_va->it, &vm->va); bo_va 512 drivers/gpu/drm/radeon/radeon_vm.c bo_va->it.start = 0; bo_va 513 drivers/gpu/drm/radeon/radeon_vm.c bo_va->it.last = 0; bo_va 514 drivers/gpu/drm/radeon/radeon_vm.c list_del_init(&bo_va->vm_status); bo_va 521 drivers/gpu/drm/radeon/radeon_vm.c bo_va->it.start = soffset; bo_va 522 drivers/gpu/drm/radeon/radeon_vm.c bo_va->it.last = eoffset; bo_va 523 drivers/gpu/drm/radeon/radeon_vm.c list_add(&bo_va->vm_status, &vm->cleared); bo_va 525 drivers/gpu/drm/radeon/radeon_vm.c interval_tree_insert(&bo_va->it, &vm->va); bo_va 528 drivers/gpu/drm/radeon/radeon_vm.c bo_va->flags = flags; bo_va 538 drivers/gpu/drm/radeon/radeon_vm.c radeon_bo_unreserve(bo_va->bo); bo_va 581 drivers/gpu/drm/radeon/radeon_vm.c radeon_bo_unreserve(bo_va->bo); bo_va 913 drivers/gpu/drm/radeon/radeon_vm.c struct radeon_bo_va *bo_va, bo_va 916 drivers/gpu/drm/radeon/radeon_vm.c struct radeon_vm *vm = bo_va->vm; bo_va 923 drivers/gpu/drm/radeon/radeon_vm.c if (!bo_va->it.start) { bo_va 925 drivers/gpu/drm/radeon/radeon_vm.c bo_va->bo, vm); bo_va 931 drivers/gpu/drm/radeon/radeon_vm.c if (list_empty(&bo_va->vm_status)) { bo_va 935 drivers/gpu/drm/radeon/radeon_vm.c list_del_init(&bo_va->vm_status); bo_va 937 drivers/gpu/drm/radeon/radeon_vm.c list_del(&bo_va->vm_status); bo_va 938 drivers/gpu/drm/radeon/radeon_vm.c list_add(&bo_va->vm_status, &vm->cleared); bo_va 942 drivers/gpu/drm/radeon/radeon_vm.c bo_va->flags &= ~RADEON_VM_PAGE_VALID; bo_va 943 drivers/gpu/drm/radeon/radeon_vm.c bo_va->flags &= ~RADEON_VM_PAGE_SYSTEM; bo_va 944 drivers/gpu/drm/radeon/radeon_vm.c bo_va->flags &= ~RADEON_VM_PAGE_SNOOPED; bo_va 945 drivers/gpu/drm/radeon/radeon_vm.c if (bo_va->bo && radeon_ttm_tt_is_readonly(bo_va->bo->tbo.ttm)) bo_va 946 drivers/gpu/drm/radeon/radeon_vm.c bo_va->flags &= ~RADEON_VM_PAGE_WRITEABLE; bo_va 951 drivers/gpu/drm/radeon/radeon_vm.c bo_va->flags |= RADEON_VM_PAGE_VALID; bo_va 954 drivers/gpu/drm/radeon/radeon_vm.c bo_va->flags |= RADEON_VM_PAGE_SYSTEM; bo_va 955 drivers/gpu/drm/radeon/radeon_vm.c if (!(bo_va->bo->flags & (RADEON_GEM_GTT_WC | RADEON_GEM_GTT_UC))) bo_va 956 drivers/gpu/drm/radeon/radeon_vm.c bo_va->flags |= RADEON_VM_PAGE_SNOOPED; bo_va 965 drivers/gpu/drm/radeon/radeon_vm.c trace_radeon_vm_bo_update(bo_va); bo_va 967 drivers/gpu/drm/radeon/radeon_vm.c nptes = bo_va->it.last - bo_va->it.start + 1; bo_va 976 drivers/gpu/drm/radeon/radeon_vm.c flags = radeon_vm_page_flags(bo_va->flags); bo_va 1005 drivers/gpu/drm/radeon/radeon_vm.c if (!(bo_va->flags & RADEON_VM_PAGE_VALID)) { bo_va 1012 drivers/gpu/drm/radeon/radeon_vm.c r = radeon_vm_update_ptes(rdev, vm, &ib, bo_va->it.start, bo_va 1013 drivers/gpu/drm/radeon/radeon_vm.c bo_va->it.last + 1, addr, bo_va 1014 drivers/gpu/drm/radeon/radeon_vm.c radeon_vm_page_flags(bo_va->flags)); bo_va 1029 drivers/gpu/drm/radeon/radeon_vm.c radeon_vm_fence_pts(vm, bo_va->it.start, bo_va->it.last + 1, ib.fence); bo_va 1030 drivers/gpu/drm/radeon/radeon_vm.c radeon_fence_unref(&bo_va->last_pt_update); bo_va 1031 drivers/gpu/drm/radeon/radeon_vm.c bo_va->last_pt_update = radeon_fence_ref(ib.fence); bo_va 1051 drivers/gpu/drm/radeon/radeon_vm.c struct radeon_bo_va *bo_va; bo_va 1056 drivers/gpu/drm/radeon/radeon_vm.c bo_va = list_first_entry(&vm->freed, bo_va 1060 drivers/gpu/drm/radeon/radeon_vm.c r = radeon_vm_bo_update(rdev, bo_va, NULL); bo_va 1061 drivers/gpu/drm/radeon/radeon_vm.c radeon_bo_unref(&bo_va->bo); bo_va 1062 drivers/gpu/drm/radeon/radeon_vm.c radeon_fence_unref(&bo_va->last_pt_update); bo_va 1064 drivers/gpu/drm/radeon/radeon_vm.c list_del(&bo_va->vm_status); bo_va 1065 drivers/gpu/drm/radeon/radeon_vm.c kfree(bo_va); bo_va 1089 drivers/gpu/drm/radeon/radeon_vm.c struct radeon_bo_va *bo_va; bo_va 1094 drivers/gpu/drm/radeon/radeon_vm.c bo_va = list_first_entry(&vm->invalidated, bo_va 1098 drivers/gpu/drm/radeon/radeon_vm.c r = radeon_vm_bo_update(rdev, bo_va, NULL); bo_va 1120 drivers/gpu/drm/radeon/radeon_vm.c struct radeon_bo_va *bo_va) bo_va 1122 drivers/gpu/drm/radeon/radeon_vm.c struct radeon_vm *vm = bo_va->vm; bo_va 1124 drivers/gpu/drm/radeon/radeon_vm.c list_del(&bo_va->bo_list); bo_va 1127 drivers/gpu/drm/radeon/radeon_vm.c if (bo_va->it.start || bo_va->it.last) bo_va 1128 drivers/gpu/drm/radeon/radeon_vm.c interval_tree_remove(&bo_va->it, &vm->va); bo_va 1131 drivers/gpu/drm/radeon/radeon_vm.c list_del(&bo_va->vm_status); bo_va 1132 drivers/gpu/drm/radeon/radeon_vm.c if (bo_va->it.start || bo_va->it.last) { bo_va 1133 drivers/gpu/drm/radeon/radeon_vm.c bo_va->bo = radeon_bo_ref(bo_va->bo); bo_va 1134 drivers/gpu/drm/radeon/radeon_vm.c list_add(&bo_va->vm_status, &vm->freed); bo_va 1136 drivers/gpu/drm/radeon/radeon_vm.c radeon_fence_unref(&bo_va->last_pt_update); bo_va 1137 drivers/gpu/drm/radeon/radeon_vm.c kfree(bo_va); bo_va 1156 drivers/gpu/drm/radeon/radeon_vm.c struct radeon_bo_va *bo_va; bo_va 1158 drivers/gpu/drm/radeon/radeon_vm.c list_for_each_entry(bo_va, &bo->va, bo_list) { bo_va 1159 drivers/gpu/drm/radeon/radeon_vm.c spin_lock(&bo_va->vm->status_lock); bo_va 1160 drivers/gpu/drm/radeon/radeon_vm.c if (list_empty(&bo_va->vm_status) && bo_va 1161 drivers/gpu/drm/radeon/radeon_vm.c (bo_va->it.start || bo_va->it.last)) bo_va 1162 drivers/gpu/drm/radeon/radeon_vm.c list_add(&bo_va->vm_status, &bo_va->vm->invalidated); bo_va 1163 drivers/gpu/drm/radeon/radeon_vm.c spin_unlock(&bo_va->vm->status_lock); bo_va 1233 drivers/gpu/drm/radeon/radeon_vm.c struct radeon_bo_va *bo_va, *tmp; bo_va 1239 drivers/gpu/drm/radeon/radeon_vm.c rbtree_postorder_for_each_entry_safe(bo_va, tmp, bo_va 1241 drivers/gpu/drm/radeon/radeon_vm.c interval_tree_remove(&bo_va->it, &vm->va); bo_va 1242 drivers/gpu/drm/radeon/radeon_vm.c r = radeon_bo_reserve(bo_va->bo, false); bo_va 1244 drivers/gpu/drm/radeon/radeon_vm.c list_del_init(&bo_va->bo_list); bo_va 1245 drivers/gpu/drm/radeon/radeon_vm.c radeon_bo_unreserve(bo_va->bo); bo_va 1246 drivers/gpu/drm/radeon/radeon_vm.c radeon_fence_unref(&bo_va->last_pt_update); bo_va 1247 drivers/gpu/drm/radeon/radeon_vm.c kfree(bo_va); bo_va 1250 drivers/gpu/drm/radeon/radeon_vm.c list_for_each_entry_safe(bo_va, tmp, &vm->freed, vm_status) { bo_va 1251 drivers/gpu/drm/radeon/radeon_vm.c radeon_bo_unref(&bo_va->bo); bo_va 1252 drivers/gpu/drm/radeon/radeon_vm.c radeon_fence_unref(&bo_va->last_pt_update); bo_va 1253 drivers/gpu/drm/radeon/radeon_vm.c kfree(bo_va);