tbo 309 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c r = amdgpu_ttm_alloc_gart(&bo->tbo); tbo 196 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 221 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct dma_resv *resv = bo->tbo.base.resv; tbo 280 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (WARN(amdgpu_ttm_tt_get_usermm(bo->tbo.ttm), tbo 286 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 313 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev); tbo 349 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev); tbo 380 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c unsigned long bo_size = bo->tbo.mem.size; tbo 452 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c entry->bo = &bo->tbo; tbo 494 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = amdgpu_ttm_tt_set_userptr(bo->tbo.ttm, user_addr, 0); tbo 507 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = amdgpu_ttm_tt_get_user_pages(bo, bo->tbo.ttm->pages); tbo 519 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 525 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_ttm_tt_get_user_pages_done(bo->tbo.ttm); tbo 582 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ctx->kfd_bo.tv.bo = &bo->tbo; tbo 645 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ctx->kfd_bo.tv.bo = &bo->tbo; tbo 815 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c sync, pd->tbo.base.resv, tbo 890 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = dma_resv_reserve_shared(vm->root.base.bo->tbo.base.resv, 1); tbo 1063 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev); tbo 1183 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c bo->tbo.sg = sg; tbo 1184 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c bo->tbo.ttm->sg = sg; tbo 1230 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c unsigned long bo_size = mem->bo->tbo.mem.size; tbo 1285 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (mem->bo->tbo.sg) { tbo 1286 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c sg_free_table(mem->bo->tbo.sg); tbo 1287 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c kfree(mem->bo->tbo.sg); tbo 1291 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c drm_gem_object_put_unlocked(&mem->bo->tbo.base); tbo 1329 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) { tbo 1338 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c bo_size = bo->tbo.mem.size; tbo 1354 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) && tbo 1355 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c bo->tbo.mem.mem_type == TTM_PL_SYSTEM) tbo 1376 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c !amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) { tbo 1414 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (!amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) && !bo->pin_count) tbo 1442 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c unsigned long bo_size = mem->bo->tbo.mem.size; tbo 1493 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c !amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm) && !mem->bo->pin_count) tbo 1527 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) { tbo 1633 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c drm_gem_object_get(&bo->tbo.base); tbo 1706 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 1734 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = amdgpu_ttm_tt_get_user_pages(bo, bo->tbo.ttm->pages); tbo 1743 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_ttm_tt_get_user_pages_done(bo->tbo.ttm); tbo 1822 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c if (bo->tbo.ttm->pages[0]) { tbo 1824 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 2032 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = amdgpu_sync_fence(NULL, &sync_obj, bo->tbo.moving, false); tbo 2137 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = dma_resv_reserve_shared(gws_bo->tbo.base.resv, 1); tbo 103 drivers/gpu/drm/amd/amdgpu/amdgpu_benchmark.c r = amdgpu_ttm_alloc_gart(&sobj->tbo); tbo 122 drivers/gpu/drm/amd/amdgpu/amdgpu_benchmark.c r = amdgpu_ttm_alloc_gart(&dobj->tbo); tbo 107 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c usermm = amdgpu_ttm_tt_get_usermm(bo->tbo.ttm); tbo 121 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c entry->tv.bo = &bo->tbo; tbo 54 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c p->uf_entry.tv.bo = &bo->tbo; tbo 66 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) { tbo 401 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 405 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c .resv = bo->tbo.base.resv, tbo 437 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 468 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 480 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c other = amdgpu_mem_type_to_domain(bo->tbo.mem.mem_type); tbo 496 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 541 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c usermm = amdgpu_ttm_tt_get_usermm(bo->tbo.ttm); tbo 545 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (amdgpu_ttm_tt_is_userptr(bo->tbo.ttm) && tbo 549 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 553 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c amdgpu_ttm_tt_set_user_pages(bo->tbo.ttm, tbo 624 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c e->user_pages = kvmalloc_array(bo->tbo.ttm->num_pages, tbo 639 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c for (i = 0; i < bo->tbo.ttm->num_pages; i++) { tbo 640 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (bo->tbo.ttm->pages[i] != e->user_pages[i]) { tbo 711 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c r = amdgpu_ttm_alloc_gart(&uf->tbo); tbo 729 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c struct dma_resv *resv = bo->tbo.base.resv; tbo 1302 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c r |= !amdgpu_ttm_tt_get_user_pages_done(bo->tbo.ttm); tbo 1731 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c if (dma_resv_locking_ctx((*bo)->tbo.base.resv) != &parser->ticket) tbo 1737 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c r = ttm_bo_validate(&(*bo)->tbo, &(*bo)->placement, &ctx); tbo 1742 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c return amdgpu_ttm_alloc_gart(&(*bo)->tbo); tbo 77 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c csa_tv.bo = &bo->tbo; tbo 3408 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c if (shadow->tbo.mem.mem_type != TTM_PL_TT || tbo 3409 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c shadow->tbo.mem.start == AMDGPU_BO_INVALID_OFFSET || tbo 3410 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c shadow->parent->tbo.mem.mem_type != TTM_PL_VRAM) tbo 202 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c r = amdgpu_ttm_alloc_gart(&new_abo->tbo); tbo 208 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c r = dma_resv_get_fences_rcu(new_abo->tbo.base.resv, &work->excl, tbo 52 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c int npages = bo->tbo.num_pages; tbo 54 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c return drm_prime_pages_to_sg(bo->tbo.ttm->pages, npages); tbo 71 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c ret = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, tbo 108 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 122 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) || tbo 198 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 219 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c r = __dma_resv_make_exclusive(bo->tbo.base.resv); tbo 254 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 286 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 303 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 339 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) || tbo 388 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c bo->tbo.sg = sg; tbo 389 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c bo->tbo.ttm->sg = sg; tbo 396 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c return &bo->tbo.base; tbo 176 drivers/gpu/drm/amd/amdgpu/amdgpu_fb.c ret = amdgpu_ttm_alloc_gart(&abo->tbo); tbo 88 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c *obj = &bo->tbo.base; tbo 125 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c struct amdgpu_device *adev = amdgpu_ttm_adev(abo->tbo.bdev); tbo 132 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c mm = amdgpu_ttm_tt_get_usermm(abo->tbo.ttm); tbo 137 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c abo->tbo.base.resv != vm->root.base.bo->tbo.base.resv) tbo 158 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 172 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c tv.bo = &bo->tbo; tbo 255 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c resv = vm->root.base.bo->tbo.base.resv; tbo 321 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c r = amdgpu_ttm_tt_set_userptr(bo->tbo.ttm, args->addr, args->flags); tbo 332 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c r = amdgpu_ttm_tt_get_user_pages(bo, bo->tbo.ttm->pages); tbo 341 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 355 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c amdgpu_ttm_tt_get_user_pages_done(bo->tbo.ttm); tbo 375 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c if (amdgpu_ttm_tt_get_usermm(robj->tbo.ttm) || tbo 438 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c ret = dma_resv_wait_timeout_rcu(robj->tbo.base.resv, true, true, tbo 603 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c tv.bo = &abo->tbo; tbo 694 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c info.bo_size = robj->tbo.base.size; tbo 695 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c info.alignment = robj->tbo.mem.page_alignment << PAGE_SHIFT; tbo 709 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c if (amdgpu_ttm_tt_get_usermm(robj->tbo.ttm)) { tbo 715 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c if (amdgpu_xgmi_same_hive(amdgpu_ttm_adev(robj->tbo.bdev), tbo 716 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c amdgpu_ttm_adev(base->vm->root.base.bo->tbo.bdev))) { tbo 805 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c domain = amdgpu_mem_type_to_domain(bo->tbo.mem.mem_type); tbo 825 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c dma_buf = READ_ONCE(bo->tbo.base.dma_buf); tbo 826 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c attachment = READ_ONCE(bo->tbo.base.import_attach); tbo 34 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.h #define gem_to_amdgpu_bo(gobj) container_of((gobj), struct amdgpu_bo, tbo.base) tbo 44 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 47 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c switch (bo->tbo.mem.mem_type) { tbo 49 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c ttm = container_of(bo->tbo.ttm, struct ttm_dma_tt, ttm); tbo 59 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c *flags = amdgpu_ttm_tt_pde_flags(bo->tbo.ttm, &bo->tbo.mem); tbo 69 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 35 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c struct ttm_buffer_object *tbo; tbo 169 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c struct ttm_buffer_object *tbo, tbo 220 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c struct ttm_buffer_object *tbo, tbo 229 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c if ((&tbo->mem == mem || tbo->mem.mem_type != TTM_PL_TT) && tbo 245 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c node->tbo = tbo; tbo 249 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c r = amdgpu_gtt_mgr_alloc(man, tbo, place, mem); tbo 321 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c r = amdgpu_ttm_recover_gart(node->tbo); tbo 1079 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c amdgpu_pasid_free_delayed(pd->tbo.base.resv, pasid); tbo 179 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c if (!amdgpu_ttm_tt_affect_userptr(bo->tbo.ttm, start, end)) tbo 182 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c r = dma_resv_wait_timeout_rcu(bo->tbo.base.resv, tbo 279 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c if (amdgpu_ttm_tt_affect_userptr(bo->tbo.ttm, tbo 378 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 436 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 64 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 66 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c if (bo->tbo.mem.mem_type == TTM_PL_VRAM) { tbo 70 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c } else if (bo->tbo.mem.mem_type == TTM_PL_TT) { tbo 75 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c static void amdgpu_bo_destroy(struct ttm_buffer_object *tbo) tbo 77 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c struct amdgpu_device *adev = amdgpu_ttm_adev(tbo->bdev); tbo 78 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c struct amdgpu_bo *bo = ttm_to_amdgpu_bo(tbo); tbo 85 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c if (bo->tbo.base.import_attach) tbo 86 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); tbo 87 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c drm_gem_object_release(&bo->tbo.base); tbo 127 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c struct amdgpu_device *adev = amdgpu_ttm_adev(abo->tbo.bdev); tbo 277 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c r = amdgpu_ttm_alloc_gart(&(*bo_ptr)->tbo); tbo 385 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c ttm_bo_mem_put(&(*bo_ptr)->tbo, &(*bo_ptr)->tbo.mem); tbo 386 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c r = ttm_bo_mem_space(&(*bo_ptr)->tbo, &(*bo_ptr)->placement, tbo 387 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c &(*bo_ptr)->tbo.mem, &ctx); tbo 551 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c drm_gem_private_object_init(adev->ddev, &bo->tbo.base, size); tbo 566 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c bo->tbo.bdev = &adev->mman.bdev; tbo 573 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c bo->tbo.priority = 1; tbo 575 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c r = ttm_bo_init_reserved(&adev->mman.bdev, &bo->tbo, size, bp->type, tbo 582 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c bo->tbo.mem.mem_type == TTM_PL_VRAM && tbo 583 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c bo->tbo.mem.start < adev->gmc.visible_vram_size >> PAGE_SHIFT) tbo 590 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c bo->tbo.mem.placement & TTM_PL_FLAG_VRAM) { tbo 593 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c r = amdgpu_fill_buffer(bo, 0, bo->tbo.base.resv, &fence); tbo 598 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c dma_fence_put(bo->tbo.moving); tbo 599 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c bo->tbo.moving = dma_fence_get(fence); tbo 616 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c dma_resv_unlock(bo->tbo.base.resv); tbo 637 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c bp.resv = bo->tbo.base.resv; tbo 678 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c WARN_ON(dma_resv_lock((*bo_ptr)->tbo.base.resv, tbo 684 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c dma_resv_unlock((*bo_ptr)->tbo.base.resv); tbo 718 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 743 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c struct amdgpu_device *adev = amdgpu_ttm_adev(shadow->tbo.bdev); tbo 781 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c r = dma_resv_wait_timeout_rcu(bo->tbo.base.resv, false, false, tbo 786 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap); tbo 838 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c ttm_bo_get(&bo->tbo); tbo 850 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c struct ttm_buffer_object *tbo; tbo 855 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c tbo = &((*bo)->tbo); tbo 856 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c ttm_bo_put(tbo); tbo 885 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 889 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) tbo 909 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c uint32_t mem_type = bo->tbo.mem.mem_type; tbo 917 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c u64 domain_start = bo->tbo.bdev->man[mem_type].gpu_offset; tbo 944 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 952 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c domain = amdgpu_mem_type_to_domain(bo->tbo.mem.mem_type); tbo 994 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 1012 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 1123 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c return ttm_fbdev_mmap(vma, &bo->tbo); tbo 1139 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 1159 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c dma_resv_assert_held(bo->tbo.base.resv); tbo 1391 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c struct dma_resv *resv = bo->tbo.base.resv; tbo 1411 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 1416 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c amdgpu_sync_resv(adev, &sync, bo->tbo.base.resv, owner, false); tbo 1435 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c WARN_ON_ONCE(bo->tbo.mem.mem_type == TTM_PL_SYSTEM); tbo 1436 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c WARN_ON_ONCE(!dma_resv_is_locked(bo->tbo.base.resv) && tbo 1437 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c !bo->pin_count && bo->tbo.type != ttm_bo_type_kernel); tbo 1438 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c WARN_ON_ONCE(bo->tbo.mem.start == AMDGPU_BO_INVALID_OFFSET); tbo 1439 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c WARN_ON_ONCE(bo->tbo.mem.mem_type == TTM_PL_VRAM && tbo 1442 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c return amdgpu_gmc_sign_extend(bo->tbo.offset); tbo 85 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h struct ttm_buffer_object tbo; tbo 111 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h static inline struct amdgpu_bo *ttm_to_amdgpu_bo(struct ttm_buffer_object *tbo) tbo 113 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h return container_of(tbo, struct amdgpu_bo, tbo); tbo 154 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 157 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h r = __ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL); tbo 168 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h ttm_bo_unreserve(&bo->tbo); tbo 173 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h return bo->tbo.num_pages << PAGE_SHIFT; tbo 178 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h return (bo->tbo.num_pages << PAGE_SHIFT) / AMDGPU_GPU_PAGE_SIZE; tbo 183 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h return (bo->tbo.mem.page_alignment << PAGE_SHIFT) / AMDGPU_GPU_PAGE_SIZE; tbo 194 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h return drm_vma_node_offset_addr(&bo->tbo.base.vma_node); tbo 202 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 204 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h struct drm_mm_node *node = bo->tbo.mem.mm_node; tbo 207 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h if (bo->tbo.mem.mem_type != TTM_PL_VRAM) tbo 210 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h for (pages_left = bo->tbo.mem.num_pages; pages_left; tbo 106 drivers/gpu/drm/amd/amdgpu/amdgpu_test.c r = amdgpu_ttm_alloc_gart(>t_obj[i]->tbo); tbo 130 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __entry->pages = bo->tbo.num_pages; tbo 131 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __entry->type = bo->tbo.mem.mem_type; tbo 231 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c return drm_vma_node_verify_access(&abo->tbo.base.vma_node, tbo 790 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c struct ttm_tt *ttm = bo->tbo.ttm; tbo 1003 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c struct ttm_buffer_object *tbo, tbo 1006 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c struct amdgpu_bo *abo = ttm_to_amdgpu_bo(tbo); tbo 1007 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c struct ttm_tt *ttm = tbo->ttm; tbo 1156 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c int amdgpu_ttm_recover_gart(struct ttm_buffer_object *tbo) tbo 1158 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c struct amdgpu_device *adev = amdgpu_ttm_adev(tbo->bdev); tbo 1162 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c if (!tbo->ttm) tbo 1165 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c flags = amdgpu_ttm_tt_pte_flags(adev, tbo->ttm, &tbo->mem); tbo 1166 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c r = amdgpu_ttm_gart_bind(adev, tbo, flags); tbo 1544 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c struct amdgpu_device *adev = amdgpu_ttm_adev(abo->tbo.bdev); tbo 1554 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c nodes = amdgpu_find_mm_node(&abo->tbo.mem, &offset); tbo 2042 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 2058 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c if (bo->tbo.mem.mem_type == TTM_PL_TT) { tbo 2059 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c r = amdgpu_ttm_alloc_gart(&bo->tbo); tbo 2064 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c num_pages = bo->tbo.num_pages; tbo 2065 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c mm_node = bo->tbo.mem.mm_node; tbo 2092 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c num_pages = bo->tbo.num_pages; tbo 2093 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c mm_node = bo->tbo.mem.mm_node; tbo 2099 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c dst_addr = amdgpu_mm_node_addr(&bo->tbo, mm_node, &bo->tbo.mem); tbo 104 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h int amdgpu_ttm_recover_gart(struct ttm_buffer_object *tbo); tbo 500 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &tctx); tbo 1041 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 1076 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c r = dma_resv_wait_timeout_rcu(bo->tbo.base.resv, tbo 1088 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c r = amdgpu_sync_resv(adev, &job->sync, bo->tbo.base.resv, tbo 610 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 206 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (bo->tbo.type == ttm_bo_type_kernel) tbo 305 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (bo->tbo.base.resv != vm->root.base.bo->tbo.base.resv) tbo 309 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (bo->tbo.type == ttm_bo_type_kernel && bo->parent) tbo 315 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c amdgpu_mem_type_to_domain(bo->tbo.mem.mem_type)) tbo 562 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c entry->tv.bo = &vm->root.base.bo->tbo; tbo 586 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (abo->tbo.base.resv == vm->root.base.bo->tbo.base.resv) tbo 622 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c ttm_bo_move_to_lru_tail(&bo->tbo, &vm->lru_bulk_move); tbo 624 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c ttm_bo_move_to_lru_tail(&bo->shadow->tbo, tbo 661 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (bo->tbo.type != ttm_bo_type_kernel) { tbo 745 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 750 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c r = ttm_bo_validate(&bo->shadow->tbo, &bo->shadow->placement, tbo 835 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c bp->resv = vm->root.base.bo->tbo.base.resv; tbo 1703 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c mem = &bo->tbo.mem; tbo 1706 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c ttm = container_of(bo->tbo.ttm, struct ttm_dma_tt, ttm); tbo 1709 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c exclusive = dma_resv_get_excl(bo->tbo.base.resv); tbo 1713 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c flags = amdgpu_ttm_tt_pte_flags(adev, bo->tbo.ttm, mem); tbo 1714 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c bo_adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 1719 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (clear || (bo && bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv)) tbo 1750 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (bo && bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv) { tbo 1751 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c uint32_t mem_type = bo->tbo.mem.mem_type; tbo 1886 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c struct dma_resv *resv = vm->root.base.bo->tbo.base.resv; tbo 2000 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c resv = bo_va->base.bo->tbo.base.resv; tbo 2054 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (bo && amdgpu_xgmi_same_hive(adev, amdgpu_ttm_adev(bo->tbo.bdev)) && tbo 2091 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (bo && bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv && tbo 2423 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (dma_resv_locking_ctx(bo->tbo.base.resv) != tbo 2451 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv) tbo 2515 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (evicted && bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv) { tbo 2524 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c if (bo->tbo.type == ttm_bo_type_kernel) tbo 2526 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c else if (bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv) tbo 2656 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c return dma_resv_wait_timeout_rcu(vm->root.base.bo->tbo.base.resv, tbo 2731 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c r = dma_resv_reserve_shared(root->tbo.base.resv, 1); tbo 40 drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c r = amdgpu_ttm_alloc_gart(&table->tbo); tbo 45 drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c r = amdgpu_ttm_alloc_gart(&table->shadow->tbo); tbo 75 drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c r = amdgpu_sync_resv(p->adev, &p->job->sync, root->tbo.base.resv, tbo 217 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); tbo 218 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c struct ttm_mem_reg *mem = &bo->tbo.mem; tbo 268 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c struct ttm_buffer_object *tbo, tbo 1078 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c memset(hpd, 0, adev->gfx.mec.hpd_eop_obj->tbo.mem.size); tbo 1745 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c memset(hpd, 0, adev->gfx.mec.hpd_eop_obj->tbo.mem.size); tbo 4513 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c adev = amdgpu_ttm_adev(rbo->tbo.bdev); tbo 4516 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c tv.bo = &rbo->tbo; tbo 4539 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c r = amdgpu_ttm_alloc_gart(&rbo->tbo); tbo 5766 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c r = dma_resv_wait_timeout_rcu(abo->tbo.base.resv, true, tbo 380 drivers/gpu/drm/qxl/qxl_cmd.c drm_gem_object_put_unlocked(&qdev->primary_bo->tbo.base); tbo 407 drivers/gpu/drm/qxl/qxl_cmd.c drm_gem_object_get(&qdev->primary_bo->tbo.base); tbo 598 drivers/gpu/drm/qxl/qxl_cmd.c ret = ttm_bo_wait(&surf->tbo, true, !stall); tbo 64 drivers/gpu/drm/qxl/qxl_debugfs.c fobj = rcu_dereference(bo->tbo.base.resv->fence); tbo 69 drivers/gpu/drm/qxl/qxl_debugfs.c (unsigned long)bo->tbo.base.size, tbo 800 drivers/gpu/drm/qxl/qxl_display.c (&qdev->dumb_shadow_bo->tbo.base); tbo 810 drivers/gpu/drm/qxl/qxl_display.c (&user_bo->shadow->tbo.base); tbo 813 drivers/gpu/drm/qxl/qxl_display.c drm_gem_object_get(&qdev->dumb_shadow_bo->tbo.base); tbo 844 drivers/gpu/drm/qxl/qxl_display.c drm_gem_object_put_unlocked(&user_bo->shadow->tbo.base); tbo 74 drivers/gpu/drm/qxl/qxl_drv.h struct ttm_buffer_object tbo; tbo 96 drivers/gpu/drm/qxl/qxl_drv.h #define gem_to_qxl_bo(gobj) container_of((gobj), struct qxl_bo, tbo.base) tbo 97 drivers/gpu/drm/qxl/qxl_drv.h #define to_qxl_bo(tobj) container_of((tobj), struct qxl_bo, tbo) tbo 310 drivers/gpu/drm/qxl/qxl_drv.h (bo->tbo.mem.mem_type == TTM_PL_VRAM) tbo 313 drivers/gpu/drm/qxl/qxl_drv.h WARN_ON_ONCE((bo->tbo.offset & slot->gpu_offset) != slot->gpu_offset); tbo 316 drivers/gpu/drm/qxl/qxl_drv.h return slot->high_bits | (bo->tbo.offset - slot->gpu_offset + offset); tbo 35 drivers/gpu/drm/qxl/qxl_gem.c struct ttm_buffer_object *tbo; tbo 41 drivers/gpu/drm/qxl/qxl_gem.c tbo = &qobj->tbo; tbo 42 drivers/gpu/drm/qxl/qxl_gem.c ttm_bo_put(tbo); tbo 66 drivers/gpu/drm/qxl/qxl_gem.c *obj = &qbo->tbo.base; tbo 331 drivers/gpu/drm/qxl/qxl_ioctl.c ret = ttm_bo_validate(&qobj->tbo, &qobj->placement, &ctx); tbo 30 drivers/gpu/drm/qxl/qxl_object.c static void qxl_ttm_bo_destroy(struct ttm_buffer_object *tbo) tbo 35 drivers/gpu/drm/qxl/qxl_object.c bo = to_qxl_bo(tbo); tbo 36 drivers/gpu/drm/qxl/qxl_object.c qdev = (struct qxl_device *)bo->tbo.base.dev->dev_private; tbo 43 drivers/gpu/drm/qxl/qxl_object.c drm_gem_object_release(&bo->tbo.base); tbo 98 drivers/gpu/drm/qxl/qxl_object.c r = drm_gem_object_init(&qdev->ddev, &bo->tbo.base, size); tbo 113 drivers/gpu/drm/qxl/qxl_object.c r = ttm_bo_init(&qdev->mman.bdev, &bo->tbo, size, type, tbo 138 drivers/gpu/drm/qxl/qxl_object.c r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap); tbo 151 drivers/gpu/drm/qxl/qxl_object.c struct ttm_mem_type_manager *man = &bo->tbo.bdev->man[bo->tbo.mem.mem_type]; tbo 156 drivers/gpu/drm/qxl/qxl_object.c if (bo->tbo.mem.mem_type == TTM_PL_VRAM) tbo 158 drivers/gpu/drm/qxl/qxl_object.c else if (bo->tbo.mem.mem_type == TTM_PL_PRIV) tbo 164 drivers/gpu/drm/qxl/qxl_object.c ret = ttm_mem_io_reserve(bo->tbo.bdev, &bo->tbo.mem); tbo 167 drivers/gpu/drm/qxl/qxl_object.c return io_mapping_map_atomic_wc(map, bo->tbo.mem.bus.offset + page_offset); tbo 196 drivers/gpu/drm/qxl/qxl_object.c struct ttm_mem_type_manager *man = &bo->tbo.bdev->man[bo->tbo.mem.mem_type]; tbo 198 drivers/gpu/drm/qxl/qxl_object.c if ((bo->tbo.mem.mem_type != TTM_PL_VRAM) && tbo 199 drivers/gpu/drm/qxl/qxl_object.c (bo->tbo.mem.mem_type != TTM_PL_PRIV)) tbo 205 drivers/gpu/drm/qxl/qxl_object.c ttm_mem_io_free(bo->tbo.bdev, &bo->tbo.mem); tbo 217 drivers/gpu/drm/qxl/qxl_object.c drm_gem_object_put_unlocked(&(*bo)->tbo.base); tbo 223 drivers/gpu/drm/qxl/qxl_object.c drm_gem_object_get(&bo->tbo.base); tbo 230 drivers/gpu/drm/qxl/qxl_object.c struct drm_device *ddev = bo->tbo.base.dev; tbo 238 drivers/gpu/drm/qxl/qxl_object.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 250 drivers/gpu/drm/qxl/qxl_object.c struct drm_device *ddev = bo->tbo.base.dev; tbo 262 drivers/gpu/drm/qxl/qxl_object.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 313 drivers/gpu/drm/qxl/qxl_object.c &bo->tbo.base, bo, (unsigned long)bo->tbo.base.size, tbo 314 drivers/gpu/drm/qxl/qxl_object.c *((unsigned long *)&bo->tbo.base.refcount)); tbo 319 drivers/gpu/drm/qxl/qxl_object.c drm_gem_object_put_unlocked(&bo->tbo.base); tbo 34 drivers/gpu/drm/qxl/qxl_object.h r = ttm_bo_reserve(&bo->tbo, true, no_wait, NULL); tbo 37 drivers/gpu/drm/qxl/qxl_object.h struct drm_device *ddev = bo->tbo.base.dev; tbo 48 drivers/gpu/drm/qxl/qxl_object.h ttm_bo_unreserve(&bo->tbo); tbo 53 drivers/gpu/drm/qxl/qxl_object.h return bo->tbo.offset; tbo 58 drivers/gpu/drm/qxl/qxl_object.h return bo->tbo.num_pages << PAGE_SHIFT; tbo 63 drivers/gpu/drm/qxl/qxl_object.h return drm_vma_node_offset_addr(&bo->tbo.base.vma_node); tbo 71 drivers/gpu/drm/qxl/qxl_object.h r = ttm_bo_reserve(&bo->tbo, true, no_wait, NULL); tbo 74 drivers/gpu/drm/qxl/qxl_object.h struct drm_device *ddev = bo->tbo.base.dev; tbo 82 drivers/gpu/drm/qxl/qxl_object.h *mem_type = bo->tbo.mem.mem_type; tbo 84 drivers/gpu/drm/qxl/qxl_object.h r = ttm_bo_wait(&bo->tbo, true, no_wait); tbo 85 drivers/gpu/drm/qxl/qxl_object.h ttm_bo_unreserve(&bo->tbo); tbo 214 drivers/gpu/drm/qxl/qxl_release.c if (entry->tv.bo == &bo->tbo) tbo 223 drivers/gpu/drm/qxl/qxl_release.c entry->tv.bo = &bo->tbo; tbo 236 drivers/gpu/drm/qxl/qxl_release.c ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 241 drivers/gpu/drm/qxl/qxl_release.c ret = dma_resv_reserve_shared(bo->tbo.base.resv, 1); tbo 246 drivers/gpu/drm/qxl/qxl_release.c ret = qxl_bo_check_id(bo->tbo.base.dev->dev_private, bo); tbo 158 drivers/gpu/drm/qxl/qxl_ttm.c return drm_vma_node_verify_access(&qbo->tbo.base.vma_node, tbo 300 drivers/gpu/drm/qxl/qxl_ttm.c qdev = qbo->tbo.base.dev->dev_private; tbo 493 drivers/gpu/drm/radeon/radeon.h struct ttm_buffer_object tbo; tbo 515 drivers/gpu/drm/radeon/radeon.h #define gem_to_radeon_bo(gobj) container_of((gobj), struct radeon_bo, tbo.base) tbo 125 drivers/gpu/drm/radeon/radeon_benchmark.c dobj->tbo.base.resv); tbo 136 drivers/gpu/drm/radeon/radeon_benchmark.c dobj->tbo.base.resv); tbo 163 drivers/gpu/drm/radeon/radeon_cs.c if (radeon_ttm_tt_has_userptr(p->relocs[i].robj->tbo.ttm)) { tbo 186 drivers/gpu/drm/radeon/radeon_cs.c p->relocs[i].tv.bo = &p->relocs[i].robj->tbo; tbo 260 drivers/gpu/drm/radeon/radeon_cs.c resv = reloc->robj->tbo.base.resv; tbo 404 drivers/gpu/drm/radeon/radeon_cs.c return (int)la->robj->tbo.num_pages - (int)lb->robj->tbo.num_pages; tbo 446 drivers/gpu/drm/radeon/radeon_cs.c drm_gem_object_put_unlocked(&bo->tbo.base); tbo 518 drivers/gpu/drm/radeon/radeon_cs.c &rdev->ring_tmp_bo.bo->tbo.mem); tbo 532 drivers/gpu/drm/radeon/radeon_cs.c r = radeon_vm_bo_update(rdev, bo_va, &bo->tbo.mem); tbo 280 drivers/gpu/drm/radeon/radeon_display.c drm_gem_object_put_unlocked(&work->old_rbo->tbo.base); tbo 538 drivers/gpu/drm/radeon/radeon_display.c work->fence = dma_fence_get(dma_resv_get_excl(new_rbo->tbo.base.resv)); tbo 612 drivers/gpu/drm/radeon/radeon_display.c drm_gem_object_put_unlocked(&work->old_rbo->tbo.base); tbo 86 drivers/gpu/drm/radeon/radeon_gem.c *obj = &robj->tbo.base; tbo 117 drivers/gpu/drm/radeon/radeon_gem.c r = dma_resv_wait_timeout_rcu(robj->tbo.base.resv, true, true, 30 * HZ); tbo 333 drivers/gpu/drm/radeon/radeon_gem.c r = radeon_ttm_tt_set_userptr(bo->tbo.ttm, args->addr, args->flags); tbo 352 drivers/gpu/drm/radeon/radeon_gem.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 422 drivers/gpu/drm/radeon/radeon_gem.c if (radeon_ttm_tt_has_userptr(robj->tbo.ttm)) { tbo 454 drivers/gpu/drm/radeon/radeon_gem.c r = dma_resv_test_signaled_rcu(robj->tbo.base.resv, true); tbo 460 drivers/gpu/drm/radeon/radeon_gem.c cur_placement = READ_ONCE(robj->tbo.mem.mem_type); tbo 483 drivers/gpu/drm/radeon/radeon_gem.c ret = dma_resv_wait_timeout_rcu(robj->tbo.base.resv, true, true, 30 * HZ); tbo 490 drivers/gpu/drm/radeon/radeon_gem.c cur_placement = READ_ONCE(robj->tbo.mem.mem_type); tbo 561 drivers/gpu/drm/radeon/radeon_gem.c tv.bo = &bo_va->bo->tbo; tbo 587 drivers/gpu/drm/radeon/radeon_gem.c r = radeon_vm_bo_update(rdev, bo_va, &bo_va->bo->tbo.mem); tbo 723 drivers/gpu/drm/radeon/radeon_gem.c if (radeon_ttm_tt_has_userptr(robj->tbo.ttm)) tbo 793 drivers/gpu/drm/radeon/radeon_gem.c domain = radeon_mem_type_to_domain(rbo->tbo.mem.mem_type); tbo 99 drivers/gpu/drm/radeon/radeon_mn.c if (!bo->tbo.ttm || bo->tbo.ttm->state != tt_bound) tbo 108 drivers/gpu/drm/radeon/radeon_mn.c r = dma_resv_wait_timeout_rcu(bo->tbo.base.resv, tbo 114 drivers/gpu/drm/radeon/radeon_mn.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 57 drivers/gpu/drm/radeon/radeon_object.c u64 size = (u64)bo->tbo.num_pages << PAGE_SHIFT; tbo 75 drivers/gpu/drm/radeon/radeon_object.c static void radeon_ttm_bo_destroy(struct ttm_buffer_object *tbo) tbo 79 drivers/gpu/drm/radeon/radeon_object.c bo = container_of(tbo, struct radeon_bo, tbo); tbo 81 drivers/gpu/drm/radeon/radeon_object.c radeon_update_memory_usage(bo, bo->tbo.mem.mem_type, -1); tbo 88 drivers/gpu/drm/radeon/radeon_object.c if (bo->tbo.base.import_attach) tbo 89 drivers/gpu/drm/radeon/radeon_object.c drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); tbo 90 drivers/gpu/drm/radeon/radeon_object.c drm_gem_object_release(&bo->tbo.base); tbo 212 drivers/gpu/drm/radeon/radeon_object.c drm_gem_private_object_init(rdev->ddev, &bo->tbo.base, size); tbo 262 drivers/gpu/drm/radeon/radeon_object.c r = ttm_bo_init(&rdev->mman.bdev, &bo->tbo, size, type, tbo 287 drivers/gpu/drm/radeon/radeon_object.c r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap); tbo 313 drivers/gpu/drm/radeon/radeon_object.c ttm_bo_get(&bo->tbo); tbo 319 drivers/gpu/drm/radeon/radeon_object.c struct ttm_buffer_object *tbo; tbo 323 drivers/gpu/drm/radeon/radeon_object.c tbo = &((*bo)->tbo); tbo 324 drivers/gpu/drm/radeon/radeon_object.c ttm_bo_put(tbo); tbo 334 drivers/gpu/drm/radeon/radeon_object.c if (radeon_ttm_tt_has_userptr(bo->tbo.ttm)) tbo 374 drivers/gpu/drm/radeon/radeon_object.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 410 drivers/gpu/drm/radeon/radeon_object.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 412 drivers/gpu/drm/radeon/radeon_object.c if (bo->tbo.mem.mem_type == TTM_PL_VRAM) tbo 445 drivers/gpu/drm/radeon/radeon_object.c &bo->tbo.base, bo, (unsigned long)bo->tbo.base.size, tbo 446 drivers/gpu/drm/radeon/radeon_object.c *((unsigned long *)&bo->tbo.base.refcount)); tbo 451 drivers/gpu/drm/radeon/radeon_object.c drm_gem_object_put_unlocked(&bo->tbo.base); tbo 556 drivers/gpu/drm/radeon/radeon_object.c radeon_mem_type_to_domain(bo->tbo.mem.mem_type); tbo 579 drivers/gpu/drm/radeon/radeon_object.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 613 drivers/gpu/drm/radeon/radeon_object.c dma_resv_assert_held(bo->tbo.base.resv); tbo 645 drivers/gpu/drm/radeon/radeon_object.c ttm_bo_unmap_virtual(&old_object->tbo); tbo 655 drivers/gpu/drm/radeon/radeon_object.c bo->tbo.mem.start << PAGE_SHIFT, tbo 656 drivers/gpu/drm/radeon/radeon_object.c bo->tbo.num_pages << PAGE_SHIFT); tbo 739 drivers/gpu/drm/radeon/radeon_object.c dma_resv_assert_held(bo->tbo.base.resv); tbo 751 drivers/gpu/drm/radeon/radeon_object.c dma_resv_assert_held(bo->tbo.base.resv); tbo 761 drivers/gpu/drm/radeon/radeon_object.c if (bo->tbo.mem.mem_type != TTM_PL_VRAM) { tbo 785 drivers/gpu/drm/radeon/radeon_object.c rbo = container_of(bo, struct radeon_bo, tbo); tbo 807 drivers/gpu/drm/radeon/radeon_object.c rbo = container_of(bo, struct radeon_bo, tbo); tbo 851 drivers/gpu/drm/radeon/radeon_object.c r = ttm_bo_reserve(&bo->tbo, true, no_wait, NULL); tbo 855 drivers/gpu/drm/radeon/radeon_object.c *mem_type = bo->tbo.mem.mem_type; tbo 857 drivers/gpu/drm/radeon/radeon_object.c r = ttm_bo_wait(&bo->tbo, true, no_wait); tbo 858 drivers/gpu/drm/radeon/radeon_object.c ttm_bo_unreserve(&bo->tbo); tbo 873 drivers/gpu/drm/radeon/radeon_object.c struct dma_resv *resv = bo->tbo.base.resv; tbo 68 drivers/gpu/drm/radeon/radeon_object.h r = ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL); tbo 79 drivers/gpu/drm/radeon/radeon_object.h ttm_bo_unreserve(&bo->tbo); tbo 93 drivers/gpu/drm/radeon/radeon_object.h return bo->tbo.offset; tbo 98 drivers/gpu/drm/radeon/radeon_object.h return bo->tbo.num_pages << PAGE_SHIFT; tbo 103 drivers/gpu/drm/radeon/radeon_object.h return (bo->tbo.num_pages << PAGE_SHIFT) / RADEON_GPU_PAGE_SIZE; tbo 108 drivers/gpu/drm/radeon/radeon_object.h return (bo->tbo.mem.page_alignment << PAGE_SHIFT) / RADEON_GPU_PAGE_SIZE; tbo 119 drivers/gpu/drm/radeon/radeon_object.h return drm_vma_node_offset_addr(&bo->tbo.base.vma_node); tbo 157 drivers/gpu/drm/radeon/radeon_pm.c if (bo->tbo.mem.mem_type == TTM_PL_VRAM) tbo 158 drivers/gpu/drm/radeon/radeon_pm.c ttm_bo_unmap_virtual(&bo->tbo); tbo 37 drivers/gpu/drm/radeon/radeon_prime.c int npages = bo->tbo.num_pages; tbo 39 drivers/gpu/drm/radeon/radeon_prime.c return drm_prime_pages_to_sg(bo->tbo.ttm->pages, npages); tbo 47 drivers/gpu/drm/radeon/radeon_prime.c ret = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, tbo 83 drivers/gpu/drm/radeon/radeon_prime.c return &bo->tbo.base; tbo 124 drivers/gpu/drm/radeon/radeon_prime.c if (radeon_ttm_tt_has_userptr(bo->tbo.ttm)) tbo 123 drivers/gpu/drm/radeon/radeon_test.c vram_obj->tbo.base.resv); tbo 127 drivers/gpu/drm/radeon/radeon_test.c vram_obj->tbo.base.resv); tbo 174 drivers/gpu/drm/radeon/radeon_test.c vram_obj->tbo.base.resv); tbo 178 drivers/gpu/drm/radeon/radeon_test.c vram_obj->tbo.base.resv); tbo 25 drivers/gpu/drm/radeon/radeon_trace.h __entry->pages = bo->tbo.num_pages; tbo 143 drivers/gpu/drm/radeon/radeon_ttm.c rbo = container_of(bo, struct radeon_bo, tbo); tbo 183 drivers/gpu/drm/radeon/radeon_ttm.c struct radeon_bo *rbo = container_of(bo, struct radeon_bo, tbo); tbo 187 drivers/gpu/drm/radeon/radeon_ttm.c return drm_vma_node_verify_access(&rbo->tbo.base.vma_node, tbo 353 drivers/gpu/drm/radeon/radeon_ttm.c rbo = container_of(bo, struct radeon_bo, tbo); tbo 480 drivers/gpu/drm/radeon/radeon_uvd.c f = dma_resv_get_excl(bo->tbo.base.resv); tbo 144 drivers/gpu/drm/radeon/radeon_vm.c list[0].tv.bo = &vm->page_directory->tbo; tbo 156 drivers/gpu/drm/radeon/radeon_vm.c list[idx].tv.bo = &list[idx].robj->tbo; tbo 400 drivers/gpu/drm/radeon/radeon_vm.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); tbo 705 drivers/gpu/drm/radeon/radeon_vm.c radeon_sync_resv(rdev, &ib.sync, pd->tbo.base.resv, true); tbo 833 drivers/gpu/drm/radeon/radeon_vm.c radeon_sync_resv(rdev, &ib->sync, pt->tbo.base.resv, true); tbo 834 drivers/gpu/drm/radeon/radeon_vm.c r = dma_resv_reserve_shared(pt->tbo.base.resv, 1); tbo 945 drivers/gpu/drm/radeon/radeon_vm.c if (bo_va->bo && radeon_ttm_tt_is_readonly(bo_va->bo->tbo.ttm)) tbo 80 drivers/gpu/drm/virtio/virtgpu_drv.h struct ttm_buffer_object tbo; tbo 382 drivers/gpu/drm/virtio/virtgpu_drv.h ttm_bo_get(&bo->tbo); tbo 388 drivers/gpu/drm/virtio/virtgpu_drv.h struct ttm_buffer_object *tbo; tbo 392 drivers/gpu/drm/virtio/virtgpu_drv.h tbo = &((*bo)->tbo); tbo 393 drivers/gpu/drm/virtio/virtgpu_drv.h ttm_bo_put(tbo); tbo 399 drivers/gpu/drm/virtio/virtgpu_drv.h return drm_vma_node_offset_addr(&bo->tbo.base.vma_node); tbo 407 drivers/gpu/drm/virtio/virtgpu_drv.h r = ttm_bo_reserve(&bo->tbo, true, no_wait, NULL); tbo 421 drivers/gpu/drm/virtio/virtgpu_drv.h ttm_bo_unreserve(&bo->tbo); tbo 74 drivers/gpu/drm/virtio/virtgpu_ioctl.c qobj = container_of(bo, struct virtio_gpu_object, tbo); tbo 92 drivers/gpu/drm/virtio/virtgpu_ioctl.c qobj = container_of(bo, struct virtio_gpu_object, tbo); tbo 188 drivers/gpu/drm/virtio/virtgpu_ioctl.c buflist[i].bo = &qobj->tbo; tbo 384 drivers/gpu/drm/virtio/virtgpu_ioctl.c ret = ttm_bo_validate(&qobj->tbo, &qobj->placement, &ctx); tbo 399 drivers/gpu/drm/virtio/virtgpu_ioctl.c dma_resv_add_excl_fence(qobj->tbo.base.resv, tbo 434 drivers/gpu/drm/virtio/virtgpu_ioctl.c ret = ttm_bo_validate(&qobj->tbo, &qobj->placement, &ctx); tbo 453 drivers/gpu/drm/virtio/virtgpu_ioctl.c dma_resv_add_excl_fence(qobj->tbo.base.resv, tbo 66 drivers/gpu/drm/virtio/virtgpu_object.c static void virtio_gpu_ttm_bo_destroy(struct ttm_buffer_object *tbo) tbo 71 drivers/gpu/drm/virtio/virtgpu_object.c bo = container_of(tbo, struct virtio_gpu_object, tbo); tbo 139 drivers/gpu/drm/virtio/virtgpu_object.c ret = ttm_bo_init(&vgdev->mman.bdev, &bo->tbo, params->size, tbo 160 drivers/gpu/drm/virtio/virtgpu_object.c mainbuf.bo = &bo->tbo; tbo 196 drivers/gpu/drm/virtio/virtgpu_object.c r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap); tbo 207 drivers/gpu/drm/virtio/virtgpu_object.c struct page **pages = bo->tbo.ttm->pages; tbo 208 drivers/gpu/drm/virtio/virtgpu_object.c int nr_pages = bo->tbo.num_pages; tbo 219 drivers/gpu/drm/virtio/virtgpu_object.c if (bo->tbo.ttm->state == tt_unpopulated) tbo 220 drivers/gpu/drm/virtio/virtgpu_object.c bo->tbo.ttm->bdev->driver->ttm_tt_populate(bo->tbo.ttm, &ctx); tbo 252 drivers/gpu/drm/virtio/virtgpu_object.c r = ttm_bo_reserve(&bo->tbo, true, no_wait, NULL); tbo 255 drivers/gpu/drm/virtio/virtgpu_object.c r = ttm_bo_wait(&bo->tbo, true, no_wait); tbo 256 drivers/gpu/drm/virtio/virtgpu_object.c ttm_bo_unreserve(&bo->tbo); tbo 215 drivers/gpu/drm/virtio/virtgpu_plane.c dma_resv_add_excl_fence(bo->tbo.base.resv, tbo 37 drivers/gpu/drm/virtio/virtgpu_prime.c if (!bo->tbo.ttm->pages || !bo->tbo.ttm->num_pages) tbo 41 drivers/gpu/drm/virtio/virtgpu_prime.c return drm_prime_pages_to_sg(bo->tbo.ttm->pages, tbo 42 drivers/gpu/drm/virtio/virtgpu_prime.c bo->tbo.ttm->num_pages); tbo 201 drivers/gpu/drm/virtio/virtgpu_ttm.c virtio_gpu_get_vgdev(gtt->obj->tbo.bdev); tbo 212 drivers/gpu/drm/virtio/virtgpu_ttm.c virtio_gpu_get_vgdev(gtt->obj->tbo.bdev); tbo 244 drivers/gpu/drm/virtio/virtgpu_ttm.c gtt->obj = container_of(bo, struct virtio_gpu_object, tbo); tbo 252 drivers/gpu/drm/virtio/virtgpu_ttm.c static void virtio_gpu_bo_swap_notify(struct ttm_buffer_object *tbo) tbo 256 drivers/gpu/drm/virtio/virtgpu_ttm.c bo = container_of(tbo, struct virtio_gpu_object, tbo);