tbo               309 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c 	r = amdgpu_ttm_alloc_gart(&bo->tbo);
tbo               196 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo               221 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct dma_resv *resv = bo->tbo.base.resv;
tbo               280 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	if (WARN(amdgpu_ttm_tt_get_usermm(bo->tbo.ttm),
tbo               286 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               313 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev);
tbo               349 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev);
tbo               380 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	unsigned long bo_size = bo->tbo.mem.size;
tbo               452 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	entry->bo = &bo->tbo;
tbo               494 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	ret = amdgpu_ttm_tt_set_userptr(bo->tbo.ttm, user_addr, 0);
tbo               507 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	ret = amdgpu_ttm_tt_get_user_pages(bo, bo->tbo.ttm->pages);
tbo               519 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               525 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	amdgpu_ttm_tt_get_user_pages_done(bo->tbo.ttm);
tbo               582 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	ctx->kfd_bo.tv.bo = &bo->tbo;
tbo               645 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	ctx->kfd_bo.tv.bo = &bo->tbo;
tbo               815 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 					sync, pd->tbo.base.resv,
tbo               890 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	ret = dma_resv_reserve_shared(vm->root.base.bo->tbo.base.resv, 1);
tbo              1063 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev);
tbo              1183 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		bo->tbo.sg = sg;
tbo              1184 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		bo->tbo.ttm->sg = sg;
tbo              1230 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	unsigned long bo_size = mem->bo->tbo.mem.size;
tbo              1285 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	if (mem->bo->tbo.sg) {
tbo              1286 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		sg_free_table(mem->bo->tbo.sg);
tbo              1287 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		kfree(mem->bo->tbo.sg);
tbo              1291 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	drm_gem_object_put_unlocked(&mem->bo->tbo.base);
tbo              1329 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) {
tbo              1338 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	bo_size = bo->tbo.mem.size;
tbo              1354 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) &&
tbo              1355 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	    bo->tbo.mem.mem_type == TTM_PL_SYSTEM)
tbo              1376 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	    !amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) {
tbo              1414 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	if (!amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) && !bo->pin_count)
tbo              1442 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	unsigned long bo_size = mem->bo->tbo.mem.size;
tbo              1493 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	    !amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm) && !mem->bo->pin_count)
tbo              1527 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) {
tbo              1633 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	drm_gem_object_get(&bo->tbo.base);
tbo              1706 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo              1734 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		ret = amdgpu_ttm_tt_get_user_pages(bo, bo->tbo.ttm->pages);
tbo              1743 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		amdgpu_ttm_tt_get_user_pages_done(bo->tbo.ttm);
tbo              1822 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		if (bo->tbo.ttm->pages[0]) {
tbo              1824 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 			ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo              2032 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		ret = amdgpu_sync_fence(NULL, &sync_obj, bo->tbo.moving, false);
tbo              2137 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	ret = dma_resv_reserve_shared(gws_bo->tbo.base.resv, 1);
tbo               103 drivers/gpu/drm/amd/amdgpu/amdgpu_benchmark.c 	r = amdgpu_ttm_alloc_gart(&sobj->tbo);
tbo               122 drivers/gpu/drm/amd/amdgpu/amdgpu_benchmark.c 	r = amdgpu_ttm_alloc_gart(&dobj->tbo);
tbo               107 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c 		usermm = amdgpu_ttm_tt_get_usermm(bo->tbo.ttm);
tbo               121 drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c 		entry->tv.bo = &bo->tbo;
tbo                54 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 	p->uf_entry.tv.bo = &bo->tbo;
tbo                66 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 	if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) {
tbo               401 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo               405 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		.resv = bo->tbo.base.resv,
tbo               437 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 	r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               468 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo               480 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		other = amdgpu_mem_type_to_domain(bo->tbo.mem.mem_type);
tbo               496 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               541 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		usermm = amdgpu_ttm_tt_get_usermm(bo->tbo.ttm);
tbo               545 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		if (amdgpu_ttm_tt_is_userptr(bo->tbo.ttm) &&
tbo               549 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 			r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               553 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 			amdgpu_ttm_tt_set_user_pages(bo->tbo.ttm,
tbo               624 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		e->user_pages = kvmalloc_array(bo->tbo.ttm->num_pages,
tbo               639 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		for (i = 0; i < bo->tbo.ttm->num_pages; i++) {
tbo               640 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 			if (bo->tbo.ttm->pages[i] != e->user_pages[i]) {
tbo               711 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		r = amdgpu_ttm_alloc_gart(&uf->tbo);
tbo               729 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		struct dma_resv *resv = bo->tbo.base.resv;
tbo              1302 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		r |= !amdgpu_ttm_tt_get_user_pages_done(bo->tbo.ttm);
tbo              1731 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 	if (dma_resv_locking_ctx((*bo)->tbo.base.resv) != &parser->ticket)
tbo              1737 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 		r = ttm_bo_validate(&(*bo)->tbo, &(*bo)->placement, &ctx);
tbo              1742 drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 	return amdgpu_ttm_alloc_gart(&(*bo)->tbo);
tbo                77 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c 	csa_tv.bo = &bo->tbo;
tbo              3408 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c 		if (shadow->tbo.mem.mem_type != TTM_PL_TT ||
tbo              3409 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c 		    shadow->tbo.mem.start == AMDGPU_BO_INVALID_OFFSET ||
tbo              3410 drivers/gpu/drm/amd/amdgpu/amdgpu_device.c 		    shadow->parent->tbo.mem.mem_type != TTM_PL_VRAM)
tbo               202 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c 	r = amdgpu_ttm_alloc_gart(&new_abo->tbo);
tbo               208 drivers/gpu/drm/amd/amdgpu/amdgpu_display.c 	r = dma_resv_get_fences_rcu(new_abo->tbo.base.resv, &work->excl,
tbo                52 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	int npages = bo->tbo.num_pages;
tbo                54 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	return drm_prime_pages_to_sg(bo->tbo.ttm->pages, npages);
tbo                71 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	ret = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages,
tbo               108 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo               122 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) ||
tbo               198 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo               219 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 		r = __dma_resv_make_exclusive(bo->tbo.base.resv);
tbo               254 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo               286 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo               303 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 		ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               339 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) ||
tbo               388 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	bo->tbo.sg = sg;
tbo               389 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	bo->tbo.ttm->sg = sg;
tbo               396 drivers/gpu/drm/amd/amdgpu/amdgpu_dma_buf.c 	return &bo->tbo.base;
tbo               176 drivers/gpu/drm/amd/amdgpu/amdgpu_fb.c 	ret = amdgpu_ttm_alloc_gart(&abo->tbo);
tbo                88 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	*obj = &bo->tbo.base;
tbo               125 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(abo->tbo.bdev);
tbo               132 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	mm = amdgpu_ttm_tt_get_usermm(abo->tbo.ttm);
tbo               137 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	    abo->tbo.base.resv != vm->root.base.bo->tbo.base.resv)
tbo               158 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo               172 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	tv.bo = &bo->tbo;
tbo               255 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 		resv = vm->root.base.bo->tbo.base.resv;
tbo               321 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	r = amdgpu_ttm_tt_set_userptr(bo->tbo.ttm, args->addr, args->flags);
tbo               332 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 		r = amdgpu_ttm_tt_get_user_pages(bo, bo->tbo.ttm->pages);
tbo               341 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 		r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               355 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 		amdgpu_ttm_tt_get_user_pages_done(bo->tbo.ttm);
tbo               375 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	if (amdgpu_ttm_tt_get_usermm(robj->tbo.ttm) ||
tbo               438 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	ret = dma_resv_wait_timeout_rcu(robj->tbo.base.resv, true, true,
tbo               603 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 		tv.bo = &abo->tbo;
tbo               694 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 		info.bo_size = robj->tbo.base.size;
tbo               695 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 		info.alignment = robj->tbo.mem.page_alignment << PAGE_SHIFT;
tbo               709 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 		if (amdgpu_ttm_tt_get_usermm(robj->tbo.ttm)) {
tbo               715 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 			if (amdgpu_xgmi_same_hive(amdgpu_ttm_adev(robj->tbo.bdev),
tbo               716 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 				amdgpu_ttm_adev(base->vm->root.base.bo->tbo.bdev))) {
tbo               805 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	domain = amdgpu_mem_type_to_domain(bo->tbo.mem.mem_type);
tbo               825 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	dma_buf = READ_ONCE(bo->tbo.base.dma_buf);
tbo               826 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c 	attachment = READ_ONCE(bo->tbo.base.import_attach);
tbo                34 drivers/gpu/drm/amd/amdgpu/amdgpu_gem.h #define gem_to_amdgpu_bo(gobj) container_of((gobj), struct amdgpu_bo, tbo.base)
tbo                44 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo                47 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c 	switch (bo->tbo.mem.mem_type) {
tbo                49 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c 		ttm = container_of(bo->tbo.ttm, struct ttm_dma_tt, ttm);
tbo                59 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c 	*flags = amdgpu_ttm_tt_pde_flags(bo->tbo.ttm, &bo->tbo.mem);
tbo                69 drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo                35 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	struct ttm_buffer_object *tbo;
tbo               169 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 				struct ttm_buffer_object *tbo,
tbo               220 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 			      struct ttm_buffer_object *tbo,
tbo               229 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	if ((&tbo->mem == mem || tbo->mem.mem_type != TTM_PL_TT) &&
tbo               245 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 	node->tbo = tbo;
tbo               249 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 		r = amdgpu_gtt_mgr_alloc(man, tbo, place, mem);
tbo               321 drivers/gpu/drm/amd/amdgpu/amdgpu_gtt_mgr.c 		r = amdgpu_ttm_recover_gart(node->tbo);
tbo              1079 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 		amdgpu_pasid_free_delayed(pd->tbo.base.resv, pasid);
tbo               179 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		if (!amdgpu_ttm_tt_affect_userptr(bo->tbo.ttm, start, end))
tbo               182 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 		r = dma_resv_wait_timeout_rcu(bo->tbo.base.resv,
tbo               279 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 			if (amdgpu_ttm_tt_affect_userptr(bo->tbo.ttm,
tbo               378 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo               436 drivers/gpu/drm/amd/amdgpu/amdgpu_mn.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo                64 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo                66 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	if (bo->tbo.mem.mem_type == TTM_PL_VRAM) {
tbo                70 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	} else if (bo->tbo.mem.mem_type == TTM_PL_TT) {
tbo                75 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c static void amdgpu_bo_destroy(struct ttm_buffer_object *tbo)
tbo                77 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(tbo->bdev);
tbo                78 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	struct amdgpu_bo *bo = ttm_to_amdgpu_bo(tbo);
tbo                85 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	if (bo->tbo.base.import_attach)
tbo                86 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 		drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg);
tbo                87 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	drm_gem_object_release(&bo->tbo.base);
tbo               127 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(abo->tbo.bdev);
tbo               277 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	r = amdgpu_ttm_alloc_gart(&(*bo_ptr)->tbo);
tbo               385 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	ttm_bo_mem_put(&(*bo_ptr)->tbo, &(*bo_ptr)->tbo.mem);
tbo               386 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	r = ttm_bo_mem_space(&(*bo_ptr)->tbo, &(*bo_ptr)->placement,
tbo               387 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 			     &(*bo_ptr)->tbo.mem, &ctx);
tbo               551 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	drm_gem_private_object_init(adev->ddev, &bo->tbo.base, size);
tbo               566 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	bo->tbo.bdev = &adev->mman.bdev;
tbo               573 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 		bo->tbo.priority = 1;
tbo               575 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	r = ttm_bo_init_reserved(&adev->mman.bdev, &bo->tbo, size, bp->type,
tbo               582 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	    bo->tbo.mem.mem_type == TTM_PL_VRAM &&
tbo               583 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	    bo->tbo.mem.start < adev->gmc.visible_vram_size >> PAGE_SHIFT)
tbo               590 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	    bo->tbo.mem.placement & TTM_PL_FLAG_VRAM) {
tbo               593 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 		r = amdgpu_fill_buffer(bo, 0, bo->tbo.base.resv, &fence);
tbo               598 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 		dma_fence_put(bo->tbo.moving);
tbo               599 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 		bo->tbo.moving = dma_fence_get(fence);
tbo               616 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 		dma_resv_unlock(bo->tbo.base.resv);
tbo               637 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	bp.resv = bo->tbo.base.resv;
tbo               678 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 			WARN_ON(dma_resv_lock((*bo_ptr)->tbo.base.resv,
tbo               684 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 			dma_resv_unlock((*bo_ptr)->tbo.base.resv);
tbo               718 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               743 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(shadow->tbo.bdev);
tbo               781 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	r = dma_resv_wait_timeout_rcu(bo->tbo.base.resv, false, false,
tbo               786 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap);
tbo               838 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	ttm_bo_get(&bo->tbo);
tbo               850 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	struct ttm_buffer_object *tbo;
tbo               855 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	tbo = &((*bo)->tbo);
tbo               856 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	ttm_bo_put(tbo);
tbo               885 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo               889 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm))
tbo               909 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 		uint32_t mem_type = bo->tbo.mem.mem_type;
tbo               917 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 			u64 domain_start = bo->tbo.bdev->man[mem_type].gpu_offset;
tbo               944 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               952 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	domain = amdgpu_mem_type_to_domain(bo->tbo.mem.mem_type);
tbo               994 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo              1012 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo              1123 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	return ttm_fbdev_mmap(vma, &bo->tbo);
tbo              1139 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo              1159 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	dma_resv_assert_held(bo->tbo.base.resv);
tbo              1391 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	struct dma_resv *resv = bo->tbo.base.resv;
tbo              1411 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo              1416 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	amdgpu_sync_resv(adev, &sync, bo->tbo.base.resv, owner, false);
tbo              1435 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	WARN_ON_ONCE(bo->tbo.mem.mem_type == TTM_PL_SYSTEM);
tbo              1436 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	WARN_ON_ONCE(!dma_resv_is_locked(bo->tbo.base.resv) &&
tbo              1437 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 		     !bo->pin_count && bo->tbo.type != ttm_bo_type_kernel);
tbo              1438 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	WARN_ON_ONCE(bo->tbo.mem.start == AMDGPU_BO_INVALID_OFFSET);
tbo              1439 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	WARN_ON_ONCE(bo->tbo.mem.mem_type == TTM_PL_VRAM &&
tbo              1442 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	return amdgpu_gmc_sign_extend(bo->tbo.offset);
tbo                85 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	struct ttm_buffer_object	tbo;
tbo               111 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h static inline struct amdgpu_bo *ttm_to_amdgpu_bo(struct ttm_buffer_object *tbo)
tbo               113 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	return container_of(tbo, struct amdgpu_bo, tbo);
tbo               154 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo               157 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	r = __ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL);
tbo               168 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	ttm_bo_unreserve(&bo->tbo);
tbo               173 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	return bo->tbo.num_pages << PAGE_SHIFT;
tbo               178 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	return (bo->tbo.num_pages << PAGE_SHIFT) / AMDGPU_GPU_PAGE_SIZE;
tbo               183 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	return (bo->tbo.mem.page_alignment << PAGE_SHIFT) / AMDGPU_GPU_PAGE_SIZE;
tbo               194 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	return drm_vma_node_offset_addr(&bo->tbo.base.vma_node);
tbo               202 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo               204 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	struct drm_mm_node *node = bo->tbo.mem.mm_node;
tbo               207 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	if (bo->tbo.mem.mem_type != TTM_PL_VRAM)
tbo               210 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 	for (pages_left = bo->tbo.mem.num_pages; pages_left;
tbo               106 drivers/gpu/drm/amd/amdgpu/amdgpu_test.c 		r = amdgpu_ttm_alloc_gart(&gtt_obj[i]->tbo);
tbo               130 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 			   __entry->pages = bo->tbo.num_pages;
tbo               131 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 			   __entry->type = bo->tbo.mem.mem_type;
tbo               231 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	return drm_vma_node_verify_access(&abo->tbo.base.vma_node,
tbo               790 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	struct ttm_tt *ttm = bo->tbo.ttm;
tbo              1003 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 				struct ttm_buffer_object *tbo,
tbo              1006 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	struct amdgpu_bo *abo = ttm_to_amdgpu_bo(tbo);
tbo              1007 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	struct ttm_tt *ttm = tbo->ttm;
tbo              1156 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c int amdgpu_ttm_recover_gart(struct ttm_buffer_object *tbo)
tbo              1158 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(tbo->bdev);
tbo              1162 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	if (!tbo->ttm)
tbo              1165 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	flags = amdgpu_ttm_tt_pte_flags(adev, tbo->ttm, &tbo->mem);
tbo              1166 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	r = amdgpu_ttm_gart_bind(adev, tbo, flags);
tbo              1544 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(abo->tbo.bdev);
tbo              1554 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	nodes = amdgpu_find_mm_node(&abo->tbo.mem, &offset);
tbo              2042 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo              2058 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	if (bo->tbo.mem.mem_type == TTM_PL_TT) {
tbo              2059 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 		r = amdgpu_ttm_alloc_gart(&bo->tbo);
tbo              2064 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	num_pages = bo->tbo.num_pages;
tbo              2065 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	mm_node = bo->tbo.mem.mm_node;
tbo              2092 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	num_pages = bo->tbo.num_pages;
tbo              2093 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	mm_node = bo->tbo.mem.mm_node;
tbo              2099 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 		dst_addr = amdgpu_mm_node_addr(&bo->tbo, mm_node, &bo->tbo.mem);
tbo               104 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h int amdgpu_ttm_recover_gart(struct ttm_buffer_object *tbo);
tbo               500 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 		r = ttm_bo_validate(&bo->tbo, &bo->placement, &tctx);
tbo              1041 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 		r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo              1076 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 		r = dma_resv_wait_timeout_rcu(bo->tbo.base.resv,
tbo              1088 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 		r = amdgpu_sync_resv(adev, &job->sync, bo->tbo.base.resv,
tbo               610 drivers/gpu/drm/amd/amdgpu/amdgpu_vce.c 	return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               206 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	if (bo->tbo.type == ttm_bo_type_kernel)
tbo               305 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	if (bo->tbo.base.resv != vm->root.base.bo->tbo.base.resv)
tbo               309 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	if (bo->tbo.type == ttm_bo_type_kernel && bo->parent)
tbo               315 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	    amdgpu_mem_type_to_domain(bo->tbo.mem.mem_type))
tbo               562 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	entry->tv.bo = &vm->root.base.bo->tbo;
tbo               586 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		if (abo->tbo.base.resv == vm->root.base.bo->tbo.base.resv)
tbo               622 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		ttm_bo_move_to_lru_tail(&bo->tbo, &vm->lru_bulk_move);
tbo               624 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 			ttm_bo_move_to_lru_tail(&bo->shadow->tbo,
tbo               661 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		if (bo->tbo.type != ttm_bo_type_kernel) {
tbo               745 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               750 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		r = ttm_bo_validate(&bo->shadow->tbo, &bo->shadow->placement,
tbo               835 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		bp->resv = vm->root.base.bo->tbo.base.resv;
tbo              1703 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		mem = &bo->tbo.mem;
tbo              1706 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 			ttm = container_of(bo->tbo.ttm, struct ttm_dma_tt, ttm);
tbo              1709 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		exclusive = dma_resv_get_excl(bo->tbo.base.resv);
tbo              1713 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		flags = amdgpu_ttm_tt_pte_flags(adev, bo->tbo.ttm, mem);
tbo              1714 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		bo_adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo              1719 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	if (clear || (bo && bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv))
tbo              1750 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	if (bo && bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv) {
tbo              1751 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		uint32_t mem_type = bo->tbo.mem.mem_type;
tbo              1886 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	struct dma_resv *resv = vm->root.base.bo->tbo.base.resv;
tbo              2000 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		resv = bo_va->base.bo->tbo.base.resv;
tbo              2054 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	if (bo && amdgpu_xgmi_same_hive(adev, amdgpu_ttm_adev(bo->tbo.bdev)) &&
tbo              2091 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	if (bo && bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv &&
tbo              2423 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 			if (dma_resv_locking_ctx(bo->tbo.base.resv) !=
tbo              2451 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		if (bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv)
tbo              2515 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		if (evicted && bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv) {
tbo              2524 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		if (bo->tbo.type == ttm_bo_type_kernel)
tbo              2526 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		else if (bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv)
tbo              2656 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	return dma_resv_wait_timeout_rcu(vm->root.base.bo->tbo.base.resv,
tbo              2731 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 	r = dma_resv_reserve_shared(root->tbo.base.resv, 1);
tbo                40 drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c 	r = amdgpu_ttm_alloc_gart(&table->tbo);
tbo                45 drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c 		r = amdgpu_ttm_alloc_gart(&table->shadow->tbo);
tbo                75 drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c 	r = amdgpu_sync_resv(p->adev, &p->job->sync, root->tbo.base.resv,
tbo               217 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
tbo               218 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c 	struct ttm_mem_reg *mem = &bo->tbo.mem;
tbo               268 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c 			       struct ttm_buffer_object *tbo,
tbo              1078 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	memset(hpd, 0, adev->gfx.mec.hpd_eop_obj->tbo.mem.size);
tbo              1745 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 	memset(hpd, 0, adev->gfx.mec.hpd_eop_obj->tbo.mem.size);
tbo              4513 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	adev = amdgpu_ttm_adev(rbo->tbo.bdev);
tbo              4516 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	tv.bo = &rbo->tbo;
tbo              4539 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	r = amdgpu_ttm_alloc_gart(&rbo->tbo);
tbo              5766 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 		r = dma_resv_wait_timeout_rcu(abo->tbo.base.resv, true,
tbo               380 drivers/gpu/drm/qxl/qxl_cmd.c 	drm_gem_object_put_unlocked(&qdev->primary_bo->tbo.base);
tbo               407 drivers/gpu/drm/qxl/qxl_cmd.c 	drm_gem_object_get(&qdev->primary_bo->tbo.base);
tbo               598 drivers/gpu/drm/qxl/qxl_cmd.c 	ret = ttm_bo_wait(&surf->tbo, true, !stall);
tbo                64 drivers/gpu/drm/qxl/qxl_debugfs.c 		fobj = rcu_dereference(bo->tbo.base.resv->fence);
tbo                69 drivers/gpu/drm/qxl/qxl_debugfs.c 			   (unsigned long)bo->tbo.base.size,
tbo               800 drivers/gpu/drm/qxl/qxl_display.c 					(&qdev->dumb_shadow_bo->tbo.base);
tbo               810 drivers/gpu/drm/qxl/qxl_display.c 					(&user_bo->shadow->tbo.base);
tbo               813 drivers/gpu/drm/qxl/qxl_display.c 			drm_gem_object_get(&qdev->dumb_shadow_bo->tbo.base);
tbo               844 drivers/gpu/drm/qxl/qxl_display.c 		drm_gem_object_put_unlocked(&user_bo->shadow->tbo.base);
tbo                74 drivers/gpu/drm/qxl/qxl_drv.h 	struct ttm_buffer_object	tbo;
tbo                96 drivers/gpu/drm/qxl/qxl_drv.h #define gem_to_qxl_bo(gobj) container_of((gobj), struct qxl_bo, tbo.base)
tbo                97 drivers/gpu/drm/qxl/qxl_drv.h #define to_qxl_bo(tobj) container_of((tobj), struct qxl_bo, tbo)
tbo               310 drivers/gpu/drm/qxl/qxl_drv.h 		(bo->tbo.mem.mem_type == TTM_PL_VRAM)
tbo               313 drivers/gpu/drm/qxl/qxl_drv.h 	WARN_ON_ONCE((bo->tbo.offset & slot->gpu_offset) != slot->gpu_offset);
tbo               316 drivers/gpu/drm/qxl/qxl_drv.h 	return slot->high_bits | (bo->tbo.offset - slot->gpu_offset + offset);
tbo                35 drivers/gpu/drm/qxl/qxl_gem.c 	struct ttm_buffer_object *tbo;
tbo                41 drivers/gpu/drm/qxl/qxl_gem.c 	tbo = &qobj->tbo;
tbo                42 drivers/gpu/drm/qxl/qxl_gem.c 	ttm_bo_put(tbo);
tbo                66 drivers/gpu/drm/qxl/qxl_gem.c 	*obj = &qbo->tbo.base;
tbo               331 drivers/gpu/drm/qxl/qxl_ioctl.c 		ret = ttm_bo_validate(&qobj->tbo, &qobj->placement, &ctx);
tbo                30 drivers/gpu/drm/qxl/qxl_object.c static void qxl_ttm_bo_destroy(struct ttm_buffer_object *tbo)
tbo                35 drivers/gpu/drm/qxl/qxl_object.c 	bo = to_qxl_bo(tbo);
tbo                36 drivers/gpu/drm/qxl/qxl_object.c 	qdev = (struct qxl_device *)bo->tbo.base.dev->dev_private;
tbo                43 drivers/gpu/drm/qxl/qxl_object.c 	drm_gem_object_release(&bo->tbo.base);
tbo                98 drivers/gpu/drm/qxl/qxl_object.c 	r = drm_gem_object_init(&qdev->ddev, &bo->tbo.base, size);
tbo               113 drivers/gpu/drm/qxl/qxl_object.c 	r = ttm_bo_init(&qdev->mman.bdev, &bo->tbo, size, type,
tbo               138 drivers/gpu/drm/qxl/qxl_object.c 	r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap);
tbo               151 drivers/gpu/drm/qxl/qxl_object.c 	struct ttm_mem_type_manager *man = &bo->tbo.bdev->man[bo->tbo.mem.mem_type];
tbo               156 drivers/gpu/drm/qxl/qxl_object.c 	if (bo->tbo.mem.mem_type == TTM_PL_VRAM)
tbo               158 drivers/gpu/drm/qxl/qxl_object.c 	else if (bo->tbo.mem.mem_type == TTM_PL_PRIV)
tbo               164 drivers/gpu/drm/qxl/qxl_object.c 	ret = ttm_mem_io_reserve(bo->tbo.bdev, &bo->tbo.mem);
tbo               167 drivers/gpu/drm/qxl/qxl_object.c 	return io_mapping_map_atomic_wc(map, bo->tbo.mem.bus.offset + page_offset);
tbo               196 drivers/gpu/drm/qxl/qxl_object.c 	struct ttm_mem_type_manager *man = &bo->tbo.bdev->man[bo->tbo.mem.mem_type];
tbo               198 drivers/gpu/drm/qxl/qxl_object.c 	if ((bo->tbo.mem.mem_type != TTM_PL_VRAM) &&
tbo               199 drivers/gpu/drm/qxl/qxl_object.c 	    (bo->tbo.mem.mem_type != TTM_PL_PRIV))
tbo               205 drivers/gpu/drm/qxl/qxl_object.c 	ttm_mem_io_free(bo->tbo.bdev, &bo->tbo.mem);
tbo               217 drivers/gpu/drm/qxl/qxl_object.c 	drm_gem_object_put_unlocked(&(*bo)->tbo.base);
tbo               223 drivers/gpu/drm/qxl/qxl_object.c 	drm_gem_object_get(&bo->tbo.base);
tbo               230 drivers/gpu/drm/qxl/qxl_object.c 	struct drm_device *ddev = bo->tbo.base.dev;
tbo               238 drivers/gpu/drm/qxl/qxl_object.c 	r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               250 drivers/gpu/drm/qxl/qxl_object.c 	struct drm_device *ddev = bo->tbo.base.dev;
tbo               262 drivers/gpu/drm/qxl/qxl_object.c 	r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               313 drivers/gpu/drm/qxl/qxl_object.c 			&bo->tbo.base, bo, (unsigned long)bo->tbo.base.size,
tbo               314 drivers/gpu/drm/qxl/qxl_object.c 			*((unsigned long *)&bo->tbo.base.refcount));
tbo               319 drivers/gpu/drm/qxl/qxl_object.c 		drm_gem_object_put_unlocked(&bo->tbo.base);
tbo                34 drivers/gpu/drm/qxl/qxl_object.h 	r = ttm_bo_reserve(&bo->tbo, true, no_wait, NULL);
tbo                37 drivers/gpu/drm/qxl/qxl_object.h 			struct drm_device *ddev = bo->tbo.base.dev;
tbo                48 drivers/gpu/drm/qxl/qxl_object.h 	ttm_bo_unreserve(&bo->tbo);
tbo                53 drivers/gpu/drm/qxl/qxl_object.h 	return bo->tbo.offset;
tbo                58 drivers/gpu/drm/qxl/qxl_object.h 	return bo->tbo.num_pages << PAGE_SHIFT;
tbo                63 drivers/gpu/drm/qxl/qxl_object.h 	return drm_vma_node_offset_addr(&bo->tbo.base.vma_node);
tbo                71 drivers/gpu/drm/qxl/qxl_object.h 	r = ttm_bo_reserve(&bo->tbo, true, no_wait, NULL);
tbo                74 drivers/gpu/drm/qxl/qxl_object.h 			struct drm_device *ddev = bo->tbo.base.dev;
tbo                82 drivers/gpu/drm/qxl/qxl_object.h 		*mem_type = bo->tbo.mem.mem_type;
tbo                84 drivers/gpu/drm/qxl/qxl_object.h 	r = ttm_bo_wait(&bo->tbo, true, no_wait);
tbo                85 drivers/gpu/drm/qxl/qxl_object.h 	ttm_bo_unreserve(&bo->tbo);
tbo               214 drivers/gpu/drm/qxl/qxl_release.c 		if (entry->tv.bo == &bo->tbo)
tbo               223 drivers/gpu/drm/qxl/qxl_release.c 	entry->tv.bo = &bo->tbo;
tbo               236 drivers/gpu/drm/qxl/qxl_release.c 		ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               241 drivers/gpu/drm/qxl/qxl_release.c 	ret = dma_resv_reserve_shared(bo->tbo.base.resv, 1);
tbo               246 drivers/gpu/drm/qxl/qxl_release.c 	ret = qxl_bo_check_id(bo->tbo.base.dev->dev_private, bo);
tbo               158 drivers/gpu/drm/qxl/qxl_ttm.c 	return drm_vma_node_verify_access(&qbo->tbo.base.vma_node,
tbo               300 drivers/gpu/drm/qxl/qxl_ttm.c 	qdev = qbo->tbo.base.dev->dev_private;
tbo               493 drivers/gpu/drm/radeon/radeon.h 	struct ttm_buffer_object	tbo;
tbo               515 drivers/gpu/drm/radeon/radeon.h #define gem_to_radeon_bo(gobj) container_of((gobj), struct radeon_bo, tbo.base)
tbo               125 drivers/gpu/drm/radeon/radeon_benchmark.c 						dobj->tbo.base.resv);
tbo               136 drivers/gpu/drm/radeon/radeon_benchmark.c 						dobj->tbo.base.resv);
tbo               163 drivers/gpu/drm/radeon/radeon_cs.c 		if (radeon_ttm_tt_has_userptr(p->relocs[i].robj->tbo.ttm)) {
tbo               186 drivers/gpu/drm/radeon/radeon_cs.c 		p->relocs[i].tv.bo = &p->relocs[i].robj->tbo;
tbo               260 drivers/gpu/drm/radeon/radeon_cs.c 		resv = reloc->robj->tbo.base.resv;
tbo               404 drivers/gpu/drm/radeon/radeon_cs.c 	return (int)la->robj->tbo.num_pages - (int)lb->robj->tbo.num_pages;
tbo               446 drivers/gpu/drm/radeon/radeon_cs.c 			drm_gem_object_put_unlocked(&bo->tbo.base);
tbo               518 drivers/gpu/drm/radeon/radeon_cs.c 				&rdev->ring_tmp_bo.bo->tbo.mem);
tbo               532 drivers/gpu/drm/radeon/radeon_cs.c 		r = radeon_vm_bo_update(rdev, bo_va, &bo->tbo.mem);
tbo               280 drivers/gpu/drm/radeon/radeon_display.c 	drm_gem_object_put_unlocked(&work->old_rbo->tbo.base);
tbo               538 drivers/gpu/drm/radeon/radeon_display.c 	work->fence = dma_fence_get(dma_resv_get_excl(new_rbo->tbo.base.resv));
tbo               612 drivers/gpu/drm/radeon/radeon_display.c 	drm_gem_object_put_unlocked(&work->old_rbo->tbo.base);
tbo                86 drivers/gpu/drm/radeon/radeon_gem.c 	*obj = &robj->tbo.base;
tbo               117 drivers/gpu/drm/radeon/radeon_gem.c 		r = dma_resv_wait_timeout_rcu(robj->tbo.base.resv, true, true, 30 * HZ);
tbo               333 drivers/gpu/drm/radeon/radeon_gem.c 	r = radeon_ttm_tt_set_userptr(bo->tbo.ttm, args->addr, args->flags);
tbo               352 drivers/gpu/drm/radeon/radeon_gem.c 		r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               422 drivers/gpu/drm/radeon/radeon_gem.c 	if (radeon_ttm_tt_has_userptr(robj->tbo.ttm)) {
tbo               454 drivers/gpu/drm/radeon/radeon_gem.c 	r = dma_resv_test_signaled_rcu(robj->tbo.base.resv, true);
tbo               460 drivers/gpu/drm/radeon/radeon_gem.c 	cur_placement = READ_ONCE(robj->tbo.mem.mem_type);
tbo               483 drivers/gpu/drm/radeon/radeon_gem.c 	ret = dma_resv_wait_timeout_rcu(robj->tbo.base.resv, true, true, 30 * HZ);
tbo               490 drivers/gpu/drm/radeon/radeon_gem.c 	cur_placement = READ_ONCE(robj->tbo.mem.mem_type);
tbo               561 drivers/gpu/drm/radeon/radeon_gem.c 	tv.bo = &bo_va->bo->tbo;
tbo               587 drivers/gpu/drm/radeon/radeon_gem.c 		r = radeon_vm_bo_update(rdev, bo_va, &bo_va->bo->tbo.mem);
tbo               723 drivers/gpu/drm/radeon/radeon_gem.c 	if (radeon_ttm_tt_has_userptr(robj->tbo.ttm))
tbo               793 drivers/gpu/drm/radeon/radeon_gem.c 		domain = radeon_mem_type_to_domain(rbo->tbo.mem.mem_type);
tbo                99 drivers/gpu/drm/radeon/radeon_mn.c 			if (!bo->tbo.ttm || bo->tbo.ttm->state != tt_bound)
tbo               108 drivers/gpu/drm/radeon/radeon_mn.c 			r = dma_resv_wait_timeout_rcu(bo->tbo.base.resv,
tbo               114 drivers/gpu/drm/radeon/radeon_mn.c 			r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo                57 drivers/gpu/drm/radeon/radeon_object.c 	u64 size = (u64)bo->tbo.num_pages << PAGE_SHIFT;
tbo                75 drivers/gpu/drm/radeon/radeon_object.c static void radeon_ttm_bo_destroy(struct ttm_buffer_object *tbo)
tbo                79 drivers/gpu/drm/radeon/radeon_object.c 	bo = container_of(tbo, struct radeon_bo, tbo);
tbo                81 drivers/gpu/drm/radeon/radeon_object.c 	radeon_update_memory_usage(bo, bo->tbo.mem.mem_type, -1);
tbo                88 drivers/gpu/drm/radeon/radeon_object.c 	if (bo->tbo.base.import_attach)
tbo                89 drivers/gpu/drm/radeon/radeon_object.c 		drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg);
tbo                90 drivers/gpu/drm/radeon/radeon_object.c 	drm_gem_object_release(&bo->tbo.base);
tbo               212 drivers/gpu/drm/radeon/radeon_object.c 	drm_gem_private_object_init(rdev->ddev, &bo->tbo.base, size);
tbo               262 drivers/gpu/drm/radeon/radeon_object.c 	r = ttm_bo_init(&rdev->mman.bdev, &bo->tbo, size, type,
tbo               287 drivers/gpu/drm/radeon/radeon_object.c 	r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap);
tbo               313 drivers/gpu/drm/radeon/radeon_object.c 	ttm_bo_get(&bo->tbo);
tbo               319 drivers/gpu/drm/radeon/radeon_object.c 	struct ttm_buffer_object *tbo;
tbo               323 drivers/gpu/drm/radeon/radeon_object.c 	tbo = &((*bo)->tbo);
tbo               324 drivers/gpu/drm/radeon/radeon_object.c 	ttm_bo_put(tbo);
tbo               334 drivers/gpu/drm/radeon/radeon_object.c 	if (radeon_ttm_tt_has_userptr(bo->tbo.ttm))
tbo               374 drivers/gpu/drm/radeon/radeon_object.c 	r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               410 drivers/gpu/drm/radeon/radeon_object.c 	r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               412 drivers/gpu/drm/radeon/radeon_object.c 		if (bo->tbo.mem.mem_type == TTM_PL_VRAM)
tbo               445 drivers/gpu/drm/radeon/radeon_object.c 			&bo->tbo.base, bo, (unsigned long)bo->tbo.base.size,
tbo               446 drivers/gpu/drm/radeon/radeon_object.c 			*((unsigned long *)&bo->tbo.base.refcount));
tbo               451 drivers/gpu/drm/radeon/radeon_object.c 		drm_gem_object_put_unlocked(&bo->tbo.base);
tbo               556 drivers/gpu/drm/radeon/radeon_object.c 				radeon_mem_type_to_domain(bo->tbo.mem.mem_type);
tbo               579 drivers/gpu/drm/radeon/radeon_object.c 			r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               613 drivers/gpu/drm/radeon/radeon_object.c 	dma_resv_assert_held(bo->tbo.base.resv);
tbo               645 drivers/gpu/drm/radeon/radeon_object.c 		ttm_bo_unmap_virtual(&old_object->tbo);
tbo               655 drivers/gpu/drm/radeon/radeon_object.c 			       bo->tbo.mem.start << PAGE_SHIFT,
tbo               656 drivers/gpu/drm/radeon/radeon_object.c 			       bo->tbo.num_pages << PAGE_SHIFT);
tbo               739 drivers/gpu/drm/radeon/radeon_object.c 	dma_resv_assert_held(bo->tbo.base.resv);
tbo               751 drivers/gpu/drm/radeon/radeon_object.c 		dma_resv_assert_held(bo->tbo.base.resv);
tbo               761 drivers/gpu/drm/radeon/radeon_object.c 	if (bo->tbo.mem.mem_type != TTM_PL_VRAM) {
tbo               785 drivers/gpu/drm/radeon/radeon_object.c 	rbo = container_of(bo, struct radeon_bo, tbo);
tbo               807 drivers/gpu/drm/radeon/radeon_object.c 	rbo = container_of(bo, struct radeon_bo, tbo);
tbo               851 drivers/gpu/drm/radeon/radeon_object.c 	r = ttm_bo_reserve(&bo->tbo, true, no_wait, NULL);
tbo               855 drivers/gpu/drm/radeon/radeon_object.c 		*mem_type = bo->tbo.mem.mem_type;
tbo               857 drivers/gpu/drm/radeon/radeon_object.c 	r = ttm_bo_wait(&bo->tbo, true, no_wait);
tbo               858 drivers/gpu/drm/radeon/radeon_object.c 	ttm_bo_unreserve(&bo->tbo);
tbo               873 drivers/gpu/drm/radeon/radeon_object.c 	struct dma_resv *resv = bo->tbo.base.resv;
tbo                68 drivers/gpu/drm/radeon/radeon_object.h 	r = ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL);
tbo                79 drivers/gpu/drm/radeon/radeon_object.h 	ttm_bo_unreserve(&bo->tbo);
tbo                93 drivers/gpu/drm/radeon/radeon_object.h 	return bo->tbo.offset;
tbo                98 drivers/gpu/drm/radeon/radeon_object.h 	return bo->tbo.num_pages << PAGE_SHIFT;
tbo               103 drivers/gpu/drm/radeon/radeon_object.h 	return (bo->tbo.num_pages << PAGE_SHIFT) / RADEON_GPU_PAGE_SIZE;
tbo               108 drivers/gpu/drm/radeon/radeon_object.h 	return (bo->tbo.mem.page_alignment << PAGE_SHIFT) / RADEON_GPU_PAGE_SIZE;
tbo               119 drivers/gpu/drm/radeon/radeon_object.h 	return drm_vma_node_offset_addr(&bo->tbo.base.vma_node);
tbo               157 drivers/gpu/drm/radeon/radeon_pm.c 		if (bo->tbo.mem.mem_type == TTM_PL_VRAM)
tbo               158 drivers/gpu/drm/radeon/radeon_pm.c 			ttm_bo_unmap_virtual(&bo->tbo);
tbo                37 drivers/gpu/drm/radeon/radeon_prime.c 	int npages = bo->tbo.num_pages;
tbo                39 drivers/gpu/drm/radeon/radeon_prime.c 	return drm_prime_pages_to_sg(bo->tbo.ttm->pages, npages);
tbo                47 drivers/gpu/drm/radeon/radeon_prime.c 	ret = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages,
tbo                83 drivers/gpu/drm/radeon/radeon_prime.c 	return &bo->tbo.base;
tbo               124 drivers/gpu/drm/radeon/radeon_prime.c 	if (radeon_ttm_tt_has_userptr(bo->tbo.ttm))
tbo               123 drivers/gpu/drm/radeon/radeon_test.c 						vram_obj->tbo.base.resv);
tbo               127 drivers/gpu/drm/radeon/radeon_test.c 						 vram_obj->tbo.base.resv);
tbo               174 drivers/gpu/drm/radeon/radeon_test.c 						vram_obj->tbo.base.resv);
tbo               178 drivers/gpu/drm/radeon/radeon_test.c 						 vram_obj->tbo.base.resv);
tbo                25 drivers/gpu/drm/radeon/radeon_trace.h 			   __entry->pages = bo->tbo.num_pages;
tbo               143 drivers/gpu/drm/radeon/radeon_ttm.c 	rbo = container_of(bo, struct radeon_bo, tbo);
tbo               183 drivers/gpu/drm/radeon/radeon_ttm.c 	struct radeon_bo *rbo = container_of(bo, struct radeon_bo, tbo);
tbo               187 drivers/gpu/drm/radeon/radeon_ttm.c 	return drm_vma_node_verify_access(&rbo->tbo.base.vma_node,
tbo               353 drivers/gpu/drm/radeon/radeon_ttm.c 	rbo = container_of(bo, struct radeon_bo, tbo);
tbo               480 drivers/gpu/drm/radeon/radeon_uvd.c 	f = dma_resv_get_excl(bo->tbo.base.resv);
tbo               144 drivers/gpu/drm/radeon/radeon_vm.c 	list[0].tv.bo = &vm->page_directory->tbo;
tbo               156 drivers/gpu/drm/radeon/radeon_vm.c 		list[idx].tv.bo = &list[idx].robj->tbo;
tbo               400 drivers/gpu/drm/radeon/radeon_vm.c 	r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx);
tbo               705 drivers/gpu/drm/radeon/radeon_vm.c 		radeon_sync_resv(rdev, &ib.sync, pd->tbo.base.resv, true);
tbo               833 drivers/gpu/drm/radeon/radeon_vm.c 		radeon_sync_resv(rdev, &ib->sync, pt->tbo.base.resv, true);
tbo               834 drivers/gpu/drm/radeon/radeon_vm.c 		r = dma_resv_reserve_shared(pt->tbo.base.resv, 1);
tbo               945 drivers/gpu/drm/radeon/radeon_vm.c 	if (bo_va->bo && radeon_ttm_tt_is_readonly(bo_va->bo->tbo.ttm))
tbo                80 drivers/gpu/drm/virtio/virtgpu_drv.h 	struct ttm_buffer_object	tbo;
tbo               382 drivers/gpu/drm/virtio/virtgpu_drv.h 	ttm_bo_get(&bo->tbo);
tbo               388 drivers/gpu/drm/virtio/virtgpu_drv.h 	struct ttm_buffer_object *tbo;
tbo               392 drivers/gpu/drm/virtio/virtgpu_drv.h 	tbo = &((*bo)->tbo);
tbo               393 drivers/gpu/drm/virtio/virtgpu_drv.h 	ttm_bo_put(tbo);
tbo               399 drivers/gpu/drm/virtio/virtgpu_drv.h 	return drm_vma_node_offset_addr(&bo->tbo.base.vma_node);
tbo               407 drivers/gpu/drm/virtio/virtgpu_drv.h 	r = ttm_bo_reserve(&bo->tbo, true, no_wait, NULL);
tbo               421 drivers/gpu/drm/virtio/virtgpu_drv.h 	ttm_bo_unreserve(&bo->tbo);
tbo                74 drivers/gpu/drm/virtio/virtgpu_ioctl.c 		qobj = container_of(bo, struct virtio_gpu_object, tbo);
tbo                92 drivers/gpu/drm/virtio/virtgpu_ioctl.c 		qobj = container_of(bo, struct virtio_gpu_object, tbo);
tbo               188 drivers/gpu/drm/virtio/virtgpu_ioctl.c 			buflist[i].bo = &qobj->tbo;
tbo               384 drivers/gpu/drm/virtio/virtgpu_ioctl.c 	ret = ttm_bo_validate(&qobj->tbo, &qobj->placement, &ctx);
tbo               399 drivers/gpu/drm/virtio/virtgpu_ioctl.c 	dma_resv_add_excl_fence(qobj->tbo.base.resv,
tbo               434 drivers/gpu/drm/virtio/virtgpu_ioctl.c 	ret = ttm_bo_validate(&qobj->tbo, &qobj->placement, &ctx);
tbo               453 drivers/gpu/drm/virtio/virtgpu_ioctl.c 		dma_resv_add_excl_fence(qobj->tbo.base.resv,
tbo                66 drivers/gpu/drm/virtio/virtgpu_object.c static void virtio_gpu_ttm_bo_destroy(struct ttm_buffer_object *tbo)
tbo                71 drivers/gpu/drm/virtio/virtgpu_object.c 	bo = container_of(tbo, struct virtio_gpu_object, tbo);
tbo               139 drivers/gpu/drm/virtio/virtgpu_object.c 	ret = ttm_bo_init(&vgdev->mman.bdev, &bo->tbo, params->size,
tbo               160 drivers/gpu/drm/virtio/virtgpu_object.c 		mainbuf.bo = &bo->tbo;
tbo               196 drivers/gpu/drm/virtio/virtgpu_object.c 	r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap);
tbo               207 drivers/gpu/drm/virtio/virtgpu_object.c 	struct page **pages = bo->tbo.ttm->pages;
tbo               208 drivers/gpu/drm/virtio/virtgpu_object.c 	int nr_pages = bo->tbo.num_pages;
tbo               219 drivers/gpu/drm/virtio/virtgpu_object.c 	if (bo->tbo.ttm->state == tt_unpopulated)
tbo               220 drivers/gpu/drm/virtio/virtgpu_object.c 		bo->tbo.ttm->bdev->driver->ttm_tt_populate(bo->tbo.ttm, &ctx);
tbo               252 drivers/gpu/drm/virtio/virtgpu_object.c 	r = ttm_bo_reserve(&bo->tbo, true, no_wait, NULL);
tbo               255 drivers/gpu/drm/virtio/virtgpu_object.c 	r = ttm_bo_wait(&bo->tbo, true, no_wait);
tbo               256 drivers/gpu/drm/virtio/virtgpu_object.c 	ttm_bo_unreserve(&bo->tbo);
tbo               215 drivers/gpu/drm/virtio/virtgpu_plane.c 			dma_resv_add_excl_fence(bo->tbo.base.resv,
tbo                37 drivers/gpu/drm/virtio/virtgpu_prime.c 	if (!bo->tbo.ttm->pages || !bo->tbo.ttm->num_pages)
tbo                41 drivers/gpu/drm/virtio/virtgpu_prime.c 	return drm_prime_pages_to_sg(bo->tbo.ttm->pages,
tbo                42 drivers/gpu/drm/virtio/virtgpu_prime.c 				     bo->tbo.ttm->num_pages);
tbo               201 drivers/gpu/drm/virtio/virtgpu_ttm.c 		virtio_gpu_get_vgdev(gtt->obj->tbo.bdev);
tbo               212 drivers/gpu/drm/virtio/virtgpu_ttm.c 		virtio_gpu_get_vgdev(gtt->obj->tbo.bdev);
tbo               244 drivers/gpu/drm/virtio/virtgpu_ttm.c 	gtt->obj = container_of(bo, struct virtio_gpu_object, tbo);
tbo               252 drivers/gpu/drm/virtio/virtgpu_ttm.c static void virtio_gpu_bo_swap_notify(struct ttm_buffer_object *tbo)
tbo               256 drivers/gpu/drm/virtio/virtgpu_ttm.c 	bo = container_of(tbo, struct virtio_gpu_object, tbo);