/linux-4.4.14/drivers/gpu/drm/radeon/ |
D | radeon_object.c | 96 void radeon_ttm_placement_from_domain(struct radeon_bo *rbo, u32 domain) in radeon_ttm_placement_from_domain() argument 100 rbo->placement.placement = rbo->placements; in radeon_ttm_placement_from_domain() 101 rbo->placement.busy_placement = rbo->placements; in radeon_ttm_placement_from_domain() 106 if ((rbo->flags & RADEON_GEM_NO_CPU_ACCESS) && in radeon_ttm_placement_from_domain() 107 rbo->rdev->mc.visible_vram_size < rbo->rdev->mc.real_vram_size) { in radeon_ttm_placement_from_domain() 108 rbo->placements[c].fpfn = in radeon_ttm_placement_from_domain() 109 rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT; in radeon_ttm_placement_from_domain() 110 rbo->placements[c++].flags = TTM_PL_FLAG_WC | in radeon_ttm_placement_from_domain() 115 rbo->placements[c].fpfn = 0; in radeon_ttm_placement_from_domain() 116 rbo->placements[c++].flags = TTM_PL_FLAG_WC | in radeon_ttm_placement_from_domain() [all …]
|
D | radeon_fb.c | 92 struct radeon_bo *rbo = gem_to_radeon_bo(gobj); in radeonfb_destroy_pinned_object() local 95 ret = radeon_bo_reserve(rbo, false); in radeonfb_destroy_pinned_object() 97 radeon_bo_kunmap(rbo); in radeonfb_destroy_pinned_object() 98 radeon_bo_unpin(rbo); in radeonfb_destroy_pinned_object() 99 radeon_bo_unreserve(rbo); in radeonfb_destroy_pinned_object() 110 struct radeon_bo *rbo = NULL; in radeonfb_create_pinned_object() local 136 rbo = gem_to_radeon_bo(gobj); in radeonfb_create_pinned_object() 154 ret = radeon_bo_set_tiling_flags(rbo, in radeonfb_create_pinned_object() 162 ret = radeon_bo_reserve(rbo, false); in radeonfb_create_pinned_object() 166 ret = radeon_bo_pin_restricted(rbo, RADEON_GEM_DOMAIN_VRAM, in radeonfb_create_pinned_object() [all …]
|
D | radeon_gem.c | 143 struct radeon_bo *rbo = gem_to_radeon_bo(obj); in radeon_gem_object_open() local 144 struct radeon_device *rdev = rbo->rdev; in radeon_gem_object_open() 155 r = radeon_bo_reserve(rbo, false); in radeon_gem_object_open() 160 bo_va = radeon_vm_bo_find(vm, rbo); in radeon_gem_object_open() 162 bo_va = radeon_vm_bo_add(rdev, vm, rbo); in radeon_gem_object_open() 166 radeon_bo_unreserve(rbo); in radeon_gem_object_open() 174 struct radeon_bo *rbo = gem_to_radeon_bo(obj); in radeon_gem_object_close() local 175 struct radeon_device *rdev = rbo->rdev; in radeon_gem_object_close() 186 r = radeon_bo_reserve(rbo, true); in radeon_gem_object_close() 192 bo_va = radeon_vm_bo_find(vm, rbo); in radeon_gem_object_close() [all …]
|
D | radeon_legacy_crtc.c | 377 struct radeon_bo *rbo; in radeon_crtc_do_set_base() local 424 rbo = gem_to_radeon_bo(obj); in radeon_crtc_do_set_base() 426 r = radeon_bo_reserve(rbo, false); in radeon_crtc_do_set_base() 430 r = radeon_bo_pin_restricted(rbo, RADEON_GEM_DOMAIN_VRAM, 1 << 27, in radeon_crtc_do_set_base() 433 radeon_bo_unreserve(rbo); in radeon_crtc_do_set_base() 453 nsize = radeon_bo_size(rbo); in radeon_crtc_do_set_base() 463 radeon_bo_get_tiling_flags(rbo, &tiling_flags, NULL); in radeon_crtc_do_set_base() 464 radeon_bo_unreserve(rbo); in radeon_crtc_do_set_base() 560 rbo = gem_to_radeon_bo(radeon_fb->obj); in radeon_crtc_do_set_base() 561 r = radeon_bo_reserve(rbo, false); in radeon_crtc_do_set_base() [all …]
|
D | radeon_ttm.c | 187 struct radeon_bo *rbo; in radeon_evict_flags() local 196 rbo = container_of(bo, struct radeon_bo, tbo); in radeon_evict_flags() 199 if (rbo->rdev->ring[radeon_copy_ring_index(rbo->rdev)].ready == false) in radeon_evict_flags() 200 radeon_ttm_placement_from_domain(rbo, RADEON_GEM_DOMAIN_CPU); in radeon_evict_flags() 201 else if (rbo->rdev->mc.visible_vram_size < rbo->rdev->mc.real_vram_size && in radeon_evict_flags() 202 bo->mem.start < (rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT)) { in radeon_evict_flags() 203 unsigned fpfn = rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT; in radeon_evict_flags() 211 radeon_ttm_placement_from_domain(rbo, RADEON_GEM_DOMAIN_VRAM | in radeon_evict_flags() 213 rbo->placement.num_busy_placement = 0; in radeon_evict_flags() 214 for (i = 0; i < rbo->placement.num_placement; i++) { in radeon_evict_flags() [all …]
|
D | radeon_uvd.c | 258 void radeon_uvd_force_into_uvd_segment(struct radeon_bo *rbo, in radeon_uvd_force_into_uvd_segment() argument 263 for (i = 0; i < rbo->placement.num_placement; ++i) { in radeon_uvd_force_into_uvd_segment() 264 rbo->placements[i].fpfn = 0 >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 265 rbo->placements[i].lpfn = (256 * 1024 * 1024) >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 273 if (rbo->placement.num_placement > 1) in radeon_uvd_force_into_uvd_segment() 277 rbo->placements[1] = rbo->placements[0]; in radeon_uvd_force_into_uvd_segment() 278 rbo->placements[1].fpfn += (256 * 1024 * 1024) >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 279 rbo->placements[1].lpfn += (256 * 1024 * 1024) >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 280 rbo->placement.num_placement++; in radeon_uvd_force_into_uvd_segment() 281 rbo->placement.num_busy_placement++; in radeon_uvd_force_into_uvd_segment()
|
D | atombios_crtc.c | 1145 struct radeon_bo *rbo; in dce4_crtc_do_set_base() local 1173 rbo = gem_to_radeon_bo(obj); in dce4_crtc_do_set_base() 1174 r = radeon_bo_reserve(rbo, false); in dce4_crtc_do_set_base() 1179 fb_location = radeon_bo_gpu_offset(rbo); in dce4_crtc_do_set_base() 1181 r = radeon_bo_pin(rbo, RADEON_GEM_DOMAIN_VRAM, &fb_location); in dce4_crtc_do_set_base() 1183 radeon_bo_unreserve(rbo); in dce4_crtc_do_set_base() 1188 radeon_bo_get_tiling_flags(rbo, &tiling_flags, NULL); in dce4_crtc_do_set_base() 1189 radeon_bo_unreserve(rbo); in dce4_crtc_do_set_base() 1438 rbo = gem_to_radeon_bo(radeon_fb->obj); in dce4_crtc_do_set_base() 1439 r = radeon_bo_reserve(rbo, false); in dce4_crtc_do_set_base() [all …]
|
D | radeon.h | 1694 void radeon_uvd_force_into_uvd_segment(struct radeon_bo *rbo, 2824 extern void radeon_ttm_placement_from_domain(struct radeon_bo *rbo, u32 domain);
|
/linux-4.4.14/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_fb.c | 92 struct amdgpu_bo *rbo = gem_to_amdgpu_bo(gobj); in amdgpufb_destroy_pinned_object() local 95 ret = amdgpu_bo_reserve(rbo, false); in amdgpufb_destroy_pinned_object() 97 amdgpu_bo_kunmap(rbo); in amdgpufb_destroy_pinned_object() 98 amdgpu_bo_unpin(rbo); in amdgpufb_destroy_pinned_object() 99 amdgpu_bo_unreserve(rbo); in amdgpufb_destroy_pinned_object() 110 struct amdgpu_bo *rbo = NULL; in amdgpufb_create_pinned_object() local 136 rbo = gem_to_amdgpu_bo(gobj); in amdgpufb_create_pinned_object() 141 ret = amdgpu_bo_reserve(rbo, false); in amdgpufb_create_pinned_object() 146 ret = amdgpu_bo_set_tiling_flags(rbo, in amdgpufb_create_pinned_object() 153 ret = amdgpu_bo_pin_restricted(rbo, AMDGPU_GEM_DOMAIN_VRAM, 0, 0, NULL); in amdgpufb_create_pinned_object() [all …]
|
D | amdgpu_gem.c | 112 struct amdgpu_bo *rbo = gem_to_amdgpu_bo(obj); in amdgpu_gem_object_open() local 113 struct amdgpu_device *adev = rbo->adev; in amdgpu_gem_object_open() 118 r = amdgpu_bo_reserve(rbo, false); in amdgpu_gem_object_open() 122 bo_va = amdgpu_vm_bo_find(vm, rbo); in amdgpu_gem_object_open() 124 bo_va = amdgpu_vm_bo_add(adev, vm, rbo); in amdgpu_gem_object_open() 128 amdgpu_bo_unreserve(rbo); in amdgpu_gem_object_open() 135 struct amdgpu_bo *rbo = gem_to_amdgpu_bo(obj); in amdgpu_gem_object_close() local 136 struct amdgpu_device *adev = rbo->adev; in amdgpu_gem_object_close() 141 r = amdgpu_bo_reserve(rbo, true); in amdgpu_gem_object_close() 147 bo_va = amdgpu_vm_bo_find(vm, rbo); in amdgpu_gem_object_close() [all …]
|
D | amdgpu_object.c | 198 void amdgpu_ttm_placement_from_domain(struct amdgpu_bo *rbo, u32 domain) in amdgpu_ttm_placement_from_domain() argument 200 amdgpu_ttm_placement_init(rbo->adev, &rbo->placement, in amdgpu_ttm_placement_from_domain() 201 rbo->placements, domain, rbo->flags); in amdgpu_ttm_placement_from_domain() 590 struct amdgpu_bo *rbo; in amdgpu_bo_move_notify() local 595 rbo = container_of(bo, struct amdgpu_bo, tbo); in amdgpu_bo_move_notify() 596 amdgpu_vm_bo_invalidate(rbo->adev, rbo); in amdgpu_bo_move_notify() 603 amdgpu_update_memory_usage(rbo->adev, &bo->mem, new_mem); in amdgpu_bo_move_notify()
|
D | amdgpu_ttm.c | 177 struct amdgpu_bo *rbo; in amdgpu_evict_flags() local 191 rbo = container_of(bo, struct amdgpu_bo, tbo); in amdgpu_evict_flags() 194 if (rbo->adev->mman.buffer_funcs_ring->ready == false) in amdgpu_evict_flags() 195 amdgpu_ttm_placement_from_domain(rbo, AMDGPU_GEM_DOMAIN_CPU); in amdgpu_evict_flags() 197 amdgpu_ttm_placement_from_domain(rbo, AMDGPU_GEM_DOMAIN_GTT); in amdgpu_evict_flags() 201 amdgpu_ttm_placement_from_domain(rbo, AMDGPU_GEM_DOMAIN_CPU); in amdgpu_evict_flags() 203 *placement = rbo->placement; in amdgpu_evict_flags() 208 struct amdgpu_bo *rbo = container_of(bo, struct amdgpu_bo, tbo); in amdgpu_verify_access() local 210 return drm_vma_node_verify_access(&rbo->gem_base.vma_node, filp); in amdgpu_verify_access()
|
D | dce_v8_0.c | 1986 struct amdgpu_bo *rbo; in dce_v8_0_crtc_do_set_base() local 2014 rbo = gem_to_amdgpu_bo(obj); in dce_v8_0_crtc_do_set_base() 2015 r = amdgpu_bo_reserve(rbo, false); in dce_v8_0_crtc_do_set_base() 2020 fb_location = amdgpu_bo_gpu_offset(rbo); in dce_v8_0_crtc_do_set_base() 2022 r = amdgpu_bo_pin(rbo, AMDGPU_GEM_DOMAIN_VRAM, &fb_location); in dce_v8_0_crtc_do_set_base() 2024 amdgpu_bo_unreserve(rbo); in dce_v8_0_crtc_do_set_base() 2029 amdgpu_bo_get_tiling_flags(rbo, &tiling_flags); in dce_v8_0_crtc_do_set_base() 2030 amdgpu_bo_unreserve(rbo); in dce_v8_0_crtc_do_set_base() 2186 rbo = gem_to_amdgpu_bo(amdgpu_fb->obj); in dce_v8_0_crtc_do_set_base() 2187 r = amdgpu_bo_reserve(rbo, false); in dce_v8_0_crtc_do_set_base() [all …]
|
D | dce_v11_0.c | 2037 struct amdgpu_bo *rbo; in dce_v11_0_crtc_do_set_base() local 2065 rbo = gem_to_amdgpu_bo(obj); in dce_v11_0_crtc_do_set_base() 2066 r = amdgpu_bo_reserve(rbo, false); in dce_v11_0_crtc_do_set_base() 2071 fb_location = amdgpu_bo_gpu_offset(rbo); in dce_v11_0_crtc_do_set_base() 2073 r = amdgpu_bo_pin(rbo, AMDGPU_GEM_DOMAIN_VRAM, &fb_location); in dce_v11_0_crtc_do_set_base() 2075 amdgpu_bo_unreserve(rbo); in dce_v11_0_crtc_do_set_base() 2080 amdgpu_bo_get_tiling_flags(rbo, &tiling_flags); in dce_v11_0_crtc_do_set_base() 2081 amdgpu_bo_unreserve(rbo); in dce_v11_0_crtc_do_set_base() 2254 rbo = gem_to_amdgpu_bo(amdgpu_fb->obj); in dce_v11_0_crtc_do_set_base() 2255 r = amdgpu_bo_reserve(rbo, false); in dce_v11_0_crtc_do_set_base() [all …]
|
D | dce_v10_0.c | 2049 struct amdgpu_bo *rbo; in dce_v10_0_crtc_do_set_base() local 2077 rbo = gem_to_amdgpu_bo(obj); in dce_v10_0_crtc_do_set_base() 2078 r = amdgpu_bo_reserve(rbo, false); in dce_v10_0_crtc_do_set_base() 2083 fb_location = amdgpu_bo_gpu_offset(rbo); in dce_v10_0_crtc_do_set_base() 2085 r = amdgpu_bo_pin(rbo, AMDGPU_GEM_DOMAIN_VRAM, &fb_location); in dce_v10_0_crtc_do_set_base() 2087 amdgpu_bo_unreserve(rbo); in dce_v10_0_crtc_do_set_base() 2092 amdgpu_bo_get_tiling_flags(rbo, &tiling_flags); in dce_v10_0_crtc_do_set_base() 2093 amdgpu_bo_unreserve(rbo); in dce_v10_0_crtc_do_set_base() 2266 rbo = gem_to_amdgpu_bo(amdgpu_fb->obj); in dce_v10_0_crtc_do_set_base() 2267 r = amdgpu_bo_reserve(rbo, false); in dce_v10_0_crtc_do_set_base() [all …]
|
D | amdgpu_uvd.c | 318 static void amdgpu_uvd_force_into_uvd_segment(struct amdgpu_bo *rbo) in amdgpu_uvd_force_into_uvd_segment() argument 321 for (i = 0; i < rbo->placement.num_placement; ++i) { in amdgpu_uvd_force_into_uvd_segment() 322 rbo->placements[i].fpfn = 0 >> PAGE_SHIFT; in amdgpu_uvd_force_into_uvd_segment() 323 rbo->placements[i].lpfn = (256 * 1024 * 1024) >> PAGE_SHIFT; in amdgpu_uvd_force_into_uvd_segment()
|
D | amdgpu.h | 2311 void amdgpu_ttm_placement_from_domain(struct amdgpu_bo *rbo, u32 domain);
|
/linux-4.4.14/drivers/gpu/drm/qxl/ |
D | qxl_release.c | 328 struct qxl_bo **rbo) in qxl_alloc_release_reserved() argument 349 if (rbo) in qxl_alloc_release_reserved() 350 *rbo = NULL; in qxl_alloc_release_reserved() 374 if (rbo) in qxl_alloc_release_reserved() 375 *rbo = bo; in qxl_alloc_release_reserved()
|
D | qxl_drv.h | 493 struct qxl_bo **rbo);
|