Lines Matching refs:rdev
30 u32 si_gpu_check_soft_reset(struct radeon_device *rdev);
41 bool si_dma_is_lockup(struct radeon_device *rdev, struct radeon_ring *ring) in si_dma_is_lockup() argument
43 u32 reset_mask = si_gpu_check_soft_reset(rdev); in si_dma_is_lockup()
52 radeon_ring_lockup_update(rdev, ring); in si_dma_is_lockup()
55 return radeon_ring_test_lockup(rdev, ring); in si_dma_is_lockup()
69 void si_dma_vm_copy_pages(struct radeon_device *rdev, in si_dma_vm_copy_pages() argument
105 void si_dma_vm_write_pages(struct radeon_device *rdev, in si_dma_vm_write_pages() argument
125 value = radeon_vm_map_gart(rdev, addr); in si_dma_vm_write_pages()
152 void si_dma_vm_set_pages(struct radeon_device *rdev, in si_dma_vm_set_pages() argument
187 void si_dma_vm_flush(struct radeon_device *rdev, struct radeon_ring *ring, in si_dma_vm_flush() argument
231 struct radeon_fence *si_copy_dma(struct radeon_device *rdev, in si_copy_dma() argument
238 int ring_index = rdev->asic->copy.dma_ring_index; in si_copy_dma()
239 struct radeon_ring *ring = &rdev->ring[ring_index]; in si_copy_dma()
248 r = radeon_ring_lock(rdev, ring, num_loops * 5 + 11); in si_copy_dma()
251 radeon_sync_free(rdev, &sync, NULL); in si_copy_dma()
255 radeon_sync_resv(rdev, &sync, resv, false); in si_copy_dma()
256 radeon_sync_rings(rdev, &sync, ring->idx); in si_copy_dma()
272 r = radeon_fence_emit(rdev, &fence, ring->idx); in si_copy_dma()
274 radeon_ring_unlock_undo(rdev, ring); in si_copy_dma()
275 radeon_sync_free(rdev, &sync, NULL); in si_copy_dma()
279 radeon_ring_unlock_commit(rdev, ring, false); in si_copy_dma()
280 radeon_sync_free(rdev, &sync, fence); in si_copy_dma()