Home
last modified time | relevance | path

Searched refs:s_fence (Results 1 – 10 of 10) sorted by relevance

/linux-4.4.14/drivers/gpu/drm/amd/scheduler/
Dgpu_scheduler.c235 struct amd_sched_fence *s_fence; in amd_sched_entity_add_dependency_cb() local
243 s_fence = to_amd_sched_fence(fence); in amd_sched_entity_add_dependency_cb()
244 if (s_fence && s_fence->sched == sched) { in amd_sched_entity_add_dependency_cb()
254 list_add_tail(&entity->cb.node, &s_fence->scheduled_cb); in amd_sched_entity_add_dependency_cb()
366 struct amd_sched_fence *s_fence = in amd_sched_process_job() local
368 struct amd_gpu_scheduler *sched = s_fence->sched; in amd_sched_process_job()
372 amd_sched_fence_signal(s_fence); in amd_sched_process_job()
374 cancel_delayed_work(&s_fence->dwork); in amd_sched_process_job()
376 list_del_init(&s_fence->list); in amd_sched_process_job()
379 trace_amd_sched_process_job(s_fence); in amd_sched_process_job()
[all …]
Dsched_fence.c60 void amd_sched_fence_scheduled(struct amd_sched_fence *s_fence) in amd_sched_fence_scheduled() argument
64 set_bit(AMD_SCHED_FENCE_SCHEDULED_BIT, &s_fence->base.flags); in amd_sched_fence_scheduled()
65 list_for_each_entry_safe(cur, tmp, &s_fence->scheduled_cb, node) { in amd_sched_fence_scheduled()
67 cur->func(&s_fence->base, cur); in amd_sched_fence_scheduled()
Dgpu_sched_trace.h29 __entry->fence = &sched_job->s_fence->base;
Dgpu_scheduler.h84 struct amd_sched_fence *s_fence; member
/linux-4.4.14/drivers/gpu/drm/amd/amdgpu/
Damdgpu_sync.c64 struct amd_sched_fence *s_fence = to_amd_sched_fence(f); in amdgpu_sync_same_dev() local
69 if (s_fence) { in amdgpu_sync_same_dev()
72 ring = container_of(s_fence->sched, struct amdgpu_ring, sched); in amdgpu_sync_same_dev()
82 struct amd_sched_fence *s_fence = to_amd_sched_fence(f); in amdgpu_sync_test_owner() local
83 if (s_fence) in amdgpu_sync_test_owner()
84 return s_fence->owner == owner; in amdgpu_sync_test_owner()
146 struct amd_sched_fence *s_fence = to_amd_sched_fence(f); in amdgpu_sync_get_owner() local
148 if (s_fence) in amdgpu_sync_get_owner()
149 return s_fence->owner; in amdgpu_sync_get_owner()
Damdgpu_sched.c87 job->base.s_fence = amd_sched_fence_create(job->base.s_entity, owner); in amdgpu_sched_ib_submit_kernel_helper()
88 if (!job->base.s_fence) { in amdgpu_sched_ib_submit_kernel_helper()
92 *f = fence_get(&job->base.s_fence->base); in amdgpu_sched_ib_submit_kernel_helper()
Damdgpu_sa.c413 struct amd_sched_fence *s_fence = to_amd_sched_fence(fence); in amdgpu_sa_bo_dump_fence() local
419 if (s_fence) { in amdgpu_sa_bo_dump_fence()
423 ring = container_of(s_fence->sched, struct amdgpu_ring, sched); in amdgpu_sa_bo_dump_fence()
425 s_fence->base.seqno, ring->idx); in amdgpu_sa_bo_dump_fence()
Damdgpu_ring.c450 struct amd_sched_fence *s_fence; in amdgpu_ring_from_fence() local
452 s_fence = to_amd_sched_fence(f); in amdgpu_ring_from_fence()
453 if (s_fence) in amdgpu_ring_from_fence()
454 return container_of(s_fence->sched, struct amdgpu_ring, sched); in amdgpu_ring_from_fence()
Damdgpu_trace.h67 __entry->fence = &job->base.s_fence->base;
92 __entry->fence = &job->base.s_fence->base;
Damdgpu_cs.c889 job->base.s_fence = fence; in amdgpu_cs_ioctl()