sched_job 183 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c static struct dma_fence *amdgpu_job_dependency(struct drm_sched_job *sched_job, sched_job 187 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c struct amdgpu_job *job = to_amdgpu_job(sched_job); sched_job 216 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c static struct dma_fence *amdgpu_job_run(struct drm_sched_job *sched_job) sched_job 218 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c struct amdgpu_ring *ring = to_amdgpu_ring(sched_job->sched); sched_job 223 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c job = to_amdgpu_job(sched_job); sched_job 35 drivers/gpu/drm/amd/amdgpu/amdgpu_job.h #define to_amdgpu_job(sched_job) \ sched_job 36 drivers/gpu/drm/amd/amdgpu/amdgpu_job.h container_of((sched_job), struct amdgpu_job, base) sched_job 468 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h TP_PROTO(struct amdgpu_job *sched_job, struct dma_fence *fence), sched_job 469 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h TP_ARGS(sched_job, fence), sched_job 471 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __string(ring, sched_job->base.sched->name); sched_job 479 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __assign_str(ring, sched_job->base.sched->name) sched_job 480 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __entry->id = sched_job->base.id; sched_job 93 drivers/gpu/drm/etnaviv/etnaviv_gem.h struct drm_sched_job sched_job; sched_job 21 drivers/gpu/drm/etnaviv/etnaviv_sched.c etnaviv_sched_dependency(struct drm_sched_job *sched_job, sched_job 24 drivers/gpu/drm/etnaviv/etnaviv_sched.c struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); sched_job 72 drivers/gpu/drm/etnaviv/etnaviv_sched.c static struct dma_fence *etnaviv_sched_run_job(struct drm_sched_job *sched_job) sched_job 74 drivers/gpu/drm/etnaviv/etnaviv_sched.c struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); sched_job 77 drivers/gpu/drm/etnaviv/etnaviv_sched.c if (likely(!sched_job->s_fence->finished.error)) sched_job 85 drivers/gpu/drm/etnaviv/etnaviv_sched.c static void etnaviv_sched_timedout_job(struct drm_sched_job *sched_job) sched_job 87 drivers/gpu/drm/etnaviv/etnaviv_sched.c struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); sched_job 112 drivers/gpu/drm/etnaviv/etnaviv_sched.c drm_sched_stop(&gpu->sched, sched_job); sched_job 114 drivers/gpu/drm/etnaviv/etnaviv_sched.c if(sched_job) sched_job 115 drivers/gpu/drm/etnaviv/etnaviv_sched.c drm_sched_increase_karma(sched_job); sched_job 127 drivers/gpu/drm/etnaviv/etnaviv_sched.c static void etnaviv_sched_free_job(struct drm_sched_job *sched_job) sched_job 129 drivers/gpu/drm/etnaviv/etnaviv_sched.c struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); sched_job 131 drivers/gpu/drm/etnaviv/etnaviv_sched.c drm_sched_job_cleanup(sched_job); sched_job 155 drivers/gpu/drm/etnaviv/etnaviv_sched.c ret = drm_sched_job_init(&submit->sched_job, sched_entity, sched_job 160 drivers/gpu/drm/etnaviv/etnaviv_sched.c submit->out_fence = dma_fence_get(&submit->sched_job.s_fence->finished); sched_job 165 drivers/gpu/drm/etnaviv/etnaviv_sched.c drm_sched_job_cleanup(&submit->sched_job); sched_job 173 drivers/gpu/drm/etnaviv/etnaviv_sched.c drm_sched_entity_push_job(&submit->sched_job, sched_entity); sched_job 14 drivers/gpu/drm/etnaviv/etnaviv_sched.h struct etnaviv_gem_submit *to_etnaviv_submit(struct drm_sched_job *sched_job) sched_job 16 drivers/gpu/drm/etnaviv/etnaviv_sched.h return container_of(sched_job, struct etnaviv_gem_submit, sched_job); sched_job 39 drivers/gpu/drm/panfrost/panfrost_job.c to_panfrost_job(struct drm_sched_job *sched_job) sched_job 41 drivers/gpu/drm/panfrost/panfrost_job.c return container_of(sched_job, struct panfrost_job, base); sched_job 302 drivers/gpu/drm/panfrost/panfrost_job.c static void panfrost_job_free(struct drm_sched_job *sched_job) sched_job 304 drivers/gpu/drm/panfrost/panfrost_job.c struct panfrost_job *job = to_panfrost_job(sched_job); sched_job 306 drivers/gpu/drm/panfrost/panfrost_job.c drm_sched_job_cleanup(sched_job); sched_job 311 drivers/gpu/drm/panfrost/panfrost_job.c static struct dma_fence *panfrost_job_dependency(struct drm_sched_job *sched_job, sched_job 314 drivers/gpu/drm/panfrost/panfrost_job.c struct panfrost_job *job = to_panfrost_job(sched_job); sched_job 339 drivers/gpu/drm/panfrost/panfrost_job.c static struct dma_fence *panfrost_job_run(struct drm_sched_job *sched_job) sched_job 341 drivers/gpu/drm/panfrost/panfrost_job.c struct panfrost_job *job = to_panfrost_job(sched_job); sched_job 377 drivers/gpu/drm/panfrost/panfrost_job.c static void panfrost_job_timedout(struct drm_sched_job *sched_job) sched_job 379 drivers/gpu/drm/panfrost/panfrost_job.c struct panfrost_job *job = to_panfrost_job(sched_job); sched_job 398 drivers/gpu/drm/panfrost/panfrost_job.c sched_job); sched_job 406 drivers/gpu/drm/panfrost/panfrost_job.c drm_sched_stop(sched, sched_job); sched_job 412 drivers/gpu/drm/panfrost/panfrost_job.c drm_sched_increase_karma(sched_job); sched_job 36 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity), sched_job 37 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h TP_ARGS(sched_job, entity), sched_job 49 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h __entry->id = sched_job->id; sched_job 50 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h __entry->fence = &sched_job->s_fence->finished; sched_job 51 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h __entry->name = sched_job->sched->name; sched_job 54 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h &sched_job->sched->hw_rq_count); sched_job 76 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h TP_PROTO(struct drm_sched_job *sched_job, struct dma_fence *fence), sched_job 77 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h TP_ARGS(sched_job, fence), sched_job 87 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h __entry->name = sched_job->sched->name; sched_job 88 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h __entry->id = sched_job->id; sched_job 32 drivers/gpu/drm/scheduler/sched_entity.c #define to_drm_sched_job(sched_job) \ sched_job 33 drivers/gpu/drm/scheduler/sched_entity.c container_of((sched_job), struct drm_sched_job, queue_node) sched_job 451 drivers/gpu/drm/scheduler/sched_entity.c struct drm_sched_job *sched_job; sched_job 453 drivers/gpu/drm/scheduler/sched_entity.c sched_job = to_drm_sched_job(spsc_queue_peek(&entity->job_queue)); sched_job 454 drivers/gpu/drm/scheduler/sched_entity.c if (!sched_job) sched_job 458 drivers/gpu/drm/scheduler/sched_entity.c sched->ops->dependency(sched_job, entity))) { sched_job 459 drivers/gpu/drm/scheduler/sched_entity.c trace_drm_sched_job_wait_dep(sched_job, entity->dependency); sched_job 467 drivers/gpu/drm/scheduler/sched_entity.c dma_fence_set_error(&sched_job->s_fence->finished, -ECANCELED); sched_job 470 drivers/gpu/drm/scheduler/sched_entity.c entity->last_scheduled = dma_fence_get(&sched_job->s_fence->finished); sched_job 473 drivers/gpu/drm/scheduler/sched_entity.c return sched_job; sched_job 518 drivers/gpu/drm/scheduler/sched_entity.c void drm_sched_entity_push_job(struct drm_sched_job *sched_job, sched_job 523 drivers/gpu/drm/scheduler/sched_entity.c trace_drm_sched_job(sched_job, entity); sched_job 526 drivers/gpu/drm/scheduler/sched_entity.c first = spsc_queue_push(&entity->job_queue, &sched_job->queue_node); sched_job 59 drivers/gpu/drm/scheduler/sched_main.c #define to_drm_sched_job(sched_job) \ sched_job 60 drivers/gpu/drm/scheduler/sched_main.c container_of((sched_job), struct drm_sched_job, queue_node) sched_job 709 drivers/gpu/drm/scheduler/sched_main.c struct drm_sched_job *sched_job; sched_job 728 drivers/gpu/drm/scheduler/sched_main.c sched_job = drm_sched_entity_pop_job(entity); sched_job 729 drivers/gpu/drm/scheduler/sched_main.c if (!sched_job) sched_job 732 drivers/gpu/drm/scheduler/sched_main.c s_fence = sched_job->s_fence; sched_job 735 drivers/gpu/drm/scheduler/sched_main.c drm_sched_job_begin(sched_job); sched_job 737 drivers/gpu/drm/scheduler/sched_main.c fence = sched->ops->run_job(sched_job); sched_job 742 drivers/gpu/drm/scheduler/sched_main.c r = dma_fence_add_callback(fence, &sched_job->cb, sched_job 745 drivers/gpu/drm/scheduler/sched_main.c drm_sched_process_job(fence, &sched_job->cb); sched_job 753 drivers/gpu/drm/scheduler/sched_main.c drm_sched_process_job(NULL, &sched_job->cb); sched_job 28 drivers/gpu/drm/v3d/v3d_sched.c to_v3d_job(struct drm_sched_job *sched_job) sched_job 30 drivers/gpu/drm/v3d/v3d_sched.c return container_of(sched_job, struct v3d_job, base); sched_job 34 drivers/gpu/drm/v3d/v3d_sched.c to_bin_job(struct drm_sched_job *sched_job) sched_job 36 drivers/gpu/drm/v3d/v3d_sched.c return container_of(sched_job, struct v3d_bin_job, base.base); sched_job 40 drivers/gpu/drm/v3d/v3d_sched.c to_render_job(struct drm_sched_job *sched_job) sched_job 42 drivers/gpu/drm/v3d/v3d_sched.c return container_of(sched_job, struct v3d_render_job, base.base); sched_job 46 drivers/gpu/drm/v3d/v3d_sched.c to_tfu_job(struct drm_sched_job *sched_job) sched_job 48 drivers/gpu/drm/v3d/v3d_sched.c return container_of(sched_job, struct v3d_tfu_job, base.base); sched_job 52 drivers/gpu/drm/v3d/v3d_sched.c to_csd_job(struct drm_sched_job *sched_job) sched_job 54 drivers/gpu/drm/v3d/v3d_sched.c return container_of(sched_job, struct v3d_csd_job, base.base); sched_job 58 drivers/gpu/drm/v3d/v3d_sched.c v3d_job_free(struct drm_sched_job *sched_job) sched_job 60 drivers/gpu/drm/v3d/v3d_sched.c struct v3d_job *job = to_v3d_job(sched_job); sched_job 62 drivers/gpu/drm/v3d/v3d_sched.c drm_sched_job_cleanup(sched_job); sched_job 73 drivers/gpu/drm/v3d/v3d_sched.c v3d_job_dependency(struct drm_sched_job *sched_job, sched_job 76 drivers/gpu/drm/v3d/v3d_sched.c struct v3d_job *job = to_v3d_job(sched_job); sched_job 88 drivers/gpu/drm/v3d/v3d_sched.c static struct dma_fence *v3d_bin_job_run(struct drm_sched_job *sched_job) sched_job 90 drivers/gpu/drm/v3d/v3d_sched.c struct v3d_bin_job *job = to_bin_job(sched_job); sched_job 141 drivers/gpu/drm/v3d/v3d_sched.c static struct dma_fence *v3d_render_job_run(struct drm_sched_job *sched_job) sched_job 143 drivers/gpu/drm/v3d/v3d_sched.c struct v3d_render_job *job = to_render_job(sched_job); sched_job 184 drivers/gpu/drm/v3d/v3d_sched.c v3d_tfu_job_run(struct drm_sched_job *sched_job) sched_job 186 drivers/gpu/drm/v3d/v3d_sched.c struct v3d_tfu_job *job = to_tfu_job(sched_job); sched_job 221 drivers/gpu/drm/v3d/v3d_sched.c v3d_csd_job_run(struct drm_sched_job *sched_job) sched_job 223 drivers/gpu/drm/v3d/v3d_sched.c struct v3d_csd_job *job = to_csd_job(sched_job); sched_job 252 drivers/gpu/drm/v3d/v3d_sched.c v3d_cache_clean_job_run(struct drm_sched_job *sched_job) sched_job 254 drivers/gpu/drm/v3d/v3d_sched.c struct v3d_job *job = to_v3d_job(sched_job); sched_job 263 drivers/gpu/drm/v3d/v3d_sched.c v3d_gpu_reset_for_timeout(struct v3d_dev *v3d, struct drm_sched_job *sched_job) sched_job 271 drivers/gpu/drm/v3d/v3d_sched.c drm_sched_stop(&v3d->queue[q].sched, sched_job); sched_job 273 drivers/gpu/drm/v3d/v3d_sched.c if (sched_job) sched_job 274 drivers/gpu/drm/v3d/v3d_sched.c drm_sched_increase_karma(sched_job); sched_job 296 drivers/gpu/drm/v3d/v3d_sched.c v3d_cl_job_timedout(struct drm_sched_job *sched_job, enum v3d_queue q, sched_job 299 drivers/gpu/drm/v3d/v3d_sched.c struct v3d_job *job = to_v3d_job(sched_job); sched_job 310 drivers/gpu/drm/v3d/v3d_sched.c v3d_gpu_reset_for_timeout(v3d, sched_job); sched_job 314 drivers/gpu/drm/v3d/v3d_sched.c v3d_bin_job_timedout(struct drm_sched_job *sched_job) sched_job 316 drivers/gpu/drm/v3d/v3d_sched.c struct v3d_bin_job *job = to_bin_job(sched_job); sched_job 318 drivers/gpu/drm/v3d/v3d_sched.c v3d_cl_job_timedout(sched_job, V3D_BIN, sched_job 323 drivers/gpu/drm/v3d/v3d_sched.c v3d_render_job_timedout(struct drm_sched_job *sched_job) sched_job 325 drivers/gpu/drm/v3d/v3d_sched.c struct v3d_render_job *job = to_render_job(sched_job); sched_job 327 drivers/gpu/drm/v3d/v3d_sched.c v3d_cl_job_timedout(sched_job, V3D_RENDER, sched_job 332 drivers/gpu/drm/v3d/v3d_sched.c v3d_generic_job_timedout(struct drm_sched_job *sched_job) sched_job 334 drivers/gpu/drm/v3d/v3d_sched.c struct v3d_job *job = to_v3d_job(sched_job); sched_job 336 drivers/gpu/drm/v3d/v3d_sched.c v3d_gpu_reset_for_timeout(job->v3d, sched_job); sched_job 340 drivers/gpu/drm/v3d/v3d_sched.c v3d_csd_job_timedout(struct drm_sched_job *sched_job) sched_job 342 drivers/gpu/drm/v3d/v3d_sched.c struct v3d_csd_job *job = to_csd_job(sched_job); sched_job 354 drivers/gpu/drm/v3d/v3d_sched.c v3d_gpu_reset_for_timeout(v3d, sched_job); sched_job 214 include/drm/gpu_scheduler.h struct dma_fence *(*dependency)(struct drm_sched_job *sched_job, sched_job 223 include/drm/gpu_scheduler.h struct dma_fence *(*run_job)(struct drm_sched_job *sched_job); sched_job 229 include/drm/gpu_scheduler.h void (*timedout_job)(struct drm_sched_job *sched_job); sched_job 235 include/drm/gpu_scheduler.h void (*free_job)(struct drm_sched_job *sched_job); sched_job 320 include/drm/gpu_scheduler.h void drm_sched_entity_push_job(struct drm_sched_job *sched_job,