sched_job         183 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c static struct dma_fence *amdgpu_job_dependency(struct drm_sched_job *sched_job,
sched_job         187 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c 	struct amdgpu_job *job = to_amdgpu_job(sched_job);
sched_job         216 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c static struct dma_fence *amdgpu_job_run(struct drm_sched_job *sched_job)
sched_job         218 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c 	struct amdgpu_ring *ring = to_amdgpu_ring(sched_job->sched);
sched_job         223 drivers/gpu/drm/amd/amdgpu/amdgpu_job.c 	job = to_amdgpu_job(sched_job);
sched_job          35 drivers/gpu/drm/amd/amdgpu/amdgpu_job.h #define to_amdgpu_job(sched_job)		\
sched_job          36 drivers/gpu/drm/amd/amdgpu/amdgpu_job.h 		container_of((sched_job), struct amdgpu_job, base)
sched_job         468 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 	    TP_PROTO(struct amdgpu_job *sched_job, struct dma_fence *fence),
sched_job         469 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 	    TP_ARGS(sched_job, fence),
sched_job         471 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 			     __string(ring, sched_job->base.sched->name);
sched_job         479 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 			   __assign_str(ring, sched_job->base.sched->name)
sched_job         480 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 			   __entry->id = sched_job->base.id;
sched_job          93 drivers/gpu/drm/etnaviv/etnaviv_gem.h 	struct drm_sched_job sched_job;
sched_job          21 drivers/gpu/drm/etnaviv/etnaviv_sched.c etnaviv_sched_dependency(struct drm_sched_job *sched_job,
sched_job          24 drivers/gpu/drm/etnaviv/etnaviv_sched.c 	struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job);
sched_job          72 drivers/gpu/drm/etnaviv/etnaviv_sched.c static struct dma_fence *etnaviv_sched_run_job(struct drm_sched_job *sched_job)
sched_job          74 drivers/gpu/drm/etnaviv/etnaviv_sched.c 	struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job);
sched_job          77 drivers/gpu/drm/etnaviv/etnaviv_sched.c 	if (likely(!sched_job->s_fence->finished.error))
sched_job          85 drivers/gpu/drm/etnaviv/etnaviv_sched.c static void etnaviv_sched_timedout_job(struct drm_sched_job *sched_job)
sched_job          87 drivers/gpu/drm/etnaviv/etnaviv_sched.c 	struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job);
sched_job         112 drivers/gpu/drm/etnaviv/etnaviv_sched.c 	drm_sched_stop(&gpu->sched, sched_job);
sched_job         114 drivers/gpu/drm/etnaviv/etnaviv_sched.c 	if(sched_job)
sched_job         115 drivers/gpu/drm/etnaviv/etnaviv_sched.c 		drm_sched_increase_karma(sched_job);
sched_job         127 drivers/gpu/drm/etnaviv/etnaviv_sched.c static void etnaviv_sched_free_job(struct drm_sched_job *sched_job)
sched_job         129 drivers/gpu/drm/etnaviv/etnaviv_sched.c 	struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job);
sched_job         131 drivers/gpu/drm/etnaviv/etnaviv_sched.c 	drm_sched_job_cleanup(sched_job);
sched_job         155 drivers/gpu/drm/etnaviv/etnaviv_sched.c 	ret = drm_sched_job_init(&submit->sched_job, sched_entity,
sched_job         160 drivers/gpu/drm/etnaviv/etnaviv_sched.c 	submit->out_fence = dma_fence_get(&submit->sched_job.s_fence->finished);
sched_job         165 drivers/gpu/drm/etnaviv/etnaviv_sched.c 		drm_sched_job_cleanup(&submit->sched_job);
sched_job         173 drivers/gpu/drm/etnaviv/etnaviv_sched.c 	drm_sched_entity_push_job(&submit->sched_job, sched_entity);
sched_job          14 drivers/gpu/drm/etnaviv/etnaviv_sched.h struct etnaviv_gem_submit *to_etnaviv_submit(struct drm_sched_job *sched_job)
sched_job          16 drivers/gpu/drm/etnaviv/etnaviv_sched.h 	return container_of(sched_job, struct etnaviv_gem_submit, sched_job);
sched_job          39 drivers/gpu/drm/panfrost/panfrost_job.c to_panfrost_job(struct drm_sched_job *sched_job)
sched_job          41 drivers/gpu/drm/panfrost/panfrost_job.c 	return container_of(sched_job, struct panfrost_job, base);
sched_job         302 drivers/gpu/drm/panfrost/panfrost_job.c static void panfrost_job_free(struct drm_sched_job *sched_job)
sched_job         304 drivers/gpu/drm/panfrost/panfrost_job.c 	struct panfrost_job *job = to_panfrost_job(sched_job);
sched_job         306 drivers/gpu/drm/panfrost/panfrost_job.c 	drm_sched_job_cleanup(sched_job);
sched_job         311 drivers/gpu/drm/panfrost/panfrost_job.c static struct dma_fence *panfrost_job_dependency(struct drm_sched_job *sched_job,
sched_job         314 drivers/gpu/drm/panfrost/panfrost_job.c 	struct panfrost_job *job = to_panfrost_job(sched_job);
sched_job         339 drivers/gpu/drm/panfrost/panfrost_job.c static struct dma_fence *panfrost_job_run(struct drm_sched_job *sched_job)
sched_job         341 drivers/gpu/drm/panfrost/panfrost_job.c 	struct panfrost_job *job = to_panfrost_job(sched_job);
sched_job         377 drivers/gpu/drm/panfrost/panfrost_job.c static void panfrost_job_timedout(struct drm_sched_job *sched_job)
sched_job         379 drivers/gpu/drm/panfrost/panfrost_job.c 	struct panfrost_job *job = to_panfrost_job(sched_job);
sched_job         398 drivers/gpu/drm/panfrost/panfrost_job.c 		sched_job);
sched_job         406 drivers/gpu/drm/panfrost/panfrost_job.c 		drm_sched_stop(sched, sched_job);
sched_job         412 drivers/gpu/drm/panfrost/panfrost_job.c 	drm_sched_increase_karma(sched_job);
sched_job          36 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h 	    TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity),
sched_job          37 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h 	    TP_ARGS(sched_job, entity),
sched_job          49 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h 			   __entry->id = sched_job->id;
sched_job          50 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h 			   __entry->fence = &sched_job->s_fence->finished;
sched_job          51 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h 			   __entry->name = sched_job->sched->name;
sched_job          54 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h 				   &sched_job->sched->hw_rq_count);
sched_job          76 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h 	    TP_PROTO(struct drm_sched_job *sched_job, struct dma_fence *fence),
sched_job          77 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h 	    TP_ARGS(sched_job, fence),
sched_job          87 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h 			   __entry->name = sched_job->sched->name;
sched_job          88 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h 			   __entry->id = sched_job->id;
sched_job          32 drivers/gpu/drm/scheduler/sched_entity.c #define to_drm_sched_job(sched_job)		\
sched_job          33 drivers/gpu/drm/scheduler/sched_entity.c 		container_of((sched_job), struct drm_sched_job, queue_node)
sched_job         451 drivers/gpu/drm/scheduler/sched_entity.c 	struct drm_sched_job *sched_job;
sched_job         453 drivers/gpu/drm/scheduler/sched_entity.c 	sched_job = to_drm_sched_job(spsc_queue_peek(&entity->job_queue));
sched_job         454 drivers/gpu/drm/scheduler/sched_entity.c 	if (!sched_job)
sched_job         458 drivers/gpu/drm/scheduler/sched_entity.c 			sched->ops->dependency(sched_job, entity))) {
sched_job         459 drivers/gpu/drm/scheduler/sched_entity.c 		trace_drm_sched_job_wait_dep(sched_job, entity->dependency);
sched_job         467 drivers/gpu/drm/scheduler/sched_entity.c 		dma_fence_set_error(&sched_job->s_fence->finished, -ECANCELED);
sched_job         470 drivers/gpu/drm/scheduler/sched_entity.c 	entity->last_scheduled = dma_fence_get(&sched_job->s_fence->finished);
sched_job         473 drivers/gpu/drm/scheduler/sched_entity.c 	return sched_job;
sched_job         518 drivers/gpu/drm/scheduler/sched_entity.c void drm_sched_entity_push_job(struct drm_sched_job *sched_job,
sched_job         523 drivers/gpu/drm/scheduler/sched_entity.c 	trace_drm_sched_job(sched_job, entity);
sched_job         526 drivers/gpu/drm/scheduler/sched_entity.c 	first = spsc_queue_push(&entity->job_queue, &sched_job->queue_node);
sched_job          59 drivers/gpu/drm/scheduler/sched_main.c #define to_drm_sched_job(sched_job)		\
sched_job          60 drivers/gpu/drm/scheduler/sched_main.c 		container_of((sched_job), struct drm_sched_job, queue_node)
sched_job         709 drivers/gpu/drm/scheduler/sched_main.c 		struct drm_sched_job *sched_job;
sched_job         728 drivers/gpu/drm/scheduler/sched_main.c 		sched_job = drm_sched_entity_pop_job(entity);
sched_job         729 drivers/gpu/drm/scheduler/sched_main.c 		if (!sched_job)
sched_job         732 drivers/gpu/drm/scheduler/sched_main.c 		s_fence = sched_job->s_fence;
sched_job         735 drivers/gpu/drm/scheduler/sched_main.c 		drm_sched_job_begin(sched_job);
sched_job         737 drivers/gpu/drm/scheduler/sched_main.c 		fence = sched->ops->run_job(sched_job);
sched_job         742 drivers/gpu/drm/scheduler/sched_main.c 			r = dma_fence_add_callback(fence, &sched_job->cb,
sched_job         745 drivers/gpu/drm/scheduler/sched_main.c 				drm_sched_process_job(fence, &sched_job->cb);
sched_job         753 drivers/gpu/drm/scheduler/sched_main.c 			drm_sched_process_job(NULL, &sched_job->cb);
sched_job          28 drivers/gpu/drm/v3d/v3d_sched.c to_v3d_job(struct drm_sched_job *sched_job)
sched_job          30 drivers/gpu/drm/v3d/v3d_sched.c 	return container_of(sched_job, struct v3d_job, base);
sched_job          34 drivers/gpu/drm/v3d/v3d_sched.c to_bin_job(struct drm_sched_job *sched_job)
sched_job          36 drivers/gpu/drm/v3d/v3d_sched.c 	return container_of(sched_job, struct v3d_bin_job, base.base);
sched_job          40 drivers/gpu/drm/v3d/v3d_sched.c to_render_job(struct drm_sched_job *sched_job)
sched_job          42 drivers/gpu/drm/v3d/v3d_sched.c 	return container_of(sched_job, struct v3d_render_job, base.base);
sched_job          46 drivers/gpu/drm/v3d/v3d_sched.c to_tfu_job(struct drm_sched_job *sched_job)
sched_job          48 drivers/gpu/drm/v3d/v3d_sched.c 	return container_of(sched_job, struct v3d_tfu_job, base.base);
sched_job          52 drivers/gpu/drm/v3d/v3d_sched.c to_csd_job(struct drm_sched_job *sched_job)
sched_job          54 drivers/gpu/drm/v3d/v3d_sched.c 	return container_of(sched_job, struct v3d_csd_job, base.base);
sched_job          58 drivers/gpu/drm/v3d/v3d_sched.c v3d_job_free(struct drm_sched_job *sched_job)
sched_job          60 drivers/gpu/drm/v3d/v3d_sched.c 	struct v3d_job *job = to_v3d_job(sched_job);
sched_job          62 drivers/gpu/drm/v3d/v3d_sched.c 	drm_sched_job_cleanup(sched_job);
sched_job          73 drivers/gpu/drm/v3d/v3d_sched.c v3d_job_dependency(struct drm_sched_job *sched_job,
sched_job          76 drivers/gpu/drm/v3d/v3d_sched.c 	struct v3d_job *job = to_v3d_job(sched_job);
sched_job          88 drivers/gpu/drm/v3d/v3d_sched.c static struct dma_fence *v3d_bin_job_run(struct drm_sched_job *sched_job)
sched_job          90 drivers/gpu/drm/v3d/v3d_sched.c 	struct v3d_bin_job *job = to_bin_job(sched_job);
sched_job         141 drivers/gpu/drm/v3d/v3d_sched.c static struct dma_fence *v3d_render_job_run(struct drm_sched_job *sched_job)
sched_job         143 drivers/gpu/drm/v3d/v3d_sched.c 	struct v3d_render_job *job = to_render_job(sched_job);
sched_job         184 drivers/gpu/drm/v3d/v3d_sched.c v3d_tfu_job_run(struct drm_sched_job *sched_job)
sched_job         186 drivers/gpu/drm/v3d/v3d_sched.c 	struct v3d_tfu_job *job = to_tfu_job(sched_job);
sched_job         221 drivers/gpu/drm/v3d/v3d_sched.c v3d_csd_job_run(struct drm_sched_job *sched_job)
sched_job         223 drivers/gpu/drm/v3d/v3d_sched.c 	struct v3d_csd_job *job = to_csd_job(sched_job);
sched_job         252 drivers/gpu/drm/v3d/v3d_sched.c v3d_cache_clean_job_run(struct drm_sched_job *sched_job)
sched_job         254 drivers/gpu/drm/v3d/v3d_sched.c 	struct v3d_job *job = to_v3d_job(sched_job);
sched_job         263 drivers/gpu/drm/v3d/v3d_sched.c v3d_gpu_reset_for_timeout(struct v3d_dev *v3d, struct drm_sched_job *sched_job)
sched_job         271 drivers/gpu/drm/v3d/v3d_sched.c 		drm_sched_stop(&v3d->queue[q].sched, sched_job);
sched_job         273 drivers/gpu/drm/v3d/v3d_sched.c 	if (sched_job)
sched_job         274 drivers/gpu/drm/v3d/v3d_sched.c 		drm_sched_increase_karma(sched_job);
sched_job         296 drivers/gpu/drm/v3d/v3d_sched.c v3d_cl_job_timedout(struct drm_sched_job *sched_job, enum v3d_queue q,
sched_job         299 drivers/gpu/drm/v3d/v3d_sched.c 	struct v3d_job *job = to_v3d_job(sched_job);
sched_job         310 drivers/gpu/drm/v3d/v3d_sched.c 	v3d_gpu_reset_for_timeout(v3d, sched_job);
sched_job         314 drivers/gpu/drm/v3d/v3d_sched.c v3d_bin_job_timedout(struct drm_sched_job *sched_job)
sched_job         316 drivers/gpu/drm/v3d/v3d_sched.c 	struct v3d_bin_job *job = to_bin_job(sched_job);
sched_job         318 drivers/gpu/drm/v3d/v3d_sched.c 	v3d_cl_job_timedout(sched_job, V3D_BIN,
sched_job         323 drivers/gpu/drm/v3d/v3d_sched.c v3d_render_job_timedout(struct drm_sched_job *sched_job)
sched_job         325 drivers/gpu/drm/v3d/v3d_sched.c 	struct v3d_render_job *job = to_render_job(sched_job);
sched_job         327 drivers/gpu/drm/v3d/v3d_sched.c 	v3d_cl_job_timedout(sched_job, V3D_RENDER,
sched_job         332 drivers/gpu/drm/v3d/v3d_sched.c v3d_generic_job_timedout(struct drm_sched_job *sched_job)
sched_job         334 drivers/gpu/drm/v3d/v3d_sched.c 	struct v3d_job *job = to_v3d_job(sched_job);
sched_job         336 drivers/gpu/drm/v3d/v3d_sched.c 	v3d_gpu_reset_for_timeout(job->v3d, sched_job);
sched_job         340 drivers/gpu/drm/v3d/v3d_sched.c v3d_csd_job_timedout(struct drm_sched_job *sched_job)
sched_job         342 drivers/gpu/drm/v3d/v3d_sched.c 	struct v3d_csd_job *job = to_csd_job(sched_job);
sched_job         354 drivers/gpu/drm/v3d/v3d_sched.c 	v3d_gpu_reset_for_timeout(v3d, sched_job);
sched_job         214 include/drm/gpu_scheduler.h 	struct dma_fence *(*dependency)(struct drm_sched_job *sched_job,
sched_job         223 include/drm/gpu_scheduler.h 	struct dma_fence *(*run_job)(struct drm_sched_job *sched_job);
sched_job         229 include/drm/gpu_scheduler.h 	void (*timedout_job)(struct drm_sched_job *sched_job);
sched_job         235 include/drm/gpu_scheduler.h 	void (*free_job)(struct drm_sched_job *sched_job);
sched_job         320 include/drm/gpu_scheduler.h void drm_sched_entity_push_job(struct drm_sched_job *sched_job,