/linux/drivers/gpu/drm/v3d/ |
A D | v3d_sched.c | 28 to_v3d_job(struct drm_sched_job *sched_job) in to_v3d_job() argument 34 to_bin_job(struct drm_sched_job *sched_job) in to_bin_job() argument 40 to_render_job(struct drm_sched_job *sched_job) in to_render_job() argument 46 to_tfu_job(struct drm_sched_job *sched_job) in to_tfu_job() argument 52 to_csd_job(struct drm_sched_job *sched_job) in to_csd_job() argument 60 struct v3d_job *job = to_v3d_job(sched_job); in v3d_sched_job_free() 247 struct v3d_job *job = to_v3d_job(sched_job); in v3d_cache_clean_job_run() 266 if (sched_job) in v3d_gpu_reset_for_timeout() 267 drm_sched_increase_karma(sched_job); in v3d_gpu_reset_for_timeout() 294 struct v3d_job *job = to_v3d_job(sched_job); in v3d_cl_job_timedout() [all …]
|
/linux/drivers/gpu/drm/etnaviv/ |
A D | etnaviv_sched.c | 21 etnaviv_sched_dependency(struct drm_sched_job *sched_job, in etnaviv_sched_dependency() argument 77 if (likely(!sched_job->s_fence->finished.error)) in etnaviv_sched_run_job() 86 *sched_job) in etnaviv_sched_timedout_job() 94 drm_sched_stop(&gpu->sched, sched_job); in etnaviv_sched_timedout_job() 115 if(sched_job) in etnaviv_sched_timedout_job() 116 drm_sched_increase_karma(sched_job); in etnaviv_sched_timedout_job() 137 drm_sched_job_cleanup(sched_job); in etnaviv_sched_free_job() 161 ret = drm_sched_job_init(&submit->sched_job, sched_entity, in etnaviv_sched_push_job() 166 drm_sched_job_arm(&submit->sched_job); in etnaviv_sched_push_job() 173 drm_sched_job_cleanup(&submit->sched_job); in etnaviv_sched_push_job() [all …]
|
A D | etnaviv_sched.h | 14 struct etnaviv_gem_submit *to_etnaviv_submit(struct drm_sched_job *sched_job) in to_etnaviv_submit() argument 16 return container_of(sched_job, struct etnaviv_gem_submit, sched_job); in to_etnaviv_submit()
|
A D | etnaviv_gem.h | 93 struct drm_sched_job sched_job; member
|
/linux/drivers/gpu/drm/scheduler/ |
A D | gpu_scheduler_trace.h | 37 TP_ARGS(sched_job, entity), 49 __entry->id = sched_job->id; 51 __entry->name = sched_job->sched->name; 54 &sched_job->sched->hw_rq_count); 64 TP_ARGS(sched_job, entity), 76 __entry->id = sched_job->id; 78 __entry->name = sched_job->sched->name; 81 &sched_job->sched->hw_rq_count); 104 TP_ARGS(sched_job, fence), 114 __entry->name = sched_job->sched->name; [all …]
|
A D | sched_entity.c | 33 #define to_drm_sched_job(sched_job) \ argument 34 container_of((sched_job), struct drm_sched_job, queue_node) 407 struct drm_sched_job *sched_job; in drm_sched_entity_pop_job() local 410 if (!sched_job) in drm_sched_entity_pop_job() 414 drm_sched_job_dependency(sched_job, entity))) { in drm_sched_entity_pop_job() 415 trace_drm_sched_job_wait_dep(sched_job, entity->dependency); in drm_sched_entity_pop_job() 423 dma_fence_set_error(&sched_job->s_fence->finished, -ECANCELED); in drm_sched_entity_pop_job() 437 return sched_job; in drm_sched_entity_pop_job() 492 void drm_sched_entity_push_job(struct drm_sched_job *sched_job) in drm_sched_entity_push_job() argument 494 struct drm_sched_entity *entity = sched_job->entity; in drm_sched_entity_push_job() [all …]
|
A D | sched_main.c | 62 #define to_drm_sched_job(sched_job) \ argument 920 struct drm_sched_job *sched_job; in drm_sched_main() local 936 sched_job = drm_sched_entity_pop_job(entity); in drm_sched_main() 938 if (!sched_job) { in drm_sched_main() 943 s_fence = sched_job->s_fence; in drm_sched_main() 946 drm_sched_job_begin(sched_job); in drm_sched_main() 948 trace_drm_run_job(sched_job, entity); in drm_sched_main() 949 fence = sched->ops->run_job(sched_job); in drm_sched_main() 955 r = dma_fence_add_callback(fence, &sched_job->cb, in drm_sched_main() 958 drm_sched_job_done(sched_job); in drm_sched_main() [all …]
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
A D | amdgpu_job.c | 212 static struct dma_fence *amdgpu_job_dependency(struct drm_sched_job *sched_job, in amdgpu_job_dependency() argument 216 struct amdgpu_job *job = to_amdgpu_job(sched_job); in amdgpu_job_dependency() 241 static struct dma_fence *amdgpu_job_run(struct drm_sched_job *sched_job) in amdgpu_job_run() argument 243 struct amdgpu_ring *ring = to_amdgpu_ring(sched_job->sched); in amdgpu_job_run() 248 job = to_amdgpu_job(sched_job); in amdgpu_job_run() 278 #define to_drm_sched_job(sched_job) \ argument 279 container_of((sched_job), struct drm_sched_job, queue_node)
|
A D | amdgpu_job.h | 35 #define to_amdgpu_job(sched_job) \ argument 36 container_of((sched_job), struct amdgpu_job, base)
|
A D | amdgpu_trace.h | 517 TP_PROTO(struct amdgpu_job *sched_job, struct dma_fence *fence), 518 TP_ARGS(sched_job, fence), 520 __string(ring, sched_job->base.sched->name) 528 __assign_str(ring, sched_job->base.sched->name); 529 __entry->id = sched_job->base.id;
|
/linux/include/drm/ |
A D | gpu_scheduler.h | 338 struct dma_fence *(*dependency)(struct drm_sched_job *sched_job, 347 struct dma_fence *(*run_job)(struct drm_sched_job *sched_job); 391 enum drm_gpu_sched_stat (*timedout_job)(struct drm_sched_job *sched_job); 397 void (*free_job)(struct drm_sched_job *sched_job); 503 void drm_sched_entity_push_job(struct drm_sched_job *sched_job);
|
/linux/drivers/gpu/drm/panfrost/ |
A D | panfrost_job.c | 42 to_panfrost_job(struct drm_sched_job *sched_job) in to_panfrost_job() argument 44 return container_of(sched_job, struct panfrost_job, base); in to_panfrost_job() 343 static void panfrost_job_free(struct drm_sched_job *sched_job) in panfrost_job_free() argument 345 struct panfrost_job *job = to_panfrost_job(sched_job); in panfrost_job_free() 347 drm_sched_job_cleanup(sched_job); in panfrost_job_free() 352 static struct dma_fence *panfrost_job_run(struct drm_sched_job *sched_job) in panfrost_job_run() argument 354 struct panfrost_job *job = to_panfrost_job(sched_job); in panfrost_job_run() 705 *sched_job) in panfrost_job_timedout() 707 struct panfrost_job *job = to_panfrost_job(sched_job); in panfrost_job_timedout() 724 sched_job); in panfrost_job_timedout() [all …]
|