Home
last modified time | relevance | path

Searched refs:s_fence (Results 1 – 19 of 19) sorted by relevance

/linux/drivers/gpu/drm/scheduler/
A Dsched_main.c174 struct drm_sched_fence *s_fence = s_job->s_fence; in drm_sched_job_done() local
182 dma_fence_get(&s_fence->finished); in drm_sched_job_done()
183 drm_sched_fence_finished(s_fence); in drm_sched_job_done()
212 struct drm_sched_fence *s_fence; in drm_sched_dependency_optimized() local
219 if (s_fence && s_fence->sched == sched) in drm_sched_dependency_optimized()
419 if (s_job->s_fence->parent && in drm_sched_stop()
538 struct drm_sched_fence *s_fence = s_job->s_fence; in drm_sched_resubmit_jobs_ext() local
559 s_job->s_fence->parent = NULL; in drm_sched_resubmit_jobs_ext()
596 if (!job->s_fence) in drm_sched_job_init()
746 job->s_fence = NULL; in drm_sched_job_cleanup()
[all …]
A Dsched_entity.c200 drm_sched_fence_finished(job->s_fence); in drm_sched_entity_kill_jobs_cb()
201 WARN_ON(job->s_fence->parent); in drm_sched_entity_kill_jobs_cb()
225 struct drm_sched_fence *s_fence = job->s_fence; in drm_sched_entity_kill_jobs() local
231 drm_sched_fence_scheduled(s_fence); in drm_sched_entity_kill_jobs()
232 dma_fence_set_error(&s_fence->finished, -ESRCH); in drm_sched_entity_kill_jobs()
365 struct drm_sched_fence *s_fence; in drm_sched_entity_add_dependency_cb() local
378 s_fence = to_drm_sched_fence(fence); in drm_sched_entity_add_dependency_cb()
379 if (s_fence && s_fence->sched == sched) { in drm_sched_entity_add_dependency_cb()
385 fence = dma_fence_get(&s_fence->scheduled); in drm_sched_entity_add_dependency_cb()
423 dma_fence_set_error(&sched_job->s_fence->finished, -ECANCELED); in drm_sched_entity_pop_job()
[all …]
A Dgpu_scheduler_trace.h50 __entry->fence = &sched_job->s_fence->finished;
77 __entry->fence = &sched_job->s_fence->finished;
/linux/drivers/gpu/drm/amd/amdgpu/
A Damdgpu_job.c139 f = job->base.s_fence ? &job->base.s_fence->finished : hw_fence; in amdgpu_job_free_resources()
187 *f = dma_fence_get(&job->base.s_fence->finished); in amdgpu_job_submit()
230 &job->base.s_fence->finished, in amdgpu_job_dependency()
249 finished = &job->base.s_fence->finished; in amdgpu_job_run()
297 struct drm_sched_fence *s_fence = s_job->s_fence; in amdgpu_job_stop_all_jobs_on_sched() local
299 dma_fence_signal(&s_fence->scheduled); in amdgpu_job_stop_all_jobs_on_sched()
300 dma_fence_set_error(&s_fence->finished, -EHWPOISON); in amdgpu_job_stop_all_jobs_on_sched()
301 dma_fence_signal(&s_fence->finished); in amdgpu_job_stop_all_jobs_on_sched()
309 struct drm_sched_fence *s_fence = s_job->s_fence; in amdgpu_job_stop_all_jobs_on_sched() local
311 dma_fence_set_error(&s_fence->finished, -EHWPOISON); in amdgpu_job_stop_all_jobs_on_sched()
[all …]
A Damdgpu_sync.c68 struct drm_sched_fence *s_fence = to_drm_sched_fence(f); in amdgpu_sync_same_dev() local
70 if (s_fence) { in amdgpu_sync_same_dev()
89 struct drm_sched_fence *s_fence; in amdgpu_sync_get_owner() local
95 s_fence = to_drm_sched_fence(f); in amdgpu_sync_get_owner()
96 if (s_fence) in amdgpu_sync_get_owner()
97 return s_fence->owner; in amdgpu_sync_get_owner()
313 struct drm_sched_fence *s_fence = to_drm_sched_fence(f); in amdgpu_sync_peek_fence() local
321 if (ring && s_fence) { in amdgpu_sync_peek_fence()
325 if (s_fence->sched == &ring->sched) { in amdgpu_sync_peek_fence()
326 if (dma_fence_is_signaled(&s_fence->scheduled)) in amdgpu_sync_peek_fence()
[all …]
A Damdgpu_ctx.c752 struct drm_sched_fence *s_fence; in amdgpu_ctx_fence_time() local
759 s_fence = to_drm_sched_fence(fence); in amdgpu_ctx_fence_time()
760 if (!dma_fence_is_signaled(&s_fence->scheduled)) { in amdgpu_ctx_fence_time()
764 t1 = s_fence->scheduled.timestamp; in amdgpu_ctx_fence_time()
769 if (dma_fence_is_signaled(&s_fence->finished) && in amdgpu_ctx_fence_time()
770 s_fence->finished.timestamp < now) in amdgpu_ctx_fence_time()
771 *total += ktime_sub(s_fence->finished.timestamp, t1); in amdgpu_ctx_fence_time()
A Damdgpu_trace.h36 job->base.s_fence->finished.ops->get_timeline_name(&job->base.s_fence->finished)
180 __entry->context = job->base.s_fence->finished.context;
181 __entry->seqno = job->base.s_fence->finished.seqno;
205 __entry->context = job->base.s_fence->finished.context;
206 __entry->seqno = job->base.s_fence->finished.seqno;
A Damdgpu_ib.c151 fence_ctx = job->base.s_fence ? in amdgpu_ib_schedule()
152 job->base.s_fence->scheduled.context : 0; in amdgpu_ib_schedule()
A Damdgpu_cs.c989 struct drm_sched_fence *s_fence; in amdgpu_cs_process_fence_dep() local
992 s_fence = to_drm_sched_fence(fence); in amdgpu_cs_process_fence_dep()
993 fence = dma_fence_get(&s_fence->scheduled); in amdgpu_cs_process_fence_dep()
1246 p->fence = dma_fence_get(&job->base.s_fence->finished); in amdgpu_cs_submit()
A Damdgpu_device.c4863 dma_fence_get(s_job->s_fence->parent); in amdgpu_device_recheck_guilty_jobs()
4866 ret = dma_fence_wait_timeout(s_job->s_fence->parent, false, ring->sched.timeout); in amdgpu_device_recheck_guilty_jobs()
4899 dma_fence_put(s_job->s_fence->parent); in amdgpu_device_recheck_guilty_jobs()
4900 dma_fence_get(&s_job->s_fence->finished); in amdgpu_device_recheck_guilty_jobs()
4901 dma_fence_signal(&s_job->s_fence->finished); in amdgpu_device_recheck_guilty_jobs()
4902 dma_fence_put(&s_job->s_fence->finished); in amdgpu_device_recheck_guilty_jobs()
5074 if (job && job->base.s_fence->parent && in amdgpu_device_gpu_recover()
5075 dma_fence_is_signaled(job->base.s_fence->parent)) { in amdgpu_device_gpu_recover()
A Damdgpu_debugfs.c1481 if (dma_fence_is_signaled(&s_job->s_fence->finished)) { in amdgpu_ib_preempt_mark_partial_job()
/linux/drivers/gpu/drm/lima/
A Dlima_trace.h25 __entry->context = task->base.s_fence->finished.context;
26 __entry->seqno = task->base.s_fence->finished.seqno;
A Dlima_sched.c173 struct dma_fence *fence = dma_fence_get(&task->base.s_fence->finished); in lima_sched_context_queue_task()
211 if (job->s_fence->finished.error < 0) in lima_sched_run_job()
/linux/drivers/gpu/drm/etnaviv/
A Detnaviv_sched.c77 if (likely(!sched_job->s_fence->finished.error)) in etnaviv_sched_run_job()
168 submit->out_fence = dma_fence_get(&submit->sched_job.s_fence->finished); in etnaviv_sched_push_job()
/linux/drivers/gpu/drm/v3d/
A Dv3d_sched.c83 if (unlikely(job->base.base.s_fence->finished.error)) in v3d_bin_job_run()
137 if (unlikely(job->base.base.s_fence->finished.error)) in v3d_render_job_run()
A Dv3d_gem.c526 job->done_fence = dma_fence_get(&job->base.s_fence->finished); in v3d_push_job()
/linux/include/drm/
A Dgpu_scheduler.h288 struct drm_sched_fence *s_fence; member
/linux/drivers/gpu/drm/panfrost/
A Dpanfrost_job.c284 job->render_done_fence = dma_fence_get(&job->base.s_fence->finished); in panfrost_job_push()
359 if (unlikely(job->base.s_fence->finished.error)) in panfrost_job_run()
/linux/drivers/gpu/drm/msm/
A Dmsm_gem_submit.c877 submit->user_fence = dma_fence_get(&submit->base.s_fence->finished); in msm_ioctl_gem_submit()

Completed in 35 milliseconds