/linux/drivers/gpu/drm/amd/amdgpu/ |
A D | amdgpu_fence.c | 162 seq = ++ring->fence_drv.sync_seq; in amdgpu_fence_emit() 169 &ring->fence_drv.lock, in amdgpu_fence_emit() 173 &ring->fence_drv.lock, in amdgpu_fence_emit() 180 ptr = &ring->fence_drv.fences[seq & ring->fence_drv.num_fences_mask]; in amdgpu_fence_emit() 343 ptr = &ring->fence_drv.fences[seq & ring->fence_drv.num_fences_mask]; in amdgpu_fence_wait_empty() 470 ring->fence_drv.gpu_addr = 0; in amdgpu_fence_driver_init_ring() 471 ring->fence_drv.sync_seq = 0; in amdgpu_fence_driver_init_ring() 481 if (!ring->fence_drv.fences) in amdgpu_fence_driver_init_ring() 559 if (ring->fence_drv.irq_src) in amdgpu_fence_driver_hw_fini() 582 kfree(ring->fence_drv.fences); in amdgpu_fence_driver_sw_fini() [all …]
|
A D | mes_v10_1.c | 104 r = amdgpu_fence_wait_polling(ring, ring->fence_drv.sync_seq, in mes_v10_1_submit_pkt_and_poll_completion() 161 mes->ring.fence_drv.gpu_addr; in mes_v10_1_add_hw_queue() 163 ++mes->ring.fence_drv.sync_seq; in mes_v10_1_add_hw_queue() 184 mes->ring.fence_drv.gpu_addr; in mes_v10_1_remove_hw_queue() 186 ++mes->ring.fence_drv.sync_seq; in mes_v10_1_remove_hw_queue() 215 mes->ring.fence_drv.gpu_addr; in mes_v10_1_query_sched_status() 217 ++mes->ring.fence_drv.sync_seq; in mes_v10_1_query_sched_status() 258 mes->ring.fence_drv.gpu_addr; in mes_v10_1_set_hw_resources() 260 ++mes->ring.fence_drv.sync_seq; in mes_v10_1_set_hw_resources()
|
A D | amdgpu_debugfs.c | 1400 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_ib_preempt_fences_swap() 1403 last_seq = atomic_read(&ring->fence_drv.last_seq); in amdgpu_ib_preempt_fences_swap() 1404 sync_seq = ring->fence_drv.sync_seq; in amdgpu_ib_preempt_fences_swap() 1461 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_ib_preempt_mark_partial_job() 1514 length = ring->fence_drv.num_fences_mask + 1; in amdgpu_debugfs_ib_preempt() 1538 if (atomic_read(&ring->fence_drv.last_seq) != in amdgpu_debugfs_ib_preempt() 1539 ring->fence_drv.sync_seq) { in amdgpu_debugfs_ib_preempt()
|
A D | amdgpu_job.c | 60 job->base.sched->name, atomic_read(&ring->fence_drv.last_seq), in amdgpu_job_timedout() 61 ring->fence_drv.sync_seq); in amdgpu_job_timedout()
|
A D | uvd_v6_0.c | 1088 uint32_t seq = ring->fence_drv.sync_seq; in uvd_v6_0_ring_emit_pipeline_sync() 1089 uint64_t addr = ring->fence_drv.gpu_addr; in uvd_v6_0_ring_emit_pipeline_sync() 1117 uint32_t seq = ring->fence_drv.sync_seq; in uvd_v6_0_enc_ring_emit_pipeline_sync() 1118 uint64_t addr = ring->fence_drv.gpu_addr; in uvd_v6_0_enc_ring_emit_pipeline_sync()
|
A D | si_dma.c | 426 uint32_t seq = ring->fence_drv.sync_seq; in si_dma_ring_emit_pipeline_sync() 427 uint64_t addr = ring->fence_drv.gpu_addr; in si_dma_ring_emit_pipeline_sync()
|
A D | vce_v3_0.c | 890 uint32_t seq = ring->fence_drv.sync_seq; in vce_v3_0_emit_pipeline_sync() 891 uint64_t addr = ring->fence_drv.gpu_addr; in vce_v3_0_emit_pipeline_sync()
|
A D | amdgpu_ring.h | 213 struct amdgpu_fence_driver fence_drv; member
|
A D | sdma_v2_4.c | 778 uint32_t seq = ring->fence_drv.sync_seq; in sdma_v2_4_ring_emit_pipeline_sync() 779 uint64_t addr = ring->fence_drv.gpu_addr; in sdma_v2_4_ring_emit_pipeline_sync()
|
A D | cik_sdma.c | 839 uint32_t seq = ring->fence_drv.sync_seq; in cik_sdma_ring_emit_pipeline_sync() 840 uint64_t addr = ring->fence_drv.gpu_addr; in cik_sdma_ring_emit_pipeline_sync()
|
A D | sdma_v3_0.c | 1049 uint32_t seq = ring->fence_drv.sync_seq; in sdma_v3_0_ring_emit_pipeline_sync() 1050 uint64_t addr = ring->fence_drv.gpu_addr; in sdma_v3_0_ring_emit_pipeline_sync()
|
A D | sdma_v5_2.c | 1150 uint32_t seq = ring->fence_drv.sync_seq; in sdma_v5_2_ring_emit_pipeline_sync() 1151 uint64_t addr = ring->fence_drv.gpu_addr; in sdma_v5_2_ring_emit_pipeline_sync()
|
/linux/drivers/gpu/drm/radeon/ |
A D | radeon_fence.c | 120 &rdev->fence_drv[ring].lockup_work, in radeon_fence_schedule_check() 269 struct radeon_fence_driver *fence_drv; in radeon_fence_check_lockup() local 275 rdev = fence_drv->rdev; in radeon_fence_check_lockup() 276 ring = fence_drv - &rdev->fence_drv[0]; in radeon_fence_check_lockup() 287 fence_drv->delayed_irq = false; in radeon_fence_check_lockup() 302 fence_drv->sync_seq[ring], ring); in radeon_fence_check_lockup() 844 rdev->fence_drv[ring].scratch_reg - in radeon_fence_driver_start_ring() 872 rdev->fence_drv[ring].gpu_addr = 0; in radeon_fence_driver_init_ring() 879 rdev->fence_drv[ring].rdev = rdev; in radeon_fence_driver_init_ring() 962 if (!rdev->fence_drv[i].initialized) in radeon_debugfs_fence_info_show() [all …]
|
A D | uvd_v2_2.c | 43 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; in uvd_v2_2_fence_emit()
|
A D | evergreen_dma.c | 44 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in evergreen_dma_fence_ring_emit()
|
A D | uvd_v1_0.c | 85 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; in uvd_v1_0_fence_emit()
|
A D | r600_dma.c | 290 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in r600_dma_fence_ring_emit()
|
A D | radeon_vce.c | 738 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; in radeon_vce_fence_emit()
|
A D | cik_sdma.c | 203 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in cik_sdma_fence_ring_emit()
|
/linux/drivers/gpu/drm/virtio/ |
A D | virtgpu_fence.c | 79 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_alloc() 105 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_emit() 130 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_event_process() 135 atomic64_set(&vgdev->fence_drv.last_fence_id, fence_id); in virtio_gpu_fence_event_process()
|
A D | virtgpu_debugfs.c | 73 (u64)atomic64_read(&vgdev->fence_drv.last_fence_id), in virtio_gpu_debugfs_irq_info() 74 vgdev->fence_drv.current_fence_id); in virtio_gpu_debugfs_irq_info()
|
A D | virtgpu_kms.c | 147 vgdev->fence_drv.context = dma_fence_context_alloc(1); in virtio_gpu_init() 148 spin_lock_init(&vgdev->fence_drv.lock); in virtio_gpu_init() 149 INIT_LIST_HEAD(&vgdev->fence_drv.fences); in virtio_gpu_init()
|
A D | virtgpu_ioctl.c | 135 fence_ctx = vgdev->fence_drv.context; in virtio_gpu_execbuffer_ioctl() 171 if (!dma_fence_match_context(in_fence, vgdev->fence_drv.context)) in virtio_gpu_execbuffer_ioctl() 347 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, 0); in virtio_gpu_resource_create_ioctl() 426 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, 0); in virtio_gpu_transfer_from_host_ioctl() 486 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, in virtio_gpu_transfer_to_host_ioctl()
|
A D | virtgpu_plane.c | 259 vgfb->fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, in virtio_gpu_plane_prepare_fb()
|
A D | virtgpu_drv.h | 238 struct virtio_gpu_fence_driver fence_drv; member
|