/linux/drivers/gpu/drm/ttm/ |
A D | ttm_execbuf_util.c | 61 if (ticket) in ttm_eu_backoff_reservation() 62 ww_acquire_fini(ticket); in ttm_eu_backoff_reservation() 88 if (ticket) in ttm_eu_reserve_buffers() 89 ww_acquire_init(ticket, &reservation_ww_class); in ttm_eu_reserve_buffers() 94 ret = ttm_bo_reserve(bo, intr, (ticket == NULL), ticket); in ttm_eu_reserve_buffers() 120 ret = ttm_bo_reserve_slowpath(bo, intr, ticket); in ttm_eu_reserve_buffers() 128 if (ticket) { in ttm_eu_reserve_buffers() 129 ww_acquire_done(ticket); in ttm_eu_reserve_buffers() 130 ww_acquire_fini(ticket); in ttm_eu_reserve_buffers() 165 if (ticket) in ttm_eu_fence_buffer_objects() [all …]
|
A D | ttm_bo.c | 645 struct ww_acquire_ctx *ticket) in ttm_mem_evict_wait_busy() argument 649 if (!busy_bo || !ticket) in ttm_mem_evict_wait_busy() 654 ticket); in ttm_mem_evict_wait_busy() 656 r = dma_resv_lock(busy_bo->base.resv, ticket); in ttm_mem_evict_wait_busy() 673 struct ww_acquire_ctx *ticket) in ttm_mem_evict_first() argument 687 if (busy && !busy_bo && ticket != in ttm_mem_evict_first() 712 ret = ttm_mem_evict_wait_busy(busy_bo, ctx, ticket); in ttm_mem_evict_first() 785 struct ww_acquire_ctx *ticket; in ttm_bo_mem_force_space() local 789 ticket = dma_resv_locking_ctx(bo->base.resv); in ttm_bo_mem_force_space() 797 ticket); in ttm_bo_mem_force_space()
|
/linux/fs/btrfs/ |
A D | space-info.c | 388 ticket->bytes = 0; in btrfs_try_granting_tickets() 390 wake_up(&ticket->wait); in btrfs_try_granting_tickets() 858 ticket->bytes = 0; in steal_from_global_rsv() 859 wake_up(&ticket->wait); in steal_from_global_rsv() 908 ticket->bytes); in maybe_fail_all_tickets() 912 ticket->error = -EIO; in maybe_fail_all_tickets() 1311 while (ticket->bytes > 0 && ticket->error == 0) { in wait_reserve_ticket() 1382 ret = ticket->error; in handle_reserve_ticket() 1383 if (ticket->bytes || ticket->error) { in handle_reserve_ticket() 1407 ASSERT(!(ticket->bytes == 0 && ticket->error)); in handle_reserve_ticket() [all …]
|
/linux/fs/xfs/ |
A D | xfs_log.c | 51 struct xlog_ticket *ticket, 2170 ticket->t_res_arr_sum, ticket->t_res_o_flow); in xlog_print_tic_res() 2174 ticket->t_res_arr_sum + ticket->t_res_o_flow + ophdr_spc); in xlog_print_tic_res() 2176 ticket->t_res_num); in xlog_print_tic_res() 3100 ticket->t_cnt--; in xfs_log_ticket_regrant() 3106 ticket->t_curr_res = ticket->t_unit_res; in xfs_log_ticket_regrant() 3112 if (!ticket->t_cnt) { in xfs_log_ticket_regrant() 3117 ticket->t_curr_res = ticket->t_unit_res; in xfs_log_ticket_regrant() 3148 ticket->t_cnt--; in xfs_log_ticket_ungrant() 3159 bytes += ticket->t_unit_res*ticket->t_cnt; in xfs_log_ticket_ungrant() [all …]
|
A D | xfs_log_cil.c | 460 if (ctx->ticket->t_curr_res == 0) { in xlog_cil_insert_items() 461 ctx_res = ctx->ticket->t_unit_res; in xlog_cil_insert_items() 462 ctx->ticket->t_curr_res = ctx_res; in xlog_cil_insert_items() 473 ctx->ticket->t_unit_res += split_res; in xlog_cil_insert_items() 474 ctx->ticket->t_curr_res += split_res; in xlog_cil_insert_items() 874 new_ctx->ticket = xlog_cil_ticket_alloc(log); in xlog_cil_push_work() 1019 tic = ctx->ticket; in xlog_cil_push_work() 1106 xfs_log_ticket_put(new_ctx->ticket); in xlog_cil_push_work() 1177 trace_xfs_log_cil_wait(log, cil->xc_ctx->ticket); in xlog_cil_push_background() 1509 if (log->l_cilp->xc_ctx->ticket) in xlog_cil_destroy() [all …]
|
A D | xfs_log.h | 124 struct xlog_ticket **ticket, 131 struct xlog_ticket *xfs_log_ticket_get(struct xlog_ticket *ticket); 132 void xfs_log_ticket_put(struct xlog_ticket *ticket);
|
A D | xfs_log_priv.h | 244 struct xlog_ticket *ticket; /* chkpt ticket */ member 517 void xlog_print_tic_res(struct xfs_mount *mp, struct xlog_ticket *ticket); 522 void xfs_log_ticket_ungrant(struct xlog *log, struct xlog_ticket *ticket); 523 void xfs_log_ticket_regrant(struct xlog *log, struct xlog_ticket *ticket);
|
/linux/net/rxrpc/ |
A D | key.c | 96 memcpy(&token->kad->ticket, &xdr[8], tktlen); in rxrpc_preparse_xdr_rxkad() 110 token->kad->ticket[0], token->kad->ticket[1], in rxrpc_preparse_xdr_rxkad() 111 token->kad->ticket[2], token->kad->ticket[3], in rxrpc_preparse_xdr_rxkad() 112 token->kad->ticket[4], token->kad->ticket[5], in rxrpc_preparse_xdr_rxkad() 113 token->kad->ticket[6], token->kad->ticket[7]); in rxrpc_preparse_xdr_rxkad() 333 v1->ticket[0], v1->ticket[1], in rxrpc_preparse() 334 v1->ticket[2], v1->ticket[3], in rxrpc_preparse() 335 v1->ticket[4], v1->ticket[5], in rxrpc_preparse() 336 v1->ticket[6], v1->ticket[7]); in rxrpc_preparse() 358 memcpy(&token->kad->ticket, v1->ticket, v1->ticket_length); in rxrpc_preparse() [all …]
|
A D | rxkad.c | 825 iov[2].iov_base = (void *)s2->ticket; in rxkad_send_response() 981 void *ticket, size_t ticket_len, in rxkad_decrypt_ticket() argument 1013 sg_init_one(&sg[0], ticket, ticket_len); in rxkad_decrypt_ticket() 1019 p = ticket; in rxkad_decrypt_ticket() 1168 void *ticket; in rxkad_verify_response() local 1229 ticket = kmalloc(ticket_len, GFP_NOFS); in rxkad_verify_response() 1230 if (!ticket) in rxkad_verify_response() 1236 ticket, ticket_len) < 0) in rxkad_verify_response() 1307 kfree(ticket); in rxkad_verify_response() 1315 kfree(ticket); in rxkad_verify_response() [all …]
|
/linux/include/drm/ttm/ |
A D | ttm_bo_driver.h | 134 struct ww_acquire_ctx *ticket) in ttm_bo_reserve() argument 140 if (WARN_ON(ticket)) in ttm_bo_reserve() 148 ret = dma_resv_lock_interruptible(bo->base.resv, ticket); in ttm_bo_reserve() 150 ret = dma_resv_lock(bo->base.resv, ticket); in ttm_bo_reserve() 168 struct ww_acquire_ctx *ticket) in ttm_bo_reserve_slowpath() argument 172 ticket); in ttm_bo_reserve_slowpath() 177 dma_resv_lock_slow(bo->base.resv, ticket); in ttm_bo_reserve_slowpath()
|
A D | ttm_execbuf_util.h | 61 void ttm_eu_backoff_reservation(struct ww_acquire_ctx *ticket, 98 int ttm_eu_reserve_buffers(struct ww_acquire_ctx *ticket, 114 void ttm_eu_fence_buffer_objects(struct ww_acquire_ctx *ticket,
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
A D | amdgpu_csa.c | 68 struct ww_acquire_ctx ticket; in amdgpu_map_static_csa() local 82 r = ttm_eu_reserve_buffers(&ticket, &list, true, NULL); in amdgpu_map_static_csa() 90 ttm_eu_backoff_reservation(&ticket, &list); in amdgpu_map_static_csa() 102 ttm_eu_backoff_reservation(&ticket, &list); in amdgpu_map_static_csa() 106 ttm_eu_backoff_reservation(&ticket, &list); in amdgpu_map_static_csa()
|
A D | amdgpu_vkms.c | 291 struct ww_acquire_ctx ticket; in amdgpu_vkms_prepare_fb() local 309 r = ttm_eu_reserve_buffers(&ticket, &list, false, NULL); in amdgpu_vkms_prepare_fb() 324 ttm_eu_backoff_reservation(&ticket, &list); in amdgpu_vkms_prepare_fb() 331 ttm_eu_backoff_reservation(&ticket, &list); in amdgpu_vkms_prepare_fb() 336 ttm_eu_backoff_reservation(&ticket, &list); in amdgpu_vkms_prepare_fb()
|
A D | amdgpu_dma_buf.c | 379 struct ww_acquire_ctx *ticket = dma_resv_locking_ctx(obj->resv); in amdgpu_dma_buf_move_notify() local 400 if (ticket) { in amdgpu_dma_buf_move_notify() 405 r = dma_resv_lock(resv, ticket); in amdgpu_dma_buf_move_notify()
|
A D | amdgpu_gem.c | 202 struct ww_acquire_ctx ticket; in amdgpu_gem_object_close() local 215 r = ttm_eu_reserve_buffers(&ticket, &list, false, &duplicates); in amdgpu_gem_object_close() 246 ttm_eu_backoff_reservation(&ticket, &list); in amdgpu_gem_object_close() 678 struct ww_acquire_ctx ticket; in amdgpu_gem_va_ioctl() local 750 r = ttm_eu_reserve_buffers(&ticket, &list, true, &duplicates); in amdgpu_gem_va_ioctl() 796 ttm_eu_backoff_reservation(&ticket, &list); in amdgpu_gem_va_ioctl()
|
/linux/arch/ia64/include/asm/ |
A D | spinlock.h | 45 int *p = (int *)&lock->lock, ticket, serve; in __ticket_spin_lock() local 47 ticket = ia64_fetchadd(1, p, acq); in __ticket_spin_lock() 49 if (!(((ticket >> TICKET_SHIFT) ^ ticket) & TICKET_MASK)) in __ticket_spin_lock() 57 if (!(((serve >> TICKET_SHIFT) ^ ticket) & TICKET_MASK)) in __ticket_spin_lock()
|
/linux/drivers/gpu/drm/etnaviv/ |
A D | etnaviv_gem_submit.c | 122 struct ww_acquire_ctx *ticket) in submit_lock_objects() argument 136 ret = dma_resv_lock_interruptible(obj->resv, ticket); in submit_lock_objects() 146 ww_acquire_done(ticket); in submit_lock_objects() 163 ret = dma_resv_lock_slow_interruptible(obj->resv, ticket); in submit_lock_objects() 436 struct ww_acquire_ctx ticket; in etnaviv_ioctl_gem_submit() local 521 ww_acquire_init(&ticket, &reservation_ww_class); in etnaviv_ioctl_gem_submit() 573 ret = submit_lock_objects(submit, &ticket); in etnaviv_ioctl_gem_submit() 609 ww_acquire_fini(&ticket); in etnaviv_ioctl_gem_submit()
|
/linux/drivers/gpu/drm/vmwgfx/ |
A D | vmwgfx_validation.h | 81 struct ww_acquire_ctx ticket; member 172 return ttm_eu_reserve_buffers(&ctx->ticket, &ctx->bo_list, intr, in vmw_validation_bo_reserve() 188 ttm_eu_fence_buffer_objects(&ctx->ticket, &ctx->bo_list, in vmw_validation_bo_fence()
|
A D | vmwgfx_resource.c | 539 vmw_resource_check_buffer(struct ww_acquire_ctx *ticket, in vmw_resource_check_buffer() argument 560 ret = ttm_eu_reserve_buffers(ticket, &val_list, interruptible, NULL); in vmw_resource_check_buffer() 578 ttm_eu_backoff_reservation(ticket, &val_list); in vmw_resource_check_buffer() 630 vmw_resource_backoff_reservation(struct ww_acquire_ctx *ticket, in vmw_resource_backoff_reservation() argument 640 ttm_eu_backoff_reservation(ticket, &val_list); in vmw_resource_backoff_reservation() 653 static int vmw_resource_do_evict(struct ww_acquire_ctx *ticket, in vmw_resource_do_evict() argument 664 ret = vmw_resource_check_buffer(ticket, res, interruptible, &val_buf); in vmw_resource_do_evict() 679 vmw_resource_backoff_reservation(ticket, &val_buf); in vmw_resource_do_evict() 923 struct ww_acquire_ctx ticket; in vmw_resource_evict_type() local 938 ret = vmw_resource_do_evict(&ticket, evict_res, false); in vmw_resource_evict_type()
|
/linux/include/keys/ |
A D | rxrpc-type.h | 31 u8 ticket[]; /* the encrypted ticket */ member 55 u8 ticket[]; member
|
/linux/drivers/gpu/drm/qxl/ |
A D | qxl_release.c | 224 ret = ttm_eu_reserve_buffers(&release->ticket, &release->bos, in qxl_release_reserve_list() 234 ttm_eu_backoff_reservation(&release->ticket, &release->bos); in qxl_release_reserve_list() 248 ttm_eu_backoff_reservation(&release->ticket, &release->bos); in qxl_release_backoff_reserve_list() 436 ww_acquire_fini(&release->ticket); in qxl_release_fence_buffer_objects()
|
/linux/arch/mips/include/asm/octeon/ |
A D | cvmx-cmd-queue.h | 154 uint64_t ticket[(CVMX_CMD_QUEUE_END >> 16) * 256]; member 278 …[ticket_ptr] "=" GCC_OFF_SMALL_ASM()(__cvmx_cmd_queue_state_ptr->ticket[__cvmx_cmd_queue_get_index… in __cvmx_cmd_queue_lock() 279 [now_serving] "=m"(qptr->now_serving), [ticket] "=r"(tmp), in __cvmx_cmd_queue_lock()
|
/linux/tools/tracing/latency/ |
A D | latency-collector.c | 185 int ticket; member 666 req->ticket = r; in printstate_next_ticket() 675 if (req->ticket > printstate.ticket_completed) in printstate_mark_req_completed() 676 printstate.ticket_completed = req->ticket; in printstate_mark_req_completed() 682 return (printstate.ticket_counter != req->ticket); in printstate_has_new_req_arrived() 764 int diff = req->ticket - req->ticket_completed_ref; in table_get_probability() 1095 sec, us, req->ticket, str); in __print_skipmessage() 1098 sec, us, req->ticket); in __print_skipmessage() 1166 sec, us, req->ticket, slept_ms); in print_tracefile() 1170 us, req->ticket); in print_tracefile()
|
/linux/arch/ia64/mm/ |
A D | tlb.c | 112 unsigned long ticket; member 118 ss->ticket = 0; in spinaphore_init() 124 unsigned long t = ia64_fetchadd(1, &ss->ticket, acq), serve; in down_spin()
|
/linux/drivers/gpu/drm/virtio/ |
A D | virtgpu_gem.c | 223 &objs->ticket); in virtio_gpu_array_lock_resv() 234 &objs->ticket); in virtio_gpu_array_unlock_resv()
|