Lines Matching refs:user_data

858 	u64				user_data;  member
1111 static bool io_cqring_fill_event(struct io_ring_ctx *ctx, u64 user_data,
1602 io_cqring_fill_event(req->ctx, req->user_data, status, 0); in io_kill_timeout()
1833 static bool io_cqring_event_overflow(struct io_ring_ctx *ctx, u64 user_data, in io_cqring_event_overflow() argument
1854 ocqe->cqe.user_data = user_data; in io_cqring_event_overflow()
1861 static inline bool __io_cqring_fill_event(struct io_ring_ctx *ctx, u64 user_data, in __io_cqring_fill_event() argument
1866 trace_io_uring_complete(ctx, user_data, res, cflags); in __io_cqring_fill_event()
1875 WRITE_ONCE(cqe->user_data, user_data); in __io_cqring_fill_event()
1880 return io_cqring_event_overflow(ctx, user_data, res, cflags); in __io_cqring_fill_event()
1884 static noinline bool io_cqring_fill_event(struct io_ring_ctx *ctx, u64 user_data, in io_cqring_fill_event() argument
1887 return __io_cqring_fill_event(ctx, user_data, res, cflags); in io_cqring_fill_event()
1896 __io_cqring_fill_event(ctx, req->user_data, res, cflags); in io_req_complete_post()
2104 io_cqring_fill_event(link->ctx, link->user_data, in io_kill_linked_timeout()
2129 io_cqring_fill_event(link->ctx, link->user_data, res, 0); in io_fail_links()
2146 io_cqring_fill_event(link->ctx, link->user_data, in io_disarm_next()
2411 __io_cqring_fill_event(ctx, req->user_data, req->result, in __io_submit_flush_completions()
2546 __io_cqring_fill_event(ctx, req->user_data, req->result, in io_do_iopoll()
5313 trace_io_uring_task_add(req->ctx, req->opcode, req->user_data, mask); in __io_async_wake()
5403 if (!io_cqring_fill_event(ctx, req->user_data, error, flags)) { in __io_poll_complete()
5563 trace_io_uring_task_run(req->ctx, req, req->opcode, req->user_data); in io_async_task_func()
5587 trace_io_uring_poll_wake(req->ctx, req->opcode, req->user_data, in io_async_wake()
5598 list = &ctx->cancel_hash[hash_long(req->user_data, ctx->cancel_hash_bits)]; in io_poll_req_insert()
5693 trace_io_uring_poll_arm(ctx, req, req->opcode, req->user_data, in io_arm_poll_handler()
5727 io_cqring_fill_event(req->ctx, req->user_data, -ECANCELED, 0); in io_poll_remove_one()
5772 if (sqe_addr != req->user_data) in io_poll_find()
5952 preq->user_data = req->poll_update.new_user_data; in io_poll_update()
5997 __u64 user_data) in io_timeout_extract() argument
6005 found = user_data == req->user_data; in io_timeout_extract()
6019 static int io_timeout_cancel(struct io_ring_ctx *ctx, __u64 user_data) in io_timeout_cancel() argument
6023 struct io_kiocb *req = io_timeout_extract(ctx, user_data); in io_timeout_cancel()
6029 io_cqring_fill_event(ctx, req->user_data, -ECANCELED, 0); in io_timeout_cancel()
6050 static int io_linked_timeout_update(struct io_ring_ctx *ctx, __u64 user_data, in io_linked_timeout_update() argument
6059 found = user_data == req->user_data; in io_linked_timeout_update()
6075 static int io_timeout_update(struct io_ring_ctx *ctx, __u64 user_data, in io_timeout_update() argument
6079 struct io_kiocb *req = io_timeout_extract(ctx, user_data); in io_timeout_update()
6274 u64 user_data; member
6282 return req->ctx == cd->ctx && req->user_data == cd->user_data; in io_cancel_cb()
6285 static int io_async_cancel_one(struct io_uring_task *tctx, u64 user_data, in io_async_cancel_one() argument
6288 struct io_cancel_data data = { .ctx = ctx, .user_data = user_data, }; in io_async_cancel_one()
6574 trace_io_uring_defer(ctx, req, req->user_data); in io_drain_req()
6934 ret = io_try_cancel_userdata(req, prev->user_data); in io_req_task_link_timeout()
7123 req->user_data = READ_ONCE(sqe->user_data); in io_init_req()
7230 trace_io_uring_submit_sqe(ctx, req, req->opcode, req->user_data, in io_submit_sqe()
10187 sqe->user_data); in __io_uring_show_fdinfo()
10196 entry & cq_mask, cqe->user_data, cqe->res, in __io_uring_show_fdinfo()
10259 cqe->user_data, cqe->res, cqe->flags); in __io_uring_show_fdinfo()
11137 BUILD_BUG_SQE_ELEM(32, __u64, user_data); in io_uring_init()