Home
last modified time | relevance | path

Searched refs:io_cq (Results 1 – 18 of 18) sorted by relevance

/linux/drivers/net/ethernet/amazon/ena/
A Dena_eth_com.h176 if (unlikely(io_cq->cq_head_db_reg)) { in ena_com_update_dev_comp_head()
177 head = io_cq->head; in ena_com_update_dev_comp_head()
184 io_cq->qid, head); in ena_com_update_dev_comp_head()
186 io_cq->last_head_update = head; in ena_com_update_dev_comp_head()
198 if (!io_cq->numa_node_cfg_reg) in ena_com_update_numa_node()
214 io_cq->head++; in ena_com_cq_inc_head()
217 if (unlikely((io_cq->head & (io_cq->q_depth - 1)) == 0)) in ena_com_cq_inc_head()
218 io_cq->phase ^= 1; in ena_com_cq_inc_head()
228 masked_head = io_cq->head & (io_cq->q_depth - 1); in ena_com_tx_comp_req_id_get()
229 expected_phase = io_cq->phase; in ena_com_tx_comp_req_id_get()
[all …]
A Dena_eth_com.c9 struct ena_com_io_cq *io_cq) in ena_com_get_next_rx_cdesc() argument
15 head_masked = io_cq->head & (io_cq->q_depth - 1); in ena_com_get_next_rx_cdesc()
16 expected_phase = io_cq->phase; in ena_com_get_next_rx_cdesc()
230 idx &= (io_cq->q_depth - 1); in ena_com_rx_cdesc_idx_to_ptr()
233 idx * io_cq->cdesc_entry_size_in_bytes); in ena_com_rx_cdesc_idx_to_ptr()
248 ena_com_cq_inc_head(io_cq); in ena_com_cdesc_rx_pkt_get()
257 count += io_cq->cur_rx_pkt_cdesc_count; in ena_com_cdesc_rx_pkt_get()
259 head_masked = io_cq->head & (io_cq->q_depth - 1); in ena_com_cdesc_rx_pkt_get()
261 io_cq->cur_rx_pkt_cdesc_count = 0; in ena_com_cdesc_rx_pkt_get()
268 io_cq->cur_rx_pkt_cdesc_count += count; in ena_com_cdesc_rx_pkt_get()
[all …]
A Dena_com.c413 memset(&io_cq->cdesc_addr, 0x0, sizeof(io_cq->cdesc_addr)); in ena_com_init_io_cq()
421 size = io_cq->cdesc_entry_size_in_bytes * io_cq->q_depth; in ena_com_init_io_cq()
425 io_cq->cdesc_addr.virt_addr = in ena_com_init_io_cq()
441 io_cq->phase = 1; in ena_com_init_io_cq()
442 io_cq->head = 0; in ena_com_init_io_cq()
947 size = io_cq->cdesc_entry_size_in_bytes * io_cq->q_depth; in ena_com_io_queue_free()
1431 io_cq->cq_head_db_reg = in ena_com_create_io_cq()
1436 io_cq->numa_node_cfg_reg = in ena_com_create_io_cq()
1441 io_cq->idx, io_cq->q_depth); in ena_com_create_io_cq()
1888 memset(io_cq, 0x0, sizeof(*io_cq)); in ena_com_create_io_queue()
[all …]
A Dena_com.h453 struct ena_com_io_cq **io_cq);
843 struct ena_com_io_cq *io_cq);
854 struct ena_com_io_cq *io_cq);
950 static inline struct ena_com_dev *ena_com_io_cq_to_ena_dev(struct ena_com_io_cq *io_cq) in ena_com_io_cq_to_ena_dev() argument
952 return container_of(io_cq, struct ena_com_dev, io_cq_queues[io_cq->qid]); in ena_com_io_cq_to_ena_dev()
/linux/block/
A Dblk-ioc.c34 struct io_cq *icq = container_of(head, struct io_cq, __rcu_head); in icq_free_icq_rcu()
43 static void ioc_exit_icq(struct io_cq *icq) in ioc_exit_icq()
60 static void ioc_destroy_icq(struct io_cq *icq) in ioc_destroy_icq()
102 struct io_cq *icq = hlist_entry(ioc->icq_list.first, in ioc_release_fn()
103 struct io_cq, ioc_node); in ioc_release_fn()
179 struct io_cq *icq; in put_io_context_active()
218 struct io_cq *icq = list_entry(icq_list->next, in __ioc_clear_queue()
219 struct io_cq, q_node); in __ioc_clear_queue()
332 struct io_cq *ioc_lookup_icq(struct io_context *ioc, struct request_queue *q) in ioc_lookup_icq()
334 struct io_cq *icq; in ioc_lookup_icq()
[all …]
A Delevator.h8 struct io_cq;
47 void (*init_icq)(struct io_cq *);
48 void (*exit_icq)(struct io_cq *);
A Dblk.h406 struct io_cq *ioc_lookup_icq(struct io_context *ioc, struct request_queue *q);
407 struct io_cq *ioc_create_icq(struct io_context *ioc, struct request_queue *q,
A Delevator.c539 if (WARN_ON(e->icq_size < sizeof(struct io_cq)) || in elv_register()
540 WARN_ON(e->icq_align < __alignof__(struct io_cq))) in elv_register()
A Dbfq-iosched.h407 struct io_cq icq; /* must be the first member */
A Dblk-mq-sched.c25 struct io_cq *icq; in blk_mq_sched_assign_ioc()
A Dbfq-iosched.c428 static struct bfq_io_cq *icq_to_bic(struct io_cq *icq) in icq_to_bic()
5228 static void bfq_exit_icq(struct io_cq *icq) in bfq_exit_icq()
/linux/include/linux/
A Diocontext.h73 struct io_cq { struct
110 struct io_cq __rcu *icq_hint; argument
A Dblk-mq.h180 struct io_cq *icq;
/linux/drivers/scsi/lpfc/
A Dlpfc_debugfs.h471 cq = phba->sli4_hba.hdwq[wqidx].io_cq; in lpfc_debug_dump_cq()
648 if (phba->sli4_hba.hdwq[cq_idx].io_cq->queue_id == qid) in lpfc_debug_dump_cq_by_id()
653 lpfc_debug_dump_q(phba->sli4_hba.hdwq[cq_idx].io_cq); in lpfc_debug_dump_cq_by_id()
A Dlpfc_sli4.h698 struct lpfc_queue *io_cq; /* Fast-path FCP & NVME compl queue */ member
A Dlpfc_init.c1298 cq = hdwq->io_cq; in lpfc_idle_stat_delay_work()
10263 phba->sli4_hba.hdwq[idx].io_cq = qdesc; in lpfc_alloc_io_wq_cq()
10710 lpfc_sli4_queue_free(hdwq[idx].io_cq); in lpfc_sli4_release_hdwq()
10713 hdwq[idx].io_cq = NULL; in lpfc_sli4_release_hdwq()
11050 qp[qidx].io_cq, in lpfc_sli4_queue_setup()
11359 lpfc_cq_destroy(phba, qp->io_cq); in lpfc_sli4_queue_unset()
A Dlpfc_debugfs.c3957 qp = phba->sli4_hba.hdwq[eqidx].io_cq; in lpfc_idiag_cqs_for_eq()
4421 qp = phba->sli4_hba.hdwq[qidx].io_cq; in lpfc_idiag_queacc_write()
A Dlpfc_sli.c6082 sli4_hba->sli4_write_cq_db(phba, qp->io_cq, 0, in lpfc_sli4_arm_cqeq_intr()
7840 cq = hdwq->io_cq; in lpfc_init_idle_stat_hb()

Completed in 100 milliseconds