/linux/drivers/net/ethernet/mellanox/mlx4/ |
A D | cq.c | 173 struct mlx4_cq_context *cq_context; in mlx4_cq_modify() local 180 cq_context = mailbox->buf; in mlx4_cq_modify() 181 cq_context->cq_max_count = cpu_to_be16(count); in mlx4_cq_modify() 195 struct mlx4_cq_context *cq_context; in mlx4_cq_resize() local 203 cq_context = mailbox->buf; in mlx4_cq_resize() 207 cq_context->mtt_base_addr_h = mtt_addr >> 32; in mlx4_cq_resize() 351 struct mlx4_cq_context *cq_context; in mlx4_cq_alloc() local 376 cq_context = mailbox->buf; in mlx4_cq_alloc() 379 cq_context->flags |= cpu_to_be32(1 << 19); in mlx4_cq_alloc() 381 cq_context->logsize_usrpage = in mlx4_cq_alloc() [all …]
|
/linux/drivers/infiniband/hw/hns/ |
A D | hns_roce_hw_v2_dfx.c | 12 struct hns_roce_v2_cq_context *cq_context; in hns_roce_v2_query_cqc_info() local 20 cq_context = mailbox->buf; in hns_roce_v2_query_cqc_info() 29 memcpy(buffer, cq_context, sizeof(*cq_context)); in hns_roce_v2_query_cqc_info()
|
A D | hns_roce_hw_v1.c | 2077 memset(cq_context, 0, sizeof(*cq_context)); in hns_roce_v1_write_cqc() 2085 roce_set_field(cq_context->cqc_byte_4, in hns_roce_v1_write_cqc() 2093 roce_set_field(cq_context->cqc_byte_12, in hns_roce_v1_write_cqc() 2097 roce_set_field(cq_context->cqc_byte_12, in hns_roce_v1_write_cqc() 2106 roce_set_field(cq_context->cqc_byte_20, in hns_roce_v1_write_cqc() 2110 roce_set_field(cq_context->cqc_byte_20, in hns_roce_v1_write_cqc() 2118 roce_set_field(cq_context->cqc_byte_20, in hns_roce_v1_write_cqc() 2128 roce_set_bit(cq_context->cqc_byte_32, in hns_roce_v1_write_cqc() 2130 roce_set_bit(cq_context->cqc_byte_32, in hns_roce_v1_write_cqc() 2132 roce_set_bit(cq_context->cqc_byte_32, in hns_roce_v1_write_cqc() [all …]
|
A D | hns_roce_hw_v2.c | 3331 struct hns_roce_v2_cq_context *cq_context; in hns_roce_v2_write_cqc() local 3333 cq_context = mb_buf; in hns_roce_v2_write_cqc() 3334 memset(cq_context, 0, sizeof(*cq_context)); in hns_roce_v2_write_cqc() 3340 hr_reg_write(cq_context, CQC_CQN, hr_cq->cqn); in hns_roce_v2_write_cqc() 3346 hr_reg_enable(cq_context, CQC_STASH); in hns_roce_v2_write_cqc() 3358 hr_reg_write(cq_context, CQC_CQE_BAR_PG_SZ, in hns_roce_v2_write_cqc() 3360 hr_reg_write(cq_context, CQC_CQE_BUF_PG_SZ, in hns_roce_v2_write_cqc() 3370 hr_reg_write(cq_context, CQC_CQ_MAX_CNT, in hns_roce_v2_write_cqc() 3372 hr_reg_write(cq_context, CQC_CQ_PERIOD, in hns_roce_v2_write_cqc() 5516 struct hns_roce_v2_cq_context *cq_context; in hns_roce_v2_modify_cq() local [all …]
|
A D | hns_roce_cq.c | 471 ibcq->comp_handler(ibcq, ibcq->cq_context); in hns_roce_cq_completion() 503 ibcq->event_handler(&event, ibcq->cq_context); in hns_roce_cq_event()
|
/linux/drivers/infiniband/hw/mthca/ |
A D | mthca_cq.c | 230 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mthca_cq_completion() 256 cq->ibcq.event_handler(&event, cq->ibcq.cq_context); in mthca_cq_event() 773 struct mthca_cq_context *cq_context; in mthca_init_cq() local 811 cq_context = mailbox->buf; in mthca_init_cq() 824 memset(cq_context, 0, sizeof *cq_context); in mthca_init_cq() 830 cq_context->logsize_usrpage |= cpu_to_be32(ctx->uar.index); in mthca_init_cq() 835 cq_context->pd = cpu_to_be32(pdn); in mthca_init_cq() 836 cq_context->lkey = cpu_to_be32(cq->buf.mr.ibmr.lkey); in mthca_init_cq() 837 cq_context->cqn = cpu_to_be32(cq->cqn); in mthca_init_cq() 840 cq_context->ci_db = cpu_to_be32(cq->set_ci_db_index); in mthca_init_cq() [all …]
|
/linux/net/smc/ |
A D | smc_wr.h | 127 void smc_wr_tx_cq_handler(struct ib_cq *ib_cq, void *cq_context); 136 void smc_wr_rx_cq_handler(struct ib_cq *ib_cq, void *cq_context);
|
A D | smc_wr.c | 166 void smc_wr_tx_cq_handler(struct ib_cq *ib_cq, void *cq_context) in smc_wr_tx_cq_handler() argument 168 struct smc_ib_device *dev = (struct smc_ib_device *)cq_context; in smc_wr_tx_cq_handler() 506 void smc_wr_rx_cq_handler(struct ib_cq *ib_cq, void *cq_context) in smc_wr_rx_cq_handler() argument 508 struct smc_ib_device *dev = (struct smc_ib_device *)cq_context; in smc_wr_rx_cq_handler()
|
/linux/drivers/infiniband/sw/rxe/ |
A D | rxe_cq.c | 54 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in rxe_send_complete() 122 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rxe_cq_post()
|
/linux/drivers/infiniband/core/ |
A D | uverbs_std_types_cq.c | 43 struct ib_uverbs_event_queue *ev_queue = cq->cq_context; in uverbs_free_cq() 125 cq->cq_context = ev_file ? &ev_file->ev_queue : NULL; in UVERBS_HANDLER()
|
A D | uverbs.h | 236 void ib_uverbs_comp_handler(struct ib_cq *cq, void *cq_context);
|
A D | cq.c | 229 cq->cq_context = private; in __ib_alloc_cq()
|
/linux/net/sunrpc/xprtrdma/ |
A D | frwr_ops.c | 374 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_fastreg() 470 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv() 490 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv_wake() 593 rpcrdma_flush_disconnect(cq->cq_context, wc); in frwr_wc_localinv_done()
|
/linux/drivers/infiniband/hw/cxgb4/ |
A D | ev.c | 115 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event() 234 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in c4iw_ev_handler()
|
A D | qp.c | 1012 schp->ibcq.cq_context); in complete_sq_drain_wr() 1063 rchp->ibcq.cq_context); in complete_rq_drain_wr() 1641 rchp->ibcq.cq_context); in __flush_qp() 1648 rchp->ibcq.cq_context); in __flush_qp() 1654 schp->ibcq.cq_context); in __flush_qp() 1678 (*rchp->ibcq.comp_handler)(&rchp->ibcq, rchp->ibcq.cq_context); in flush_qp() 1684 schp->ibcq.cq_context); in flush_qp()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
A D | cq.c | 92 int eqn = MLX5_GET(cqc, MLX5_ADDR_OF(create_cq_in, in, cq_context), in mlx5_core_create_cq() 209 cqc = MLX5_ADDR_OF(modify_cq_in, in, cq_context); in mlx5_core_modify_cq_moderation()
|
A D | debugfs.c | 340 ctx = MLX5_ADDR_OF(query_cq_out, out, cq_context); in cq_read_field()
|
/linux/drivers/infiniband/hw/mlx5/ |
A D | cq.c | 45 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx5_ib_cq_comp() 65 ibcq->event_handler(&event, ibcq->cq_context); in mlx5_ib_cq_event() 776 cqc = MLX5_ADDR_OF(create_cq_in, *cqb, cq_context); in create_cq_user() 903 cqc = MLX5_ADDR_OF(create_cq_in, *cqb, cq_context); in create_cq_kernel() 931 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in notify_soft_wc_handler() 992 cqc = MLX5_ADDR_OF(create_cq_in, cqb, cq_context); in mlx5_ib_create_cq() 1339 cqc = MLX5_ADDR_OF(modify_cq_in, in, cq_context); in mlx5_ib_resize_cq()
|
/linux/drivers/infiniband/sw/rdmavt/ |
A D | cq.c | 74 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rvt_cq_enter() 140 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in send_complete()
|
/linux/net/9p/ |
A D | trans_rdma.c | 293 struct p9_client *client = cq->cq_context; in recv_done() 344 struct p9_client *client = cq->cq_context; in send_done()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
A D | health.c | 49 cqc = MLX5_ADDR_OF(query_cq_out, out, cq_context); in mlx5e_health_cq_diag_fmsg()
|
/linux/drivers/infiniband/hw/qedr/ |
A D | qedr_roce_cm.c | 93 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_tx_packet() 122 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_rx_packet()
|
A D | main.c | 470 (&cq->ibcq, cq->ibcq.cq_context); in qedr_irq_handler() 725 ibcq->event_handler(&event, ibcq->cq_context); in qedr_affiliated_event()
|
/linux/drivers/infiniband/hw/mlx4/ |
A D | cq.c | 46 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx4_ib_cq_comp() 65 ibcq->event_handler(&event, ibcq->cq_context); in mlx4_ib_cq_event()
|
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
A D | pvrdma_main.c | 328 ibcq->event_handler(&e, ibcq->cq_context); in pvrdma_cq_event() 493 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in pvrdma_intrx_handler()
|