/linux/drivers/infiniband/hw/irdma/ |
A D | ctrl.c | 432 cqp = qp->dev->cqp; in irdma_sc_qp_create() 486 cqp = qp->dev->cqp; in irdma_sc_qp_modify() 555 cqp = qp->dev->cqp; in irdma_sc_qp_destroy() 1053 cqp = dev->cqp; in irdma_sc_alloc_stag() 1131 cqp = dev->cqp; in irdma_sc_mr_reg_non_shared() 1196 cqp = dev->cqp; in irdma_sc_dealloc_stag() 1236 cqp = dev->cqp; in irdma_sc_mw_alloc() 2318 cqp = dev->cqp; in irdma_sc_qp_upload_context() 2512 cqp = cq->dev->cqp; in irdma_sc_cq_create() 3376 info->cqp = cqp; in irdma_sc_ccq_get_cqe_info() [all …]
|
A D | uda.h | 41 enum irdma_status_code irdma_access_mcast_grp(struct irdma_sc_cqp *cqp, 50 static inline enum irdma_status_code irdma_sc_create_ah(struct irdma_sc_cqp *cqp, in irdma_sc_create_ah() argument 54 return irdma_sc_access_ah(cqp, info, IRDMA_CQP_OP_CREATE_ADDR_HANDLE, in irdma_sc_create_ah() 58 static inline enum irdma_status_code irdma_sc_destroy_ah(struct irdma_sc_cqp *cqp, in irdma_sc_destroy_ah() argument 62 return irdma_sc_access_ah(cqp, info, IRDMA_CQP_OP_DESTROY_ADDR_HANDLE, in irdma_sc_destroy_ah() 66 static inline enum irdma_status_code irdma_sc_create_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_sc_create_mcast_grp() argument 70 return irdma_access_mcast_grp(cqp, info, IRDMA_CQP_OP_CREATE_MCAST_GRP, in irdma_sc_create_mcast_grp() 74 static inline enum irdma_status_code irdma_sc_modify_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_sc_modify_mcast_grp() argument 78 return irdma_access_mcast_grp(cqp, info, IRDMA_CQP_OP_MODIFY_MCAST_GRP, in irdma_sc_modify_mcast_grp() 82 static inline enum irdma_status_code irdma_sc_destroy_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_sc_destroy_mcast_grp() argument [all …]
|
A D | utils.c | 525 struct irdma_cqp *cqp = &rf->cqp; in irdma_cleanup_pending_cqp_op() local 830 struct irdma_sc_cqp *cqp = dev->cqp; in irdma_cqp_qp_suspend_resume() local 841 cqp_info->in.u.suspend_resume.cqp = cqp; in irdma_cqp_qp_suspend_resume() 958 cqp_info->in.u.query_fpm_val.cqp = dev->cqp; in irdma_cqp_query_fpm_val_cmd() 993 cqp_info->in.u.commit_fpm_val.cqp = dev->cqp; in irdma_cqp_commit_fpm_val_cmd() 1098 cqp_info->in.u.manage_push_page.cqp = &rf->cqp.sc_cqp; in irdma_dealloc_push_page() 1885 cqp_info->in.u.stats_gather.cqp = &rf->cqp.sc_cqp; in irdma_cqp_gather_stats_cmd() 1926 cqp_info->in.u.stats_manage.cqp = &rf->cqp.sc_cqp; in irdma_cqp_stats_inst_cmd() 2027 cqp_info->in.u.ws_node.cqp = cqp; in irdma_cqp_ws_node_cmd() 2084 cqp_info->in.u.ah_create.cqp = &rf->cqp.sc_cqp; in irdma_ah_cqp_op() [all …]
|
A D | hw.c | 569 struct irdma_cqp *cqp = &rf->cqp; in irdma_destroy_cqp() local 579 dma_free_coherent(dev->hw->device, cqp->sq.size, cqp->sq.va, in irdma_destroy_cqp() 580 cqp->sq.pa); in irdma_destroy_cqp() 581 cqp->sq.va = NULL; in irdma_destroy_cqp() 909 struct irdma_cqp *cqp = &rf->cqp; in irdma_create_cqp() local 923 dev->cqp = &cqp->sc_cqp; in irdma_create_cqp() 927 cqp->sq.va = dma_alloc_coherent(dev->hw->device, cqp->sq.size, in irdma_create_cqp() 984 list_add_tail(&cqp->cqp_requests[i].list, &cqp->cqp_avail_reqs); in irdma_create_cqp() 2287 cqp_info->in.u.manage_apbvt_entry.cqp = &iwdev->rf->cqp.sc_cqp; in irdma_cqp_manage_apbvt_cmd() 2397 cqp_info->in.u.add_arp_cache_entry.cqp = &rf->cqp.sc_cqp; in irdma_manage_arp_cache() [all …]
|
A D | uda.c | 19 enum irdma_status_code irdma_sc_access_ah(struct irdma_sc_cqp *cqp, in irdma_sc_access_ah() argument 26 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch); in irdma_sc_access_ah() 69 FIELD_PREP(IRDMA_UDA_CQPSQ_MAV_WQEVALID, cqp->polarity) | in irdma_sc_access_ah() 78 irdma_sc_cqp_post_sq(cqp); in irdma_sc_access_ah() 118 enum irdma_status_code irdma_access_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_access_mcast_grp() argument 126 ibdev_dbg(to_ibdev(cqp->dev), "WQE: mg_id out of range\n"); in irdma_access_mcast_grp() 130 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch); in irdma_access_mcast_grp() 132 ibdev_dbg(to_ibdev(cqp->dev), "WQE: ring full\n"); in irdma_access_mcast_grp() 163 FIELD_PREP(IRDMA_UDA_CQPSQ_MG_WQEVALID, cqp->polarity) | in irdma_access_mcast_grp() 174 irdma_sc_cqp_post_sq(cqp); in irdma_access_mcast_grp()
|
A D | type.h | 644 struct irdma_sc_cqp *cqp; member 708 struct irdma_sc_cqp *cqp; member 1337 struct irdma_sc_cqp *cqp; member 1343 struct irdma_sc_cqp *cqp; member 1349 struct irdma_sc_cqp *cqp; member 1355 struct irdma_sc_cqp *cqp; member 1362 struct irdma_sc_cqp *cqp; member 1367 struct irdma_sc_cqp *cqp; member 1417 struct irdma_sc_cqp *cqp; member 1425 struct irdma_sc_cqp *cqp; member [all …]
|
A D | protos.h | 19 void irdma_sc_cqp_post_sq(struct irdma_sc_cqp *cqp); 20 __le64 *irdma_sc_cqp_get_next_send_wqe(struct irdma_sc_cqp *cqp, u64 scratch); 85 irdma_sc_static_hmc_pages_allocated(struct irdma_sc_cqp *cqp, u64 scratch,
|
A D | puda.c | 609 struct irdma_sc_cqp *cqp; in irdma_puda_qp_wqe() local 615 cqp = dev->cqp; in irdma_puda_qp_wqe() 616 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, 0); in irdma_puda_qp_wqe() 628 FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity); in irdma_puda_qp_wqe() 635 irdma_sc_cqp_post_sq(cqp); in irdma_puda_qp_wqe() 729 struct irdma_sc_cqp *cqp; in irdma_puda_cq_wqe() local 734 cqp = dev->cqp; in irdma_puda_cq_wqe() 735 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, 0); in irdma_puda_cq_wqe() 754 FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity); in irdma_puda_cq_wqe() 761 irdma_sc_cqp_post_sq(dev->cqp); in irdma_puda_cq_wqe() [all …]
|
A D | main.h | 300 struct irdma_cqp cqp; member 476 struct irdma_cqp_request *irdma_alloc_and_get_cqp_request(struct irdma_cqp *cqp, 478 void irdma_free_cqp_request(struct irdma_cqp *cqp, 480 void irdma_put_cqp_request(struct irdma_cqp *cqp,
|
A D | verbs.c | 272 cqp_info->in.u.manage_push_page.cqp = &iwdev->rf->cqp.sc_cqp; in irdma_alloc_push_page() 282 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_alloc_push_page() 688 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_cqp_create_qp_cmd() 1884 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_resize_cq() 2094 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_create_cq() 2453 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_hw_alloc_mw() 2515 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_dealloc_mw() 2553 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_hw_alloc_stag() 2711 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_hwreg_mr() 3030 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_dereg_mr() [all …]
|
A D | hmc.c | 137 return dev->cqp->process_cqp_sds(dev, &sdinfo); in irdma_hmc_sd_one() 177 ret_code = dev->cqp->process_cqp_sds(dev, &sdinfo); in irdma_hmc_sd_grp() 189 ret_code = dev->cqp->process_cqp_sds(dev, &sdinfo); in irdma_hmc_sd_grp()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
A D | params.h | 22 struct mlx5e_cq_param cqp; member 29 struct mlx5e_cq_param cqp; member
|
A D | params.c | 589 mlx5e_build_rx_cq_param(mdev, params, xsk, ¶m->cqp); in mlx5e_build_rq_param() 648 mlx5e_build_tx_cq_param(mdev, params, ¶m->cqp); in mlx5e_build_sq_param() 766 mlx5e_build_ico_cq_param(mdev, log_wq_size, ¶m->cqp); in mlx5e_build_icosq_param() 783 mlx5e_build_ico_cq_param(mdev, log_wq_size, ¶m->cqp); in mlx5e_build_async_icosq_param() 796 mlx5e_build_tx_cq_param(mdev, params, ¶m->cqp); in mlx5e_build_xdpsq_param()
|
A D | ptp.c | 378 cq_param = &cparams->txq_sq_param.cqp; in mlx5e_ptp_open_tx_cqs() 425 cq_param = &cparams->rq_param.cqp; in mlx5e_ptp_open_rx_cq() 453 mlx5e_build_tx_cq_param(mdev, params, ¶m->cqp); in mlx5e_ptp_build_sq_param()
|
A D | trap.c | 69 err = mlx5e_open_cq(priv, trap_moder, &rq_param->cqp, &ccp, &rq->cq); in mlx5e_open_trap_rq()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
A D | setup.c | 113 err = mlx5e_open_cq(c->priv, params->rx_cq_moderation, &cparam->rq.cqp, &ccp, in mlx5e_open_xsk() 122 err = mlx5e_open_cq(c->priv, params->tx_cq_moderation, &cparam->xdp_sq.cqp, &ccp, in mlx5e_open_xsk()
|
/linux/drivers/scsi/lpfc/ |
A D | lpfc_sli4.h | 1077 int lpfc_cq_create_set(struct lpfc_hba *phba, struct lpfc_queue **cqp, 1087 struct lpfc_queue **drqp, struct lpfc_queue **cqp,
|
A D | lpfc_sli.c | 16400 lpfc_cq_create_set(struct lpfc_hba *phba, struct lpfc_queue **cqp, in lpfc_cq_create_set() argument 16417 if (!cqp || !hdwq || !numcq) in lpfc_cq_create_set() 16425 length += ((numcq * cqp[0]->page_count) * in lpfc_cq_create_set() 16443 cq = cqp[idx]; in lpfc_cq_create_set() 16617 cq = cqp[idx]; in lpfc_cq_create_set() 17419 struct lpfc_queue **drqp, struct lpfc_queue **cqp, in lpfc_mrq_create() argument 17434 if (!hrqp || !drqp || !cqp || !numrq) in lpfc_mrq_create() 17470 cq = cqp[idx]; in lpfc_mrq_create()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
A D | en_main.c | 1871 err = mlx5e_open_cq(c->priv, params->tx_cq_moderation, &cparam->txq_sq.cqp, in mlx5e_open_tx_cqs() 2066 err = mlx5e_open_cq(c->priv, icocq_moder, &cparam->async_icosq.cqp, &ccp, in mlx5e_open_queues() 2071 err = mlx5e_open_cq(c->priv, icocq_moder, &cparam->icosq.cqp, &ccp, in mlx5e_open_queues() 2080 err = mlx5e_open_cq(c->priv, params->tx_cq_moderation, &cparam->xdp_sq.cqp, &ccp, in mlx5e_open_queues() 2085 err = mlx5e_open_cq(c->priv, params->rx_cq_moderation, &cparam->rq.cqp, &ccp, in mlx5e_open_queues() 2090 err = c->xdp ? mlx5e_open_cq(c->priv, params->tx_cq_moderation, &cparam->xdp_sq.cqp, in mlx5e_open_queues()
|