Home
last modified time | relevance | path

Searched refs:recv_cq (Results 1 – 25 of 60) sorted by relevance

123

/linux/drivers/infiniband/hw/hns/
A Dhns_roce_qp.c282 hr_recv_cq = recv_cq ? to_hr_cq(recv_cq) : NULL; in add_qp_to_list()
313 init_attr->recv_cq); in hns_roce_qp_store()
1350 __acquire(&recv_cq->lock); in hns_roce_lock_cqs()
1353 __acquire(&recv_cq->lock); in hns_roce_lock_cqs()
1359 __acquire(&recv_cq->lock); in hns_roce_lock_cqs()
1371 __releases(&recv_cq->lock) in hns_roce_unlock_cqs()
1374 __release(&recv_cq->lock); in hns_roce_unlock_cqs()
1377 __release(&recv_cq->lock); in hns_roce_unlock_cqs()
1381 spin_unlock(&recv_cq->lock); in hns_roce_unlock_cqs()
1383 __release(&recv_cq->lock); in hns_roce_unlock_cqs()
[all …]
A Dhns_roce_hw_v1.c378 hr_qp->ibqp.recv_cq)) { in hns_roce_v1_post_recv()
879 hr_qp->ibqp.recv_cq = cq; in hns_roce_v1_rsv_lp_qp()
2645 to_hr_cq(ibqp->recv_cq)->cqn); in hns_roce_v1_m_sqp()
2692 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_sqp()
3312 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_qp()
3612 struct hns_roce_cq *send_cq, *recv_cq; in hns_roce_v1_destroy_qp() local
3620 recv_cq = hr_qp->ibqp.recv_cq ? to_hr_cq(hr_qp->ibqp.recv_cq) : NULL; in hns_roce_v1_destroy_qp()
3622 hns_roce_lock_cqs(send_cq, recv_cq); in hns_roce_v1_destroy_qp()
3624 if (recv_cq) in hns_roce_v1_destroy_qp()
3630 if (send_cq && send_cq != recv_cq) in hns_roce_v1_destroy_qp()
[all …]
A Dhns_roce_hw_v2.c865 hr_qp->ibqp.recv_cq))) { in hns_roce_v2_post_recv()
4963 if (ibqp->recv_cq && ibqp->recv_cq != ibqp->send_cq) in clear_qp()
4964 hns_roce_v2_cq_clean(to_hr_cq(ibqp->recv_cq), in clear_qp()
5201 qp_init_attr->recv_cq = ibqp->recv_cq; in hns_roce_v2_query_qp()
5226 struct hns_roce_cq *send_cq, *recv_cq; in hns_roce_v2_destroy_qp_common() local
5241 recv_cq = hr_qp->ibqp.recv_cq ? to_hr_cq(hr_qp->ibqp.recv_cq) : NULL; in hns_roce_v2_destroy_qp_common()
5244 hns_roce_lock_cqs(send_cq, recv_cq); in hns_roce_v2_destroy_qp_common()
5247 if (recv_cq) in hns_roce_v2_destroy_qp_common()
5248 __hns_roce_v2_cq_clean(recv_cq, hr_qp->qpn, in hns_roce_v2_destroy_qp_common()
5253 if (send_cq && send_cq != recv_cq) in hns_roce_v2_destroy_qp_common()
[all …]
/linux/drivers/infiniband/ulp/ipoib/
A Dipoib_verbs.c178 priv->recv_cq = ib_create_cq(priv->ca, ipoib_ib_rx_completion, NULL, in ipoib_transport_dev_init()
180 if (IS_ERR(priv->recv_cq)) { in ipoib_transport_dev_init()
194 if (ib_req_notify_cq(priv->recv_cq, IB_CQ_NEXT_COMP)) in ipoib_transport_dev_init()
198 init_attr.recv_cq = priv->recv_cq; in ipoib_transport_dev_init()
247 ib_destroy_cq(priv->recv_cq); in ipoib_transport_dev_init()
267 ib_destroy_cq(priv->recv_cq); in ipoib_transport_dev_cleanup()
A Dipoib_ethtool.c103 ret = rdma_set_cq_moderation(priv->recv_cq, in ipoib_set_coalesce()
A Dipoib_ib.c466 n = ib_poll_cq(priv->recv_cq, t, priv->ibwc); in ipoib_rx_poll()
488 if (unlikely(ib_req_notify_cq(priv->recv_cq, in ipoib_rx_poll()
843 ib_req_notify_cq(priv->recv_cq, IB_CQ_NEXT_COMP); in ipoib_ib_dev_stop_default()
976 n = ib_poll_cq(priv->recv_cq, IPOIB_NUM_WC, priv->ibwc); in ipoib_drain_cq()
A Dipoib_cm.c255 .send_cq = priv->recv_cq, /* For drain WR */ in ipoib_cm_create_rx_qp()
256 .recv_cq = priv->recv_cq, in ipoib_cm_create_rx_qp()
1062 .recv_cq = priv->recv_cq, in ipoib_cm_create_tx_qp()
/linux/drivers/infiniband/hw/mlx5/
A Dqp.c1233 recv_cq ? get_ts_format(dev, recv_cq, fr_sup, rt_sup) : in get_qp_ts_format()
2131 if (recv_cq) in create_dci()
2334 if (recv_cq) in create_user_qp()
2463 if (recv_cq) in create_kernel_qp()
2479 if (recv_cq) { in mlx5_ib_lock_cqs()
2509 if (recv_cq) { in mlx5_ib_unlock_cqs()
2540 *recv_cq = NULL; in get_cqs()
2604 &recv_cq); in destroy_qp_common()
2613 if (recv_cq) in destroy_qp_common()
4146 if (recv_cq) in __mlx5_ib_modify_qp()
[all …]
A Dgsi.c206 .recv_cq = gsi->rx_qp->recv_cq, in create_gsi_ud_qp()
A Dmem.c246 qp_init_attr.recv_cq = cq; in mlx5_ib_test_wc()
/linux/drivers/infiniband/hw/mthca/
A Dmthca_qp.c1165 struct mthca_cq *recv_cq, in mthca_alloc_qp_common() argument
1294 struct mthca_cq *recv_cq, in mthca_alloc_qp() argument
1339 if (send_cq == recv_cq) { in mthca_lock_cqs()
1341 __acquire(&recv_cq->lock); in mthca_lock_cqs()
1354 if (send_cq == recv_cq) { in mthca_unlock_cqs()
1355 __release(&recv_cq->lock); in mthca_unlock_cqs()
1358 spin_unlock(&recv_cq->lock); in mthca_unlock_cqs()
1369 struct mthca_cq *recv_cq, in mthca_alloc_sqp() argument
1449 struct mthca_cq *recv_cq; in mthca_free_qp() local
1452 recv_cq = to_mcq(qp->ibqp.recv_cq); in mthca_free_qp()
[all …]
A Dmthca_dev.h536 struct mthca_cq *recv_cq,
545 struct mthca_cq *recv_cq,
/linux/drivers/infiniband/core/
A Duverbs_std_types_qp.c94 struct ib_cq *recv_cq = NULL; in UVERBS_HANDLER() local
181 recv_cq = uverbs_attr_get_obj(attrs, in UVERBS_HANDLER()
183 if (IS_ERR(recv_cq)) in UVERBS_HANDLER()
184 return PTR_ERR(recv_cq); in UVERBS_HANDLER()
235 attr.recv_cq = recv_cq; in UVERBS_HANDLER()
A Dverbs.c1183 qp->send_cq = qp->recv_cq = NULL; in create_xrc_qp_user()
1236 qp->recv_cq = attr->recv_cq; in create_qp()
1250 qp->recv_cq = attr->recv_cq; in create_qp()
1311 if (qp->recv_cq) in ib_qp_usecnt_inc()
1312 atomic_inc(&qp->recv_cq->usecnt); in ib_qp_usecnt_inc()
1326 if (qp->recv_cq) in ib_qp_usecnt_dec()
1327 atomic_dec(&qp->recv_cq->usecnt); in ib_qp_usecnt_dec()
2791 struct ib_cq *cq = qp->recv_cq; in __ib_drain_rq()
2874 trace_cq_drain_complete(qp->recv_cq); in ib_drain_rq()
/linux/drivers/infiniband/hw/mlx4/
A Dqp.c1266 if (send_cq == recv_cq) { in mlx4_ib_lock_cqs()
1268 __acquire(&recv_cq->lock); in mlx4_ib_lock_cqs()
1273 spin_lock(&recv_cq->lock); in mlx4_ib_lock_cqs()
1281 if (send_cq == recv_cq) { in mlx4_ib_unlock_cqs()
1317 *recv_cq = *send_cq; in get_cqs()
1321 *recv_cq = *send_cq; in get_cqs()
1324 *recv_cq = (src == MLX4_IB_QP_SRC) ? to_mcq(qp->ibqp.recv_cq) : in get_cqs()
1327 *recv_cq; in get_cqs()
1408 if (send_cq != recv_cq) in destroy_qp_common()
2320 recv_cq = send_cq; in __mlx4_ib_modify_qp()
[all …]
/linux/fs/ksmbd/
A Dtransport_rdma.c104 struct ib_cq *recv_cq; member
437 if (t->recv_cq) in free_transport()
438 ib_free_cq(t->recv_cq); in free_transport()
1827 t->recv_cq = ib_alloc_cq(t->cm_id->device, t, in smb_direct_create_qpair()
1830 if (IS_ERR(t->recv_cq)) { in smb_direct_create_qpair()
1832 ret = PTR_ERR(t->recv_cq); in smb_direct_create_qpair()
1833 t->recv_cq = NULL; in smb_direct_create_qpair()
1844 qp_attr.recv_cq = t->recv_cq; in smb_direct_create_qpair()
1862 if (t->recv_cq) { in smb_direct_create_qpair()
1863 ib_destroy_cq(t->recv_cq); in smb_direct_create_qpair()
[all …]
/linux/net/sunrpc/xprtrdma/
A Dverbs.c354 if (ep->re_attr.recv_cq) in rpcrdma_ep_destroy()
355 ib_free_cq(ep->re_attr.recv_cq); in rpcrdma_ep_destroy()
356 ep->re_attr.recv_cq = NULL; in rpcrdma_ep_destroy()
442 ep->re_attr.recv_cq = ib_alloc_cq_any(device, r_xprt, in rpcrdma_ep_create()
445 if (IS_ERR(ep->re_attr.recv_cq)) { in rpcrdma_ep_create()
446 rc = PTR_ERR(ep->re_attr.recv_cq); in rpcrdma_ep_create()
1391 rep->rr_cid.ci_queue_id = ep->re_attr.recv_cq->res.id; in rpcrdma_post_recvs()
/linux/drivers/infiniband/hw/vmw_pvrdma/
A Dpvrdma_qp.c59 struct pvrdma_cq **recv_cq) in get_cqs() argument
62 *recv_cq = to_vcq(qp->ibqp.recv_cq); in get_cqs()
360 cmd->recv_cq_handle = to_vcq(init_attr->recv_cq)->cq_handle; in pvrdma_create_qp()
1043 init_attr->recv_cq = qp->ibqp.recv_cq; in pvrdma_query_qp()
/linux/fs/cifs/
A Dsmbdirect.c1392 ib_free_cq(info->recv_cq); in smbd_destroy()
1579 info->recv_cq = NULL; in _smbd_get_connection()
1588 info->recv_cq = in _smbd_get_connection()
1591 if (IS_ERR(info->recv_cq)) { in _smbd_get_connection()
1592 info->recv_cq = NULL; in _smbd_get_connection()
1607 qp_attr.recv_cq = info->recv_cq; in _smbd_get_connection()
1724 if (info->recv_cq) in _smbd_get_connection()
1725 ib_free_cq(info->recv_cq); in _smbd_get_connection()
A Dsmbdirect.h59 struct ib_cq *send_cq, *recv_cq; member
/linux/drivers/infiniband/sw/rxe/
A Drxe_qp.c76 if (!init->recv_cq || !init->send_cq) { in rxe_qp_chk_init()
330 struct rxe_cq *rcq = to_rcq(init->recv_cq); in rxe_qp_from_init()
384 init->recv_cq = qp->ibqp.recv_cq; in rxe_qp_to_init()
/linux/drivers/infiniband/ulp/srp/
A Dib_srp.c534 struct ib_cq *recv_cq, *send_cq; in srp_create_ch_ib() local
545 recv_cq = ib_alloc_cq(dev->dev, ch, target->queue_size + 1, in srp_create_ch_ib()
547 if (IS_ERR(recv_cq)) { in srp_create_ch_ib()
548 ret = PTR_ERR(recv_cq); in srp_create_ch_ib()
567 init_attr->recv_cq = recv_cq; in srp_create_ch_ib()
602 if (ch->recv_cq) in srp_create_ch_ib()
603 ib_free_cq(ch->recv_cq); in srp_create_ch_ib()
608 ch->recv_cq = recv_cq; in srp_create_ch_ib()
630 ib_free_cq(recv_cq); in srp_create_ch_ib()
673 ib_free_cq(ch->recv_cq); in srp_free_ch_ib()
[all …]
A Dib_srp.h148 struct ib_cq *recv_cq; member
/linux/drivers/net/ethernet/ibm/ehea/
A Dehea_main.c198 arr[i++].fwh = pr->recv_cq->fw_handle; in ehea_update_firmware_handles()
889 ehea_reset_cq_ep(pr->recv_cq); in ehea_poll()
891 ehea_reset_cq_n1(pr->recv_cq); in ehea_poll()
1466 pr->recv_cq = ehea_create_cq(adapter, pr_cfg->max_entries_rcq, in ehea_init_port_res()
1469 if (!pr->recv_cq) { in ehea_init_port_res()
1485 pr->recv_cq->attr.act_nr_of_cqes); in ehea_init_port_res()
1510 init_attr->recv_cq_handle = pr->recv_cq->fw_handle; in ehea_init_port_res()
1560 ehea_destroy_cq(pr->recv_cq); in ehea_init_port_res()
1577 ehea_destroy_cq(pr->recv_cq); in ehea_clean_portres()
/linux/drivers/infiniband/sw/siw/
A Dsiw_verbs.c345 if (!attrs->send_cq || (!attrs->recv_cq && !attrs->srq)) { in siw_create_qp()
397 qp->rcq = to_siw_cq(attrs->recv_cq); in siw_create_qp()
528 qp_init_attr->recv_cq = base_qp->recv_cq; in siw_query_qp()

Completed in 97 milliseconds

123