Home
last modified time | relevance | path

Searched refs:ibqp (Results 1 – 25 of 116) sorted by relevance

12345

/linux/drivers/infiniband/hw/mlx5/
A Dwr.h44 int mlx5_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
46 int mlx5_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
49 static inline int mlx5_ib_post_send_nodrain(struct ib_qp *ibqp, in mlx5_ib_post_send_nodrain() argument
53 return mlx5_ib_post_send(ibqp, wr, bad_wr, false); in mlx5_ib_post_send_nodrain()
56 static inline int mlx5_ib_post_send_drain(struct ib_qp *ibqp, in mlx5_ib_post_send_drain() argument
60 return mlx5_ib_post_send(ibqp, wr, bad_wr, true); in mlx5_ib_post_send_drain()
63 static inline int mlx5_ib_post_recv_nodrain(struct ib_qp *ibqp, in mlx5_ib_post_recv_nodrain() argument
67 return mlx5_ib_post_recv(ibqp, wr, bad_wr, false); in mlx5_ib_post_recv_nodrain()
70 static inline int mlx5_ib_post_recv_drain(struct ib_qp *ibqp, in mlx5_ib_post_recv_drain() argument
74 return mlx5_ib_post_recv(ibqp, wr, bad_wr, true); in mlx5_ib_post_recv_drain()
/linux/drivers/infiniband/hw/hfi1/
A Dqp.c163 struct ib_qp *ibqp = &qp->ibqp; in hfi1_check_modify_qp() local
225 struct ib_qp *ibqp = &qp->ibqp; in hfi1_modify_qp() local
270 switch (qp->ibqp.qp_type) { in hfi1_setup_wqe()
546 switch (qp->ibqp.qp_type) { in qp_to_sdma_engine()
568 switch (qp->ibqp.qp_type) { in qp_to_send_context()
618 qp->ibqp.qp_num, in qp_iter_print()
788 ev.device = qp->ibqp.device; in hfi1_migrate_qp()
789 ev.element.qp = &qp->ibqp; in hfi1_migrate_qp()
791 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in hfi1_migrate_qp()
904 ev.element.qp = &qp->ibqp; in hfi1_qp_iter_cb()
[all …]
A Dud.c58 IB_QPT_UD : qp->ibqp.qp_type; in ud_loopback()
69 if (qp->ibqp.qp_num > 1) { in ud_loopback()
82 sqp->ibqp.qp_num, qp->ibqp.qp_num, in ud_loopback()
93 if (qp->ibqp.qp_num) { in ud_loopback()
131 if (qp->ibqp.qp_num == 0) in ud_loopback()
196 wc.qp = &qp->ibqp; in ud_loopback()
198 if (qp->ibqp.qp_type == IB_QPT_GSI || qp->ibqp.qp_type == IB_QPT_SMI) { in ud_loopback()
242 if (qp->ibqp.qp_type == IB_QPT_GSI || qp->ibqp.qp_type == IB_QPT_SMI) in hfi1_make_bth_deth()
870 if (qp->ibqp.qp_num) { in hfi1_ud_rcv()
988 wc.qp = &qp->ibqp; in hfi1_ud_rcv()
[all …]
A Dtrace_tid.h208 __entry->qpn = qp->ibqp.qp_num;
266 __entry->qpn = qp->ibqp.qp_num;
318 __entry->qpn = qp->ibqp.qp_num;
430 __entry->qpn = qp->ibqp.qp_num;
468 __entry->qpn = qp->ibqp.qp_num;
515 __entry->qpn = qp->ibqp.qp_num;
653 __entry->qpn = qp->ibqp.qp_num;
695 __entry->qpn = qp->ibqp.qp_num;
769 __entry->qpn = qp->ibqp.qp_num;
890 __entry->qpn = qp->ibqp.qp_num;
[all …]
A Dtrace_rc.h21 DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device))
32 DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device));
33 __entry->qpn = qp->ibqp.qp_num;
83 DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device))
92 DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device));
93 __entry->qpn = qp->ibqp.qp_num;
/linux/drivers/infiniband/sw/rdmavt/
A Dqp.c719 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_qp_mr_clean()
1318 wc.qp = &qp->ibqp; in rvt_error_qp()
1652 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_modify_qp()
1658 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_modify_qp()
1790 !qp->ibqp.srq; in rvt_post_recv()
2279 pd = ibpd_to_rvtpd(qp->ibqp.srq ? qp->ibqp.srq->pd : qp->ibqp.pd); in init_sge()
2461 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_comm_est()
2481 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_rc_error()
2511 struct ib_qp *ibqp = &qp->ibqp; in rvt_add_retry_timer_ext() local
2972 qp->ibqp.qp_type != sqp->ibqp.qp_type) { in rvt_ruc_loopback()
[all …]
A Dqp.h13 int rvt_create_qp(struct ib_qp *ibqp, struct ib_qp_init_attr *init_attr,
15 int rvt_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
17 int rvt_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata);
18 int rvt_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,
20 int rvt_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr,
22 int rvt_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
A Dtrace_qp.h21 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device))
26 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device));
27 __entry->qpn = qp->ibqp.qp_num;
51 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device))
58 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device));
59 __entry->qpn = qp->ibqp.qp_num;
A Dtrace_tx.h52 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device))
72 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device));
75 __entry->qpn = qp->ibqp.qp_num;
76 __entry->qpt = qp->ibqp.qp_type;
119 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device))
131 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device));
134 __entry->qpn = qp->ibqp.qp_num;
135 __entry->qpt = qp->ibqp.qp_type;
A Dmcast.c238 int rvt_attach_mcast(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in rvt_attach_mcast() argument
240 struct rvt_qp *qp = ibqp_to_rvtqp(ibqp); in rvt_attach_mcast()
241 struct rvt_dev_info *rdi = ib_to_rvt(ibqp->device); in rvt_attach_mcast()
247 if (ibqp->qp_num <= 1 || qp->state == IB_QPS_RESET) in rvt_attach_mcast()
301 int rvt_detach_mcast(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in rvt_detach_mcast() argument
303 struct rvt_qp *qp = ibqp_to_rvtqp(ibqp); in rvt_detach_mcast()
304 struct rvt_dev_info *rdi = ib_to_rvt(ibqp->device); in rvt_detach_mcast()
312 if (ibqp->qp_num <= 1) in rvt_detach_mcast()
/linux/drivers/infiniband/hw/qib/
A Dqib_ud.c87 if (qp->ibqp.qp_num > 1) { in qib_ud_loopback()
99 sqp->ibqp.qp_num, qp->ibqp.qp_num, in qib_ud_loopback()
111 if (qp->ibqp.qp_num) { in qib_ud_loopback()
149 if (qp->ibqp.qp_num == 0) in qib_ud_loopback()
204 wc.qp = &qp->ibqp; in qib_ud_loopback()
205 wc.src_qp = sqp->ibqp.qp_num; in qib_ud_loopback()
464 if (qp->ibqp.qp_num) { in qib_ud_rcv()
468 if (qp->ibqp.qp_num > 1) { in qib_ud_rcv()
509 if (qp->ibqp.qp_num > 1 && in qib_ud_rcv()
539 if (qp->ibqp.qp_num == 0) in qib_ud_rcv()
[all …]
A Dqib_ruc.c53 ev.device = qp->ibqp.device; in qib_migrate_qp()
54 ev.element.qp = &qp->ibqp; in qib_migrate_qp()
56 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in qib_migrate_qp()
114 0, qp->ibqp.qp_num, in qib_ruc_check_hdr()
153 0, qp->ibqp.qp_num, in qib_ruc_check_hdr()
210 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_make_ruc_header()
266 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_do_send()
271 if ((qp->ibqp.qp_type == IB_QPT_RC || in qib_do_send()
272 qp->ibqp.qp_type == IB_QPT_UC) && in qib_do_send()
279 if (qp->ibqp.qp_type == IB_QPT_RC) in qib_do_send()
[all …]
/linux/drivers/infiniband/hw/vmw_pvrdma/
A Dpvrdma_qp.c61 *send_cq = to_vcq(qp->ibqp.send_cq); in get_cqs()
62 *recv_cq = to_vcq(qp->ibqp.recv_cq); in get_cqs()
194 struct pvrdma_qp *qp = to_vqp(ibqp); in pvrdma_create_qp()
394 qp->ibqp.qp_num = resp_v2->qpn; in pvrdma_create_qp()
397 qp->ibqp.qp_num = resp->qpn; in pvrdma_create_qp()
406 qp_resp.qpn = qp->ibqp.qp_num; in pvrdma_create_qp()
529 struct pvrdma_qp *qp = to_vqp(ibqp); in pvrdma_modify_qp()
674 struct pvrdma_qp *qp = to_vqp(ibqp); in pvrdma_post_send()
774 switch (qp->ibqp.qp_type) { in pvrdma_post_send()
886 struct pvrdma_qp *qp = to_vqp(ibqp); in pvrdma_post_recv()
[all …]
/linux/drivers/infiniband/hw/mthca/
A Dmthca_qp.c261 event.element.qp = &qp->ibqp; in mthca_qp_event()
262 if (qp->ibqp.event_handler) in mthca_qp_event()
263 qp->ibqp.event_handler(&event, qp->ibqp.qp_context); in mthca_qp_event()
626 if (qp->ibqp.uobject) in __mthca_modify_qp()
759 if (ibqp->srq) in __mthca_modify_qp()
784 if (ibqp->srq) in __mthca_modify_qp()
836 qp->ibqp.srq ? to_msrq(qp->ibqp.srq) : NULL); in __mthca_modify_qp()
837 if (qp->ibqp.send_cq != qp->ibqp.recv_cq) in __mthca_modify_qp()
1479 if (!qp->ibqp.uobject) { in mthca_free_qp()
1481 qp->ibqp.srq ? to_msrq(qp->ibqp.srq) : NULL); in mthca_free_qp()
[all …]
A Dmthca_mcg.c120 int mthca_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mthca_multicast_attach() argument
122 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_multicast_attach()
165 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) { in mthca_multicast_attach()
167 ibqp->qp_num); in mthca_multicast_attach()
171 mgm->qp[i] = cpu_to_be32(ibqp->qp_num | (1 << 31)); in mthca_multicast_attach()
214 int mthca_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mthca_multicast_detach() argument
216 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_multicast_detach()
242 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) in mthca_multicast_detach()
249 mthca_err(dev, "QP %06x not found in MGM\n", ibqp->qp_num); in mthca_multicast_detach()
/linux/drivers/infiniband/hw/mlx4/
A Dqp.c206 struct ib_qp *ibqp = &to_mibqp(qp)->ibqp; in mlx4_ib_qp_event() local
211 if (ibqp->event_handler) { in mlx4_ib_qp_event()
213 event.element.qp = ibqp; in mlx4_ib_qp_event()
245 ibqp->event_handler(&event, ibqp->qp_context); in mlx4_ib_qp_event()
1407 qp->ibqp.srq ? to_msrq(qp->ibqp.srq): NULL); in destroy_qp_common()
2103 struct ib_qp *ibqp; in __mlx4_ib_modify_qp() local
3156 !qp->ibqp.qp_num ? in build_mlx_header()
3167 if (!qp->ibqp.qp_num) in build_mlx_header()
3980 if (ibqp->rwq_ind_tbl) in mlx4_ib_query_qp()
4010 if (qp->ibqp.qp_type == IB_QPT_RC || qp->ibqp.qp_type == IB_QPT_UC || in mlx4_ib_query_qp()
[all …]
/linux/drivers/infiniband/hw/hns/
A Dhns_roce_hw_v1.c878 hr_qp->ibqp.pd = pd; in hns_roce_v1_rsv_lp_qp()
2691 ibqp->srq ? to_hr_srq(ibqp->srq) : NULL); in hns_roce_v1_m_sqp()
2692 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_sqp()
2825 if (ibqp->srq) in hns_roce_v1_m_qp()
2890 if (ibqp->srq) in hns_roce_v1_m_qp()
3291 if (ibqp->uobject) { in hns_roce_v1_m_qp()
3311 ibqp->srq ? to_hr_srq(ibqp->srq) : NULL); in hns_roce_v1_m_qp()
3312 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_qp()
3334 if (ibqp->qp_type == IB_QPT_GSI || ibqp->qp_type == IB_QPT_SMI) in hns_roce_v1_modify_qp()
3619 send_cq = hr_qp->ibqp.send_cq ? to_hr_cq(hr_qp->ibqp.send_cq) : NULL; in hns_roce_v1_destroy_qp()
[all …]
A Dhns_roce_hw_v2.c353 struct ib_qp *ibqp = &hr_qp->ibqp; in check_send_valid() local
359 ibqp->qp_type); in check_send_valid()
775 struct ib_qp *ibqp = &hr_qp->ibqp; in check_recv_valid() local
4125 if (ibqp->qp_type != IB_QPT_UD && ibqp->qp_type != IB_QPT_GSI) in modify_qp_reset_to_init()
4131 if (ibqp->srq) { in modify_qp_reset_to_init()
4170 if (ibqp->srq) { in modify_qp_init_to_init()
4310 if (ibqp->qp_type == IB_QPT_GSI || ibqp->qp_type == IB_QPT_UD) in get_mtu()
4957 struct ib_qp *ibqp = &hr_qp->ibqp; in clear_qp() local
4963 if (ibqp->recv_cq && ibqp->recv_cq != ibqp->send_cq) in clear_qp()
4995 if (ibqp->srq || ibqp->qp_type == IB_QPT_XRC_INI) /* no RQ */ in v2_set_flushed_fields()
[all …]
A Dhns_roce_qp.c133 struct ib_qp *ibqp = &hr_qp->ibqp; in hns_roce_ib_qp_event() local
136 if (ibqp->event_handler) { in hns_roce_ib_qp_event()
137 event.device = ibqp->device; in hns_roce_ib_qp_event()
138 event.element.qp = ibqp; in hns_roce_ib_qp_event()
171 ibqp->event_handler(&event, ibqp->qp_context); in hns_roce_ib_qp_event()
221 if (hr_qp->ibqp.qp_type == IB_QPT_GSI) { in alloc_qpn()
425 if (hr_qp->ibqp.qp_type == IB_QPT_GSI) in free_qpn()
507 hr_qp->ibqp.qp_type != IB_QPT_GSI) in set_rq_size()
521 if (qp->ibqp.qp_type == IB_QPT_GSI || qp->ibqp.qp_type == IB_QPT_UD) in get_wqe_ext_sge_cnt()
1115 hr_qp->ibqp.qp_num = hr_qp->qpn; in hns_roce_create_qp_common()
[all …]
/linux/drivers/infiniband/hw/usnic/
A Dusnic_ib_qp_grp.h49 struct ib_qp ibqp; member
105 struct usnic_ib_qp_grp *to_uqp_grp(struct ib_qp *ibqp) in to_uqp_grp() argument
107 return container_of(ibqp, struct usnic_ib_qp_grp, ibqp); in to_uqp_grp()
/linux/drivers/infiniband/sw/rxe/
A Drxe_verbs.c410 struct rxe_dev *rxe = to_rdev(ibqp->device); in rxe_create_qp()
411 struct rxe_pd *pd = to_rpd(ibqp->pd); in rxe_create_qp()
412 struct rxe_qp *qp = to_rqp(ibqp); in rxe_create_qp()
459 struct rxe_qp *qp = to_rqp(ibqp); in rxe_modify_qp()
481 struct rxe_qp *qp = to_rqp(ibqp); in rxe_query_qp()
491 struct rxe_qp *qp = to_rqp(ibqp); in rxe_destroy_qp()
712 struct rxe_qp *qp = to_rqp(ibqp); in rxe_post_send()
736 struct rxe_qp *qp = to_rqp(ibqp); in rxe_post_recv()
1006 struct rxe_qp *qp = to_rqp(ibqp); in rxe_attach_mcast()
1023 struct rxe_qp *qp = to_rqp(ibqp); in rxe_detach_mcast()
[all …]
A Drxe_comp.c216 struct rxe_dev *rxe = to_rdev(qp->ibqp.device); in check_ack()
391 wc->qp = &qp->ibqp; in make_send_cqe()
395 uwc->qp_num = qp->ibqp.qp_num; in make_send_cqe()
425 struct rxe_dev *rxe = to_rdev(qp->ibqp.device); in do_complete()
481 if (qp->ibqp.event_handler) { in complete_ack()
484 ev.device = qp->ibqp.device; in complete_ack()
485 ev.element.qp = &qp->ibqp; in complete_ack()
487 qp->ibqp.event_handler(&ev, in complete_ack()
488 qp->ibqp.qp_context); in complete_ack()
533 ib_device_put(qp->ibqp.device); in rxe_drain_resp_pkts()
[all …]
/linux/include/rdma/
A Drdmavt_qp.h333 struct ib_qp ibqp; member
464 static inline struct rvt_qp *ibqp_to_rvtqp(struct ib_qp *ibqp) in ibqp_to_rvtqp() argument
466 return container_of(ibqp, struct rvt_qp, ibqp); in ibqp_to_rvtqp()
711 if (qp->ibqp.qp_num == qpn) in rvt_lookup_qpn()
725 struct ib_qp *ibqp = &qp->ibqp; in rvt_mod_retry_timer_ext() local
726 struct rvt_dev_info *rdi = ib_to_rvt(ibqp->device); in rvt_mod_retry_timer_ext()
785 struct rvt_cq *cq = ibcq_to_rvtcq(qp->ibqp.recv_cq); in rvt_recv_cq()
805 struct rvt_cq *cq = ibcq_to_rvtcq(qp->ibqp.send_cq); in rvt_send_cq()
859 .qp = &qp->ibqp, in rvt_qp_complete_swqe()
968 struct rvt_dev_info *rdi = ib_to_rvt(qp->ibqp.device); in rvt_to_iport()
/linux/drivers/infiniband/hw/cxgb4/
A Dev.c109 event.element.qp = &qhp->ibqp; in post_qp_event()
110 if (qhp->ibqp.event_handler) in post_qp_event()
111 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context); in post_qp_event()
153 c4iw_qp_add_ref(&qhp->ibqp); in c4iw_ev_dispatch()
217 c4iw_qp_rem_ref(&qhp->ibqp); in c4iw_ev_dispatch()
/linux/drivers/infiniband/hw/efa/
A Defa.h102 struct ib_qp ibqp; member
139 int efa_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr,
148 int efa_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata);
149 int efa_create_qp(struct ib_qp *ibqp, struct ib_qp_init_attr *init_attr,
173 int efa_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr,

Completed in 81 milliseconds

12345