/linux/drivers/infiniband/hw/mlx5/ |
A D | wr.h | 44 int mlx5_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, 46 int mlx5_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, 49 static inline int mlx5_ib_post_send_nodrain(struct ib_qp *ibqp, in mlx5_ib_post_send_nodrain() argument 53 return mlx5_ib_post_send(ibqp, wr, bad_wr, false); in mlx5_ib_post_send_nodrain() 56 static inline int mlx5_ib_post_send_drain(struct ib_qp *ibqp, in mlx5_ib_post_send_drain() argument 60 return mlx5_ib_post_send(ibqp, wr, bad_wr, true); in mlx5_ib_post_send_drain() 63 static inline int mlx5_ib_post_recv_nodrain(struct ib_qp *ibqp, in mlx5_ib_post_recv_nodrain() argument 67 return mlx5_ib_post_recv(ibqp, wr, bad_wr, false); in mlx5_ib_post_recv_nodrain() 70 static inline int mlx5_ib_post_recv_drain(struct ib_qp *ibqp, in mlx5_ib_post_recv_drain() argument 74 return mlx5_ib_post_recv(ibqp, wr, bad_wr, true); in mlx5_ib_post_recv_drain()
|
/linux/drivers/infiniband/hw/hfi1/ |
A D | qp.c | 163 struct ib_qp *ibqp = &qp->ibqp; in hfi1_check_modify_qp() local 225 struct ib_qp *ibqp = &qp->ibqp; in hfi1_modify_qp() local 270 switch (qp->ibqp.qp_type) { in hfi1_setup_wqe() 546 switch (qp->ibqp.qp_type) { in qp_to_sdma_engine() 568 switch (qp->ibqp.qp_type) { in qp_to_send_context() 618 qp->ibqp.qp_num, in qp_iter_print() 788 ev.device = qp->ibqp.device; in hfi1_migrate_qp() 789 ev.element.qp = &qp->ibqp; in hfi1_migrate_qp() 791 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in hfi1_migrate_qp() 904 ev.element.qp = &qp->ibqp; in hfi1_qp_iter_cb() [all …]
|
A D | ud.c | 58 IB_QPT_UD : qp->ibqp.qp_type; in ud_loopback() 69 if (qp->ibqp.qp_num > 1) { in ud_loopback() 82 sqp->ibqp.qp_num, qp->ibqp.qp_num, in ud_loopback() 93 if (qp->ibqp.qp_num) { in ud_loopback() 131 if (qp->ibqp.qp_num == 0) in ud_loopback() 196 wc.qp = &qp->ibqp; in ud_loopback() 198 if (qp->ibqp.qp_type == IB_QPT_GSI || qp->ibqp.qp_type == IB_QPT_SMI) { in ud_loopback() 242 if (qp->ibqp.qp_type == IB_QPT_GSI || qp->ibqp.qp_type == IB_QPT_SMI) in hfi1_make_bth_deth() 870 if (qp->ibqp.qp_num) { in hfi1_ud_rcv() 988 wc.qp = &qp->ibqp; in hfi1_ud_rcv() [all …]
|
A D | trace_tid.h | 208 __entry->qpn = qp->ibqp.qp_num; 266 __entry->qpn = qp->ibqp.qp_num; 318 __entry->qpn = qp->ibqp.qp_num; 430 __entry->qpn = qp->ibqp.qp_num; 468 __entry->qpn = qp->ibqp.qp_num; 515 __entry->qpn = qp->ibqp.qp_num; 653 __entry->qpn = qp->ibqp.qp_num; 695 __entry->qpn = qp->ibqp.qp_num; 769 __entry->qpn = qp->ibqp.qp_num; 890 __entry->qpn = qp->ibqp.qp_num; [all …]
|
A D | trace_rc.h | 21 DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device)) 32 DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device)); 33 __entry->qpn = qp->ibqp.qp_num; 83 DD_DEV_ENTRY(dd_from_ibdev(qp->ibqp.device)) 92 DD_DEV_ASSIGN(dd_from_ibdev(qp->ibqp.device)); 93 __entry->qpn = qp->ibqp.qp_num;
|
/linux/drivers/infiniband/sw/rdmavt/ |
A D | qp.c | 719 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_qp_mr_clean() 1318 wc.qp = &qp->ibqp; in rvt_error_qp() 1652 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_modify_qp() 1658 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_modify_qp() 1790 !qp->ibqp.srq; in rvt_post_recv() 2279 pd = ibpd_to_rvtpd(qp->ibqp.srq ? qp->ibqp.srq->pd : qp->ibqp.pd); in init_sge() 2461 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_comm_est() 2481 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in rvt_rc_error() 2511 struct ib_qp *ibqp = &qp->ibqp; in rvt_add_retry_timer_ext() local 2972 qp->ibqp.qp_type != sqp->ibqp.qp_type) { in rvt_ruc_loopback() [all …]
|
A D | qp.h | 13 int rvt_create_qp(struct ib_qp *ibqp, struct ib_qp_init_attr *init_attr, 15 int rvt_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, 17 int rvt_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata); 18 int rvt_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, 20 int rvt_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, 22 int rvt_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr,
|
A D | trace_qp.h | 21 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device)) 26 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device)); 27 __entry->qpn = qp->ibqp.qp_num; 51 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device)) 58 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device)); 59 __entry->qpn = qp->ibqp.qp_num;
|
A D | trace_tx.h | 52 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device)) 72 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device)); 75 __entry->qpn = qp->ibqp.qp_num; 76 __entry->qpt = qp->ibqp.qp_type; 119 RDI_DEV_ENTRY(ib_to_rvt(qp->ibqp.device)) 131 RDI_DEV_ASSIGN(ib_to_rvt(qp->ibqp.device)); 134 __entry->qpn = qp->ibqp.qp_num; 135 __entry->qpt = qp->ibqp.qp_type;
|
A D | mcast.c | 238 int rvt_attach_mcast(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in rvt_attach_mcast() argument 240 struct rvt_qp *qp = ibqp_to_rvtqp(ibqp); in rvt_attach_mcast() 241 struct rvt_dev_info *rdi = ib_to_rvt(ibqp->device); in rvt_attach_mcast() 247 if (ibqp->qp_num <= 1 || qp->state == IB_QPS_RESET) in rvt_attach_mcast() 301 int rvt_detach_mcast(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in rvt_detach_mcast() argument 303 struct rvt_qp *qp = ibqp_to_rvtqp(ibqp); in rvt_detach_mcast() 304 struct rvt_dev_info *rdi = ib_to_rvt(ibqp->device); in rvt_detach_mcast() 312 if (ibqp->qp_num <= 1) in rvt_detach_mcast()
|
/linux/drivers/infiniband/hw/qib/ |
A D | qib_ud.c | 87 if (qp->ibqp.qp_num > 1) { in qib_ud_loopback() 99 sqp->ibqp.qp_num, qp->ibqp.qp_num, in qib_ud_loopback() 111 if (qp->ibqp.qp_num) { in qib_ud_loopback() 149 if (qp->ibqp.qp_num == 0) in qib_ud_loopback() 204 wc.qp = &qp->ibqp; in qib_ud_loopback() 205 wc.src_qp = sqp->ibqp.qp_num; in qib_ud_loopback() 464 if (qp->ibqp.qp_num) { in qib_ud_rcv() 468 if (qp->ibqp.qp_num > 1) { in qib_ud_rcv() 509 if (qp->ibqp.qp_num > 1 && in qib_ud_rcv() 539 if (qp->ibqp.qp_num == 0) in qib_ud_rcv() [all …]
|
A D | qib_ruc.c | 53 ev.device = qp->ibqp.device; in qib_migrate_qp() 54 ev.element.qp = &qp->ibqp; in qib_migrate_qp() 56 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in qib_migrate_qp() 114 0, qp->ibqp.qp_num, in qib_ruc_check_hdr() 153 0, qp->ibqp.qp_num, in qib_ruc_check_hdr() 210 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_make_ruc_header() 266 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_do_send() 271 if ((qp->ibqp.qp_type == IB_QPT_RC || in qib_do_send() 272 qp->ibqp.qp_type == IB_QPT_UC) && in qib_do_send() 279 if (qp->ibqp.qp_type == IB_QPT_RC) in qib_do_send() [all …]
|
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
A D | pvrdma_qp.c | 61 *send_cq = to_vcq(qp->ibqp.send_cq); in get_cqs() 62 *recv_cq = to_vcq(qp->ibqp.recv_cq); in get_cqs() 194 struct pvrdma_qp *qp = to_vqp(ibqp); in pvrdma_create_qp() 394 qp->ibqp.qp_num = resp_v2->qpn; in pvrdma_create_qp() 397 qp->ibqp.qp_num = resp->qpn; in pvrdma_create_qp() 406 qp_resp.qpn = qp->ibqp.qp_num; in pvrdma_create_qp() 529 struct pvrdma_qp *qp = to_vqp(ibqp); in pvrdma_modify_qp() 674 struct pvrdma_qp *qp = to_vqp(ibqp); in pvrdma_post_send() 774 switch (qp->ibqp.qp_type) { in pvrdma_post_send() 886 struct pvrdma_qp *qp = to_vqp(ibqp); in pvrdma_post_recv() [all …]
|
/linux/drivers/infiniband/hw/mthca/ |
A D | mthca_qp.c | 261 event.element.qp = &qp->ibqp; in mthca_qp_event() 262 if (qp->ibqp.event_handler) in mthca_qp_event() 263 qp->ibqp.event_handler(&event, qp->ibqp.qp_context); in mthca_qp_event() 626 if (qp->ibqp.uobject) in __mthca_modify_qp() 759 if (ibqp->srq) in __mthca_modify_qp() 784 if (ibqp->srq) in __mthca_modify_qp() 836 qp->ibqp.srq ? to_msrq(qp->ibqp.srq) : NULL); in __mthca_modify_qp() 837 if (qp->ibqp.send_cq != qp->ibqp.recv_cq) in __mthca_modify_qp() 1479 if (!qp->ibqp.uobject) { in mthca_free_qp() 1481 qp->ibqp.srq ? to_msrq(qp->ibqp.srq) : NULL); in mthca_free_qp() [all …]
|
A D | mthca_mcg.c | 120 int mthca_multicast_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mthca_multicast_attach() argument 122 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_multicast_attach() 165 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) { in mthca_multicast_attach() 167 ibqp->qp_num); in mthca_multicast_attach() 171 mgm->qp[i] = cpu_to_be32(ibqp->qp_num | (1 << 31)); in mthca_multicast_attach() 214 int mthca_multicast_detach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid) in mthca_multicast_detach() argument 216 struct mthca_dev *dev = to_mdev(ibqp->device); in mthca_multicast_detach() 242 if (mgm->qp[i] == cpu_to_be32(ibqp->qp_num | (1 << 31))) in mthca_multicast_detach() 249 mthca_err(dev, "QP %06x not found in MGM\n", ibqp->qp_num); in mthca_multicast_detach()
|
/linux/drivers/infiniband/hw/mlx4/ |
A D | qp.c | 206 struct ib_qp *ibqp = &to_mibqp(qp)->ibqp; in mlx4_ib_qp_event() local 211 if (ibqp->event_handler) { in mlx4_ib_qp_event() 213 event.element.qp = ibqp; in mlx4_ib_qp_event() 245 ibqp->event_handler(&event, ibqp->qp_context); in mlx4_ib_qp_event() 1407 qp->ibqp.srq ? to_msrq(qp->ibqp.srq): NULL); in destroy_qp_common() 2103 struct ib_qp *ibqp; in __mlx4_ib_modify_qp() local 3156 !qp->ibqp.qp_num ? in build_mlx_header() 3167 if (!qp->ibqp.qp_num) in build_mlx_header() 3980 if (ibqp->rwq_ind_tbl) in mlx4_ib_query_qp() 4010 if (qp->ibqp.qp_type == IB_QPT_RC || qp->ibqp.qp_type == IB_QPT_UC || in mlx4_ib_query_qp() [all …]
|
/linux/drivers/infiniband/hw/hns/ |
A D | hns_roce_hw_v1.c | 878 hr_qp->ibqp.pd = pd; in hns_roce_v1_rsv_lp_qp() 2691 ibqp->srq ? to_hr_srq(ibqp->srq) : NULL); in hns_roce_v1_m_sqp() 2692 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_sqp() 2825 if (ibqp->srq) in hns_roce_v1_m_qp() 2890 if (ibqp->srq) in hns_roce_v1_m_qp() 3291 if (ibqp->uobject) { in hns_roce_v1_m_qp() 3311 ibqp->srq ? to_hr_srq(ibqp->srq) : NULL); in hns_roce_v1_m_qp() 3312 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_qp() 3334 if (ibqp->qp_type == IB_QPT_GSI || ibqp->qp_type == IB_QPT_SMI) in hns_roce_v1_modify_qp() 3619 send_cq = hr_qp->ibqp.send_cq ? to_hr_cq(hr_qp->ibqp.send_cq) : NULL; in hns_roce_v1_destroy_qp() [all …]
|
A D | hns_roce_hw_v2.c | 353 struct ib_qp *ibqp = &hr_qp->ibqp; in check_send_valid() local 359 ibqp->qp_type); in check_send_valid() 775 struct ib_qp *ibqp = &hr_qp->ibqp; in check_recv_valid() local 4125 if (ibqp->qp_type != IB_QPT_UD && ibqp->qp_type != IB_QPT_GSI) in modify_qp_reset_to_init() 4131 if (ibqp->srq) { in modify_qp_reset_to_init() 4170 if (ibqp->srq) { in modify_qp_init_to_init() 4310 if (ibqp->qp_type == IB_QPT_GSI || ibqp->qp_type == IB_QPT_UD) in get_mtu() 4957 struct ib_qp *ibqp = &hr_qp->ibqp; in clear_qp() local 4963 if (ibqp->recv_cq && ibqp->recv_cq != ibqp->send_cq) in clear_qp() 4995 if (ibqp->srq || ibqp->qp_type == IB_QPT_XRC_INI) /* no RQ */ in v2_set_flushed_fields() [all …]
|
A D | hns_roce_qp.c | 133 struct ib_qp *ibqp = &hr_qp->ibqp; in hns_roce_ib_qp_event() local 136 if (ibqp->event_handler) { in hns_roce_ib_qp_event() 137 event.device = ibqp->device; in hns_roce_ib_qp_event() 138 event.element.qp = ibqp; in hns_roce_ib_qp_event() 171 ibqp->event_handler(&event, ibqp->qp_context); in hns_roce_ib_qp_event() 221 if (hr_qp->ibqp.qp_type == IB_QPT_GSI) { in alloc_qpn() 425 if (hr_qp->ibqp.qp_type == IB_QPT_GSI) in free_qpn() 507 hr_qp->ibqp.qp_type != IB_QPT_GSI) in set_rq_size() 521 if (qp->ibqp.qp_type == IB_QPT_GSI || qp->ibqp.qp_type == IB_QPT_UD) in get_wqe_ext_sge_cnt() 1115 hr_qp->ibqp.qp_num = hr_qp->qpn; in hns_roce_create_qp_common() [all …]
|
/linux/drivers/infiniband/hw/usnic/ |
A D | usnic_ib_qp_grp.h | 49 struct ib_qp ibqp; member 105 struct usnic_ib_qp_grp *to_uqp_grp(struct ib_qp *ibqp) in to_uqp_grp() argument 107 return container_of(ibqp, struct usnic_ib_qp_grp, ibqp); in to_uqp_grp()
|
/linux/drivers/infiniband/sw/rxe/ |
A D | rxe_verbs.c | 410 struct rxe_dev *rxe = to_rdev(ibqp->device); in rxe_create_qp() 411 struct rxe_pd *pd = to_rpd(ibqp->pd); in rxe_create_qp() 412 struct rxe_qp *qp = to_rqp(ibqp); in rxe_create_qp() 459 struct rxe_qp *qp = to_rqp(ibqp); in rxe_modify_qp() 481 struct rxe_qp *qp = to_rqp(ibqp); in rxe_query_qp() 491 struct rxe_qp *qp = to_rqp(ibqp); in rxe_destroy_qp() 712 struct rxe_qp *qp = to_rqp(ibqp); in rxe_post_send() 736 struct rxe_qp *qp = to_rqp(ibqp); in rxe_post_recv() 1006 struct rxe_qp *qp = to_rqp(ibqp); in rxe_attach_mcast() 1023 struct rxe_qp *qp = to_rqp(ibqp); in rxe_detach_mcast() [all …]
|
A D | rxe_comp.c | 216 struct rxe_dev *rxe = to_rdev(qp->ibqp.device); in check_ack() 391 wc->qp = &qp->ibqp; in make_send_cqe() 395 uwc->qp_num = qp->ibqp.qp_num; in make_send_cqe() 425 struct rxe_dev *rxe = to_rdev(qp->ibqp.device); in do_complete() 481 if (qp->ibqp.event_handler) { in complete_ack() 484 ev.device = qp->ibqp.device; in complete_ack() 485 ev.element.qp = &qp->ibqp; in complete_ack() 487 qp->ibqp.event_handler(&ev, in complete_ack() 488 qp->ibqp.qp_context); in complete_ack() 533 ib_device_put(qp->ibqp.device); in rxe_drain_resp_pkts() [all …]
|
/linux/include/rdma/ |
A D | rdmavt_qp.h | 333 struct ib_qp ibqp; member 464 static inline struct rvt_qp *ibqp_to_rvtqp(struct ib_qp *ibqp) in ibqp_to_rvtqp() argument 466 return container_of(ibqp, struct rvt_qp, ibqp); in ibqp_to_rvtqp() 711 if (qp->ibqp.qp_num == qpn) in rvt_lookup_qpn() 725 struct ib_qp *ibqp = &qp->ibqp; in rvt_mod_retry_timer_ext() local 726 struct rvt_dev_info *rdi = ib_to_rvt(ibqp->device); in rvt_mod_retry_timer_ext() 785 struct rvt_cq *cq = ibcq_to_rvtcq(qp->ibqp.recv_cq); in rvt_recv_cq() 805 struct rvt_cq *cq = ibcq_to_rvtcq(qp->ibqp.send_cq); in rvt_send_cq() 859 .qp = &qp->ibqp, in rvt_qp_complete_swqe() 968 struct rvt_dev_info *rdi = ib_to_rvt(qp->ibqp.device); in rvt_to_iport()
|
/linux/drivers/infiniband/hw/cxgb4/ |
A D | ev.c | 109 event.element.qp = &qhp->ibqp; in post_qp_event() 110 if (qhp->ibqp.event_handler) in post_qp_event() 111 (*qhp->ibqp.event_handler)(&event, qhp->ibqp.qp_context); in post_qp_event() 153 c4iw_qp_add_ref(&qhp->ibqp); in c4iw_ev_dispatch() 217 c4iw_qp_rem_ref(&qhp->ibqp); in c4iw_ev_dispatch()
|
/linux/drivers/infiniband/hw/efa/ |
A D | efa.h | 102 struct ib_qp ibqp; member 139 int efa_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, 148 int efa_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata); 149 int efa_create_qp(struct ib_qp *ibqp, struct ib_qp_init_attr *init_attr, 173 int efa_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr,
|