Home
last modified time | relevance | path

Searched refs:qp (Results 1 – 25 of 347) sorted by relevance

12345678910>>...14

/linux/drivers/infiniband/hw/qib/
A Dqib_rc.c253 rvt_send_complete(qp, wqe, qp->s_last != qp->s_acked ? in qib_make_rc_req()
474 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in qib_make_rc_req()
499 if (qp->s_cur >= qp->s_size) in qib_make_rc_req()
513 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in qib_make_rc_req()
538 if (qp->s_cur >= qp->s_size) in qib_make_rc_req()
980 qp->s_acked = qp->s_cur; in do_rc_completion()
1881 wc.qp = &qp->ibqp; in qib_rc_rcv()
2084 qp->r_ack_psn = qp->r_psn; in qib_rc_rcv()
2096 qp->r_ack_psn = qp->r_psn; in qib_rc_rcv()
2110 qp->r_ack_psn = qp->r_psn; in qib_rc_rcv()
[all …]
A Dqib_uc.c71 wqe = rvt_get_swqe_ptr(qp, qp->s_last); in qib_make_uc_req()
85 wqe = rvt_get_swqe_ptr(qp, qp->s_cur); in qib_make_uc_req()
125 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req()
154 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req()
183 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req()
208 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req()
214 qp->s_cur_sge = &qp->s_sge; in qib_make_uc_req()
337 qp->r_sge = qp->s_rdma_read_sge; in qib_uc_rcv()
348 qp->s_rdma_read_sge = qp->r_sge; in qib_uc_rcv()
394 wc.qp = &qp->ibqp; in qib_uc_rcv()
[all …]
A Dqib_ruc.c49 qp->remote_ah_attr = qp->alt_ah_attr; in qib_migrate_qp()
50 qp->port_num = rdma_ah_get_port_num(&qp->alt_ah_attr); in qib_migrate_qp()
51 qp->s_pkey_index = qp->s_alt_pkey_index; in qib_migrate_qp()
54 ev.element.qp = &qp->ibqp; in qib_migrate_qp()
56 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in qib_migrate_qp()
210 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_make_ruc_header()
220 qp->s_hdrwords += in qib_make_ruc_header()
252 qib_do_send(qp); in _qib_do_send()
266 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_do_send()
304 if (qib_verbs_send(qp, priv->s_hdr, qp->s_hdrwords, in qib_do_send()
[all …]
A Dqib_ud.c68 if (!qp) { in qib_ud_loopback()
166 rvt_copy_sge(qp, &qp->r_sge, &grh, in qib_ud_loopback()
204 wc.qp = &qp->ibqp; in qib_ud_loopback()
251 if (qp->s_last == READ_ONCE(qp->s_head)) in qib_make_ud_req()
258 wqe = rvt_get_swqe_ptr(qp, qp->s_last); in qib_make_ud_req()
264 if (qp->s_cur == READ_ONCE(qp->s_head)) in qib_make_ud_req()
267 wqe = rvt_get_swqe_ptr(qp, qp->s_cur); in qib_make_ud_req()
273 ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_make_ud_req()
315 qp->s_cur_sge = &qp->s_sge; in qib_make_ud_req()
550 rvt_copy_sge(qp, &qp->r_sge, &hdr->u.l.grh, in qib_ud_rcv()
[all …]
/linux/drivers/infiniband/sw/rxe/
A Drxe_qp.c168 qp->mtu = ib_mtu_enum_to_int(qp->attr.path_mtu); in rxe_qp_init_misc()
211 qp->sk->sk->sk_user_data = qp; in rxe_qp_init_req()
239 qp->sq.queue->buf, qp->sq.queue->buf_size, in rxe_qp_init_req()
259 rxe_init_task(rxe, &qp->req.task, qp, in rxe_qp_init_req()
261 rxe_init_task(rxe, &qp->comp.task, qp, in rxe_qp_init_req()
288 qp_num(qp), qp->rq.max_wr, qp->rq.max_sge, wqe_size); in rxe_qp_init_resp()
297 qp->rq.queue->buf, qp->rq.queue->buf_size, in rxe_qp_init_resp()
312 rxe_init_task(rxe, &qp->resp.task, qp, in rxe_qp_init_resp()
669 qp->resp.psn = qp->attr.rq_psn; in rxe_qp_from_attr()
682 qp->req.psn = qp->attr.sq_psn; in rxe_qp_from_attr()
[all …]
A Drxe_resp.c321 qp->resp.wqe = &qp->resp.srq_wqe.wqe; in get_srq_wqe()
352 qp->resp.wqe = queue_head(qp->rq.queue, in check_resource()
380 qp->resp.wqe = queue_head(qp->rq.queue, in check_resource()
538 err = rxe_mr_copy(qp->resp.mr, qp->resp.va + qp->resp.offset, in write_data_in()
618 ack->qp = qp; in prepare_ack_packet()
663 res = &qp->resp.resources[qp->resp.res_head]; in read_reply()
819 qp->resp.ack_psn = qp->resp.psn; in execute()
854 wc->qp = &qp->ibqp; in do_complete()
976 res = &qp->resp.resources[qp->resp.res_head]; in send_atomic_ack()
1003 send_ack(qp, pkt, qp->resp.aeth_syndrome, pkt->psn); in acknowledge()
[all …]
A Drxe_comp.c115 struct rxe_qp *qp = from_timer(qp, t, retrans_timer); in retransmit_timer() local
117 if (qp->valid) { in retransmit_timer()
166 qp->comp.retry_cnt = qp->attr.retry_cnt; in reset_retry_counters()
167 qp->comp.rnr_retry = qp->attr.rnr_retry; in reset_retry_counters()
391 wc->qp = &qp->ibqp; in make_send_cqe()
477 (qp->comp.psn == qp->req.psn)) { in complete_ack()
485 ev.element.qp = &qp->ibqp; in complete_ack()
549 struct rxe_qp *qp = pkt->qp; in free_pkt() local
569 if (!qp->valid || qp->req.state == QP_STATE_ERROR || in rxe_completer()
571 rxe_drain_resp_pkts(qp, qp->valid && in rxe_completer()
[all …]
A Drxe_req.c27 qp->req.opcode = next_opcode(qp, wqe, in retry_first_write_send()
56 qp->req.psn = qp->comp.psn; in req_retry()
93 qp->mtu; in req_retry()
104 struct rxe_qp *qp = from_timer(qp, t, rnr_nak_timer); in rnr_nak_timer() local
151 ev.element.qp = &qp->ibqp; in req_next_wqe()
385 pkt->qp = qp; in init_req_packet()
523 qp->req.psn = (qp->req.psn + 1) & BTH_PSN_MASK; in update_wqe_psn()
554 qp->req.wqe_index = queue_next_index(qp->sq.queue, in update_state()
605 qp->req.wqe_index = queue_next_index(qp->sq.queue, qp->req.wqe_index); in rxe_do_local_ops()
663 psn_compare(qp->req.psn, (qp->comp.psn + in rxe_requester()
[all …]
/linux/drivers/infiniband/hw/hfi1/
A Drc.c906 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in hfi1_make_rc_req()
937 if (qp->s_cur >= qp->s_size) in hfi1_make_rc_req()
951 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in hfi1_make_rc_req()
977 if (qp->s_cur >= qp->s_size) in hfi1_make_rc_req()
1885 qp->s_acked = qp->s_cur; in do_rc_completion()
1894 if (qp->state == IB_QPS_SQD && qp->s_acked == qp->s_cur) in do_rc_completion()
2945 wc.qp = &qp->ibqp; in hfi1_rc_rcv()
3176 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv()
3184 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv()
3194 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv()
[all …]
A Dqp.c449 qp = tx->qp; in iowait_sleep()
587 qp->s_last == qp->s_acked && in qp_idle()
588 qp->s_acked == qp->s_cur && in qp_idle()
589 qp->s_cur == qp->s_tail && in qp_idle()
590 qp->s_tail == qp->s_head; in qp_idle()
601 struct rvt_qp *qp = iter->qp; in qp_iter_print() local
631 qp->s_psn, qp->s_next_psn, in qp_iter_print()
634 qp->s_last, qp->s_acked, qp->s_cur, in qp_iter_print()
635 qp->s_tail, qp->s_head, qp->s_size, in qp_iter_print()
789 ev.element.qp = &qp->ibqp; in hfi1_migrate_qp()
[all …]
A Duc.c49 wqe = rvt_get_swqe_ptr(qp, qp->s_last); in hfi1_make_uc_req()
72 wqe = rvt_get_swqe_ptr(qp, qp->s_cur); in hfi1_make_uc_req()
93 if (qp->s_last != qp->s_cur) in hfi1_make_uc_req()
95 if (++qp->s_cur == qp->s_size) in hfi1_make_uc_req()
138 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req()
167 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req()
197 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req()
223 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req()
264 struct rvt_qp *qp = packet->qp; in hfi1_uc_rcv() local
412 wc.qp = &qp->ibqp; in hfi1_uc_rcv()
[all …]
A Dtid_rdma.c1669 req->qp = qp; in hfi1_init_trdma_req()
1709 struct rvt_qp *qp = req->qp; in hfi1_build_tid_rdma_read_packet() local
2006 qp->r_ack_psn = qp->r_psn; in tid_rdma_rcv_error()
2333 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_read_req()
2341 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_read_req()
2892 packet->qp = qp; in hfi1_handle_kdeth_eflags()
3217 s_prev = (qp->s_cur == 0 ? qp->s_size : qp->s_cur) - 1; in hfi1_tid_rdma_wqe_interlock()
3603 qp->r_ack_psn = qp->r_psn; in hfi1_tid_write_alloc_resources()
3824 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_write_req()
3832 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_write_req()
[all …]
A Dtrace_tid.h196 TP_ARGS(qp),
226 TP_ARGS(qp)
232 TP_ARGS(qp)
238 TP_ARGS(qp)
244 TP_ARGS(qp)
250 TP_ARGS(qp)
360 __entry->qpn = qp ? qp->ibqp.qp_num : 0;
931 TP_ARGS(qp)
937 TP_ARGS(qp)
943 TP_ARGS(qp)
[all …]
A Druc.c31 struct rvt_qp *qp = packet->qp; in hfi1_ruc_check_hdr() local
74 hfi1_migrate_qp(qp); in hfi1_ruc_check_hdr()
168 clear_ahg(qp); in build_ahg()
171 if (qp->s_ahgidx < 0) in build_ahg()
174 qp->s_ahgpsn = npsn; in build_ahg()
282 build_ahg(qp, bth2); in hfi1_make_ruc_header_16B()
360 build_ahg(qp, bth2); in hfi1_make_ruc_header_9B()
481 hfi1_do_send(qp, true); in _hfi1_do_send()
500 ps.ibp = to_iport(qp->ibqp.device, qp->port_num); in hfi1_do_send()
512 rvt_ruc_loopback(qp); in hfi1_do_send()
[all …]
/linux/drivers/infiniband/sw/rdmavt/
A Dqp.c717 ev.element.qp = &qp->ibqp; in rvt_qp_mr_clean()
1318 wc.qp = &qp->ibqp; in rvt_error_qp()
1383 rcu_assign_pointer(rvp->qp[qp->ibqp.qp_num], qp); in rvt_insert_qp()
1544 qp->s_draining = qp->s_last != qp->s_cur; in rvt_modify_qp()
1650 ev.element.qp = &qp->ibqp; in rvt_modify_qp()
1821 wc.qp = &qp->ibqp; in rvt_post_recv()
1950 qp->ibqp.qp_num, qp->s_size, qp->s_avail, in rvt_qp_is_avail()
1951 qp->s_head, qp->s_tail, qp->s_cur, in rvt_qp_is_avail()
2310 wc.qp = &qp->ibqp; in init_sge()
2725 iter->qp = qp; in rvt_qp_iter_next()
[all …]
/linux/drivers/ntb/
A Dntb_transport.c472 if (!qp || !qp->link_is_up) in debugfs_read()
661 entry->qp = qp; in ntb_transport_setup_qp_mw()
946 qp->event_handler(qp->cb_data, qp->link_is_up); in ntb_qp_link_cleanup()
1147 qp->event_handler(qp->cb_data, qp->link_is_up); in ntb_qp_link_work()
1217 qp->debugfs_dir, qp, in ntb_transport_init_queue()
1463 qp->rx_handler(qp, qp->cb_data, cb_data, len); in ntb_complete_rxc()
1620 offset = qp->rx_buff + qp->rx_max_frame * qp->rx_index; in ntb_process_rxc()
1774 qp->tx_handler(qp, qp->cb_data, entry->cb_data, in ntb_tx_copy_callback()
1904 qp->tx_handler(qp, qp->cb_data, NULL, -EIO); in ntb_process_tx()
2056 entry->qp = qp; in ntb_transport_create_queue()
[all …]
/linux/drivers/infiniband/sw/siw/
A Dsiw_qp.c168 if (qp->cep) { in siw_qp_llp_close()
170 qp->cep = NULL; in siw_qp_llp_close()
368 memset(&qp->term_info, 0, sizeof(qp->term_info)); in siw_init_terminate()
685 qp->attrs.orq_size, qp->attrs.irq_size); in siw_qp_nextstate_from_idle()
926 if (qp->attrs.orq_size && !siw_orq_empty(qp)) { in siw_activate_tx_from_sq()
962 qp->sq_get++; in siw_activate_tx_from_sq()
985 irqe = &qp->irq[qp->irq_get % qp->attrs.irq_size]; in siw_activate_tx()
1198 sqe = &qp->orq[qp->orq_get % qp->attrs.orq_size]; in siw_sq_flush()
1234 sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_flush()
1298 &qp->recvq[qp->rq_get % qp->attrs.rq_size]; in siw_rq_flush()
[all …]
/linux/drivers/infiniband/hw/mthca/
A Dmthca_qp.c257 qp->port = qp->alt_port; in mthca_qp_event()
261 event.element.qp = &qp->ibqp; in mthca_qp_event()
841 qp->sq.last = get_send_wqe(qp, qp->sq.max - 1); in __mthca_modify_qp()
844 qp->rq.last = get_recv_wqe(qp, qp->rq.max - 1); in __mthca_modify_qp()
1052 qp->send_wqe_offset = ALIGN(qp->rq.max << qp->rq.wqe_shift, in mthca_alloc_wqe_buf()
1072 &qp->queue, &qp->is_direct, pd, 0, &qp->mr); in mthca_alloc_wqe_buf()
1088 &qp->queue, qp->is_direct, &qp->mr); in mthca_free_wqe_buf()
1138 qp->qpn, &qp->rq.db); in mthca_alloc_memfree()
1143 qp->qpn, &qp->sq.db); in mthca_alloc_memfree()
1246 qp->sq.last = get_send_wqe(qp, qp->sq.max - 1); in mthca_alloc_qp_common()
[all …]
/linux/drivers/net/ethernet/qlogic/qed/
A Dqed_roce.c210 if (qp->vlan_id) { in qed_roce_get_qp_tc()
217 qp->icid, tc, qp->vlan_id ? "enabled" : "disabled"); in qed_roce_get_qp_tc()
244 if (!qp->irq) { in qed_roce_sp_create_responder()
343 qp->cq_prod = 0; in qed_roce_sp_create_responder()
355 qp->irq, qp->irq_phys_addr); in qed_roce_sp_create_responder()
382 if (!qp->orq) { in qed_roce_sp_create_requester()
486 qp->orq, qp->orq_phys_addr); in qed_roce_sp_create_requester()
736 qp->irq, qp->irq_phys_addr); in qed_roce_sp_destroy_qp_responder()
798 qp->orq, qp->orq_phys_addr); in qed_roce_sp_destroy_qp_requester()
828 if ((!(qp->resp_offloaded)) && (!(qp->req_offloaded))) { in qed_roce_query_qp()
[all …]
/linux/drivers/infiniband/hw/irdma/
A Duk.c130 writel(qp->qp_id, qp->wqe_alloc_db); in irdma_uk_qp_post_wr()
136 writel(qp->qp_id, qp->wqe_alloc_db); in irdma_uk_qp_post_wr()
140 writel(qp->qp_id, qp->wqe_alloc_db); in irdma_uk_qp_post_wr()
145 qp->initial_ring.head = qp->sq_ring.head; in irdma_uk_qp_post_wr()
157 qp->initial_ring.head = qp->sq_ring.head; in irdma_qp_ring_push_db()
217 irdma_qp_push_wqe(qp, qp->sq_base[nop_wqe_idx].elem, in irdma_qp_get_next_send_wqe()
223 qp->swqe_polarity = !qp->swqe_polarity; in irdma_qp_get_next_send_wqe()
258 qp->rwqe_polarity = !qp->rwqe_polarity; in irdma_qp_get_next_recv_wqe()
1114 if (!qp || qp->destroy_pending) { in irdma_uk_cq_poll_cmpl()
1165 if (wqe_idx < qp->conn_wqes && qp->sq_ring.head == qp->sq_ring.tail) { in irdma_uk_cq_poll_cmpl()
[all …]
/linux/drivers/infiniband/hw/mlx4/
A Dqp.c173 return get_wqe(qp, qp->rq.offset + (n << qp->rq.wqe_shift)); in get_recv_wqe()
178 return get_wqe(qp, qp->sq.offset + (n << qp->sq.wqe_shift)); in get_send_wqe()
390 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in set_kernel_sq_size()
394 qp->sq.offset = qp->rq.wqe_cnt << qp->rq.wqe_shift; in set_kernel_sq_size()
396 qp->rq.offset = qp->sq.wqe_cnt << qp->sq.wqe_shift; in set_kernel_sq_size()
425 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in set_user_sq_size()
735 qp->ibqp.qp_num = qp->mqp.qpn; in _mlx4_ib_create_qp_rss()
879 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in create_rq()
1373 if (qp->pri.smac || (!qp->pri.smac && qp->pri.smac_port)) { in destroy_qp_common()
2555 if (qp->pri.smac || (!qp->pri.smac && qp->pri.smac_port)) { in __mlx4_ib_modify_qp()
[all …]
/linux/drivers/infiniband/hw/vmw_pvrdma/
A Dpvrdma_qp.c150 qp->npages_recv = (qp->rq.wqe_cnt * qp->rq.wqe_size + PAGE_SIZE - 1) / in pvrdma_set_rq_size()
300 qp->npages = qp->npages_send + qp->npages_recv; in pvrdma_create_qp()
312 qp->npages = qp->npages_send + qp->npages_recv; in pvrdma_create_qp()
318 qp->rq.offset = qp->npages_send * PAGE_SIZE; in pvrdma_create_qp()
344 qp->sq.ring = qp->pdir.pages[0]; in pvrdma_create_qp()
345 qp->rq.ring = is_srq ? NULL : &qp->sq.ring[1]; in pvrdma_create_qp()
542 qp->state; in pvrdma_modify_qp()
637 qp->sq.offset + n * qp->sq.wqe_size); in get_sq_wqe()
643 qp->rq.offset + n * qp->rq.wqe_size); in get_rq_wqe()
696 qp->sq.ring, qp->sq.wqe_cnt, &tail))) { in pvrdma_post_send()
[all …]
/linux/drivers/infiniband/hw/bnxt_re/
A Dqplib_fp.c96 if (qp->scq == qp->rcq) in bnxt_qplib_acquire_cq_flush_locks()
106 if (qp->scq == qp->rcq) in bnxt_qplib_release_cq_flush_locks()
141 __clean_cq(qp->scq, (u64)(unsigned long)qp); in bnxt_qplib_clean_qp()
144 __clean_cq(qp->rcq, (u64)(unsigned long)qp); in bnxt_qplib_clean_qp()
182 qp->rq_hdr_buf, qp->rq_hdr_buf_map); in bnxt_qplib_free_qp_hdr_buf()
186 qp->sq_hdr_buf, qp->sq_hdr_buf_map); in bnxt_qplib_free_qp_hdr_buf()
1332 qp->cur_qp_state = qp->state; in bnxt_qplib_modify_qp()
1406 qp->sq.max_wqe = qp->sq.hwq.max_elements; in bnxt_qplib_query_qp()
1724 qp->id, qp->state); in bnxt_qplib_post_send()
1972 qp->id, qp->state); in bnxt_qplib_post_recv()
[all …]
/linux/net/ipv4/
A Dip_fragment.c90 qp->ecn = 0; in ip4_frag_init()
98 struct ipq *qp; in ip4_frag_free() local
101 if (qp->peer) in ip4_frag_free()
155 ipq_kill(qp); in ip_expire()
197 ipq_put(qp); in ip_expire()
265 qp->iif = 0; in ip_frag_reinit()
266 qp->ecn = 0; in ip_frag_reinit()
367 qp->q.meat == qp->q.len) { in ip_frag_queue()
411 ipq_kill(qp); in ip_frag_reasm()
447 if (qp->max_df_size == qp->q.max_size) { in ip_frag_reasm()
[all …]
/linux/include/rdma/
A Drdmavt_qp.h510 struct rvt_qp *qp; member
558 return !!qp->pid; in rvt_is_user_qp()
576 if (qp && atomic_dec_and_test(&qp->refcount)) in rvt_put_qp()
606 struct rvt_qp *qp, in rvt_qp_wqe_reserve() argument
662 return (len + qp->pmtu - 1) >> qp->log_pmtu; in rvt_div_round_up_mtu()
705 qp = rcu_dereference(rvp->qp[qpn]); in rvt_lookup_qpn()
709 for (qp = rcu_dereference(rdi->qp_dev->qp_table[n]); qp; in rvt_lookup_qpn()
710 qp = rcu_dereference(qp->next)) in rvt_lookup_qpn()
714 return qp; in rvt_lookup_qpn()
851 last = rvt_qp_swqe_incr(qp, qp->s_last); in rvt_qp_complete_swqe()
[all …]

Completed in 136 milliseconds

12345678910>>...14