/linux/drivers/infiniband/hw/qib/ |
A D | qib_rc.c | 253 rvt_send_complete(qp, wqe, qp->s_last != qp->s_acked ? in qib_make_rc_req() 474 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in qib_make_rc_req() 499 if (qp->s_cur >= qp->s_size) in qib_make_rc_req() 513 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in qib_make_rc_req() 538 if (qp->s_cur >= qp->s_size) in qib_make_rc_req() 980 qp->s_acked = qp->s_cur; in do_rc_completion() 1881 wc.qp = &qp->ibqp; in qib_rc_rcv() 2084 qp->r_ack_psn = qp->r_psn; in qib_rc_rcv() 2096 qp->r_ack_psn = qp->r_psn; in qib_rc_rcv() 2110 qp->r_ack_psn = qp->r_psn; in qib_rc_rcv() [all …]
|
A D | qib_uc.c | 71 wqe = rvt_get_swqe_ptr(qp, qp->s_last); in qib_make_uc_req() 85 wqe = rvt_get_swqe_ptr(qp, qp->s_cur); in qib_make_uc_req() 125 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req() 154 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req() 183 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req() 208 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req() 214 qp->s_cur_sge = &qp->s_sge; in qib_make_uc_req() 337 qp->r_sge = qp->s_rdma_read_sge; in qib_uc_rcv() 348 qp->s_rdma_read_sge = qp->r_sge; in qib_uc_rcv() 394 wc.qp = &qp->ibqp; in qib_uc_rcv() [all …]
|
A D | qib_ruc.c | 49 qp->remote_ah_attr = qp->alt_ah_attr; in qib_migrate_qp() 50 qp->port_num = rdma_ah_get_port_num(&qp->alt_ah_attr); in qib_migrate_qp() 51 qp->s_pkey_index = qp->s_alt_pkey_index; in qib_migrate_qp() 54 ev.element.qp = &qp->ibqp; in qib_migrate_qp() 56 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in qib_migrate_qp() 210 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_make_ruc_header() 220 qp->s_hdrwords += in qib_make_ruc_header() 252 qib_do_send(qp); in _qib_do_send() 266 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_do_send() 304 if (qib_verbs_send(qp, priv->s_hdr, qp->s_hdrwords, in qib_do_send() [all …]
|
A D | qib_ud.c | 68 if (!qp) { in qib_ud_loopback() 166 rvt_copy_sge(qp, &qp->r_sge, &grh, in qib_ud_loopback() 204 wc.qp = &qp->ibqp; in qib_ud_loopback() 251 if (qp->s_last == READ_ONCE(qp->s_head)) in qib_make_ud_req() 258 wqe = rvt_get_swqe_ptr(qp, qp->s_last); in qib_make_ud_req() 264 if (qp->s_cur == READ_ONCE(qp->s_head)) in qib_make_ud_req() 267 wqe = rvt_get_swqe_ptr(qp, qp->s_cur); in qib_make_ud_req() 273 ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_make_ud_req() 315 qp->s_cur_sge = &qp->s_sge; in qib_make_ud_req() 550 rvt_copy_sge(qp, &qp->r_sge, &hdr->u.l.grh, in qib_ud_rcv() [all …]
|
/linux/drivers/infiniband/sw/rxe/ |
A D | rxe_qp.c | 168 qp->mtu = ib_mtu_enum_to_int(qp->attr.path_mtu); in rxe_qp_init_misc() 211 qp->sk->sk->sk_user_data = qp; in rxe_qp_init_req() 239 qp->sq.queue->buf, qp->sq.queue->buf_size, in rxe_qp_init_req() 259 rxe_init_task(rxe, &qp->req.task, qp, in rxe_qp_init_req() 261 rxe_init_task(rxe, &qp->comp.task, qp, in rxe_qp_init_req() 288 qp_num(qp), qp->rq.max_wr, qp->rq.max_sge, wqe_size); in rxe_qp_init_resp() 297 qp->rq.queue->buf, qp->rq.queue->buf_size, in rxe_qp_init_resp() 312 rxe_init_task(rxe, &qp->resp.task, qp, in rxe_qp_init_resp() 669 qp->resp.psn = qp->attr.rq_psn; in rxe_qp_from_attr() 682 qp->req.psn = qp->attr.sq_psn; in rxe_qp_from_attr() [all …]
|
A D | rxe_resp.c | 321 qp->resp.wqe = &qp->resp.srq_wqe.wqe; in get_srq_wqe() 352 qp->resp.wqe = queue_head(qp->rq.queue, in check_resource() 380 qp->resp.wqe = queue_head(qp->rq.queue, in check_resource() 538 err = rxe_mr_copy(qp->resp.mr, qp->resp.va + qp->resp.offset, in write_data_in() 618 ack->qp = qp; in prepare_ack_packet() 663 res = &qp->resp.resources[qp->resp.res_head]; in read_reply() 819 qp->resp.ack_psn = qp->resp.psn; in execute() 854 wc->qp = &qp->ibqp; in do_complete() 976 res = &qp->resp.resources[qp->resp.res_head]; in send_atomic_ack() 1003 send_ack(qp, pkt, qp->resp.aeth_syndrome, pkt->psn); in acknowledge() [all …]
|
A D | rxe_comp.c | 115 struct rxe_qp *qp = from_timer(qp, t, retrans_timer); in retransmit_timer() local 117 if (qp->valid) { in retransmit_timer() 166 qp->comp.retry_cnt = qp->attr.retry_cnt; in reset_retry_counters() 167 qp->comp.rnr_retry = qp->attr.rnr_retry; in reset_retry_counters() 391 wc->qp = &qp->ibqp; in make_send_cqe() 477 (qp->comp.psn == qp->req.psn)) { in complete_ack() 485 ev.element.qp = &qp->ibqp; in complete_ack() 549 struct rxe_qp *qp = pkt->qp; in free_pkt() local 569 if (!qp->valid || qp->req.state == QP_STATE_ERROR || in rxe_completer() 571 rxe_drain_resp_pkts(qp, qp->valid && in rxe_completer() [all …]
|
A D | rxe_req.c | 27 qp->req.opcode = next_opcode(qp, wqe, in retry_first_write_send() 56 qp->req.psn = qp->comp.psn; in req_retry() 93 qp->mtu; in req_retry() 104 struct rxe_qp *qp = from_timer(qp, t, rnr_nak_timer); in rnr_nak_timer() local 151 ev.element.qp = &qp->ibqp; in req_next_wqe() 385 pkt->qp = qp; in init_req_packet() 523 qp->req.psn = (qp->req.psn + 1) & BTH_PSN_MASK; in update_wqe_psn() 554 qp->req.wqe_index = queue_next_index(qp->sq.queue, in update_state() 605 qp->req.wqe_index = queue_next_index(qp->sq.queue, qp->req.wqe_index); in rxe_do_local_ops() 663 psn_compare(qp->req.psn, (qp->comp.psn + in rxe_requester() [all …]
|
/linux/drivers/infiniband/hw/hfi1/ |
A D | rc.c | 906 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in hfi1_make_rc_req() 937 if (qp->s_cur >= qp->s_size) in hfi1_make_rc_req() 951 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in hfi1_make_rc_req() 977 if (qp->s_cur >= qp->s_size) in hfi1_make_rc_req() 1885 qp->s_acked = qp->s_cur; in do_rc_completion() 1894 if (qp->state == IB_QPS_SQD && qp->s_acked == qp->s_cur) in do_rc_completion() 2945 wc.qp = &qp->ibqp; in hfi1_rc_rcv() 3176 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv() 3184 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv() 3194 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv() [all …]
|
A D | qp.c | 449 qp = tx->qp; in iowait_sleep() 587 qp->s_last == qp->s_acked && in qp_idle() 588 qp->s_acked == qp->s_cur && in qp_idle() 589 qp->s_cur == qp->s_tail && in qp_idle() 590 qp->s_tail == qp->s_head; in qp_idle() 601 struct rvt_qp *qp = iter->qp; in qp_iter_print() local 631 qp->s_psn, qp->s_next_psn, in qp_iter_print() 634 qp->s_last, qp->s_acked, qp->s_cur, in qp_iter_print() 635 qp->s_tail, qp->s_head, qp->s_size, in qp_iter_print() 789 ev.element.qp = &qp->ibqp; in hfi1_migrate_qp() [all …]
|
A D | uc.c | 49 wqe = rvt_get_swqe_ptr(qp, qp->s_last); in hfi1_make_uc_req() 72 wqe = rvt_get_swqe_ptr(qp, qp->s_cur); in hfi1_make_uc_req() 93 if (qp->s_last != qp->s_cur) in hfi1_make_uc_req() 95 if (++qp->s_cur == qp->s_size) in hfi1_make_uc_req() 138 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req() 167 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req() 197 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req() 223 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req() 264 struct rvt_qp *qp = packet->qp; in hfi1_uc_rcv() local 412 wc.qp = &qp->ibqp; in hfi1_uc_rcv() [all …]
|
A D | tid_rdma.c | 1669 req->qp = qp; in hfi1_init_trdma_req() 1709 struct rvt_qp *qp = req->qp; in hfi1_build_tid_rdma_read_packet() local 2006 qp->r_ack_psn = qp->r_psn; in tid_rdma_rcv_error() 2333 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_read_req() 2341 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_read_req() 2892 packet->qp = qp; in hfi1_handle_kdeth_eflags() 3217 s_prev = (qp->s_cur == 0 ? qp->s_size : qp->s_cur) - 1; in hfi1_tid_rdma_wqe_interlock() 3603 qp->r_ack_psn = qp->r_psn; in hfi1_tid_write_alloc_resources() 3824 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_write_req() 3832 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_write_req() [all …]
|
A D | trace_tid.h | 196 TP_ARGS(qp), 226 TP_ARGS(qp) 232 TP_ARGS(qp) 238 TP_ARGS(qp) 244 TP_ARGS(qp) 250 TP_ARGS(qp) 360 __entry->qpn = qp ? qp->ibqp.qp_num : 0; 931 TP_ARGS(qp) 937 TP_ARGS(qp) 943 TP_ARGS(qp) [all …]
|
A D | ruc.c | 31 struct rvt_qp *qp = packet->qp; in hfi1_ruc_check_hdr() local 74 hfi1_migrate_qp(qp); in hfi1_ruc_check_hdr() 168 clear_ahg(qp); in build_ahg() 171 if (qp->s_ahgidx < 0) in build_ahg() 174 qp->s_ahgpsn = npsn; in build_ahg() 282 build_ahg(qp, bth2); in hfi1_make_ruc_header_16B() 360 build_ahg(qp, bth2); in hfi1_make_ruc_header_9B() 481 hfi1_do_send(qp, true); in _hfi1_do_send() 500 ps.ibp = to_iport(qp->ibqp.device, qp->port_num); in hfi1_do_send() 512 rvt_ruc_loopback(qp); in hfi1_do_send() [all …]
|
/linux/drivers/infiniband/sw/rdmavt/ |
A D | qp.c | 717 ev.element.qp = &qp->ibqp; in rvt_qp_mr_clean() 1318 wc.qp = &qp->ibqp; in rvt_error_qp() 1383 rcu_assign_pointer(rvp->qp[qp->ibqp.qp_num], qp); in rvt_insert_qp() 1544 qp->s_draining = qp->s_last != qp->s_cur; in rvt_modify_qp() 1650 ev.element.qp = &qp->ibqp; in rvt_modify_qp() 1821 wc.qp = &qp->ibqp; in rvt_post_recv() 1950 qp->ibqp.qp_num, qp->s_size, qp->s_avail, in rvt_qp_is_avail() 1951 qp->s_head, qp->s_tail, qp->s_cur, in rvt_qp_is_avail() 2310 wc.qp = &qp->ibqp; in init_sge() 2725 iter->qp = qp; in rvt_qp_iter_next() [all …]
|
/linux/drivers/ntb/ |
A D | ntb_transport.c | 472 if (!qp || !qp->link_is_up) in debugfs_read() 661 entry->qp = qp; in ntb_transport_setup_qp_mw() 946 qp->event_handler(qp->cb_data, qp->link_is_up); in ntb_qp_link_cleanup() 1147 qp->event_handler(qp->cb_data, qp->link_is_up); in ntb_qp_link_work() 1217 qp->debugfs_dir, qp, in ntb_transport_init_queue() 1463 qp->rx_handler(qp, qp->cb_data, cb_data, len); in ntb_complete_rxc() 1620 offset = qp->rx_buff + qp->rx_max_frame * qp->rx_index; in ntb_process_rxc() 1774 qp->tx_handler(qp, qp->cb_data, entry->cb_data, in ntb_tx_copy_callback() 1904 qp->tx_handler(qp, qp->cb_data, NULL, -EIO); in ntb_process_tx() 2056 entry->qp = qp; in ntb_transport_create_queue() [all …]
|
/linux/drivers/infiniband/sw/siw/ |
A D | siw_qp.c | 168 if (qp->cep) { in siw_qp_llp_close() 170 qp->cep = NULL; in siw_qp_llp_close() 368 memset(&qp->term_info, 0, sizeof(qp->term_info)); in siw_init_terminate() 685 qp->attrs.orq_size, qp->attrs.irq_size); in siw_qp_nextstate_from_idle() 926 if (qp->attrs.orq_size && !siw_orq_empty(qp)) { in siw_activate_tx_from_sq() 962 qp->sq_get++; in siw_activate_tx_from_sq() 985 irqe = &qp->irq[qp->irq_get % qp->attrs.irq_size]; in siw_activate_tx() 1198 sqe = &qp->orq[qp->orq_get % qp->attrs.orq_size]; in siw_sq_flush() 1234 sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_flush() 1298 &qp->recvq[qp->rq_get % qp->attrs.rq_size]; in siw_rq_flush() [all …]
|
/linux/drivers/infiniband/hw/mthca/ |
A D | mthca_qp.c | 257 qp->port = qp->alt_port; in mthca_qp_event() 261 event.element.qp = &qp->ibqp; in mthca_qp_event() 841 qp->sq.last = get_send_wqe(qp, qp->sq.max - 1); in __mthca_modify_qp() 844 qp->rq.last = get_recv_wqe(qp, qp->rq.max - 1); in __mthca_modify_qp() 1052 qp->send_wqe_offset = ALIGN(qp->rq.max << qp->rq.wqe_shift, in mthca_alloc_wqe_buf() 1072 &qp->queue, &qp->is_direct, pd, 0, &qp->mr); in mthca_alloc_wqe_buf() 1088 &qp->queue, qp->is_direct, &qp->mr); in mthca_free_wqe_buf() 1138 qp->qpn, &qp->rq.db); in mthca_alloc_memfree() 1143 qp->qpn, &qp->sq.db); in mthca_alloc_memfree() 1246 qp->sq.last = get_send_wqe(qp, qp->sq.max - 1); in mthca_alloc_qp_common() [all …]
|
/linux/drivers/net/ethernet/qlogic/qed/ |
A D | qed_roce.c | 210 if (qp->vlan_id) { in qed_roce_get_qp_tc() 217 qp->icid, tc, qp->vlan_id ? "enabled" : "disabled"); in qed_roce_get_qp_tc() 244 if (!qp->irq) { in qed_roce_sp_create_responder() 343 qp->cq_prod = 0; in qed_roce_sp_create_responder() 355 qp->irq, qp->irq_phys_addr); in qed_roce_sp_create_responder() 382 if (!qp->orq) { in qed_roce_sp_create_requester() 486 qp->orq, qp->orq_phys_addr); in qed_roce_sp_create_requester() 736 qp->irq, qp->irq_phys_addr); in qed_roce_sp_destroy_qp_responder() 798 qp->orq, qp->orq_phys_addr); in qed_roce_sp_destroy_qp_requester() 828 if ((!(qp->resp_offloaded)) && (!(qp->req_offloaded))) { in qed_roce_query_qp() [all …]
|
/linux/drivers/infiniband/hw/irdma/ |
A D | uk.c | 130 writel(qp->qp_id, qp->wqe_alloc_db); in irdma_uk_qp_post_wr() 136 writel(qp->qp_id, qp->wqe_alloc_db); in irdma_uk_qp_post_wr() 140 writel(qp->qp_id, qp->wqe_alloc_db); in irdma_uk_qp_post_wr() 145 qp->initial_ring.head = qp->sq_ring.head; in irdma_uk_qp_post_wr() 157 qp->initial_ring.head = qp->sq_ring.head; in irdma_qp_ring_push_db() 217 irdma_qp_push_wqe(qp, qp->sq_base[nop_wqe_idx].elem, in irdma_qp_get_next_send_wqe() 223 qp->swqe_polarity = !qp->swqe_polarity; in irdma_qp_get_next_send_wqe() 258 qp->rwqe_polarity = !qp->rwqe_polarity; in irdma_qp_get_next_recv_wqe() 1114 if (!qp || qp->destroy_pending) { in irdma_uk_cq_poll_cmpl() 1165 if (wqe_idx < qp->conn_wqes && qp->sq_ring.head == qp->sq_ring.tail) { in irdma_uk_cq_poll_cmpl() [all …]
|
/linux/drivers/infiniband/hw/mlx4/ |
A D | qp.c | 173 return get_wqe(qp, qp->rq.offset + (n << qp->rq.wqe_shift)); in get_recv_wqe() 178 return get_wqe(qp, qp->sq.offset + (n << qp->sq.wqe_shift)); in get_send_wqe() 390 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in set_kernel_sq_size() 394 qp->sq.offset = qp->rq.wqe_cnt << qp->rq.wqe_shift; in set_kernel_sq_size() 396 qp->rq.offset = qp->sq.wqe_cnt << qp->sq.wqe_shift; in set_kernel_sq_size() 425 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in set_user_sq_size() 735 qp->ibqp.qp_num = qp->mqp.qpn; in _mlx4_ib_create_qp_rss() 879 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in create_rq() 1373 if (qp->pri.smac || (!qp->pri.smac && qp->pri.smac_port)) { in destroy_qp_common() 2555 if (qp->pri.smac || (!qp->pri.smac && qp->pri.smac_port)) { in __mlx4_ib_modify_qp() [all …]
|
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
A D | pvrdma_qp.c | 150 qp->npages_recv = (qp->rq.wqe_cnt * qp->rq.wqe_size + PAGE_SIZE - 1) / in pvrdma_set_rq_size() 300 qp->npages = qp->npages_send + qp->npages_recv; in pvrdma_create_qp() 312 qp->npages = qp->npages_send + qp->npages_recv; in pvrdma_create_qp() 318 qp->rq.offset = qp->npages_send * PAGE_SIZE; in pvrdma_create_qp() 344 qp->sq.ring = qp->pdir.pages[0]; in pvrdma_create_qp() 345 qp->rq.ring = is_srq ? NULL : &qp->sq.ring[1]; in pvrdma_create_qp() 542 qp->state; in pvrdma_modify_qp() 637 qp->sq.offset + n * qp->sq.wqe_size); in get_sq_wqe() 643 qp->rq.offset + n * qp->rq.wqe_size); in get_rq_wqe() 696 qp->sq.ring, qp->sq.wqe_cnt, &tail))) { in pvrdma_post_send() [all …]
|
/linux/drivers/infiniband/hw/bnxt_re/ |
A D | qplib_fp.c | 96 if (qp->scq == qp->rcq) in bnxt_qplib_acquire_cq_flush_locks() 106 if (qp->scq == qp->rcq) in bnxt_qplib_release_cq_flush_locks() 141 __clean_cq(qp->scq, (u64)(unsigned long)qp); in bnxt_qplib_clean_qp() 144 __clean_cq(qp->rcq, (u64)(unsigned long)qp); in bnxt_qplib_clean_qp() 182 qp->rq_hdr_buf, qp->rq_hdr_buf_map); in bnxt_qplib_free_qp_hdr_buf() 186 qp->sq_hdr_buf, qp->sq_hdr_buf_map); in bnxt_qplib_free_qp_hdr_buf() 1332 qp->cur_qp_state = qp->state; in bnxt_qplib_modify_qp() 1406 qp->sq.max_wqe = qp->sq.hwq.max_elements; in bnxt_qplib_query_qp() 1724 qp->id, qp->state); in bnxt_qplib_post_send() 1972 qp->id, qp->state); in bnxt_qplib_post_recv() [all …]
|
/linux/net/ipv4/ |
A D | ip_fragment.c | 90 qp->ecn = 0; in ip4_frag_init() 98 struct ipq *qp; in ip4_frag_free() local 101 if (qp->peer) in ip4_frag_free() 155 ipq_kill(qp); in ip_expire() 197 ipq_put(qp); in ip_expire() 265 qp->iif = 0; in ip_frag_reinit() 266 qp->ecn = 0; in ip_frag_reinit() 367 qp->q.meat == qp->q.len) { in ip_frag_queue() 411 ipq_kill(qp); in ip_frag_reasm() 447 if (qp->max_df_size == qp->q.max_size) { in ip_frag_reasm() [all …]
|
/linux/include/rdma/ |
A D | rdmavt_qp.h | 510 struct rvt_qp *qp; member 558 return !!qp->pid; in rvt_is_user_qp() 576 if (qp && atomic_dec_and_test(&qp->refcount)) in rvt_put_qp() 606 struct rvt_qp *qp, in rvt_qp_wqe_reserve() argument 662 return (len + qp->pmtu - 1) >> qp->log_pmtu; in rvt_div_round_up_mtu() 705 qp = rcu_dereference(rvp->qp[qpn]); in rvt_lookup_qpn() 709 for (qp = rcu_dereference(rdi->qp_dev->qp_table[n]); qp; in rvt_lookup_qpn() 710 qp = rcu_dereference(qp->next)) in rvt_lookup_qpn() 714 return qp; in rvt_lookup_qpn() 851 last = rvt_qp_swqe_incr(qp, qp->s_last); in rvt_qp_complete_swqe() [all …]
|