/linux/drivers/infiniband/hw/hns/ |
A D | hns_roce_qp.c | 470 hr_qp->rq.wqe_cnt = 0; in set_rq_size() 472 hr_qp->rq_inl_buf.wqe_cnt = 0; in set_rq_size() 504 hr_qp->rq.wqe_cnt = cnt; in set_rq_size() 508 hr_qp->rq_inl_buf.wqe_cnt = cnt; in set_rq_size() 510 hr_qp->rq_inl_buf.wqe_cnt = 0; in set_rq_size() 602 hr_qp->sq.wqe_cnt = cnt; in set_user_sq_size() 678 hr_qp->sq.wqe_cnt = cnt; in set_kernel_sq_size() 711 u32 wqe_cnt = hr_qp->rq_inl_buf.wqe_cnt; in alloc_rq_inline_buf() local 729 for (i = 1; i < wqe_cnt; i++) in alloc_rq_inline_buf() 956 if (hr_qp->rq.wqe_cnt) { in alloc_kernel_wrid() [all …]
|
A D | hns_roce_srq.c | 175 buf_attr.region[0].size = to_hr_hem_entries_size(srq->wqe_cnt, in alloc_srq_idx() 190 idx_que->bitmap = bitmap_zalloc(srq->wqe_cnt, GFP_KERNEL); in alloc_srq_idx() 230 buf_attr.region[0].size = to_hr_hem_entries_size(srq->wqe_cnt, in alloc_srq_wqe_buf() 253 srq->wrid = kvmalloc_array(srq->wqe_cnt, sizeof(u64), GFP_KERNEL); in alloc_srq_wrid() 306 srq->wqe_cnt = roundup_pow_of_two(attr->max_wr); in set_srq_basic_param() 309 attr->max_wr = srq->wqe_cnt; in set_srq_basic_param()
|
A D | hns_roce_hw_v1.c | 2284 ((*cur_qp)->sq.wqe_cnt-1)); in hns_roce_v1_poll_one() 2320 (wq->wqe_cnt - 1); in hns_roce_v1_poll_one() 2599 ilog2((unsigned int)hr_qp->sq.wqe_cnt)); in hns_roce_v1_m_sqp() 2603 ilog2((unsigned int)hr_qp->rq.wqe_cnt)); in hns_roce_v1_m_sqp() 2806 ilog2((unsigned int)hr_qp->sq.wqe_cnt)); in hns_roce_v1_m_qp() 2810 ilog2((unsigned int)hr_qp->rq.wqe_cnt)); in hns_roce_v1_m_qp() 2871 ilog2((unsigned int)hr_qp->sq.wqe_cnt)); in hns_roce_v1_m_qp() 3443 qp_attr->cap.max_recv_wr = hr_qp->rq.wqe_cnt; in hns_roce_v1_q_sqp() 3445 qp_attr->cap.max_send_wr = hr_qp->sq.wqe_cnt; in hns_roce_v1_q_sqp() 3578 qp_attr->cap.max_recv_wr = hr_qp->rq.wqe_cnt; in hns_roce_v1_q_qp() [all …]
|
A D | hns_roce_hw_v2.c | 832 if (hr_qp->rq_inl_buf.wqe_cnt) { in fill_rq_wqe() 951 if (unlikely(pos == srq->wqe_cnt)) in get_srq_wqe_idx() 965 head = idx_que->head & (srq->wqe_cnt - 1); in fill_wqe_idx() 3412 wr_cnt = wr_num & (qp->rq.wqe_cnt - 1); in hns_roce_handle_recv_inl_wqe() 3753 (wq->wqe_cnt - 1); in hns_roce_v2_poll_one() 4190 if (hr_qp->rq.wqe_cnt && count < 1) { in config_qp_rq_buf() 4210 hr_qp->sq.wqe_cnt)); in config_qp_rq_buf() 4220 hr_qp->rq.wqe_cnt)); in config_qp_rq_buf() 5420 srq->wqe_cnt)); in hns_roce_v2_write_srqc() 5451 if (srq_attr->srq_limit > srq->wqe_cnt) in hns_roce_v2_modify_srq() [all …]
|
A D | hns_roce_device.h | 374 u32 wqe_cnt; /* WQE num */ member 480 u32 wqe_cnt; member 615 u32 wqe_cnt; member
|
/linux/drivers/infiniband/hw/mlx5/ |
A D | qp.c | 363 qp->rq.wqe_cnt = 0; in set_rq_size() 538 qp->sq.wqe_cnt, in calc_sq_size() 578 qp->sq.wqe_cnt, in set_user_buf_size() 1055 if (qp->rq.wqe_cnt) in _create_kernel_qp() 1059 if (qp->sq.wqe_cnt) { in _create_kernel_qp() 1499 if (!qp->sq.wqe_cnt && !qp->rq.wqe_cnt) in create_raw_packet_qp() 1501 if (qp->sq.wqe_cnt) { in create_raw_packet_qp() 1522 if (qp->rq.wqe_cnt) { in create_raw_packet_qp() 1569 if (!qp->sq.wqe_cnt) in create_raw_packet_qp() 1585 if (qp->rq.wqe_cnt) { in destroy_raw_packet_qp() [all …]
|
A D | srq.h | 21 u32 wqe_cnt; member
|
A D | cq.c | 197 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in handle_responder() 341 idx = tail & (qp->sq.wqe_cnt - 1); in handle_atomics() 407 idx &= (wq->wqe_cnt - 1); in sw_comp() 498 idx = wqe_ctr & (wq->wqe_cnt - 1); in mlx5_poll_one() 526 idx = wqe_ctr & (wq->wqe_cnt - 1); in mlx5_poll_one() 539 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx5_poll_one()
|
A D | wr.c | 43 idx = (sq->cur_post + (wqe_sz >> 2)) & (sq->wqe_cnt - 1); in handle_post_send_edge() 937 idx = (idx + 1) & (qp->sq.wqe_cnt - 1); in dump_wqe() 954 *idx = qp->sq.cur_post & (qp->sq.wqe_cnt - 1); in __begin_wqe() 1007 (qp->sq.wqe_cnt - 1)) : in finish_wqe() 1486 ind = qp->rq.head & (qp->rq.wqe_cnt - 1); in mlx5_ib_post_recv() 1521 ind = (ind + 1) & (qp->rq.wqe_cnt - 1); in mlx5_ib_post_recv()
|
A D | mem.c | 120 idx = qp->sq.cur_post & (qp->sq.wqe_cnt - 1); in post_send_nop()
|
A D | mlx5_ib.h | 353 int wqe_cnt; member
|
/linux/drivers/infiniband/hw/mlx4/ |
A D | qp.c | 314 qp->rq.wqe_cnt = qp->rq.max_gs = 0; in set_rq_size() 391 (qp->sq.wqe_cnt << qp->sq.wqe_shift); in set_kernel_sq_size() 401 qp->sq.wqe_cnt - qp->sq_spare_wqes; in set_kernel_sq_size() 641 qp->sq.wqe_cnt = 1; in create_qp_rss() 877 qp->sq.wqe_cnt = 1; in create_rq() 1436 if (qp->rq.wqe_cnt) { in destroy_qp_common() 1452 if (qp->rq.wqe_cnt) in destroy_qp_common() 2171 if (qp->rq.wqe_cnt) in __mlx4_ib_modify_qp() 2176 if (qp->sq.wqe_cnt) in __mlx4_ib_modify_qp() 2396 if (qp->rq.wqe_cnt && in __mlx4_ib_modify_qp() [all …]
|
A D | cq.c | 621 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx4_ib_qp_sw_comp() 737 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx4_ib_poll_one() 751 tail = wq->tail & (wq->wqe_cnt - 1); in mlx4_ib_poll_one()
|
A D | mlx4_ib.h | 165 int wqe_cnt; member
|
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
A D | pvrdma_qp.c | 140 qp->rq.wqe_cnt = roundup_pow_of_two(max(1U, req_cap->max_recv_wr)); in pvrdma_set_rq_size() 144 req_cap->max_recv_wr = qp->rq.wqe_cnt; in pvrdma_set_rq_size() 150 qp->npages_recv = (qp->rq.wqe_cnt * qp->rq.wqe_size + PAGE_SIZE - 1) / in pvrdma_set_rq_size() 165 qp->sq.wqe_cnt = roundup_pow_of_two(max(1U, req_cap->max_send_wr)); in pvrdma_set_sq_size() 169 req_cap->max_send_wr = qp->sq.wqe_cnt; in pvrdma_set_sq_size() 177 (qp->sq.wqe_cnt * qp->sq.wqe_size + PAGE_SIZE - 1) / in pvrdma_set_sq_size() 696 qp->sq.ring, qp->sq.wqe_cnt, &tail))) { in pvrdma_post_send() 857 qp->sq.wqe_cnt); in pvrdma_post_send() 922 qp->rq.ring, qp->rq.wqe_cnt, &tail))) { in pvrdma_post_recv() 948 qp->rq.wqe_cnt); in pvrdma_post_recv()
|
A D | pvrdma.h | 154 int wqe_cnt; member 169 int wqe_cnt; member
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
A D | dr_send.c | 60 (dr_cq->qp->sq.wqe_cnt - 1); in dr_parse_cqe() 66 (dr_cq->qp->sq.wqe_cnt - 1); in dr_parse_cqe() 126 dr_qp->rq.wqe_cnt = 4; in dr_create_rc_qp() 129 dr_qp->sq.wqe_cnt = roundup_pow_of_two(attr->max_send_wr); in dr_create_rc_qp() 132 MLX5_SET(qpc, temp_qpc, log_rq_size, ilog2(dr_qp->rq.wqe_cnt)); in dr_create_rc_qp() 133 MLX5_SET(qpc, temp_qpc, log_sq_size, ilog2(dr_qp->sq.wqe_cnt)); in dr_create_rc_qp() 141 dr_qp->sq.wqe_head = kcalloc(dr_qp->sq.wqe_cnt, in dr_create_rc_qp() 172 MLX5_SET(qpc, qpc, log_rq_size, ilog2(dr_qp->rq.wqe_cnt)); in dr_create_rc_qp() 174 MLX5_SET(qpc, qpc, log_sq_size, ilog2(dr_qp->sq.wqe_cnt)); in dr_create_rc_qp() 240 idx = dr_qp->sq.pc & (dr_qp->sq.wqe_cnt - 1); in dr_rdma_segments()
|
A D | dr_types.h | 1248 unsigned int wqe_cnt; member 1254 unsigned int wqe_cnt; member
|
/linux/drivers/infiniband/hw/mthca/ |
A D | mthca_srq.c | 56 __be16 wqe_cnt; member 68 __be16 wqe_cnt; member
|
/linux/drivers/net/ethernet/microsoft/mana/ |
A D | gdma.h | 81 u64 wqe_cnt : 8; member
|
A D | gdma_main.c | 247 e.rq.wqe_cnt = num_req; in mana_gd_ring_doorbell()
|
/linux/drivers/infiniband/hw/bnxt_re/ |
A D | qplib_fp.h | 293 u64 wqe_cnt; member
|
A D | ib_verbs.c | 2543 qp->qplib_qp.wqe_cnt == BNXT_RE_UD_QP_HW_STALL) { in bnxt_ud_qp_hw_stall_workaround() 2550 qp->qplib_qp.wqe_cnt = 0; in bnxt_ud_qp_hw_stall_workaround()
|
A D | qplib_fp.c | 1929 qp->wqe_cnt++; in bnxt_qplib_post_send()
|
/linux/drivers/net/ethernet/mellanox/mlx4/ |
A D | mlx4.h | 376 __be16 wqe_cnt; member
|