/linux/block/ |
A D | blk-wbt.c | 238 struct rq_depth *rqd = &rwb->rq_depth; in latency_exceeded() 291 struct rq_depth *rqd = &rwb->rq_depth; in rwb_trace_step() 301 } else if (rwb->rq_depth.max_depth <= 2) { in calc_wb_limits() 302 rwb->wb_normal = rwb->rq_depth.max_depth; in calc_wb_limits() 305 rwb->wb_normal = (rwb->rq_depth.max_depth + 1) / 2; in calc_wb_limits() 312 if (!rq_depth_scale_up(&rwb->rq_depth)) in scale_up() 331 struct rq_depth *rqd = &rwb->rq_depth; in rwb_arm_timer() 356 struct rq_depth *rqd = &rwb->rq_depth; in wb_timer_fn() 414 struct rq_depth *rqd = &rwb->rq_depth; in wbt_update_limits() 477 limit = rwb->rq_depth.max_depth; in get_limit() [all …]
|
A D | blk-iolatency.c | 139 struct rq_depth rq_depth; member 370 unsigned long old = iolat->rq_depth.max_depth; in scale_change() 382 iolat->rq_depth.max_depth = old; in scale_change() 387 iolat->rq_depth.max_depth = max(old, 1UL); in scale_change() 453 iolat->rq_depth.max_depth = UINT_MAX; in check_scale_change() 908 if (iolat->rq_depth.max_depth == UINT_MAX) in iolatency_ssd_stat() 916 iolat->rq_depth.max_depth); in iolatency_ssd_stat() 934 if (iolat->rq_depth.max_depth == UINT_MAX) in iolatency_pd_stat() 939 iolat->rq_depth.max_depth, avg_lat, cur_win); in iolatency_pd_stat() 985 iolat->rq_depth.max_depth = UINT_MAX; in iolatency_pd_init() [all …]
|
A D | blk-rq-qos.h | 52 struct rq_depth { struct 142 bool rq_depth_scale_up(struct rq_depth *rqd); 143 bool rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle); 144 bool rq_depth_calc_max_depth(struct rq_depth *rqd);
|
A D | blk-wbt.h | 70 struct rq_depth rq_depth; member
|
A D | blk-rq-qos.c | 116 bool rq_depth_calc_max_depth(struct rq_depth *rqd) in rq_depth_calc_max_depth() 164 bool rq_depth_scale_up(struct rq_depth *rqd) in rq_depth_scale_up() 183 bool rq_depth_scale_down(struct rq_depth *rqd, bool hard_throttle) in rq_depth_scale_down()
|
/linux/net/9p/ |
A D | trans_rdma.c | 86 int rq_depth; member 127 int rq_depth; member 159 if (rdma->rq_depth != P9_RDMA_RQ_DEPTH) in p9_rdma_show_options() 160 seq_printf(m, ",rq=%u", rdma->rq_depth); in p9_rdma_show_options() 184 opts->rq_depth = P9_RDMA_RQ_DEPTH; in parse_opts() 221 opts->rq_depth = option; in parse_opts() 234 opts->rq_depth = max(opts->rq_depth, opts->sq_depth); in parse_opts() 574 rdma->rq_depth = opts->rq_depth; in alloc_rdma() 579 sema_init(&rdma->rq_sem, rdma->rq_depth); in alloc_rdma() 690 opts.sq_depth + opts.rq_depth + 1, in rdma_create_trans() [all …]
|
/linux/net/sunrpc/xprtrdma/ |
A D | svc_rdma_transport.c | 373 unsigned int ctxts, rq_depth; in svc_rdma_accept() local 408 rq_depth = newxprt->sc_max_requests + newxprt->sc_max_bc_requests + in svc_rdma_accept() 410 if (rq_depth > dev->attrs.max_qp_wr) { in svc_rdma_accept() 413 rq_depth = dev->attrs.max_qp_wr; in svc_rdma_accept() 415 newxprt->sc_max_requests = rq_depth - 2; in svc_rdma_accept() 421 newxprt->sc_sq_depth = rq_depth + ctxts; in svc_rdma_accept() 439 ib_alloc_cq_any(dev, newxprt, rq_depth, IB_POLL_WORKQUEUE); in svc_rdma_accept() 449 qp_attr.cap.max_recv_wr = rq_depth; in svc_rdma_accept()
|
/linux/drivers/net/ethernet/huawei/hinic/ |
A D | hinic_hw_dev.c | 278 unsigned int rq_depth) in set_hw_ioctxt() argument 292 hw_ioctxt.rq_depth = ilog2(rq_depth); in set_hw_ioctxt() 445 int hinic_hwdev_ifup(struct hinic_hwdev *hwdev, u16 sq_depth, u16 rq_depth) in hinic_hwdev_ifup() argument 469 func_to_io->rq_depth = rq_depth; in hinic_hwdev_ifup() 496 err = set_hw_ioctxt(hwdev, sq_depth, rq_depth); in hinic_hwdev_ifup()
|
A D | hinic_hw_io.h | 74 u16 rq_depth; member
|
A D | hinic_dev.h | 96 u16 rq_depth; member
|
A D | hinic_hw_dev.h | 220 u16 rq_depth; member 561 int hinic_hwdev_ifup(struct hinic_hwdev *hwdev, u16 sq_depth, u16 rq_depth);
|
A D | hinic_ethtool.c | 556 ring->rx_pending = nic_dev->rq_depth; in hinic_get_ringparam() 597 new_rq_depth == nic_dev->rq_depth) in hinic_set_ringparam() 602 nic_dev->sq_depth, nic_dev->rq_depth, in hinic_set_ringparam() 606 nic_dev->rq_depth = new_rq_depth; in hinic_set_ringparam()
|
A D | hinic_hw_mbox.c | 1294 ((hw_ctxt)->rq_depth >= HINIC_QUEUE_MIN_DEPTH && \ 1295 (hw_ctxt)->rq_depth <= HINIC_QUEUE_MAX_DEPTH && \ 1305 if (!hw_ctxt->rq_depth && !hw_ctxt->sq_depth && in hw_ctxt_qps_param_valid()
|
A D | hinic_main.c | 439 nic_dev->rq_depth); in hinic_open() 1231 nic_dev->rq_depth = HINIC_RQ_DEPTH; in nic_dev_init()
|
A D | hinic_port.c | 476 rq_num.rq_depth = ilog2(nic_dev->rq_depth); in hinic_set_max_qnum()
|
A D | hinic_hw_io.c | 293 func_to_io->rq_depth, HINIC_RQ_WQE_SIZE); in init_qp()
|
A D | hinic_port.h | 317 u32 rq_depth; member
|
/linux/drivers/infiniband/hw/efa/ |
A D | efa_com_cmd.h | 26 u32 rq_depth; member
|
A D | efa_com_cmd.c | 32 params->rq_depth; in efa_com_create_qp()
|
A D | efa_verbs.c | 693 create_qp_params.rq_depth = init_attr->cap.max_recv_wr; in efa_create_qp()
|
/linux/drivers/infiniband/hw/hns/ |
A D | hns_roce_hw_v2.h | 1280 __le16 rq_depth; member
|