/linux/net/ipv4/ |
A D | inet_connection_sock.c | 706 if (!nreq) { in inet_reqsk_clone() 735 rcu_assign_pointer(tcp_sk(nreq->sk)->fastopen_rsk, nreq); in inet_reqsk_clone() 737 return nreq; in inet_reqsk_clone() 814 if (!nreq) in reqsk_timer_handler() 826 req = nreq; in reqsk_timer_handler() 876 if (!nreq) in reqsk_timer_handler() 890 reqsk_put(nreq); in reqsk_timer_handler() 898 if (nreq) { in reqsk_timer_handler() 903 __reqsk_free(nreq); in reqsk_timer_handler() 1137 if (!nreq) { in inet_csk_complete_hashdance() [all …]
|
/linux/drivers/infiniband/hw/mlx5/ |
A D | wr.c | 93 return cur + nreq >= wq->max_post; in mlx5_wq_overflow() 973 void **cur_edge, int nreq) in begin_wqe() argument 1294 int nreq; in mlx5_ib_post_send() local 1312 for (nreq = 0; wr; nreq++, wr = wr->next) { in mlx5_ib_post_send() 1329 nreq); in mlx5_ib_post_send() 1429 if (likely(nreq)) { in mlx5_ib_post_send() 1430 qp->sq.head += nreq; in mlx5_ib_post_send() 1471 int nreq; in mlx5_ib_post_recv() local 1488 for (nreq = 0; wr; nreq++, wr = wr->next) { in mlx5_ib_post_recv() 1525 if (likely(nreq)) { in mlx5_ib_post_recv() [all …]
|
A D | srq.c | 413 int nreq; in mlx5_ib_post_srq_recv() local 424 for (nreq = 0; wr; nreq++, wr = wr->next) { in mlx5_ib_post_srq_recv() 456 if (likely(nreq)) { in mlx5_ib_post_srq_recv() 457 srq->wqe_ctr += nreq; in mlx5_ib_post_srq_recv()
|
/linux/drivers/infiniband/hw/mthca/ |
A D | mthca_qp.c | 1634 int nreq; in mthca_tavor_post_send() local 1655 for (nreq = 0; wr; ++nreq, wr = wr->next) { in mthca_tavor_post_send() 1830 int nreq; in mthca_tavor_post_receive() local 1890 if (!nreq) in mthca_tavor_post_receive() 1897 ++nreq; in mthca_tavor_post_receive() 1899 nreq = 0; in mthca_tavor_post_receive() 1938 int nreq; in mthca_arbel_post_send() local 1959 for (nreq = 0; wr; ++nreq, wr = wr->next) { in mthca_arbel_post_send() 1961 nreq = 0; in mthca_arbel_post_send() 2169 int nreq; in mthca_arbel_post_receive() local [all …]
|
A D | mthca_srq.c | 493 int nreq; in mthca_tavor_post_srq_recv() local 502 for (nreq = 0; wr; wr = wr->next) { in mthca_tavor_post_srq_recv() 543 ++nreq; in mthca_tavor_post_srq_recv() 544 if (unlikely(nreq == MTHCA_TAVOR_MAX_WQES_PER_RECV_DB)) { in mthca_tavor_post_srq_recv() 545 nreq = 0; in mthca_tavor_post_srq_recv() 561 if (likely(nreq)) { in mthca_tavor_post_srq_recv() 568 mthca_write64(first_ind << srq->wqe_shift, (srq->srqn << 8) | nreq, in mthca_tavor_post_srq_recv() 586 int nreq; in mthca_arbel_post_srq_recv() local 592 for (nreq = 0; wr; ++nreq, wr = wr->next) { in mthca_arbel_post_srq_recv() 627 if (likely(nreq)) { in mthca_arbel_post_srq_recv() [all …]
|
/linux/crypto/ |
A D | echainiv.c | 45 SYNC_SKCIPHER_REQUEST_ON_STACK(nreq, ctx->sknull); in echainiv_encrypt() 47 skcipher_request_set_sync_tfm(nreq, ctx->sknull); in echainiv_encrypt() 48 skcipher_request_set_callback(nreq, req->base.flags, in echainiv_encrypt() 50 skcipher_request_set_crypt(nreq, req->src, req->dst, in echainiv_encrypt() 54 err = crypto_skcipher_encrypt(nreq); in echainiv_encrypt()
|
A D | seqiv.c | 69 SYNC_SKCIPHER_REQUEST_ON_STACK(nreq, ctx->sknull); in seqiv_aead_encrypt() 71 skcipher_request_set_sync_tfm(nreq, ctx->sknull); in seqiv_aead_encrypt() 72 skcipher_request_set_callback(nreq, req->base.flags, in seqiv_aead_encrypt() 74 skcipher_request_set_crypt(nreq, req->src, req->dst, in seqiv_aead_encrypt() 78 err = crypto_skcipher_encrypt(nreq); in seqiv_aead_encrypt()
|
A D | gcm.c | 966 SYNC_SKCIPHER_REQUEST_ON_STACK(nreq, ctx->null); in crypto_rfc4543_copy_src_to_dst() 968 skcipher_request_set_sync_tfm(nreq, ctx->null); in crypto_rfc4543_copy_src_to_dst() 969 skcipher_request_set_callback(nreq, req->base.flags, NULL, NULL); in crypto_rfc4543_copy_src_to_dst() 970 skcipher_request_set_crypt(nreq, req->src, req->dst, nbytes, NULL); in crypto_rfc4543_copy_src_to_dst() 972 return crypto_skcipher_encrypt(nreq); in crypto_rfc4543_copy_src_to_dst()
|
/linux/drivers/infiniband/hw/mlx4/ |
A D | srq.c | 315 int nreq; in mlx4_ib_post_srq_recv() local 323 nreq = 0; in mlx4_ib_post_srq_recv() 327 for (nreq = 0; wr; ++nreq, wr = wr->next) { in mlx4_ib_post_srq_recv() 359 if (likely(nreq)) { in mlx4_ib_post_srq_recv() 360 srq->wqe_ctr += nreq; in mlx4_ib_post_srq_recv()
|
A D | qp.c | 3253 return cur + nreq >= wq->max_post; in mlx4_wq_overflow() 3491 int nreq; in _mlx4_ib_post_send() local 3527 nreq = 0; in _mlx4_ib_post_send() 3533 for (nreq = 0; wr; ++nreq, wr = wr->next) { in _mlx4_ib_post_send() 3776 if (likely(nreq)) { in _mlx4_ib_post_send() 3777 qp->sq.head += nreq; in _mlx4_ib_post_send() 3811 int nreq; in _mlx4_ib_post_recv() local 3824 nreq = 0; in _mlx4_ib_post_recv() 3830 for (nreq = 0; wr; ++nreq, wr = wr->next) { in _mlx4_ib_post_recv() 3875 if (likely(nreq)) { in _mlx4_ib_post_recv() [all …]
|
/linux/fs/nfs/ |
A D | pnfs_nfs.c | 459 unsigned int nreq = 0; in pnfs_bucket_alloc_ds_commits() local 471 nreq++; in pnfs_bucket_alloc_ds_commits() 475 return nreq; in pnfs_bucket_alloc_ds_commits() 480 return nreq; in pnfs_bucket_alloc_ds_commits() 515 unsigned int nreq = 0; in pnfs_generic_commit_pagelist() local 522 nreq++; in pnfs_generic_commit_pagelist() 525 nreq += pnfs_alloc_ds_commits_list(&list, fl_cinfo, cinfo); in pnfs_generic_commit_pagelist() 526 if (nreq == 0) in pnfs_generic_commit_pagelist()
|
/linux/drivers/crypto/inside-secure/ |
A D | safexcel.c | 816 int ret, nreq = 0, cdesc = 0, rdesc = 0, commands, results; in safexcel_dequeue() local 856 nreq++; in safexcel_dequeue() 867 if (!nreq) in safexcel_dequeue() 872 priv->ring[ring].requests += nreq; in safexcel_dequeue() 1019 int ret, i, nreq, ndesc, tot_descs, handled = 0; in safexcel_handle_result_descriptor() local 1025 nreq = readl(EIP197_HIA_RDR(priv, ring) + EIP197_HIA_xDR_PROC_COUNT); in safexcel_handle_result_descriptor() 1026 nreq >>= EIP197_xDR_PROC_xD_PKT_OFFSET; in safexcel_handle_result_descriptor() 1027 nreq &= EIP197_xDR_PROC_xD_PKT_MASK; in safexcel_handle_result_descriptor() 1028 if (!nreq) in safexcel_handle_result_descriptor() 1031 for (i = 0; i < nreq; i++) { in safexcel_handle_result_descriptor() [all …]
|
/linux/drivers/dma/ |
A D | bcm-sba-raid.c | 297 struct sba_request *nreq; in sba_free_chained_requests() local 303 list_for_each_entry(nreq, &req->next, next) in sba_free_chained_requests() 304 _sba_free_request(sba, nreq); in sba_free_chained_requests() 420 struct sba_request *nreq, *first = req->first; in sba_process_received_request() local 442 list_for_each_entry(nreq, &first->next, next) in sba_process_received_request() 443 _sba_free_request(sba, nreq); in sba_process_received_request() 528 struct sba_request *req, *nreq; in sba_tx_submit() local 540 list_for_each_entry(nreq, &req->next, next) in sba_tx_submit() 541 _sba_pending_request(sba, nreq); in sba_tx_submit()
|
/linux/fs/nilfs2/ |
A D | btree.c | 1727 union nilfs_bmap_ptr_req *nreq, in nilfs_btree_prepare_convert_and_insert() argument 1750 if (nreq != NULL) { in nilfs_btree_prepare_convert_and_insert() 1751 nreq->bpr_ptr = dreq->bpr_ptr + 1; in nilfs_btree_prepare_convert_and_insert() 1752 ret = nilfs_bmap_prepare_alloc_ptr(btree, nreq, dat); in nilfs_btree_prepare_convert_and_insert() 1769 nilfs_bmap_abort_alloc_ptr(btree, nreq, dat); in nilfs_btree_prepare_convert_and_insert() 1783 union nilfs_bmap_ptr_req *nreq, in nilfs_btree_commit_convert_and_insert() argument 1801 if (nreq != NULL) { in nilfs_btree_commit_convert_and_insert() 1803 nilfs_bmap_commit_alloc_ptr(btree, nreq, dat); in nilfs_btree_commit_convert_and_insert() 1819 tmpptr = nreq->bpr_ptr; in nilfs_btree_commit_convert_and_insert() 1855 union nilfs_bmap_ptr_req dreq, nreq, *di, *ni; in nilfs_btree_convert_and_insert() local [all …]
|
/linux/drivers/crypto/qat/qat_common/ |
A D | qat_algs.c | 1224 struct skcipher_request *nreq = skcipher_request_ctx(req); in qat_alg_skcipher_xts_encrypt() local 1230 memcpy(nreq, req, sizeof(*req)); in qat_alg_skcipher_xts_encrypt() 1231 skcipher_request_set_tfm(nreq, ctx->ftfm); in qat_alg_skcipher_xts_encrypt() 1232 return crypto_skcipher_encrypt(nreq); in qat_alg_skcipher_xts_encrypt() 1294 struct skcipher_request *nreq = skcipher_request_ctx(req); in qat_alg_skcipher_xts_decrypt() local 1300 memcpy(nreq, req, sizeof(*req)); in qat_alg_skcipher_xts_decrypt() 1301 skcipher_request_set_tfm(nreq, ctx->ftfm); in qat_alg_skcipher_xts_decrypt() 1302 return crypto_skcipher_decrypt(nreq); in qat_alg_skcipher_xts_decrypt()
|
/linux/drivers/infiniband/hw/hns/ |
A D | hns_roce_hw_v2.c | 707 u32 nreq; in hns_roce_v2_post_send() local 715 nreq = 0; in hns_roce_v2_post_send() 721 for (nreq = 0; wr; ++nreq, wr = wr->next) { in hns_roce_v2_post_send() 756 if (likely(nreq)) { in hns_roce_v2_post_send() 757 qp->sq.head += nreq; in hns_roce_v2_post_send() 858 nreq = 0; in hns_roce_v2_post_recv() 863 for (nreq = 0; wr; ++nreq, wr = wr->next) { in hns_roce_v2_post_recv() 885 if (likely(nreq)) { in hns_roce_v2_post_recv() 992 u32 nreq; in hns_roce_v2_post_srq_recv() local 997 for (nreq = 0; wr; ++nreq, wr = wr->next) { in hns_roce_v2_post_srq_recv() [all …]
|
A D | hns_roce_hw_v1.c | 97 int nreq; in hns_roce_v1_post_send() local 108 for (nreq = 0; wr; ++nreq, wr = wr->next) { in hns_roce_v1_post_send() 109 if (hns_roce_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { in hns_roce_v1_post_send() 115 wqe_idx = (qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1); in hns_roce_v1_post_send() 331 if (likely(nreq)) { in hns_roce_v1_post_send() 332 qp->sq.head += nreq; in hns_roce_v1_post_send() 370 int nreq; in hns_roce_v1_post_recv() local 376 for (nreq = 0; wr; ++nreq, wr = wr->next) { in hns_roce_v1_post_recv() 377 if (hns_roce_wq_overflow(&hr_qp->rq, nreq, in hns_roce_v1_post_recv() 410 if (likely(nreq)) { in hns_roce_v1_post_recv() [all …]
|
A D | hns_roce_qp.c | 1414 bool hns_roce_wq_overflow(struct hns_roce_wq *hr_wq, u32 nreq, in hns_roce_wq_overflow() argument 1421 if (likely(cur + nreq < hr_wq->wqe_cnt)) in hns_roce_wq_overflow() 1429 return cur + nreq >= hr_wq->wqe_cnt; in hns_roce_wq_overflow()
|
A D | hns_roce_device.h | 1246 bool hns_roce_wq_overflow(struct hns_roce_wq *hr_wq, u32 nreq,
|
/linux/drivers/net/ethernet/mellanox/mlx4/ |
A D | main.c | 2943 int nreq = min3(dev->caps.num_ports * in mlx4_enable_msi_x() local 2949 nreq = min_t(int, nreq, msi_x); in mlx4_enable_msi_x() 2951 entries = kcalloc(nreq, sizeof(*entries), GFP_KERNEL); in mlx4_enable_msi_x() 2955 for (i = 0; i < nreq; ++i) in mlx4_enable_msi_x() 2958 nreq = pci_enable_msix_range(dev->persist->pdev, entries, 2, in mlx4_enable_msi_x() 2959 nreq); in mlx4_enable_msi_x() 2961 if (nreq < 0 || nreq < MLX4_EQ_ASYNC) { in mlx4_enable_msi_x() 2966 dev->caps.num_comp_vectors = nreq - 1; in mlx4_enable_msi_x()
|
/linux/drivers/usb/isp1760/ |
A D | isp1760-udc.c | 841 struct isp1760_request *req, *nreq; in isp1760_ep_disable() local 868 list_for_each_entry_safe(req, nreq, &req_list, queue) { in isp1760_ep_disable()
|
/linux/drivers/infiniband/sw/rdmavt/ |
A D | qp.c | 2152 unsigned nreq = 0; in rvt_post_send() local 2179 nreq++; in rvt_post_send() 2183 if (nreq) { in rvt_post_send() 2188 if (nreq == 1 && call_send) in rvt_post_send()
|
/linux/drivers/net/ethernet/neterion/vxge/ |
A D | vxge-config.c | 2321 u32 nreq = 0, i; in __vxge_hw_blockpool_blocks_add() local 2325 nreq = VXGE_HW_INCR_DMA_BLOCK_POOL_SIZE; in __vxge_hw_blockpool_blocks_add() 2326 blockpool->req_out += nreq; in __vxge_hw_blockpool_blocks_add() 2329 for (i = 0; i < nreq; i++) in __vxge_hw_blockpool_blocks_add()
|
/linux/drivers/infiniband/hw/hfi1/ |
A D | tid_rdma.c | 5198 struct tid_rdma_request *req, *nreq; in make_tid_rdma_ack() local 5272 nreq = ack_to_tid_req(&qp->s_ack_queue[next]); in make_tid_rdma_ack() 5273 if (!nreq->comp_seg || nreq->ack_seg == nreq->comp_seg) in make_tid_rdma_ack()
|
/linux/drivers/nvme/host/ |
A D | fc.c | 90 struct nvme_request nreq; /* member 2448 op->nreq.flags |= NVME_REQ_CANCELLED; in nvme_fc_terminate_exchange()
|