Home
last modified time | relevance | path

Searched refs:nq (Results 1 – 25 of 33) sorted by relevance

12

/linux/drivers/block/null_blk/
A Dmain.c677 wake_up(&nq->wait); in put_tag()
685 tag = find_first_zero_bit(nq->tag_map, nq->queue_depth); in get_tag()
705 tag = get_tag(nq); in __alloc_cmd()
710 cmd->nq = nq; in __alloc_cmd()
1629 cmd->nq = nq; in null_queue_rq()
1662 kfree(nq->tag_map); in cleanup_queue()
1663 kfree(nq->cmds); in cleanup_queue()
1793 nq->cmds = kcalloc(nq->queue_depth, sizeof(*cmd), GFP_KERNEL); in setup_commands()
1794 if (!nq->cmds) in setup_commands()
1799 if (!nq->tag_map) { in setup_commands()
[all …]
A Dzoned.c369 struct nullb_device *dev = cmd->nq->dev; in null_zone_write()
603 struct nullb_device *dev = cmd->nq->dev; in null_zone_mgmt()
672 dev = cmd->nq->dev; in null_process_zoned_cmd()
A Dnull_blk.h23 struct nullb_queue *nq; member
/linux/drivers/infiniband/hw/bnxt_re/
A Dqplib_fp.c158 struct bnxt_qplib_nq *nq = nq_work->nq; in bnxt_qpn_cqn_sched_task() local
166 nq->cqn_handler(nq, cq); in bnxt_qpn_cqn_sched_task()
300 struct bnxt_qplib_nq *nq = from_tasklet(nq, t, nq_tasklet); in bnxt_qplib_service_nq() local
342 if (!nq->cqn_handler(nq, (cq))) in bnxt_qplib_service_nq()
363 if (!nq->srqn_handler(nq, in bnxt_qplib_service_nq()
416 free_irq(nq->msix_vec, nq); in bnxt_qplib_nq_stop_irq()
456 rc = request_irq(nq->msix_vec, bnxt_qplib_nq_irq, 0, nq->name, nq); in bnxt_qplib_nq_start_irq()
462 rc = irq_set_affinity_hint(nq->msix_vec, &nq->mask); in bnxt_qplib_nq_start_irq()
547 bnxt_qplib_free_hwq(nq->res, &nq->hwq); in bnxt_qplib_free_nq()
1935 nq_work->nq = qp->scq->nq; in bnxt_qplib_post_send()
[all …]
A Dmain.c301 struct bnxt_qplib_nq *nq; in bnxt_re_stop_irq() local
305 nq = &rdev->nq[indx - 1]; in bnxt_re_stop_irq()
306 bnxt_qplib_nq_stop_irq(nq, false); in bnxt_re_stop_irq()
317 struct bnxt_qplib_nq *nq; in bnxt_re_start_irq() local
339 nq = &rdev->nq[indx - 1]; in bnxt_re_start_irq()
984 bnxt_qplib_free_nq(&rdev->nq[i]); in bnxt_re_free_nq_res()
985 rdev->nq[i].res = NULL; in bnxt_re_free_nq_res()
1030 struct bnxt_qplib_nq *nq; in bnxt_re_alloc_res() local
1032 nq = &rdev->nq[i]; in bnxt_re_alloc_res()
1042 rattr.pages = nq->hwq.pbl[rdev->nq[i].hwq.level].pg_count; in bnxt_re_alloc_res()
[all …]
A Dqplib_fp.h404 struct bnxt_qplib_nq *nq; member
466 typedef int (*cqn_handler_t)(struct bnxt_qplib_nq *nq,
468 typedef int (*srqn_handler_t)(struct bnxt_qplib_nq *nq,
491 struct bnxt_qplib_nq *nq; member
495 void bnxt_qplib_nq_stop_irq(struct bnxt_qplib_nq *nq, bool kill);
496 void bnxt_qplib_disable_nq(struct bnxt_qplib_nq *nq);
497 int bnxt_qplib_nq_start_irq(struct bnxt_qplib_nq *nq, int nq_indx,
499 int bnxt_qplib_enable_nq(struct pci_dev *pdev, struct bnxt_qplib_nq *nq,
540 void bnxt_qplib_free_nq(struct bnxt_qplib_nq *nq);
541 int bnxt_qplib_alloc_nq(struct bnxt_qplib_res *res, struct bnxt_qplib_nq *nq);
A Dib_verbs.c1564 nq = qplib_srq->cq->nq; in bnxt_re_destroy_srq()
1568 if (nq) in bnxt_re_destroy_srq()
1569 nq->budget--; in bnxt_re_destroy_srq()
1651 nq = &rdev->nq[0]; in bnxt_re_create_srq()
1677 if (nq) in bnxt_re_create_srq()
1678 nq->budget++; in bnxt_re_create_srq()
2800 nq = cq->qplib_cq.nq; in bnxt_re_destroy_cq()
2806 nq->budget--; in bnxt_re_destroy_cq()
2874 nq = &rdev->nq[nq_alloc_cnt % (rdev->num_msix - 1)]; in bnxt_re_create_cq()
2877 cq->qplib_cq.nq = nq; in bnxt_re_create_cq()
[all …]
A Dbnxt_re.h155 struct bnxt_qplib_nq nq[BNXT_RE_MAX_MSIX]; member
/linux/lib/crypto/
A Dcurve25519-hacl64.c551 point_swap_conditional(nq, nqpq, bit0); in ladder_smallloop_cmult_small_loop_step()
552 addanddouble_fmonty(nq2, nqpq2, nq, nqpq, q); in ladder_smallloop_cmult_small_loop_step()
572 ladder_smallloop_cmult_small_loop_double_step(nq, nqpq, nq2, in ladder_smallloop_cmult_small_loop()
585 ladder_smallloop_cmult_small_loop(nq, nqpq, nq2, nqpq2, q, in ladder_bigloop_cmult_big_loop()
593 u64 *nq = point_buf; in ladder_cmult() local
598 nq[0] = 1; in ladder_cmult()
600 point_copy(result, nq); in ladder_cmult()
778 u64 *nq = buf; in curve25519_generic() local
779 u64 *x = nq; in curve25519_generic()
781 ladder_cmult(nq, scalar, q); in curve25519_generic()
[all …]
/linux/fs/xfs/
A Dxfs_trans_dquot.c78 struct xfs_dqtrx *oq, *nq; in xfs_trans_dup_dqinfo() local
97 nq = &nqa[i]; in xfs_trans_dup_dqinfo()
102 nq->qt_dquot = oq->qt_dquot; in xfs_trans_dup_dqinfo()
103 nq->qt_bcount_delta = nq->qt_icount_delta = 0; in xfs_trans_dup_dqinfo()
104 nq->qt_rtbcount_delta = 0; in xfs_trans_dup_dqinfo()
109 nq->qt_blk_res = oq->qt_blk_res - blk_res_used; in xfs_trans_dup_dqinfo()
112 nq->qt_rtblk_res = oq->qt_rtblk_res - in xfs_trans_dup_dqinfo()
116 nq->qt_ino_res = oq->qt_ino_res - oq->qt_ino_res_used; in xfs_trans_dup_dqinfo()
/linux/arch/x86/crypto/
A Dcurve25519-x86_64.c766 u64 *nq = p01_tmp1; in point_add_and_double() local
770 u64 *x2 = nq; in point_add_and_double()
771 u64 *z2 = nq + (u32)4U; in point_add_and_double()
813 fmul2(nq, dc1, ab1, tmp2); in point_add_and_double()
817 static void point_double(u64 *nq, u64 *tmp1, u64 *tmp2) in point_double() argument
819 u64 *x2 = nq; in point_double()
820 u64 *z2 = nq + (u32)4U; in point_double()
837 fmul2(nq, dc, ab, tmp2); in point_double()
/linux/drivers/net/ethernet/chelsio/cxgb4/
A Dcxgb4_uld.c111 unsigned int nq = rxq_info->nrxq + rxq_info->nciq; in alloc_uld_rxqs() local
125 for (i = 0; i < nq; i++, q++) { in alloc_uld_rxqs()
403 int nq = txq_info->ntxq; in free_sge_txq_uld() local
406 for (i = 0; i < nq; i++) { in free_sge_txq_uld()
426 int nq = txq_info->ntxq; in alloc_sge_txq_uld() local
429 j = nq / adap->params.nports; in alloc_sge_txq_uld()
430 for (i = 0; i < nq; i++) { in alloc_sge_txq_uld()
/linux/drivers/net/ethernet/marvell/
A Dmvneta.c1891 netif_tx_wake_queue(nq); in mvneta_txq_done()
2097 struct netdev_queue *nq; in mvneta_xdp_xmit_back() local
2110 __netif_tx_lock(nq, cpu); in mvneta_xdp_xmit_back()
2125 __netif_tx_unlock(nq); in mvneta_xdp_xmit_back()
2139 struct netdev_queue *nq; in mvneta_xdp_xmit() local
2151 __netif_tx_lock(nq, cpu); in mvneta_xdp_xmit()
2163 __netif_tx_unlock(nq); in mvneta_xdp_xmit()
2849 netif_tx_stop_queue(nq); in mvneta_tx()
2892 struct netdev_queue *nq; in mvneta_tx_done_gbe() local
2899 __netif_tx_lock(nq, cpu); in mvneta_tx_done_gbe()
[all …]
A Dmv643xx_eth.c497 if (netif_tx_queue_stopped(nq)) { in txq_maybe_wake()
498 __netif_tx_lock(nq, smp_processor_id()); in txq_maybe_wake()
500 netif_tx_wake_queue(nq); in txq_maybe_wake()
501 __netif_tx_unlock(nq); in txq_maybe_wake()
997 struct netdev_queue *nq; in mv643xx_eth_xmit() local
1001 nq = netdev_get_tx_queue(dev, queue); in mv643xx_eth_xmit()
1020 netif_tx_stop_queue(nq); in mv643xx_eth_xmit()
1038 __netif_tx_lock(nq, smp_processor_id()); in txq_kick()
1051 __netif_tx_unlock(nq); in txq_kick()
1062 __netif_tx_lock_bh(nq); in txq_reclaim()
[all …]
/linux/drivers/net/ethernet/chelsio/cxgb4vf/
A Dt4vf_hw.c1272 int nq = min(n, 32); in t4vf_config_rss_range() local
1279 cmd.niqid = cpu_to_be16(nq); in t4vf_config_rss_range()
1285 start += nq; in t4vf_config_rss_range()
1286 n -= nq; in t4vf_config_rss_range()
1293 while (nq > 0) { in t4vf_config_rss_range()
1302 int nqbuf = min(3, nq); in t4vf_config_rss_range()
1304 nq -= nqbuf; in t4vf_config_rss_range()
/linux/kernel/rcu/
A Dtasks.h852 int nq = READ_ONCE(t->trc_reader_special.b.need_qs); in rcu_read_unlock_trace_special() local
858 if (nq) in rcu_read_unlock_trace_special()
861 if (nq && atomic_dec_and_test(&trc_n_readers_need_end)) in rcu_read_unlock_trace_special()
/linux/arch/s390/mm/
A Dpgtable.c787 unsigned char key, bool nq) in set_guest_storage_key() argument
842 page_set_storage_key(paddr, skey, !nq); in set_guest_storage_key()
866 bool nq, bool mr, bool mc) in cond_set_guest_storage_key() argument
885 rc = set_guest_storage_key(current->mm, addr, key, nq); in cond_set_guest_storage_key()
/linux/Documentation/devicetree/bindings/display/
A Darm,komeda.txt18 Required properties for sub-node: pipeline@nq
/linux/drivers/net/
A Dtap.c190 struct tap_queue *nq; in tap_disable_queue() local
201 nq = rtnl_dereference(tap->taps[tap->numvtaps - 1]); in tap_disable_queue()
202 nq->queue_index = index; in tap_disable_queue()
204 rcu_assign_pointer(tap->taps[index], nq); in tap_disable_queue()
/linux/drivers/net/ethernet/intel/igc/
A Digc_main.c2198 struct netdev_queue *nq; in igc_xdp_xmit_back() local
2206 nq = txring_txq(ring); in igc_xdp_xmit_back()
2208 __netif_tx_lock(nq, cpu); in igc_xdp_xmit_back()
2210 __netif_tx_unlock(nq); in igc_xdp_xmit_back()
2277 struct netdev_queue *nq; in igc_finalize_xdp() local
2282 nq = txring_txq(ring); in igc_finalize_xdp()
2286 __netif_tx_unlock(nq); in igc_finalize_xdp()
2647 __netif_tx_unlock(nq); in igc_xdp_xmit_zc()
6022 struct netdev_queue *nq; in igc_xdp_xmit() local
6033 nq = txring_txq(ring); in igc_xdp_xmit()
[all …]
/linux/net/sched/
A Dsch_api.c318 struct netdev_queue *nq; in qdisc_lookup_rcu() local
327 nq = dev_ingress_queue_rcu(dev); in qdisc_lookup_rcu()
328 if (nq) in qdisc_lookup_rcu()
329 q = qdisc_match_from_root(nq->qdisc_sleeping, handle); in qdisc_lookup_rcu()
/linux/drivers/net/ethernet/freescale/
A Dfec_main.c833 struct netdev_queue *nq; in fec_enet_start_xmit() local
838 nq = netdev_get_tx_queue(ndev, queue); in fec_enet_start_xmit()
849 netif_tx_stop_queue(nq); in fec_enet_start_xmit()
1286 struct netdev_queue *nq; in fec_enet_tx_queue() local
1294 nq = netdev_get_tx_queue(ndev, queue_id); in fec_enet_tx_queue()
1374 if (netif_tx_queue_stopped(nq)) { in fec_enet_tx_queue()
1377 netif_tx_wake_queue(nq); in fec_enet_tx_queue()
/linux/arch/s390/kvm/
A Dpriv.c1018 bool mr = false, mc = false, nq; in handle_pfmf() local
1048 nq = vcpu->run->s.regs.gprs[reg1] & PFMF_NQ; in handle_pfmf()
1100 key, NULL, nq, mr, mc); in handle_pfmf()
/linux/drivers/net/ethernet/stmicro/stmmac/
A Dstmmac_main.c2397 nq->trans_start = jiffies; in stmmac_xdp_xmit_zc()
4688 struct netdev_queue *nq; in stmmac_xdp_xmit_back() local
4696 nq = netdev_get_tx_queue(priv->dev, queue); in stmmac_xdp_xmit_back()
4698 __netif_tx_lock(nq, cpu); in stmmac_xdp_xmit_back()
4700 nq->trans_start = jiffies; in stmmac_xdp_xmit_back()
4706 __netif_tx_unlock(nq); in stmmac_xdp_xmit_back()
6323 struct netdev_queue *nq; in stmmac_xdp_xmit() local
6334 nq = netdev_get_tx_queue(priv->dev, queue); in stmmac_xdp_xmit()
6336 __netif_tx_lock(nq, cpu); in stmmac_xdp_xmit()
6338 nq->trans_start = jiffies; in stmmac_xdp_xmit()
[all …]
/linux/drivers/net/ethernet/freescale/dpaa2/
A Ddpaa2-eth.c1109 struct netdev_queue *nq; in __dpaa2_eth_tx() local
1180 nq = netdev_get_tx_queue(net_dev, queue_mapping); in __dpaa2_eth_tx()
1181 netdev_tx_sent_queue(nq, fd_len); in __dpaa2_eth_tx()
1196 netdev_tx_completed_queue(nq, 1, fd_len); in __dpaa2_eth_tx()
1563 struct netdev_queue *nq; in dpaa2_eth_poll() local
1630 nq = netdev_get_tx_queue(priv->net_dev, txc_fq->flowid); in dpaa2_eth_poll()
1631 netdev_tx_completed_queue(nq, txc_fq->dq_frames, in dpaa2_eth_poll()

Completed in 179 milliseconds

12