Home
last modified time | relevance | path

Searched refs:queues (Results 1 – 25 of 372) sorted by relevance

12345678910>>...15

/linux/net/sched/
A Dsch_multiq.c25 struct Qdisc **queues; member
54 return q->queues[0]; in multiq_classify()
56 return q->queues[band]; in multiq_classify()
167 qdisc_put(q->queues[band]); in multiq_destroy()
169 kfree(q->queues); in multiq_destroy()
221 old = q->queues[i]; in multiq_tune()
222 q->queues[i] = child; in multiq_tune()
242 q->queues = NULL; in multiq_init()
254 if (!q->queues) in multiq_init()
300 return q->queues[band]; in multiq_leaf()
[all …]
A Dsch_prio.c26 struct Qdisc *queues[TCQ_PRIO_BANDS]; member
63 return q->queues[q->prio2band[0]]; in prio_classify()
65 return q->queues[band]; in prio_classify()
137 qdisc_reset(q->queues[prio]); in prio_reset()
175 qdisc_put(q->queues[prio]); in prio_destroy()
203 if (!queues[i]) { in prio_tune()
205 qdisc_put(queues[--i]); in prio_tune()
219 q->queues[i] = queues[i]; in prio_tune()
227 qdisc_put(q->queues[i]); in prio_tune()
324 return q->queues[band]; in prio_leaf()
[all …]
/linux/drivers/staging/wfx/
A Dqueue.c246 WARN_ON(num_queues >= ARRAY_SIZE(queues)); in wfx_tx_queues_get_skb()
247 queues[num_queues] = &wvif->tx_queue[i]; in wfx_tx_queues_get_skb()
249 if (wfx_tx_queue_get_weight(queues[j]) < in wfx_tx_queues_get_skb()
251 swap(queues[j - 1], queues[j]); in wfx_tx_queues_get_skb()
261 skb = skb_dequeue(&queues[i]->cab); in wfx_tx_queues_get_skb()
270 WARN_ON(queues[i] != in wfx_tx_queues_get_skb()
272 atomic_inc(&queues[i]->pending_frames); in wfx_tx_queues_get_skb()
273 trace_queues_stats(wdev, queues[i]); in wfx_tx_queues_get_skb()
282 skb = skb_dequeue(&queues[i]->normal); in wfx_tx_queues_get_skb()
284 atomic_inc(&queues[i]->pending_frames); in wfx_tx_queues_get_skb()
[all …]
/linux/drivers/scsi/aacraid/
A Dcomminit.c373 struct aac_entry * queues; in aac_comm_init() local
399 queues += HOST_NORM_CMD_ENTRIES; in aac_comm_init()
406 queues += HOST_HIGH_CMD_ENTRIES; in aac_comm_init()
413 queues += ADAP_NORM_CMD_ENTRIES; in aac_comm_init()
420 queues += ADAP_HIGH_CMD_ENTRIES; in aac_comm_init()
426 queues += HOST_NORM_RESP_ENTRIES; in aac_comm_init()
433 queues += HOST_HIGH_RESP_ENTRIES; in aac_comm_init()
440 queues += ADAP_NORM_RESP_ENTRIES; in aac_comm_init()
638 if (dev->queues == NULL) { in aac_init_adapter()
644 kfree(dev->queues); in aac_init_adapter()
[all …]
/linux/Documentation/ABI/testing/
A Dsysfs-class-net-queues1 What: /sys/class/<iface>/queues/rx-<queue>/rps_cpus
11 What: /sys/class/<iface>/queues/rx-<queue>/rps_flow_cnt
19 What: /sys/class/<iface>/queues/tx-<queue>/tx_timeout
27 What: /sys/class/<iface>/queues/tx-<queue>/tx_maxrate
35 What: /sys/class/<iface>/queues/tx-<queue>/xps_cpus
45 What: /sys/class/<iface>/queues/tx-<queue>/xps_rxqs
56 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/hold_time
65 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/inflight
73 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/limit
82 What: /sys/class/<iface>/queues/tx-<queue>/byte_queue_limits/limit_max
[all …]
/linux/drivers/nvme/target/
A Dloop.c30 struct nvme_loop_queue *queues; member
71 return queue - queue->ctrl->queues; in nvme_loop_queue_idx()
198 iod->queue = &ctrl->queues[queue_idx]; in nvme_loop_init_iod()
268 nvmet_sq_destroy(&ctrl->queues[0].nvme_sq); in nvme_loop_destroy_admin_queue()
289 kfree(ctrl->queues); in nvme_loop_free_ctrl()
320 ctrl->queues[i].ctrl = ctrl; in nvme_loop_init_io_queues()
365 ctrl->queues[0].ctrl = ctrl; in nvme_loop_configure_admin_queue()
419 nvmet_sq_destroy(&ctrl->queues[0].nvme_sq); in nvme_loop_configure_admin_queue()
614 ctrl->queues = kcalloc(opts->nr_io_queues + 1, sizeof(*ctrl->queues), in nvme_loop_create_ctrl()
616 if (!ctrl->queues) in nvme_loop_create_ctrl()
[all …]
/linux/Documentation/devicetree/bindings/soc/ti/
A Dkeystone-navigator-qmss.txt32 -- managed-queues : the actual queues managed by each queue manager
33 instance, specified as <"base queue #" "# of queues">.
51 - qpend : pool of qpend(interruptible) queues
53 as free descriptor queues or the
54 transmit DMA queues.
58 <"base queue #" "# of queues">.
60 for interruptible queues. The driver additionally sets
67 queues looking for descriptors that have been pushed
90 monitor up to 32 queues starting at the base queue #.
132 managed-queues = <0 0x2000>;
[all …]
/linux/sound/virtio/
A Dvirtio_card.h51 struct virtio_snd_queue queues[VIRTIO_SND_VQ_MAX]; member
70 return &snd->queues[VIRTIO_SND_VQ_CONTROL]; in virtsnd_control_queue()
76 return &snd->queues[VIRTIO_SND_VQ_EVENT]; in virtsnd_event_queue()
82 return &snd->queues[VIRTIO_SND_VQ_TX]; in virtsnd_tx_queue()
88 return &snd->queues[VIRTIO_SND_VQ_RX]; in virtsnd_rx_queue()
/linux/Documentation/networking/device_drivers/ethernet/ti/
A Dcpsw.rst28 - CBS shapers should be used with rated queues
156 // Check if num of queues is set correctly:
175 // Leave last 2 tx queues not rated.
176 $ echo 40 > /sys/class/net/eth0/queues/tx-0/tx_maxrate
181 // Check maximum rate of tx (cpdma) queues:
182 $ cat /sys/class/net/eth0/queues/tx-*/tx_maxrate
206 map 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 queues 1@4 hw 0
387 // Check if num of queues is set correctly:
407 // Leave last 4 tx queues as not rated
415 // Check maximum rate of tx (cpdma) queues:
[all …]
/linux/tools/perf/util/
A Dintel-bts.c46 struct auxtrace_queues queues; member
222 if (bts->queues.new_data) { in intel_bts_update_queues()
223 bts->queues.new_data = false; in intel_bts_update_queues()
539 struct auxtrace_queues *queues = &bts->queues; in intel_bts_process_tid_exit() local
542 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_process_tid_exit()
710 struct auxtrace_queues *queues = &bts->queues; in intel_bts_free_events() local
713 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_free_events()
715 queues->queue_array[i].priv = NULL; in intel_bts_free_events()
717 auxtrace_queues__free(queues); in intel_bts_free_events()
928 if (bts->queues.populated) in intel_bts_process_auxtrace_info()
[all …]
A Darm-spe.c41 struct auxtrace_queues queues; member
628 if (spe->queues.new_data) { in arm_spe__update_queues()
629 spe->queues.new_data = false; in arm_spe__update_queues()
712 struct auxtrace_queues *queues = &spe->queues; in arm_spe_process_timeless_queues() local
716 for (i = 0; i < queues->nr_queues; i++) { in arm_spe_process_timeless_queues()
884 struct auxtrace_queues *queues = &spe->queues; in arm_spe_free_events() local
887 for (i = 0; i < queues->nr_queues; i++) { in arm_spe_free_events()
889 queues->queue_array[i].priv = NULL; in arm_spe_free_events()
891 auxtrace_queues__free(queues); in arm_spe_free_events()
1190 if (spe->queues.populated) in arm_spe_process_auxtrace_info()
[all …]
A Ds390-cpumsf.c169 struct auxtrace_queues queues; member
205 q = &sf->queues.queue_array[sample->cpu]; in s390_cpumcf_dumpctr()
835 if (!sf->queues.new_data) in s390_cpumsf_update_queues()
838 sf->queues.new_data = false; in s390_cpumsf_update_queues()
1015 struct auxtrace_queues *queues = &sf->queues; in s390_cpumsf_free_queues() local
1018 for (i = 0; i < queues->nr_queues; i++) { in s390_cpumsf_free_queues()
1020 queues->queue_array[i].priv; in s390_cpumsf_free_queues()
1032 zfree(&queues->queue_array[i].priv); in s390_cpumsf_free_queues()
1034 auxtrace_queues__free(queues); in s390_cpumsf_free_queues()
1164 if (sf->queues.populated) in s390_cpumsf_process_auxtrace_info()
[all …]
A Dauxtrace.c214 queues->queue_array = auxtrace_alloc_queue_array(queues->nr_queues); in auxtrace_queues__init()
215 if (!queues->queue_array) in auxtrace_queues__init()
249 queues->nr_queues = nr_queues; in auxtrace_queues__grow()
250 queues->queue_array = queue_array; in auxtrace_queues__grow()
284 if (idx >= queues->nr_queues) { in auxtrace_queues__queue_buffer()
290 queue = &queues->queue_array[idx]; in auxtrace_queues__queue_buffer()
302 queues->new_data = true; in auxtrace_queues__queue_buffer()
303 queues->populated = true; in auxtrace_queues__queue_buffer()
456 zfree(&queues->queue_array); in auxtrace_queues__free()
457 queues->nr_queues = 0; in auxtrace_queues__free()
[all …]
/linux/drivers/target/
A Dtarget_core_tmr.c118 flush_work(&dev->queues[i].sq.work); in core_tmr_abort_task()
120 spin_lock_irqsave(&dev->queues[i].lock, flags); in core_tmr_abort_task()
121 list_for_each_entry_safe(se_cmd, next, &dev->queues[i].state_list, in core_tmr_abort_task()
148 spin_unlock_irqrestore(&dev->queues[i].lock, flags); in core_tmr_abort_task()
163 spin_unlock_irqrestore(&dev->queues[i].lock, flags); in core_tmr_abort_task()
292 flush_work(&dev->queues[i].sq.work); in core_tmr_drain_state_list()
294 spin_lock_irqsave(&dev->queues[i].lock, flags); in core_tmr_drain_state_list()
295 list_for_each_entry_safe(cmd, next, &dev->queues[i].state_list, in core_tmr_drain_state_list()
324 spin_unlock_irqrestore(&dev->queues[i].lock, flags); in core_tmr_drain_state_list()
/linux/include/linux/
A Dptr_ring.h625 void ***queues; in ptr_ring_resize_multiple() local
628 queues = kmalloc_array(nrings, sizeof(*queues), gfp); in ptr_ring_resize_multiple()
629 if (!queues) in ptr_ring_resize_multiple()
633 queues[i] = __ptr_ring_init_queue_alloc(size, gfp); in ptr_ring_resize_multiple()
634 if (!queues[i]) in ptr_ring_resize_multiple()
641 queues[i] = __ptr_ring_swap_queue(rings[i], queues[i], in ptr_ring_resize_multiple()
648 kvfree(queues[i]); in ptr_ring_resize_multiple()
650 kfree(queues); in ptr_ring_resize_multiple()
656 kvfree(queues[i]); in ptr_ring_resize_multiple()
658 kfree(queues); in ptr_ring_resize_multiple()
/linux/Documentation/arm/keystone/
A Dknav-qmss.rst15 management of the packet queues. Packets are queued/de-queued by writing or
24 knav qmss driver provides a set of APIs to drivers to open/close qmss queues,
25 allocate descriptor pools, map the descriptors, push/pop to queues etc. For
31 Accumulator QMSS queues using PDSP firmware
34 queue or multiple contiguous queues. drivers/soc/ti/knav_qmss_acc.c is the
37 1 or 32 queues per channel. More description on the firmware is available in
56 Use of accumulated queues requires the firmware image to be present in the
57 file system. The driver doesn't acc queues to the supported queue range if
/linux/Documentation/devicetree/bindings/misc/
A Dintel,ixp4xx-ahb-queue-manager.yaml14 The IXP4xx AHB Queue Manager maintains queues as circular buffers in
17 IXP4xx for accelerating queues, especially for networking. Clients pick
18 queues from the queue manager with foo-queue = <&qmgr N> where the
33 - description: Interrupt for queues 0-31
34 - description: Interrupt for queues 32-63
/linux/drivers/vdpa/virtio_pci/
A Dvp_vdpa.c40 int queues; member
96 for (i = 0; i < vp_vdpa->queues; i++) { in vp_vdpa_free_irq()
142 int queues = vp_vdpa->queues; in vp_vdpa_request_irq() local
143 int vectors = queues + 1; in vp_vdpa_request_irq()
155 for (i = 0; i < queues; i++) { in vp_vdpa_request_irq()
174 irq = pci_irq_vector(pdev, queues); in vp_vdpa_request_irq()
182 vp_modern_config_vector(mdev, queues); in vp_vdpa_request_irq()
480 vp_vdpa->queues = vp_modern_get_num_queues(mdev); in vp_vdpa_probe()
489 vp_vdpa->vring = devm_kcalloc(&pdev->dev, vp_vdpa->queues, in vp_vdpa_probe()
498 for (i = 0; i < vp_vdpa->queues; i++) { in vp_vdpa_probe()
[all …]
/linux/drivers/vdpa/alibaba/
A Deni_vdpa.c45 int queues; member
111 for (i = 0; i < eni_vdpa->queues; i++) { in eni_vdpa_free_irq()
157 int queues = eni_vdpa->queues; in eni_vdpa_request_irq() local
158 int vectors = queues + 1; in eni_vdpa_request_irq()
170 for (i = 0; i < queues; i++) { in eni_vdpa_request_irq()
188 irq = pci_irq_vector(pdev, queues); in eni_vdpa_request_irq()
195 vp_legacy_config_vector(ldev, queues); in eni_vdpa_request_irq()
489 eni_vdpa->queues = eni_vdpa_get_num_queues(eni_vdpa); in eni_vdpa_probe()
498 eni_vdpa->vring = devm_kcalloc(&pdev->dev, eni_vdpa->queues, in eni_vdpa_probe()
507 for (i = 0; i < eni_vdpa->queues; i++) { in eni_vdpa_probe()
[all …]
/linux/Documentation/block/
A Dblk-mq.rst37 spawns multiple queues with individual entry points local to the CPU, removing
49 blk-mq has two group of queues: software staging queues and hardware dispatch
50 queues. When the request arrives at the block layer, it will try the shortest
56 Then, after the requests are processed by software queues, they will be placed
62 Software staging queues
65 The block IO subsystem adds requests in the software staging queues
71 the number of queues is defined by a per-CPU or per-node basis.
99 queue (a.k.a. run the hardware queue), the software queues mapped to that
102 Hardware dispatch queues
116 requests that were ready to be sent first. The number of hardware queues
[all …]
/linux/Documentation/networking/
A Dmultiqueue.rst18 the subqueue memory, as well as netdev configuration of where the queues
21 The base driver will also need to manage the queues as it does the global
33 A new round-robin qdisc, sch_multiq also supports multiple hardware queues. The
35 bands and queues based on the value in skb->queue_mapping. Use this field in
42 On qdisc load, the number of bands is based on the number of queues on the
56 The qdisc will allocate the number of bands to equal the number of queues that
58 queues, the band mapping would look like::
A Dscaling.rst47 Some advanced NICs allow steering packets to queues based on
159 /sys/class/net/<dev>/queues/rx-<n>/rps_cpus
328 /sys/class/net/<dev>/queues/rx-<n>/rps_flow_cnt
406 The goal of this mapping is usually to assign queues
416 2. XPS using receive queues map
420 queues can be mapped to a set of transmit queues (many:many), although
438 mapping, from CPUs to transmit queues or from receive-queues to transmit
474 /sys/class/net/<dev>/queues/tx-<n>/xps_cpus
476 For selection based on receive-queues map::
478 /sys/class/net/<dev>/queues/tx-<n>/xps_rxqs
[all …]
/linux/drivers/staging/qlge/
A DTODO13 * rename "rx" queues to "completion" queues. Calling tx completion queues "rx
14 queues" is confusing.
18 frames, resets the link, device and driver buffer queues become
/linux/arch/arm64/boot/dts/rockchip/
A Drk3568.dtsi64 gmac0_mtl_rx_setup: rx-queues-config {
65 snps,rx-queues-to-use = <1>;
69 gmac0_mtl_tx_setup: tx-queues-config {
70 snps,tx-queues-to-use = <1>;
/linux/Documentation/networking/device_drivers/ethernet/freescale/dpaa2/
A Dethernet-driver.rst25 - queues, channels
32 hardware resources, like queues, do not have a corresponding MC object and
99 queues ---------------------- | | Buffer pool |
109 Frames are transmitted and received through hardware frame queues, which can be
111 enqueues TX frames on egress queues and after transmission is complete a TX
114 When frames are available on ingress queues, a data availability notification
116 queues in the same channel have available frames, only one notification is sent.
119 Each network interface can have multiple Rx, Tx and confirmation queues affined

Completed in 46 milliseconds

12345678910>>...15