Home
last modified time | relevance | path

Searched refs:nr_queues (Results 1 – 25 of 39) sorted by relevance

12

/linux/block/
A Dblk-mq-cpumap.c19 unsigned int nr_queues, const int q) in queue_index() argument
21 return qmap->queue_offset + (q % nr_queues); in queue_index()
38 unsigned int nr_queues = qmap->nr_queues; in blk_mq_map_queues() local
49 if (q >= nr_queues) in blk_mq_map_queues()
51 map[cpu] = queue_index(qmap, nr_queues, q++); in blk_mq_map_queues()
63 if (q < nr_queues) { in blk_mq_map_queues()
64 map[cpu] = queue_index(qmap, nr_queues, q++); in blk_mq_map_queues()
68 map[cpu] = queue_index(qmap, nr_queues, q++); in blk_mq_map_queues()
A Dblk-mq-pci.c32 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_pci_map_queues()
44 WARN_ON_ONCE(qmap->nr_queues > 1); in blk_mq_pci_map_queues()
A Dblk-mq-rdma.c30 for (queue = 0; queue < map->nr_queues; queue++) { in blk_mq_rdma_map_queues()
A Dblk-mq-virtio.c33 for (queue = 0; queue < qmap->nr_queues; queue++) { in blk_mq_virtio_map_queues()
/linux/drivers/crypto/cavium/cpt/
A Dcptvf_main.c42 if (cptvf->nr_queues) { in init_worker_threads()
44 cptvf->nr_queues); in init_worker_threads()
71 cptvf->nr_queues); in cleanup_worker_threads()
104 u32 nr_queues) in alloc_pending_queues() argument
111 pqinfo->nr_queues = nr_queues; in alloc_pending_queues()
144 if (!nr_queues) in init_pending_queues()
150 nr_queues); in init_pending_queues()
323 nr_queues = min_t(u32, nr_queues, max_dev_queues); in cptvf_sw_init()
324 cptvf->nr_queues = nr_queues; in cptvf_sw_init()
329 nr_queues); in cptvf_sw_init()
[all …]
A Dcptvf.h85 u32 nr_queues; /* Number of queues supported */ member
91 for (i = 0, q = &qinfo->queue[i]; i < qinfo->nr_queues; i++, \
110 u32 nr_queues; member
A Dcptvf_reqmanager.c233 if (unlikely(qno >= cptvf->nr_queues)) { in send_cpt_command()
235 qno, cptvf->nr_queues); in send_cpt_command()
545 if (unlikely(qno > cptvf->nr_queues)) { in vq_post_process()
/linux/drivers/crypto/cavium/nitrox/
A Dnitrox_sriov.c58 int nr_queues = 0; in vf_mode_to_nr_queues() local
62 nr_queues = MAX_PF_QUEUES; in vf_mode_to_nr_queues()
65 nr_queues = 8; in vf_mode_to_nr_queues()
68 nr_queues = 4; in vf_mode_to_nr_queues()
71 nr_queues = 2; in vf_mode_to_nr_queues()
74 nr_queues = 1; in vf_mode_to_nr_queues()
78 return nr_queues; in vf_mode_to_nr_queues()
A Dnitrox_lib.c91 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_free_aqm_queues()
102 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_alloc_aqm_queues()
142 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_free_pktin_queues()
155 ndev->pkt_inq = kcalloc_node(ndev->nr_queues, in nitrox_alloc_pktin_queues()
161 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_alloc_pktin_queues()
A Dnitrox_dev.h160 int nr_queues; member
250 u16 nr_queues; member
A Dnitrox_mbx.c67 vfdev->nr_queues = vfdev->msg.data; in pf2vf_send_response()
75 vfdev->nr_queues = 0; in pf2vf_send_response()
A Dnitrox_hal.c124 for (i = 0; i < ndev->nr_queues; i++) { in nitrox_config_pkt_input_rings()
240 for (i = 0; i < ndev->nr_queues; i++) in nitrox_config_pkt_solicit_ports()
356 for (ring = 0; ring < ndev->nr_queues; ring++) { in nitrox_config_aqm_rings()
/linux/tools/perf/util/
A Dauxtrace.c196 if (nr_queues > max_nr_queues) in auxtrace_alloc_queue_array()
203 for (i = 0; i < nr_queues; i++) { in auxtrace_alloc_queue_array()
223 unsigned int nr_queues = queues->nr_queues; in auxtrace_queues__grow() local
227 if (!nr_queues) in auxtrace_queues__grow()
230 while (nr_queues && nr_queues < new_nr_queues) in auxtrace_queues__grow()
231 nr_queues <<= 1; in auxtrace_queues__grow()
233 if (nr_queues < queues->nr_queues || nr_queues < new_nr_queues) in auxtrace_queues__grow()
249 queues->nr_queues = nr_queues; in auxtrace_queues__grow()
284 if (idx >= queues->nr_queues) { in auxtrace_queues__queue_buffer()
457 queues->nr_queues = 0; in auxtrace_queues__free()
[all …]
A Dintel-bts.c211 for (i = 0; i < bts->queues.nr_queues; i++) { in intel_bts_setup_queues()
542 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_process_tid_exit()
713 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_free_events()
A Ds390-cpumsf.c202 if (!sf->use_logfile || sf->queues.nr_queues <= sample->cpu) in s390_cpumcf_dumpctr()
824 for (i = 0; i < sf->queues.nr_queues; i++) { in s390_cpumsf_setup_queues()
1018 for (i = 0; i < queues->nr_queues; i++) { in s390_cpumsf_free_queues()
A Darm-spe.c617 for (i = 0; i < spe->queues.nr_queues; i++) { in arm_spe__setup_queues()
716 for (i = 0; i < queues->nr_queues; i++) { in arm_spe_process_timeless_queues()
887 for (i = 0; i < queues->nr_queues; i++) { in arm_spe_free_events()
A Dintel-pt.c1286 for (i = 0; i < pt->queues.nr_queues; i++) { in intel_pt_first_timestamp()
1455 for (i = 0; i < pt->queues.nr_queues; i++) { in intel_pt_setup_queues()
2449 for (i = 0; i < pt->queues.nr_queues; i++) { in intel_pt_enable_sync_switch()
2685 for (i = 0; i < queues->nr_queues; i++) { in intel_pt_process_timeless_queues()
2758 if (cpu < 0 || !pt->queues.nr_queues) in intel_pt_cpu_to_ptq()
2761 if ((unsigned)cpu >= pt->queues.nr_queues) in intel_pt_cpu_to_ptq()
2762 i = pt->queues.nr_queues - 1; in intel_pt_cpu_to_ptq()
2774 for (; j < pt->queues.nr_queues; j++) { in intel_pt_cpu_to_ptq()
3140 for (i = 0; i < queues->nr_queues; i++) { in intel_pt_free_events()
/linux/drivers/block/rnbd/
A Drnbd-clt.c1194 set->map[HCTX_TYPE_DEFAULT].nr_queues = num_online_cpus(); in rnbd_rdma_map_queues()
1196 set->map[HCTX_TYPE_READ].nr_queues = num_online_cpus(); in rnbd_rdma_map_queues()
1203 set->map[HCTX_TYPE_POLL].nr_queues = sess->nr_poll_queues; in rnbd_rdma_map_queues()
1205 set->map[HCTX_TYPE_READ].nr_queues; in rnbd_rdma_map_queues()
1209 set->map[HCTX_TYPE_DEFAULT].nr_queues, in rnbd_rdma_map_queues()
1210 set->map[HCTX_TYPE_READ].nr_queues, in rnbd_rdma_map_queues()
1211 set->map[HCTX_TYPE_POLL].nr_queues); in rnbd_rdma_map_queues()
1215 set->map[HCTX_TYPE_DEFAULT].nr_queues, in rnbd_rdma_map_queues()
1216 set->map[HCTX_TYPE_READ].nr_queues); in rnbd_rdma_map_queues()
/linux/drivers/block/null_blk/
A Dmain.c1467 if (nullb->nr_queues != 1) in nullb_to_queue()
1468 index = raw_smp_processor_id() / ((nr_cpu_ids + nullb->nr_queues - 1) / nullb->nr_queues); in nullb_to_queue()
1541 map->nr_queues = submit_queues; in null_map_queues()
1544 map->nr_queues = 0; in null_map_queues()
1547 map->nr_queues = poll_queues; in null_map_queues()
1551 qoff += map->nr_queues; in null_map_queues()
1670 for (i = 0; i < nullb->nr_queues; i++) in cleanup_queues()
1681 nullb->nr_queues--; in null_exit_hctx()
1707 nullb->nr_queues++; in null_init_hctx()
1841 nullb->nr_queues++; in init_driver_queues()
A Dnull_blk.h124 unsigned int nr_queues; member
/linux/drivers/char/
A Dvirtio_console.c1846 u32 i, j, nr_ports, nr_queues; in init_vqs() local
1850 nr_queues = use_multiport(portdev) ? (nr_ports + 1) * 2 : 2; in init_vqs()
1852 vqs = kmalloc_array(nr_queues, sizeof(struct virtqueue *), GFP_KERNEL); in init_vqs()
1853 io_callbacks = kmalloc_array(nr_queues, sizeof(vq_callback_t *), in init_vqs()
1855 io_names = kmalloc_array(nr_queues, sizeof(char *), GFP_KERNEL); in init_vqs()
1893 err = virtio_find_vqs(portdev->vdev, nr_queues, vqs, in init_vqs()
/linux/drivers/s390/cio/
A Dqdio_setup.c100 static int __qdio_allocate_qs(struct qdio_q **irq_ptr_qs, int nr_queues) in __qdio_allocate_qs() argument
105 for (i = 0; i < nr_queues; i++) { in __qdio_allocate_qs()
/linux/drivers/nvme/host/
A Dpci.c461 map->nr_queues = dev->io_queues[i]; in nvme_pci_map_queues()
462 if (!map->nr_queues) { in nvme_pci_map_queues()
476 qoff += map->nr_queues; in nvme_pci_map_queues()
477 offset += map->nr_queues; in nvme_pci_map_queues()
2427 int nr_queues = dev->online_queues - 1, sent = 0; in __nvme_disable_io_queues() local
2432 while (nr_queues > 0) { in __nvme_disable_io_queues()
2433 if (nvme_delete_queue(&dev->queues[nr_queues], opcode)) in __nvme_disable_io_queues()
2435 nr_queues--; in __nvme_disable_io_queues()
2439 struct nvme_queue *nvmeq = &dev->queues[nr_queues + sent]; in __nvme_disable_io_queues()
2447 if (nr_queues) in __nvme_disable_io_queues()
A Drdma.c2178 set->map[HCTX_TYPE_DEFAULT].nr_queues = in nvme_rdma_map_queues()
2181 set->map[HCTX_TYPE_READ].nr_queues = in nvme_rdma_map_queues()
2187 set->map[HCTX_TYPE_DEFAULT].nr_queues = in nvme_rdma_map_queues()
2190 set->map[HCTX_TYPE_READ].nr_queues = in nvme_rdma_map_queues()
2201 set->map[HCTX_TYPE_POLL].nr_queues = in nvme_rdma_map_queues()
A Dtcp.c2396 set->map[HCTX_TYPE_DEFAULT].nr_queues = in nvme_tcp_map_queues()
2399 set->map[HCTX_TYPE_READ].nr_queues = in nvme_tcp_map_queues()
2405 set->map[HCTX_TYPE_DEFAULT].nr_queues = in nvme_tcp_map_queues()
2408 set->map[HCTX_TYPE_READ].nr_queues = in nvme_tcp_map_queues()
2417 set->map[HCTX_TYPE_POLL].nr_queues = in nvme_tcp_map_queues()

Completed in 67 milliseconds

12