/linux/drivers/net/ethernet/microsoft/mana/ |
A D | mana_ethtool.c | 21 unsigned int num_queues = apc->num_queues; in mana_get_sset_count() local 32 unsigned int num_queues = apc->num_queues; in mana_get_strings() local 44 for (i = 0; i < num_queues; i++) { in mana_get_strings() 51 for (i = 0; i < num_queues; i++) { in mana_get_strings() 63 unsigned int num_queues = apc->num_queues; in mana_get_ethtool_stats() local 76 for (q = 0; q < num_queues; q++) { in mana_get_ethtool_stats() 89 for (q = 0; q < num_queues; q++) { in mana_get_ethtool_stats() 110 cmd->data = apc->num_queues; in mana_get_rxnfc() 164 if (indir[i] >= apc->num_queues) in mana_set_rxfh() 220 apc->num_queues = new_count; in mana_set_channels() [all …]
|
A D | mana_en.c | 299 unsigned int num_queues = apc->num_queues; in mana_get_stats64() local 310 for (q = 0; q < num_queues; q++) { in mana_get_stats64() 323 for (q = 0; q < num_queues; q++) { in mana_get_stats64() 1140 for (i = 0; i < apc->num_queues; i++) { in mana_destroy_txq() 1191 for (i = 0; i < apc->num_queues; i++) { in mana_create_txq() 1513 for (i = 0; i < apc->num_queues; i++) { in mana_add_rx_queues() 1568 ethtool_rxfh_indir_default(i, apc->num_queues); in mana_rss_table_init() 1611 if (apc->num_queues > apc->max_queues) in mana_init_port() 1612 apc->num_queues = apc->max_queues; in mana_init_port() 1709 for (i = 0; i < apc->num_queues; i++) { in mana_dealloc_queues() [all …]
|
/linux/drivers/net/xen-netback/ |
A D | interface.c | 204 num_queues = READ_ONCE(vif->num_queues); in xenvif_select_queue() 205 if (num_queues < 1) in xenvif_select_queue() 236 num_queues = READ_ONCE(vif->num_queues); in xenvif_start_xmit() 237 if (num_queues < 1) in xenvif_start_xmit() 245 index %= num_queues; in xenvif_start_xmit() 295 num_queues = READ_ONCE(vif->num_queues); in xenvif_get_stats() 319 unsigned int num_queues = vif->num_queues; in xenvif_up() local 335 unsigned int num_queues = vif->num_queues; in xenvif_down() local 447 num_queues = READ_ONCE(vif->num_queues); in xenvif_get_ethtool_stats() 821 unsigned int num_queues = vif->num_queues; in xenvif_disconnect_data() local [all …]
|
A D | xenbus.c | 177 for (i = 0; i < vif->num_queues; ++i) { in xenvif_debugfs_addif() 255 unsigned int num_queues = vif->num_queues; in backend_disconnect() local 267 vif->num_queues = 0; in backend_disconnect() 270 for (queue_index = 0; queue_index < num_queues; ++queue_index) in backend_disconnect() 773 be->vif->num_queues = requested_num_queues; in connect() 791 be->vif->num_queues = queue_index; in connect() 807 be->vif->num_queues = queue_index; in connect() 842 if (be->vif->num_queues > 0) in connect() 848 be->vif->num_queues = 0; in connect() 858 unsigned int num_queues = queue->vif->num_queues; in connect_data_rings() local [all …]
|
/linux/drivers/net/ethernet/google/gve/ |
A D | gve_ethtool.c | 90 for (i = 0; i < priv->rx_cfg.num_queues; i++) { in gve_get_strings() 172 ring < priv->rx_cfg.num_queues; ring++) { in gve_get_ethtool_stats() 195 ring < priv->tx_cfg.num_queues; ring++) { in gve_get_ethtool_stats() 384 cmd->rx_count = priv->rx_cfg.num_queues; in gve_get_channels() 385 cmd->tx_count = priv->tx_cfg.num_queues; in gve_get_channels() 410 priv->tx_cfg.num_queues = new_tx; in gve_set_channels() 411 priv->rx_cfg.num_queues = new_rx; in gve_set_channels() 415 new_tx_cfg.num_queues = new_tx; in gve_set_channels() 416 new_rx_cfg.num_queues = new_rx; in gve_set_channels() 517 priv->tx_cfg.num_queues; in gve_set_priv_flags() [all …]
|
A D | gve_main.c | 138 priv->tx_cfg.num_queues; in gve_alloc_stats_report() 140 priv->rx_cfg.num_queues; in gve_alloc_stats_report() 574 priv->tx_cfg.num_queues); in gve_create_rings() 581 priv->tx_cfg.num_queues); in gve_create_rings() 586 priv->rx_cfg.num_queues); in gve_create_rings() 593 priv->rx_cfg.num_queues); in gve_create_rings() 602 for (i = 0; i < priv->rx_cfg.num_queues; i++) in gve_create_rings() 1136 if (txqueue > priv->tx_cfg.num_queues) in gve_tx_timeout() 1405 priv->tx_cfg.num_queues); in gve_init_priv() 1407 priv->rx_cfg.num_queues); in gve_init_priv() [all …]
|
A D | gve_adminq.c | 516 int gve_adminq_create_tx_queues(struct gve_priv *priv, u32 num_queues) in gve_adminq_create_tx_queues() argument 521 for (i = 0; i < num_queues; i++) { in gve_adminq_create_tx_queues() 572 int gve_adminq_create_rx_queues(struct gve_priv *priv, u32 num_queues) in gve_adminq_create_rx_queues() argument 577 for (i = 0; i < num_queues; i++) { in gve_adminq_create_rx_queues() 604 int gve_adminq_destroy_tx_queues(struct gve_priv *priv, u32 num_queues) in gve_adminq_destroy_tx_queues() argument 609 for (i = 0; i < num_queues; i++) { in gve_adminq_destroy_tx_queues() 636 int gve_adminq_destroy_rx_queues(struct gve_priv *priv, u32 num_queues) in gve_adminq_destroy_rx_queues() argument 641 for (i = 0; i < num_queues; i++) { in gve_adminq_destroy_rx_queues()
|
/linux/drivers/crypto/marvell/octeontx/ |
A D | otx_cptvf_main.c | 40 cptvf->num_queues); in init_worker_threads() 65 cptvf->num_queues); in cleanup_worker_threads() 94 u32 num_queues) in alloc_pending_queues() argument 101 pqinfo->num_queues = num_queues; in alloc_pending_queues() 133 if (!num_queues) in init_pending_queues() 139 num_queues); in init_pending_queues() 153 cptvf->num_queues); in cleanup_pending_queues() 307 num_queues = min_t(u32, num_queues, max_dev_queues); in cptvf_sw_init() 308 cptvf->num_queues = num_queues; in cptvf_sw_init() 313 num_queues); in cptvf_sw_init() [all …]
|
A D | otx_cptvf.h | 52 u32 num_queues; /* Number of queues supported */ member 57 for (i = 0, q = &qinfo->queue[i]; i < qinfo->num_queues; i++, \ 87 u32 num_queues; member
|
/linux/arch/mips/cavium-octeon/executive/ |
A D | cvmx-pko.c | 71 const int num_queues = 1; in __cvmx_pko_iport_config() local 76 for (queue = 0; queue < num_queues; queue++) { in __cvmx_pko_iport_config() 85 config.s.tail = (queue == (num_queues - 1)); in __cvmx_pko_iport_config() 101 num_queues, queue); in __cvmx_pko_iport_config() 327 uint64_t num_queues, in cvmx_pko_config_port() argument 350 (unsigned long long)(base_queue + num_queues)); in cvmx_pko_config_port() 360 for (queue = 0; queue < num_queues; queue++) { in cvmx_pko_config_port() 374 && queue == num_queues - 1) in cvmx_pko_config_port() 417 cvmx_dprintf("num queues: %d (%lld,%lld)\n", num_queues, in cvmx_pko_config_port() 422 for (queue = 0; queue < num_queues; queue++) { in cvmx_pko_config_port() [all …]
|
/linux/drivers/net/wireless/intel/iwlwifi/fw/ |
A D | init.c | 107 int i, num_queues, size, ret; in iwl_configure_rxq() local 125 num_queues = fwrt->trans->num_rx_queues - 1; in iwl_configure_rxq() 127 size = struct_size(cmd, data, num_queues); in iwl_configure_rxq() 133 cmd->num_queues = num_queues; in iwl_configure_rxq() 135 for (i = 0; i < num_queues; i++) { in iwl_configure_rxq()
|
/linux/drivers/net/netdevsim/ |
A D | bus.c | 146 nsim_bus_dev_new(unsigned int id, unsigned int port_count, unsigned int num_queues); 151 unsigned int id, port_count, num_queues; in new_device_store() local 155 err = sscanf(buf, "%u %u %u", &id, &port_count, &num_queues); in new_device_store() 161 num_queues = 1; in new_device_store() 181 nsim_bus_dev = nsim_bus_dev_new(id, port_count, num_queues); in new_device_store() 282 nsim_bus_dev_new(unsigned int id, unsigned int port_count, unsigned int num_queues) in nsim_bus_dev_new() argument 298 nsim_bus_dev->num_queues = num_queues; in nsim_bus_dev_new()
|
/linux/drivers/staging/wfx/ |
A D | queue.c | 237 int i, j, num_queues = 0; in wfx_tx_queues_get_skb() local 246 WARN_ON(num_queues >= ARRAY_SIZE(queues)); in wfx_tx_queues_get_skb() 247 queues[num_queues] = &wvif->tx_queue[i]; in wfx_tx_queues_get_skb() 248 for (j = num_queues; j > 0; j--) in wfx_tx_queues_get_skb() 252 num_queues++; in wfx_tx_queues_get_skb() 260 for (i = 0; i < num_queues; i++) { in wfx_tx_queues_get_skb() 281 for (i = 0; i < num_queues; i++) { in wfx_tx_queues_get_skb()
|
/linux/drivers/net/ |
A D | xen-netfront.c | 363 for (i = 0; i < num_queues; ++i) { in xennet_open() 588 if (num_queues == 1) { in xennet_select_queue() 592 queue_idx = hash % num_queues; in xennet_select_queue() 694 if (num_queues < 1) in xennet_start_xmit() 2168 unsigned int *num_queues) in xennet_create_queues() argument 2188 *num_queues = i; in xennet_create_queues() 2196 *num_queues = i; in xennet_create_queues() 2208 if (*num_queues == 0) { in xennet_create_queues() 2226 unsigned int num_queues = 1; in talk_to_netback() local 2301 if (num_queues == 1) { in talk_to_netback() [all …]
|
/linux/drivers/scsi/ |
A D | virtio_scsi.c | 78 u32 num_queues; member 205 num_vqs = vscsi->num_queues; in virtscsi_poll_requests() 796 num_vqs = vscsi->num_queues + VIRTIO_SCSI_VQ_BASE; in virtscsi_init() 848 u32 num_queues; in virtscsi_probe() local 857 num_queues = virtscsi_config_get(vdev, num_queues) ? : 1; in virtscsi_probe() 858 num_queues = min_t(unsigned int, nr_cpu_ids, num_queues); in virtscsi_probe() 863 struct_size(vscsi, req_vqs, num_queues)); in virtscsi_probe() 871 vscsi->num_queues = num_queues; in virtscsi_probe() 891 shost->nr_hw_queues = num_queues; in virtscsi_probe()
|
/linux/drivers/soc/ti/ |
A D | knav_qmss_acc.c | 33 for (queue = 0; queue < range->num_queues; queue++) { in __knav_acc_notify() 150 queue >= range_base + range->num_queues) { in knav_acc_int_handler() 154 range_base + range->num_queues); in knav_acc_int_handler() 313 queue_mask = BIT(range->num_queues) - 1; in knav_acc_setup_cmd() 379 for (queue = 0; queue < range->num_queues; queue++) { in knav_acc_init_range() 439 channels = range->num_queues; in knav_acc_free_range() 523 channels = range->num_queues; in knav_init_acc_range() 533 if (range->num_queues > 32) { in knav_init_acc_range()
|
A D | knav_qmss_queue.c | 208 kdev->base_id + kdev->num_queues > id) { in knav_queue_find_by_id() 473 kdev->base_id + kdev->num_queues - 1); in knav_queue_debug_show() 1231 range->num_queues = temp[1]; in knav_setup_queue_range() 1286 range->queue_base + range->num_queues); in knav_setup_queue_range() 1299 range->queue_base + range->num_queues - 1, in knav_setup_queue_range() 1304 kdev->num_queues_in_use += range->num_queues; in knav_setup_queue_range() 1412 qmgr->num_queues = temp[1]; in knav_queue_init_qmgrs() 1420 qmgr->start_queue, qmgr->num_queues); in knav_queue_init_qmgrs() 1485 qmgr->start_queue, qmgr->num_queues, in knav_queue_init_qmgrs() 1677 (id < qmgr->start_queue + qmgr->num_queues)) in knav_find_qmgr() [all …]
|
A D | knav_qmss.h | 140 unsigned num_queues; member 297 unsigned num_queues; member 332 unsigned num_queues; member
|
/linux/drivers/net/ethernet/qlogic/qede/ |
A D | qede.h | 186 #define QEDE_MAX_RSS_CNT(edev) ((edev)->dev_info.num_queues) 187 #define QEDE_MAX_TSS_CNT(edev) ((edev)->dev_info.num_queues) 200 u16 num_queues; member 203 #define QEDE_QUEUE_CNT(edev) ((edev)->num_queues) 204 #define QEDE_RSS_COUNT(edev) ((edev)->num_queues - (edev)->fp_num_tx) 206 #define QEDE_TSS_COUNT(edev) ((edev)->num_queues - (edev)->fp_num_rx) 606 #define for_each_queue(i) for (i = 0; i < edev->num_queues; i++)
|
/linux/drivers/net/ethernet/cavium/liquidio/ |
A D | cn23xx_vf_device.c | 50 static int cn23xx_vf_reset_io_queues(struct octeon_device *oct, u32 num_queues) in cn23xx_vf_reset_io_queues() argument 57 for (q_no = 0; q_no < num_queues; q_no++) { in cn23xx_vf_reset_io_queues() 67 for (q_no = 0; q_no < num_queues; q_no++) { in cn23xx_vf_reset_io_queues() 361 u32 num_queues = oct->num_iqs; in cn23xx_disable_vf_io_queues() local 366 if (num_queues < oct->num_oqs) in cn23xx_disable_vf_io_queues() 367 num_queues = oct->num_oqs; in cn23xx_disable_vf_io_queues() 369 cn23xx_vf_reset_io_queues(oct, num_queues); in cn23xx_disable_vf_io_queues()
|
/linux/drivers/gpu/drm/amd/amdkfd/ |
A D | kfd_pm4_headers.h | 88 uint32_t num_queues:10; member 141 uint32_t num_queues:10; member
|
A D | kfd_pm4_headers_vi.h | 191 uint32_t num_queues:10; member 240 uint32_t num_queues:3; member 361 uint32_t num_queues:3; member
|
A D | kfd_packet_manager_v9.c | 49 packet->bitfields14.num_queues = (qpd->is_debug) ? 0 : qpd->queue_count; in pm_map_process_v9() 94 packet->bitfields14.num_queues = (qpd->is_debug) ? 0 : qpd->queue_count; in pm_map_process_aldebaran() 192 packet->bitfields2.num_queues = 1; in pm_map_queues_v9() 298 packet->bitfields2.num_queues = 1; in pm_unmap_queues_v9()
|
A D | kfd_pm4_headers_ai.h | 184 uint32_t num_queues:10; member 285 uint32_t num_queues:3; member 408 uint32_t num_queues:3; member
|
/linux/drivers/net/ethernet/cadence/ |
A D | macb_main.c | 3489 cmd->data = bp->num_queues; in gem_get_rxnfc() 3517 || (cmd->fs.ring_cookie >= bp->num_queues)) { in gem_set_rxnfc() 3730 unsigned int *num_queues) in macb_probe_queues() argument 3733 *num_queues = 1; in macb_probe_queues() 3746 *num_queues = hweight32(*queue_mask); in macb_probe_queues() 4668 unsigned int queue_mask, num_queues; in macb_probe() local 4704 dev = alloc_etherdev_mq(sizeof(*bp), num_queues); in macb_probe() 4726 bp->num_queues = num_queues; in macb_probe() 4889 for (q = 0, queue = bp->queues; q < bp->num_queues; in macb_suspend() 4932 for (q = 0, queue = bp->queues; q < bp->num_queues; in macb_suspend() [all …]
|