/linux/drivers/net/ethernet/aquantia/atlantic/ |
A D | aq_vec.c | 23 unsigned int tx_rings; member 47 self->tx_rings > i; ++i, ring = self->ring[i]) { in aq_vec_poll() 124 self->tx_rings = 0; in aq_vec_alloc() 152 ++self->tx_rings; in aq_vec_ring_alloc() 186 self->tx_rings > i; ++i, ring = self->ring[i]) { in aq_vec_init() 228 self->tx_rings > i; ++i, ring = self->ring[i]) { in aq_vec_start() 252 self->tx_rings > i; ++i, ring = self->ring[i]) { in aq_vec_stop() 272 self->tx_rings > i; ++i, ring = self->ring[i]) { in aq_vec_deinit() 301 self->tx_rings > i; ++i, ring = self->ring[i]) { in aq_vec_ring_free() 307 self->tx_rings = 0; in aq_vec_ring_free() [all …]
|
A D | aq_hw.h | 66 u8 tx_rings; member
|
/linux/drivers/net/wireless/realtek/rtw88/ |
A D | pci.c | 174 tx_ring = &rtwpci->tx_rings[i]; in rtw_pci_free_trx_ring() 335 tx_ring = &rtwpci->tx_rings[i]; in rtw_pci_init_trx_ring() 357 tx_ring = &rtwpci->tx_rings[i]; in rtw_pci_init_trx_ring() 425 rtwpci->tx_rings[RTW_TX_QUEUE_BK].r.rp = 0; in rtw_pci_reset_buf_desc() 426 rtwpci->tx_rings[RTW_TX_QUEUE_BK].r.wp = 0; in rtw_pci_reset_buf_desc() 550 tx_ring = &rtwpci->tx_rings[queue]; in rtw_pci_dma_release() 626 tx_ring = &rtwpci->tx_rings[queue]; in rtw_pci_deep_ps_enter() 807 ring = &rtwpci->tx_rings[queue]; in rtw_pci_tx_kick_off_queue() 842 ring = &rtwpci->tx_rings[queue]; in rtw_pci_tx_write_data() 955 ring = &rtwpci->tx_rings[queue]; in rtw_pci_tx_write() [all …]
|
A D | pci.h | 223 struct rtw_pci_tx_ring tx_rings[RTK_MAX_TX_QUEUE_NUM]; member
|
/linux/drivers/net/ethernet/intel/ice/ |
A D | ice_ethtool.c | 2781 tx_rings = kcalloc(vsi->num_txq, sizeof(*tx_rings), GFP_KERNEL); in ice_set_ringparam() 2782 if (!tx_rings) { in ice_set_ringparam() 2789 tx_rings[i] = *vsi->tx_rings[i]; in ice_set_ringparam() 2791 tx_rings[i].desc = NULL; in ice_set_ringparam() 2792 tx_rings[i].tx_buf = NULL; in ice_set_ringparam() 2797 kfree(tx_rings); in ice_set_ringparam() 2882 if (tx_rings) { in ice_set_ringparam() 2885 *vsi->tx_rings[i] = tx_rings[i]; in ice_set_ringparam() 2887 kfree(tx_rings); in ice_set_ringparam() 2924 if (tx_rings) { in ice_set_ringparam() [all …]
|
A D | ice_lib.c | 84 if (!vsi->tx_rings) in ice_vsi_alloc_arrays() 134 devm_kfree(dev, vsi->tx_rings); in ice_vsi_alloc_arrays() 334 if (vsi->tx_rings) { in ice_vsi_free_arrays() 336 vsi->tx_rings = NULL; in ice_vsi_free_arrays() 1367 if (vsi->tx_rings) { in ice_vsi_clear_rings() 1369 if (vsi->tx_rings[i]) { in ice_vsi_clear_rings() 1843 if (q_idx >= vsi->alloc_txq || !tx_rings || !tx_rings[q_idx]) in ice_vsi_cfg_single_txq() 2861 if (!vsi->tx_rings) in ice_vsi_free_tx_rings() 2865 if (vsi->tx_rings[i] && vsi->tx_rings[i]->desc) in ice_vsi_free_tx_rings() 2973 if (vsi->tx_rings) { in ice_vsi_dis_irq() [all …]
|
A D | ice_xsk.c | 29 memset(&vsi->tx_rings[q_idx]->stats, 0, in ice_qp_reset_stats() 30 sizeof(vsi->tx_rings[q_idx]->stats)); in ice_qp_reset_stats() 43 ice_clean_tx_ring(vsi->tx_rings[q_idx]); in ice_qp_clean_rings() 162 tx_ring = vsi->tx_rings[q_idx]; in ice_qp_dis() 227 tx_ring = vsi->tx_rings[q_idx]; in ice_qp_ena()
|
A D | ice_dcb_lib.c | 207 return vsi->tx_rings[queue_index]->dcb_tc; in ice_dcb_get_tc() 224 tx_ring = vsi->tx_rings[i]; in ice_vsi_cfg_dcb_rings() 241 vsi->tx_rings[i]->dcb_tc = n; in ice_vsi_cfg_dcb_rings() 263 vsi->tx_rings[i]->dcb_tc = first_droptc; in ice_vsi_cfg_dcb_rings()
|
A D | ice_lib.h | 17 int ice_vsi_cfg_single_txq(struct ice_vsi *vsi, struct ice_tx_ring **tx_rings, u16 q_idx);
|
A D | ice_base.c | 255 if (vsi->tx_rings[i] == ring) in ice_eswitch_calc_txq_handle() 724 struct ice_tx_ring *tx_ring = vsi->tx_rings[q_id]; in ice_vsi_map_rings_to_vectors()
|
A D | ice_main.c | 116 struct ice_tx_ring *tx_ring = vsi->tx_rings[i]; in ice_check_for_hang_subtask() 5496 q_handle = vsi->tx_rings[queue_index]->q_handle; in ice_set_tx_maxrate() 5983 ice_update_vsi_tx_ring_stats(vsi, vsi_stats, vsi->tx_rings, in ice_update_vsi_ring_stats() 6315 ice_clean_tx_ring(vsi->tx_rings[i]); in ice_down() 6346 struct ice_tx_ring *ring = vsi->tx_rings[i]; in ice_vsi_setup_tx_rings() 7313 if (vsi->tx_rings[i] && vsi->tx_rings[i]->desc) in ice_tx_timeout() 7314 if (txqueue == vsi->tx_rings[i]->q_index) { in ice_tx_timeout() 7315 tx_ring = vsi->tx_rings[i]; in ice_tx_timeout() 7610 tx_ring = vsi->tx_rings[ch->base_q + i]; in ice_chnl_cfg_res() 7894 tx_ring = vsi->tx_rings[ch->base_q + i]; in ice_remove_q_channels()
|
/linux/drivers/thunderbolt/ |
A D | nhi.c | 471 if (!nhi->tx_rings[i]) { in nhi_alloc_hop() 489 if (ring->is_tx && nhi->tx_rings[ring->hop]) { in nhi_alloc_hop() 502 nhi->tx_rings[ring->hop] = ring; in nhi_alloc_hop() 753 ring->nhi->tx_rings[ring->hop] = NULL; in tb_ring_free() 875 ring = nhi->tx_rings[hop]; in nhi_interrupt_work() 1073 if (nhi->tx_rings[i]) in nhi_shutdown() 1214 nhi->tx_rings = devm_kcalloc(&pdev->dev, nhi->hop_count, in nhi_probe() 1215 sizeof(*nhi->tx_rings), GFP_KERNEL); in nhi_probe() 1218 if (!nhi->tx_rings || !nhi->rx_rings) in nhi_probe()
|
/linux/drivers/net/ethernet/intel/i40e/ |
A D | i40e_ethtool.c | 2022 if (!tx_rings) { in i40e_set_ringparam() 2031 tx_rings[i] = *vsi->tx_rings[i]; in i40e_set_ringparam() 2046 kfree(tx_rings); in i40e_set_ringparam() 2047 tx_rings = NULL; in i40e_set_ringparam() 2113 if (tx_rings) { in i40e_set_ringparam() 2117 *vsi->tx_rings[i] = tx_rings[i]; in i40e_set_ringparam() 2120 kfree(tx_rings); in i40e_set_ringparam() 2121 tx_rings = NULL; in i40e_set_ringparam() 2150 if (tx_rings) { in i40e_set_ringparam() 2155 kfree(tx_rings); in i40e_set_ringparam() [all …]
|
A D | i40e_main.c | 330 if (vsi->tx_rings[i] && vsi->tx_rings[i]->desc) { in i40e_tx_timeout() 333 tx_ring = vsi->tx_rings[i]; in i40e_tx_timeout() 441 if (!vsi->tx_rings) in i40e_get_netdev_stats_struct() 3097 if (vsi->tx_rings) { in i40e_vsi_free_tx_resources() 3099 if (vsi->tx_rings[i] && vsi->tx_rings[i]->desc) in i40e_vsi_free_tx_resources() 3520 tx_ring = vsi->tx_rings[i]; in i40e_vsi_config_dcb_rings() 11136 if (!vsi->tx_rings) in i40e_vsi_alloc_arrays() 11157 kfree(vsi->tx_rings); in i40e_vsi_alloc_arrays() 11267 kfree(vsi->tx_rings); in i40e_vsi_free_arrays() 11268 vsi->tx_rings = NULL; in i40e_vsi_free_arrays() [all …]
|
A D | i40e_debugfs.c | 291 struct i40e_ring *tx_ring = READ_ONCE(vsi->tx_rings[i]); in i40e_dbg_dump_vsi_seid() 572 if (!vsi->tx_rings || !vsi->tx_rings[0]->desc) { in i40e_dbg_dump_desc() 584 ring = kmemdup(vsi->tx_rings[ring_id], sizeof(*ring), GFP_KERNEL); in i40e_dbg_dump_desc()
|
/linux/drivers/net/ethernet/broadcom/genet/ |
A D | bcmgenet.c | 1028 tx_rings[num].packets), \ 1030 tx_rings[num].bytes), \ 2066 ring = &priv->tx_rings[index]; in bcmgenet_xmit() 2734 ring = &priv->tx_rings[i]; in bcmgenet_enable_tx_napi() 2739 ring = &priv->tx_rings[DESC_INDEX]; in bcmgenet_enable_tx_napi() 2750 ring = &priv->tx_rings[i]; in bcmgenet_disable_tx_napi() 2754 ring = &priv->tx_rings[DESC_INDEX]; in bcmgenet_disable_tx_napi() 2764 ring = &priv->tx_rings[i]; in bcmgenet_fini_tx_napi() 2768 ring = &priv->tx_rings[DESC_INDEX]; in bcmgenet_fini_tx_napi() 3163 tx_ring = &priv->tx_rings[index]; in bcmgenet_isr1() [all …]
|
A D | bcmgenet.h | 584 struct bcmgenet_tx_ring tx_rings[DESC_INDEX + 1]; member
|
/linux/drivers/net/ethernet/broadcom/ |
A D | bcmsysport.c | 458 ring = &priv->tx_rings[q]; in bcm_sysport_update_tx_stats() 524 ring = &priv->tx_rings[i]; in bcm_sysport_get_stats() 656 bcm_sysport_set_tx_coalesce(&priv->tx_rings[i], ec); in bcm_sysport_set_coalesce() 1012 bcm_sysport_tx_reclaim(priv, &priv->tx_rings[q]); in bcm_sysport_tx_reclaim_all() 1166 txr = &priv->tx_rings[ring]; in bcm_sysport_rx_isr() 1198 txr = &priv->tx_rings[ring]; in bcm_sysport_tx_isr() 1323 ring = &priv->tx_rings[queue]; in bcm_sysport_xmit() 2358 ring = &priv->tx_rings[q]; in bcm_sysport_map_queues() 2390 ring = &priv->tx_rings[q]; in bcm_sysport_unmap_queues() 2510 priv->tx_rings = devm_kcalloc(&pdev->dev, txq, in bcm_sysport_probe() [all …]
|
/linux/drivers/net/ethernet/netronome/nfp/ |
A D | nfp_net_common.c | 1001 tx_ring = &dp->tx_rings[qidx]; in nfp_net_tx() 2487 dp->tx_rings = kcalloc(dp->num_tx_rings, sizeof(*dp->tx_rings), in nfp_net_tx_rings_prepare() 2489 if (!dp->tx_rings) in nfp_net_tx_rings_prepare() 2501 if (nfp_net_tx_ring_alloc(dp, &dp->tx_rings[r])) in nfp_net_tx_rings_prepare() 2512 nfp_net_tx_ring_bufs_free(dp, &dp->tx_rings[r]); in nfp_net_tx_rings_prepare() 2514 nfp_net_tx_ring_free(&dp->tx_rings[r]); in nfp_net_tx_rings_prepare() 2516 kfree(dp->tx_rings); in nfp_net_tx_rings_prepare() 2525 nfp_net_tx_ring_bufs_free(dp, &dp->tx_rings[r]); in nfp_net_tx_rings_free() 2526 nfp_net_tx_ring_free(&dp->tx_rings[r]); in nfp_net_tx_rings_free() 2529 kfree(dp->tx_rings); in nfp_net_tx_rings_free() [all …]
|
/linux/drivers/net/wireless/realtek/rtw89/ |
A D | pci.c | 111 struct rtw89_pci_tx_ring *tx_ring = &rtwpci->tx_rings[RTW89_TXCH_CH12]; in rtw89_pci_reclaim_tx_fwcmd() 460 tx_ring = &rtwpci->tx_rings[txch]; in rtw89_pci_release_rpp() 795 struct rtw89_pci_tx_ring *tx_ring = &rtwpci->tx_rings[txch]; in __rtw89_pci_check_and_reclaim_tx_resource() 861 struct rtw89_pci_tx_ring *tx_ring = &rtwpci->tx_rings[txch]; in rtw89_pci_ops_tx_kick_off() 871 struct rtw89_pci_tx_ring *tx_ring = &rtwpci->tx_rings[txch]; in __pci_flush_txch() 1071 tx_ring = &rtwpci->tx_rings[txch]; in rtw89_pci_tx_write() 1140 tx_ring = &rtwpci->tx_rings[i]; in rtw89_pci_reset_trx_rings() 1194 rtw89_pci_release_tx_ring(rtwdev, &rtwpci->tx_rings[txch]); in rtw89_pci_ops_reset() 2072 tx_ring = &rtwpci->tx_rings[i]; in rtw89_pci_free_tx_rings() 2275 tx_ring = &rtwpci->tx_rings[i]; in rtw89_pci_alloc_tx_rings() [all …]
|
/linux/drivers/net/ethernet/intel/iavf/ |
A D | iavf_main.c | 364 struct iavf_ring *tx_ring = &adapter->tx_rings[t_idx]; in iavf_map_vector_to_txq() 1148 kfree(adapter->tx_rings); in iavf_free_queues() 1149 adapter->tx_rings = NULL; in iavf_free_queues() 1182 adapter->tx_rings = kcalloc(num_active_queues, in iavf_alloc_queues() 1184 if (!adapter->tx_rings) in iavf_alloc_queues() 1195 tx_ring = &adapter->tx_rings[i]; in iavf_alloc_queues() 2537 if (!adapter->tx_rings) in iavf_free_all_tx_resources() 2541 if (adapter->tx_rings[i].desc) in iavf_free_all_tx_resources() 2542 iavf_free_tx_resources(&adapter->tx_rings[i]); in iavf_free_all_tx_resources() 2560 adapter->tx_rings[i].count = adapter->tx_desc_count; in iavf_setup_all_tx_resources() [all …]
|
A D | iavf.h | 244 struct iavf_ring *tx_rings; member
|
/linux/drivers/net/ethernet/sun/ |
A D | niu.c | 3591 index = (rp - np->tx_rings); in niu_tx_work() 4308 if (np->tx_rings) { in niu_free_channels() 4314 kfree(np->tx_rings); in niu_free_channels() 4315 np->tx_rings = NULL; in niu_free_channels() 4456 struct tx_ring_info *tx_rings; in niu_alloc_channels() local 4512 if (!tx_rings) in niu_alloc_channels() 4517 np->tx_rings = tx_rings; in niu_alloc_channels() 6257 tx_rings = READ_ONCE(np->tx_rings); in niu_get_tx_stats() 6258 if (!tx_rings) in niu_get_tx_stats() 6463 if (np->tx_rings) { in niu_reset_buffers() [all …]
|
/linux/include/linux/ |
A D | thunderbolt.h | 478 struct tb_ring **tx_rings; member
|
/linux/drivers/net/ethernet/broadcom/bnxt/ |
A D | bnxt.c | 6005 *tx_rings = le16_to_cpu(resp->alloc_tx_rings); in __bnxt_hwrm_get_tx_rings() 6024 enables |= tx_rings ? FUNC_CFG_REQ_ENABLES_NUM_TX_RINGS : 0; in __bnxt_hwrm_reserve_pf_rings() 6025 req->num_tx_rings = cpu_to_le16(tx_rings); in __bnxt_hwrm_reserve_pf_rings() 6031 enables |= tx_rings + ring_grps ? in __bnxt_hwrm_reserve_pf_rings() 6046 req->num_cmpl_rings = cpu_to_le16(tx_rings + ring_grps); in __bnxt_hwrm_reserve_pf_rings() 6081 enables |= tx_rings + ring_grps ? in __bnxt_hwrm_reserve_vf_rings() 6093 req->num_tx_rings = cpu_to_le16(tx_rings); in __bnxt_hwrm_reserve_vf_rings() 6096 req->num_cmpl_rings = cpu_to_le16(tx_rings + ring_grps); in __bnxt_hwrm_reserve_vf_rings() 6132 bp->hw_resc.resv_tx_rings = tx_rings; in bnxt_hwrm_reserve_pf_rings() 6145 bp->hw_resc.resv_tx_rings = tx_rings; in bnxt_hwrm_reserve_vf_rings() [all …]
|