Home
last modified time | relevance | path

Searched refs:txqs (Results 1 – 25 of 54) sorted by relevance

123

/linux/drivers/net/wireless/intel/iwlwifi/
A Diwl-trans.c46 trans->txqs.tfd.addr_size = 64; in iwl_trans_alloc()
47 trans->txqs.tfd.max_tbs = IWL_TFH_NUM_TBS; in iwl_trans_alloc()
50 trans->txqs.tfd.addr_size = 36; in iwl_trans_alloc()
51 trans->txqs.tfd.max_tbs = IWL_NUM_OF_TBS; in iwl_trans_alloc()
92 trans->txqs.bc_tbl_size, in iwl_trans_init()
94 if (!trans->txqs.bc_pool) in iwl_trans_init()
99 WARN_ON(trans->txqs.tfd.addr_size != in iwl_trans_init()
112 if (!trans->txqs.tso_hdr_page) { in iwl_trans_init()
127 if (trans->txqs.tso_hdr_page) { in iwl_trans_free()
130 per_cpu_ptr(trans->txqs.tso_hdr_page, i); in iwl_trans_free()
[all …]
A Diwl-trans.h170 #define IWL_TRANS_MAX_FRAGS(trans) ((trans)->txqs.tfd.max_tbs - 3)
579 void (*freeze_txq_timer)(struct iwl_trans *trans, unsigned long txqs,
1010 struct iwl_trans_txqs txqs; member
1266 unsigned long txqs, in iwl_trans_freeze_txq_timer() argument
1275 trans->ops->freeze_txq_timer(trans, txqs, freeze); in iwl_trans_freeze_txq_timer()
1291 u32 txqs) in iwl_trans_wait_tx_queues_empty() argument
1302 return trans->ops->wait_tx_queues_empty(trans, txqs); in iwl_trans_wait_tx_queues_empty()
/linux/drivers/net/wireless/intel/iwlwifi/pcie/
A Dtx.c167 trans->txqs.tfd.max_tbs); in iwl_pcie_txq_build_tfd()
274 trans->txqs.tfd.size * in iwl_pcie_txq_free()
306 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); in iwl_pcie_tx_start()
329 trans->txqs.cmd.fifo, in iwl_pcie_tx_start()
330 trans->txqs.cmd.wdg_timeout); in iwl_pcie_tx_start()
443 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); in iwl_pcie_tx_stop()
467 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); in iwl_pcie_tx_free()
904 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_enqueue_hcmd()
1185 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_hcmd_complete()
1313 trans->txqs.tfd.size, in iwl_fill_data_tbs_amsdu()
[all …]
A Dtrans-gen2.c239 if (iwl_txq_gen2_init(trans, trans->txqs.cmd.q_id, queue_size)) in iwl_pcie_gen2_nic_init()
325 memset(trans->txqs.queue_stopped, 0, in iwl_trans_pcie_gen2_fw_alive()
326 sizeof(trans->txqs.queue_stopped)); in iwl_trans_pcie_gen2_fw_alive()
327 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); in iwl_trans_pcie_gen2_fw_alive()
A Dtx-gen2.c31 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_gen2_enqueue_hcmd()
143 cpu_to_le16(QUEUE_TO_SEQ(trans->txqs.cmd.q_id) | in iwl_pcie_gen2_enqueue_hcmd()
191 cmd_size, txq->write_ptr, idx, trans->txqs.cmd.q_id); in iwl_pcie_gen2_enqueue_hcmd()
A Dtrans.c1914 trans->txqs.cmd.q_id = trans_cfg->cmd_queue; in iwl_trans_pcie_configure()
1915 trans->txqs.cmd.fifo = trans_cfg->cmd_fifo; in iwl_trans_pcie_configure()
2256 struct iwl_txq *txq = trans->txqs.txq[i]; in iwl_trans_pcie_block_txq_ptrs()
2258 if (i == trans->txqs.cmd.q_id) in iwl_trans_pcie_block_txq_ptrs()
2310 txq = trans->txqs.txq[txq_idx]; in iwl_trans_pcie_wait_txq_empty()
2366 if (cnt == trans->txqs.cmd.q_id) in iwl_trans_pcie_wait_txqs_empty()
2368 if (!test_bit(cnt, trans->txqs.queue_used)) in iwl_trans_pcie_wait_txqs_empty()
2556 if (state->pos == trans->txqs.cmd.q_id) in iwl_dbgfs_tx_queue_seq_show()
3247 struct iwl_txq *cmdq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_trans_pcie_dump_data()
3316 u16 tfd_size = trans->txqs.tfd.size; in iwl_trans_pcie_dump_data()
[all …]
A Dctxt-info.c221 cpu_to_le64(trans->txqs.txq[trans->txqs.cmd.q_id]->dma_addr); in iwl_pcie_ctxt_info_init()
A Dctxt-info-gen3.c183 cpu_to_le64(trans->txqs.txq[trans->txqs.cmd.q_id]->dma_addr); in iwl_pcie_ctxt_info_gen3_init()
/linux/drivers/net/wireless/intel/iwlwifi/queue/
A Dtx.c47 WARN_ON(trans->txqs.bc_table_dword); in iwl_pcie_gen2_update_byte_tbl()
167 trans->txqs.tfd.max_tbs); in iwl_txq_gen2_set_tb()
872 txq = trans->txqs.txq[txq_id]; in iwl_txq_gen2_free()
889 trans->txqs.txq[txq_id] = NULL; in iwl_txq_gen2_free()
1170 trans->txqs.txq[qid] = txq; in iwl_txq_alloc_response()
1251 memset(trans->txqs.queue_used, 0, sizeof(trans->txqs.queue_used)); in iwl_txq_gen2_tx_free()
1255 if (!trans->txqs.txq[i]) in iwl_txq_gen2_tx_free()
1268 if (!trans->txqs.txq[txq_id]) { in iwl_txq_gen2_init()
1415 if (trans->txqs.bc_table_dword) in iwl_txq_gen1_update_byte_cnt_tbl()
1620 trans->txqs.dev_cmd_offs); in iwl_txq_reclaim()
[all …]
A Dtx.h32 if (test_and_clear_bit(txq->id, trans->txqs.queue_stopped)) { in iwl_wake_queue()
44 return txq->tfds + trans->txqs.tfd.size * idx; in iwl_txq_get_tfd()
63 if (!test_and_set_bit(txq->id, trans->txqs.queue_stopped)) { in iwl_txq_stop()
179 void iwl_trans_txq_freeze_timer(struct iwl_trans *trans, unsigned long txqs,
/linux/drivers/net/wireless/ath/ath5k/
A Dmac80211-ops.c69 ath5k_tx_queue(hw, skb, &ah->txqs[qnum], control); in ath5k_tx()
745 *tx = ah->txqs[AR5K_TX_QUEUE_ID_DATA_MIN].txq_max; in ath5k_get_ringparam()
765 for (qnum = 0; qnum < ARRAY_SIZE(ah->txqs); qnum++) { in ath5k_set_ringparam()
766 if (!ah->txqs[qnum].setup) in ath5k_set_ringparam()
768 if (ah->txqs[qnum].qnum < AR5K_TX_QUEUE_ID_DATA_MIN || in ath5k_set_ringparam()
769 ah->txqs[qnum].qnum > AR5K_TX_QUEUE_ID_DATA_MAX) in ath5k_set_ringparam()
772 ah->txqs[qnum].txq_max = tx; in ath5k_set_ringparam()
773 if (ah->txqs[qnum].txq_len >= ah->txqs[qnum].txq_max) in ath5k_set_ringparam()
774 ieee80211_stop_queue(hw, ah->txqs[qnum].qnum); in ath5k_set_ringparam()
A Dbase.c1009 txq = &ah->txqs[qnum]; in ath5k_txq_setup()
1021 return &ah->txqs[qnum]; in ath5k_txq_setup()
1115 for (i = 0; i < ARRAY_SIZE(ah->txqs); i++) { in ath5k_drain_tx_buffs()
1116 if (ah->txqs[i].setup) { in ath5k_drain_tx_buffs()
1117 txq = &ah->txqs[i]; in ath5k_drain_tx_buffs()
1140 struct ath5k_txq *txq = ah->txqs; in ath5k_txq_release()
1143 for (i = 0; i < ARRAY_SIZE(ah->txqs); i++, txq++) in ath5k_txq_release()
1793 ath5k_tx_processq(ah, &ah->txqs[i]); in ath5k_tasklet_tx()
2473 for (i = 0; i < ARRAY_SIZE(ah->txqs); i++) { in ath5k_tx_complete_poll_work()
2474 if (ah->txqs[i].setup) { in ath5k_tx_complete_poll_work()
[all …]
/linux/drivers/net/ethernet/huawei/hinic/
A Dhinic_main.c149 if (nic_dev->txqs) in create_txqs()
154 if (!nic_dev->txqs) in create_txqs()
180 hinic_clean_txq(&nic_dev->txqs[i]); in create_txqs()
184 hinic_clean_txq(&nic_dev->txqs[j]); in create_txqs()
199 napi_enable(&nic_dev->txqs[i].napi); in enable_txqs_napi()
208 napi_disable(&nic_dev->txqs[i].napi); in disable_txqs_napi()
220 if (!nic_dev->txqs) in free_txqs()
225 hinic_clean_txq(&nic_dev->txqs[i]); in free_txqs()
231 nic_dev->txqs = NULL; in free_txqs()
852 nic_dev->txqs[q_id].napi.state); in hinic_tx_timeout()
[all …]
A Dhinic_dev.h93 struct hinic_txq *txqs; member
/linux/drivers/infiniband/hw/hfi1/
A Dipoib_tx.c695 priv->txqs = kcalloc_node(dev->num_tx_queues, in hfi1_ipoib_txreq_init()
699 if (!priv->txqs) in hfi1_ipoib_txreq_init()
703 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_init()
746 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_init()
752 kfree(priv->txqs); in hfi1_ipoib_txreq_init()
753 priv->txqs = NULL; in hfi1_ipoib_txreq_init()
786 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_txreq_deinit()
796 kfree(priv->txqs); in hfi1_ipoib_txreq_deinit()
797 priv->txqs = NULL; in hfi1_ipoib_txreq_deinit()
806 struct hfi1_ipoib_txq *txq = &priv->txqs[i]; in hfi1_ipoib_napi_tx_enable()
[all …]
A Dipoib.h125 struct hfi1_ipoib_txq *txqs; member
/linux/net/
A Ddevres.c22 unsigned int txqs, unsigned int rxqs) in devm_alloc_etherdev_mqs() argument
30 dr->ndev = alloc_etherdev_mqs(sizeof_priv, txqs, rxqs); in devm_alloc_etherdev_mqs()
/linux/drivers/net/ethernet/netronome/nfp/abm/
A Dmain.c85 unsigned int txqs; in nfp_abm_spawn_repr() local
90 txqs = 1; in nfp_abm_spawn_repr()
93 txqs = alink->vnic->max_rx_rings; in nfp_abm_spawn_repr()
96 netdev = nfp_repr_alloc_mqs(app, txqs, 1); in nfp_abm_spawn_repr()
/linux/drivers/net/ethernet/intel/ice/
A Dice_base.c773 u8 buf_len = struct_size(qg_buf, txqs, 1); in ice_vsi_cfg_txq()
789 qg_buf->txqs[0].txq_id = cpu_to_le16(pf_q); in ice_vsi_cfg_txq()
790 ice_set_ctx(hw, (u8 *)&tlan_ctx, qg_buf->txqs[0].txq_ctx, in ice_vsi_cfg_txq()
833 txq = &qg_buf->txqs[0]; in ice_vsi_cfg_txq()
/linux/drivers/net/ethernet/netronome/nfp/
A Dnfp_net_repr.h102 nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs);
/linux/include/linux/
A Detherdevice.h56 struct net_device *alloc_etherdev_mqs(int sizeof_priv, unsigned int txqs,
62 unsigned int txqs,
/linux/drivers/net/can/dev/
A Ddev.c241 unsigned int txqs, unsigned int rxqs) in alloc_candev_mqs() argument
268 txqs, rxqs); in alloc_candev_mqs()
/linux/net/ethernet/
A Deth.c391 struct net_device *alloc_etherdev_mqs(int sizeof_priv, unsigned int txqs, in alloc_etherdev_mqs() argument
395 ether_setup, txqs, rxqs); in alloc_etherdev_mqs()
/linux/drivers/net/ethernet/marvell/
A Dmvneta.c472 struct mvneta_tx_queue *txqs; member
1223 struct mvneta_tx_queue *txq = &pp->txqs[queue]; in mvneta_port_up()
1827 return &pp->txqs[queue]; in mvneta_tx_done_policy()
2107 txq = &pp->txqs[cpu % txq_number]; in mvneta_xdp_xmit_back()
2148 txq = &pp->txqs[cpu % txq_number]; in mvneta_xdp_xmit()
2781 struct mvneta_tx_queue *txq = &pp->txqs[txq_id]; in mvneta_tx()
3290 mvneta_txq_done_force(pp, &pp->txqs[queue]); in mvneta_tx_reset()
3515 mvneta_txq_deinit(pp, &pp->txqs[queue]); in mvneta_cleanup_txqs()
4505 struct mvneta_tx_queue *txq = &pp->txqs[queue]; in mvneta_ethtool_set_coalesce()
5027 pp->txqs = devm_kcalloc(dev, txq_number, sizeof(*pp->txqs), GFP_KERNEL); in mvneta_init()
[all …]
/linux/include/linux/can/
A Ddev.h152 unsigned int txqs, unsigned int rxqs);

Completed in 59 milliseconds

123