/linux/drivers/infiniband/hw/hfi1/ |
A D | ipoib_tx.c | 57 netif_stop_subqueue(txq->priv->netdev, txq->q_idx); in hfi1_ipoib_stop_txq() 64 netif_wake_subqueue(txq->priv->netdev, txq->q_idx); in hfi1_ipoib_wake_txq() 82 if (hfi1_ipoib_used(txq) >= hfi1_ipoib_ring_hwat(txq) && in hfi1_ipoib_check_queue_depth() 352 struct hfi1_ipoib_txq *txq = txp->txq; in hfi1_ipoib_send_dma_common() local 377 tx->txq = txq; in hfi1_ipoib_send_dma_common() 449 iowait_starve_clear(txq->pkts_sent, &txq->wait); in hfi1_ipoib_submit_tx() 459 struct hfi1_ipoib_txq *txq = txp->txq; in hfi1_ipoib_send_dma_single() local 510 struct hfi1_ipoib_txq *txq = txp->txq; in hfi1_ipoib_send_dma_list() local 735 txq->tx_ring.avail = hfi1_ipoib_ring_hwat(txq); in hfi1_ipoib_txreq_init() 838 txq->sde ? txq->sde->this_idx : 0); in hfi1_ipoib_tx_timeout() [all …]
|
A D | trace_tx.h | 914 __entry->txq = txq; 949 TP_ARGS(txq) 955 TP_ARGS(txq) 961 TP_ARGS(txq) 967 TP_ARGS(txq) 973 TP_ARGS(txq) 979 TP_ARGS(txq) 985 TP_ARGS(txq) 991 TP_ARGS(txq) 1015 __entry->txq = tx->txq; [all …]
|
/linux/drivers/net/wireless/intel/iwlwifi/queue/ |
A D | tx.c | 782 if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) in iwl_txq_gen2_tx() 842 txq->tfds, txq->dma_addr); in iwl_txq_gen2_free_memory() 872 txq = trans->txqs.txq[txq_id]; in iwl_txq_gen2_free() 981 txq->read_ptr, txq->write_ptr); in iwl_txq_log_scd_error() 994 txq->read_ptr, txq->write_ptr, in iwl_txq_log_scd_error() 1009 if (txq->read_ptr == txq->write_ptr) { in iwl_txq_stuck_timer() 1170 trans->txqs.txq[qid] = txq; in iwl_txq_alloc_response() 1514 if (txq->read_ptr == txq->write_ptr) in iwl_txq_progress() 1556 txq->write_ptr, txq->read_ptr); in iwl_txq_reclaim() 1649 txq->read_ptr = txq->write_ptr; in iwl_txq_set_q_ptrs() [all …]
|
A D | tx.h | 16 iwl_txq_get_first_tb_dma(struct iwl_txq *txq, int idx) in iwl_txq_get_first_tb_dma() argument 18 return txq->first_tb_dma + in iwl_txq_get_first_tb_dma() 30 struct iwl_txq *txq) in iwl_wake_queue() argument 39 struct iwl_txq *txq, int idx) in iwl_txq_get_tfd() argument 42 idx = iwl_txq_get_cmd_index(txq, idx); in iwl_txq_get_tfd() 44 return txq->tfds + trans->txqs.tfd.size * idx; in iwl_txq_get_tfd() 68 txq->id); in iwl_txq_stop() 170 struct iwl_txq *txq, int index); 172 struct iwl_txq *txq); 174 struct iwl_txq *txq, u16 byte_cnt, [all …]
|
/linux/drivers/net/wireless/intel/iwlwifi/pcie/ |
A D | tx.c | 121 struct iwl_txq *txq = trans->txqs.txq[i]; in iwl_pcie_txq_check_wrptrs() local 213 while (txq->write_ptr != txq->read_ptr) { in iwl_pcie_txq_unmap() 228 if (txq->read_ptr == txq->write_ptr && in iwl_pcie_txq_unmap() 276 txq->tfds, txq->dma_addr); in iwl_pcie_txq_free() 291 memset(txq, 0, sizeof(*txq)); in iwl_pcie_txq_free() 684 txq->write_ptr, txq->read_ptr); in iwl_pcie_cmdq_reclaim() 699 if (txq->read_ptr == txq->write_ptr) in iwl_pcie_cmdq_reclaim() 1147 if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) in iwl_pcie_enqueue_hcmd() 1452 txq = trans->txqs.txq[txq_id]; in iwl_trans_pcie_tx() 1502 txq->entries[txq->write_ptr].skb = skb; in iwl_trans_pcie_tx() [all …]
|
A D | tx-gen2.c | 31 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_gen2_enqueue_hcmd() local 112 spin_lock_irqsave(&txq->lock, flags); in iwl_pcie_gen2_enqueue_hcmd() 114 idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_pcie_gen2_enqueue_hcmd() 115 tfd = iwl_txq_get_tfd(trans, txq, txq->write_ptr); in iwl_pcie_gen2_enqueue_hcmd() 127 out_cmd = txq->entries[idx].cmd; in iwl_pcie_gen2_enqueue_hcmd() 128 out_meta = &txq->entries[idx].meta; in iwl_pcie_gen2_enqueue_hcmd() 144 INDEX_TO_SEQ(txq->write_ptr)); in iwl_pcie_gen2_enqueue_hcmd() 244 if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) in iwl_pcie_gen2_enqueue_hcmd() 245 mod_timer(&txq->stuck_timer, jiffies + txq->wd_timeout); in iwl_pcie_gen2_enqueue_hcmd() 249 txq->write_ptr = iwl_txq_inc_wrap(trans, txq->write_ptr); in iwl_pcie_gen2_enqueue_hcmd() [all …]
|
/linux/drivers/net/ethernet/huawei/hinic/ |
A D | hinic_tx.c | 534 hinic_sq_prepare_wqe(txq->sq, prod_idx, sq_wqe, txq->sges, nr_sges); in hinic_lb_xmit_frame() 618 hinic_sq_prepare_wqe(txq->sq, prod_idx, sq_wqe, txq->sges, nr_sges); in hinic_xmit_frame() 832 err = request_irq(sq->irq, tx_irq, 0, txq->irq_name, txq); in tx_request_irq() 867 txq->netdev = netdev; in hinic_init_txq() 868 txq->sq = sq; in hinic_init_txq() 870 txq_stats_init(txq); in hinic_init_txq() 874 sges_size = txq->max_sges * sizeof(*txq->sges); in hinic_init_txq() 876 if (!txq->sges) in hinic_init_txq() 879 sges_size = txq->max_sges * sizeof(*txq->free_sges); in hinic_init_txq() 888 if (!txq->irq_name) { in hinic_init_txq() [all …]
|
/linux/drivers/net/wireless/ath/ath9k/ |
A D | xmit.c | 129 struct ath_txq *txq = tid->txq; in ath9k_wake_tx_queue() local 193 .txq = tid->txq, in ath_tid_pull() 234 struct ath_txq *txq = tid->txq; in ath_tx_flush_tid() local 1528 txq = txtid->txq; in ath_tx_aggr_start() 1560 struct ath_txq *txq = txtid->txq; in ath_tx_aggr_stop() local 1599 txq = tid->txq; in ath_tx_aggr_wakeup() 1904 txq = &sc->tx.txq[i]; in ath_drain_all_txq() 2003 txq->axq_qnum, txq->axq_depth); in ath_tx_txqaddbuf() 2015 txq->axq_qnum, txq->axq_link, in ath_tx_txqaddbuf() 2689 txq = &sc->tx.txq[ts.qid]; in ath_tx_edma_tasklet() [all …]
|
/linux/drivers/net/ethernet/marvell/ |
A D | mv643xx_eth.c | 447 return container_of(txq, struct mv643xx_eth_private, txq[txq->index]); in txq_to_mp() 735 if (txq->tx_curr_desc == txq->tx_ring_size) in txq_put_data_tso() 813 if (txq->tx_curr_desc == txq->tx_ring_size) in txq_put_hdr_tso() 954 if (txq->tx_curr_desc == txq->tx_ring_size) in txq_submit_skb() 1000 txq = mp->txq + queue; in mv643xx_eth_xmit() 1261 struct tx_queue *txq = mp->txq + i; in mv643xx_eth_get_stats() local 2031 struct tx_queue *txq = mp->txq + index; in txq_init() local 2122 txq_reclaim(txq, txq->tx_ring_size, 1); in txq_deinit() 2137 txq->tso_hdrs, txq->tso_hdrs_dma); in txq_deinit() 2207 struct tx_queue *txq = mp->txq + i; in handle_link_event() local [all …]
|
A D | mvneta.c | 740 if (txq->txq_get_index == txq->size) in mvneta_txq_inc_get() 748 if (txq->txq_put_index == txq->size) in mvneta_txq_inc_put() 1467 for (txq = 0; txq < txq_number; txq++) in mvneta_defaults_set() 1890 if (txq->count <= txq->tx_wake_threshold) in mvneta_txq_done() 2055 if (txq->count >= txq->tx_stop_threshold) in mvneta_xdp_submit_frame() 2060 buf = &txq->buf[txq->txq_put_index]; in mvneta_xdp_submit_frame() 2848 if (txq->count >= txq->tx_stop_threshold) in mvneta_tx() 3410 txq->last_desc = txq->size - 1; in mvneta_txq_sw_init() 3412 txq->buf = kmalloc_array(txq->size, sizeof(*txq->buf), GFP_KERNEL); in mvneta_txq_sw_init() 3476 txq->tso_hdrs, txq->tso_hdrs_phys); in mvneta_txq_sw_deinit() [all …]
|
/linux/drivers/net/ethernet/qlogic/qede/ |
A D | qede_fp.c | 89 idx, txq->sw_tx_cons, txq->sw_tx_prod); in qede_free_tx_pkt() 296 writel(txq->tx_db.raw, txq->doorbell_addr); in qede_update_tx_producer() 314 txq->stopped_cnt++; in qede_xdp_xmit() 330 xdp = txq->sw_tx_ring.xdp + txq->sw_tx_prod; in qede_xdp_xmit() 335 txq->sw_tx_prod = (txq->sw_tx_prod + 1) % txq->num_tx_buffers; in qede_xdp_xmit() 428 txq->sw_tx_cons = (txq->sw_tx_cons + 1) % txq->num_tx_buffers; in qede_xdp_tx_int() 429 txq->xmit_pkts++; in qede_xdp_tx_int() 458 txq->sw_tx_cons = (txq->sw_tx_cons + 1) % txq->num_tx_buffers; in qede_tx_int() 459 txq->xmit_pkts++; in qede_tx_int() 1706 txq->sw_tx_prod = (txq->sw_tx_prod + 1) % txq->num_tx_buffers; in qede_start_xmit() [all …]
|
A D | qede_main.c | 516 txq->index, le16_to_cpu(*txq->hw_cons_ptr), in qede_tx_log_print() 535 txq = &edev->fp_array[txqueue].txq[cos]; in qede_tx_timeout() 950 if (!fp->txq) in qede_alloc_fp_array() 1547 if (txq->is_xdp) in qede_free_mem_txq() 1572 size = sizeof(*txq->sw_tx_ring.xdp) * txq->num_tx_buffers; in qede_alloc_mem_txq() 1577 size = sizeof(*txq->sw_tx_ring.skbs) * txq->num_tx_buffers; in qede_alloc_mem_txq() 1695 txq->index, qed_chain_get_cons_idx(&txq->tx_pbl), in qede_empty_tx_queue() 1782 struct qede_tx_queue *txq = &fp->txq[cos]; in qede_init_fp() local 1959 while (txq->sw_tx_cons != txq->sw_tx_prod) { in qede_drain_txq() 1972 txq->index, txq->sw_tx_prod, in qede_drain_txq() [all …]
|
/linux/drivers/net/ethernet/atheros/alx/ |
A D | main.c | 160 if (txq->write_idx >= txq->read_idx) in alx_tpd_avail() 161 return txq->count + txq->read_idx - txq->write_idx - 1; in alx_tpd_avail() 201 alx_tpd_avail(txq) > txq->count / 4) in alx_clean_tx_irq() 766 np->txq = txq; in alx_alloc_napis() 1468 if (++txq->write_idx == txq->count) in alx_map_tx_skb() 1471 tpd = &txq->tpd[txq->write_idx]; in alx_map_tx_skb() 1492 if (++txq->write_idx == txq->count) in alx_map_tx_skb() 1494 tpd = &txq->tpd[txq->write_idx]; in alx_map_tx_skb() 1512 txq->bufs[txq->write_idx].skb = skb; in alx_map_tx_skb() 1514 if (++txq->write_idx == txq->count) in alx_map_tx_skb() [all …]
|
/linux/include/trace/events/ |
A D | qdisc.h | 16 TP_PROTO(struct Qdisc *qdisc, const struct netdev_queue *txq, 19 TP_ARGS(qdisc, txq, packets, skb), 23 __field(const struct netdev_queue *, txq ) 35 __entry->txq = txq; 38 __entry->ifindex = txq->dev ? txq->dev->ifindex : 0; 41 __entry->txq_state = txq->state; 51 TP_PROTO(struct Qdisc *qdisc, const struct netdev_queue *txq, struct sk_buff *skb), 53 TP_ARGS(qdisc, txq, skb), 57 __field(const struct netdev_queue *, txq) 66 __entry->txq = txq; [all …]
|
/linux/drivers/net/wireless/mediatek/mt76/ |
A D | tx.c | 11 if (!txq->sta) in mt76_txq_get_qid() 14 return txq->ac; in mt76_txq_get_qid() 21 struct ieee80211_txq *txq; in mt76_tx_check_agg_ssn() local 30 txq = sta->txq[tid]; in mt76_tx_check_agg_ssn() 399 struct ieee80211_txq *txq = sta->txq[i]; in mt76_release_buffered_frames() local 463 ieee80211_get_tx_rates(txq->vif, txq->sta, skb, in mt76_txq_send_burst() 483 ieee80211_get_tx_rates(txq->vif, txq->sta, skb, in mt76_txq_send_burst() 520 if (!txq) in mt76_txq_schedule_list() 535 u8 tid = txq->tid; in mt76_txq_schedule_list() 615 struct ieee80211_txq *txq = sta->txq[i]; in mt76_stop_tx_queues() local [all …]
|
/linux/drivers/net/ethernet/freescale/ |
A D | fec_main.c | 310 (addr < txq->tso_hdrs_dma + txq->bd.ring_size * TSO_HEADER_SIZE)) 339 (const char *)txq->bd.cur) >> txq->bd.dsize_log2) - 1; in fec_enet_get_free_txdesc_num() 374 bdp = txq->bd.base; in fec_dump() 492 bdp = txq->bd.cur; in fec_enet_txq_submit_frag_skb() 530 bdp = txq->bd.cur; in fec_enet_txq_submit_skb() 620 txq->bd.cur = bdp; in fec_enet_txq_submit_skb() 809 txq->bd.cur = bdp; in fec_enet_txq_submit_tso() 3017 txq = kzalloc(sizeof(*txq), GFP_KERNEL); in fec_enet_alloc_queue() 3018 if (!txq) { in fec_enet_alloc_queue() 3029 (txq->bd.ring_size - txq->tx_stop_threshold) / 2; in fec_enet_alloc_queue() [all …]
|
/linux/drivers/vdpa/vdpa_sim/ |
A D | vdpa_sim_net.c | 36 struct vdpasim_virtqueue *txq = &vdpasim->vqs[1]; in vdpasim_net_work() local 48 if (!txq->ready || !rxq->ready) in vdpasim_net_work() 53 err = vringh_getdesc_iotlb(&txq->vring, &txq->out_iov, NULL, in vdpasim_net_work() 54 &txq->head, GFP_ATOMIC); in vdpasim_net_work() 61 vringh_complete_iotlb(&txq->vring, txq->head, 0); in vdpasim_net_work() 66 read = vringh_iov_pull_iotlb(&txq->vring, &txq->out_iov, in vdpasim_net_work() 83 vringh_complete_iotlb(&txq->vring, txq->head, 0); in vdpasim_net_work() 90 if (vringh_need_notify_iotlb(&txq->vring) > 0) in vdpasim_net_work() 91 vringh_notify(&txq->vring); in vdpasim_net_work()
|
/linux/drivers/net/ethernet/chelsio/cxgb4vf/ |
A D | sge.c | 1135 netif_tx_stop_queue(txq->txq); in txq_stop() 1266 wr = (void *)&txq->q.desc[txq->q.pidx]; in t4vf_eth_xmit() 1322 txq->tso++; in t4vf_eth_xmit() 2136 if (reclaimable(&txq->q) && __netif_tx_trylock(txq->txq)) { in sge_tx_timer_cb() 2144 __netif_tx_unlock(txq->txq); in sge_tx_timer_cb() 2433 txq->q.desc = alloc_ring(adapter->pdev_dev, txq->q.size, in t4vf_sge_alloc_eth_txq() 2436 &txq->q.phys_addr, &txq->q.sdesc, s->stat_len); in t4vf_sge_alloc_eth_txq() 2485 txq->q.desc, txq->q.phys_addr); in t4vf_sge_alloc_eth_txq() 2493 txq->q.stat = (void *)&txq->q.desc[txq->q.size]; in t4vf_sge_alloc_eth_txq() 2501 txq->txq = devq; in t4vf_sge_alloc_eth_txq() [all …]
|
/linux/drivers/net/ethernet/hisilicon/ |
A D | hisi_femac.c | 155 struct hisi_femac_queue *txq = &priv->txq; in hisi_femac_xmit_reclaim() local 163 skb = txq->skb[txq->tail]; in hisi_femac_xmit_reclaim() 177 txq->skb[txq->tail] = NULL; in hisi_femac_xmit_reclaim() 178 txq->tail = (txq->tail + 1) % txq->num; in hisi_femac_xmit_reclaim() 387 struct hisi_femac_queue *txq = &priv->txq; in hisi_femac_free_skb_rings() local 412 pos = txq->tail; in hisi_femac_free_skb_rings() 503 struct hisi_femac_queue *txq = &priv->txq; in hisi_femac_net_xmit() local 517 if (unlikely(!CIRC_SPACE(txq->head, txq->tail, in hisi_femac_net_xmit() 533 txq->dma_phys[txq->head] = addr; in hisi_femac_net_xmit() 535 txq->skb[txq->head] = skb; in hisi_femac_net_xmit() [all …]
|
/linux/drivers/net/wireless/ath/ath5k/ |
A D | base.c | 831 txq->txq_len++; in ath5k_txbuf_setup() 1010 if (!txq->setup) { in ath5k_txq_setup() 1011 txq->qnum = qnum; in ath5k_txq_setup() 1012 txq->link = NULL; in ath5k_txq_setup() 1015 txq->setup = true; in ath5k_txq_setup() 1016 txq->txq_len = 0; in ath5k_txq_setup() 1127 txq->txq_len--; in ath5k_drain_tx_buffs() 1144 if (txq->setup) { in ath5k_txq_release() 1630 if (txq->txq_len >= txq->txq_max && in ath5k_tx_queue() 1776 txq->txq_len--; in ath5k_tx_processq() [all …]
|
/linux/drivers/net/ethernet/microsoft/mana/ |
A D | mana_en.c | 150 txq = &apc->tx_qp[txq_idx].txq; in mana_start_xmit() 343 int txq; in mana_get_tx_queue() local 351 return txq; in mana_get_tx_queue() 357 int txq; in mana_select_queue() local 368 txq = mana_get_tx_queue(ndev, skb, txq); in mana_select_queue() 371 return txq; in mana_select_queue() 787 struct mana_txq *txq = cq->txq; in mana_poll_tx_cq() local 1195 txq = &apc->tx_qp[i].txq; in mana_create_txq() 1215 cq->txq = txq; in mana_create_txq() 1252 txq->gdma_txq_id = txq->gdma_sq->id; in mana_create_txq() [all …]
|
/linux/drivers/net/ethernet/chelsio/cxgb3/ |
A D | sge.c | 694 if (q->txq[i].desc) { in t3_free_qset() 704 q->txq[i].size * in t3_free_qset() 706 q->txq[i].desc, q->txq[i].phys_addr); in t3_free_qset() 1284 q = &qs->txq[TXQ_ETH]; in t3_eth_xmit() 2259 struct sge_txq *txq = &qs->txq[TXQ_ETH]; in check_ring_db() local 2261 if (txq->cleaned + txq->in_use != txq->processed && in check_ring_db() 2270 struct sge_txq *txq = &qs->txq[TXQ_OFLD]; in check_ring_db() local 2272 if (txq->cleaned + txq->in_use != txq->processed && in check_ring_db() 3080 if (!q->txq[i].desc) in t3_sge_alloc_qset() 3083 q->txq[i].gen = 1; in t3_sge_alloc_qset() [all …]
|
/linux/net/sched/ |
A D | sch_generic.c | 86 txq = skb_get_tx_queue(txq->dev, skb); in __skb_dequeue_bad_txq() 98 qdisc_maybe_clear_missed(q, txq); in __skb_dequeue_bad_txq() 259 txq = skb_get_tx_queue(txq->dev, skb); in dequeue_skb() 282 qdisc_maybe_clear_missed(q, txq); in dequeue_skb() 346 HARD_TX_UNLOCK(dev, txq); in sch_direct_xmit() 391 struct netdev_queue *txq; in qdisc_restart() local 405 txq = skb_get_tx_queue(dev, skb); in qdisc_restart() 462 struct netdev_queue *txq; in dev_watchdog() local 466 if (netif_xmit_stopped(txq) && in dev_watchdog() 470 txq->trans_timeout++; in dev_watchdog() [all …]
|
/linux/drivers/net/wireless/intel/iwlegacy/ |
A D | common.c | 2749 struct il_tx_queue *txq = &il->txq[txq_id]; in il_tx_queue_unmap() local 2773 struct il_tx_queue *txq = &il->txq[txq_id]; in il_tx_queue_free() local 2788 txq->tfds, txq->q.dma_addr); in il_tx_queue_free() 2801 memset(txq, 0, sizeof(*txq)); in il_tx_queue_free() 2868 txq->tfds, txq->q.dma_addr); in il_cmd_queue_free() 2877 memset(txq, 0, sizeof(*txq)); in il_cmd_queue_free() 3005 struct il_tx_queue *txq = &il->txq[txq_id]; in il_tx_queue_init() local 3028 if (!txq->meta || !txq->cmd) in il_tx_queue_init() 3081 struct il_tx_queue *txq = &il->txq[txq_id]; in il_tx_queue_reset() local 3237 struct il_tx_queue *txq = &il->txq[txq_id]; in il_hcmd_queue_reclaim() local [all …]
|
/linux/drivers/atm/ |
A D | ambassador.c | 613 amb_txq * txq = &dev->txq; in tx_give() local 623 if (txq->pending < txq->maximum) { in tx_give() 627 txq->pending++; in tx_give() 628 txq->in.ptr = NEXTQ (txq->in.ptr, txq->in.start, txq->in.limit); in tx_give() 633 if (txq->pending > txq->high) in tx_give() 634 txq->high = txq->pending; in tx_give() 638 txq->filled++; in tx_give() 645 amb_txq * txq = &dev->txq; in tx_take() local 652 if (txq->pending && txq->out.ptr->handle) { in tx_take() 659 txq->out.ptr = NEXTQ (txq->out.ptr, txq->out.start, txq->out.limit); in tx_take() [all …]
|