/linux/drivers/net/ethernet/moxa/ |
A D | moxart_ether.c | 139 priv->tx_head = 0; in moxart_mac_setup_desc_ring() 293 unsigned int tx_head = priv->tx_head; in moxart_tx_finished() local 296 while (tx_tail != tx_head) { in moxart_tx_finished() 340 unsigned int tx_head; in moxart_mac_start_xmit() local 346 tx_head = priv->tx_head; in moxart_mac_start_xmit() 347 desc = priv->tx_desc_base + (TX_REG_DESC_SIZE * tx_head); in moxart_mac_start_xmit() 368 priv->tx_len[tx_head] = len; in moxart_mac_start_xmit() 369 priv->tx_skb[tx_head] = skb; in moxart_mac_start_xmit() 371 moxart_desc_write(priv->tx_mapping[tx_head], in moxart_mac_start_xmit() 386 if (tx_head == TX_DESC_NUM_MASK) in moxart_mac_start_xmit() [all …]
|
/linux/drivers/net/ethernet/amd/ |
A D | sun3lance.c | 395 (dvma_vtob(MEM->tx_head) >> 16); in lance_probe() 399 (dvma_vtob(MEM->tx_head)))); in lance_probe() 461 MEM->tx_head[i].flag = 0; in lance_init_ring() 462 MEM->tx_head[i].base_hi = in lance_init_ring() 464 MEM->tx_head[i].length = 0; in lance_init_ring() 465 MEM->tx_head[i].misc = 0; in lance_init_ring() 493 (dvma_vtob(MEM->tx_head) >> 16); in lance_init_ring() 549 i, MEM->tx_head[i].base, in lance_start_xmit() 550 -MEM->tx_head[i].length, in lance_start_xmit() 551 MEM->tx_head[i].misc ); in lance_start_xmit() [all …]
|
A D | atarilance.c | 157 struct lance_tx_head tx_head[TX_RING_SIZE]; member 703 MEM->tx_head[i].base = offset; in lance_init_ring() 704 MEM->tx_head[i].flag = TMD1_OWN_HOST; in lance_init_ring() 705 MEM->tx_head[i].base_hi = 0; in lance_init_ring() 706 MEM->tx_head[i].length = 0; in lance_init_ring() 707 MEM->tx_head[i].misc = 0; in lance_init_ring() 754 i, MEM->tx_head[i].base, in lance_tx_timeout() 755 -MEM->tx_head[i].length, in lance_tx_timeout() 756 MEM->tx_head[i].misc )); in lance_tx_timeout() 812 head = &(MEM->tx_head[entry]); in lance_start_xmit() [all …]
|
A D | au1000_eth.c | 703 aup->tx_head = (aup->tx_dma_ring[0]->buff_stat & 0xC) >> 2; in au1000_init() 704 aup->tx_tail = aup->tx_head; in au1000_init() 967 skb->data, aup->tx_head); in au1000_tx() 969 ptxd = aup->tx_dma_ring[aup->tx_head]; in au1000_tx() 986 pDB = aup->tx_db_inuse[aup->tx_head]; in au1000_tx() 1002 aup->tx_head = (aup->tx_head + 1) & (NUM_TX_DMA - 1); in au1000_tx()
|
A D | au1000_eth.h | 78 u32 tx_head; member
|
/linux/drivers/infiniband/hw/hfi1/ |
A D | iowait.h | 52 struct list_head tx_head; member 289 if (!list_empty(&wait->tx_head)) { in iowait_get_txhead() 291 &wait->tx_head, in iowait_get_txhead() 304 if (!list_empty(&w->tx_head)) { in iowait_get_desc() 305 tx = list_first_entry(&w->tx_head, struct sdma_txreq, in iowait_get_desc() 327 if (!list_empty(&w->tx_head)) { in iowait_update_priority() 328 tx = list_first_entry(&w->tx_head, struct sdma_txreq, in iowait_update_priority() 409 return !list_empty(&wait->tx_head); in iowait_packet_queued()
|
A D | trace_tx.h | 371 __field(u16, tx_head) 378 __entry->tx_head = sde->tx_head; 391 __entry->tx_head, 407 __field(u16, tx_head) 414 __entry->tx_head = sde->tx_head; 425 __entry->tx_head,
|
/linux/drivers/staging/vt6656/ |
A D | rxtx.c | 330 mic_hdr = &tx_buffer->tx_head.tx_rts.tx.mic.hdr; in vnt_fill_txkey() 332 mic_hdr = &tx_buffer->tx_head.tx_cts.tx.mic.hdr; in vnt_fill_txkey() 334 mic_hdr = &tx_buffer->tx_head.tx_ab.tx.mic.hdr; in vnt_fill_txkey() 381 union vnt_tx_head *tx_head = &tx_buffer->tx_head; in vnt_rxtx_rts() local 382 struct vnt_rrv_time_rts *buf = &tx_head->tx_rts.rts; in vnt_rxtx_rts() 394 head = &tx_head->tx_rts.tx.mic.head; in vnt_rxtx_rts() 404 union vnt_tx_head *tx_head = &tx_buffer->tx_head; in vnt_rxtx_cts() local 415 head = &tx_head->tx_cts.tx.mic.head; in vnt_rxtx_cts() 425 union vnt_tx_head *tx_head = &tx_buffer->tx_head; in vnt_rxtx_ab() local 426 struct vnt_rrv_time_ab *buf = &tx_head->tx_ab.ab; in vnt_rxtx_ab() [all …]
|
/linux/drivers/net/ethernet/i825xx/ |
A D | ether1.c | 502 priv(dev)->tx_head = NOP_ADDR + NOP_SIZE; in ether1_init_for_open() 619 if (priv(dev)->tx_head + size > TX_AREA_END) { in ether1_txalloc() 620 if (tail > priv(dev)->tx_head) in ether1_txalloc() 625 priv(dev)->tx_head = start + size; in ether1_txalloc() 627 if (priv(dev)->tx_head < tail && (priv(dev)->tx_head + size) > tail) in ether1_txalloc() 629 start = priv(dev)->tx_head; in ether1_txalloc() 630 priv(dev)->tx_head += size; in ether1_txalloc() 729 tmp = priv(dev)->tx_head; in ether1_sendpacket() 731 priv(dev)->tx_head = tmp; in ether1_sendpacket() 834 caddr = priv(dev)->tx_head; in ether1_xmit_done() [all …]
|
/linux/drivers/net/ethernet/hisilicon/ |
A D | hip04_eth.c | 230 unsigned int tx_head; member 456 count = tx_count(READ_ONCE(priv->tx_head), tx_tail); in hip04_tx_reclaim() 510 unsigned int tx_head = priv->tx_head, count; in hip04_mac_start_xmit() local 511 struct tx_desc *desc = &priv->tx_desc[tx_head]; in hip04_mac_start_xmit() 515 count = tx_count(tx_head, READ_ONCE(priv->tx_tail)); in hip04_mac_start_xmit() 527 priv->tx_skb[tx_head] = skb; in hip04_mac_start_xmit() 528 priv->tx_phys[tx_head] = phys; in hip04_mac_start_xmit() 540 phys = priv->tx_desc_dma + tx_head * sizeof(struct tx_desc); in hip04_mac_start_xmit() 548 priv->tx_head = TX_NEXT(tx_head); in hip04_mac_start_xmit() 730 priv->tx_head = 0; in hip04_mac_open()
|
/linux/drivers/net/can/ |
A D | ti_hecc.c | 197 u32 tx_head; member 204 return priv->tx_head & HECC_TX_MB_MASK; in get_tx_head_mb() 214 return (priv->tx_head >> HECC_TX_PRIO_SHIFT) & MAX_TX_PRIO; in get_tx_head_prio() 370 priv->tx_head = HECC_TX_MASK; in ti_hecc_start() 493 priv->tx_head, priv->tx_tail); in ti_hecc_xmit() 519 --priv->tx_head; in ti_hecc_xmit() 521 (priv->tx_head & HECC_TX_MASK) == HECC_TX_MASK) { in ti_hecc_xmit() 748 while (priv->tx_tail - priv->tx_head > 0) { in ti_hecc_interrupt() 767 if ((priv->tx_head == priv->tx_tail && in ti_hecc_interrupt() 768 ((priv->tx_head & HECC_TX_MASK) != HECC_TX_MASK)) || in ti_hecc_interrupt() [all …]
|
/linux/drivers/net/ethernet/qualcomm/ |
A D | qca_uart.c | 57 unsigned char *tx_head; /* pointer to next XMIT byte */ member 151 written = serdev_device_write_buf(qca->serdev, qca->tx_head, in qcauart_transmit() 155 qca->tx_head += written; in qcauart_transmit() 240 qca->tx_head = qca->tx_buffer + written; in qcauart_netdev_xmit()
|
/linux/drivers/net/can/usb/etas_es58x/ |
A D | es58x_core.c | 280 u32 num_echo_skb = priv->tx_head - priv->tx_tail; in es58x_is_echo_skb_threshold_reached() 332 rcv_packet_idx, priv->tx_tail, priv->tx_head, in es58x_can_get_echo_skb_recovery() 333 priv->tx_head - priv->tx_tail, in es58x_can_get_echo_skb_recovery() 341 } else if ((s32)(rcv_packet_idx - priv->tx_head) >= 0) { in es58x_can_get_echo_skb_recovery() 353 if (priv->tx_tail == priv->tx_head) in es58x_can_get_echo_skb_recovery() 477 priv->tx_head = 0; in es58x_can_reset_echo_fifo() 506 priv->tx_head--; in es58x_flush_pending_tx_msg() 508 can_free_echo_skb(netdev, priv->tx_head & fifo_mask, in es58x_flush_pending_tx_msg() 544 priv->tx_head - priv->tx_tail, in es58x_tx_ack_msg() 1957 priv->tx_head & es58x_dev->param->fifo_mask, in es58x_start_xmit() [all …]
|
/linux/drivers/net/ethernet/seeq/ |
A D | ether3.c | 335 priv(dev)->tx_head = 0; in ether3_init_for_open() 466 priv(dev)->tx_head, priv(dev)->tx_tail); in ether3_timeout() 474 priv(dev)->tx_head = priv(dev)->tx_tail = 0; in ether3_timeout() 502 next_ptr = (priv(dev)->tx_head + 1) & 15; in ether3_sendpacket() 511 ptr = 0x600 * priv(dev)->tx_head; in ether3_sendpacket() 512 priv(dev)->tx_head = next_ptr; in ether3_sendpacket() 534 next_ptr = (priv(dev)->tx_head + 1) & 15; in ether3_sendpacket()
|
/linux/drivers/net/ethernet/dec/tulip/ |
A D | de2104x.c | 96 (((CP)->tx_tail <= (CP)->tx_head) ? \ 98 (CP)->tx_tail - (CP)->tx_head - 1) 292 unsigned tx_head; member 541 unsigned tx_head = de->tx_head; in de_tx() local 544 while (tx_tail != tx_head) { in de_tx() 622 entry = de->tx_head; in de_start_xmit() 644 de->tx_head = NEXT_TX(entry); in de_start_xmit() 753 entry = de->tx_head; in __de_set_rx_mode() 791 de->tx_head = NEXT_TX(entry); in __de_set_rx_mode() 895 de->tx_head = de->tx_tail = 0; in de_stop_hw() [all …]
|
/linux/drivers/net/ethernet/actions/ |
A D | owl-emac.c | 496 unsigned int tx_head; in owl_emac_setup_frame_xmit() local 515 tx_head = ring->head; in owl_emac_setup_frame_xmit() 516 desc = &ring->descs[tx_head]; in owl_emac_setup_frame_xmit() 530 ring->skbs[tx_head] = skb; in owl_emac_setup_frame_xmit() 531 ring->skbs_dma[tx_head] = dma_addr; in owl_emac_setup_frame_xmit() 572 unsigned int tx_head; in owl_emac_ndo_start_xmit() local 586 tx_head = ring->head; in owl_emac_ndo_start_xmit() 587 desc = &ring->descs[tx_head]; in owl_emac_ndo_start_xmit() 605 ring->skbs[tx_head] = skb; in owl_emac_ndo_start_xmit() 606 ring->skbs_dma[tx_head] = dma_addr; in owl_emac_ndo_start_xmit()
|
/linux/drivers/net/ethernet/silan/ |
A D | sc92031.c | 280 unsigned tx_head; member 410 while (priv->tx_head - priv->tx_tail > 0) { in _sc92031_tx_clear() 414 priv->tx_head = priv->tx_tail = 0; in _sc92031_tx_clear() 657 while (priv->tx_head - priv->tx_tail > 0) { in _sc92031_tx_tasklet() 953 BUG_ON(priv->tx_head - priv->tx_tail >= NUM_TX_DESC); in sc92031_start_xmit() 955 entry = priv->tx_head++ % NUM_TX_DESC; in sc92031_start_xmit() 979 if (priv->tx_head - priv->tx_tail >= NUM_TX_DESC) in sc92031_start_xmit() 1010 priv->tx_head = priv->tx_tail = 0; in sc92031_open()
|
/linux/drivers/net/can/rcar/ |
A D | rcar_can.c | 98 u32 tx_head; member 383 if (priv->tx_head - priv->tx_tail <= unsent) in rcar_can_tx_done() 619 priv->tx_dlc[priv->tx_head % RCAR_CAN_FIFO_DEPTH] = cf->len; in rcar_can_start_xmit() 620 can_put_echo_skb(skb, ndev, priv->tx_head % RCAR_CAN_FIFO_DEPTH, 0); in rcar_can_start_xmit() 621 priv->tx_head++; in rcar_can_start_xmit() 628 if (priv->tx_head - priv->tx_tail >= RCAR_CAN_FIFO_DEPTH) in rcar_can_start_xmit()
|
/linux/drivers/net/ethernet/realtek/ |
A D | 8139cp.c | 116 (((CP)->tx_tail <= (CP)->tx_head) ? \ 117 (CP)->tx_tail + (CP_TX_RING_SIZE - 1) - (CP)->tx_head : \ 118 (CP)->tx_tail - (CP)->tx_head - 1) 341 unsigned tx_head ____cacheline_aligned; 650 unsigned tx_head = cp->tx_head; in cp_tx() local 654 while (tx_tail != tx_head) { in cp_tx() 751 entry = cp->tx_head; in cp_start_xmit() 867 cp->tx_head = NEXT_TX(entry); in cp_start_xmit() 966 cp->tx_head = cp->tx_tail = 0; in cp_stop_hw() 1099 cp->tx_head = cp->tx_tail = 0; in cp_init_rings_index() [all …]
|
/linux/drivers/infiniband/ulp/ipoib/ |
A D | ipoib_ib.c | 626 tx_req = &priv->tx_ring[priv->tx_head & (ipoib_sendq_size - 1)]; in ipoib_send() 653 rc = post_send(priv, priv->tx_head & (ipoib_sendq_size - 1), in ipoib_send() 666 rc = priv->tx_head; in ipoib_send() 667 ++priv->tx_head; in ipoib_send() 796 while (priv->tx_head != priv->tx_tail || recvs_pending(dev)) { in ipoib_ib_dev_stop_default() 800 priv->tx_head - priv->tx_tail, in ipoib_ib_dev_stop_default() 807 while ((int)priv->tx_tail - (int)priv->tx_head < 0) { in ipoib_ib_dev_stop_default()
|
/linux/drivers/net/ethernet/cadence/ |
A D | macb_main.c | 1152 queue->tx_head = 0; in macb_tx_error_task() 1183 head = queue->tx_head; in macb_tx_interrupt() 1876 unsigned int len, entry, i, tx_head = queue->tx_head; in macb_tx_map() local 1920 tx_head++; in macb_tx_map() 1950 tx_head++; in macb_tx_map() 1970 i = tx_head; in macb_tx_map() 2006 if (i == queue->tx_head) { in macb_tx_map() 2027 queue->tx_head = tx_head; in macb_tx_map() 2034 for (i = queue->tx_head; i != tx_head; i++) { in macb_tx_map() 2432 queue->tx_head = 0; in gem_init_rings() [all …]
|
/linux/drivers/net/ethernet/calxeda/ |
A D | xgmac.c | 362 unsigned int tx_head; member 401 dma_ring_space((p)->tx_head, (p)->tx_tail, DMA_TX_RING_SZ) 770 priv->tx_head = 0; in xgmac_dma_desc_rings_init() 868 while (dma_ring_cnt(priv->tx_head, priv->tx_tail, DMA_TX_RING_SZ)) { in xgmac_tx_complete() 878 priv->tx_head, priv->tx_tail); in xgmac_tx_complete() 925 priv->tx_head = 0; in xgmac_tx_timeout_work() 1091 entry = priv->tx_head; in xgmac_xmit() 1136 priv->tx_head = dma_ring_incr(entry, DMA_TX_RING_SZ); in xgmac_xmit() 1150 entry = priv->tx_head; in xgmac_xmit()
|
/linux/arch/sparc/kernel/ |
A D | ldc.c | 123 unsigned long tx_head; member 227 if (t == lp->tx_head) in handshake_get_tx_packet() 247 return lp->tx_head; in head_for_data() 529 &lp->tx_head, in ldc_abort() 926 &lp->tx_head, in ldc_tx() 931 orig_state, lp->chan_state, lp->tx_head, lp->tx_tail); in ldc_tx() 1339 &lp->tx_head, in ldc_bind() 1346 lp->tx_acked = lp->tx_head; in ldc_bind() 1495 lp->tx_head, lp->tx_tail, lp->tx_num_entries, in __ldc_print() 1506 hv_err = sun4v_ldc_tx_get_state(lp->id, &lp->tx_head, &lp->tx_tail, in write_raw() [all …]
|
/linux/drivers/net/can/mscan/ |
A D | mscan.c | 135 INIT_LIST_HEAD(&priv->tx_head); in mscan_start() 271 list_add_tail(&priv->tx_queue[buf_id].list, &priv->tx_head); in mscan_start_xmit() 440 list_for_each_safe(pos, tmp, &priv->tx_head) { in mscan_isr() 456 if (list_empty(&priv->tx_head)) { in mscan_isr()
|
/linux/drivers/net/ethernet/ti/ |
A D | tlan.c | 1061 dev->name, priv->tx_head, priv->tx_tail); in tlan_start_tx() 1357 priv->tx_head, priv->tx_tail); in tlan_handle_tx_eof() 1358 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof() 1381 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof() 1391 priv->tx_head, priv->tx_tail); in tlan_handle_tx_eof() 1392 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof() 1394 + sizeof(struct tlan_list)*priv->tx_head; in tlan_handle_tx_eof() 1633 priv->tx_head, priv->tx_tail); in tlan_handle_tx_eoc() 1634 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eoc() 1636 + sizeof(struct tlan_list)*priv->tx_head; in tlan_handle_tx_eoc() [all …]
|