/linux/drivers/net/ethernet/intel/igc/ |
A D | igc_dump.c | 144 dma_unmap_len(buffer_info, len), in igc_rings_dump() 195 dma_unmap_len(buffer_info, len), in igc_rings_dump() 204 dma_unmap_len(buffer_info, len), in igc_rings_dump()
|
/linux/drivers/net/ethernet/google/gve/ |
A D | gve_tx_dqo.c | 91 dma_unmap_len(cur_state, len[j]), in gve_tx_clean_pending_packets() 96 dma_unmap_len(cur_state, len[j]), in gve_tx_clean_pending_packets() 553 dma_unmap_len(pkt, len[i]), in gve_tx_add_skb_no_copy_dqo() 558 dma_unmap_len(pkt, len[i]), in gve_tx_add_skb_no_copy_dqo() 751 dma_unmap_len(pkt, len[0]), DMA_TO_DEVICE); in gve_unmap_packet() 754 dma_unmap_len(pkt, len[i]), DMA_TO_DEVICE); in gve_unmap_packet()
|
A D | gve_tx.c | 308 dma_unmap_len(info, len), in gve_tx_unmap_buf() 313 dma_unmap_len(info, len), in gve_tx_unmap_buf()
|
/linux/drivers/net/ethernet/intel/ice/ |
A D | ice_txrx.c | 120 if (dma_unmap_len(tx_buf, len)) in ice_unmap_and_free_tx_buf() 123 dma_unmap_len(tx_buf, len), in ice_unmap_and_free_tx_buf() 125 } else if (dma_unmap_len(tx_buf, len)) { in ice_unmap_and_free_tx_buf() 128 dma_unmap_len(tx_buf, len), in ice_unmap_and_free_tx_buf() 256 dma_unmap_len(tx_buf, len), in ice_clean_tx_irq() 276 if (dma_unmap_len(tx_buf, len)) { in ice_clean_tx_irq() 279 dma_unmap_len(tx_buf, len), in ice_clean_tx_irq() 2440 if (dma_unmap_len(tx_buf, len)) in ice_clean_ctrl_tx_irq() 2443 dma_unmap_len(tx_buf, len), in ice_clean_ctrl_tx_irq()
|
A D | ice_txrx_lib.c | 246 dma_unmap_len(tx_buf, len), DMA_TO_DEVICE); in ice_clean_xdp_irq()
|
A D | ice_xsk.c | 676 dma_unmap_len(tx_buf, len), DMA_TO_DEVICE); in ice_clean_xdp_tx_buf()
|
/linux/drivers/net/ethernet/arc/ |
A D | emac_main.c | 144 dma_unmap_len(tx_buff, len), DMA_TO_DEVICE); in arc_emac_tx_clean() 241 dma_unmap_len(rx_buff, len), DMA_FROM_DEVICE); in arc_emac_rx() 560 dma_unmap_len(tx_buff, len), in arc_free_tx_queue() 591 dma_unmap_len(rx_buff, len), in arc_free_rx_queue()
|
/linux/drivers/net/ethernet/intel/iavf/ |
A D | iavf_txrx.c | 35 if (dma_unmap_len(tx_buffer, len)) in iavf_unmap_and_free_tx_resource() 38 dma_unmap_len(tx_buffer, len), in iavf_unmap_and_free_tx_resource() 40 } else if (dma_unmap_len(tx_buffer, len)) { in iavf_unmap_and_free_tx_resource() 43 dma_unmap_len(tx_buffer, len), in iavf_unmap_and_free_tx_resource() 232 dma_unmap_len(tx_buf, len), in iavf_clean_tx_irq() 254 if (dma_unmap_len(tx_buf, len)) { in iavf_clean_tx_irq() 257 dma_unmap_len(tx_buf, len), in iavf_clean_tx_irq()
|
/linux/drivers/net/ethernet/chelsio/cxgb/ |
A D | sge.c | 516 dma_unmap_len(ce, dma_len), DMA_FROM_DEVICE); in free_freelQ_buffers() 630 if (likely(dma_unmap_len(ce, dma_len))) { in free_cmdQ_buffers() 633 dma_unmap_len(ce, dma_len), in free_cmdQ_buffers() 1064 dma_unmap_len(ce, dma_len), in get_packet() 1069 dma_unmap_len(ce, dma_len), in get_packet() 1082 dma_unmap_len(ce, dma_len), DMA_FROM_DEVICE); in get_packet() 1106 dma_unmap_len(ce, dma_len), DMA_FROM_DEVICE); in unexpected_offload()
|
/linux/drivers/net/ethernet/intel/fm10k/ |
A D | fm10k_netdev.c | 149 if (dma_unmap_len(tx_buffer, len)) in fm10k_unmap_and_free_tx_resource() 152 dma_unmap_len(tx_buffer, len), in fm10k_unmap_and_free_tx_resource() 154 } else if (dma_unmap_len(tx_buffer, len)) { in fm10k_unmap_and_free_tx_resource() 157 dma_unmap_len(tx_buffer, len), in fm10k_unmap_and_free_tx_resource()
|
A D | fm10k_main.c | 1215 dma_unmap_len(tx_buffer, len), in fm10k_clean_tx_irq() 1234 if (dma_unmap_len(tx_buffer, len)) { in fm10k_clean_tx_irq() 1237 dma_unmap_len(tx_buffer, len), in fm10k_clean_tx_irq()
|
/linux/drivers/net/ethernet/qlogic/ |
A D | qla3xxx.c | 1948 dma_unmap_len(&tx_cb->map[0], maplen), DMA_TO_DEVICE); in ql_process_mac_tx_intr() 1954 dma_unmap_len(&tx_cb->map[i], maplen), in ql_process_mac_tx_intr() 2024 dma_unmap_len(lrg_buf_cb2, maplen), DMA_FROM_DEVICE); in ql_process_mac_rx_intr() 2069 dma_unmap_len(lrg_buf_cb2, maplen), DMA_FROM_DEVICE); in ql_process_macip_rx_intr() 2424 dma_unmap_len(&tx_cb->map[seg], maplen), in ql_send_map() 2432 dma_unmap_len(&tx_cb->map[seg], maplen), in ql_send_map() 2714 dma_unmap_len(lrg_buf_cb, maplen), in ql_free_large_buffers() 3631 dma_unmap_len(&tx_cb->map[0], maplen), in ql_reset_work() 3636 dma_unmap_len(&tx_cb->map[j], maplen), in ql_reset_work()
|
/linux/include/linux/ |
A D | dma-mapping.h | 588 #define dma_unmap_len(PTR, LEN_NAME) ((PTR)->LEN_NAME) macro 595 #define dma_unmap_len(PTR, LEN_NAME) (0) macro
|
/linux/drivers/net/ethernet/intel/ixgbevf/ |
A D | ixgbevf_main.c | 313 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_irq() 331 if (dma_unmap_len(tx_buffer, len)) { in ixgbevf_clean_tx_irq() 334 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_irq() 2401 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_ring() 2420 if (dma_unmap_len(tx_buffer, len)) in ixgbevf_clean_tx_ring() 2423 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_ring() 4051 if (dma_unmap_len(tx_buffer, len)) in ixgbevf_tx_map() 4054 dma_unmap_len(tx_buffer, len), in ixgbevf_tx_map() 4063 if (dma_unmap_len(tx_buffer, len)) in ixgbevf_tx_map() 4066 dma_unmap_len(tx_buffer, len), in ixgbevf_tx_map()
|
/linux/drivers/net/ethernet/intel/i40e/ |
A D | i40e_txrx.c | 766 if (dma_unmap_len(tx_buffer, len)) in i40e_unmap_and_free_tx_resource() 769 dma_unmap_len(tx_buffer, len), in i40e_unmap_and_free_tx_resource() 771 } else if (dma_unmap_len(tx_buffer, len)) { in i40e_unmap_and_free_tx_resource() 774 dma_unmap_len(tx_buffer, len), in i40e_unmap_and_free_tx_resource() 978 dma_unmap_len(tx_buf, len), in i40e_clean_tx_irq() 1000 if (dma_unmap_len(tx_buf, len)) { in i40e_clean_tx_irq() 1003 dma_unmap_len(tx_buf, len), in i40e_clean_tx_irq()
|
A D | i40e_xsk.c | 508 dma_unmap_len(tx_bi, len), DMA_TO_DEVICE); in i40e_clean_xdp_tx_buffer()
|
/linux/drivers/net/ethernet/alacritech/ |
A D | slicoss.c | 369 dma_unmap_len(buff, map_len), DMA_TO_DEVICE); in slic_xmit_complete() 588 dma_unmap_len(buff, map_len), in slic_handle_receive() 908 dma_unmap_len(buff, map_len), DMA_TO_DEVICE); in slic_free_tx_queue() 950 dma_unmap_len(buff, map_len), in slic_free_rx_queue()
|
/linux/drivers/net/ethernet/broadcom/bnxt/ |
A D | bnxt_xdp.c | 89 dma_unmap_len(tx_buf, len), in bnxt_tx_int_xdp()
|
/linux/drivers/net/ethernet/atheros/alx/ |
A D | main.c | 58 if (dma_unmap_len(txb, size)) { in alx_free_txbuf() 61 dma_unmap_len(txb, size), in alx_free_txbuf() 245 dma_unmap_len(rxb, size), in alx_clean_rx_irq() 520 dma_unmap_len(cur_buf, size), in alx_free_rxring_buf()
|
/linux/drivers/net/ethernet/mediatek/ |
A D | mtk_eth_soc.c | 895 dma_unmap_len(tx_buf, dma_len0), in mtk_tx_unmap() 900 dma_unmap_len(tx_buf, dma_len0), in mtk_tx_unmap() 904 if (dma_unmap_len(tx_buf, dma_len0)) { in mtk_tx_unmap() 907 dma_unmap_len(tx_buf, dma_len0), in mtk_tx_unmap() 911 if (dma_unmap_len(tx_buf, dma_len1)) { in mtk_tx_unmap() 914 dma_unmap_len(tx_buf, dma_len1), in mtk_tx_unmap()
|
/linux/drivers/net/ethernet/marvell/ |
A D | skge.c | 999 dma_unmap_len(e, maplen), in skge_rx_clean() 2841 dma_unmap_len(e, maplen), DMA_TO_DEVICE); in skge_xmit_frame() 2845 dma_unmap_len(e, maplen), DMA_TO_DEVICE); in skge_xmit_frame() 2863 dma_unmap_len(e, maplen), DMA_TO_DEVICE); in skge_tx_unmap() 2866 dma_unmap_len(e, maplen), DMA_TO_DEVICE); in skge_tx_unmap() 3079 dma_unmap_len(e, maplen), in skge_rx_get() 3084 dma_unmap_len(e, maplen), in skge_rx_get() 3107 dma_unmap_len(&ee, maplen), DMA_FROM_DEVICE); in skge_rx_get()
|
/linux/drivers/net/ethernet/intel/igb/ |
A D | igb_main.c | 397 dma_unmap_len(buffer_info, len), in igb_dump() 446 dma_unmap_len(buffer_info, len), in igb_dump() 455 dma_unmap_len(buffer_info, len), in igb_dump() 4828 dma_unmap_len(tx_buffer, len), in igb_clean_tx_ring() 4847 if (dma_unmap_len(tx_buffer, len)) in igb_clean_tx_ring() 6228 if (dma_unmap_len(tx_buffer, len)) in igb_tx_map() 6231 dma_unmap_len(tx_buffer, len), in igb_tx_map() 6240 if (dma_unmap_len(tx_buffer, len)) in igb_tx_map() 6243 dma_unmap_len(tx_buffer, len), in igb_tx_map() 8086 dma_unmap_len(tx_buffer, len), in igb_clean_tx_irq() [all …]
|
/linux/drivers/net/ethernet/intel/ixgbe/ |
A D | ixgbe_xsk.c | 441 dma_unmap_len(tx_bi, len), DMA_TO_DEVICE); in ixgbe_clean_xdp_tx_buffer()
|
A D | ixgbe_main.c | 559 dma_unmap_len(tx_buffer, len), in ixgbe_print_buffer() 677 if (dma_unmap_len(tx_buffer, len) > 0) { in ixgbe_dump() 694 dma_unmap_len(tx_buffer, len), in ixgbe_dump() 705 dma_unmap_len(tx_buffer, len), in ixgbe_dump() 1162 dma_unmap_len(tx_buffer, len), in ixgbe_clean_tx_irq() 1180 if (dma_unmap_len(tx_buffer, len)) { in ixgbe_clean_tx_irq() 1183 dma_unmap_len(tx_buffer, len), in ixgbe_clean_tx_irq() 6017 dma_unmap_len(tx_buffer, len), in ixgbe_clean_tx_ring() 6036 if (dma_unmap_len(tx_buffer, len)) in ixgbe_clean_tx_ring() 8333 if (dma_unmap_len(tx_buffer, len)) in ixgbe_tx_map() [all …]
|
/linux/drivers/net/ethernet/alteon/ |
A D | acenic.c | 2044 if (dma_unmap_len(info, maplen)) { in ace_tx_int() 2047 dma_unmap_len(info, maplen), in ace_tx_int() 2326 if (dma_unmap_len(info, maplen)) { in ace_close() 2339 dma_unmap_len(info, maplen), in ace_close()
|