Home
last modified time | relevance | path

Searched refs:tx_desc (Results 1 – 25 of 114) sorted by relevance

12345

/linux/drivers/infiniband/ulp/isert/
A Dib_isert.c805 struct iser_tx_desc *tx_desc) in __isert_create_send_desc() argument
811 tx_desc->num_sge = 1; in __isert_create_send_desc()
848 tx_desc->dma_addr = dma_addr; in isert_init_tx_hdrs()
849 tx_desc->tx_sg[0].addr = tx_desc->dma_addr; in isert_init_tx_hdrs()
854 tx_desc->tx_sg[0].addr, tx_desc->tx_sg[0].length, in isert_init_tx_hdrs()
855 tx_desc->tx_sg[0].lkey); in isert_init_tx_hdrs()
864 struct iser_tx_desc *tx_desc = &isert_cmd->tx_desc; in isert_init_send_wr() local
942 tx_desc->num_sge = 2; in isert_put_login_tx()
1504 if (tx_desc->dma_addr != 0) { in isert_unmap_tx_desc()
1508 tx_desc->dma_addr = 0; in isert_unmap_tx_desc()
[all …]
/linux/drivers/crypto/ccp/
A Dccp-dmaengine.c148 desc->tx_desc.cookie, cmd); in ccp_issue_next_cmd()
212 tx_desc = &desc->tx_desc; in ccp_handle_active_desc()
214 tx_desc = NULL; in ccp_handle_active_desc()
227 dma_cookie_complete(tx_desc); in ccp_handle_active_desc()
228 dma_descriptor_unmap(tx_desc); in ccp_handle_active_desc()
235 if (tx_desc) { in ccp_handle_active_desc()
238 dma_run_dependencies(tx_desc); in ccp_handle_active_desc()
300 tx_desc); in ccp_tx_submit()
341 desc->tx_desc.flags = flags; in ccp_alloc_dma_desc()
503 return &desc->tx_desc; in ccp_prep_dma_memcpy()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx4/
A Den_tx.c240 __be32 *ptr = (__be32 *)tx_desc; in mlx4_en_stamp_wqe()
244 if (likely((void *)tx_desc + in mlx4_en_stamp_wqe()
308 if (likely((void *)tx_desc + in mlx4_en_free_tx_desc()
401 struct mlx4_en_tx_desc *tx_desc; in mlx4_en_handle_err_cqe() local
767 tx_desc->ctrl.qpn_vlan = qpn_vlan; in mlx4_en_tx_write_desc()
868 struct mlx4_en_tx_desc *tx_desc; in mlx4_en_xmit() local
944 data = &tx_desc->data; in mlx4_en_xmit()
1132 tx_desc->data.lkey = ring->mr_key; in mlx4_en_init_tx_xdp_ring_descs()
1143 struct mlx4_en_tx_desc *tx_desc; in mlx4_en_xmit_frame() local
1163 data = &tx_desc->data; in mlx4_en_xmit_frame()
[all …]
/linux/drivers/infiniband/ulp/iser/
A Diser_initiator.c164 struct iser_tx_desc *tx_desc) in iser_create_send_desc() argument
172 tx_desc->iser_header.flags = ISER_VER; in iser_create_send_desc()
173 tx_desc->num_sge = 1; in iser_create_send_desc()
379 tx_desc->type = ISCSI_TX_SCSI_COMMAND; in iser_send_command()
380 tx_desc->cqe.done = iser_cmd_comp; in iser_send_command()
443 struct iser_tx_desc *tx_desc; in iser_send_data_out() local
459 if (!tx_desc) in iser_send_data_out()
462 tx_desc->type = ISCSI_TX_DATAOUT; in iser_send_data_out()
463 tx_desc->cqe.done = iser_dataout_comp; in iser_send_data_out()
473 tx_dsg = &tx_desc->tx_sg[1]; in iser_send_data_out()
[all …]
A Discsi_iser.c203 struct iser_tx_desc *tx_desc) in iser_initialize_task_headers() argument
218 tx_desc->inv_wr.next = NULL; in iser_initialize_task_headers()
219 tx_desc->reg_wr.wr.next = NULL; in iser_initialize_task_headers()
220 tx_desc->mapped = true; in iser_initialize_task_headers()
221 tx_desc->dma_addr = dma_addr; in iser_initialize_task_headers()
222 tx_desc->tx_sg[0].addr = tx_desc->dma_addr; in iser_initialize_task_headers()
223 tx_desc->tx_sg[0].length = ISER_HEADERS_LEN; in iser_initialize_task_headers()
224 tx_desc->tx_sg[0].lkey = device->pd->local_dma_lkey; in iser_initialize_task_headers()
377 struct iser_tx_desc *tx_desc = &iser_task->desc; in iscsi_iser_cleanup_task() local
385 if (likely(tx_desc->mapped)) { in iscsi_iser_cleanup_task()
[all …]
A Diser_memory.c239 struct iser_tx_desc *tx_desc = &iser_task->desc; in iser_reg_sig_mr() local
243 struct ib_reg_wr *wr = &tx_desc->reg_wr; in iser_reg_sig_mr()
254 iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr); in iser_reg_sig_mr()
267 wr->wr.next = &tx_desc->send_wr; in iser_reg_sig_mr()
296 struct iser_tx_desc *tx_desc = &iser_task->desc; in iser_fast_reg_mr() local
299 struct ib_reg_wr *wr = &tx_desc->reg_wr; in iser_fast_reg_mr()
303 iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr); in iser_fast_reg_mr()
314 wr->wr.next = &tx_desc->send_wr; in iser_fast_reg_mr()
/linux/drivers/net/ethernet/intel/fm10k/
A Dfm10k_main.c739 struct fm10k_tx_desc *tx_desc; in fm10k_tso() local
769 tx_desc->hdrlen = hdrlen; in fm10k_tso()
786 struct fm10k_tx_desc *tx_desc; in fm10k_tx_csum() local
863 tx_desc->hdrlen = 0; in fm10k_tx_csum()
864 tx_desc->mss = 0; in fm10k_tx_csum()
894 tx_desc->flags = desc_flags; in fm10k_tx_desc_push()
930 struct fm10k_tx_desc *tx_desc; in fm10k_tx_map() local
945 tx_desc->vlan = 0; in fm10k_tx_map()
1225 tx_desc++; in fm10k_clean_tx_irq()
1245 tx_desc++; in fm10k_clean_tx_irq()
[all …]
/linux/drivers/net/ethernet/intel/ice/
A Dice_txrx.c36 struct ice_tx_desc *tx_desc; in ice_prgm_fdir_fltr() local
219 struct ice_tx_desc *tx_desc; in ice_clean_tx_irq() local
267 tx_desc++; in ice_clean_tx_irq()
288 tx_desc++; in ice_clean_tx_irq()
296 prefetch(tx_desc); in ice_clean_tx_irq()
1621 tx_desc++; in ice_tx_map()
1642 tx_desc++; in ice_tx_map()
2426 tx_desc->buf_addr = 0; in ice_clean_ctrl_tx_irq()
2431 tx_desc++; in ice_clean_ctrl_tx_irq()
2453 tx_desc->buf_addr = 0; in ice_clean_ctrl_tx_irq()
[all …]
A Dice_txrx_lib.c272 struct ice_tx_desc *tx_desc; in ice_xmit_xdp_ring() local
297 tx_desc = ICE_TX_DESC(xdp_ring, i); in ice_xmit_xdp_ring()
298 tx_desc->buf_addr = cpu_to_le64(dma); in ice_xmit_xdp_ring()
299 tx_desc->cmd_type_offset_bsz = ice_build_ctob(ICE_TX_DESC_CMD_EOP, 0, in ice_xmit_xdp_ring()
305 tx_desc = ICE_TX_DESC(xdp_ring, xdp_ring->next_rs); in ice_xmit_xdp_ring()
306 tx_desc->cmd_type_offset_bsz |= in ice_xmit_xdp_ring()
313 tx_desc = ICE_TX_DESC(xdp_ring, xdp_ring->next_rs); in ice_xmit_xdp_ring()
314 tx_desc->cmd_type_offset_bsz |= in ice_xmit_xdp_ring()
A Dice_xsk.c623 struct ice_tx_desc *tx_desc = NULL; in ice_xmit_zc() local
649 tx_desc->buf_addr = cpu_to_le64(dma); in ice_xmit_zc()
650 tx_desc->cmd_type_offset_bsz = in ice_xmit_zc()
658 if (tx_desc) { in ice_xmit_zc()
691 struct ice_tx_desc *tx_desc; in ice_clean_tx_irq_zc() local
696 tx_desc = ICE_TX_DESC(xdp_ring, ntc); in ice_clean_tx_irq_zc()
701 if (!(tx_desc->cmd_type_offset_bsz & in ice_clean_tx_irq_zc()
715 tx_desc->cmd_type_offset_bsz = 0; in ice_clean_tx_irq_zc()
717 tx_desc++; in ice_clean_tx_irq_zc()
723 tx_desc = ICE_TX_DESC(xdp_ring, 0); in ice_clean_tx_irq_zc()
[all …]
/linux/drivers/net/ethernet/intel/ixgbe/
A Dixgbe_xsk.c385 union ixgbe_adv_tx_desc *tx_desc = NULL; in ixgbe_xmit_zc() local
410 tx_desc = IXGBE_TX_DESC(xdp_ring, xdp_ring->next_to_use); in ixgbe_xmit_zc()
411 tx_desc->read.buffer_addr = cpu_to_le64(dma); in ixgbe_xmit_zc()
418 tx_desc->read.cmd_type_len = cpu_to_le32(cmd_type); in ixgbe_xmit_zc()
419 tx_desc->read.olinfo_status = in ixgbe_xmit_zc()
427 if (tx_desc) { in ixgbe_xmit_zc()
451 union ixgbe_adv_tx_desc *tx_desc; in ixgbe_clean_xdp_tx_irq() local
456 tx_desc = IXGBE_TX_DESC(tx_ring, ntc); in ixgbe_clean_xdp_tx_irq()
473 tx_desc++; in ixgbe_clean_xdp_tx_irq()
478 tx_desc = IXGBE_TX_DESC(tx_ring, 0); in ixgbe_clean_xdp_tx_irq()
[all …]
/linux/drivers/dma/ptdma/
A Dptdma-dmaengine.c81 struct dma_async_tx_descriptor *tx_desc; in pt_handle_active_desc() local
94 tx_desc = &desc->vd.tx; in pt_handle_active_desc()
97 tx_desc = NULL; in pt_handle_active_desc()
106 dma_cookie_complete(tx_desc); in pt_handle_active_desc()
107 dma_descriptor_unmap(tx_desc); in pt_handle_active_desc()
115 if (tx_desc) { in pt_handle_active_desc()
116 dmaengine_desc_get_callback_invoke(tx_desc, NULL); in pt_handle_active_desc()
117 dma_run_dependencies(tx_desc); in pt_handle_active_desc()
/linux/drivers/staging/rtl8712/
A Drtl8712_xmit.c251 struct tx_desc *ptx_desc = (struct tx_desc *)pxmitbuf->pbuf; in r8712_construct_txaggr_cmd_desc()
285 struct tx_desc *ptx_desc = (struct tx_desc *)pxmitbuf->pbuf; in r8712_append_mpdu_unit()
344 (((struct tx_desc *)pxmitbuf->pbuf)->txdw0 & 0x0000ffff); in r8712_xmitframe_aggr_next()
351 (((struct tx_desc *)pxmitbuf->pbuf)->txdw0 & 0x0000ffff); in r8712_xmitframe_aggr_next()
359 struct tx_desc *ptxdesc = pxmitbuf->pbuf; in r8712_dump_aggr_xframe()
404 struct tx_desc *ptxdesc = (struct tx_desc *)pmem; in update_txdesc()
412 struct tx_desc txdesc_mp; in update_txdesc()
414 memcpy(&txdesc_mp, ptxdesc, sizeof(struct tx_desc)); in update_txdesc()
415 memset(ptxdesc, 0, sizeof(struct tx_desc)); in update_txdesc()
536 struct tx_desc *ptxdesc_mp; in update_txdesc()
A Drtl8712_xmit.h39 #define tx_cmd tx_desc
78 struct tx_desc { struct
91 struct tx_desc txdesc; argument
/linux/drivers/staging/mt7621-dma/
A Dhsdma-mt7621.c232 struct hsdma_desc *tx_desc; in hsdma_dump_desc() local
240 tx_desc = &chan->tx_ring[i]; in hsdma_dump_desc()
245 i, tx_desc->addr0, tx_desc->flags, in hsdma_dump_desc()
317 struct hsdma_desc *tx_desc, *rx_desc; in mtk_hsdma_start_transfer() local
329 tx_desc = &chan->tx_ring[chan->tx_idx]; in mtk_hsdma_start_transfer()
337 tx_desc->addr1 = src; in mtk_hsdma_start_transfer()
338 tx_desc->flags |= HSDMA_DESC_PLEN1(tlen); in mtk_hsdma_start_transfer()
340 tx_desc->addr0 = src; in mtk_hsdma_start_transfer()
341 tx_desc->flags = HSDMA_DESC_PLEN0(tlen); in mtk_hsdma_start_transfer()
351 tx_desc->flags |= HSDMA_DESC_LS0; in mtk_hsdma_start_transfer()
[all …]
/linux/drivers/net/ethernet/intel/i40e/
A Di40e_xsk.c401 struct i40e_tx_desc *tx_desc; in i40e_xmit_pkt() local
407 tx_desc = I40E_TX_DESC(xdp_ring, xdp_ring->next_to_use++); in i40e_xmit_pkt()
408 tx_desc->buffer_addr = cpu_to_le64(dma); in i40e_xmit_pkt()
409 tx_desc->cmd_type_offset_bsz = build_ctob(I40E_TX_DESC_CMD_ICRC | I40E_TX_DESC_CMD_EOP, in i40e_xmit_pkt()
419 struct i40e_tx_desc *tx_desc; in i40e_xmit_pkt_batch() local
427 tx_desc = I40E_TX_DESC(xdp_ring, ntu++); in i40e_xmit_pkt_batch()
428 tx_desc->buffer_addr = cpu_to_le64(dma); in i40e_xmit_pkt_batch()
429 tx_desc->cmd_type_offset_bsz = build_ctob(I40E_TX_DESC_CMD_ICRC | in i40e_xmit_pkt_batch()
455 struct i40e_tx_desc *tx_desc; in i40e_set_rs_bit() local
457 tx_desc = I40E_TX_DESC(xdp_ring, ntu); in i40e_set_rs_bit()
[all …]
A Di40e_txrx.c90 struct i40e_tx_desc *tx_desc; in i40e_program_fdir_filter() local
142 tx_desc->cmd_type_offset_bsz = in i40e_program_fdir_filter()
151 first->next_to_watch = tx_desc; in i40e_program_fdir_filter()
937 struct i40e_tx_desc *tx_desc; in i40e_clean_tx_irq() local
959 if (tx_head == tx_desc) in i40e_clean_tx_irq()
986 while (tx_desc != eop_desc) { in i40e_clean_tx_irq()
991 tx_desc++; in i40e_clean_tx_irq()
1011 tx_desc++; in i40e_clean_tx_irq()
1019 prefetch(tx_desc); in i40e_clean_tx_irq()
3528 tx_desc++; in i40e_tx_map()
[all …]
/linux/drivers/net/ethernet/seeq/
A Dsgiseeq.c96 struct sgiseeq_tx_desc *tx_desc; member
197 dma_sync_desc_dev(dev, &sp->tx_desc[i]); in seeq_init_ring()
230 if (sp->tx_desc[i].skb) { in seeq_purge_ring()
231 dev_kfree_skb(sp->tx_desc[i].skb); in seeq_purge_ring()
232 sp->tx_desc[i].skb = NULL; in seeq_purge_ring()
253 struct sgiseeq_tx_desc *t = gpriv->tx_desc; in sgiseeq_dump_rings()
443 td = &sp->tx_desc[i]; in kick_tx()
448 td = &sp->tx_desc[i]; in kick_tx()
480 td = &sp->tx_desc[j]; in sgiseeq_tx()
613 td = &sp->tx_desc[entry]; in sgiseeq_start_xmit()
[all …]
/linux/drivers/net/ethernet/intel/iavf/
A Diavf_txrx.c195 struct iavf_tx_desc *tx_desc; in iavf_clean_tx_irq() local
240 while (tx_desc != eop_desc) { in iavf_clean_tx_irq()
242 tx_ring, tx_desc, tx_buf); in iavf_clean_tx_irq()
245 tx_desc++; in iavf_clean_tx_irq()
265 tx_desc++; in iavf_clean_tx_irq()
273 prefetch(tx_desc); in iavf_clean_tx_irq()
2263 struct iavf_tx_desc *tx_desc; in iavf_tx_map() local
2300 tx_desc++; in iavf_tx_map()
2321 tx_desc++; in iavf_tx_map()
2350 tx_desc->cmd_type_offset_bsz = in iavf_tx_map()
[all …]
/linux/drivers/net/ethernet/oki-semi/pch_gbe/
A Dpch_gbe_main.c1072 struct pch_gbe_tx_desc *tx_desc; in pch_gbe_tx_queue() local
1159 tx_desc->length = (tmp_skb->len); in pch_gbe_tx_queue()
1430 struct pch_gbe_tx_desc *tx_desc; in pch_gbe_alloc_tx_buffers() local
1458 struct pch_gbe_tx_desc *tx_desc; in pch_gbe_clean_tx() local
1472 tx_desc->gbec_status, tx_desc->dma_status); in pch_gbe_clean_tx()
1498 tx_desc->gbec_status); in pch_gbe_clean_tx()
1505 tx_desc->gbec_status); in pch_gbe_clean_tx()
1523 } else if ((tx_desc->gbec_status & in pch_gbe_clean_tx()
1547 tx_desc->gbec_status = DSC_INIT16; in pch_gbe_clean_tx()
1708 struct pch_gbe_tx_desc *tx_desc; in pch_gbe_setup_tx_resources() local
[all …]
/linux/drivers/net/ethernet/marvell/mvpp2/
A Dmvpp2_main.c223 struct mvpp2_tx_desc *tx_desc, in mvpp2_txdesc_size_set() argument
233 struct mvpp2_tx_desc *tx_desc, in mvpp2_txdesc_txq_set() argument
237 tx_desc->pp21.phys_txq = txq; in mvpp2_txdesc_txq_set()
239 tx_desc->pp22.phys_txq = txq; in mvpp2_txdesc_txq_set()
243 struct mvpp2_tx_desc *tx_desc, in mvpp2_txdesc_cmd_set() argument
2439 return txq->descs + tx_desc; in mvpp2_txq_next_desc_get()
3650 struct mvpp2_tx_desc *tx_desc; in mvpp2_xdp_submit_frame() local
4144 tx_desc->pp22.ptp_descriptor &= in mvpp2_tx_hw_tstamp()
4161 struct mvpp2_tx_desc *tx_desc; in mvpp2_tx_frag_process() local
4202 tx_desc = txq->descs + i; in mvpp2_tx_frag_process()
[all …]
/linux/drivers/net/ethernet/hisilicon/
A Dhip04_eth.c173 struct tx_desc { struct
226 struct tx_desc *tx_desc; member
451 struct tx_desc *desc; in hip04_tx_reclaim()
461 desc = &priv->tx_desc[tx_tail]; in hip04_tx_reclaim()
511 struct tx_desc *desc = &priv->tx_desc[tx_head]; in hip04_mac_start_xmit()
542 offsetof(struct tx_desc, send_addr)); in hip04_mac_start_xmit()
859 priv->tx_desc = dma_alloc_coherent(d, in hip04_alloc_ring()
860 TX_DESC_NUM * sizeof(struct tx_desc), in hip04_alloc_ring()
862 if (!priv->tx_desc) in hip04_alloc_ring()
889 dma_free_coherent(d, TX_DESC_NUM * sizeof(struct tx_desc), in hip04_free_ring()
[all …]
/linux/drivers/spi/
A Dspi-pxa2xx-dma.c136 struct dma_async_tx_descriptor *tx_desc, *rx_desc; in pxa2xx_spi_dma_prepare() local
139 tx_desc = pxa2xx_spi_dma_prepare_one(drv_data, DMA_MEM_TO_DEV, xfer); in pxa2xx_spi_dma_prepare()
140 if (!tx_desc) { in pxa2xx_spi_dma_prepare()
158 dmaengine_submit(tx_desc); in pxa2xx_spi_dma_prepare()
/linux/drivers/net/ethernet/
A Dec_bhf.c103 struct tx_desc { struct
144 struct tx_desc *tx_descs;
170 static void ec_bhf_send_packet(struct ec_bhf_priv *priv, struct tx_desc *desc) in ec_bhf_send_packet()
178 static int ec_bhf_desc_sent(struct tx_desc *desc) in ec_bhf_desc_sent()
286 struct tx_desc *desc; in ec_bhf_start_xmit()
361 priv->tx_dcount = priv->tx_buf.len / sizeof(struct tx_desc); in ec_bhf_setup_tx_descs()
362 priv->tx_descs = (struct tx_desc *)priv->tx_buf.buf; in ec_bhf_setup_tx_descs()
409 FIFO_SIZE * sizeof(struct tx_desc)); in ec_bhf_open()
/linux/drivers/dma/xilinx/
A Dxilinx_dpdma.c588 struct xilinx_dpdma_tx_desc *tx_desc) in xilinx_dpdma_chan_dump_tx_desc() argument
634 struct xilinx_dpdma_tx_desc *tx_desc; in xilinx_dpdma_chan_alloc_tx_desc() local
636 tx_desc = kzalloc(sizeof(*tx_desc), GFP_NOWAIT); in xilinx_dpdma_chan_alloc_tx_desc()
637 if (!tx_desc) in xilinx_dpdma_chan_alloc_tx_desc()
640 INIT_LIST_HEAD(&tx_desc->descriptors); in xilinx_dpdma_chan_alloc_tx_desc()
641 tx_desc->chan = chan; in xilinx_dpdma_chan_alloc_tx_desc()
642 tx_desc->error = false; in xilinx_dpdma_chan_alloc_tx_desc()
644 return tx_desc; in xilinx_dpdma_chan_alloc_tx_desc()
686 struct xilinx_dpdma_tx_desc *tx_desc; in xilinx_dpdma_chan_prep_interleaved_dma() local
700 if (!tx_desc) in xilinx_dpdma_chan_prep_interleaved_dma()
[all …]

Completed in 96 milliseconds

12345