Searched refs:sw_tail (Results 1 – 11 of 11) sorted by relevance
/linux/drivers/crypto/cavium/zip/ |
A D | zip_device.c | 59 return ((zip_dev->iq[queue].sw_head - zip_dev->iq[queue].sw_tail) * in zip_cmd_queue_consumed() 110 zip_dbg("sw_tail : %lx", zip_dev->iq[queue].sw_tail); in zip_load_instr() 128 zip_dev->iq[queue].sw_head = zip_dev->iq[queue].sw_tail; in zip_load_instr() 151 zip_dev->iq[queue].sw_head, zip_dev->iq[queue].sw_tail, in zip_load_instr() 197 zip_dev->iq[queue].sw_head, zip_dev->iq[queue].sw_tail, in zip_update_cmd_bufs()
|
A D | zip_mem.c | 78 zip_dbg("Freeing cmd_qbuf 0x%lx\n", zip->iq[q].sw_tail); in zip_cmd_qbuf_free() 80 free_pages((u64)zip->iq[q].sw_tail, get_order(ZIP_CMD_QBUF_SIZE)); in zip_cmd_qbuf_free()
|
A D | zip_main.h | 87 u64 *sw_tail; member
|
A D | zip_main.c | 180 zip->iq[q].sw_tail = zip->iq[q].sw_head; in zip_init_hw() 198 zip->iq[q].sw_head, zip->iq[q].sw_tail, in zip_init_hw()
|
/linux/drivers/net/ethernet/aquantia/atlantic/ |
A D | aq_ring.h | 131 unsigned int sw_tail; member 165 return (((self->sw_tail >= self->sw_head)) ? in aq_ring_avail_dx() 166 (self->size - 1) - self->sw_tail + self->sw_head : in aq_ring_avail_dx() 167 self->sw_head - self->sw_tail - 1); in aq_ring_avail_dx()
|
A D | aq_ring.c | 224 self->sw_tail = 0; in aq_ring_init() 542 self->sw_tail = aq_ring_next_dx(self, self->sw_tail)) { in aq_ring_rx_fill() 543 buff = &self->buff_ring[self->sw_tail]; in aq_ring_rx_fill() 565 for (; self->sw_head != self->sw_tail; in aq_ring_rx_deinit()
|
A D | aq_vec.c | 79 sw_tail_old = ring[AQ_VEC_RX_ID].sw_tail; in aq_vec_poll()
|
A D | aq_nic.c | 592 dx = ring->sw_tail; in aq_nic_map_skb() 703 for (dx = ring->sw_tail; in aq_nic_map_skb()
|
A D | aq_ptp.c | 706 sw_tail_old = aq_ptp->ptp_rx.sw_tail; in aq_ptp_poll()
|
/linux/drivers/net/ethernet/aquantia/atlantic/hw_atl/ |
A D | hw_atl_b0.c | 656 hw_atl_reg_tx_dma_desc_tail_ptr_set(self, ring->sw_tail, ring->idx); in hw_atl_b0_hw_tx_ring_tail_update() 672 buff = &ring->buff_ring[ring->sw_tail]; in hw_atl_b0_hw_ring_tx_xmit() 676 txd = (struct hw_atl_txd_s *)&ring->dx_ring[ring->sw_tail * in hw_atl_b0_hw_ring_tx_xmit() 682 buff = &ring->buff_ring[ring->sw_tail]; in hw_atl_b0_hw_ring_tx_xmit() 741 ring->sw_tail = aq_ring_next_dx(ring, ring->sw_tail); in hw_atl_b0_hw_ring_tx_xmit() 823 for (; sw_tail_old != ring->sw_tail; in hw_atl_b0_hw_ring_rx_fill() 846 ring->sw_tail = aq_ring_next_dx(ring, ring->sw_tail)) { in hw_atl_b0_hw_ring_hwts_rx_fill() 849 &ring->dx_ring[ring->sw_tail * HW_ATL_B0_RXD_SIZE]; in hw_atl_b0_hw_ring_hwts_rx_fill() 857 hw_atl_reg_rx_dma_desc_tail_ptr_set(self, ring->sw_tail, ring->idx); in hw_atl_b0_hw_ring_hwts_rx_fill() 865 while (ring->hw_head != ring->sw_tail) { in hw_atl_b0_hw_ring_hwts_rx_receive() [all …]
|
A D | hw_atl_a0.c | 432 hw_atl_reg_tx_dma_desc_tail_ptr_set(self, ring->sw_tail, ring->idx); in hw_atl_a0_hw_tx_ring_tail_update() 448 buff = &ring->buff_ring[ring->sw_tail]; in hw_atl_a0_hw_ring_tx_xmit() 452 txd = (struct hw_atl_txd_s *)&ring->dx_ring[ring->sw_tail * in hw_atl_a0_hw_ring_tx_xmit() 458 buff = &ring->buff_ring[ring->sw_tail]; in hw_atl_a0_hw_ring_tx_xmit() 505 ring->sw_tail = aq_ring_next_dx(ring, ring->sw_tail); in hw_atl_a0_hw_ring_tx_xmit() 589 for (; sw_tail_old != ring->sw_tail; in hw_atl_a0_hw_ring_rx_fill() 626 for (; ring->hw_head != ring->sw_tail; in hw_atl_a0_hw_ring_rx_receive()
|
Completed in 23 milliseconds