/linux/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
A D | ktls_tx.c | 55 struct tls_offload_context_tx *tx_ctx; member 111 priv_tx->tx_ctx = tls_offload_ctx_tx(tls_ctx); in mlx5e_ktls_add_tx() 235 struct tls_offload_context_tx *tx_ctx = priv_tx->tx_ctx; in tx_sync_info_get() local 242 spin_lock_irqsave(&tx_ctx->lock, flags); in tx_sync_info_get() 243 record = tls_get_record(tx_ctx, tcp_seq, &info->rcd_sn); in tx_sync_info_get() 282 spin_unlock_irqrestore(&tx_ctx->lock, flags); in tx_sync_info_get()
|
A D | tls.c | 130 struct mlx5e_tls_offload_context_tx *tx_ctx = in mlx5e_tls_add() local 133 tx_ctx->swid = htonl(swid); in mlx5e_tls_add() 134 tx_ctx->expected_seq = start_offload_tcp_sn; in mlx5e_tls_add()
|
/linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/ |
A D | chcr_ktls.c | 364 struct chcr_ktls_ofld_ctx_tx *tx_ctx = in chcr_ktls_dev_del() local 399 tx_ctx->chcr_info = NULL; in chcr_ktls_dev_del() 420 struct chcr_ktls_ofld_ctx_tx *tx_ctx; in chcr_ktls_dev_add() local 443 if (tx_ctx->chcr_info) in chcr_ktls_dev_add() 569 tx_ctx->chcr_info = tx_info; in chcr_ktls_dev_add() 650 struct chcr_ktls_ofld_ctx_tx *tx_ctx; in chcr_ktls_cpl_act_open_rpl() local 1929 struct chcr_ktls_ofld_ctx_tx *tx_ctx; in chcr_ktls_xmit() local 1952 tx_info = tx_ctx->chcr_info; in chcr_ktls_xmit() 2184 struct chcr_ktls_ofld_ctx_tx *tx_ctx; in ch_ktls_reset_all_conn() local 2189 tx_info = tx_ctx->chcr_info; in ch_ktls_reset_all_conn() [all …]
|
/linux/drivers/infiniband/sw/siw/ |
A D | siw_qp.c | 134 qp->tx_ctx.tx_suspend = 1; in siw_qp_llp_close() 229 struct siw_iwarp_tx *c_tx = &qp->tx_ctx; in siw_qp_enable_crc() 583 if (qp->tx_ctx.mpa_crc_hd) { in siw_send_terminate() 584 crypto_shash_init(qp->tx_ctx.mpa_crc_hd); in siw_send_terminate() 585 if (crypto_shash_update(qp->tx_ctx.mpa_crc_hd, in siw_send_terminate() 591 if (crypto_shash_update(qp->tx_ctx.mpa_crc_hd, in siw_send_terminate() 660 qp->tx_ctx.ddp_msn[RDMAP_UNTAGGED_QN_SEND] = 0; in siw_qp_nextstate_from_idle() 661 qp->tx_ctx.ddp_msn[RDMAP_UNTAGGED_QN_RDMA_READ] = 0; in siw_qp_nextstate_from_idle() 662 qp->tx_ctx.ddp_msn[RDMAP_UNTAGGED_QN_TERMINATE] = 0; in siw_qp_nextstate_from_idle() 927 qp->tx_ctx.orq_fence = 1; in siw_activate_tx_from_sq() [all …]
|
A D | siw_qp_tx.c | 705 struct siw_iwarp_tx *c_tx = &qp->tx_ctx; in siw_prepare_fpdu() 793 struct siw_iwarp_tx *c_tx = &qp->tx_ctx; in siw_qp_sq_proc_tx() 795 int rv = 0, burst_len = qp->tx_ctx.burst; in siw_qp_sq_proc_tx() 915 qp->tx_ctx.burst = burst_len; in siw_qp_sq_proc_tx() 1035 if (unlikely(qp->tx_ctx.tx_suspend)) { in siw_qp_sq_process() 1093 qp->tx_ctx.ctrl_sent, qp->tx_ctx.ctrl_len, in siw_qp_sq_process() 1094 qp->tx_ctx.bytes_unsent); in siw_qp_sq_process() 1130 if (!qp->tx_ctx.tx_suspend) in siw_qp_sq_process() 1175 !qp->tx_ctx.tx_suspend)) { in siw_sq_resume() 1183 if (!qp->tx_ctx.tx_suspend) in siw_sq_resume()
|
A D | siw.h | 433 struct siw_iwarp_tx tx_ctx; /* Transmit context */ member 472 #define tx_qp(tx) container_of(tx, struct siw_qp, tx_ctx) 473 #define tx_wqe(qp) (&(qp)->tx_ctx.wqe_active)
|
A D | siw_verbs.c | 427 qp->tx_ctx.gso_seg_limit = 1; in siw_create_qp() 428 qp->tx_ctx.zcopy_tx = zcopy_tx; in siw_create_qp() 569 qp->tx_ctx.tx_suspend = 1; in siw_verbs_modify_qp() 618 kfree(qp->tx_ctx.mpa_crc_hd); in siw_destroy_qp() 928 qp->tx_ctx.in_syscall = 1; in siw_post_send() 930 if (siw_qp_sq_process(qp) != 0 && !(qp->tx_ctx.tx_suspend)) in siw_post_send() 933 qp->tx_ctx.in_syscall = 0; in siw_post_send()
|
A D | siw_cm.c | 377 qp->tx_ctx.tx_suspend = 1; in siw_qp_cm_drop() 755 qp->tx_ctx.gso_seg_limit = 0; in siw_proc_mpareply() 1291 cep->qp->tx_ctx.tx_suspend = 1; in siw_cm_llp_state_change() 1577 qp->tx_ctx.gso_seg_limit = 0; in siw_accept()
|
A D | siw_qp_rx.c | 1161 if (qp->tx_ctx.orq_fence) { in siw_check_tx_fence() 1180 qp->tx_ctx.orq_fence = 0; in siw_check_tx_fence() 1184 qp->tx_ctx.orq_fence = 0; in siw_check_tx_fence()
|
/linux/net/wireless/ |
A D | lib80211_crypt_wep.c | 35 struct arc4_ctx tx_ctx; member 138 arc4_setkey(&wep->tx_ctx, key, klen); in lib80211_wep_encrypt() 139 arc4_crypt(&wep->tx_ctx, pos, pos, len + 4); in lib80211_wep_encrypt()
|
/linux/net/tipc/ |
A D | crypto.c | 738 struct tipc_crypto_tx_ctx *tx_ctx; in tipc_aead_encrypt() local 784 ctx = tipc_aead_mem_alloc(tfm, sizeof(*tx_ctx), &iv, &req, &sg, nsg); in tipc_aead_encrypt() 820 tx_ctx = (struct tipc_crypto_tx_ctx *)ctx; in tipc_aead_encrypt() 821 tx_ctx->aead = aead; in tipc_aead_encrypt() 822 tx_ctx->bearer = b; in tipc_aead_encrypt() 823 memcpy(&tx_ctx->dst, dst, sizeof(*dst)); in tipc_aead_encrypt() 847 struct tipc_crypto_tx_ctx *tx_ctx = TIPC_SKB_CB(skb)->crypto_ctx; in tipc_aead_encrypt_done() local 848 struct tipc_bearer *b = tx_ctx->bearer; in tipc_aead_encrypt_done() 849 struct tipc_aead *aead = tx_ctx->aead; in tipc_aead_encrypt_done() 858 b->media->send_msg(net, skb, b, &tx_ctx->dst); in tipc_aead_encrypt_done() [all …]
|
/linux/drivers/net/ethernet/intel/i40e/ |
A D | i40e_virtchnl_pf.c | 608 struct i40e_hmc_obj_txq tx_ctx; in i40e_config_vsi_tx_queue() local 626 memset(&tx_ctx, 0, sizeof(struct i40e_hmc_obj_txq)); in i40e_config_vsi_tx_queue() 629 tx_ctx.base = info->dma_ring_addr / 128; in i40e_config_vsi_tx_queue() 630 tx_ctx.qlen = info->ring_len; in i40e_config_vsi_tx_queue() 631 tx_ctx.rdylist = le16_to_cpu(vsi->info.qs_handle[0]); in i40e_config_vsi_tx_queue() 632 tx_ctx.rdylist_act = 0; in i40e_config_vsi_tx_queue() 633 tx_ctx.head_wb_ena = info->headwb_enabled; in i40e_config_vsi_tx_queue() 634 tx_ctx.head_wb_addr = info->dma_headwb_addr; in i40e_config_vsi_tx_queue() 647 ret = i40e_set_lan_tx_queue_context(hw, pf_queue_id, &tx_ctx); in i40e_config_vsi_tx_queue()
|
A D | i40e_main.c | 3201 struct i40e_hmc_obj_txq tx_ctx; in i40e_configure_tx_ring() local 3220 memset(&tx_ctx, 0, sizeof(tx_ctx)); in i40e_configure_tx_ring() 3222 tx_ctx.new_context = 1; in i40e_configure_tx_ring() 3223 tx_ctx.base = (ring->dma / 128); in i40e_configure_tx_ring() 3224 tx_ctx.qlen = ring->count; in i40e_configure_tx_ring() 3227 tx_ctx.timesync_ena = !!(vsi->back->flags & I40E_FLAG_PTP); in i40e_configure_tx_ring() 3230 tx_ctx.head_wb_ena = 1; in i40e_configure_tx_ring() 3231 tx_ctx.head_wb_addr = ring->dma + in i40e_configure_tx_ring() 3246 tx_ctx.rdylist = in i40e_configure_tx_ring() 3252 tx_ctx.rdylist_act = 0; in i40e_configure_tx_ring() [all …]
|
/linux/net/tls/ |
A D | tls_sw.c | 2306 struct tls_sw_context_tx *tx_ctx = tls_sw_ctx_tx(ctx); in tls_sw_write_space() local 2309 if (is_tx_ready(tx_ctx) && in tls_sw_write_space() 2310 !test_and_set_bit(BIT_TX_SCHEDULED, &tx_ctx->tx_bitmask)) in tls_sw_write_space() 2311 schedule_delayed_work(&tx_ctx->tx_work.work, 0); in tls_sw_write_space()
|