/linux/drivers/md/ |
A D | dm-crypt.c | 173 unsigned int iv_size; member 316 memset(iv, 0, cc->iv_size); in crypt_iv_plain_gen() 325 memset(iv, 0, cc->iv_size); in crypt_iv_plain64_gen() 334 memset(iv, 0, cc->iv_size); in crypt_iv_plain64be_gen() 348 memset(iv, 0, cc->iv_size); in crypt_iv_essiv_gen() 691 if (cc->iv_size > 8) in crypt_iv_tcw_gen() 693 cc->iv_size - 8); in crypt_iv_tcw_gen() 1257 u8 *ptr = iv_of_dmreq(cc, dmreq) + cc->iv_size + cc->iv_size; in org_sector_of_dmreq() 2761 if (cc->iv_size) in crypt_ctr_ivmode() 2763 cc->iv_size = max(cc->iv_size, in crypt_ctr_ivmode() [all …]
|
/linux/drivers/crypto/keembay/ |
A D | keembay-ocs-aes-core.c | 259 int iv_size = crypto_skcipher_ivsize(tfm); in kmb_ocs_sk_validate_input() local 406 int iv_size = crypto_skcipher_ivsize(tfm); in kmb_ocs_sk_prepare_inplace() local 418 req->cryptlen - iv_size, iv_size, 0); in kmb_ocs_sk_prepare_inplace() 554 req->cryptlen, req->iv, iv_size); in kmb_ocs_sk_run() 574 req->cryptlen - iv_size, in kmb_ocs_sk_run() 575 iv_size, 0); in kmb_ocs_sk_run() 583 req->cryptlen - iv_size, in kmb_ocs_sk_run() 584 iv_size, 0); in kmb_ocs_sk_run() 603 int iv_size = crypto_aead_ivsize(tfm); in kmb_ocs_aead_validate_input() local 615 if (iv_size != GCM_AES_IV_SIZE) in kmb_ocs_aead_validate_input() [all …]
|
A D | ocs-aes.c | 601 const u8 *iv, u32 iv_size, in ocs_aes_validate_inputs() argument 664 if (!iv || iv_size != AES_BLOCK_SIZE) in ocs_aes_validate_inputs() 680 if (!iv || iv_size != AES_BLOCK_SIZE) in ocs_aes_validate_inputs() 696 if (!iv || iv_size != AES_BLOCK_SIZE) in ocs_aes_validate_inputs() 703 if (!iv || iv_size != GCM_AES_IV_SIZE) in ocs_aes_validate_inputs() 730 if (!iv || iv_size != AES_BLOCK_SIZE) in ocs_aes_validate_inputs() 804 u32 iv_size) in ocs_aes_op() argument 809 rc = ocs_aes_validate_inputs(src_dma_list, src_size, iv, iv_size, 0, 0, in ocs_aes_op()
|
A D | ocs-aes.h | 78 u32 iv_size);
|
/linux/include/net/ |
A D | tls.h | 235 u16 iv_size; member 518 prot->iv_size); in tls_advance_record_sn() 527 size_t pkt_len, iv_size = prot->iv_size; in tls_fill_prepend() local 532 pkt_len += iv_size; in tls_fill_prepend() 535 ctx->tx.iv + prot->salt_size, iv_size); in tls_fill_prepend()
|
/linux/drivers/s390/cio/ |
A D | airq.c | 116 static inline unsigned long iv_size(unsigned long bits) in iv_size() function 138 size = iv_size(bits); in airq_iv_create() 208 cio_dma_free(iv->vector, iv_size(iv->bits)); in airq_iv_release()
|
/linux/net/tls/ |
A D | tls_sw.c | 522 prot->iv_size + prot->salt_size); in tls_do_encryption() 1490 prot->iv_size); in decrypt_internal() 2115 cipher_overhead += prot->iv_size; in tls_read_size() 2399 iv_size = TLS_CIPHER_AES_GCM_128_IV_SIZE; in tls_set_sw_offload() 2416 iv_size = TLS_CIPHER_AES_GCM_256_IV_SIZE; in tls_set_sw_offload() 2433 iv_size = TLS_CIPHER_AES_CCM_128_IV_SIZE; in tls_set_sw_offload() 2451 iv_size = TLS_CIPHER_CHACHA20_POLY1305_IV_SIZE; in tls_set_sw_offload() 2468 iv_size = TLS_CIPHER_SM4_GCM_IV_SIZE; in tls_set_sw_offload() 2485 iv_size = TLS_CIPHER_SM4_CCM_IV_SIZE; in tls_set_sw_offload() 2523 prot->iv_size = iv_size; in tls_set_sw_offload() [all …]
|
A D | tls_device.c | 1013 u16 nonce_size, tag_size, iv_size, rec_seq_size, salt_size; in tls_set_device_offload() local 1051 iv_size = TLS_CIPHER_AES_GCM_128_IV_SIZE; in tls_set_device_offload() 1074 prot->iv_size = iv_size; in tls_set_device_offload() 1076 ctx->tx.iv = kmalloc(iv_size + TLS_CIPHER_AES_GCM_128_SALT_SIZE, in tls_set_device_offload() 1083 memcpy(ctx->tx.iv + TLS_CIPHER_AES_GCM_128_SALT_SIZE, iv, iv_size); in tls_set_device_offload()
|
/linux/drivers/crypto/hisilicon/sec2/ |
A D | sec_crypto.c | 1339 u32 iv_size = req->ctx->c_ctx.ivsize; in sec_update_iv() local 1359 sz = sg_pcopy_to_buffer(sgl, sg_nents(sgl), iv, iv_size, in sec_update_iv() 1360 cryptlen - iv_size); in sec_update_iv() 1361 if (unlikely(sz != iv_size)) in sec_update_iv() 1364 sz = cryptlen / iv_size; in sec_update_iv() 1365 if (cryptlen % iv_size) in sec_update_iv() 1367 ctr_iv_inc(iv, iv_size, sz); in sec_update_iv() 2121 .ivsize = iv_size,\ 2125 max_key_size, blk_size, iv_size) \ argument 2341 ctx_exit, blk_size, iv_size, max_authsize)\ argument [all …]
|
/linux/drivers/crypto/ |
A D | sa2ul.c | 107 u8 iv_size; member 584 if (cfg->iv_size) in sa_format_cmdl_gen() 585 auth_offset += cfg->iv_size; in sa_format_cmdl_gen() 600 if (cfg->iv_size) { in sa_format_cmdl_gen() 604 upd_info->enc_iv.size = cfg->iv_size; in sa_format_cmdl_gen() 607 SA_CMDL_HEADER_SIZE_BYTES + cfg->iv_size; in sa_format_cmdl_gen() 610 (SA_CTX_ENC_AUX2_OFFSET | (cfg->iv_size >> 3)); in sa_format_cmdl_gen() 611 total += SA_CMDL_HEADER_SIZE_BYTES + cfg->iv_size; in sa_format_cmdl_gen() 902 cfg.iv_size = crypto_skcipher_ivsize(tfm); in sa_cipher_setkey() 1466 cfg.iv_size = 0; in sa_sha_setup() [all …]
|
/linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/ |
A D | chcr_ktls.h | 50 u32 iv_size; member
|
A D | chcr_ktls.c | 89 tx_info->iv_size = TLS_CIPHER_AES_GCM_128_IV_SIZE; in chcr_ktls_save_keys() 1187 cipher_start = TLS_HEADER_SIZE + tx_info->iv_size + 1; in chcr_ktls_xmit_wr_complete() 1366 (!tls_rec_offset ? TLS_HEADER_SIZE + tx_info->iv_size : 0); in chcr_ktls_xmit_wr_short() 1423 memcpy(pos + tx_info->salt_size, &iv_record, tx_info->iv_size); in chcr_ktls_xmit_wr_short() 1424 *(__be32 *)(pos + tx_info->salt_size + tx_info->iv_size) = in chcr_ktls_xmit_wr_short() 1426 (TLS_HEADER_SIZE + tx_info->iv_size)) / AES_BLOCK_LEN) : 0)); in chcr_ktls_xmit_wr_short() 1806 if (tls_rec_offset + data_len <= (TLS_HEADER_SIZE + tx_info->iv_size)) { in chcr_short_record_handler() 1824 if (tls_rec_offset < (TLS_HEADER_SIZE + tx_info->iv_size)) { in chcr_short_record_handler() 1831 (TLS_HEADER_SIZE + tx_info->iv_size)) in chcr_short_record_handler()
|
/linux/net/tipc/ |
A D | crypto.c | 690 unsigned int iv_size, req_size; in tipc_aead_mem_alloc() local 694 iv_size = crypto_aead_ivsize(tfm); in tipc_aead_mem_alloc() 698 len += iv_size; in tipc_aead_mem_alloc() 711 *req = (struct aead_request *)PTR_ALIGN(*iv + iv_size, in tipc_aead_mem_alloc()
|