/linux/drivers/crypto/allwinner/sun4i-ss/ |
A D | sun4i-ss-cipher.c | 45 if (!areq->src || !areq->dst) { in sun4i_ss_opti_poll() 68 if (areq->iv) { in sun4i_ss_opti_poll() 83 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_ss_opti_poll() 111 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), in sun4i_ss_opti_poll() 140 scatterwalk_map_and_copy(areq->iv, areq->dst, areq->cryptlen - ivsize, in sun4i_ss_opti_poll() 167 areq->base.complete, areq->base.data); in sun4i_ss_cipher_poll_fallback() 169 areq->cryptlen, areq->iv); in sun4i_ss_cipher_poll_fallback() 216 if (!areq->src || !areq->dst) { in sun4i_ss_cipher_poll() 278 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_ss_cipher_poll() 332 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), in sun4i_ss_cipher_poll() [all …]
|
A D | sun4i-ss-hash.c | 91 sun4i_hash_init(areq); in sun4i_hash_import_md5() 134 sun4i_hash_init(areq); in sun4i_hash_import_sha1() 220 copied = sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in sun4i_hash() 247 if (end > areq->nbytes || areq->nbytes - end > 63) { in sun4i_hash() 249 end, areq->nbytes); in sun4i_hash() 255 if (areq->nbytes < 4) in sun4i_hash() 272 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_hash() 513 return sun4i_hash(areq); in sun4i_hash_final() 521 return sun4i_hash(areq); in sun4i_hash_update() 530 return sun4i_hash(areq); in sun4i_hash_finup() [all …]
|
A D | sun4i-ss.h | 200 int sun4i_hash_init(struct ahash_request *areq); 201 int sun4i_hash_update(struct ahash_request *areq); 202 int sun4i_hash_final(struct ahash_request *areq); 203 int sun4i_hash_finup(struct ahash_request *areq); 204 int sun4i_hash_digest(struct ahash_request *areq); 210 int sun4i_ss_cbc_aes_encrypt(struct skcipher_request *areq); 211 int sun4i_ss_cbc_aes_decrypt(struct skcipher_request *areq); 212 int sun4i_ss_ecb_aes_encrypt(struct skcipher_request *areq); 213 int sun4i_ss_ecb_aes_decrypt(struct skcipher_request *areq); 215 int sun4i_ss_cbc_des_encrypt(struct skcipher_request *areq); [all …]
|
/linux/drivers/crypto/gemini/ |
A D | sl3516-ce-cipher.c | 30 if (areq->cryptlen == 0 || areq->cryptlen % 16) { in sl3516_ce_need_fallback() 49 sg = areq->src; in sl3516_ce_need_fallback() 65 sg = areq->dst; in sl3516_ce_need_fallback() 113 areq->base.complete, areq->base.data); in sl3516_ce_cipher_fallback() 115 areq->cryptlen, areq->iv); in sl3516_ce_cipher_fallback() 149 if (areq->src == areq->dst) { in sl3516_ce_cipher() 150 nr_sgs = dma_map_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher() 246 if (areq->src == areq->dst) { in sl3516_ce_cipher() 247 dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher() 250 dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher() [all …]
|
/linux/drivers/crypto/allwinner/sun8i-ce/ |
A D | sun8i-ce-cipher.c | 34 if (areq->cryptlen == 0 || areq->cryptlen % 16) in sun8i_ce_cipher_need_fallback() 37 sg = areq->src; in sun8i_ce_cipher_need_fallback() 43 sg = areq->dst; in sun8i_ce_cipher_need_fallback() 68 areq->base.complete, areq->base.data); in sun8i_ce_cipher_fallback() 70 areq->cryptlen, areq->iv); in sun8i_ce_cipher_fallback() 180 if (areq->src == areq->dst) { in sun8i_ce_cipher_prepare() 242 if (areq->src == areq->dst) { in sun8i_ce_cipher_prepare() 243 dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src), in sun8i_ce_cipher_prepare() 247 dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src), in sun8i_ce_cipher_prepare() 249 dma_unmap_sg(ce->dev, areq->dst, sg_nents(areq->dst), in sun8i_ce_cipher_prepare() [all …]
|
A D | sun8i-ce-hash.c | 144 rctx->fallback_req.src = areq->src; in sun8i_ce_hash_update() 164 rctx->fallback_req.src = areq->src; in sun8i_ce_hash_finup() 203 if (areq->nbytes == 0) in sun8i_ce_hash_need_fallback() 208 sg = areq->src; in sun8i_ce_hash_need_fallback() 231 nr_sgs = sg_nents(areq->src); in sun8i_ce_hash_digest() 307 dev_dbg(ce->dev, "%s %s len=%d\n", __func__, crypto_tfm_alg_name(areq->base.tfm), areq->nbytes); in sun8i_ce_hash_run() 320 nr_sgs = dma_map_sg(ce->dev, areq->src, sg_nents(areq->src), DMA_TO_DEVICE); in sun8i_ce_hash_run() 327 len = areq->nbytes; in sun8i_ce_hash_run() 348 byte_count = areq->nbytes; in sun8i_ce_hash_run() 403 chan->timeout = areq->nbytes; in sun8i_ce_hash_run() [all …]
|
A D | sun8i-ce.h | 342 int sun8i_ce_enqueue(struct crypto_async_request *areq, u32 type); 350 int sun8i_ce_skdecrypt(struct skcipher_request *areq); 351 int sun8i_ce_skencrypt(struct skcipher_request *areq); 359 int sun8i_ce_hash_init(struct ahash_request *areq); 360 int sun8i_ce_hash_export(struct ahash_request *areq, void *out); 361 int sun8i_ce_hash_import(struct ahash_request *areq, const void *in); 362 int sun8i_ce_hash(struct ahash_request *areq); 363 int sun8i_ce_hash_final(struct ahash_request *areq); 364 int sun8i_ce_hash_update(struct ahash_request *areq); 365 int sun8i_ce_hash_finup(struct ahash_request *areq); [all …]
|
/linux/drivers/crypto/amlogic/ |
A D | amlogic-gxl-cipher.c | 73 areq->base.complete, areq->base.data); in meson_cipher_do_fallback() 74 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in meson_cipher_do_fallback() 75 areq->cryptlen, areq->iv); in meson_cipher_do_fallback() 108 areq->cryptlen, in meson_cipher() 177 if (areq->src == areq->dst) { in meson_cipher() 178 nr_sgs = dma_map_sg(mc->dev, areq->src, sg_nents(areq->src), in meson_cipher() 187 nr_sgs = dma_map_sg(mc->dev, areq->src, sg_nents(areq->src), in meson_cipher() 194 nr_sgd = dma_map_sg(mc->dev, areq->dst, sg_nents(areq->dst), in meson_cipher() 203 src_sg = areq->src; in meson_cipher() 238 if (areq->src == areq->dst) { in meson_cipher() [all …]
|
/linux/drivers/crypto/allwinner/sun8i-ss/ |
A D | sun8i-ss-cipher.c | 28 if (areq->cryptlen == 0 || areq->cryptlen % 16) in sun8i_ss_need_fallback() 31 if (sg_nents(areq->src) > 8 || sg_nents(areq->dst) > 8) in sun8i_ss_need_fallback() 34 sg = areq->src; in sun8i_ss_need_fallback() 44 sg = areq->dst; in sun8i_ss_need_fallback() 85 areq->base.complete, areq->base.data); in sun8i_ss_cipher_fallback() 87 areq->cryptlen, areq->iv); in sun8i_ss_cipher_fallback() 161 if (areq->src == areq->dst) { in sun8i_ss_cipher() 234 if (areq->src == areq->dst) { in sun8i_ss_cipher() 235 dma_unmap_sg(ss->dev, areq->src, sg_nents(areq->src), in sun8i_ss_cipher() 238 dma_unmap_sg(ss->dev, areq->src, sg_nents(areq->src), in sun8i_ss_cipher() [all …]
|
A D | sun8i-ss-hash.c | 144 rctx->fallback_req.src = areq->src; in sun8i_ss_hash_update() 164 rctx->fallback_req.src = areq->src; in sun8i_ss_hash_finup() 189 rctx->fallback_req.src = areq->src; in sun8i_ss_hash_digest_fb() 262 if (areq->nbytes == 0) in sun8i_ss_hash_need_fallback() 265 if (sg_nents(areq->src) > MAX_SG - 1) in sun8i_ss_hash_need_fallback() 267 sg = areq->src; in sun8i_ss_hash_need_fallback() 296 nr_sgs = sg_nents(areq->src); in sun8i_ss_hash_digest() 367 nr_sgs = dma_map_sg(ss->dev, areq->src, sg_nents(areq->src), DMA_TO_DEVICE); in sun8i_ss_hash_run() 381 len = areq->nbytes; in sun8i_ss_hash_run() 396 byte_count = areq->nbytes; in sun8i_ss_hash_run() [all …]
|
A D | sun8i-ss.h | 279 int sun8i_ss_enqueue(struct crypto_async_request *areq, u32 type); 287 int sun8i_ss_skdecrypt(struct skcipher_request *areq); 288 int sun8i_ss_skencrypt(struct skcipher_request *areq); 301 int sun8i_ss_hash_init(struct ahash_request *areq); 302 int sun8i_ss_hash_export(struct ahash_request *areq, void *out); 303 int sun8i_ss_hash_import(struct ahash_request *areq, const void *in); 304 int sun8i_ss_hash_final(struct ahash_request *areq); 305 int sun8i_ss_hash_update(struct ahash_request *areq); 306 int sun8i_ss_hash_finup(struct ahash_request *areq); 307 int sun8i_ss_hash_digest(struct ahash_request *areq);
|
/linux/drivers/crypto/cavium/nitrox/ |
A D | nitrox_aead.c | 202 areq->base.complete(&areq->base, err); in nitrox_aead_callback() 229 rctx->srclen = areq->assoclen + areq->cryptlen; in nitrox_aes_gcm_enc() 244 areq); in nitrox_aes_gcm_enc() 263 rctx->srclen = areq->cryptlen + areq->assoclen; in nitrox_aes_gcm_dec() 278 areq); in nitrox_aes_gcm_dec() 400 if (areq->assoclen != 16 && areq->assoclen != 20) in nitrox_rfc4106_set_aead_rctx_sglist() 406 sg = scatterwalk_ffwd(rctx->src + 1, areq->src, areq->assoclen); in nitrox_rfc4106_set_aead_rctx_sglist() 410 if (areq->src != areq->dst) { in nitrox_rfc4106_set_aead_rctx_sglist() 413 sg = scatterwalk_ffwd(rctx->dst + 1, areq->dst, areq->assoclen); in nitrox_rfc4106_set_aead_rctx_sglist() 437 areq->base.complete(&areq->base, err); in nitrox_rfc4106_callback() [all …]
|
/linux/drivers/crypto/inside-secure/ |
A D | safexcel_hash.c | 285 areq->nbytes = 0; in safexcel_handle_req_result() 343 sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in safexcel_ahash_send_req() 425 areq->nbytes), in safexcel_ahash_send_req() 679 sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in safexcel_ahash_cache() 681 areq->nbytes, 0); in safexcel_ahash_cache() 748 if (!areq->nbytes) in safexcel_ahash_update() 799 memcpy(areq->result, in safexcel_ahash_final() 827 crypto_cipher_encrypt_one(ctx->kaes, areq->result, areq->result); in safexcel_ahash_final() 1157 if (!areq) { in __safexcel_hmac_setkey() 1956 return safexcel_crc32_init(areq) ?: safexcel_ahash_finup(areq); in safexcel_crc32_digest() [all …]
|
/linux/crypto/ |
A D | algif_skcipher.c | 73 if (IS_ERR(areq)) in _skcipher_recvmsg() 74 return PTR_ERR(areq); in _skcipher_recvmsg() 93 if (!areq->tsgl_entries) in _skcipher_recvmsg() 94 areq->tsgl_entries = 1; in _skcipher_recvmsg() 95 areq->tsgl = sock_kmalloc(sk, array_size(sizeof(*areq->tsgl), in _skcipher_recvmsg() 96 areq->tsgl_entries), in _skcipher_recvmsg() 98 if (!areq->tsgl) { in _skcipher_recvmsg() 102 sg_init_table(areq->tsgl, areq->tsgl_entries); in _skcipher_recvmsg() 107 skcipher_request_set_crypt(&areq->cra_u.skcipher_req, areq->tsgl, in _skcipher_recvmsg() 113 areq->iocb = msg->msg_iocb; in _skcipher_recvmsg() [all …]
|
A D | algif_aead.c | 155 if (IS_ERR(areq)) in _aead_recvmsg() 156 return PTR_ERR(areq); in _aead_recvmsg() 252 if (!areq->tsgl_entries) in _aead_recvmsg() 253 areq->tsgl_entries = 1; in _aead_recvmsg() 254 areq->tsgl = sock_kmalloc(sk, array_size(sizeof(*areq->tsgl), in _aead_recvmsg() 255 areq->tsgl_entries), in _aead_recvmsg() 257 if (!areq->tsgl) { in _aead_recvmsg() 261 sg_init_table(areq->tsgl, areq->tsgl_entries); in _aead_recvmsg() 273 areq->tsgl); in _aead_recvmsg() 276 rsgl_src = areq->tsgl; in _aead_recvmsg() [all …]
|
A D | af_alg.c | 680 tsgl = areq->tsgl; in af_alg_free_areq_sgls() 1050 sock_kfree_s(sk, areq, areq->areqlen); in af_alg_free_resources() 1122 if (unlikely(!areq)) in af_alg_alloc_areq() 1125 areq->areqlen = areqlen; in af_alg_alloc_areq() 1126 areq->sk = sk; in af_alg_alloc_areq() 1127 areq->last_rsgl = NULL; in af_alg_alloc_areq() 1129 areq->tsgl = NULL; in af_alg_alloc_areq() 1130 areq->tsgl_entries = 0; in af_alg_alloc_areq() 1132 return areq; in af_alg_alloc_areq() 1187 if (areq->last_rsgl) in af_alg_get_rsgl() [all …]
|
A D | chacha20poly1305.c | 120 async_done_continue(areq->data, err, poly_verify_tag); in chacha_decrypt_done() 166 async_done_continue(areq->data, err, poly_tail_continue); in poly_tail_done() 196 async_done_continue(areq->data, err, poly_tail); in poly_cipherpad_done() 225 async_done_continue(areq->data, err, poly_cipherpad); in poly_cipher_done() 255 async_done_continue(areq->data, err, poly_cipher); in poly_adpad_done() 282 static void poly_ad_done(struct crypto_async_request *areq, int err) in poly_ad_done() argument 284 async_done_continue(areq->data, err, poly_adpad); in poly_ad_done() 308 async_done_continue(areq->data, err, poly_ad); in poly_setkey_done() 334 async_done_continue(areq->data, err, poly_setkey); in poly_init_done() 357 async_done_continue(areq->data, err, poly_init); in poly_genkey_done() [all …]
|
A D | ahash.c | 287 struct ahash_request *areq = req->data; in ahash_op_unaligned_done() local 290 ahash_notify_einprogress(areq); in ahash_op_unaligned_done() 304 ahash_restore_req(areq, err); in ahash_op_unaligned_done() 307 areq->base.complete(&areq->base, err); in ahash_op_unaligned_done() 387 struct ahash_request *areq = req->data; in ahash_def_finup_done2() local 392 ahash_restore_req(areq, err); in ahash_def_finup_done2() 394 areq->base.complete(&areq->base, err); in ahash_def_finup_done2() 415 struct ahash_request *areq = req->data; in ahash_def_finup_done1() local 418 ahash_notify_einprogress(areq); in ahash_def_finup_done1() 425 if (areq->priv) in ahash_def_finup_done1() [all …]
|
/linux/drivers/crypto/ |
A D | talitos.c | 1226 (areq->src == areq->dst) ? in ipsec_esp() 1264 if (areq->src != areq->dst) { in ipsec_esp() 1538 talitos_sg_unmap(dev, edesc, areq->src, areq->dst, areq->cryptlen, 0); in common_nonsnoop_unmap() 1563 areq->base.complete(&areq->base, err); in skcipher_done() 1599 (areq->src == areq->dst) ? in common_nonsnoop() 1610 if (areq->src != areq->dst) { in common_nonsnoop() 1647 areq->iv, 0, areq->cryptlen, 0, ivsize, 0, in skcipher_edesc_alloc() 1762 areq->base.complete(&areq->base, err); in ahash_done() 2099 return ahash_process_req(areq, areq->nbytes); in ahash_update() 2117 return ahash_process_req(areq, areq->nbytes); in ahash_finup() [all …]
|
A D | atmel-i2c.h | 163 void (*cbk)(struct atmel_i2c_work_data *work_data, void *areq, 165 void *areq; member 174 void *areq, int status), 175 void *areq);
|
/linux/drivers/crypto/stm32/ |
A D | stm32-cryp.c | 142 struct aead_request *areq; member 509 if (cryp->areq->assoclen) in stm32_cryp_ccm_init() 609 if (cryp->areq->assoclen) { in stm32_cryp_hw_init() 691 void *areq); 707 void *areq); 924 if (!req && !areq) in stm32_cryp_prepare_req() 946 cryp->areq = NULL; in stm32_cryp_prepare_req() 971 cryp->areq = areq; in stm32_cryp_prepare_req() 974 cryp->total_in = areq->assoclen + areq->cryptlen; in stm32_cryp_prepare_req() 1022 void *areq) in stm32_cryp_prepare_cipher_req() argument [all …]
|
/linux/drivers/crypto/xilinx/ |
A D | zynqmp-aes-gcm.c | 207 struct aead_request *areq = in zynqmp_handle_aes_req() local 211 struct zynqmp_aead_req_ctx *rq_ctx = aead_request_ctx(areq); in zynqmp_handle_aes_req() 216 need_fallback = zynqmp_fallback_check(tfm_ctx, areq); in zynqmp_handle_aes_req() 221 aead_request_set_callback(subreq, areq->base.flags, in zynqmp_handle_aes_req() 223 aead_request_set_crypt(subreq, areq->src, areq->dst, in zynqmp_handle_aes_req() 224 areq->cryptlen, areq->iv); in zynqmp_handle_aes_req() 225 aead_request_set_ad(subreq, areq->assoclen); in zynqmp_handle_aes_req() 231 err = zynqmp_aes_aead_cipher(areq); in zynqmp_handle_aes_req() 234 crypto_finalize_aead_request(engine, areq, err); in zynqmp_handle_aes_req()
|
/linux/drivers/crypto/axis/ |
A D | artpec6_crypto.c | 1439 sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in artpec6_crypto_prepare_hash() 1676 req_ctx = skcipher_request_ctx(areq); in artpec6_crypto_prepare_crypto() 1799 areq->cryptlen; in artpec6_crypto_prepare_crypto() 1904 input_length = areq->cryptlen; in artpec6_crypto_prepare_aead() 1910 __cpu_to_be64(8*areq->assoclen); in artpec6_crypto_prepare_aead() 1930 count = areq->assoclen; in artpec6_crypto_prepare_aead() 1935 if (!IS_ALIGNED(areq->assoclen, 16)) { in artpec6_crypto_prepare_aead() 1968 size_t output_len = areq->cryptlen; in artpec6_crypto_prepare_aead() 2193 sg_pcopy_to_buffer(areq->src, in artpec6_crypto_complete_aead() 2194 sg_nents(areq->src), in artpec6_crypto_complete_aead() [all …]
|
/linux/drivers/crypto/hisilicon/hpre/ |
A D | hpre_crypto.c | 137 } areq; member 425 areq = req->areq.dh; in hpre_dh_cb() 432 hpre_hw_data_clr_all(ctx, req, areq->dst, areq->src); in hpre_dh_cb() 451 areq = req->areq.rsa; in hpre_rsa_cb() 453 hpre_hw_data_clr_all(ctx, req, areq->dst, areq->src); in hpre_rsa_cb() 530 h_req->areq.dh = kreq; in hpre_msg_request_set() 1454 areq = req->areq.ecdh; in hpre_ecdh_cb() 1461 p = sg_virt(areq->dst); in hpre_ecdh_cb() 1465 hpre_ecdh_hw_data_clr_all(ctx, req, areq->dst, areq->src); in hpre_ecdh_cb() 1762 areq = req->areq.curve25519; in hpre_curve25519_cb() [all …]
|
/linux/drivers/crypto/marvell/octeontx2/ |
A D | otx2_cptvf_reqmgr.c | 82 pentry->areq = NULL; in free_pentry() 245 gfp = (req->areq->flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? GFP_KERNEL : in process_request() 291 pentry->areq = req->areq; in process_request() 440 struct crypto_async_request *areq; in process_pending_queue() local 501 areq = resume_pentry->areq; in process_pending_queue() 510 callback(-EINPROGRESS, areq, info); in process_pending_queue() 516 areq = pentry->areq; in process_pending_queue() 529 callback(res_code, areq, info); in process_pending_queue()
|