/linux/lib/crypto/ |
A D | chacha20poly1305.c | 76 poly1305_update(&poly1305_state, dst, src_len); in __chacha20poly1305_encrypt() 77 if (src_len & 0xf) in __chacha20poly1305_encrypt() 81 b.lens[1] = cpu_to_le64(src_len); in __chacha20poly1305_encrypt() 84 poly1305_final(&poly1305_state, dst + src_len); in __chacha20poly1305_encrypt() 138 if (unlikely(src_len < POLY1305_DIGEST_SIZE)) in __chacha20poly1305_decrypt() 148 dst_len = src_len - POLY1305_DIGEST_SIZE; in __chacha20poly1305_decrypt() 209 const size_t src_len, in chacha20poly1305_crypt_sg_inplace() argument 236 if (WARN_ON(src_len > INT_MAX)) in chacha20poly1305_crypt_sg_inplace() 297 if (src_len & 0xf) in chacha20poly1305_crypt_sg_inplace() 301 b.lens[1] = cpu_to_le64(src_len); in chacha20poly1305_crypt_sg_inplace() [all …]
|
/linux/arch/s390/include/asm/ |
A D | cpacf.h | 255 s.odd = (unsigned long)src_len; in cpacf_km() 266 return src_len - s.odd; in cpacf_km() 287 s.odd = (unsigned long)src_len; in cpacf_kmc() 298 return src_len - s.odd; in cpacf_kmc() 315 s.odd = (unsigned long)src_len; in cpacf_kimd() 340 s.odd = (unsigned long)src_len; in cpacf_klmd() 368 s.odd = (unsigned long)src_len; in cpacf_kmac() 379 return src_len - s.odd; in cpacf_kmac() 401 s.odd = (unsigned long)src_len; in cpacf_kmctr() 414 return src_len - s.odd; in cpacf_kmctr() [all …]
|
/linux/lib/ |
A D | decompress_unlzo.c | 105 u32 src_len, dst_len; in unlzo() local 207 src_len = get_unaligned_be32(in_buf); in unlzo() 211 if (src_len <= 0 || src_len > dst_len) { in unlzo() 217 if (fill && in_len < src_len) { in unlzo() 218 skip = fill(in_buf + in_len, src_len - in_len); in unlzo() 222 if (in_len < src_len) { in unlzo() 231 if (unlikely(dst_len == src_len)) in unlzo() 232 memcpy(out_buf, in_buf, src_len); in unlzo() 248 *posp += src_len + 12; in unlzo() 250 in_buf += src_len; in unlzo() [all …]
|
/linux/net/decnet/ |
A D | dn_rules.c | 44 unsigned char src_len; member 149 if (frh->src_len) in dn_fib_rule_configure() 155 r->src_len = frh->src_len; in dn_fib_rule_configure() 156 r->srcmask = dnet_make_mask(r->src_len); in dn_fib_rule_configure() 169 if (frh->src_len && (r->src_len != frh->src_len)) in dn_fib_rule_compare() 175 if (frh->src_len && (r->src != nla_get_le16(tb[FRA_SRC]))) in dn_fib_rule_compare() 208 frh->src_len = r->src_len; in dn_fib_rule_fill() 213 (r->src_len && in dn_fib_rule_fill()
|
/linux/include/crypto/ |
A D | akcipher.h | 37 unsigned int src_len; member 251 unsigned int src_len, in akcipher_request_set_crypt() argument 256 req->src_len = src_len; in akcipher_request_set_crypt() 292 unsigned int src_len = req->src_len; in crypto_akcipher_encrypt() local 297 crypto_stats_akcipher_encrypt(src_len, ret, calg); in crypto_akcipher_encrypt() 316 unsigned int src_len = req->src_len; in crypto_akcipher_decrypt() local 321 crypto_stats_akcipher_decrypt(src_len, ret, calg); in crypto_akcipher_decrypt()
|
A D | chacha20poly1305.h | 18 void chacha20poly1305_encrypt(u8 *dst, const u8 *src, const size_t src_len, 24 chacha20poly1305_decrypt(u8 *dst, const u8 *src, const size_t src_len, 28 void xchacha20poly1305_encrypt(u8 *dst, const u8 *src, const size_t src_len, 34 u8 *dst, const u8 *src, const size_t src_len, const u8 *ad, 38 bool chacha20poly1305_encrypt_sg_inplace(struct scatterlist *src, size_t src_len, 43 bool chacha20poly1305_decrypt_sg_inplace(struct scatterlist *src, size_t src_len,
|
/linux/net/smc/ |
A D | smc_tx.c | 345 sge[srcchunk].length = src_len; in smcr_tx_rdma_writes() 348 src_off += src_len; in smcr_tx_rdma_writes() 355 src_len = dst_len - src_len; /* remainder */ in smcr_tx_rdma_writes() 356 src_len_sum += src_len; in smcr_tx_rdma_writes() 370 src_len_sum = src_len; in smcr_tx_rdma_writes() 392 dst_off += src_len; in smcd_tx_rdma_writes() 393 src_off += src_len; in smcd_tx_rdma_writes() 400 src_len = dst_len - src_len; /* remainder */ in smcd_tx_rdma_writes() 401 src_len_sum += src_len; in smcd_tx_rdma_writes() 410 src_len_sum = src_len; in smcd_tx_rdma_writes() [all …]
|
/linux/net/ipv4/ |
A D | fib_rules.c | 37 u8 src_len; member 52 if (r->dst_len || r->src_len || r->tos) in fib4_rule_matchall() 257 if (frh->src_len) in fib4_rule_configure() 274 rule4->src_len = frh->src_len; in fib4_rule_configure() 275 rule4->srcmask = inet_make_mask(rule4->src_len); in fib4_rule_configure() 315 if (frh->src_len && (rule4->src_len != frh->src_len)) in fib4_rule_compare() 329 if (frh->src_len && (rule4->src != nla_get_in_addr(tb[FRA_SRC]))) in fib4_rule_compare() 344 frh->src_len = rule4->src_len; in fib4_rule_fill() 349 (rule4->src_len && in fib4_rule_fill()
|
/linux/include/linux/ |
A D | ccp.h | 183 u64 src_len; /* In bytes */ member 243 u64 src_len; /* In bytes */ member 292 u64 src_len; /* In bytes */ member 354 u64 src_len; /* In bytes */ member 382 u32 src_len; /* In bytes */ member 441 u64 src_len; /* In bytes */ member 470 u64 src_len; /* In bytes */ member
|
A D | lzo.h | 24 int lzo1x_1_compress(const unsigned char *src, size_t src_len, 28 int lzorle1x_1_compress(const unsigned char *src, size_t src_len, 32 int lzo1x_decompress_safe(const unsigned char *src, size_t src_len,
|
A D | crypto.h | 518 void crypto_stats_akcipher_encrypt(unsigned int src_len, int ret, struct crypto_alg *alg); 519 void crypto_stats_akcipher_decrypt(unsigned int src_len, int ret, struct crypto_alg *alg); 544 static inline void crypto_stats_akcipher_encrypt(unsigned int src_len, int ret, struct crypto_alg *… in crypto_stats_akcipher_encrypt() argument 546 static inline void crypto_stats_akcipher_decrypt(unsigned int src_len, int ret, struct crypto_alg *… in crypto_stats_akcipher_decrypt() argument
|
/linux/drivers/base/regmap/ |
A D | regcache-lzo.c | 21 size_t src_len; member 47 ret = lzo1x_1_compress(lzo_ctx->src, lzo_ctx->src_len, in regcache_lzo_compress() 61 ret = lzo1x_decompress_safe(lzo_ctx->src, lzo_ctx->src_len, in regcache_lzo_decompress() 181 lzo_blocks[i]->src_len = end - p; in regcache_lzo_init() 183 lzo_blocks[i]->src_len = blksize; in regcache_lzo_init() 189 lzo_blocks[i]->src_len; in regcache_lzo_init() 249 lzo_block->src_len = lzo_block->dst_len; in regcache_lzo_read() 286 lzo_block->src_len = lzo_block->dst_len; in regcache_lzo_write() 303 lzo_block->src_len = lzo_block->dst_len; in regcache_lzo_write()
|
/linux/drivers/gpu/drm/ |
A D | drm_format_helper.c | 167 size_t src_len = width * sizeof(u32); in drm_fb_xrgb8888_to_rgb332() local 172 sbuf = kmalloc(src_len, GFP_KERNEL); in drm_fb_xrgb8888_to_rgb332() 178 memcpy(sbuf, src, src_len); in drm_fb_xrgb8888_to_rgb332() 225 size_t src_len = linepixels * sizeof(u32); in drm_fb_xrgb8888_to_rgb565() local 234 sbuf = kmalloc(src_len, GFP_KERNEL); in drm_fb_xrgb8888_to_rgb565() 240 memcpy(sbuf, vaddr, src_len); in drm_fb_xrgb8888_to_rgb565() 320 size_t src_len = width * sizeof(u32); in drm_fb_xrgb8888_to_rgb888() local 325 sbuf = kmalloc(src_len, GFP_KERNEL); in drm_fb_xrgb8888_to_rgb888() 331 memcpy(sbuf, src, src_len); in drm_fb_xrgb8888_to_rgb888()
|
/linux/crypto/ |
A D | rsa-pkcs1pad.c | 240 if (req->src_len > ctx->key_size - 11) in pkcs1pad_encrypt() 253 ps_end = ctx->key_size - req->src_len - 2; in pkcs1pad_encrypt() 260 ctx->key_size - 1 - req->src_len, req->src); in pkcs1pad_encrypt() 353 if (!ctx->key_size || req->src_len != ctx->key_size) in pkcs1pad_decrypt() 369 req_ctx->out_sg, req->src_len, in pkcs1pad_decrypt() 396 if (req->src_len + digest_size > ctx->key_size - 11) in pkcs1pad_sign() 409 ps_end = ctx->key_size - digest_size - req->src_len - 2; in pkcs1pad_sign() 419 ctx->key_size - 1 - req->src_len, req->src); in pkcs1pad_sign() 496 req->src_len + req->dst_len), in pkcs1pad_verify_complete() 541 !ctx->key_size || req->src_len < ctx->key_size) in pkcs1pad_verify() [all …]
|
A D | ecrdsa.c | 99 req->dst_len * 2 != req->src_len || in ecrdsa_verify() 100 WARN_ON(req->src_len > sizeof(sig)) || in ecrdsa_verify() 104 sg_copy_to_buffer(req->src, sg_nents_for_len(req->src, req->src_len), in ecrdsa_verify() 105 sig, req->src_len); in ecrdsa_verify() 108 req->src_len + req->dst_len), in ecrdsa_verify() 109 digest, req->dst_len, req->src_len); in ecrdsa_verify()
|
A D | ecdsa.c | 154 buffer = kmalloc(req->src_len + req->dst_len, GFP_KERNEL); in ecdsa_verify() 159 sg_nents_for_len(req->src, req->src_len + req->dst_len), in ecdsa_verify() 160 buffer, req->src_len + req->dst_len, 0); in ecdsa_verify() 163 buffer, req->src_len); in ecdsa_verify() 172 memcpy(&rawhash[diff], buffer + req->src_len, req->dst_len); in ecdsa_verify() 175 memcpy(&rawhash, buffer + req->src_len, keylen); in ecdsa_verify()
|
A D | sm2.c | 360 buffer = kmalloc(req->src_len + req->dst_len, GFP_KERNEL); in sm2_verify() 365 sg_nents_for_len(req->src, req->src_len + req->dst_len), in sm2_verify() 366 buffer, req->src_len + req->dst_len, 0); in sm2_verify() 371 buffer, req->src_len); in sm2_verify() 376 hash = mpi_read_raw_data(buffer + req->src_len, req->dst_len); in sm2_verify()
|
/linux/drivers/staging/r8188eu/os_dep/ |
A D | osdep_service.c | 184 void rtw_buf_update(u8 **buf, u32 *buf_len, u8 *src, u32 src_len) in rtw_buf_update() argument 193 if (!src || !src_len) in rtw_buf_update() 197 dup = kmalloc(src_len, GFP_ATOMIC); in rtw_buf_update() 199 dup_len = src_len; in rtw_buf_update()
|
/linux/drivers/staging/rtl8723bs/os_dep/ |
A D | osdep_service.c | 173 void rtw_buf_update(u8 **buf, u32 *buf_len, u8 *src, u32 src_len) in rtw_buf_update() argument 182 if (!src || !src_len) in rtw_buf_update() 186 dup = rtw_malloc(src_len); in rtw_buf_update() 188 dup_len = src_len; in rtw_buf_update()
|
/linux/tools/bpf/bpftool/ |
A D | tracelog.c | 42 size_t src_len; in find_tracefs_mnt_single() local 47 src_len = strlen(mntpt); in find_tracefs_mnt_single() 48 if (src_len + 1 >= PATH_MAX) { in find_tracefs_mnt_single()
|
/linux/drivers/crypto/ccp/ |
A D | ccp-dmaengine.c | 366 unsigned int src_offset, src_len; in ccp_create_desc() local 384 src_len = sg_dma_len(src_sg); in ccp_create_desc() 391 if (!src_len) { in ccp_create_desc() 400 src_len = sg_dma_len(src_sg); in ccp_create_desc() 419 len = min(dst_len, src_len); in ccp_create_desc() 435 ccp_pt->src_len = len; in ccp_create_desc() 445 &ccp_pt->dst_dma, ccp_pt->src_len); in ccp_create_desc() 449 src_len -= len; in ccp_create_desc()
|
A D | ccp-ops.c | 487 if (aes->src_len & (AES_BLOCK_SIZE - 1)) in ccp_run_aes_cmac_cmd() 685 ilen = aes->src_len; in ccp_run_aes_gcm_cmd() 689 ilen = aes->src_len - authsize; in ccp_run_aes_gcm_cmd() 930 (aes->src_len & (AES_BLOCK_SIZE - 1))) in ccp_run_aes_cmd() 1314 (des3->src_len & (DES3_EDE_BLOCK_SIZE - 1))) in ccp_run_des3_cmd() 1535 if (!sha->final && (sha->src_len & (block_size - 1))) in ccp_run_sha_cmd() 1541 if (!sha->src_len) { in ccp_run_sha_cmd() 1638 if (sha->src_len && !sha->src) in ccp_run_sha_cmd() 1807 hmac_cmd.u.sha.src_len = block_size + digest_size; in ccp_run_sha_cmd() 2146 op.src.u.dma.length = pt->src_len; in ccp_run_passthru_nomap_cmd() [all …]
|
/linux/net/ipv6/ |
A D | fib6_rules.c | 368 if (frh->src_len) in fib6_rule_configure() 374 rule6->src.plen = frh->src_len; in fib6_rule_configure() 403 if (frh->src_len && (rule6->src.plen != frh->src_len)) in fib6_rule_compare() 412 if (frh->src_len && in fib6_rule_compare() 429 frh->src_len = rule6->src.plen; in fib6_rule_fill()
|
/linux/arch/x86/xen/ |
A D | setup.c | 673 phys_addr_t dest_off, src_off, dest_len, src_len, len; in xen_phys_memcpy() local 682 src_len = n; in xen_phys_memcpy() 683 if (src_len > (NR_FIX_BTMAPS << PAGE_SHIFT) - src_off) in xen_phys_memcpy() 684 src_len = (NR_FIX_BTMAPS << PAGE_SHIFT) - src_off; in xen_phys_memcpy() 685 len = min(dest_len, src_len); in xen_phys_memcpy() 687 from = early_memremap(src - src_off, src_len + src_off); in xen_phys_memcpy() 690 early_memunmap(from, src_len + src_off); in xen_phys_memcpy()
|
/linux/drivers/target/ |
A D | target_core_rd.c | 428 u32 src_len; in rd_execute_rw() local 460 src_len = PAGE_SIZE - rd_offset; in rd_execute_rw() 475 len = min((u32)m.length, src_len); in rd_execute_rw() 495 src_len -= len; in rd_execute_rw() 496 if (src_len) { in rd_execute_rw() 504 src_len = PAGE_SIZE; in rd_execute_rw()
|