/linux/arch/arm64/crypto/ |
A D | aes-ce.S | 28 .macro load_round_keys, rounds, rk 29 cmp \rounds, #12 40 .macro enc_prepare, rounds, rk, temp 42 load_round_keys \rounds, \temp 46 .macro enc_switch_key, rounds, rk, temp 48 load_round_keys \rounds, \temp 52 .macro dec_prepare, rounds, rk, temp 54 load_round_keys \rounds, \temp 113 cmp \rounds, #12 127 do_block_Nx e, \rounds, \in [all …]
|
A D | aes-neonbs-glue.c | 29 int rounds, int blocks); 31 int rounds, int blocks); 34 int rounds, int blocks, u8 iv[]); 40 int rounds, int blocks, u8 iv[]); 46 int rounds, int blocks); 58 int rounds; member 83 ctx->rounds = 6 + key_len / 4; in aesbs_setkey() 94 int rounds, int blocks)) in __ecb_crypt() argument 112 ctx->rounds, blocks); in __ecb_crypt() 142 ctx->key.rounds = 6 + key_len / 4; in aesbs_cbc_setkey() [all …]
|
A D | aes-glue.c | 77 int rounds, int blocks); 79 int rounds, int blocks); 82 int rounds, int blocks, u8 iv[]); 84 int rounds, int blocks, u8 iv[]); 188 ctx->key_enc, rounds, blocks); in ecb_encrypt() 208 ctx->key_dec, rounds, blocks); in ecb_decrypt() 531 ctx->key1.key_enc, rounds, nbytes, in xts_encrypt() 603 ctx->key1.key_dec, rounds, nbytes, in xts_decrypt() 753 int rounds = 6 + key_len / 4; in cmac_setkey() local 763 rounds, 1); in cmac_setkey() [all …]
|
A D | aes-cipher-glue.c | 12 asmlinkage void __aes_arm64_encrypt(u32 *rk, u8 *out, const u8 *in, int rounds); 13 asmlinkage void __aes_arm64_decrypt(u32 *rk, u8 *out, const u8 *in, int rounds); 18 int rounds = 6 + ctx->key_length / 4; in aes_arm64_encrypt() local 20 __aes_arm64_encrypt(ctx->key_enc, out, in, rounds); in aes_arm64_encrypt() 26 int rounds = 6 + ctx->key_length / 4; in aes_arm64_decrypt() local 28 __aes_arm64_decrypt(ctx->key_dec, out, in, rounds); in aes_arm64_decrypt()
|
A D | aes-neon.S | 98 .macro do_block, enc, in, rounds, rk, rkp, i 101 mov \i, \rounds 114 .macro encrypt_block, in, rounds, rk, rkp, i 115 do_block 1, \in, \rounds, \rk, \rkp, \i 118 .macro decrypt_block, in, rounds, rk, rkp, i 119 do_block 0, \in, \rounds, \rk, \rkp, \i 208 mov \i, \rounds 231 .macro encrypt_block4x, in0, in1, in2, in3, rounds, rk, rkp, i 232 do_block_4x 1, \in0, \in1, \in2, \in3, \rounds, \rk, \rkp, \i 235 .macro decrypt_block4x, in0, in1, in2, in3, rounds, rk, rkp, i [all …]
|
A D | aes-cipher-core.S | 17 rounds .req x3 102 tbnz rounds, #1, 1f 107 1: subs rounds, rounds, #4
|
A D | aes-ce-ccm-glue.c | 31 u32 macp, u32 const rk[], u32 rounds); 34 u32 const rk[], u32 rounds, u8 mac[], 38 u32 const rk[], u32 rounds, u8 mac[], 42 u32 rounds);
|
/linux/arch/powerpc/crypto/ |
A D | aes-spe-glue.c | 44 u32 rounds; member 51 u32 rounds; member 99 ctx->rounds = 4; in ppc_aes_setkey() 103 ctx->rounds = 5; in ppc_aes_setkey() 107 ctx->rounds = 6; in ppc_aes_setkey() 139 ctx->rounds = 4; in ppc_xts_setkey() 144 ctx->rounds = 5; in ppc_xts_setkey() 149 ctx->rounds = 6; in ppc_xts_setkey() 236 ctx->key_enc, ctx->rounds, nbytes, in ppc_cbc_crypt() 240 ctx->key_dec, ctx->rounds, nbytes, in ppc_cbc_crypt() [all …]
|
/linux/arch/arm/crypto/ |
A D | aes-cipher-glue.c | 13 asmlinkage void __aes_arm_encrypt(u32 *rk, int rounds, const u8 *in, u8 *out); 14 asmlinkage void __aes_arm_decrypt(u32 *rk, int rounds, const u8 *in, u8 *out); 19 int rounds = 6 + ctx->key_length / 4; in aes_arm_encrypt() local 21 __aes_arm_encrypt(ctx->key_enc, rounds, in, out); in aes_arm_encrypt() 27 int rounds = 6 + ctx->key_length / 4; in aes_arm_decrypt() local 29 __aes_arm_decrypt(ctx->key_dec, rounds, in, out); in aes_arm_decrypt()
|
A D | aes-neonbs-glue.c | 32 int rounds, int blocks); 34 int rounds, int blocks); 37 int rounds, int blocks, u8 iv[]); 43 int rounds, int blocks, u8 iv[], int); 45 int rounds, int blocks, u8 iv[], int); 48 int rounds; member 79 ctx->rounds = 6 + key_len / 4; in aesbs_setkey() 90 int rounds, int blocks)) in __ecb_crypt() argument 108 ctx->rounds, blocks); in __ecb_crypt() 138 ctx->key.rounds = 6 + key_len / 4; in aesbs_cbc_setkey() [all …]
|
A D | aes-ce-glue.c | 30 int rounds, int blocks); 32 int rounds, int blocks); 35 int rounds, int blocks, u8 iv[]); 37 int rounds, int blocks, u8 iv[]); 39 int rounds, int bytes, u8 const iv[]); 41 int rounds, int bytes, u8 const iv[]); 44 int rounds, int blocks, u8 ctr[]); 47 int rounds, int bytes, u8 iv[], 50 int rounds, int bytes, u8 iv[], 448 int err, first, rounds = num_rounds(&ctx->key1); in xts_encrypt() local [all …]
|
A D | aes-cipher-core.S | 17 rounds .req r1 146 tst rounds, #2 152 1: subs rounds, rounds, #4 171 pop {rounds} // oldcpsr 172 \round r4, r5, r6, r7, r8, r9, r10, r11, \bsz, b, rounds
|
A D | aes-neonbs-core.S | 24 rounds .req ip 507 sub rounds, rounds, #1 521 subs rounds, rounds, #1 551 add bskey, bskey, rounds, lsl #7 576 sub rounds, rounds, #1 590 subs rounds, rounds, #1 643 mov rounds, r3 704 mov rounds, r3 819 mov rounds, r3 967 mov rounds, r3
|
/linux/tools/power/cpupower/bench/ |
A D | benchmark.c | 37 unsigned int rounds = 0; in calculate_timespace() local 53 rounds = (unsigned int)(load * estimated / timed); in calculate_timespace() 54 dprintf("calibrating with %u rounds\n", rounds); in calculate_timespace() 56 ROUNDS(rounds); in calculate_timespace() 60 estimated = rounds; in calculate_timespace() 91 for (_round = 1; _round <= config->rounds; _round++) in start_benchmark() 95 for (_round = 0; _round < config->rounds; _round++) { in start_benchmark()
|
A D | README-BENCH | 47 rounds=5 66 This shows expected results of the first two test run rounds from 117 -r, --rounds<int> load/sleep rounds
|
A D | main.c | 126 sscanf(optarg, "%u", &config->rounds); in main() 174 config->rounds, in main()
|
/linux/drivers/crypto/vmx/ |
A D | aesp8-ppc.pl | 419 srwi $rounds,$rounds,1 480 srwi $rounds,$rounds,1 483 subi $rounds,$rounds,1 573 srwi $rounds,$rounds,1 575 subi $rounds,$rounds,1 1343 srwi $rounds,$rounds,1 1345 subi $rounds,$rounds,1 2009 srwi $rounds,$rounds,1 2010 subi $rounds,$rounds,1 2053 srwi $rounds,$rounds,1 [all …]
|
/linux/arch/sparc/crypto/ |
A D | sha1_glue.c | 27 unsigned int rounds); 52 const unsigned int rounds = (len - done) / SHA1_BLOCK_SIZE; in __sha1_sparc64_update() local 54 sha1_sparc64_transform(sctx->state, data + done, rounds); in __sha1_sparc64_update() 55 done += rounds * SHA1_BLOCK_SIZE; in __sha1_sparc64_update()
|
A D | md5_glue.c | 30 unsigned int rounds); 58 const unsigned int rounds = (len - done) / MD5_HMAC_BLOCK_SIZE; in __md5_sparc64_update() local 60 md5_sparc64_transform(sctx->hash, data + done, rounds); in __md5_sparc64_update() 61 done += rounds * MD5_HMAC_BLOCK_SIZE; in __md5_sparc64_update()
|
A D | sha512_glue.c | 26 unsigned int rounds); 73 const unsigned int rounds = (len - done) / SHA512_BLOCK_SIZE; in __sha512_sparc64_update() local 75 sha512_sparc64_transform(sctx->state, data + done, rounds); in __sha512_sparc64_update() 76 done += rounds * SHA512_BLOCK_SIZE; in __sha512_sparc64_update()
|
A D | sha256_glue.c | 27 unsigned int rounds); 73 const unsigned int rounds = (len - done) / SHA256_BLOCK_SIZE; in __sha256_sparc64_update() local 75 sha256_sparc64_transform(sctx->state, data + done, rounds); in __sha256_sparc64_update() 76 done += rounds * SHA256_BLOCK_SIZE; in __sha256_sparc64_update()
|
/linux/drivers/md/ |
A D | dm-verity-fec.c | 38 return offset + mod * (v->fec->rounds << v->data_dev_block_bits); in fec_interleave() 450 res = div64_u64(offset, v->fec->rounds << v->data_dev_block_bits); in verity_fec_decode() 456 rsb = offset - res * (v->fec->rounds << v->data_dev_block_bits); in verity_fec_decode() 721 f->rounds = f->blocks; in verity_fec_ctr() 722 if (sector_div(f->rounds, f->rsn)) in verity_fec_ctr() 723 f->rounds++; in verity_fec_ctr() 729 if (f->blocks < v->data_blocks + hash_blocks || !f->rounds) { in verity_fec_ctr() 760 fec_blocks = div64_u64(f->rounds * f->roots, v->fec->roots << SECTOR_SHIFT); in verity_fec_ctr()
|
/linux/lib/crypto/ |
A D | aes.c | 261 int rounds = 6 + ctx->key_length / 4; in aes_encrypt() local 287 if (round == rounds - 2) in aes_encrypt() 312 int rounds = 6 + ctx->key_length / 4; in aes_decrypt() local 338 if (round == rounds - 2) in aes_decrypt()
|
/linux/arch/x86/crypto/ |
A D | sha512-ssse3-asm.S | 151 # Compute rounds t-2 and t-1 154 # Two rounds are computed based on the values for K[t-2]+W[t-2] and 163 # The computation of the message schedule and the rounds are tightly 165 # For clarity, integer instructions (for the rounds calculation) are indented 260 movdqa %xmm0, WK_2(idx) # Store W[t]+K[t] for next rounds 309 # (80 rounds) / (2 rounds/iteration) + (1 iteration) 318 movdqa %xmm0, WK_2(t) # Store into WK for rounds
|
A D | sha512-avx-asm.S | 156 # Compute rounds t-2 and t-1 159 # Two rounds are computed based on the values for K[t-2]+W[t-2] and 168 # The computation of the message schedule and the rounds are tightly 250 vmovdqa %xmm0, WK_2(idx) # Store W[t]+K[t] for next rounds 307 # (80 rounds) / (2 rounds/iteration) + (1 iteration) 316 vmovdqa %xmm0, WK_2(t) # Store into WK for rounds
|