/linux-master/tools/power/cpupower/bench/ |
H A D | parse.c | 131 config->rounds = 50; 198 else if (strcmp("rounds", opt) == 0) 199 sscanf(val, "%u", &config->rounds);
|
H A D | system.c | 132 for (round = 0; round < config->rounds; round++) {
|
/linux-master/arch/arm64/crypto/ |
H A D | aes-ce-glue.c | 27 asmlinkage void __aes_ce_encrypt(u32 *rk, u8 *out, const u8 *in, int rounds); 28 asmlinkage void __aes_ce_decrypt(u32 *rk, u8 *out, const u8 *in, int rounds); 37 * # of rounds specified by AES: 38 * 128 bit key 10 rounds 39 * 192 bit key 12 rounds 40 * 256 bit key 14 rounds 41 * => n byte key => 6 + (n/4) rounds
|
H A D | aes-neonbs-core.S | 23 rounds .req x11 382 * void aesbs_convert_key(u8 out[], u32 const rk[], int rounds) 456 sub rounds, rounds, #1 464 subs rounds, rounds, #1 496 lsl x9, rounds, #7 522 sub rounds, rounds, #1 530 subs rounds, round [all...] |
H A D | ghash-ce-glue.c | 60 u32 const rk[], int rounds, u8 tag[]); 63 u32 const rk[], int rounds, const u8 l[], 223 * # of rounds specified by AES: 224 * 128 bit key 10 rounds 225 * 192 bit key 12 rounds 226 * 256 bit key 14 rounds 227 * => n byte key => 6 + (n/4) rounds
|
H A D | ghash-ce-core.S | 386 .macro load_round_keys, rounds, rk, tmp 390 add \tmp, \rk, \rounds, lsl #4 407 .macro enc_block, state, rounds, rk, tmp 414 tbnz \rounds, #2, .Lnot128_\@ 432 tbz \rounds, #1, .Lout192_\@ 602 * int rounds, u8 tag) 611 * int rounds, u8 tag)
|
/linux-master/drivers/md/ |
H A D | dm-verity-fec.h | 42 sector_t rounds; /* number of interleaving rounds */ member in struct:dm_verity_fec
|
H A D | dm-verity-fec.c | 39 return offset + mod * (v->fec->rounds << v->data_dev_block_bits); 454 res = div64_u64(offset, v->fec->rounds << v->data_dev_block_bits); 460 rsb = offset - res * (v->fec->rounds << v->data_dev_block_bits); 725 f->rounds = f->blocks; 726 if (sector_div(f->rounds, f->rsn)) 727 f->rounds++; 733 if (f->blocks < v->data_blocks + hash_blocks || !f->rounds) { 764 fec_blocks = div64_u64(f->rounds * f->roots, v->fec->roots << SECTOR_SHIFT);
|
/linux-master/arch/arm/crypto/ |
H A D | sha1_neon_glue.c | 30 const u8 *data, int rounds);
|
H A D | ghash-ce-glue.c | 48 int rounds; member in struct:gcm_key 367 const char *iv, int rounds, u32 counter); 371 const char *iv, int rounds, u32 counter); 375 const char *iv, int rounds, u32 counter); 379 const char *iv, int rounds, u32 counter, 397 ctx->rounds = 6 + keylen / 4; 520 ctx->rounds, counter); 561 ctx->rounds, counter); 616 ctx->rounds, counter); 648 ctx->rounds, counte [all...] |
H A D | aes-neonbs-core.S | 24 rounds .req ip 432 * void aesbs_convert_key(u8 out[], u32 const rk[], int rounds) 507 sub rounds, rounds, #1 521 subs rounds, rounds, #1 551 add bskey, bskey, rounds, lsl #7 576 sub rounds, rounds, #1 590 subs rounds, round [all...] |
H A D | aes-ce-core.S | 110 blo 0f @ AES-128: 10 rounds 113 beq 1f @ AES-192: 12 rounds 135 * r3 : number of rounds 162 .macro prepare_key, rk, rounds 163 add ip, \rk, \rounds, lsl #4 169 * aes_ecb_encrypt(u8 out[], u8 const in[], u32 const rk[], int rounds, 171 * aes_ecb_decrypt(u8 out[], u8 const in[], u32 const rk[], int rounds, 227 * aes_cbc_encrypt(u8 out[], u8 const in[], u32 const rk[], int rounds, 229 * aes_cbc_decrypt(u8 out[], u8 const in[], u32 const rk[], int rounds, 291 * int rounds, in [all...] |
/linux-master/arch/powerpc/crypto/ |
H A D | aesp8-ppc.pl | 120 my ($inp,$bits,$out,$ptr,$cnt,$rounds)=map("r$_",(3..8)); 262 li $rounds,10 336 li $rounds,12 344 li $rounds,14 398 stw $rounds,0($out) 418 slwi $cnt,$rounds,4 420 srwi $rounds,$rounds,1 422 mtctr $rounds 460 my ($inp,$out,$key,$rounds, [all...] |
H A D | chacha-p10le-8x.S | 13 # do rounds, 8 quarter rounds
|
H A D | aes-gcm-p10-glue.c | 46 u64 rounds; member in struct:aes_key
|
H A D | aes-gcm-p10.S | 571 # rounds is at offset 240 in rk 605 # load rounds - 10 (128), 12 (192), 14 (256) 1146 # load rounds - 10 (128), 12 (192), 14 (256)
|
/linux-master/arch/x86/crypto/ |
H A D | sha512-ssse3-asm.S | 152 # Compute rounds t-2 and t-1 155 # Two rounds are computed based on the values for K[t-2]+W[t-2] and 164 # The computation of the message schedule and the rounds are tightly 166 # For clarity, integer instructions (for the rounds calculation) are indented 261 movdqa %xmm0, WK_2(idx) # Store W[t]+K[t] for next rounds 310 # (80 rounds) / (2 rounds/iteration) + (1 iteration) 319 movdqa %xmm0, WK_2(t) # Store into WK for rounds
|
H A D | sha512-avx-asm.S | 157 # Compute rounds t-2 and t-1 160 # Two rounds are computed based on the values for K[t-2]+W[t-2] and 169 # The computation of the message schedule and the rounds are tightly 251 vmovdqa %xmm0, WK_2(idx) # Store W[t]+K[t] for next rounds 308 # (80 rounds) / (2 rounds/iteration) + (1 iteration) 317 vmovdqa %xmm0, WK_2(t) # Store into WK for rounds
|
H A D | twofish-x86_64-asm_64-3way.S | 47 # used only before/after all rounds 52 # used only during rounds
|
H A D | sha1_ssse3_glue.c | 261 int rounds);
|
H A D | sha256_ssse3_glue.c | 334 const u8 *data, int rounds);
|
/linux-master/lib/crypto/ |
H A D | aes.c | 179 * The expanded key size is 240 bytes (max of 14 rounds with a unique 16 bytes 261 int rounds = 6 + ctx->key_length / 4; local 287 if (round == rounds - 2) 312 int rounds = 6 + ctx->key_length / 4; local 338 if (round == rounds - 2)
|
/linux-master/drivers/crypto/ |
H A D | padlock-aes.c | 45 rounds:4, member in struct:cword 134 ctx->cword.encrypt.rounds = 10 + (key_len - 16) / 4; 135 ctx->cword.decrypt.rounds = ctx->cword.encrypt.rounds;
|
/linux-master/arch/x86/kernel/ |
H A D | asm-offsets.c | 120 OFFSET(ARIA_CTX_rounds, aria_ctx, rounds);
|
/linux-master/drivers/iio/chemical/ |
H A D | sgp40.c | 101 static u32 sgp40_exp(int exp, u32 power, u32 rounds) argument 122 for (i = 1; i <= rounds; i++) {
|