/linux-master/crypto/ |
H A D | sm4.c | 104 static inline u32 sm4_round(u32 x0, u32 x1, u32 x2, u32 x3, u32 rk) argument 106 return x0 ^ sm4_enc_sub(x1 ^ x2 ^ x3 ^ rk); 122 u32 rk[4]; local 129 rk[0] = get_unaligned_be32(&key[0]) ^ fk[0]; 130 rk[1] = get_unaligned_be32(&key[1]) ^ fk[1]; 131 rk[2] = get_unaligned_be32(&key[2]) ^ fk[2]; 132 rk[3] = get_unaligned_be32(&key[3]) ^ fk[3]; 135 rk[0] ^= sm4_key_sub(rk[1] ^ rk[ 160 sm4_crypt_block(const u32 *rk, u8 *out, const u8 *in) argument [all...] |
/linux-master/include/crypto/ |
H A D | sm4.h | 42 * @rk: The rkey_enc for encrypt or rkey_dec for decrypt 46 void sm4_crypt_block(const u32 *rk, u8 *out, const u8 *in);
|
H A D | aria.h | 407 static inline void aria_add_round_key(u32 *rk, u32 *t0, u32 *t1, u32 *t2, argument 410 *t0 ^= rk[0]; 411 *t1 ^= rk[1]; 412 *t2 ^= rk[2]; 413 *t3 ^= rk[3]; 434 static inline void aria_gsrk(u32 *rk, u32 *x, u32 *y, u32 n) argument 439 rk[0] = (x[0]) ^ 442 rk[1] = (x[1]) ^ 445 rk[2] = (x[2]) ^ 448 rk[ [all...] |
/linux-master/arch/x86/crypto/ |
H A D | sm4-avx.h | 8 typedef void (*sm4_crypt_func)(const u32 *rk, u8 *dst, const u8 *src, u8 *iv);
|
H A D | sm4_aesni_avx2_glue.c | 22 asmlinkage void sm4_aesni_avx2_ctr_enc_blk16(const u32 *rk, u8 *dst, 24 asmlinkage void sm4_aesni_avx2_cbc_dec_blk16(const u32 *rk, u8 *dst,
|
H A D | aria-gfni-avx512-asm_64.S | 272 t0, rk, round) \ 274 vpbroadcastb ((round * 16) + 3)(rk), t0; \ 276 vpbroadcastb ((round * 16) + 2)(rk), t0; \ 278 vpbroadcastb ((round * 16) + 1)(rk), t0; \ 280 vpbroadcastb ((round * 16) + 0)(rk), t0; \ 282 vpbroadcastb ((round * 16) + 7)(rk), t0; \ 284 vpbroadcastb ((round * 16) + 6)(rk), t0; \ 286 vpbroadcastb ((round * 16) + 5)(rk), t0; \ 288 vpbroadcastb ((round * 16) + 4)(rk), t0; \ 290 vpbroadcastb ((round * 16) + 11)(rk), t [all...] |
H A D | sm4_aesni_avx_glue.c | 22 asmlinkage void sm4_aesni_avx_crypt4(const u32 *rk, u8 *dst, 24 asmlinkage void sm4_aesni_avx_crypt8(const u32 *rk, u8 *dst, 26 asmlinkage void sm4_aesni_avx_ctr_enc_blk8(const u32 *rk, u8 *dst, 28 asmlinkage void sm4_aesni_avx_cbc_dec_blk8(const u32 *rk, u8 *dst,
|
H A D | aria-aesni-avx2-asm_64.S | 286 t0, rk, idx, round) \ 288 vpbroadcastb ((round * 16) + idx + 3)(rk), t0; \ 290 vpbroadcastb ((round * 16) + idx + 2)(rk), t0; \ 292 vpbroadcastb ((round * 16) + idx + 1)(rk), t0; \ 294 vpbroadcastb ((round * 16) + idx + 0)(rk), t0; \ 296 vpbroadcastb ((round * 16) + idx + 7)(rk), t0; \ 298 vpbroadcastb ((round * 16) + idx + 6)(rk), t0; \ 300 vpbroadcastb ((round * 16) + idx + 5)(rk), t0; \ 302 vpbroadcastb ((round * 16) + idx + 4)(rk), t0; \ 464 mem_tmp, rk, roun [all...] |
H A D | aria-aesni-avx-asm_64.S | 270 t0, t1, t2, rk, \ 273 vbroadcastss ((round * 16) + idx + 0)(rk), t0; \ 285 vbroadcastss ((round * 16) + idx + 4)(rk), t0; \ 423 mem_tmp, rk, round) \ 426 y0, y7, y2, rk, 8, round); \ 441 y0, y7, y2, rk, 0, round); \ 478 mem_tmp, rk, round) \ 481 y0, y7, y2, rk, 8, round); \ 496 y0, y7, y2, rk, 0, round); \ 533 mem_tmp, rk, roun [all...] |
/linux-master/arch/arm64/crypto/ |
H A D | aes-neonbs-glue.c | 26 asmlinkage void aesbs_convert_key(u8 out[], u32 const rk[], int rounds); 28 asmlinkage void aesbs_ecb_encrypt(u8 out[], u8 const in[], u8 const rk[], 30 asmlinkage void aesbs_ecb_decrypt(u8 out[], u8 const in[], u8 const rk[], 33 asmlinkage void aesbs_cbc_decrypt(u8 out[], u8 const in[], u8 const rk[], 36 asmlinkage void aesbs_ctr_encrypt(u8 out[], u8 const in[], u8 const rk[], 39 asmlinkage void aesbs_xts_encrypt(u8 out[], u8 const in[], u8 const rk[], 41 asmlinkage void aesbs_xts_decrypt(u8 out[], u8 const in[], u8 const rk[], 45 asmlinkage void neon_aes_ecb_encrypt(u8 out[], u8 const in[], u32 const rk[], 47 asmlinkage void neon_aes_cbc_encrypt(u8 out[], u8 const in[], u32 const rk[], 49 asmlinkage void neon_aes_ctr_encrypt(u8 out[], u8 const in[], u32 const rk[], 59 u8 rk[13 * (8 * AES_BLOCK_SIZE) + 32]; member in struct:aesbs_ctx 78 struct crypto_aes_ctx rk; local 94 __ecb_crypt(struct skcipher_request *req, void (*fn)(u8 out[], u8 const in[], u8 const rk[], int rounds, int blocks)) argument 137 struct crypto_aes_ctx rk; local 255 struct crypto_aes_ctx rk; local 276 __xts_crypt(struct skcipher_request *req, bool encrypt, void (*fn)(u8 out[], u8 const in[], u8 const rk[], int rounds, int blocks, u8 iv[])) argument [all...] |
H A D | aes-ce.S | 28 .macro load_round_keys, rounds, rk 32 ld1 {v17.4s-v18.4s}, [\rk], #32 33 1111: ld1 {v19.4s-v20.4s}, [\rk], #32 34 2222: ld1 {v21.4s-v24.4s}, [\rk], #64 35 ld1 {v25.4s-v28.4s}, [\rk], #64 36 ld1 {v29.4s-v31.4s}, [\rk] 39 /* prepare for encryption with key in rk[] */ 40 .macro enc_prepare, rounds, rk, temp 41 mov \temp, \rk 45 /* prepare for encryption (again) but with new key in rk[] */ [all...] |
H A D | aes-cipher-glue.c | 12 asmlinkage void __aes_arm64_encrypt(u32 *rk, u8 *out, const u8 *in, int rounds); 13 asmlinkage void __aes_arm64_decrypt(u32 *rk, u8 *out, const u8 *in, int rounds);
|
H A D | aes-neon.S | 98 .macro do_block, enc, in, rounds, rk, rkp, i 99 ld1 {v15.4s}, [\rk] 100 add \rkp, \rk, #16 114 .macro encrypt_block, in, rounds, rk, rkp, i 115 do_block 1, \in, \rounds, \rk, \rkp, \i 118 .macro decrypt_block, in, rounds, rk, rkp, i 119 do_block 0, \in, \rounds, \rk, \rkp, \i 205 .macro do_block_4x, enc, in0, in1, in2, in3, rounds, rk, rkp, i 206 ld1 {v15.4s}, [\rk] 207 add \rkp, \rk, #1 [all...] |
H A D | aes-cipher-core.S | 14 rk .req x0 57 ldp \out0, \out1, [rk], #8 87 ldp w8, w9, [rk], #16 88 ldp w10, w11, [rk, #-8]
|
H A D | aes-ce-ccm-glue.c | 35 asmlinkage u32 ce_aes_mac_update(u8 const in[], u32 const rk[], int rounds, 40 u32 const rk[], u32 rounds, u8 mac[], 44 u32 const rk[], u32 rounds, u8 mac[], 101 u32 macp, u32 const rk[], u32 rounds) 109 u32 rem = ce_aes_mac_update(in, rk, rounds, blocks, mac, 100 ce_aes_ccm_auth_data(u8 mac[], u8 const in[], u32 abytes, u32 macp, u32 const rk[], u32 rounds) argument
|
H A D | aes-ce-ccm-core.S | 17 .macro load_round_keys, rk, nr, tmp 19 add \tmp, \rk, w\tmp, sxtw #4 20 ld1 {v10.4s-v13.4s}, [\rk] 73 eor v0.16b, v0.16b, v2.16b /* xor mac with pt ^ rk[last] */ 120 * u8 const rk[], u32 rounds, u8 mac[], 123 * u8 const rk[], u32 rounds, u8 mac[],
|
H A D | aes-ce-glue.c | 27 asmlinkage void __aes_ce_encrypt(u32 *rk, u8 *out, const u8 *in, int rounds); 28 asmlinkage void __aes_ce_decrypt(u32 *rk, u8 *out, const u8 *in, int rounds);
|
H A D | sm4-ce-cipher-glue.c | 18 asmlinkage void sm4_ce_do_crypt(const u32 *rk, void *out, const void *in);
|
/linux-master/arch/arm/crypto/ |
H A D | aes-neonbs-glue.c | 29 asmlinkage void aesbs_convert_key(u8 out[], u32 const rk[], int rounds); 31 asmlinkage void aesbs_ecb_encrypt(u8 out[], u8 const in[], u8 const rk[], 33 asmlinkage void aesbs_ecb_decrypt(u8 out[], u8 const in[], u8 const rk[], 36 asmlinkage void aesbs_cbc_decrypt(u8 out[], u8 const in[], u8 const rk[], 39 asmlinkage void aesbs_ctr_encrypt(u8 out[], u8 const in[], u8 const rk[], 42 asmlinkage void aesbs_xts_encrypt(u8 out[], u8 const in[], u8 const rk[], 44 asmlinkage void aesbs_xts_decrypt(u8 out[], u8 const in[], u8 const rk[], 49 u8 rk[13 * (8 * AES_BLOCK_SIZE) + 32] __aligned(AES_BLOCK_SIZE); 72 struct crypto_aes_ctx rk; local 75 err = aes_expandkey(&rk, in_ke 88 __ecb_crypt(struct skcipher_request *req, void (*fn)(u8 out[], u8 const in[], u8 const rk[], int rounds, int blocks)) argument 131 struct crypto_aes_ctx rk; local 337 __xts_crypt(struct skcipher_request *req, bool encrypt, void (*fn)(u8 out[], u8 const in[], u8 const rk[], int rounds, int blocks, u8 iv[], int)) argument [all...] |
H A D | aes-cipher-glue.c | 13 asmlinkage void __aes_arm_encrypt(u32 *rk, int rounds, const u8 *in, u8 *out); 14 asmlinkage void __aes_arm_decrypt(u32 *rk, int rounds, const u8 *in, u8 *out);
|
H A D | sha2-ce-core.S | 18 rk .req r3 35 vld1.32 {k\ev}, [rk, :128]! 90 adr rk, .Lsha256_rcon 91 vld1.32 {k0}, [rk, :128]!
|
H A D | aes-cipher-core.S | 16 rk .req r0 84 ldm rk!, {t1, t2} 106 ldm rk!, {r8-r11}
|
H A D | aes-ce-glue.c | 29 asmlinkage void ce_aes_ecb_encrypt(u8 out[], u8 const in[], u32 const rk[], 31 asmlinkage void ce_aes_ecb_decrypt(u8 out[], u8 const in[], u32 const rk[], 34 asmlinkage void ce_aes_cbc_encrypt(u8 out[], u8 const in[], u32 const rk[], 36 asmlinkage void ce_aes_cbc_decrypt(u8 out[], u8 const in[], u32 const rk[], 38 asmlinkage void ce_aes_cbc_cts_encrypt(u8 out[], u8 const in[], u32 const rk[], 40 asmlinkage void ce_aes_cbc_cts_decrypt(u8 out[], u8 const in[], u32 const rk[], 43 asmlinkage void ce_aes_ctr_encrypt(u8 out[], u8 const in[], u32 const rk[],
|
/linux-master/arch/loongarch/include/asm/ |
H A D | inst.h | 341 unsigned int rk : 5; member in struct:reg3_format 348 unsigned int rk : 5; member in struct:reg3sa2_format 505 u32 larch_insn_gen_or(enum loongarch_gpr rd, enum loongarch_gpr rj, enum loongarch_gpr rk); 703 enum loongarch_gpr rk) \ 708 insn->reg3_format.rk = rk; \ 754 enum loongarch_gpr rk, \ 761 insn->reg3sa2_format.rk = rk; \
|
/linux-master/drivers/net/ethernet/stmicro/stmmac/ |
H A D | Makefile | 23 obj-$(CONFIG_DWMAC_ROCKCHIP) += dwmac-rk.o
|