/linux-master/arch/arm/crypto/ |
H A D | ghash-ce-glue.c | 490 static int gcm_encrypt(struct aead_request *req, const u8 *iv, u32 assoclen) argument 510 if (assoclen) 511 gcm_calculate_auth_mac(req, dg, assoclen); 543 lengths.a = cpu_to_be64(assoclen * 8); 574 scatterwalk_map_and_copy(tag, req->dst, req->assoclen + req->cryptlen, 580 static int gcm_decrypt(struct aead_request *req, const u8 *iv, u32 assoclen) argument 599 req->assoclen + req->cryptlen - authsize, 606 if (assoclen) 607 gcm_calculate_auth_mac(req, dg, assoclen); 638 lengths.a = cpu_to_be64(assoclen * [all...] |
/linux-master/arch/arm64/crypto/ |
H A D | aes-ce-ccm-glue.c | 93 if (req->assoclen) 141 u32 len = req->assoclen; 202 if (req->assoclen) 239 scatterwalk_map_and_copy(mac, req->dst, req->assoclen + req->cryptlen, 269 if (req->assoclen) 307 req->assoclen + req->cryptlen - authsize,
|
H A D | ghash-ce-glue.c | 335 static int gcm_encrypt(struct aead_request *req, char *iv, int assoclen) argument 347 lengths.a = cpu_to_be64(assoclen * 8); 350 if (assoclen) 351 gcm_calculate_auth_mac(req, dg, assoclen); 392 scatterwalk_map_and_copy(tag, req->dst, req->assoclen + req->cryptlen, 398 static int gcm_decrypt(struct aead_request *req, char *iv, int assoclen) argument 413 lengths.a = cpu_to_be64(assoclen * 8); 416 if (assoclen) 417 gcm_calculate_auth_mac(req, dg, assoclen); 422 req->assoclen [all...] |
H A D | sm4-ce-ccm-glue.c | 75 if (req->assoclen) 96 u32 assoclen = req->assoclen; local 100 if (assoclen < 0xff00) { 101 aadlen.l = cpu_to_be16(assoclen); 105 put_unaligned_be32(assoclen, &aadlen.h); 115 u32 n = scatterwalk_clamp(&walk, assoclen); 120 n = scatterwalk_clamp(&walk, assoclen); 124 assoclen -= n; 158 scatterwalk_done(&walk, 0, assoclen); [all...] |
H A D | sm4-ce-gcm-glue.c | 78 u32 assoclen = req->assoclen; local 85 u32 n = scatterwalk_clamp(&walk, assoclen); 90 n = scatterwalk_clamp(&walk, assoclen); 94 assoclen -= n; 127 scatterwalk_done(&walk, 0, assoclen); 128 } while (assoclen); 151 lengths.a = cpu_to_be64(req->assoclen * 8); 159 if (req->assoclen) 211 scatterwalk_map_and_copy(ghash, req->dst, req->assoclen [all...] |
/linux-master/arch/powerpc/crypto/ |
H A D | aes-gcm-p10-glue.c | 114 struct Hash_ctx *hash, u8 *assoc, unsigned int assoclen) 135 gctx->aadLen = assoclen; 137 if (assoclen) 138 set_aad(gctx, hash, assoc, assoclen); 213 unsigned int assoclen = req->assoclen; local 228 if (req->src->length >= assoclen && req->src->length) { 236 assocmem = kmalloc(assoclen, flags); 241 scatterwalk_map_and_copy(assoc, req->src, 0, assoclen, 0); 245 gcmp10_init(gctx, iv, (unsigned char *) &ctx->enc_key, hash, assoc, assoclen); 113 gcmp10_init(struct gcm_ctx *gctx, u8 *iv, unsigned char *rdkey, struct Hash_ctx *hash, u8 *assoc, unsigned int assoclen) argument [all...] |
/linux-master/arch/s390/crypto/ |
H A D | aes_s390.c | 832 unsigned int aadlen = req->assoclen;
|
/linux-master/arch/x86/crypto/ |
H A D | aegis128-aesni-glue.c | 45 unsigned int assoclen); 71 unsigned int assoclen) 78 while (assoclen != 0) { 79 unsigned int size = scatterwalk_clamp(&walk, assoclen); 104 assoclen -= size; 108 scatterwalk_done(&walk, 0, assoclen); 180 crypto_aegis128_aesni_process_ad(&state, req->src, req->assoclen); 182 crypto_aegis128_aesni_final(&state, tag_xor, req->assoclen, cryptlen); 203 req->assoclen + cryptlen, authsize, 1); 223 req->assoclen 69 crypto_aegis128_aesni_process_ad( struct aegis_state *state, struct scatterlist *sg_src, unsigned int assoclen) argument [all...] |
H A D | aesni-intel_glue.c | 669 unsigned int assoclen, u8 *hash_subkey, 690 if (req->src->length >= assoclen && req->src->length) { 698 assocmem = kmalloc(assoclen, flags); 703 scatterwalk_map_and_copy(assoc, req->src, 0, assoclen, 0); 709 assoclen); 712 assoclen); 714 aesni_gcm_init(aes_ctx, data, iv, hash_subkey, assoc, assoclen); 778 static int gcmaes_encrypt(struct aead_request *req, unsigned int assoclen, argument 786 err = gcmaes_crypt_by_sg(true, req, assoclen, hash_subkey, iv, aes_ctx, 792 req->assoclen 668 gcmaes_crypt_by_sg(bool enc, struct aead_request *req, unsigned int assoclen, u8 *hash_subkey, u8 *iv, void *aes_ctx, u8 *auth_tag, unsigned long auth_tag_len) argument 797 gcmaes_decrypt(struct aead_request *req, unsigned int assoclen, u8 *hash_subkey, u8 *iv, void *aes_ctx) argument [all...] |
/linux-master/crypto/ |
H A D | aegis-neon.h | 13 unsigned int assoclen,
|
H A D | aegis.h | 42 unsigned int assoclen,
|
H A D | aegis128-core.c | 278 unsigned int assoclen, 286 while (assoclen != 0) { 287 unsigned int size = scatterwalk_clamp(&walk, assoclen); 310 assoclen -= size; 313 scatterwalk_done(&walk, 0, assoclen); 346 u64 assoclen, u64 cryptlen) 348 u64 assocbits = assoclen * 8; 400 crypto_aegis128_process_ad(&state, req->src, req->assoclen, false); 403 crypto_aegis128_final(&state, &tag, req->assoclen, cryptlen); 405 scatterwalk_map_and_copy(tag.bytes, req->dst, req->assoclen 276 crypto_aegis128_process_ad(struct aegis_state *state, struct scatterlist *sg_src, unsigned int assoclen, bool do_simd) argument 344 crypto_aegis128_final(struct aegis_state *state, union aegis_block *tag_xor, u64 assoclen, u64 cryptlen) argument [all...] |
H A D | aegis128-neon-inner.c | 318 unsigned int assoclen, 328 v = st.v[3] ^ (uint8x16_t)vcombine_u64(vmov_n_u64(8ULL * assoclen), 317 crypto_aegis128_final_neon(void *state, void *tag_xor, unsigned int assoclen, unsigned int cryptlen, unsigned int authsize) argument
|
H A D | aegis128-neon.c | 57 unsigned int assoclen, 64 ret = crypto_aegis128_final_neon(state, tag_xor, assoclen, cryptlen, 55 crypto_aegis128_final_simd(struct aegis_state *state, union aegis_block *tag_xor, unsigned int assoclen, unsigned int cryptlen, unsigned int authsize) argument
|
H A D | authenc.c | 125 req->assoclen + req->cryptlen, 146 req->assoclen + req->cryptlen); 154 scatterwalk_map_and_copy(hash, req->dst, req->assoclen + req->cryptlen, 182 skcipher_request_set_crypt(skreq, req->src, req->dst, req->assoclen, 202 src = scatterwalk_ffwd(areq_ctx->src, req->src, req->assoclen); 210 dst = scatterwalk_ffwd(areq_ctx->dst, req->dst, req->assoclen); 245 src = scatterwalk_ffwd(areq_ctx->src, req->src, req->assoclen); 249 dst = scatterwalk_ffwd(areq_ctx->dst, req->dst, req->assoclen); 288 req->assoclen + req->cryptlen - authsize);
|
H A D | authencesn.c | 93 unsigned int assoclen = req->assoclen; local 100 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 0); 103 scatterwalk_map_and_copy(hash, dst, assoclen + cryptlen, authsize, 1); 125 unsigned int assoclen = req->assoclen; local 136 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 1); 142 ahash_request_set_crypt(ahreq, dst, hash, assoclen + cryptlen); 183 unsigned int assoclen = req->assoclen; local 225 unsigned int assoclen = req->assoclen; local 271 unsigned int assoclen = req->assoclen; local [all...] |
H A D | ccm.c | 144 if (req->assoclen) 176 unsigned int assoclen = req->assoclen; local 191 if (assoclen) { 192 ilen = format_adata(idata, assoclen); 202 ahash_request_set_crypt(ahreq, sg, NULL, assoclen + ilen + 16); 211 ilen = 16 - (assoclen + ilen) % 16; 237 req->assoclen + req->cryptlen, 271 sg = scatterwalk_ffwd(pctx->src + 1, req->src, req->assoclen); 278 sg = scatterwalk_ffwd(pctx->dst + 1, req->dst, req->assoclen); [all...] |
H A D | chacha20poly1305.c | 38 __le64 assoclen; member in struct:poly_req::__anon125 61 unsigned int assoclen; member in struct:chachapoly_req_ctx 101 req->assoclen + rctx->cryptlen, 113 req->assoclen + rctx->cryptlen, 136 src = scatterwalk_ffwd(rctx->src, req->src, req->assoclen); 139 dst = scatterwalk_ffwd(rctx->dst, req->dst, req->assoclen); 177 preq->tail.assoclen = cpu_to_le64(rctx->assoclen); 239 crypt = scatterwalk_ffwd(rctx->src, crypt, req->assoclen); 266 padlen = -rctx->assoclen [all...] |
H A D | cryptd.c | 763 aead_request_set_ad(subreq, req->assoclen);
|
H A D | echainiv.c | 51 req->assoclen + req->cryptlen, 63 aead_request_set_ad(subreq, req->assoclen); 69 scatterwalk_map_and_copy(info, req->dst, req->assoclen, ivsize, 1); 105 aead_request_set_ad(subreq, req->assoclen + ivsize); 107 scatterwalk_map_and_copy(req->iv, req->src, req->assoclen, ivsize, 0);
|
H A D | essiv.c | 202 req->assoclen - crypto_aead_ivsize(tfm), 207 int ssize = req->assoclen - ivsize; 239 sg = scatterwalk_ffwd(rctx->sg + 2, req->src, req->assoclen); 247 aead_request_set_ad(subreq, req->assoclen);
|
H A D | gcm.c | 164 sg = scatterwalk_ffwd(pctx->src + 1, req->src, req->assoclen); 171 sg = scatterwalk_ffwd(pctx->dst + 1, req->dst, req->assoclen); 230 lengths.a = cpu_to_be64(req->assoclen * 8); 347 remain = gcm_remain(req->assoclen); 373 if (req->assoclen) 375 req->src, req->assoclen, flags) ?: 417 req->assoclen + req->cryptlen, 474 req->assoclen + cryptlen, authsize, 0); 729 scatterwalk_map_and_copy(iv + GCM_AES_IV_SIZE, req->src, 0, req->assoclen - 8, 0); 735 sg_set_buf(rctx->src, iv + GCM_AES_IV_SIZE, req->assoclen [all...] |
H A D | pcrypt.c | 115 aead_request_set_ad(creq, req->assoclen); 164 aead_request_set_ad(creq, req->assoclen);
|
H A D | seqiv.c | 74 req->assoclen + req->cryptlen, 97 aead_request_set_ad(subreq, req->assoclen + ivsize); 100 scatterwalk_map_and_copy(info, req->dst, req->assoclen, ivsize, 1); 128 aead_request_set_ad(subreq, req->assoclen + ivsize); 130 scatterwalk_map_and_copy(req->iv, req->src, req->assoclen, ivsize, 0);
|
H A D | skcipher.c | 545 scatterwalk_copychunks(NULL, &walk->in, req->assoclen, 2); 546 scatterwalk_copychunks(NULL, &walk->out, req->assoclen, 2);
|