/linux-master/drivers/crypto/xilinx/ |
H A D | zynqmp-aes-gcm.c | 190 if (req->assoclen != 0 || 225 aead_request_set_ad(subreq, areq->assoclen);
|
/linux-master/drivers/crypto/bcm/ |
H A D | spu.h | 271 unsigned int assoclen,
|
H A D | spu2.h | 213 unsigned int assoclen, unsigned int chunksize,
|
H A D | cipher.h | 400 unsigned int assoclen, unsigned int chunksize,
|
H A D | spu.c | 1135 * @assoclen: Length of AAD data 1143 unsigned int assoclen, 1188 adata = (assoclen > 0); /* adata = 1 if any associated data */ 1141 spum_ccm_update_iv(unsigned int digestsize, struct spu_cipher_parms *cipher_parms, unsigned int assoclen, unsigned int chunksize, bool is_encrypt, bool is_esp) argument
|
H A D | spu2.c | 1345 * @assoclen: Length of AAD data 1353 unsigned int assoclen, unsigned int chunksize, 1351 spu2_ccm_update_iv(unsigned int digestsize, struct spu_cipher_parms *cipher_parms, unsigned int assoclen, unsigned int chunksize, bool is_encrypt, bool is_esp) argument
|
/linux-master/drivers/crypto/ccp/ |
H A D | ccp-crypto-aes-galois.c | 129 rctx->cmd.u.aes.aad_len = req->assoclen;
|
/linux-master/drivers/crypto/starfive/ |
H A D | jh7110-cryp.h | 193 size_t assoclen; member in struct:starfive_cryp_dev
|
/linux-master/drivers/crypto/amcc/ |
H A D | crypto4xx_core.h | 161 const unsigned int assoclen,
|
/linux-master/drivers/crypto/caam/ |
H A D | caamalg_qi.c | 797 * @assoclen: associated data length, in CAAM endianness 798 * @assoclen_dma: bus physical mapped address of req->assoclen 808 unsigned int assoclen; member in struct:aead_edesc 974 src_len = req->assoclen + req->cryptlen + 993 src_len = req->assoclen + req->cryptlen; 1044 * Create S/G table: req->assoclen, [IV,] req->src [, req->dst]. 1099 edesc->assoclen = cpu_to_caam32(req->assoclen); 1100 edesc->assoclen_dma = dma_map_single(qidev, &edesc->assoclen, 4, 1103 dev_err(qidev, "unable to map assoclen\ [all...] |
H A D | caamalg.c | 1101 append_seq_in_ptr(desc, src_dma, req->assoclen + req->cryptlen, 1124 req->assoclen + req->cryptlen + authsize, 1128 req->assoclen + req->cryptlen - authsize, 1144 append_math_add_imm_u32(desc, REG3, ZERO, IMM, req->assoclen); 1148 if (encrypt && generic_gcm && !(req->assoclen + req->cryptlen)) 1168 unsigned int assoclen = req->assoclen; local 1182 assoclen -= ivsize; 1185 append_math_add_imm_u32(desc, REG3, ZERO, IMM, assoclen); 1231 * {REG3, DPOVRD} = assoclen, dependin [all...] |
H A D | caamalg_qi2.c | 382 src_len = req->assoclen + req->cryptlen; 427 src_len = req->assoclen + req->cryptlen + 451 * Create S/G table: req->assoclen, [IV,] req->src [, req->dst]. 510 edesc->assoclen = cpu_to_caam32(req->assoclen - ivsize); 512 edesc->assoclen = cpu_to_caam32(req->assoclen); 513 edesc->assoclen_dma = dma_map_single(dev, &edesc->assoclen, 4, 516 dev_err(dev, "unable to map assoclen\n"); 548 out_len = req->assoclen [all...] |
/linux-master/drivers/crypto/ |
H A D | talitos.c | 1005 cryptlen + authsize, areq->assoclen); 1015 areq->assoclen + cryptlen - ivsize); 1223 areq->assoclen + cryptlen); 1230 ret = talitos_sg_map(dev, areq->src, areq->assoclen, edesc, 1255 sg_count, areq->assoclen, tbl_off, elen, 1275 sg_count, areq->assoclen, tbl_off, elen, 1295 sg_count, areq->assoclen + cryptlen, tbl_off); 1323 unsigned int assoclen, 1346 src_len = assoclen + cryptlen + authsize; 1356 src_len = assoclen 1319 talitos_edesc_alloc(struct device *dev, struct scatterlist *src, struct scatterlist *dst, u8 *iv, unsigned int assoclen, unsigned int cryptlen, unsigned int authsize, unsigned int ivsize, int icv_stashing, u32 cryptoflags, bool encrypt) argument [all...] |
H A D | atmel-sha.c | 2119 u32 assoclen; member in struct:atmel_sha_authenc_reqctx 2275 struct scatterlist *assoc, unsigned int assoclen, 2286 if (unlikely(!IS_ALIGNED(assoclen, sizeof(u32)))) 2292 authctx->assoclen = assoclen; 2325 msg_size = authctx->assoclen + authctx->textlen; 2332 return atmel_sha_cpu_start(dd, authctx->assoc, authctx->assoclen, 2274 atmel_sha_authenc_init(struct ahash_request *req, struct scatterlist *assoc, unsigned int assoclen, unsigned int textlen, atmel_aes_authenc_fn_t cb, struct atmel_aes_dev *aes_dev) argument
|
H A D | sa2ul.c | 1707 start = req->assoclen + req->cryptlen; 1917 auth_size = req->assoclen + req->cryptlen; 1935 aead_request_set_ad(subreq, req->assoclen); 1942 sa_req.enc_offset = req->assoclen;
|
/linux-master/drivers/crypto/qce/ |
H A D | common.c | 514 totallen = rctx->cryptlen + rctx->assoclen; 521 qce_write(qce, REG_ENCR_SEG_START, rctx->assoclen & 0xffff);
|
/linux-master/drivers/crypto/axis/ |
H A D | artpec6_crypto.c | 1914 __cpu_to_be64(8*areq->assoclen); 1934 count = areq->assoclen; 1939 if (!IS_ALIGNED(areq->assoclen, 16)) { 1940 size_t assoc_pad = 16 - (areq->assoclen % 16); 1980 count = artpec6_crypto_walk_advance(&walk, areq->assoclen); 2201 areq->assoclen + areq->cryptlen -
|
/linux-master/drivers/crypto/cavium/nitrox/ |
H A D | nitrox_req.h | 223 * @assoclen: AAD length 237 unsigned int assoclen; member in struct:nitrox_aead_rctx
|
/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | qat_algs.c | 816 cipher_param->cipher_offset = areq->assoclen; 820 auth_param->auth_len = areq->assoclen + cipher_param->cipher_length; 863 cipher_param->cipher_offset = areq->assoclen; 866 auth_param->auth_len = areq->assoclen + areq->cryptlen;
|
/linux-master/drivers/crypto/intel/ixp4xx/ |
H A D | ixp4xx_crypto.c | 357 int decryptlen = req->assoclen + req->cryptlen - authsize; 1133 crypt->auth_len = req->assoclen + cryptlen; 1304 return aead_perform(req, 1, req->assoclen, req->cryptlen, req->iv); 1309 return aead_perform(req, 0, req->assoclen, req->cryptlen, req->iv);
|
/linux-master/crypto/ |
H A D | skcipher.c | 545 scatterwalk_copychunks(NULL, &walk->in, req->assoclen, 2); 546 scatterwalk_copychunks(NULL, &walk->out, req->assoclen, 2);
|
H A D | cryptd.c | 763 aead_request_set_ad(subreq, req->assoclen);
|
H A D | testmgr.c | 2122 req->assoclen != vec->alen || 2134 if (req->assoclen != vec->alen) 2135 pr_err("alg: aead: changed 'req->assoclen'\n");
|
/linux-master/arch/s390/crypto/ |
H A D | aes_s390.c | 832 unsigned int aadlen = req->assoclen;
|