/linux-master/net/ipv6/ |
H A D | esp6.c | 575 int assoclen; local 585 assoclen = sizeof(struct ip_esp_hdr); 590 assoclen += sizeof(__be32); 617 assoclen + ivlen + esp->clen + alen); 645 assoclen + ivlen + esp->clen + alen); 656 aead_request_set_ad(req, assoclen); 930 int assoclen; local 948 assoclen = sizeof(struct ip_esp_hdr); 953 assoclen += seqhilen; 1004 aead_request_set_ad(req, assoclen); [all...] |
/linux-master/net/ipv4/ |
H A D | esp4.c | 538 int assoclen; local 548 assoclen = sizeof(struct ip_esp_hdr); 553 assoclen += sizeof(__be32); 580 assoclen + ivlen + esp->clen + alen); 608 assoclen + ivlen + esp->clen + alen); 619 aead_request_set_ad(req, assoclen); 886 int assoclen; local 900 assoclen = sizeof(struct ip_esp_hdr); 905 assoclen += seqhilen; 956 aead_request_set_ad(req, assoclen); [all...] |
/linux-master/drivers/crypto/xilinx/ |
H A D | zynqmp-aes-gcm.c | 190 if (req->assoclen != 0 || 225 aead_request_set_ad(subreq, areq->assoclen);
|
/linux-master/drivers/crypto/hisilicon/sec2/ |
H A D | sec_crypto.c | 901 copy_size = aead_req->cryptlen + aead_req->assoclen; 934 copy_size = c_req->c_len + aead_req->assoclen; 955 skip_size = aead_req->assoclen + aead_req->cryptlen - authsize; 1490 if (aead_req->assoclen) 1557 sec_sqe->type2.alen_ivllen = cpu_to_le32(aq->assoclen); 1559 sec_sqe->type2.cipher_src_offset = cpu_to_le16((u16)aq->assoclen); 1583 sqe3->a_len_key = cpu_to_le32(aq->assoclen); 1585 sqe3->cipher_src_offset = cpu_to_le16((u16)aq->assoclen); 1615 sec_sqe->type2.alen_ivllen = cpu_to_le32(c_req->c_len + aq->assoclen); 1617 sec_sqe->type2.cipher_src_offset = cpu_to_le16((u16)aq->assoclen); [all...] |
/linux-master/crypto/ |
H A D | testmgr.c | 2122 req->assoclen != vec->alen || 2134 if (req->assoclen != vec->alen) 2135 pr_err("alg: aead: changed 'req->assoclen'\n");
|
H A D | skcipher.c | 545 scatterwalk_copychunks(NULL, &walk->in, req->assoclen, 2); 546 scatterwalk_copychunks(NULL, &walk->out, req->assoclen, 2);
|
/linux-master/arch/arm64/crypto/ |
H A D | aes-ce-ccm-glue.c | 93 if (req->assoclen) 141 u32 len = req->assoclen; 202 if (req->assoclen) 239 scatterwalk_map_and_copy(mac, req->dst, req->assoclen + req->cryptlen, 269 if (req->assoclen) 307 req->assoclen + req->cryptlen - authsize,
|
/linux-master/net/mac802154/ |
H A D | llsec.c | 662 int authlen, assoclen, datalen, rc; local 673 assoclen = skb->mac_len; 680 sg_init_one(&sg, skb_mac_header(skb), assoclen + datalen + authlen); 683 assoclen += datalen; 689 aead_request_set_ad(req, assoclen); 871 int authlen, datalen, assoclen, rc; local 882 assoclen = skb->mac_len; 887 sg_init_one(&sg, skb_mac_header(skb), assoclen + datalen); 890 assoclen += datalen - authlen; 896 aead_request_set_ad(req, assoclen); [all...] |
/linux-master/drivers/crypto/caam/ |
H A D | caamalg_qi2.c | 382 src_len = req->assoclen + req->cryptlen; 427 src_len = req->assoclen + req->cryptlen + 451 * Create S/G table: req->assoclen, [IV,] req->src [, req->dst]. 510 edesc->assoclen = cpu_to_caam32(req->assoclen - ivsize); 512 edesc->assoclen = cpu_to_caam32(req->assoclen); 513 edesc->assoclen_dma = dma_map_single(dev, &edesc->assoclen, 4, 516 dev_err(dev, "unable to map assoclen\n"); 548 out_len = req->assoclen [all...] |
/linux-master/drivers/crypto/stm32/ |
H A D | stm32-cryp.c | 553 if (cryp->areq->assoclen) { 603 u32 alen = cryp->areq->assoclen; 654 if (cryp->areq->assoclen) 685 if (cryp->areq->assoclen) { 1190 * <- assoclen -> <- cryptlen -> 1193 * <- assoclen -> <-- cryptlen --> <- authsize -> 1197 * <- assoclen ---> <---------- cryptlen ----------> 1200 * <- assoclen -> <- cryptlen - authsize -> 1207 cryp->header_in = areq->assoclen; 1211 cryp->header_in = areq->assoclen; [all...] |
/linux-master/drivers/crypto/starfive/ |
H A D | jh7110-cryp.h | 193 size_t assoclen; member in struct:starfive_cryp_dev
|
H A D | jh7110-aes.c | 120 writel(upper_32_bits(cryp->assoclen), cryp->base + STARFIVE_AES_ALEN0); 121 writel(lower_32_bits(cryp->assoclen), cryp->base + STARFIVE_AES_ALEN1); 221 if (cryp->assoclen) 396 total_len = ALIGN(cryp->assoclen, AES_BLOCK_SIZE) / sizeof(unsigned int); 425 total_len = cryp->assoclen; 474 cryp->assoclen = 0; 478 cryp->assoclen = areq->assoclen; 495 if (cryp->assoclen) { 496 rctx->adata = kzalloc(cryp->assoclen [all...] |
/linux-master/drivers/crypto/ |
H A D | sa2ul.c | 1707 start = req->assoclen + req->cryptlen; 1917 auth_size = req->assoclen + req->cryptlen; 1935 aead_request_set_ad(subreq, req->assoclen); 1942 sa_req.enc_offset = req->assoclen;
|
H A D | atmel-aes.c | 1364 if (likely(req->assoclen != 0 || ctx->textlen != 0)) 1385 atmel_aes_write(dd, AES_AADLENR, req->assoclen); 1389 if (unlikely(req->assoclen == 0)) { 1395 padlen = atmel_aes_padlen(req->assoclen, AES_BLOCK_SIZE); 1396 if (unlikely(req->assoclen + padlen > dd->buflen)) 1398 sg_copy_to_buffer(req->src, sg_nents(req->src), dd->buf, req->assoclen); 1402 dd->datalen = req->assoclen + padlen; 1433 src = scatterwalk_ffwd(ctx->src, req->src, req->assoclen); 1435 scatterwalk_ffwd(ctx->dst, req->dst, req->assoclen)); 1473 data[0] = cpu_to_be64(req->assoclen * [all...] |
/linux-master/drivers/crypto/marvell/octeontx2/ |
H A D | otx2_cptvf_algs.c | 1052 rctx->ctrl_word.e.enc_data_offset = req->assoclen; 1056 if (req->assoclen > 248 || !IS_ALIGNED(req->assoclen, 8)) 1076 if (crypto_ipsec_check_assoclen(req->assoclen)) 1086 rctx->ctrl_word.e.iv_offset = req->assoclen - AES_GCM_IV_OFFSET; 1102 req_info->req.param2 = req->cryptlen + req->assoclen; 1106 req_info->req.param2 = req->cryptlen + req->assoclen - mac_len; 1161 u32 inputlen = req->cryptlen + req->assoclen; 1181 outputlen = req->cryptlen + req->assoclen + mac_len; 1183 outputlen = req->cryptlen + req->assoclen [all...] |
/linux-master/drivers/crypto/marvell/octeontx/ |
H A D | otx_cptvf_algs.c | 971 rctx->ctrl_word.e.enc_data_offset = req->assoclen; 999 rctx->ctrl_word.e.iv_offset = req->assoclen - AES_GCM_IV_OFFSET; 1015 req_info->req.param2 = req->cryptlen + req->assoclen; 1019 req_info->req.param2 = req->cryptlen + req->assoclen - mac_len; 1076 u32 inputlen = req->cryptlen + req->assoclen; 1096 outputlen = req->cryptlen + req->assoclen + mac_len; 1098 outputlen = req->cryptlen + req->assoclen - mac_len; 1114 inputlen = req->cryptlen + req->assoclen; 1116 inputlen = req->cryptlen + req->assoclen - mac_len; 1136 inputlen = req->cryptlen + req->assoclen; [all...] |
/linux-master/drivers/crypto/inside-secure/ |
H A D | safexcel_cipher.c | 164 u32 cryptlen, u32 assoclen, u32 digestsize) 168 int assocadj = assoclen - ctx->aadskip, aadalign; 676 unsigned int cryptlen, unsigned int assoclen, 689 unsigned int totlen_src = cryptlen + assoclen; 826 assoclen, digestsize); 844 if (assoclen) { 845 if (assoclen >= len) { 846 assoclen -= len; 851 assoclen, 852 len - assoclen); 160 safexcel_aead_token(struct safexcel_cipher_ctx *ctx, u8 *iv, struct safexcel_command_desc *cdesc, struct safexcel_token *atoken, enum safexcel_cipher_direction direction, u32 cryptlen, u32 assoclen, u32 digestsize) argument 673 safexcel_send_req(struct crypto_async_request *base, int ring, struct safexcel_cipher_req *sreq, struct scatterlist *src, struct scatterlist *dst, unsigned int cryptlen, unsigned int assoclen, unsigned int digestsize, u8 *iv, int *commands, int *results) argument [all...] |
/linux-master/drivers/crypto/ccree/ |
H A D | cc_aead.c | 236 areq->cryptlen, areq->assoclen); 736 areq_ctx->assoclen, NS_BIT); 794 * if assoclen is ZERO only IV perform 1089 if (areq_ctx->assoclen > 0) 1317 unsigned int assoclen = areq_ctx->assoclen; local 1340 if (!IS_ALIGNED(assoclen, sizeof(u32))) 1351 if (!IS_ALIGNED(assoclen, DES_BLOCK_SIZE)) 1476 if (req_ctx->assoclen > 0) { 1568 if (req_ctx->assoclen > [all...] |
/linux-master/drivers/crypto/bcm/ |
H A D | cipher.c | 1319 aead_parms.assoc_size = req->assoclen; 1392 spu->spu_ccm_update_iv(digestsize, &cipher_parms, req->assoclen, 1421 req->assoclen + rctx->total_sent - 1546 if (req->assoclen) 1548 req->assoclen); 1557 icv_offset = req->assoclen + rctx->total_sent; 1566 dump_sg(req->dst, req->assoclen, result_len); 2518 (req->assoclen == 0)) { 2542 (req->assoclen == 0)) { 2555 req->assoclen ! [all...] |
/linux-master/drivers/crypto/axis/ |
H A D | artpec6_crypto.c | 1914 __cpu_to_be64(8*areq->assoclen); 1934 count = areq->assoclen; 1939 if (!IS_ALIGNED(areq->assoclen, 16)) { 1940 size_t assoc_pad = 16 - (areq->assoclen % 16); 1980 count = artpec6_crypto_walk_advance(&walk, areq->assoclen); 2201 areq->assoclen + areq->cryptlen -
|
/linux-master/drivers/crypto/amcc/ |
H A D | crypto4xx_core.c | 600 aead_req->assoclen + aead_req->cryptlen - 685 const unsigned int assoclen, 721 tmp = sg_nents_for_len(src, assoclen + datalen); 730 if (assoclen) { 731 nbytes += assoclen; 732 dst = scatterwalk_ffwd(_dst, dst, assoclen); 822 sa->sa_command_1.bf.hash_crypto_offset = (assoclen >> 2); 921 pd->pd_ctl_len.w = 0x00400000 | (assoclen + datalen); 677 crypto4xx_build_pd(struct crypto_async_request *req, struct crypto4xx_ctx *ctx, struct scatterlist *src, struct scatterlist *dst, const unsigned int datalen, const __le32 *iv, const u32 iv_len, const struct dynamic_sa_ctl *req_sa, const unsigned int sa_len, const unsigned int assoclen, struct scatterlist *_dst) argument
|
H A D | crypto4xx_core.h | 161 const unsigned int assoclen,
|
H A D | crypto4xx_alg.c | 327 if (req->assoclen & 0x3 || req->assoclen > 1020) 347 aead_request_set_ad(subreq, req->assoclen); 456 sa, ctx->sa_len, req->assoclen, rctx->dst); 593 ctx->sa_len, req->assoclen, rctx->dst);
|
/linux-master/arch/s390/crypto/ |
H A D | aes_s390.c | 832 unsigned int aadlen = req->assoclen;
|
/linux-master/arch/powerpc/crypto/ |
H A D | aes-gcm-p10-glue.c | 114 struct Hash_ctx *hash, u8 *assoc, unsigned int assoclen) 135 gctx->aadLen = assoclen; 137 if (assoclen) 138 set_aad(gctx, hash, assoc, assoclen); 213 unsigned int assoclen = req->assoclen; local 228 if (req->src->length >= assoclen && req->src->length) { 236 assocmem = kmalloc(assoclen, flags); 241 scatterwalk_map_and_copy(assoc, req->src, 0, assoclen, 0); 245 gcmp10_init(gctx, iv, (unsigned char *) &ctx->enc_key, hash, assoc, assoclen); 113 gcmp10_init(struct gcm_ctx *gctx, u8 *iv, unsigned char *rdkey, struct Hash_ctx *hash, u8 *assoc, unsigned int assoclen) argument [all...] |