Searched refs:cryptlen (Results 1 - 25 of 110) sorted by relevance

12345

/linux-master/crypto/
H A Daegis-neon.h14 unsigned int cryptlen,
H A Dkeywrap.c130 u64 t = 6 * ((req->cryptlen) >> 3);
138 if (req->cryptlen < (2 * SEMIBSIZE) || req->cryptlen % SEMIBSIZE)
154 unsigned int nbytes = req->cryptlen;
208 if (req->cryptlen < (2 * SEMIBSIZE) || req->cryptlen % SEMIBSIZE)
227 unsigned int nbytes = req->cryptlen;
H A Dauthencesn.c94 unsigned int cryptlen = req->cryptlen; local
100 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 0);
103 scatterwalk_map_and_copy(hash, dst, assoclen + cryptlen, authsize, 1);
126 unsigned int cryptlen = req->cryptlen; local
136 scatterwalk_map_and_copy(tmp + 1, dst, assoclen + cryptlen, 4, 1);
142 ahash_request_set_crypt(ahreq, dst, hash, assoclen + cryptlen);
184 unsigned int cryptlen = req->cryptlen; local
224 unsigned int cryptlen = req->cryptlen - authsize; local
272 unsigned int cryptlen = req->cryptlen; local
[all...]
H A Daegis128-neon.c58 unsigned int cryptlen,
64 ret = crypto_aegis128_final_neon(state, tag_xor, assoclen, cryptlen,
55 crypto_aegis128_final_simd(struct aegis_state *state, union aegis_block *tag_xor, unsigned int assoclen, unsigned int cryptlen, unsigned int authsize) argument
H A Dechainiv.c37 if (req->cryptlen < ivsize)
51 req->assoclen + req->cryptlen,
62 req->cryptlen, info);
94 if (req->cryptlen < ivsize)
104 req->cryptlen - ivsize, req->iv);
H A Daegis128-core.c346 u64 assoclen, u64 cryptlen)
349 u64 cryptbits = cryptlen * 8;
394 unsigned int cryptlen = req->cryptlen; local
403 crypto_aegis128_final(&state, &tag, req->assoclen, cryptlen);
405 scatterwalk_map_and_copy(tag.bytes, req->dst, req->assoclen + cryptlen,
416 unsigned int cryptlen = req->cryptlen - authsize; local
421 scatterwalk_map_and_copy(tag.bytes, req->src, req->assoclen + cryptlen,
429 crypto_aegis128_final(&state, &tag, req->assoclen, cryptlen);
344 crypto_aegis128_final(struct aegis_state *state, union aegis_block *tag_xor, u64 assoclen, u64 cryptlen) argument
455 unsigned int cryptlen = req->cryptlen; local
479 unsigned int cryptlen = req->cryptlen - authsize; local
[all...]
H A Dccm.c129 unsigned int cryptlen)
147 return set_msg_len(info + 16 - l, cryptlen, l);
170 unsigned int cryptlen)
183 err = format_input(odata, req, cryptlen);
219 cryptlen += ilen;
222 ahash_request_set_crypt(ahreq, plain, odata, cryptlen);
237 req->assoclen + req->cryptlen,
293 unsigned int cryptlen = req->cryptlen; local
302 err = crypto_ccm_auth(req, sg_next(pctx->src), cryptlen);
128 format_input(u8 *info, struct aead_request *req, unsigned int cryptlen) argument
169 crypto_ccm_auth(struct aead_request *req, struct scatterlist *plain, unsigned int cryptlen) argument
330 unsigned int cryptlen = req->cryptlen - authsize; local
353 unsigned int cryptlen = req->cryptlen; local
[all...]
H A Dxts.c87 const bool cts = (req->cryptlen % XTS_BLOCK_SIZE);
164 int offset = req->cryptlen & ~(XTS_BLOCK_SIZE - 1);
167 int tail = req->cryptlen % XTS_BLOCK_SIZE;
209 if (!err && unlikely(req->cryptlen % XTS_BLOCK_SIZE)) {
229 if (!err && unlikely(req->cryptlen % XTS_BLOCK_SIZE)) {
247 if (req->cryptlen < XTS_BLOCK_SIZE)
253 req->cryptlen & ~(XTS_BLOCK_SIZE - 1), NULL);
272 if (err || likely((req->cryptlen % XTS_BLOCK_SIZE) == 0))
289 if (err || likely((req->cryptlen % XTS_BLOCK_SIZE) == 0))
H A Dseqiv.c58 if (req->cryptlen < ivsize)
74 req->assoclen + req->cryptlen,
96 req->cryptlen - ivsize, info);
117 if (req->cryptlen < ivsize + crypto_aead_authsize(geniv))
127 req->cryptlen - ivsize, req->iv);
H A Dchacha20poly1305.c39 __le64 cryptlen; member in struct:poly_req::__anon125
59 unsigned int cryptlen; member in struct:chachapoly_req_ctx
101 req->assoclen + rctx->cryptlen,
113 req->assoclen + rctx->cryptlen,
131 if (rctx->cryptlen == 0)
145 rctx->cryptlen, creq->iv);
158 if (rctx->cryptlen == req->cryptlen) /* encrypting */
178 preq->tail.cryptlen = cpu_to_le64(rctx->cryptlen);
[all...]
H A Dgcm.c58 unsigned int cryptlen; member in struct:crypto_gcm_ghash_ctx
178 unsigned int cryptlen)
190 cryptlen + sizeof(pctx->auth_tag),
231 lengths.b = cpu_to_be64(gctx->cryptlen * 8);
291 remain = gcm_remain(gctx->cryptlen);
320 if (gctx->cryptlen)
322 gctx->src, gctx->cryptlen, flags) ?:
417 req->assoclen + req->cryptlen,
428 gctx->cryptlen = req->cryptlen;
177 crypto_gcm_init_crypt(struct aead_request *req, unsigned int cryptlen) argument
470 unsigned int cryptlen = req->cryptlen - authsize; local
505 unsigned int cryptlen = req->cryptlen; local
[all...]
/linux-master/arch/x86/crypto/
H A Daegis128-aesni-glue.c44 void *state, void *tag_xor, unsigned int cryptlen,
167 unsigned int cryptlen,
182 crypto_aegis128_aesni_final(&state, tag_xor, req->assoclen, cryptlen);
198 unsigned int cryptlen = req->cryptlen; local
200 crypto_aegis128_aesni_crypt(req, &tag, cryptlen, &OPS);
203 req->assoclen + cryptlen, authsize, 1);
220 unsigned int cryptlen = req->cryptlen - authsize; local
223 req->assoclen + cryptlen, authsiz
165 crypto_aegis128_aesni_crypt(struct aead_request *req, struct aegis_block *tag_xor, unsigned int cryptlen, const struct aegis_crypt_ops *ops) argument
[all...]
/linux-master/drivers/crypto/qce/
H A Dcipher.h32 * @cryptlen: crypto length
44 unsigned int cryptlen; member in struct:qce_cipher_reqctx
H A Daead.c71 totallen = req->cryptlen + req->assoclen;
80 totallen = req->cryptlen + req->assoclen - ctx->authsize;
124 totallen = rctx->cryptlen + assoclen;
152 sg = qce_sgtable_add(&rctx->dst_tbl, msg_sg, rctx->cryptlen);
157 totallen = rctx->cryptlen + rctx->assoclen;
192 unsigned int adata_header_len, cryptlen, totallen; local
198 cryptlen = rctx->cryptlen + ctx->authsize;
200 cryptlen = rctx->cryptlen;
336 unsigned int cryptlen; local
[all...]
H A Daead.h38 unsigned int cryptlen; member in struct:qce_aead_reqctx
/linux-master/drivers/crypto/allwinner/sun8i-ce/
H A Dsun8i-ce-cipher.c34 if (sg_nents_for_len(areq->src, areq->cryptlen) > MAX_SG ||
35 sg_nents_for_len(areq->dst, areq->cryptlen) > MAX_SG) {
40 if (areq->cryptlen < crypto_skcipher_ivsize(tfm)) {
45 if (areq->cryptlen == 0) {
50 if (areq->cryptlen % 16) {
55 len = areq->cryptlen;
71 len = areq->cryptlen;
112 areq->cryptlen, areq->iv);
138 int ns = sg_nents_for_len(areq->src, areq->cryptlen);
139 int nd = sg_nents_for_len(areq->dst, areq->cryptlen);
[all...]
/linux-master/drivers/crypto/
H A Domap-aes-gcm.c94 int alen, clen, cryptlen, assoclen, ret; local
102 cryptlen = req->cryptlen;
108 cryptlen -= authlen;
111 clen = ALIGN(cryptlen, AES_BLOCK_SIZE);
113 nsg = !!(assoclen && cryptlen);
131 if (cryptlen) {
137 ret = omap_crypto_align_sg(&tmp, cryptlen,
149 dd->total = cryptlen;
162 if (cryptlen) {
[all...]
/linux-master/drivers/crypto/allwinner/sun8i-ss/
H A Dsun8i-ss-cipher.c33 if (areq->cryptlen == 0 || areq->cryptlen % 16) {
38 if (sg_nents_for_len(areq->src, areq->cryptlen) > 8 ||
39 sg_nents_for_len(areq->dst, areq->cryptlen) > 8) {
44 len = areq->cryptlen;
59 len = areq->cryptlen;
112 areq->cryptlen, areq->iv);
128 unsigned int len = areq->cryptlen;
137 offset = areq->cryptlen - ivsize;
197 int nsgs = sg_nents_for_len(areq->src, areq->cryptlen);
[all...]
/linux-master/include/crypto/
H A Daead.h84 * @cryptlen: Length of data to be encrypted or decrypted
94 unsigned int cryptlen; member in struct:aead_request
514 * @cryptlen: number of bytes to process from @src
542 unsigned int cryptlen, u8 *iv)
546 req->cryptlen = cryptlen;
539 aead_request_set_crypt(struct aead_request *req, struct scatterlist *src, struct scatterlist *dst, unsigned int cryptlen, u8 *iv) argument
/linux-master/arch/arm/crypto/
H A Daes-ce-glue.c271 int cbc_blocks = DIV_ROUND_UP(req->cryptlen, AES_BLOCK_SIZE) - 2;
282 if (req->cryptlen <= AES_BLOCK_SIZE) {
283 if (req->cryptlen < AES_BLOCK_SIZE)
298 if (req->cryptlen == AES_BLOCK_SIZE)
301 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen);
304 subreq.cryptlen);
309 req->cryptlen - cbc_blocks * AES_BLOCK_SIZE,
329 int cbc_blocks = DIV_ROUND_UP(req->cryptlen, AES_BLOCK_SIZE) - 2;
340 if (req->cryptlen <= AES_BLOCK_SIZE) {
341 if (req->cryptlen < AES_BLOCK_SIZ
[all...]
/linux-master/drivers/crypto/gemini/
H A Dsl3516-ce-cipher.c34 if (areq->cryptlen == 0 || areq->cryptlen % 16) {
119 areq->cryptlen, areq->iv);
147 areq->cryptlen,
179 len = areq->cryptlen;
189 areq->cryptlen, i, rctx->t_src[i].len, sg->offset, todo);
196 dev_err(ce->dev, "remaining len %d/%u nr_sgs=%d\n", len, areq->cryptlen, nr_sgs);
201 len = areq->cryptlen;
211 areq->cryptlen, i, rctx->t_dst[i].len, sg->offset, todo);
234 ecb->cipher.algorithm_len = areq->cryptlen;
[all...]
/linux-master/drivers/crypto/cavium/nitrox/
H A Dnitrox_aead.c165 creq->gph.param0 = cpu_to_be16(rctx->cryptlen);
166 creq->gph.param1 = cpu_to_be16(rctx->cryptlen + rctx->assoclen);
227 rctx->cryptlen = areq->cryptlen;
229 rctx->srclen = areq->assoclen + areq->cryptlen;
261 rctx->cryptlen = areq->cryptlen - aead->authsize;
263 rctx->srclen = areq->cryptlen + areq->assoclen;
449 aead_rctx->cryptlen = areq->cryptlen;
[all...]
/linux-master/drivers/crypto/stm32/
H A Dstm32-cryp.c528 return is_encrypt(cryp) ? cryp->areq->cryptlen :
529 cryp->areq->cryptlen - cryp->authsize;
965 if (req->cryptlen % AES_BLOCK_SIZE)
968 if (req->cryptlen == 0)
976 if (req->cryptlen % AES_BLOCK_SIZE)
979 if (req->cryptlen == 0)
987 if (req->cryptlen % AES_BLOCK_SIZE)
990 if (req->cryptlen == 0)
998 if (req->cryptlen % AES_BLOCK_SIZE)
1001 if (req->cryptlen
[all...]
/linux-master/arch/arm64/crypto/
H A Daes-glue.c282 int cbc_blocks = DIV_ROUND_UP(req->cryptlen, AES_BLOCK_SIZE) - 2;
292 if (req->cryptlen <= AES_BLOCK_SIZE) {
293 if (req->cryptlen < AES_BLOCK_SIZE)
308 if (req->cryptlen == AES_BLOCK_SIZE)
311 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen);
314 subreq.cryptlen);
319 req->cryptlen - cbc_blocks * AES_BLOCK_SIZE,
339 int cbc_blocks = DIV_ROUND_UP(req->cryptlen, AES_BLOCK_SIZE) - 2;
349 if (req->cryptlen <= AES_BLOCK_SIZE) {
350 if (req->cryptlen < AES_BLOCK_SIZ
[all...]
/linux-master/drivers/crypto/nx/
H A Dnx-aes-cbc.c71 to_process = req->cryptlen - processed;
95 } while (processed < req->cryptlen);

Completed in 215 milliseconds

12345