Lines Matching defs:areq_ctx

211 	struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(areq);
222 areq->iv = areq_ctx->backup_iv;
227 if (areq_ctx->gen_ctx.op_type == DRV_CRYPTO_DIRECTION_DECRYPT) {
228 if (memcmp(areq_ctx->mac_buf, areq_ctx->icv_virt_addr,
240 } else if (areq_ctx->is_icv_fragmented) {
241 u32 skip = areq->cryptlen + areq_ctx->dst_offset;
243 cc_copy_sg_portion(dev, areq_ctx->mac_buf, areq_ctx->dst_sgl,
726 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(areq);
727 enum cc_req_dma_buf_type assoc_dma_type = areq_ctx->assoc_buff_type;
736 areq_ctx->assoclen, NS_BIT);
739 areq_ctx->cryptlen > 0)
745 set_din_type(&desc[idx], DMA_MLLI, areq_ctx->assoc.sram_addr,
746 areq_ctx->assoc.mlli_nents, NS_BIT);
749 areq_ctx->cryptlen > 0)
765 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(areq);
766 enum cc_req_dma_buf_type data_dma_type = areq_ctx->data_buff_type;
777 areq_ctx->dst_sgl : areq_ctx->src_sgl;
781 areq_ctx->dst_offset : areq_ctx->src_offset;
786 areq_ctx->cryptlen, NS_BIT);
796 u32 mlli_addr = areq_ctx->assoc.sram_addr;
797 u32 mlli_nents = areq_ctx->assoc.mlli_nents;
799 if (areq_ctx->is_single_pass) {
801 mlli_addr = areq_ctx->dst.sram_addr;
802 mlli_nents = areq_ctx->dst.mlli_nents;
804 mlli_addr = areq_ctx->src.sram_addr;
805 mlli_nents = areq_ctx->src.mlli_nents;
830 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(areq);
831 enum cc_req_dma_buf_type data_dma_type = areq_ctx->data_buff_type;
836 if (areq_ctx->cryptlen == 0)
844 (sg_dma_address(areq_ctx->src_sgl) +
845 areq_ctx->src_offset), areq_ctx->cryptlen,
848 (sg_dma_address(areq_ctx->dst_sgl) +
849 areq_ctx->dst_offset),
850 areq_ctx->cryptlen, NS_BIT, 0);
856 set_din_type(&desc[idx], DMA_MLLI, areq_ctx->src.sram_addr,
857 areq_ctx->src.mlli_nents, NS_BIT);
858 set_dout_mlli(&desc[idx], areq_ctx->dst.sram_addr,
859 areq_ctx->dst.mlli_nents, NS_BIT, 0);
1085 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(req);
1089 if (areq_ctx->assoclen > 0)
1315 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(req);
1317 unsigned int assoclen = areq_ctx->assoclen;
1325 areq_ctx->is_single_pass = true; /*defaulted to fast flow*/
1335 if (areq_ctx->plaintext_authenticate_only)
1336 areq_ctx->is_single_pass = false;
1341 areq_ctx->is_single_pass = false;
1345 areq_ctx->is_single_pass = false;
1352 areq_ctx->is_single_pass = false;
1594 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(req);
1597 memset(areq_ctx->ctr_iv, 0, AES_BLOCK_SIZE);
1601 areq_ctx->ctr_iv[0] = 3;
1606 memcpy(areq_ctx->ctr_iv + CCM_BLOCK_NONCE_OFFSET, ctx->ctr_nonce,
1608 memcpy(areq_ctx->ctr_iv + CCM_BLOCK_IV_OFFSET, req->iv,
1610 req->iv = areq_ctx->ctr_iv;
1882 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(req);
1884 memcpy(areq_ctx->ctr_iv + GCM_BLOCK_RFC4_NONCE_OFFSET,
1886 memcpy(areq_ctx->ctr_iv + GCM_BLOCK_RFC4_IV_OFFSET, req->iv,
1888 req->iv = areq_ctx->ctr_iv;
1899 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(req);
1913 req->cryptlen, areq_ctx->assoclen);
1922 areq_ctx->gen_ctx.op_type = direct;
1923 areq_ctx->req_authsize = ctx->authsize;
1924 areq_ctx->cipher_mode = ctx->cipher_mode;
1932 memcpy(areq_ctx->ctr_iv, ctx->ctr_nonce,
1934 memcpy(areq_ctx->ctr_iv + CTR_RFC3686_NONCE_SIZE, req->iv,
1937 *(__be32 *)(areq_ctx->ctr_iv + CTR_RFC3686_NONCE_SIZE +
1941 req->iv = areq_ctx->ctr_iv;
1942 areq_ctx->hw_iv_size = CTR_RFC3686_BLOCK_SIZE;
1945 areq_ctx->hw_iv_size = AES_BLOCK_SIZE;
1946 if (areq_ctx->ctr_iv != req->iv) {
1947 memcpy(areq_ctx->ctr_iv, req->iv,
1949 req->iv = areq_ctx->ctr_iv;
1952 areq_ctx->hw_iv_size = crypto_aead_ivsize(tfm);
1963 areq_ctx->ccm_hdr_size = ccm_header_size_null;
2022 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(req);
2025 memset(areq_ctx, 0, sizeof(*areq_ctx));
2028 areq_ctx->backup_iv = req->iv;
2029 areq_ctx->assoclen = req->assoclen;
2033 req->iv = areq_ctx->backup_iv;
2042 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(req);
2049 memset(areq_ctx, 0, sizeof(*areq_ctx));
2052 areq_ctx->backup_iv = req->iv;
2053 areq_ctx->assoclen = req->assoclen - CCM_BLOCK_IV_SIZE;
2059 req->iv = areq_ctx->backup_iv;
2066 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(req);
2069 memset(areq_ctx, 0, sizeof(*areq_ctx));
2072 areq_ctx->backup_iv = req->iv;
2073 areq_ctx->assoclen = req->assoclen;
2077 req->iv = areq_ctx->backup_iv;
2084 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(req);
2091 memset(areq_ctx, 0, sizeof(*areq_ctx));
2094 areq_ctx->backup_iv = req->iv;
2095 areq_ctx->assoclen = req->assoclen - CCM_BLOCK_IV_SIZE;
2101 req->iv = areq_ctx->backup_iv;
2196 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(req);
2203 memset(areq_ctx, 0, sizeof(*areq_ctx));
2206 areq_ctx->backup_iv = req->iv;
2207 areq_ctx->assoclen = req->assoclen - GCM_BLOCK_RFC4_IV_SIZE;
2213 req->iv = areq_ctx->backup_iv;
2220 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(req);
2227 memset(areq_ctx, 0, sizeof(*areq_ctx));
2230 areq_ctx->plaintext_authenticate_only = true;
2233 areq_ctx->backup_iv = req->iv;
2234 areq_ctx->assoclen = req->assoclen;
2240 req->iv = areq_ctx->backup_iv;
2247 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(req);
2254 memset(areq_ctx, 0, sizeof(*areq_ctx));
2257 areq_ctx->backup_iv = req->iv;
2258 areq_ctx->assoclen = req->assoclen - GCM_BLOCK_RFC4_IV_SIZE;
2264 req->iv = areq_ctx->backup_iv;
2271 struct aead_req_ctx *areq_ctx = aead_request_ctx_dma(req);
2278 memset(areq_ctx, 0, sizeof(*areq_ctx));
2281 areq_ctx->plaintext_authenticate_only = true;
2284 areq_ctx->backup_iv = req->iv;
2285 areq_ctx->assoclen = req->assoclen;
2291 req->iv = areq_ctx->backup_iv;