/linux-master/drivers/crypto/nx/ |
H A D | nx-aes-xcbc.c | 63 struct nx_sg *in_sg, *out_sg; local 81 in_sg = nx_build_sg_list(nx_ctx->in_sg, (u8 *) keys, &len, 93 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); 108 in_sg = nx_build_sg_list(nx_ctx->in_sg, (u8 *) keys[1], &len, 121 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg); 172 struct nx_sg *in_sg; local 297 struct nx_sg *in_sg, *out_sg; local [all...] |
H A D | nx-sha512.c | 109 struct nx_sg *in_sg = nx_ctx->in_sg; local 113 in_sg = nx_build_sg_list(in_sg, 121 used_sgs = in_sg - nx_ctx->in_sg; 136 in_sg = nx_build_sg_list(in_sg, (u8 *) data, 139 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeo 189 struct nx_sg *in_sg, *out_sg; local [all...] |
H A D | nx-sha256.c | 109 struct nx_sg *in_sg = nx_ctx->in_sg; local 113 in_sg = nx_build_sg_list(in_sg, 122 used_sgs = in_sg - nx_ctx->in_sg; 137 in_sg = nx_build_sg_list(in_sg, (u8 *) data, 140 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeo 186 struct nx_sg *in_sg, *out_sg; local [all...] |
H A D | nx-aes-gcm.c | 107 struct nx_sg *nx_sg = nx_ctx->in_sg; 137 nx_sg = nx_walk_and_build(nx_ctx->in_sg, max_sg_len, 145 nx_ctx->op_aead.inlen = (nx_ctx->in_sg - nx_sg) 204 nx_sg = nx_walk_and_build(nx_ctx->in_sg, max_sg_len, 212 nx_ctx->op.inlen = (nx_ctx->in_sg - nx_sg) 249 struct nx_sg *in_sg, *out_sg; local 269 in_sg = nx_build_sg_list(nx_ctx->in_sg, (u8 *) iv, 282 nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeo [all...] |
H A D | nx.c | 269 struct nx_sg *nx_insg = nx_ctx->in_sg; 294 nx_ctx->op.inlen = trim_sg_list(nx_ctx->in_sg, nx_insg, delta, nbytes); 314 nx_ctx->op.in = __pa(nx_ctx->in_sg); 322 nx_ctx->op_aead.in = __pa(nx_ctx->in_sg); 677 nx_ctx->in_sg = (struct nx_sg *)((u8 *)nx_ctx->csbcpb + NX_PAGE_SIZE); 678 nx_ctx->out_sg = (struct nx_sg *)((u8 *)nx_ctx->in_sg + NX_PAGE_SIZE); 753 nx_ctx->in_sg = NULL;
|
H A D | nx-aes-ccm.c | 161 struct nx_sg *nx_insg = nx_ctx->in_sg; 249 nx_ctx->op.inlen = (nx_ctx->in_sg - nx_insg) * 282 nx_insg = nx_walk_and_build(nx_ctx->in_sg, 296 nx_ctx->op_aead.inlen = (nx_ctx->in_sg - nx_insg) *
|
H A D | nx.h | 126 struct nx_sg *in_sg; /* aligned pointer into kmem to an sg list */ member in struct:nx_crypto_ctx
|
/linux-master/lib/ |
H A D | sg_split.c | 80 struct scatterlist *in_sg, *out_sg; local 84 in_sg = split->in_sg0; 87 *out_sg = *in_sg; 96 in_sg = sg_next(in_sg); 106 struct scatterlist *in_sg, *out_sg; local 110 in_sg = split->in_sg0; 113 sg_dma_address(out_sg) = sg_dma_address(in_sg); 114 sg_dma_len(out_sg) = sg_dma_len(in_sg); 119 in_sg [all...] |
/linux-master/drivers/crypto/ |
H A D | omap-des.c | 147 struct scatterlist *in_sg; member in struct:omap_des_dev 372 struct scatterlist *in_sg, struct scatterlist *out_sg, 382 scatterwalk_start(&dd->in_walk, dd->in_sg); 391 dma_sync_sg_for_device(dd->dev, dd->in_sg, in_sg_len, DMA_TO_DEVICE); 410 tx_in = dmaengine_prep_slave_sg(dd->dma_lch_in, in_sg, in_sg_len, 461 err = dma_map_sg(dd->dev, dd->in_sg, dd->in_sg_len, 476 err = omap_des_crypt_dma(tfm, dd->in_sg, dd->out_sg, dd->in_sg_len, 479 dma_unmap_sg(dd->dev, dd->in_sg, dd->in_sg_len, DMA_TO_DEVICE); 533 dd->in_sg = req->src; 541 ret = omap_crypto_align_sg(&dd->in_sg, d 371 omap_des_crypt_dma(struct crypto_tfm *tfm, struct scatterlist *in_sg, struct scatterlist *out_sg, int in_sg_len, int out_sg_len) argument [all...] |
H A D | omap-aes.c | 266 struct scatterlist *in_sg, 275 scatterwalk_start(&dd->in_walk, dd->in_sg); 285 dma_sync_sg_for_device(dd->dev, dd->in_sg, in_sg_len, DMA_TO_DEVICE); 304 tx_in = dmaengine_prep_slave_sg(dd->dma_lch_in, in_sg, in_sg_len, 367 err = dma_map_sg(dd->dev, dd->in_sg, dd->in_sg_len, 384 err = omap_aes_crypt_dma(dd, dd->in_sg, dd->out_sg, dd->in_sg_len, 387 dma_unmap_sg(dd->dev, dd->in_sg, dd->in_sg_len, DMA_TO_DEVICE); 440 dd->in_sg = req->src; 448 ret = omap_crypto_align_sg(&dd->in_sg, dd->total, AES_BLOCK_SIZE, 460 dd->in_sg_len = sg_nents_for_len(dd->in_sg, d 265 omap_aes_crypt_dma(struct omap_aes_dev *dd, struct scatterlist *in_sg, struct scatterlist *out_sg, int in_sg_len, int out_sg_len) argument [all...] |
H A D | atmel-tdes.c | 105 struct scatterlist *in_sg; member in struct:atmel_tdes_dev 295 dma_unmap_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); 460 in = IS_ALIGNED((u32)dd->in_sg->offset, sizeof(u32)) && 461 IS_ALIGNED(dd->in_sg->length, dd->ctx->block_size); 466 if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg)) 472 count = min_t(size_t, dd->total, sg_dma_len(dd->in_sg)); 475 err = dma_map_sg(dd->dev, dd->in_sg, 1, DMA_TO_DEVICE); 485 dma_unmap_sg(dd->dev, dd->in_sg, 1, 490 addr_in = sg_dma_address(dd->in_sg); 497 count = atmel_tdes_sg_copy(&dd->in_sg, [all...] |
H A D | sahara.c | 163 * @in_sg: scatterlist for input data 178 struct scatterlist *in_sg; member in struct:sahara_sha_reqctx 212 struct scatterlist *in_sg; member in struct:sahara_dev 458 dev->nb_in_sg = sg_nents_for_len(dev->in_sg, dev->total); 474 ret = dma_map_sg(dev->device, dev->in_sg, dev->nb_in_sg, 490 sg = dev->in_sg; 534 dma_unmap_sg(dev->device, dev->in_sg, dev->nb_in_sg, 571 dev->in_sg = req->src; 605 dma_unmap_sg(dev->device, dev->in_sg, dev->nb_in_sg, 765 dev->in_sg [all...] |
H A D | omap-aes-gcm.c | 36 dd->in_sg = NULL; 59 dma_unmap_sg(dd->dev, dd->in_sg, dd->in_sg_len, DMA_TO_DEVICE); 148 dd->in_sg = dd->in_sgl; 171 dd->in_sg_len = sg_nents_for_len(dd->in_sg, alen + clen);
|
H A D | omap-aes.h | 181 struct scatterlist *in_sg; member in struct:omap_aes_dev
|
/linux-master/drivers/crypto/gemini/ |
H A D | sl3516-ce-cipher.c | 30 struct scatterlist *in_sg; local 87 in_sg = areq->src; 89 while (in_sg && out_sg) { 90 if (in_sg->length != out_sg->length) { 94 in_sg = sg_next(in_sg); 97 if (in_sg || out_sg)
|
/linux-master/drivers/crypto/allwinner/sun8i-ss/ |
H A D | sun8i-ss-cipher.c | 28 struct scatterlist *in_sg = areq->src; local 76 in_sg = areq->src; 78 while (in_sg && out_sg) { 79 if (in_sg->length != out_sg->length) 81 in_sg = sg_next(in_sg); 84 if (in_sg || out_sg)
|
/linux-master/drivers/crypto/allwinner/sun4i-ss/ |
H A D | sun4i-ss-hash.c | 198 struct scatterlist *in_sg = areq->src; local 263 while (in_sg && i == 1) { 264 if (in_sg->length % 4) 266 in_sg = sg_next(in_sg);
|
H A D | sun4i-ss-cipher.c | 182 struct scatterlist *in_sg = areq->src; local 225 while (in_sg && no_chunk == 1) { 226 if ((in_sg->length | in_sg->offset) & 3u) 228 in_sg = sg_next(in_sg);
|
/linux-master/drivers/crypto/starfive/ |
H A D | jh7110-hash.c | 96 dma_unmap_sg(cryp->dev, rctx->in_sg, rctx->in_sg_len, DMA_TO_DEVICE); 131 sg_dma_len(rctx->in_sg) = total_len; 139 ret = dma_map_sg(cryp->dev, rctx->in_sg, rctx->in_sg_len, DMA_TO_DEVICE); 152 in_desc = dmaengine_prep_slave_sg(cryp->tx, rctx->in_sg, 367 rctx->in_sg = req->src; 370 rctx->in_sg_len = sg_nents_for_len(rctx->in_sg, rctx->total); 373 if (starfive_hash_check_aligned(rctx->in_sg, rctx->total, rctx->blksize))
|
H A D | jh7110-rsa.c | 255 rctx->total = sg_copy_to_buffer(rctx->in_sg, rctx->nents, 303 rctx->in_sg = req->src; 306 rctx->nents = sg_nents(rctx->in_sg); 335 rctx->in_sg = req->src;
|
H A D | jh7110-cryp.h | 217 struct scatterlist *in_sg; member in struct:starfive_cryp_request_ctx
|
/linux-master/crypto/ |
H A D | rsa-pkcs1pad.c | 125 struct scatterlist in_sg[2], out_sg[1]; member in struct:pkcs1pad_request 279 pkcs1pad_sg_set_buf(req_ctx->in_sg, req_ctx->in_buf, 287 akcipher_request_set_crypt(&req_ctx->child_req, req_ctx->in_sg, 436 pkcs1pad_sg_set_buf(req_ctx->in_sg, req_ctx->in_buf, 444 akcipher_request_set_crypt(&req_ctx->child_req, req_ctx->in_sg,
|
/linux-master/arch/um/drivers/ |
H A D | virt-pci.c | 77 struct scatterlist out_sg, extra_sg, in_sg; local 80 [1] = extra ? &extra_sg : &in_sg, 81 [2] = extra ? &in_sg : NULL, 132 sg_init_one(&in_sg, out, out_size);
|
/linux-master/net/vmw_vsock/ |
H A D | virtio_transport.c | 114 int ret, in_sg = 0, out_sg = 0; local 163 ret = virtqueue_add_sgs(vq, sgs, out_sg, in_sg, skb, GFP_KERNEL);
|
/linux-master/drivers/crypto/aspeed/ |
H A D | aspeed-hace-crypto.c | 185 struct scatterlist *in_sg; local 190 in_sg = req->src; 192 nbytes = sg_copy_to_buffer(in_sg, rctx->src_nents,
|