/linux-master/net/rxrpc/ |
H A D | rxkad.c | 235 struct crypto_skcipher *tfm = &call->conn->rxkad.cipher->base; local 237 return skcipher_request_alloc(tfm, GFP_NOFS); 1216 struct crypto_sync_skcipher *tfm; local 1221 tfm = crypto_alloc_sync_skcipher("pcbc(fcrypt)", 0, 0); 1222 if (IS_ERR(tfm)) 1223 return PTR_ERR(tfm); 1225 req = skcipher_request_alloc(&tfm->base, GFP_KERNEL); 1230 rxkad_ci = tfm; 1234 crypto_free_sync_skcipher(tfm);
|
/linux-master/drivers/nvme/target/ |
H A D | tcp.c | 856 struct crypto_ahash *tfm = crypto_ahash_reqtfm(queue->rcv_hash); local 860 crypto_free_ahash(tfm); 865 struct crypto_ahash *tfm; local 867 tfm = crypto_alloc_ahash("crc32c", 0, CRYPTO_ALG_ASYNC); 868 if (IS_ERR(tfm)) 869 return PTR_ERR(tfm); 871 queue->snd_hash = ahash_request_alloc(tfm, GFP_KERNEL); 876 queue->rcv_hash = ahash_request_alloc(tfm, GFP_KERNEL); 885 crypto_free_ahash(tfm);
|
H A D | auth.c | 332 shash->tfm = shash_tfm; 441 shash->tfm = shash_tfm;
|
/linux-master/drivers/nvme/host/ |
H A D | tcp.c | 1303 struct crypto_ahash *tfm = crypto_ahash_reqtfm(queue->rcv_hash); local 1307 crypto_free_ahash(tfm); 1312 struct crypto_ahash *tfm; local 1314 tfm = crypto_alloc_ahash("crc32c", 0, CRYPTO_ALG_ASYNC); 1315 if (IS_ERR(tfm)) 1316 return PTR_ERR(tfm); 1318 queue->snd_hash = ahash_request_alloc(tfm, GFP_KERNEL); 1323 queue->rcv_hash = ahash_request_alloc(tfm, GFP_KERNEL); 1332 crypto_free_ahash(tfm);
|
/linux-master/arch/s390/crypto/ |
H A D | paes_s390.c | 160 static int ecb_paes_init(struct crypto_skcipher *tfm) argument 162 struct s390_paes_ctx *ctx = crypto_skcipher_ctx(tfm); 170 static void ecb_paes_exit(struct crypto_skcipher *tfm) argument 172 struct s390_paes_ctx *ctx = crypto_skcipher_ctx(tfm); 197 static int ecb_paes_set_key(struct crypto_skcipher *tfm, const u8 *in_key, argument 201 struct s390_paes_ctx *ctx = crypto_skcipher_ctx(tfm); 213 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req); local 214 struct s390_paes_ctx *ctx = crypto_skcipher_ctx(tfm); 275 static int cbc_paes_init(struct crypto_skcipher *tfm) argument 277 struct s390_paes_ctx *ctx = crypto_skcipher_ctx(tfm); 285 cbc_paes_exit(struct crypto_skcipher *tfm) argument 312 cbc_paes_set_key(struct crypto_skcipher *tfm, const u8 *in_key, unsigned int key_len) argument 328 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req); local 395 xts_paes_init(struct crypto_skcipher *tfm) argument 406 xts_paes_exit(struct crypto_skcipher *tfm) argument 454 xts_paes_set_key(struct crypto_skcipher *tfm, const u8 *in_key, unsigned int xts_key_len) argument 494 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req); local 575 ctr_paes_init(struct crypto_skcipher *tfm) argument 585 ctr_paes_exit(struct crypto_skcipher *tfm) argument 613 ctr_paes_set_key(struct crypto_skcipher *tfm, const u8 *in_key, unsigned int key_len) argument 644 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req); local [all...] |
/linux-master/drivers/dma/idxd/ |
H A D | idxd.h | 392 struct crypto_tfm *tfm; member in struct:crypto_ctx
|
/linux-master/net/ipv4/ |
H A D | tcp_ao.c | 1314 struct crypto_ahash *tfm; local 1374 tfm = crypto_ahash_reqtfm(hp.req); 1384 err = crypto_ahash_setkey(tfm, scratch, 16); 1403 err = crypto_ahash_setkey(tfm, key->key, key->keylen); 1525 struct crypto_ahash *tfm; local 1549 tfm = crypto_ahash_reqtfm(hp.req); 1550 digest_size = crypto_ahash_digestsize(tfm);
|
/linux-master/include/net/ |
H A D | macsec.h | 65 * @tfm: crypto struct, key storage 70 struct crypto_aead *tfm; member in struct:macsec_key
|
/linux-master/drivers/soc/mediatek/ |
H A D | mtk-svs.c | 1918 struct svs_fusemap tfm = { 0, 24 }; local 1927 val = svs_get_fuse_val(svsp->tefuse, &tfm, 8);
|
/linux-master/drivers/net/ |
H A D | macsec.c | 193 crypto_free_aead(sa->key.tfm); 221 crypto_free_aead(sa->key.tfm); 551 static struct aead_request *macsec_alloc_req(struct crypto_aead *tfm, argument 560 size = sizeof(struct aead_request) + crypto_aead_reqsize(tfm); 576 aead_request_set_tfm(req, tfm); 668 req = macsec_alloc_req(tx_sa->key.tfm, &iv, &sg, ret); 894 req = macsec_alloc_req(rx_sa->key.tfm, &iv, &sg, ret); 1359 struct crypto_aead *tfm; local 1362 tfm = crypto_alloc_aead("gcm(aes)", 0, 0); 1364 if (IS_ERR(tfm)) [all...] |
/linux-master/fs/btrfs/ |
H A D | scrub.c | 649 shash->tfm = fs_info->csum_shash;
|
H A D | inode.c | 3300 shash->tfm = fs_info->csum_shash;
|
/linux-master/fs/smb/client/ |
H A D | smb2transport.c | 263 rc = crypto_shash_setkey(shash->tfm, ses->auth_key.response, 332 rc = crypto_shash_setkey(server->secmech.hmacsha256->tfm, 590 rc = crypto_shash_setkey(shash->tfm, key, SMB2_CMACAES_SIZE); 916 struct crypto_aead *tfm; local 921 tfm = crypto_alloc_aead("gcm(aes)", 0, 0); 923 tfm = crypto_alloc_aead("ccm(aes)", 0, 0); 924 if (IS_ERR(tfm)) { 927 return PTR_ERR(tfm); 929 server->secmech.enc = tfm; 935 tfm [all...] |
H A D | smb2ops.c | 4102 static void *smb2_aead_req_alloc(struct crypto_aead *tfm, const struct smb_rqst *rqst, argument 4107 unsigned int req_size = sizeof(**req) + crypto_aead_reqsize(tfm); 4108 unsigned int iv_size = crypto_aead_ivsize(tfm); 4117 len += crypto_aead_alignmask(tfm) & ~(crypto_tfm_ctx_alignment() - 1); 4128 *iv = (u8 *)PTR_ALIGN(p, crypto_aead_alignmask(tfm) + 1); 4136 static void *smb2_get_aead_req(struct crypto_aead *tfm, struct smb_rqst *rqst, argument 4146 p = smb2_aead_req_alloc(tfm, rqst, num_rqst, sig, iv, req, &sgtable, 4234 struct crypto_aead *tfm; local 4252 tfm = enc ? server->secmech.enc : server->secmech.dec; 4256 rc = crypto_aead_setkey(tfm, ke [all...] |
H A D | misc.c | 1084 (*sdesc)->tfm = alg; 1100 if ((*sdesc)->tfm) { 1101 crypto_free_shash((*sdesc)->tfm); 1102 (*sdesc)->tfm = NULL;
|
/linux-master/arch/powerpc/crypto/ |
H A D | chacha-p10-glue.c | 126 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req); local 127 struct chacha_ctx *ctx = crypto_skcipher_ctx(tfm); 134 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req); local 135 struct chacha_ctx *ctx = crypto_skcipher_ctx(tfm);
|
/linux-master/drivers/infiniband/sw/rxe/ |
H A D | rxe.c | 34 if (rxe->tfm) 35 crypto_free_shash(rxe->tfm);
|
/linux-master/fs/bcachefs/ |
H A D | checksum.c | 97 static inline int do_encrypt_sg(struct crypto_sync_skcipher *tfm, argument 101 SYNC_SKCIPHER_REQUEST_ON_STACK(req, tfm); 104 skcipher_request_set_sync_tfm(req, tfm); 114 static inline int do_encrypt(struct crypto_sync_skcipher *tfm, argument 127 return do_encrypt_sg(tfm, nonce, &sg, len); 149 ret = do_encrypt_sg(tfm, nonce, sg, orig_len); 194 desc->tfm = c->poly1305;
|
/linux-master/drivers/md/ |
H A D | dm-integrity.c | 507 desc->tfm = ic->journal_mac; 786 desc->tfm = ic->journal_mac; 1649 req->tfm = ic->internal_hash;
|
/linux-master/net/tls/ |
H A D | tls_sw.c | 2699 struct crypto_tfm *tfm; local 2764 tfm = crypto_aead_tfm(sw_ctx_rx->aead_recv); 2769 !!(tfm->__crt_alg->cra_flags & CRYPTO_ALG_ASYNC);
|
/linux-master/net/sunrpc/auth_gss/ |
H A D | gss_krb5_crypto.c | 88 * @tfm: initialized cipher transform 104 struct crypto_sync_skcipher *tfm, 113 SYNC_SKCIPHER_REQUEST_ON_STACK(req, tfm); 115 if (length % crypto_sync_skcipher_blocksize(tfm) != 0) 118 if (crypto_sync_skcipher_ivsize(tfm) > GSS_KRB5_MAX_BLOCKSIZE) { 119 dprintk("RPC: gss_k5encrypt: tfm iv size too large %d\n", 120 crypto_sync_skcipher_ivsize(tfm)); 125 memcpy(local_iv, iv, crypto_sync_skcipher_ivsize(tfm)); 130 skcipher_request_set_sync_tfm(req, tfm); 143 * @tfm 103 krb5_encrypt( struct crypto_sync_skcipher *tfm, void * iv, void * in, void * out, int length) argument 158 krb5_decrypt( struct crypto_sync_skcipher *tfm, void * iv, void * in, void * out, int length) argument 215 struct crypto_ahash *tfm; local 314 gss_krb5_checksum(struct crypto_ahash *tfm, char *header, int hdrlen, const struct xdr_buf *body, int body_offset, struct xdr_netobj *cksumout) argument 385 struct crypto_sync_skcipher *tfm = local 446 gss_encrypt_xdr_buf(struct crypto_sync_skcipher *tfm, struct xdr_buf *buf, int offset, struct page **pages) argument 487 struct crypto_sync_skcipher *tfm = local 529 gss_decrypt_xdr_buf(struct crypto_sync_skcipher *tfm, struct xdr_buf *buf, int offset) argument 929 krb5_etm_checksum(struct crypto_sync_skcipher *cipher, struct crypto_ahash *tfm, const struct xdr_buf *body, int body_offset, struct xdr_netobj *cksumout) argument [all...] |
/linux-master/drivers/crypto/intel/iaa/ |
H A D | iaa_crypto_main.c | 1017 static int iaa_compress_verify(struct crypto_tfm *tfm, struct acomp_req *req, 1038 compression_ctx = crypto_tfm_ctx(ctx->tfm); 1099 ret = iaa_compress_verify(ctx->tfm, ctx->req, iaa_wq->wq, src_addr, 1127 static int iaa_compress(struct crypto_tfm *tfm, struct acomp_req *req, argument 1135 struct iaa_compression_ctx *ctx = crypto_tfm_ctx(tfm); 1179 idxd_desc->crypto.tfm = tfm; 1280 static int iaa_compress_verify(struct crypto_tfm *tfm, struct acomp_req *req, argument 1287 struct iaa_compression_ctx *ctx = crypto_tfm_ctx(tfm); 1367 static int iaa_decompress(struct crypto_tfm *tfm, struc argument 1493 struct crypto_tfm *tfm = req->base.tfm; local 1619 struct crypto_tfm *tfm = req->base.tfm; local 1711 struct crypto_tfm *tfm = req->base.tfm; local 1804 struct crypto_tfm *tfm = crypto_acomp_tfm(acomp_tfm); local [all...] |
/linux-master/crypto/asymmetric_keys/ |
H A D | public_key.c | 163 struct crypto_akcipher *tfm; local 214 tfm = crypto_alloc_akcipher(alg_name, 0, 0); 215 if (IS_ERR(tfm)) { 216 ret = PTR_ERR(tfm); 221 ret = crypto_akcipher_set_priv_key(tfm, key, pkey->keylen); 223 ret = crypto_akcipher_set_pub_key(tfm, key, pkey->keylen); 227 len = crypto_akcipher_maxsize(tfm); 266 crypto_free_akcipher(tfm); 281 struct crypto_akcipher *tfm; local 323 tfm 386 struct crypto_sig *tfm; local [all...] |
/linux-master/drivers/thunderbolt/ |
H A D | domain.c | 711 struct crypto_shash *tfm; local 728 tfm = crypto_alloc_shash("hmac(sha256)", 0, 0); 729 if (IS_ERR(tfm)) 730 return PTR_ERR(tfm); 732 ret = crypto_shash_setkey(tfm, sw->key, TB_SWITCH_KEY_SIZE); 736 shash = kzalloc(sizeof(*shash) + crypto_shash_descsize(tfm), 743 shash->tfm = tfm; 756 crypto_free_shash(tfm); 764 crypto_free_shash(tfm); [all...] |
/linux-master/drivers/base/firmware_loader/ |
H A D | main.c | 829 shash->tfm = alg;
|