/linux-master/fs/nfs/ |
H A D | fscache.c | 277 static struct nfs_netfs_io_data *nfs_netfs_alloc(struct netfs_io_subrequest *sreq) argument 284 netfs->sreq = sreq; 289 static bool nfs_netfs_clamp_length(struct netfs_io_subrequest *sreq) argument 291 size_t rsize = NFS_SB(sreq->rreq->inode->i_sb)->rsize; 293 sreq->len = min(sreq->len, rsize); 297 static void nfs_netfs_issue_read(struct netfs_io_subrequest *sreq) argument 301 struct inode *inode = sreq->rreq->inode; 302 struct nfs_open_context *ctx = sreq 358 struct netfs_io_subrequest *sreq; local [all...] |
H A D | fscache.h | 46 struct netfs_io_subrequest *sreq; member in struct:nfs_netfs_io_data 77 final_len = min_t(s64, netfs->sreq->len, atomic64_read(&netfs->transferred)); 78 netfs_subreq_terminated(netfs->sreq, netfs->error ?: final_len, false);
|
/linux-master/include/trace/events/ |
H A D | netfs.h | 253 TP_PROTO(struct netfs_io_subrequest *sreq, 256 TP_ARGS(sreq, what), 271 __entry->rreq = sreq->rreq->debug_id; 272 __entry->index = sreq->debug_index; 273 __entry->error = sreq->error; 274 __entry->flags = sreq->flags; 275 __entry->source = sreq->source; 277 __entry->len = sreq->len; 278 __entry->transferred = sreq->transferred; 279 __entry->start = sreq [all...] |
/linux-master/drivers/crypto/marvell/cesa/ |
H A D | cipher.c | 86 struct mv_cesa_skcipher_std_req *sreq = &creq->std; local 88 size_t len = min_t(size_t, req->cryptlen - sreq->offset, 91 mv_cesa_adjust_op(engine, &sreq->op); 93 memcpy(engine->sram_pool, &sreq->op, sizeof(sreq->op)); 95 memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op)); 99 sreq->offset); 101 sreq->size = len; 102 mv_cesa_set_crypt_op_len(&sreq 127 struct mv_cesa_skcipher_std_req *sreq = &creq->std; local 179 struct mv_cesa_skcipher_std_req *sreq = &creq->std; local 411 struct mv_cesa_skcipher_std_req *sreq = &creq->std; local [all...] |
H A D | hash.c | 161 struct mv_cesa_ahash_std_req *sreq = &creq->req.std; local 178 if (!sreq->offset) { 194 len = min_t(size_t, req->nbytes + creq->cache_ptr - sreq->offset, 203 sreq->offset += mv_cesa_sg_copy_to_sram( 206 len - creq->cache_ptr, sreq->offset); 212 if (creq->last_req && sreq->offset == req->nbytes && 287 struct mv_cesa_ahash_std_req *sreq = &creq->req.std; local 289 if (sreq->offset < (req->nbytes - creq->cache_ptr)) 306 struct mv_cesa_ahash_std_req *sreq = &creq->req.std; local 308 sreq [all...] |
/linux-master/drivers/crypto/inside-secure/ |
H A D | safexcel_cipher.c | 507 struct safexcel_cipher_req *sreq, 529 (sreq->direction == SAFEXCEL_ENCRYPT ? 544 if (sreq->direction == SAFEXCEL_ENCRYPT && 549 else if (sreq->direction == SAFEXCEL_ENCRYPT) 559 if (sreq->direction == SAFEXCEL_ENCRYPT) 612 struct safexcel_cipher_req *sreq, 623 if (unlikely(!sreq->rdescs)) 626 while (sreq->rdescs--) { 644 if (sreq->nr_src > 0) 645 dma_unmap_sg(priv->dev, src, sreq 505 safexcel_context_control(struct safexcel_cipher_ctx *ctx, struct crypto_async_request *async, struct safexcel_cipher_req *sreq, struct safexcel_command_desc *cdesc) argument 607 safexcel_handle_req_result(struct safexcel_crypto_priv *priv, int ring, struct crypto_async_request *async, struct scatterlist *src, struct scatterlist *dst, unsigned int cryptlen, struct safexcel_cipher_req *sreq, bool *should_complete, int *ret) argument 673 safexcel_send_req(struct crypto_async_request *base, int ring, struct safexcel_cipher_req *sreq, struct scatterlist *src, struct scatterlist *dst, unsigned int cryptlen, unsigned int assoclen, unsigned int digestsize, u8 *iv, int *commands, int *results) argument 917 safexcel_handle_inv_result(struct safexcel_crypto_priv *priv, int ring, struct crypto_async_request *base, struct safexcel_cipher_req *sreq, bool *should_complete, int *ret) argument 982 struct safexcel_cipher_req *sreq = skcipher_request_ctx(req); local 1005 struct safexcel_cipher_req *sreq = aead_request_ctx(req); local 1044 struct safexcel_cipher_req *sreq = skcipher_request_ctx(req); local 1077 struct safexcel_cipher_req *sreq = aead_request_ctx(req); local 1094 safexcel_cipher_exit_inv(struct crypto_tfm *tfm, struct crypto_async_request *base, struct safexcel_cipher_req *sreq, struct crypto_wait *result) argument 1130 struct safexcel_cipher_req *sreq = skcipher_request_ctx(req); local 1145 struct safexcel_cipher_req *sreq = aead_request_ctx(req); local 1157 safexcel_queue_req(struct crypto_async_request *base, struct safexcel_cipher_req *sreq, enum safexcel_cipher_direction dir) argument [all...] |
H A D | safexcel_hash.c | 234 struct safexcel_ahash_req *sreq = ahash_request_ctx_dma(areq); local 251 if (sreq->nents) { 252 dma_unmap_sg(priv->dev, areq->src, sreq->nents, DMA_TO_DEVICE); 253 sreq->nents = 0; 256 if (sreq->result_dma) { 257 dma_unmap_single(priv->dev, sreq->result_dma, sreq->digest_sz, 259 sreq->result_dma = 0; 262 if (sreq->cache_dma) { 263 dma_unmap_single(priv->dev, sreq [all...] |
/linux-master/drivers/net/wireless/mediatek/mt76/ |
H A D | mt76_connac_mcu.c | 1654 struct cfg80211_scan_request *sreq = &scan_req->req; local 1656 int ext_channels_num = max_t(int, sreq->n_channels - 32, 0); 1657 struct ieee80211_channel **scan_list = sreq->channels; 1677 req->scan_type = sreq->n_ssids ? 1 : 0; 1678 req->probe_req_num = sreq->n_ssids ? 2 : 0; 1681 for (i = 0; i < sreq->n_ssids; i++) { 1682 if (!sreq->ssids[i].ssid_len) 1685 req->ssids[i].ssid_len = cpu_to_le32(sreq->ssids[i].ssid_len); 1686 memcpy(req->ssids[i].ssid, sreq->ssids[i].ssid, 1687 sreq 1781 mt76_connac_mcu_sched_scan_req(struct mt76_phy *phy, struct ieee80211_vif *vif, struct cfg80211_sched_scan_request *sreq) argument [all...] |
H A D | mt76_connac_mcu.h | 1938 struct cfg80211_sched_scan_request *sreq);
|
/linux-master/drivers/crypto/hisilicon/sec2/ |
H A D | sec_crypto.c | 2009 struct sec_req *sreq) 2011 u32 cryptlen = sreq->c_req.sk_req->cryptlen; 2043 static int sec_skcipher_param_check(struct sec_ctx *ctx, struct sec_req *sreq) argument 2045 struct skcipher_request *sk_req = sreq->c_req.sk_req; 2054 sreq->c_req.c_len = sk_req->cryptlen; 2057 sreq->use_pbuf = true; 2059 sreq->use_pbuf = false; 2068 return sec_skcipher_cryptlen_check(ctx, sreq); 2077 struct skcipher_request *sreq, bool encrypt) 2092 skcipher_request_set_callback(subreq, sreq 2008 sec_skcipher_cryptlen_check(struct sec_ctx *ctx, struct sec_req *sreq) argument 2076 sec_skcipher_soft_crypto(struct sec_ctx *ctx, struct skcipher_request *sreq, bool encrypt) argument 2230 sec_aead_spec_check(struct sec_ctx *ctx, struct sec_req *sreq) argument 2278 sec_aead_param_check(struct sec_ctx *ctx, struct sec_req *sreq) argument [all...] |
/linux-master/drivers/net/wireless/mediatek/mt76/mt7925/ |
H A D | mcu.c | 2389 struct cfg80211_scan_request *sreq = &scan_req->req; local 2391 struct ieee80211_channel **scan_list = sreq->channels; 2423 req->scan_type = sreq->n_ssids ? 1 : 0; 2424 req->probe_req_num = sreq->n_ssids ? 2 : 0; 2428 if (!sreq->n_ssids) 2430 req->timeout_value = cpu_to_le16(sreq->n_channels * duration); 2436 for (i = 0; i < sreq->n_ssids; i++) { 2437 if (!sreq->ssids[i].ssid_len) 2440 ssid->ssids[i].ssid_len = cpu_to_le32(sreq->ssids[i].ssid_len); 2441 memcpy(ssid->ssids[i].ssid, sreq 2501 mt7925_mcu_sched_scan_req(struct mt76_phy *phy, struct ieee80211_vif *vif, struct cfg80211_sched_scan_request *sreq) argument [all...] |
H A D | mcu.h | 575 struct cfg80211_sched_scan_request *sreq);
|
/linux-master/drivers/crypto/marvell/octeontx/ |
H A D | otx_cptvf_algs.c | 147 struct skcipher_request *sreq; local 153 sreq = container_of(areq, struct skcipher_request, base); 154 stfm = crypto_skcipher_reqtfm(sreq); 158 rctx = skcipher_request_ctx_dma(sreq); 161 start = sreq->cryptlen - ivsize; 164 scatterwalk_map_and_copy(sreq->iv, sreq->dst, start, 167 if (sreq->src != sreq->dst) { 168 scatterwalk_map_and_copy(sreq [all...] |
/linux-master/drivers/crypto/marvell/octeontx2/ |
H A D | otx2_cptvf_algs.c | 131 struct skcipher_request *sreq; local 136 sreq = container_of(areq, struct skcipher_request, base); 137 stfm = crypto_skcipher_reqtfm(sreq); 141 rctx = skcipher_request_ctx_dma(sreq); 144 start = sreq->cryptlen - ivsize; 147 scatterwalk_map_and_copy(sreq->iv, sreq->dst, start, 150 if (sreq->src != sreq->dst) { 151 scatterwalk_map_and_copy(sreq [all...] |
/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | qat_algs.c | 683 struct skcipher_request *sreq = qat_req->skcipher_req; local 688 memcpy(qat_req->iv, sreq->iv, AES_BLOCK_SIZE); 694 iv_lo += DIV_ROUND_UP(sreq->cryptlen, AES_BLOCK_SIZE); 704 struct skcipher_request *sreq = qat_req->skcipher_req; local 705 int offset = sreq->cryptlen - AES_BLOCK_SIZE; 709 sgl = sreq->dst; 711 sgl = sreq->src; 741 struct skcipher_request *sreq = qat_req->skcipher_req; local 752 memcpy(sreq->iv, qat_req->iv, AES_BLOCK_SIZE); 754 skcipher_request_complete(sreq, re [all...] |
/linux-master/include/linux/netfilter/ |
H A D | nf_conntrack_pptp.h | 288 struct PptpStartSessionRequest sreq; member in union:pptp_ctrl_union
|
/linux-master/drivers/crypto/starfive/ |
H A D | jh7110-aes.c | 294 starfive_aes_write_iv(ctx, (void *)cryp->req.sreq->iv); 342 starfive_aes_get_iv(cryp, (void *)cryp->req.sreq->iv); 352 crypto_finalize_skcipher_request(cryp->engine, cryp->req.sreq, 471 cryp->req.sreq = req;
|
H A D | jh7110-cryp.h | 206 struct skcipher_request *sreq; member in union:starfive_cryp_dev::__anon136
|
/linux-master/arch/s390/kernel/ |
H A D | perf_cpum_sf.c | 159 struct hws_lsctl_request_block sreq; local 161 memset(&sreq, 0, sizeof(sreq)); 162 return lsctl(&sreq);
|