/linux-master/fs/netfs/ |
H A D | io.c | 39 static void netfs_read_from_cache(struct netfs_io_request *rreq, argument 43 struct netfs_cache_resources *cres = &rreq->cache_resources; 53 static void netfs_fill_with_zeroes(struct netfs_io_request *rreq, argument 77 static void netfs_read_from_server(struct netfs_io_request *rreq, argument 82 if (rreq->origin != NETFS_DIO_READ && 85 rreq->debug_id, subreq->debug_index, 88 rreq->netfs_ops->issue_read(subreq); 94 static void netfs_rreq_completed(struct netfs_io_request *rreq, bool was_async) argument 96 trace_netfs_rreq(rreq, netfs_rreq_trace_done); 97 netfs_clear_subrequests(rreq, was_asyn 107 netfs_rreq_unmark_after_write(struct netfs_io_request *rreq, bool was_async) argument 144 struct netfs_io_request *rreq = subreq->rreq; local 167 netfs_rreq_do_write_to_cache(struct netfs_io_request *rreq) argument 227 struct netfs_io_request *rreq = local 233 netfs_rreq_write_to_cache(struct netfs_io_request *rreq) argument 243 netfs_rreq_short_read(struct netfs_io_request *rreq, struct netfs_io_subrequest *subreq) argument 263 netfs_reset_subreq_iter(struct netfs_io_request *rreq, struct netfs_io_subrequest *subreq) argument 288 netfs_rreq_perform_resubmissions(struct netfs_io_request *rreq) argument 330 netfs_rreq_is_still_valid(struct netfs_io_request *rreq) argument 349 netfs_rreq_assess_dio(struct netfs_io_request *rreq) argument 389 netfs_rreq_assess(struct netfs_io_request *rreq, bool was_async) argument 420 struct netfs_io_request *rreq = local 429 netfs_rreq_terminated(struct netfs_io_request *rreq, bool was_async) argument 463 struct netfs_io_request *rreq = subreq->rreq; local 553 struct netfs_io_request *rreq = subreq->rreq; local 567 netfs_rreq_prepare_read(struct netfs_io_request *rreq, struct netfs_io_subrequest *subreq, struct iov_iter *io_iter) argument 645 netfs_rreq_submit_slice(struct netfs_io_request *rreq, struct iov_iter *io_iter, unsigned int *_debug_index) argument 707 netfs_begin_read(struct netfs_io_request *rreq, bool sync) argument [all...] |
H A D | objects.c | 22 struct netfs_io_request *rreq; local 29 rreq = kzalloc(ctx->ops->io_request_size ?: sizeof(struct netfs_io_request), 31 if (!rreq) 34 rreq->start = start; 35 rreq->len = len; 36 rreq->upper_len = len; 37 rreq->origin = origin; 38 rreq->netfs_ops = ctx->ops; 39 rreq->mapping = mapping; 40 rreq 66 netfs_get_request(struct netfs_io_request *rreq, enum netfs_rreq_ref_trace what) argument 74 netfs_clear_subrequests(struct netfs_io_request *rreq, bool was_async) argument 89 struct netfs_io_request *rreq = local 113 netfs_put_request(struct netfs_io_request *rreq, bool was_async, enum netfs_rreq_ref_trace what) argument 139 netfs_alloc_subrequest(struct netfs_io_request *rreq) argument 171 struct netfs_io_request *rreq = subreq->rreq; local [all...] |
H A D | buffered_read.c | 16 void netfs_rreq_unlock_folios(struct netfs_io_request *rreq) argument 21 pgoff_t start_page = rreq->start / PAGE_SIZE; 22 pgoff_t last_page = ((rreq->start + rreq->len) / PAGE_SIZE) - 1; 26 XA_STATE(xas, &rreq->mapping->i_pages, start_page); 28 if (test_bit(NETFS_RREQ_FAILED, &rreq->flags)) { 29 __clear_bit(NETFS_RREQ_COPY_TO_CACHE, &rreq->flags); 30 list_for_each_entry(subreq, &rreq->subrequests, rreq_link) { 41 subreq = list_first_entry(&rreq->subrequests, 45 trace_netfs_rreq(rreq, netfs_rreq_trace_unloc 118 netfs_cache_expand_readahead(struct netfs_io_request *rreq, loff_t *_start, size_t *_len, loff_t i_size) argument 127 netfs_rreq_expand(struct netfs_io_request *rreq, struct readahead_control *ractl) argument 165 netfs_begin_cache_read(struct netfs_io_request *rreq, struct netfs_inode *ctx) argument 187 struct netfs_io_request *rreq; local 250 struct netfs_io_request *rreq; local 419 struct netfs_io_request *rreq; local 520 struct netfs_io_request *rreq; local [all...] |
H A D | direct_read.c | 31 struct netfs_io_request *rreq; local 46 rreq = netfs_alloc_request(iocb->ki_filp->f_mapping, iocb->ki_filp, 49 if (IS_ERR(rreq)) 50 return PTR_ERR(rreq); 53 trace_netfs_read(rreq, rreq->start, rreq->len, netfs_read_trace_dio_read); 65 ret = netfs_extract_user_iter(iter, rreq->len, &rreq->iter, 0); 68 rreq [all...] |
H A D | main.c | 47 struct netfs_io_request *rreq; local 57 rreq = list_entry(v, struct netfs_io_request, proc_link); 60 rreq->debug_id, 61 netfs_origins[rreq->origin], 62 refcount_read(&rreq->ref), 63 rreq->flags, 64 rreq->error, 65 atomic_read(&rreq->nr_outstanding), 66 rreq->start, rreq [all...] |
H A D | internal.h | 25 void netfs_rreq_unlock_folios(struct netfs_io_request *rreq); 32 int netfs_begin_read(struct netfs_io_request *rreq, bool sync); 42 static inline void netfs_proc_add_rreq(struct netfs_io_request *rreq) argument 45 list_add_tail_rcu(&rreq->proc_link, &netfs_io_requests); 48 static inline void netfs_proc_del_rreq(struct netfs_io_request *rreq) argument 50 if (!list_empty(&rreq->proc_link)) { 52 list_del_rcu(&rreq->proc_link); 57 static inline void netfs_proc_add_rreq(struct netfs_io_request *rreq) {} argument 58 static inline void netfs_proc_del_rreq(struct netfs_io_request *rreq) {} argument 81 void netfs_get_request(struct netfs_io_request *rreq, enu 87 netfs_see_request(struct netfs_io_request *rreq, enum netfs_rreq_ref_trace what) argument [all...] |
H A D | output.c | 156 struct netfs_io_request *wreq = subreq->rreq; 247 struct netfs_io_request *wreq = subreq->rreq;
|
/linux-master/fs/9p/ |
H A D | vfs_addr.c | 31 struct p9_fid *fid = subreq->rreq->netfs_priv; 67 struct netfs_io_request *rreq = subreq->rreq; local 68 struct p9_fid *fid = rreq->netfs_priv; 83 * @rreq: The read request 86 static int v9fs_init_request(struct netfs_io_request *rreq, struct file *file) argument 89 bool writing = (rreq->origin == NETFS_READ_FOR_WRITE || 90 rreq->origin == NETFS_WRITEBACK || 91 rreq->origin == NETFS_WRITETHROUGH || 92 rreq 124 v9fs_free_request(struct netfs_io_request *rreq) argument [all...] |
/linux-master/include/trace/events/ |
H A D | netfs.h | 193 TP_PROTO(struct netfs_io_request *rreq, 197 TP_ARGS(rreq, start, len, what), 200 __field(unsigned int, rreq ) 209 __entry->rreq = rreq->debug_id; 210 __entry->cookie = rreq->cache_resources.debug_id; 214 __entry->netfs_inode = rreq->inode->i_ino; 218 __entry->rreq, 226 TP_PROTO(struct netfs_io_request *rreq, 229 TP_ARGS(rreq, wha [all...] |
/linux-master/tools/net/ynl/samples/ |
H A D | ethtool.c | 14 struct ethtool_rings_get_req_dump rreq = {}; local 41 rreq._present.header = 1; /* ethtool needs an empty nest.. */ 42 rings = ethtool_rings_get_dump(ys, &rreq);
|
/linux-master/fs/nfs/ |
H A D | fscache.c | 264 static int nfs_netfs_init_request(struct netfs_io_request *rreq, struct file *file) argument 266 rreq->netfs_priv = get_nfs_open_context(nfs_file_open_context(file)); 267 rreq->debug_id = atomic_inc_return(&nfs_netfs_debug_id); 272 static void nfs_netfs_free_request(struct netfs_io_request *rreq) argument 274 put_nfs_open_context(rreq->netfs_priv); 291 size_t rsize = NFS_SB(sreq->rreq->inode->i_sb)->rsize; 301 struct inode *inode = sreq->rreq->inode; 302 struct nfs_open_context *ctx = sreq->rreq->netfs_priv; 319 xa_for_each_range(&sreq->rreq->mapping->i_pages, idx, page, start, last) {
|
/linux-master/drivers/infiniband/sw/siw/ |
H A D | siw_qp.c | 294 struct siw_sqe *rreq = NULL; local 301 rreq = orq_get_free(qp); 302 if (rreq) { 303 siw_read_to_orq(rreq, &wqe->sqe); 464 struct iwarp_rdma_rreq *rreq; local 472 rreq = (struct iwarp_rdma_rreq *)err_hdr; 474 memcpy(&rreq->ctrl, 478 rreq->rsvd = 0; 479 rreq->ddp_qn = 483 rreq 872 siw_read_to_orq(struct siw_sqe *rreq, struct siw_sqe *sqe) argument 937 struct siw_sqe *rreq; local [all...] |
H A D | siw_qp_rx.c | 634 pr_warn("siw: [QP %u]: rreq with mpa len %d\n", qp_id(qp), 661 uint64_t raddr = be64_to_cpu(srx->hdr.rreq.sink_to), 662 laddr = be64_to_cpu(srx->hdr.rreq.source_to); 663 uint32_t length = be32_to_cpu(srx->hdr.rreq.read_size), 664 lkey = be32_to_cpu(srx->hdr.rreq.source_stag), 665 rkey = be32_to_cpu(srx->hdr.rreq.sink_stag), 666 msn = be32_to_cpu(srx->hdr.rreq.ddp_msn); 1133 struct siw_sqe *rreq; local 1140 rreq = orq_get_current(qp); 1141 WRITE_ONCE(rreq [all...] |
H A D | siw_qp_tx.c | 130 c_tx->pkt.rreq.rsvd = 0; 131 c_tx->pkt.rreq.ddp_qn = htonl(RDMAP_UNTAGGED_QN_RDMA_READ); 132 c_tx->pkt.rreq.ddp_msn = 134 c_tx->pkt.rreq.ddp_mo = 0; 135 c_tx->pkt.rreq.sink_stag = htonl(wqe->sqe.sge[0].lkey); 136 c_tx->pkt.rreq.sink_to = 138 c_tx->pkt.rreq.source_stag = htonl(wqe->sqe.rkey); 139 c_tx->pkt.rreq.source_to = cpu_to_be64(wqe->sqe.raddr); 140 c_tx->pkt.rreq.read_size = htonl(wqe->sqe.sge[0].length);
|
H A D | siw.h | 353 struct iwarp_rdma_rreq rreq; member in struct:siw_rreq_pkt 373 struct iwarp_rdma_rreq rreq; member in union:siw_iwarp_tx::__anon1020 523 void siw_read_to_orq(struct siw_sqe *rreq, struct siw_sqe *sqe);
|
H A D | iwarp.h | 276 struct iwarp_rdma_rreq rreq; member in union:iwarp_hdr
|
/linux-master/net/9p/ |
H A D | trans_fd.c | 96 * @rreq: read request 119 struct p9_req_t *rreq; member in struct:p9_conn 313 if ((!m->rreq) && (m->rc.offset == m->rc.capacity)) { 329 m->rreq = p9_tag_lookup(m->client, m->rc.tag); 330 if (!m->rreq || (m->rreq->status != REQ_STATUS_SENT)) { 337 if (m->rc.size > m->rreq->rc.capacity) { 340 m->rc.size, m->rc.tag, m->rreq->rc.capacity); 345 if (!m->rreq->rc.sdata) { 348 m->rc.tag, m->rreq); [all...] |
/linux-master/include/linux/ |
H A D | netfs.h | 208 struct netfs_io_request *rreq; /* Supervising I/O request */ member in struct:netfs_io_subrequest 210 struct list_head rreq_link; /* Link in rreq->subrequests */ 300 int (*init_request)(struct netfs_io_request *rreq, struct file *file); 301 void (*free_request)(struct netfs_io_request *rreq); 302 void (*free_subrequest)(struct netfs_io_subrequest *rreq); 305 void (*expand_readahead)(struct netfs_io_request *rreq); 308 bool (*is_still_valid)(struct netfs_io_request *rreq); 311 void (*done)(struct netfs_io_request *rreq);
|
/linux-master/fs/afs/ |
H A D | file.c | 310 struct afs_vnode *vnode = AFS_FS_I(subreq->rreq->inode); 320 fsreq->key = key_get(subreq->rreq->netfs_priv); 352 static int afs_init_request(struct netfs_io_request *rreq, struct file *file) argument 355 rreq->netfs_priv = key_get(afs_file_key(file)); 356 rreq->rsize = 256 * 1024; 357 rreq->wsize = 256 * 1024; 369 static void afs_free_request(struct netfs_io_request *rreq) argument 371 key_put(rreq->netfs_priv);
|
H A D | write.c | 164 struct afs_vnode *vnode = AFS_FS_I(subreq->rreq->inode); 168 subreq->rreq->debug_id, subreq->debug_index, subreq->io_iter.count); 172 subreq->rreq->origin == NETFS_LAUNDER_WRITE);
|
/linux-master/fs/ceph/ |
H A D | addr.c | 165 static void ceph_netfs_expand_readahead(struct netfs_io_request *rreq) argument 167 struct inode *inode = rreq->inode; 171 loff_t end = rreq->start + rreq->len, new_end; 172 struct ceph_netfs_request_data *priv = rreq->netfs_priv; 196 new_end = min(round_up(end, lo->stripe_unit), rreq->i_size); 197 if (new_end > end && new_end <= rreq->start + max_len) 198 rreq->len = new_end - rreq->start; 201 div_u64_rem(rreq 273 struct netfs_io_request *rreq = subreq->rreq; local 328 struct netfs_io_request *rreq = subreq->rreq; local 419 ceph_init_request(struct netfs_io_request *rreq, struct file *file) argument 478 ceph_netfs_free_request(struct netfs_io_request *rreq) argument [all...] |
/linux-master/fs/cachefiles/ |
H A D | io.c | 498 return cachefiles_do_prepare_read(&subreq->rreq->cache_resources, 500 &subreq->flags, subreq->rreq->inode->i_ino);
|
/linux-master/drivers/md/ |
H A D | raid5.h | 259 /* rreq and rvec are used for the replacement device when 262 struct bio req, rreq; member in struct:stripe_head::r5dev
|
H A D | raid5.c | 1184 rbi = &dev->rreq; /* For writing to replacement */ 2859 if (bi == &sh->dev[i].rreq) {
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | tid_rdma.c | 1712 struct tid_rdma_read_req *rreq = &ohdr->u.tid_rdma.r_req; local 1746 KDETH_RESET(rreq->kdeth0, KVER, 0x1); 1747 KDETH_RESET(rreq->kdeth1, JKEY, remote->jkey); 1748 rreq->reth.vaddr = cpu_to_be64(wqe->rdma_wr.remote_addr + 1750 rreq->reth.rkey = cpu_to_be32(wqe->rdma_wr.rkey); 1751 rreq->reth.length = cpu_to_be32(*len); 1752 rreq->tid_flow_psn = 1757 rreq->tid_flow_qp = 1762 rreq->verbs_qp = cpu_to_be32(qp->remote_qpn);
|