Searched refs:req (Results 1 - 25 of 2166) sorted by relevance

1234567891011>>

/linux-master/io_uring/
H A Dxattr.h3 void io_xattr_cleanup(struct io_kiocb *req);
5 int io_fsetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
6 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags);
8 int io_setxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
9 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags);
11 int io_fgetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
12 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags);
14 int io_getxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
15 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags);
H A Dfs.h3 int io_renameat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_renameat(struct io_kiocb *req, unsigned int issue_flags);
5 void io_renameat_cleanup(struct io_kiocb *req);
7 int io_unlinkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
8 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags);
9 void io_unlinkat_cleanup(struct io_kiocb *req);
11 int io_mkdirat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
12 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags);
13 void io_mkdirat_cleanup(struct io_kiocb *req);
15 int io_symlinkat_prep(struct io_kiocb *req, cons
[all...]
H A Dsync.h3 int io_sfr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags);
6 int io_fsync_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
7 int io_fsync(struct io_kiocb *req, unsigned int issue_flags);
9 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags);
10 int io_fallocate_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
H A Depoll.h4 int io_epoll_ctl_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
5 int io_epoll_ctl(struct io_kiocb *req, unsigned int issue_flags);
H A Dnop.h3 int io_nop_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_nop(struct io_kiocb *req, unsigned int issue_flags);
H A Dtruncate.h3 int io_ftruncate_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_ftruncate(struct io_kiocb *req, unsigned int issue_flags);
H A Dadvise.h3 int io_madvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_madvise(struct io_kiocb *req, unsigned int issue_flags);
6 int io_fadvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
7 int io_fadvise(struct io_kiocb *req, unsigned int issue_flags);
H A Dsplice.h3 int io_tee_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_tee(struct io_kiocb *req, unsigned int issue_flags);
6 int io_splice_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
7 int io_splice(struct io_kiocb *req, unsigned int issue_flags);
H A Dstatx.h3 int io_statx_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_statx(struct io_kiocb *req, unsigned int issue_flags);
5 void io_statx_cleanup(struct io_kiocb *req);
H A Dmsg_ring.h3 int io_msg_ring_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
4 int io_msg_ring(struct io_kiocb *req, unsigned int issue_flags);
5 void io_msg_ring_cleanup(struct io_kiocb *req);
H A Drefs.h11 #define req_ref_zero_or_close_to_overflow(req) \
12 ((unsigned int) atomic_read(&(req->refs)) + 127u <= 127u)
14 static inline bool req_ref_inc_not_zero(struct io_kiocb *req) argument
16 WARN_ON_ONCE(!(req->flags & REQ_F_REFCOUNT));
17 return atomic_inc_not_zero(&req->refs);
20 static inline bool req_ref_put_and_test(struct io_kiocb *req) argument
22 if (likely(!(req->flags & REQ_F_REFCOUNT)))
25 WARN_ON_ONCE(req_ref_zero_or_close_to_overflow(req));
26 return atomic_dec_and_test(&req->refs);
29 static inline void req_ref_get(struct io_kiocb *req) argument
36 req_ref_put(struct io_kiocb *req) argument
43 __io_req_set_refcount(struct io_kiocb *req, int nr) argument
51 io_req_set_refcount(struct io_kiocb *req) argument
[all...]
H A Drw.h15 int io_prep_read_fixed(struct io_kiocb *req, const struct io_uring_sqe *sqe);
16 int io_prep_write_fixed(struct io_kiocb *req, const struct io_uring_sqe *sqe);
17 int io_prep_readv(struct io_kiocb *req, const struct io_uring_sqe *sqe);
18 int io_prep_writev(struct io_kiocb *req, const struct io_uring_sqe *sqe);
19 int io_prep_read(struct io_kiocb *req, const struct io_uring_sqe *sqe);
20 int io_prep_write(struct io_kiocb *req, const struct io_uring_sqe *sqe);
21 int io_read(struct io_kiocb *req, unsigned int issue_flags);
22 int io_write(struct io_kiocb *req, unsigned int issue_flags);
23 void io_readv_writev_cleanup(struct io_kiocb *req);
24 void io_rw_fail(struct io_kiocb *req);
[all...]
H A Dnet.h23 int io_shutdown_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
24 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags);
26 void io_sendmsg_recvmsg_cleanup(struct io_kiocb *req);
27 int io_sendmsg_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
28 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags);
30 int io_send(struct io_kiocb *req, unsigned int issue_flags);
32 int io_recvmsg_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
33 int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags);
34 int io_recv(struct io_kiocb *req, unsigned int issue_flags);
36 void io_sendrecv_fail(struct io_kiocb *req);
[all...]
H A Dkbuf.h59 void __user *io_buffer_select(struct io_kiocb *req, size_t *len,
61 int io_buffers_select(struct io_kiocb *req, struct buf_sel_arg *arg,
63 int io_buffers_peek(struct io_kiocb *req, struct buf_sel_arg *arg);
66 int io_remove_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
67 int io_remove_buffers(struct io_kiocb *req, unsigned int issue_flags);
69 int io_provide_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
70 int io_provide_buffers(struct io_kiocb *req, unsigned int issue_flags);
76 void __io_put_kbuf(struct io_kiocb *req, unsigned issue_flags);
78 bool io_kbuf_recycle_legacy(struct io_kiocb *req, unsigned issue_flags);
85 static inline bool io_kbuf_recycle_ring(struct io_kiocb *req) argument
102 io_do_buffer_select(struct io_kiocb *req) argument
109 io_kbuf_recycle(struct io_kiocb *req, unsigned issue_flags) argument
120 __io_put_kbuf_ring(struct io_kiocb *req, int nr) argument
134 __io_put_kbuf_list(struct io_kiocb *req, struct list_head *list) argument
146 io_kbuf_drop(struct io_kiocb *req) argument
156 __io_put_kbufs(struct io_kiocb *req, int nbufs, unsigned issue_flags) argument
172 io_put_kbuf(struct io_kiocb *req, unsigned issue_flags) argument
178 io_put_kbufs(struct io_kiocb *req, int nbufs, unsigned issue_flags) argument
[all...]
H A Dtimeout.h4 struct io_kiocb *req; member in struct:io_timeout_data
11 struct io_kiocb *__io_disarm_linked_timeout(struct io_kiocb *req,
14 static inline struct io_kiocb *io_disarm_linked_timeout(struct io_kiocb *req) argument
16 struct io_kiocb *link = req->link;
19 return __io_disarm_linked_timeout(req, link);
29 void io_queue_linked_timeout(struct io_kiocb *req);
30 void io_disarm_next(struct io_kiocb *req);
32 int io_timeout_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
33 int io_link_timeout_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
34 int io_timeout(struct io_kiocb *req, unsigne
[all...]
H A Dopenclose.h6 int io_openat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
7 int io_openat(struct io_kiocb *req, unsigned int issue_flags);
8 void io_open_cleanup(struct io_kiocb *req);
10 int io_openat2_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
11 int io_openat2(struct io_kiocb *req, unsigned int issue_flags);
13 int io_close_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
14 int io_close(struct io_kiocb *req, unsigned int issue_flags);
16 int io_install_fixed_fd_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
17 int io_install_fixed_fd(struct io_kiocb *req, unsigned int issue_flags);
/linux-master/drivers/usb/gadget/
H A Du_f.c16 struct usb_request *req; local
18 req = usb_ep_alloc_request(ep, GFP_ATOMIC);
19 if (req) {
20 req->length = usb_endpoint_dir_out(ep->desc) ?
22 req->buf = kmalloc(req->length, GFP_ATOMIC);
23 if (!req->buf) {
24 usb_ep_free_request(ep, req);
25 req = NULL;
28 return req;
[all...]
/linux-master/drivers/crypto/intel/qat/qat_common/
H A Dqat_algs_send.c10 static int qat_alg_send_message_retry(struct qat_alg_req *req) argument
15 ret = adf_send_message(req->tx_ring, req->fw_req);
26 struct qat_alg_req *req, *tmp; local
29 list_for_each_entry_safe(req, tmp, &backlog->list, list) {
30 if (adf_send_message(req->tx_ring, req->fw_req)) {
37 list_del(&req->list);
38 crypto_request_complete(req->base, -EINPROGRESS);
43 static bool qat_alg_try_enqueue(struct qat_alg_req *req) argument
65 qat_alg_send_message_maybacklog(struct qat_alg_req *req) argument
83 qat_alg_send_message(struct qat_alg_req *req) argument
[all...]
/linux-master/drivers/media/mc/
H A Dmc-request.c39 static void media_request_clean(struct media_request *req) argument
44 WARN_ON(req->state != MEDIA_REQUEST_STATE_CLEANING);
45 WARN_ON(req->updating_count);
46 WARN_ON(req->access_count);
48 list_for_each_entry_safe(obj, obj_safe, &req->objects, list) {
53 req->updating_count = 0;
54 req->access_count = 0;
55 WARN_ON(req->num_incomplete_objects);
56 req->num_incomplete_objects = 0;
57 wake_up_interruptible_all(&req
62 struct media_request *req = local
79 media_request_put(struct media_request *req) argument
87 struct media_request *req = filp->private_data; local
96 struct media_request *req = filp->private_data; local
119 media_request_ioctl_queue(struct media_request *req) argument
188 media_request_ioctl_reinit(struct media_request *req) argument
224 struct media_request *req = filp->private_data; local
250 struct media_request *req; local
290 struct media_request *req; local
355 struct media_request *req = obj->req; local
363 media_request_object_find(struct media_request *req, const struct media_request_object_ops *ops, void *priv) argument
404 media_request_object_bind(struct media_request *req, const struct media_request_object_ops *ops, void *priv, bool is_buffer, struct media_request_object *obj) argument
440 struct media_request *req = obj->req; local
485 struct media_request *req = obj->req; local
[all...]
/linux-master/drivers/staging/greybus/
H A Daudio_apbridgea.c16 struct audio_apbridgea_set_config_request req; local
18 req.hdr.type = AUDIO_APBRIDGEA_TYPE_SET_CONFIG;
19 req.hdr.i2s_port = cpu_to_le16(i2s_port);
20 req.format = cpu_to_le32(format);
21 req.rate = cpu_to_le32(rate);
22 req.mclk_freq = cpu_to_le32(mclk_freq);
24 return gb_hd_output(connection->hd, &req, sizeof(req),
33 struct audio_apbridgea_register_cport_request req; local
36 req
54 struct audio_apbridgea_unregister_cport_request req; local
74 struct audio_apbridgea_set_tx_data_size_request req; local
88 struct audio_apbridgea_prepare_tx_request req; local
101 struct audio_apbridgea_start_tx_request req; local
114 struct audio_apbridgea_stop_tx_request req; local
127 struct audio_apbridgea_shutdown_tx_request req; local
140 struct audio_apbridgea_set_rx_data_size_request req; local
154 struct audio_apbridgea_prepare_rx_request req; local
167 struct audio_apbridgea_start_rx_request req; local
179 struct audio_apbridgea_stop_rx_request req; local
192 struct audio_apbridgea_shutdown_rx_request req; local
[all...]
/linux-master/net/handshake/
H A Drequest.c67 static bool handshake_req_hash_add(struct handshake_req *req) argument
72 &req->hr_rhash,
77 static void handshake_req_destroy(struct handshake_req *req) argument
79 if (req->hr_proto->hp_destroy)
80 req->hr_proto->hp_destroy(req);
81 rhashtable_remove_fast(&handshake_rhashtbl, &req->hr_rhash,
83 kfree(req);
89 struct handshake_req *req; local
91 req
112 struct handshake_req *req; local
138 handshake_req_private(struct handshake_req *req) argument
144 __add_pending_locked(struct handshake_net *hn, struct handshake_req *req) argument
154 __remove_pending_locked(struct handshake_net *hn, struct handshake_req *req) argument
167 remove_pending(struct handshake_net *hn, struct handshake_req *req) argument
183 struct handshake_req *req, *pos; local
223 handshake_req_submit(struct socket *sock, struct handshake_req *req, gfp_t flags) argument
286 handshake_complete(struct handshake_req *req, unsigned int status, struct genl_info *info) argument
315 struct handshake_req *req; local
[all...]
/linux-master/net/sunrpc/
H A Dbackchannel_rqst.c40 static void xprt_free_allocation(struct rpc_rqst *req) argument
44 dprintk("RPC: free allocations for req= %p\n", req);
45 WARN_ON_ONCE(test_bit(RPC_BC_PA_IN_USE, &req->rq_bc_pa_state));
46 xbufp = &req->rq_rcv_buf;
48 xbufp = &req->rq_snd_buf;
50 kfree(req);
78 struct rpc_rqst *req; local
81 req = kzalloc(sizeof(*req), gfp_flag
133 struct rpc_rqst *req; local
211 struct rpc_rqst *req = NULL, *tmp = NULL; local
239 struct rpc_rqst *req = NULL; local
267 xprt_free_bc_request(struct rpc_rqst *req) argument
274 xprt_free_bc_rqst(struct rpc_rqst *req) argument
326 struct rpc_rqst *req, *new = NULL; local
354 xprt_complete_bc_request(struct rpc_rqst *req, uint32_t copied) argument
[all...]
/linux-master/tools/net/ynl/samples/
H A Dovs.c19 struct ovs_datapath_new_req *req; local
21 req = ovs_datapath_new_req_alloc();
22 if (!req)
25 ovs_datapath_new_req_set_upcall_pid(req, 1);
26 ovs_datapath_new_req_set_name(req, argv[1]);
28 err = ovs_datapath_new(ys, req);
29 ovs_datapath_new_req_free(req);
33 struct ovs_datapath_get_req_dump *req; local
37 req = ovs_datapath_get_req_dump_alloc();
39 dps = ovs_datapath_get_dump(ys, req);
[all...]
/linux-master/drivers/nvme/target/
H A Dio-cmd-file.c76 static ssize_t nvmet_file_submit_bvec(struct nvmet_req *req, loff_t pos, argument
79 struct kiocb *iocb = &req->f.iocb;
84 if (req->cmd->rw.opcode == nvme_cmd_write) {
85 if (req->cmd->rw.control & cpu_to_le16(NVME_RW_FUA))
87 call_iter = req->ns->file->f_op->write_iter;
90 call_iter = req->ns->file->f_op->read_iter;
94 iov_iter_bvec(&iter, rw, req->f.bvec, nr_segs, count);
97 iocb->ki_filp = req->ns->file;
105 struct nvmet_req *req = container_of(iocb, struct nvmet_req, f.iocb); local
108 if (req
120 nvmet_file_execute_io(struct nvmet_req *req, int ki_flags) argument
207 struct nvmet_req *req = container_of(w, struct nvmet_req, f.work); local
212 nvmet_file_submit_buffered_io(struct nvmet_req *req) argument
218 nvmet_file_execute_rw(struct nvmet_req *req) argument
253 nvmet_file_flush(struct nvmet_req *req) argument
260 struct nvmet_req *req = container_of(w, struct nvmet_req, f.work); local
265 nvmet_file_execute_flush(struct nvmet_req *req) argument
273 nvmet_file_execute_discard(struct nvmet_req *req) argument
310 struct nvmet_req *req = container_of(w, struct nvmet_req, f.work); local
325 nvmet_file_execute_dsm(struct nvmet_req *req) argument
335 struct nvmet_req *req = container_of(w, struct nvmet_req, f.work); local
355 nvmet_file_execute_write_zeroes(struct nvmet_req *req) argument
363 nvmet_file_parse_io_cmd(struct nvmet_req *req) argument
[all...]
/linux-master/arch/arm/kernel/
H A Diwmmxt.h18 .macro wldrd, reg:req, base:req, offset:req variable
22 .macro wldrw, reg:req, base:req, offset:req variable
26 .macro wstrd, reg:req, base:req, offset:req variable
30 .macro wstrw, reg:req, bas variable
38 .macro tmrc, dest:req, control:req variable
42 .macro tmcr, control:req, src:req variable
[all...]

Completed in 615 milliseconds

1234567891011>>