/linux-master/io_uring/ |
H A D | xattr.h | 3 void io_xattr_cleanup(struct io_kiocb *req); 5 int io_fsetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 6 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags); 8 int io_setxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 9 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags); 11 int io_fgetxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 12 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags); 14 int io_getxattr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 15 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags);
|
H A D | fs.h | 3 int io_renameat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 4 int io_renameat(struct io_kiocb *req, unsigned int issue_flags); 5 void io_renameat_cleanup(struct io_kiocb *req); 7 int io_unlinkat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 8 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags); 9 void io_unlinkat_cleanup(struct io_kiocb *req); 11 int io_mkdirat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 12 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags); 13 void io_mkdirat_cleanup(struct io_kiocb *req); 15 int io_symlinkat_prep(struct io_kiocb *req, cons [all...] |
H A D | sync.h | 3 int io_sfr_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 4 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags); 6 int io_fsync_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 7 int io_fsync(struct io_kiocb *req, unsigned int issue_flags); 9 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags); 10 int io_fallocate_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
H A D | epoll.h | 4 int io_epoll_ctl_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 5 int io_epoll_ctl(struct io_kiocb *req, unsigned int issue_flags);
|
H A D | nop.h | 3 int io_nop_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 4 int io_nop(struct io_kiocb *req, unsigned int issue_flags);
|
H A D | truncate.h | 3 int io_ftruncate_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 4 int io_ftruncate(struct io_kiocb *req, unsigned int issue_flags);
|
H A D | advise.h | 3 int io_madvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 4 int io_madvise(struct io_kiocb *req, unsigned int issue_flags); 6 int io_fadvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 7 int io_fadvise(struct io_kiocb *req, unsigned int issue_flags);
|
H A D | splice.h | 3 int io_tee_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 4 int io_tee(struct io_kiocb *req, unsigned int issue_flags); 6 int io_splice_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 7 int io_splice(struct io_kiocb *req, unsigned int issue_flags);
|
H A D | statx.h | 3 int io_statx_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 4 int io_statx(struct io_kiocb *req, unsigned int issue_flags); 5 void io_statx_cleanup(struct io_kiocb *req);
|
H A D | msg_ring.h | 3 int io_msg_ring_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 4 int io_msg_ring(struct io_kiocb *req, unsigned int issue_flags); 5 void io_msg_ring_cleanup(struct io_kiocb *req);
|
H A D | refs.h | 11 #define req_ref_zero_or_close_to_overflow(req) \ 12 ((unsigned int) atomic_read(&(req->refs)) + 127u <= 127u) 14 static inline bool req_ref_inc_not_zero(struct io_kiocb *req) argument 16 WARN_ON_ONCE(!(req->flags & REQ_F_REFCOUNT)); 17 return atomic_inc_not_zero(&req->refs); 20 static inline bool req_ref_put_and_test(struct io_kiocb *req) argument 22 if (likely(!(req->flags & REQ_F_REFCOUNT))) 25 WARN_ON_ONCE(req_ref_zero_or_close_to_overflow(req)); 26 return atomic_dec_and_test(&req->refs); 29 static inline void req_ref_get(struct io_kiocb *req) argument 36 req_ref_put(struct io_kiocb *req) argument 43 __io_req_set_refcount(struct io_kiocb *req, int nr) argument 51 io_req_set_refcount(struct io_kiocb *req) argument [all...] |
H A D | rw.h | 15 int io_prep_read_fixed(struct io_kiocb *req, const struct io_uring_sqe *sqe); 16 int io_prep_write_fixed(struct io_kiocb *req, const struct io_uring_sqe *sqe); 17 int io_prep_readv(struct io_kiocb *req, const struct io_uring_sqe *sqe); 18 int io_prep_writev(struct io_kiocb *req, const struct io_uring_sqe *sqe); 19 int io_prep_read(struct io_kiocb *req, const struct io_uring_sqe *sqe); 20 int io_prep_write(struct io_kiocb *req, const struct io_uring_sqe *sqe); 21 int io_read(struct io_kiocb *req, unsigned int issue_flags); 22 int io_write(struct io_kiocb *req, unsigned int issue_flags); 23 void io_readv_writev_cleanup(struct io_kiocb *req); 24 void io_rw_fail(struct io_kiocb *req); [all...] |
H A D | net.h | 23 int io_shutdown_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 24 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags); 26 void io_sendmsg_recvmsg_cleanup(struct io_kiocb *req); 27 int io_sendmsg_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 28 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags); 30 int io_send(struct io_kiocb *req, unsigned int issue_flags); 32 int io_recvmsg_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 33 int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags); 34 int io_recv(struct io_kiocb *req, unsigned int issue_flags); 36 void io_sendrecv_fail(struct io_kiocb *req); [all...] |
H A D | kbuf.h | 59 void __user *io_buffer_select(struct io_kiocb *req, size_t *len, 61 int io_buffers_select(struct io_kiocb *req, struct buf_sel_arg *arg, 63 int io_buffers_peek(struct io_kiocb *req, struct buf_sel_arg *arg); 66 int io_remove_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 67 int io_remove_buffers(struct io_kiocb *req, unsigned int issue_flags); 69 int io_provide_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 70 int io_provide_buffers(struct io_kiocb *req, unsigned int issue_flags); 76 void __io_put_kbuf(struct io_kiocb *req, unsigned issue_flags); 78 bool io_kbuf_recycle_legacy(struct io_kiocb *req, unsigned issue_flags); 85 static inline bool io_kbuf_recycle_ring(struct io_kiocb *req) argument 102 io_do_buffer_select(struct io_kiocb *req) argument 109 io_kbuf_recycle(struct io_kiocb *req, unsigned issue_flags) argument 120 __io_put_kbuf_ring(struct io_kiocb *req, int nr) argument 134 __io_put_kbuf_list(struct io_kiocb *req, struct list_head *list) argument 146 io_kbuf_drop(struct io_kiocb *req) argument 156 __io_put_kbufs(struct io_kiocb *req, int nbufs, unsigned issue_flags) argument 172 io_put_kbuf(struct io_kiocb *req, unsigned issue_flags) argument 178 io_put_kbufs(struct io_kiocb *req, int nbufs, unsigned issue_flags) argument [all...] |
H A D | timeout.h | 4 struct io_kiocb *req; member in struct:io_timeout_data 11 struct io_kiocb *__io_disarm_linked_timeout(struct io_kiocb *req, 14 static inline struct io_kiocb *io_disarm_linked_timeout(struct io_kiocb *req) argument 16 struct io_kiocb *link = req->link; 19 return __io_disarm_linked_timeout(req, link); 29 void io_queue_linked_timeout(struct io_kiocb *req); 30 void io_disarm_next(struct io_kiocb *req); 32 int io_timeout_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 33 int io_link_timeout_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 34 int io_timeout(struct io_kiocb *req, unsigne [all...] |
H A D | openclose.h | 6 int io_openat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 7 int io_openat(struct io_kiocb *req, unsigned int issue_flags); 8 void io_open_cleanup(struct io_kiocb *req); 10 int io_openat2_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 11 int io_openat2(struct io_kiocb *req, unsigned int issue_flags); 13 int io_close_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 14 int io_close(struct io_kiocb *req, unsigned int issue_flags); 16 int io_install_fixed_fd_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 17 int io_install_fixed_fd(struct io_kiocb *req, unsigned int issue_flags);
|
/linux-master/drivers/usb/gadget/ |
H A D | u_f.c | 16 struct usb_request *req; local 18 req = usb_ep_alloc_request(ep, GFP_ATOMIC); 19 if (req) { 20 req->length = usb_endpoint_dir_out(ep->desc) ? 22 req->buf = kmalloc(req->length, GFP_ATOMIC); 23 if (!req->buf) { 24 usb_ep_free_request(ep, req); 25 req = NULL; 28 return req; [all...] |
/linux-master/drivers/crypto/intel/qat/qat_common/ |
H A D | qat_algs_send.c | 10 static int qat_alg_send_message_retry(struct qat_alg_req *req) argument 15 ret = adf_send_message(req->tx_ring, req->fw_req); 26 struct qat_alg_req *req, *tmp; local 29 list_for_each_entry_safe(req, tmp, &backlog->list, list) { 30 if (adf_send_message(req->tx_ring, req->fw_req)) { 37 list_del(&req->list); 38 crypto_request_complete(req->base, -EINPROGRESS); 43 static bool qat_alg_try_enqueue(struct qat_alg_req *req) argument 65 qat_alg_send_message_maybacklog(struct qat_alg_req *req) argument 83 qat_alg_send_message(struct qat_alg_req *req) argument [all...] |
/linux-master/drivers/media/mc/ |
H A D | mc-request.c | 39 static void media_request_clean(struct media_request *req) argument 44 WARN_ON(req->state != MEDIA_REQUEST_STATE_CLEANING); 45 WARN_ON(req->updating_count); 46 WARN_ON(req->access_count); 48 list_for_each_entry_safe(obj, obj_safe, &req->objects, list) { 53 req->updating_count = 0; 54 req->access_count = 0; 55 WARN_ON(req->num_incomplete_objects); 56 req->num_incomplete_objects = 0; 57 wake_up_interruptible_all(&req 62 struct media_request *req = local 79 media_request_put(struct media_request *req) argument 87 struct media_request *req = filp->private_data; local 96 struct media_request *req = filp->private_data; local 119 media_request_ioctl_queue(struct media_request *req) argument 188 media_request_ioctl_reinit(struct media_request *req) argument 224 struct media_request *req = filp->private_data; local 250 struct media_request *req; local 290 struct media_request *req; local 355 struct media_request *req = obj->req; local 363 media_request_object_find(struct media_request *req, const struct media_request_object_ops *ops, void *priv) argument 404 media_request_object_bind(struct media_request *req, const struct media_request_object_ops *ops, void *priv, bool is_buffer, struct media_request_object *obj) argument 440 struct media_request *req = obj->req; local 485 struct media_request *req = obj->req; local [all...] |
/linux-master/drivers/staging/greybus/ |
H A D | audio_apbridgea.c | 16 struct audio_apbridgea_set_config_request req; local 18 req.hdr.type = AUDIO_APBRIDGEA_TYPE_SET_CONFIG; 19 req.hdr.i2s_port = cpu_to_le16(i2s_port); 20 req.format = cpu_to_le32(format); 21 req.rate = cpu_to_le32(rate); 22 req.mclk_freq = cpu_to_le32(mclk_freq); 24 return gb_hd_output(connection->hd, &req, sizeof(req), 33 struct audio_apbridgea_register_cport_request req; local 36 req 54 struct audio_apbridgea_unregister_cport_request req; local 74 struct audio_apbridgea_set_tx_data_size_request req; local 88 struct audio_apbridgea_prepare_tx_request req; local 101 struct audio_apbridgea_start_tx_request req; local 114 struct audio_apbridgea_stop_tx_request req; local 127 struct audio_apbridgea_shutdown_tx_request req; local 140 struct audio_apbridgea_set_rx_data_size_request req; local 154 struct audio_apbridgea_prepare_rx_request req; local 167 struct audio_apbridgea_start_rx_request req; local 179 struct audio_apbridgea_stop_rx_request req; local 192 struct audio_apbridgea_shutdown_rx_request req; local [all...] |
/linux-master/net/handshake/ |
H A D | request.c | 67 static bool handshake_req_hash_add(struct handshake_req *req) argument 72 &req->hr_rhash, 77 static void handshake_req_destroy(struct handshake_req *req) argument 79 if (req->hr_proto->hp_destroy) 80 req->hr_proto->hp_destroy(req); 81 rhashtable_remove_fast(&handshake_rhashtbl, &req->hr_rhash, 83 kfree(req); 89 struct handshake_req *req; local 91 req 112 struct handshake_req *req; local 138 handshake_req_private(struct handshake_req *req) argument 144 __add_pending_locked(struct handshake_net *hn, struct handshake_req *req) argument 154 __remove_pending_locked(struct handshake_net *hn, struct handshake_req *req) argument 167 remove_pending(struct handshake_net *hn, struct handshake_req *req) argument 183 struct handshake_req *req, *pos; local 223 handshake_req_submit(struct socket *sock, struct handshake_req *req, gfp_t flags) argument 286 handshake_complete(struct handshake_req *req, unsigned int status, struct genl_info *info) argument 315 struct handshake_req *req; local [all...] |
/linux-master/net/sunrpc/ |
H A D | backchannel_rqst.c | 40 static void xprt_free_allocation(struct rpc_rqst *req) argument 44 dprintk("RPC: free allocations for req= %p\n", req); 45 WARN_ON_ONCE(test_bit(RPC_BC_PA_IN_USE, &req->rq_bc_pa_state)); 46 xbufp = &req->rq_rcv_buf; 48 xbufp = &req->rq_snd_buf; 50 kfree(req); 78 struct rpc_rqst *req; local 81 req = kzalloc(sizeof(*req), gfp_flag 133 struct rpc_rqst *req; local 211 struct rpc_rqst *req = NULL, *tmp = NULL; local 239 struct rpc_rqst *req = NULL; local 267 xprt_free_bc_request(struct rpc_rqst *req) argument 274 xprt_free_bc_rqst(struct rpc_rqst *req) argument 326 struct rpc_rqst *req, *new = NULL; local 354 xprt_complete_bc_request(struct rpc_rqst *req, uint32_t copied) argument [all...] |
/linux-master/tools/net/ynl/samples/ |
H A D | ovs.c | 19 struct ovs_datapath_new_req *req; local 21 req = ovs_datapath_new_req_alloc(); 22 if (!req) 25 ovs_datapath_new_req_set_upcall_pid(req, 1); 26 ovs_datapath_new_req_set_name(req, argv[1]); 28 err = ovs_datapath_new(ys, req); 29 ovs_datapath_new_req_free(req); 33 struct ovs_datapath_get_req_dump *req; local 37 req = ovs_datapath_get_req_dump_alloc(); 39 dps = ovs_datapath_get_dump(ys, req); [all...] |
/linux-master/drivers/nvme/target/ |
H A D | io-cmd-file.c | 76 static ssize_t nvmet_file_submit_bvec(struct nvmet_req *req, loff_t pos, argument 79 struct kiocb *iocb = &req->f.iocb; 84 if (req->cmd->rw.opcode == nvme_cmd_write) { 85 if (req->cmd->rw.control & cpu_to_le16(NVME_RW_FUA)) 87 call_iter = req->ns->file->f_op->write_iter; 90 call_iter = req->ns->file->f_op->read_iter; 94 iov_iter_bvec(&iter, rw, req->f.bvec, nr_segs, count); 97 iocb->ki_filp = req->ns->file; 105 struct nvmet_req *req = container_of(iocb, struct nvmet_req, f.iocb); local 108 if (req 120 nvmet_file_execute_io(struct nvmet_req *req, int ki_flags) argument 207 struct nvmet_req *req = container_of(w, struct nvmet_req, f.work); local 212 nvmet_file_submit_buffered_io(struct nvmet_req *req) argument 218 nvmet_file_execute_rw(struct nvmet_req *req) argument 253 nvmet_file_flush(struct nvmet_req *req) argument 260 struct nvmet_req *req = container_of(w, struct nvmet_req, f.work); local 265 nvmet_file_execute_flush(struct nvmet_req *req) argument 273 nvmet_file_execute_discard(struct nvmet_req *req) argument 310 struct nvmet_req *req = container_of(w, struct nvmet_req, f.work); local 325 nvmet_file_execute_dsm(struct nvmet_req *req) argument 335 struct nvmet_req *req = container_of(w, struct nvmet_req, f.work); local 355 nvmet_file_execute_write_zeroes(struct nvmet_req *req) argument 363 nvmet_file_parse_io_cmd(struct nvmet_req *req) argument [all...] |
/linux-master/arch/arm/kernel/ |
H A D | iwmmxt.h | 18 .macro wldrd, reg:req, base:req, offset:req variable 22 .macro wldrw, reg:req, base:req, offset:req variable 26 .macro wstrd, reg:req, base:req, offset:req variable 30 .macro wstrw, reg:req, bas variable 38 .macro tmrc, dest:req, control:req variable 42 .macro tmcr, control:req, src:req variable [all...] |