/linux-master/io_uring/ |
H A D | register.h | 5 int io_eventfd_unregister(struct io_ring_ctx *ctx); 6 int io_unregister_personality(struct io_ring_ctx *ctx, unsigned id);
|
H A D | tctx.h | 6 struct io_ring_ctx *ctx; 10 struct io_ring_ctx *ctx); 12 int __io_uring_add_tctx_node(struct io_ring_ctx *ctx); 13 int __io_uring_add_tctx_node_from_submit(struct io_ring_ctx *ctx); 17 int io_ringfd_register(struct io_ring_ctx *ctx, void __user *__arg, 19 int io_ringfd_unregister(struct io_ring_ctx *ctx, void __user *__arg, 25 static inline int io_uring_add_tctx_node(struct io_ring_ctx *ctx)
|
H A D | napi.h | 12 void io_napi_init(struct io_ring_ctx *ctx); 13 void io_napi_free(struct io_ring_ctx *ctx); 15 int io_register_napi(struct io_ring_ctx *ctx, void __user *arg); 16 int io_unregister_napi(struct io_ring_ctx *ctx, void __user *arg); 18 void __io_napi_add(struct io_ring_ctx *ctx, struct socket *sock); 20 void __io_napi_adjust_timeout(struct io_ring_ctx *ctx, 22 void __io_napi_busy_loop(struct io_ring_ctx *ctx, struct io_wait_queue *iowq); 23 int io_napi_sqpoll_busy_poll(struct io_ring_ctx *ctx); 25 static inline bool io_napi(struct io_ring_ctx *ctx) 30 static inline void io_napi_adjust_timeout(struct io_ring_ctx *ct [all...] |
H A D | futex.h | 12 int io_futex_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd, 14 bool io_futex_remove_all(struct io_ring_ctx *ctx, struct task_struct *task, 16 void io_futex_cache_init(struct io_ring_ctx *ctx); 17 void io_futex_cache_free(struct io_ring_ctx *ctx); 19 static inline int io_futex_cancel(struct io_ring_ctx *ctx, 25 static inline bool io_futex_remove_all(struct io_ring_ctx *ctx, 30 static inline void io_futex_cache_init(struct io_ring_ctx *ctx) 33 static inline void io_futex_cache_free(struct io_ring_ctx *ctx)
|
H A D | sqpoll.h | 24 int io_sq_offload_create(struct io_ring_ctx *ctx, struct io_uring_params *p); 25 void io_sq_thread_finish(struct io_ring_ctx *ctx); 30 void io_sqpoll_wait_sq(struct io_ring_ctx *ctx); 31 int io_sqpoll_wq_cpu_affinity(struct io_ring_ctx *ctx, cpumask_var_t mask);
|
H A D | waitid.h | 12 int io_waitid_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd, 14 bool io_waitid_remove_all(struct io_ring_ctx *ctx, struct task_struct *task,
|
H A D | rsrc.h | 27 typedef void (rsrc_put_fn)(struct io_ring_ctx *ctx, struct io_rsrc_put *prsrc); 30 struct io_ring_ctx *ctx; 41 struct io_ring_ctx *ctx; 59 void io_rsrc_node_destroy(struct io_ring_ctx *ctx, struct io_rsrc_node *ref_node); 60 struct io_rsrc_node *io_rsrc_node_alloc(struct io_ring_ctx *ctx); 67 void __io_sqe_buffers_unregister(struct io_ring_ctx *ctx); 68 int io_sqe_buffers_unregister(struct io_ring_ctx *ctx); 69 int io_sqe_buffers_register(struct io_ring_ctx *ctx, void __user *arg, 71 void __io_sqe_files_unregister(struct io_ring_ctx *ctx); 72 int io_sqe_files_unregister(struct io_ring_ctx *ct [all...] |
H A D | io_uring.h | 40 struct io_ring_ctx *ctx; 53 struct io_ring_ctx *ctx = iowq->ctx; 64 bool io_cqe_cache_refill(struct io_ring_ctx *ctx, bool overflow); 66 int io_run_task_work_sig(struct io_ring_ctx *ctx); 69 bool io_post_aux_cqe(struct io_ring_ctx *ctx, u64 user_data, s32 res, u32 cflags); 71 void __io_commit_cqring_flush(struct io_ring_ctx *ctx); 91 struct io_ring_ctx *ctx); 97 int io_submit_sqes(struct io_ring_ctx *ctx, unsigned int nr); 98 int io_do_iopoll(struct io_ring_ctx *ctx, bool force_nonspin); 99 void __io_submit_flush_completions(struct io_ring_ctx *ct [all...] |
H A D | timeout.h | 24 __cold void io_flush_timeouts(struct io_ring_ctx *ctx); 26 int io_timeout_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd); 27 __cold bool io_kill_timeouts(struct io_ring_ctx *ctx, struct task_struct *tsk,
|
H A D | filetable.h | 13 int __io_fixed_fd_install(struct io_ring_ctx *ctx, struct file *file, 15 int io_fixed_fd_remove(struct io_ring_ctx *ctx, unsigned int offset); 17 int io_register_file_alloc_range(struct io_ring_ctx *ctx, 69 static inline void io_reset_alloc_hint(struct io_ring_ctx *ctx) 74 static inline void io_file_table_set_alloc_range(struct io_ring_ctx *ctx,
|
H A D | poll.h | 43 int io_poll_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd, 46 bool io_poll_remove_all(struct io_ring_ctx *ctx, struct task_struct *tsk,
|
H A D | cancel.h | 8 struct io_ring_ctx *ctx; 25 int io_sync_cancel(struct io_ring_ctx *ctx, void __user *arg);
|
H A D | napi.c | 36 void __io_napi_add(struct io_ring_ctx *ctx, struct socket *sock) 83 static void __io_napi_remove_stale(struct io_ring_ctx *ctx) 99 static inline void io_napi_remove_stale(struct io_ring_ctx *ctx, bool is_stale) 133 static bool __io_napi_do_busy_loop(struct io_ring_ctx *ctx, 154 static void io_napi_blocking_busy_loop(struct io_ring_ctx *ctx, 182 void io_napi_init(struct io_ring_ctx *ctx) 196 void io_napi_free(struct io_ring_ctx *ctx) 217 int io_register_napi(struct io_ring_ctx *ctx, void __user *arg) 247 int io_unregister_napi(struct io_ring_ctx *ctx, void __user *arg) 271 void __io_napi_adjust_timeout(struct io_ring_ctx *ct [all...] |
H A D | kbuf.h | 46 void io_destroy_buffers(struct io_ring_ctx *ctx); 54 int io_register_pbuf_ring(struct io_ring_ctx *ctx, void __user *arg); 55 int io_unregister_pbuf_ring(struct io_ring_ctx *ctx, void __user *arg); 56 int io_register_pbuf_status(struct io_ring_ctx *ctx, void __user *arg); 58 void io_kbuf_mmap_list_free(struct io_ring_ctx *ctx); 64 void io_put_bl(struct io_ring_ctx *ctx, struct io_buffer_list *bl); 65 struct io_buffer_list *io_pbuf_get_bl(struct io_ring_ctx *ctx,
|
H A D | filetable.c | 16 static int io_file_bitmap_get(struct io_ring_ctx *ctx) 63 static int io_install_fixed_file(struct io_ring_ctx *ctx, struct file *file, 96 int __io_fixed_fd_install(struct io_ring_ctx *ctx, struct file *file, 123 struct io_ring_ctx *ctx = req->ctx; 135 int io_fixed_fd_remove(struct io_ring_ctx *ctx, unsigned int offset) 160 int io_register_file_alloc_range(struct io_ring_ctx *ctx,
|
H A D | kbuf.c | 41 static inline struct io_buffer_list *__io_buffer_get_list(struct io_ring_ctx *ctx, 47 static inline struct io_buffer_list *io_buffer_get_list(struct io_ring_ctx *ctx, 55 static int io_buffer_add_list(struct io_ring_ctx *ctx, 70 struct io_ring_ctx *ctx = req->ctx; 101 struct io_ring_ctx *ctx = req->ctx; 184 struct io_ring_ctx *ctx = req->ctx; 204 static void io_kbuf_mark_free(struct io_ring_ctx *ctx, struct io_buffer_list *bl) 219 static int __io_remove_buffers(struct io_ring_ctx *ctx, 269 void io_put_bl(struct io_ring_ctx *ctx, struct io_buffer_list *bl) 277 void io_destroy_buffers(struct io_ring_ctx *ct [all...] |
H A D | notif.h | 22 struct io_kiocb *io_alloc_notif(struct io_ring_ctx *ctx); 42 struct io_ring_ctx *ctx = notif->ctx;
|
H A D | openclose.h | 3 int __io_close_fixed(struct io_ring_ctx *ctx, unsigned int issue_flags,
|
H A D | msg_ring.c | 36 static void io_double_unlock_ctx(struct io_ring_ctx *octx) 41 static int io_double_lock_ctx(struct io_ring_ctx *octx, 69 static inline bool io_msg_need_remote(struct io_ring_ctx *target_ctx) 78 struct io_ring_ctx *ctx = req->file->private_data; 96 struct io_ring_ctx *target_ctx = req->file->private_data; 128 struct io_ring_ctx *target_ctx = req->file->private_data; 163 struct io_ring_ctx *ctx = req->ctx; 180 struct io_ring_ctx *target_ctx = req->file->private_data; 225 struct io_ring_ctx *target_ctx = req->file->private_data; 227 struct io_ring_ctx *ct [all...] |
H A D | rsrc.c | 26 static void io_rsrc_buf_put(struct io_ring_ctx *ctx, struct io_rsrc_put *prsrc); 27 static int io_sqe_buffer_register(struct io_ring_ctx *ctx, struct iovec *iov, 61 static void io_unaccount_mem(struct io_ring_ctx *ctx, unsigned long nr_pages) 70 static int io_account_mem(struct io_ring_ctx *ctx, unsigned long nr_pages) 86 static int io_copy_iov(struct io_ring_ctx *ctx, struct iovec *dst, 135 static void io_buffer_unmap(struct io_ring_ctx *ctx, struct io_mapped_ubuf **slot) 170 void io_rsrc_node_destroy(struct io_ring_ctx *ctx, struct io_rsrc_node *node) 179 struct io_ring_ctx *ctx = node->ctx; 197 struct io_rsrc_node *io_rsrc_node_alloc(struct io_ring_ctx *ctx) 218 struct io_ring_ctx *ct [all...] |
H A D | futex.c | 37 void io_futex_cache_init(struct io_ring_ctx *ctx) 48 void io_futex_cache_free(struct io_ring_ctx *ctx) 63 struct io_ring_ctx *ctx = req->ctx; 99 static bool __io_futex_cancel(struct io_ring_ctx *ctx, struct io_kiocb *req) 122 int io_futex_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd, 150 bool io_futex_remove_all(struct io_ring_ctx *ctx, struct task_struct *task, 260 static struct io_futex_data *io_alloc_ifd(struct io_ring_ctx *ctx) 275 struct io_ring_ctx *ctx = req->ctx; 329 struct io_ring_ctx *ctx = req->ctx;
|
H A D | io_uring.c | 143 static bool io_uring_try_cancel_requests(struct io_ring_ctx *ctx, 177 static inline void io_submit_flush_completions(struct io_ring_ctx *ctx) 184 static inline unsigned int __io_cqring_events(struct io_ring_ctx *ctx) 189 static inline unsigned int __io_cqring_events_user(struct io_ring_ctx *ctx) 220 struct io_ring_ctx *ctx = head->ctx; 238 static inline void io_req_add_to_cache(struct io_kiocb *req, struct io_ring_ctx *ctx) 245 struct io_ring_ctx *ctx = container_of(ref, struct io_ring_ctx, refs); 252 struct io_ring_ctx *ctx = container_of(work, struct io_ring_ctx, [all...] |
H A D | timeout.c | 72 struct io_ring_ctx *ctx = req->ctx; 111 __cold void io_flush_timeouts(struct io_ring_ctx *ctx) 205 struct io_ring_ctx *ctx = req->ctx; 242 struct io_ring_ctx *ctx = req->ctx; 260 static struct io_kiocb *io_timeout_extract(struct io_ring_ctx *ctx, 287 int io_timeout_cancel(struct io_ring_ctx *ctx, struct io_cancel_data *cd) 333 struct io_ring_ctx *ctx = req->ctx; 374 static int io_linked_timeout_update(struct io_ring_ctx *ctx, __u64 user_data, 402 static int io_timeout_update(struct io_ring_ctx *ctx, __u64 user_data, 466 struct io_ring_ctx *ct [all...] |
H A D | register.c | 34 static int io_eventfd_register(struct io_ring_ctx *ctx, void __user *arg, 72 int io_eventfd_unregister(struct io_ring_ctx *ctx) 89 static __cold int io_probe(struct io_ring_ctx *ctx, void __user *arg, 129 int io_unregister_personality(struct io_ring_ctx *ctx, unsigned id) 143 static int io_register_personality(struct io_ring_ctx *ctx) 160 static __cold int io_register_restrictions(struct io_ring_ctx *ctx, 230 static int io_register_enable_rings(struct io_ring_ctx *ctx) 254 static __cold int __io_register_iowq_aff(struct io_ring_ctx *ctx, 270 static __cold int io_register_iowq_aff(struct io_ring_ctx *ctx, 302 static __cold int io_unregister_iowq_aff(struct io_ring_ctx *ct [all...] |
H A D | tctx.c | 15 static struct io_wq *io_init_wq_offload(struct io_ring_ctx *ctx, 61 struct io_ring_ctx *ctx) 94 int __io_uring_add_tctx_node(struct io_ring_ctx *ctx) 136 int __io_uring_add_tctx_node_from_submit(struct io_ring_ctx *ctx) 253 int io_ringfd_register(struct io_ring_ctx *ctx, void __user *__arg, 312 int io_ringfd_unregister(struct io_ring_ctx *ctx, void __user *__arg,
|