Searched refs:uring_lock (Results 1 - 18 of 18) sorted by relevance
/linux-master/io_uring/ |
H A D | tctx.c | 22 mutex_lock(&ctx->uring_lock); 27 mutex_unlock(&ctx->uring_lock); 34 mutex_unlock(&ctx->uring_lock); 129 mutex_lock(&ctx->uring_lock); 131 mutex_unlock(&ctx->uring_lock); 169 mutex_lock(&node->ctx->uring_lock); 171 mutex_unlock(&node->ctx->uring_lock); 191 * uring_lock) to avoid race with io_uring_try_cancel_iowq(). 264 mutex_unlock(&ctx->uring_lock); 266 mutex_lock(&ctx->uring_lock); [all...] |
H A D | io_uring.h | 130 lockdep_assert_held(&ctx->uring_lock); 238 lockdep_assert_held(&ctx->uring_lock); 240 mutex_unlock(&ctx->uring_lock); 247 * "Normal" inline submissions always hold the uring_lock, since we 253 mutex_lock(&ctx->uring_lock); 254 lockdep_assert_held(&ctx->uring_lock); 349 mutex_lock(&ctx->uring_lock); 356 * Protected by ->uring_lock and can only be used either with 360 __must_hold(&req->ctx->uring_lock) 364 lockdep_assert_held(&req->ctx->uring_lock); [all...] |
H A D | register.c | 42 lockdep_is_held(&ctx->uring_lock)); 77 lockdep_is_held(&ctx->uring_lock)); 262 mutex_unlock(&ctx->uring_lock); 264 mutex_lock(&ctx->uring_lock); 309 __must_hold(&ctx->uring_lock) 327 * Observe the correct sqd->lock -> ctx->uring_lock 328 * ordering. Fine to drop uring_lock here, we hold 332 mutex_unlock(&ctx->uring_lock); 334 mutex_lock(&ctx->uring_lock); 392 __releases(ctx->uring_lock) [all...] |
H A D | notif.h | 31 __must_hold(¬if->ctx->uring_lock)
|
H A D | msg_ring.c | 38 mutex_unlock(&octx->uring_lock); 50 if (!mutex_trylock(&octx->uring_lock)) 54 mutex_lock(&octx->uring_lock); 109 * holding the uring_lock for posting completions. Other ring 114 mutex_lock(&target_ctx->uring_lock); 118 mutex_unlock(&target_ctx->uring_lock);
|
H A D | io_uring.c | 259 mutex_lock(&ctx->uring_lock); 265 mutex_unlock(&ctx->uring_lock); 325 mutex_init(&ctx->uring_lock); 719 /* iopoll syncs against uring_lock, not completion_lock */ 721 mutex_lock(&ctx->uring_lock); 724 mutex_unlock(&ctx->uring_lock); 894 __must_hold(&ctx->uring_lock) 899 lockdep_assert_held(&ctx->uring_lock); 952 lockdep_assert_held(&ctx->uring_lock); 1034 mutex_lock(&ctx->uring_lock); [all...] |
H A D | rsrc.h | 85 lockdep_assert_held(&ctx->uring_lock); 106 lockdep_assert_held(&ctx->uring_lock);
|
H A D | cancel.c | 252 /* fixed must be grabbed every time since we drop the uring_lock */ 267 __must_hold(&ctx->uring_lock) 329 mutex_unlock(&ctx->uring_lock); 341 mutex_lock(&ctx->uring_lock); 345 mutex_lock(&ctx->uring_lock);
|
H A D | notif.c | 66 __must_hold(&ctx->uring_lock)
|
H A D | kbuf.c | 50 lockdep_assert_held(&ctx->uring_lock); 61 * always under the ->uring_lock, but the RCU lookup from mmap does. 92 * ctx->uring_lock. If we already hold this lock, add back to this 107 lockdep_assert_held(&req->ctx->uring_lock); 254 lockdep_assert_held(&ctx->uring_lock); 390 * Completions that don't happen inline (eg not under uring_lock) will 617 lockdep_assert_held(&ctx->uring_lock); 676 lockdep_assert_held(&ctx->uring_lock); 730 * the uring_lock. This means the buffer_list could be simultaneously
|
H A D | filetable.c | 65 __must_hold(&req->ctx->uring_lock)
|
H A D | fdinfo.c | 79 * since we get cached_sq_head and cached_cq_tail without uring_lock 145 has_lock = mutex_trylock(&ctx->uring_lock); 213 mutex_unlock(&ctx->uring_lock);
|
H A D | waitid.c | 126 lockdep_assert_held(&req->ctx->uring_lock); 194 lockdep_assert_held(&ctx->uring_lock);
|
H A D | rsrc.c | 177 __must_hold(&node->ctx->uring_lock) 224 /* As We may drop ->uring_lock, other task may have started quiesce */ 249 mutex_unlock(&ctx->uring_lock); 253 mutex_lock(&ctx->uring_lock); 261 mutex_lock(&ctx->uring_lock); 479 lockdep_assert_held(&ctx->uring_lock); 689 * Quiesce may unlock ->uring_lock, and while it's not held 795 * Quiesce may unlock ->uring_lock, and while it's not held
|
H A D | sqpoll.c | 185 mutex_lock(&ctx->uring_lock); 196 mutex_unlock(&ctx->uring_lock);
|
H A D | futex.c | 157 lockdep_assert_held(&ctx->uring_lock);
|
H A D | poll.c | 147 lockdep_assert_held(&req->ctx->uring_lock); 158 * ->cancel_table_locked is protected by ->uring_lock in 607 /* io-wq doesn't hold uring_lock */ 796 __must_hold(&ctx->uring_lock) 977 * If sqpoll or single issuer, there is no contention for ->uring_lock
|
/linux-master/include/linux/ |
H A D | io_uring_types.h | 200 /* inline/task_work completion list, under ->uring_lock */ 256 struct mutex uring_lock; member in struct:io_ring_ctx::__anon48 276 * uring_lock, and updated through io_uring_register(2) 282 * ->iopoll_list is protected by the ctx->uring_lock for 374 /* deferred free list, protected by ->uring_lock */ 386 /* protected by ->uring_lock */ 441 /* ->uring_lock is taken, callbacks can use io_tw_lock to lock it */ 549 /* hashed into ->cancel_hash_locked, protected by ->uring_lock */
|
Completed in 173 milliseconds