/linux-master/virt/kvm/ |
H A D | pfncache.c | 32 read_lock_irq(&gpc->lock); 37 read_unlock_irq(&gpc->lock); 43 * is still necessary once the write lock has been 47 write_lock_irq(&gpc->lock); 51 write_unlock_irq(&gpc->lock); 55 read_unlock_irq(&gpc->lock); 137 * is not protected by gpc->lock. It is guaranteed to 138 * be elevated before the mmu_notifier acquires gpc->lock, and 165 lockdep_assert_held_write(&gpc->lock); 168 * Invalidate the cache prior to dropping gpc->lock, th [all...] |
H A D | kvm_main.c | 102 * kvm->lock --> kvm->slots_lock --> kvm->irq_lock 828 * each cache's lock. There are relatively few caches in existence at 1202 mutex_init(&kvm->lock); 1328 * We do not need to take the kvm->lock here, because nobody else 1848 * slots_arch_lock in kvm_swap_active_memslots() and re-acquiring the lock 1930 * (and without a lock), a window would exist between effecting the 3991 * (preempted lock holder), indicated by @in_spin_loop. 4000 * to preempted lock-holder could result in wrong VCPU selection and CPU 4001 * burning. Giving priority for a potential lock-holder increases lock [all...] |
/linux-master/kernel/sched/ |
H A D | sched.h | 270 /* nests inside the rq lock: */ 302 raw_spinlock_t lock; member in struct:dl_bw 354 raw_spinlock_t lock; member in struct:cfs_bandwidth 616 raw_spinlock_t lock ____cacheline_aligned; 713 /* Nests inside the rq lock: */ 897 /* These atomics are updated outside of a lock */ 981 * Locking rule: those places that want to lock multiple runqueues 982 * (such as the load balancing or the thread migration code), lock 986 /* runqueue lock: */ 1031 * it on another CPU. Always updated under the runqueue lock [all...] |
/linux-master/fs/smb/server/ |
H A D | smb2pdu.c | 7077 /* Checking for wrong flag combination during lock request*/ 7117 struct ksmbd_lock *lock; local 7119 lock = kzalloc(sizeof(struct ksmbd_lock), GFP_KERNEL); 7120 if (!lock) 7123 lock->cmd = cmd; 7124 lock->fl = flock; 7125 lock->start = flock->fl_start; 7126 lock->end = flock->fl_end; 7127 lock->flags = flags; 7128 if (lock [all...] |
/linux-master/fs/smb/common/ |
H A D | smb2pdu.h | 846 * the lock sequence number (0 to 64). See MS-SMB2 2.2.26 853 struct smb2_lock_element lock; member in union:smb2_lock_req::__anon91
|
/linux-master/fs/smb/client/ |
H A D | smb2pdu.c | 187 * as cifs_put_tcp_session takes a higher lock 6011 cifs_dbg(FYI, "smb2_lockv num lock %d\n", num_lock); 6067 struct smb2_lock_element lock; local 6069 lock.Offset = cpu_to_le64(offset); 6070 lock.Length = cpu_to_le64(length); 6071 lock.Flags = cpu_to_le32(lock_flags); 6073 lock.Flags |= cpu_to_le32(SMB2_LOCKFLAG_FAIL_IMMEDIATELY); 6075 return smb2_lockv(xid, tcon, persist_fid, volatile_fid, pid, 1, &lock);
|
H A D | smb2ops.c | 195 /* don't log while holding the lock */ 590 /* do not query too frequently, this time with lock held */ 2864 __u64 length, __u32 type, int lock, int unlock, bool wait) 2866 if (unlock && !lock) 2863 smb2_mand_lock(const unsigned int xid, struct cifsFileInfo *cfile, __u64 offset, __u64 length, __u32 type, int lock, int unlock, bool wait) argument
|
H A D | cifsproto.h | 315 extern void cifs_del_lock_waiters(struct cifsLockInfo *lock);
|
H A D | cifsfs.c | 1534 .lock = cifs_lock, 1554 .lock = cifs_lock, 1574 .lock = cifs_lock,
|
/linux-master/fs/bcachefs/ |
H A D | thread_with_file.c | 94 spin_lock_init(&buf->lock); 131 spin_lock_irq(&buf->lock); 143 spin_unlock_irq(&buf->lock); 183 spin_lock(&buf->lock); 195 spin_unlock(&buf->lock); 345 spin_lock(&buf->lock); 352 spin_unlock(&buf->lock); 374 spin_lock(&buf->lock); 387 spin_unlock(&buf->lock); 429 spin_lock_irqsave(&buf->lock, flag [all...] |
H A D | journal_io.c | 103 struct mutex lock; member in struct:journal_list 1088 mutex_lock(&jlist->lock); 1097 mutex_unlock(&jlist->lock); 1151 mutex_lock(&jlist->lock); 1169 mutex_unlock(&jlist->lock); 1204 mutex_lock(&jlist->lock); 1206 mutex_unlock(&jlist->lock); 1224 mutex_init(&jlist.lock); 1568 /* we aren't holding j->lock: */ 1586 spin_lock(&j->lock); [all...] |
H A D | btree_update_interior.c | 239 mutex_lock(&c->btree_cache.lock); 241 mutex_unlock(&c->btree_cache.lock); 254 six_unlock_write(&b->c.lock); 269 struct prealloc_nodes *p = &as->prealloc_nodes[b->c.lock.readers != NULL]; 284 mutex_lock(&c->btree_cache.lock); 287 mutex_unlock(&c->btree_cache.lock); 292 six_unlock_intent(&b->c.lock); 364 six_unlock_write(&b->c.lock); 365 six_unlock_intent(&b->c.lock); 516 six_unlock_write(&b->c.lock); [all...] |
H A D | btree_types.h | 66 struct six_lock lock; member in struct:btree_bkey_cached_common 111 * fails because the lock sequence number has changed - i.e. the 157 struct mutex lock; member in struct:btree_cache 172 * to satisfy the allocation - lock to guarantee only one thread does
|
H A D | btree_node_scan.c | 75 six_unlock_read(&b->c.lock); 166 mutex_lock(&f->lock); 176 mutex_unlock(&f->lock); 333 mutex_init(&f->lock);
|
H A D | btree_io.c | 485 EBUG_ON(!six_lock_counts(&b->c.lock).n[SIX_LOCK_write]); 1753 mutex_lock(&c->btree_cache.lock); 1755 mutex_unlock(&c->btree_cache.lock); 1763 six_unlock_write(&b->c.lock); 1764 six_unlock_intent(&b->c.lock); 1838 six_unlock_read(&b->c.lock); 1991 * We may only have a read lock on the btree node - the dirty bit is our 1992 * "lock" against racing with other threads that may be trying to start 1994 * dirty bit requires a write lock, we can't race with other threads 2223 * Work that must be done with write lock hel [all...] |
H A D | btree_key_cache.c | 56 if (!six_trylock_intent(&ck->c.lock)) 60 six_unlock_intent(&ck->c.lock); 64 if (!six_trylock_write(&ck->c.lock)) { 65 six_unlock_intent(&ck->c.lock); 92 if (ck->c.lock.readers) { 105 six_unlock_write(&ck->c.lock); 106 six_unlock_intent(&ck->c.lock); 134 if (!ck->c.lock.readers) { 149 mutex_lock(&bc->lock); 161 mutex_unlock(&bc->lock); [all...] |
H A D | btree_gc.c | 114 mutex_lock(&c->btree_cache.lock); 122 mutex_unlock(&c->btree_cache.lock); 123 six_unlock_read(&b->c.lock); 202 mutex_lock(&c->btree_cache.lock); 208 mutex_unlock(&c->btree_cache.lock); 390 six_unlock_read(&cur->c.lock); 407 six_unlock_read(&cur->c.lock); 418 six_unlock_read(&prev->c.lock); 449 six_unlock_read(&prev->c.lock); 452 six_unlock_read(&cur->c.lock); [all...] |
/linux-master/drivers/usb/typec/ucsi/ |
H A D | ucsi.c | 223 mutex_lock(&con->lock); 227 mutex_unlock(&con->lock); 241 mutex_unlock(&con->lock); 1168 mutex_lock(&con->lock); 1237 mutex_unlock(&con->lock); 1375 mutex_lock(&con->lock); 1398 mutex_unlock(&con->lock); 1407 mutex_unlock(&con->lock); 1419 mutex_lock(&con->lock); 1440 mutex_unlock(&con->lock); [all...] |
/linux-master/drivers/usb/typec/tcpm/ |
H A D | tcpm.c | 315 struct mutex lock; /* tcpm state machine lock */ member in struct:tcpm_port 404 struct mutex swap_lock; /* swap command lock */ 556 struct mutex logbuffer_lock; /* log buffer access lock */ 970 mutex_unlock(&port->lock); 973 mutex_lock(&port->lock); 1537 WARN_ON(!mutex_is_locked(&port->lock)); 1567 mutex_lock(&port->lock); 1569 mutex_unlock(&port->lock); 2264 * advance because we are dropping the lock bu [all...] |
/linux-master/drivers/usb/misc/ |
H A D | onboard_usb_hub.c | 64 struct mutex lock; member in struct:onboard_hub 124 mutex_lock(&hub->lock); 136 mutex_unlock(&hub->lock); 165 mutex_lock(&hub->lock); 182 mutex_unlock(&hub->lock); 190 mutex_unlock(&hub->lock); 203 mutex_lock(&hub->lock); 213 mutex_unlock(&hub->lock); 294 mutex_init(&hub->lock); 324 mutex_lock(&hub->lock); [all...] |
/linux-master/drivers/usb/host/ |
H A D | xhci-ring.c | 396 * This must be called with command ring stopped and xhci->lock held. 433 /* Must be called with xhci->lock held, releases and aquires lock back */ 481 spin_unlock_irqrestore(&xhci->lock, flags); 484 spin_lock_irqsave(&xhci->lock, flags); 789 * Must be called with xhci->lock held in interrupt context, 790 * releases and re-acquires xhci->lock 986 * We have the xHCI lock, so nothing can modify this list until we drop it. 1271 * Call with xhci->lock held. 1272 * lock i [all...] |
/linux-master/drivers/usb/gadget/udc/ |
H A D | fsl_udc_core.c | 155 __releases(ep->udc->lock) 156 __acquires(ep->udc->lock) 192 spin_unlock(&ep->udc->lock); 196 spin_lock(&ep->udc->lock); 590 spin_lock_irqsave(&udc->lock, flags); 612 spin_unlock_irqrestore(&udc->lock, flags); 657 spin_lock_irqsave(&udc->lock, flags); 664 spin_unlock_irqrestore(&udc->lock, flags); 905 spin_lock_irqsave(&udc->lock, flags); 914 spin_unlock_irqrestore(&udc->lock, flag [all...] |
/linux-master/drivers/usb/gadget/function/ |
H A D | f_ncm.c | 67 spinlock_t lock; member in struct:f_ncm 532 * Context: ncm->lock held 590 spin_unlock(&ncm->lock); 592 spin_lock(&ncm->lock); 600 * Context: ncm->lock held 624 spin_lock(&ncm->lock); 643 spin_unlock(&ncm->lock); 922 spin_lock(&ncm->lock); 924 spin_unlock(&ncm->lock); 1407 spin_lock(&ncm->lock); [all...] |
H A D | f_fs.c | 141 spinlock_t lock; member in struct:ffs_dmabuf_priv 308 __releases(&ffs->ev.waitq.lock) 314 spin_unlock_irq(&ffs->ev.waitq.lock); 320 spin_unlock_irq(&ffs->ev.waitq.lock); 434 spin_lock_irq(&ffs->ev.waitq.lock); 450 spin_unlock_irq(&ffs->ev.waitq.lock); 458 spin_unlock_irq(&ffs->ev.waitq.lock); 466 spin_lock_irq(&ffs->ev.waitq.lock); 484 spin_unlock_irq(&ffs->ev.waitq.lock); 501 /* Called with ffs->ev.waitq.lock an [all...] |
/linux-master/drivers/usb/dwc3/ |
H A D | ep0.c | 201 spin_lock_irqsave(&dwc->lock, flags); 218 spin_unlock_irqrestore(&dwc->lock, flags); 271 spin_lock_irqsave(&dwc->lock, flags); 273 spin_unlock_irqrestore(&dwc->lock, flags); 629 spin_unlock(&dwc->lock); 631 spin_lock(&dwc->lock);
|