/linux-master/drivers/cxl/ |
H A D | cxlmem.h | 271 * @lock: Protect reads of the poison list 282 struct mutex lock; /* Protect reads of poison list */ member in struct:cxl_poison_state
|
/linux-master/drivers/accel/ivpu/ |
H A D | ivpu_mmu.c | 682 mutex_lock(&mmu->lock); 692 mutex_unlock(&mmu->lock); 743 mutex_lock(&mmu->lock); 753 mutex_unlock(&mmu->lock); 793 ret = drmm_mutex_init(&vdev->drm, &mmu->lock); 829 mutex_lock(&mmu->lock); 851 mutex_unlock(&mmu->lock); 856 mutex_unlock(&mmu->lock); 864 mutex_lock(&mmu->lock); 866 mutex_unlock(&mmu->lock); [all...] |
H A D | ivpu_ipc.c | 219 mutex_lock(&ipc->lock); 233 mutex_unlock(&ipc->lock); 504 ret = drmm_mutex_init(&vdev->drm, &ipc->lock); 506 ivpu_err(vdev, "Failed to initialize ipc->lock, ret %d\n", ret); 535 mutex_lock(&ipc->lock); 537 mutex_unlock(&ipc->lock); 548 mutex_lock(&ipc->lock); 550 mutex_unlock(&ipc->lock); 571 mutex_lock(&ipc->lock); 578 mutex_unlock(&ipc->lock); [all...] |
H A D | ivpu_drv.c | 72 mutex_lock(&file_priv->lock); 83 mutex_unlock(&file_priv->lock); 100 mutex_destroy(&file_priv->lock); 234 mutex_init(&file_priv->lock); 263 mutex_destroy(&file_priv->lock);
|
H A D | ivpu_drv.h | 151 struct mutex lock; /* Protects cmdq */ member in struct:ivpu_file_priv
|
/linux-master/fs/btrfs/ |
H A D | transaction.c | 368 spin_lock_init(&cur_trans->delayed_refs.lock); 435 * take the relocation lock in btrfs_record_root_in_trans 450 * lock. smp_wmb() makes sure that all the writes above are 2097 * Currently are holding the fs freeze lock, if we do an async flush 2099 * wait for the fs freeze lock. Using the direct flushing we benefit 2101 * have to re-take the fs freeze lock. 2104 * if it can read lock sb->s_umount. It will always be able to lock it, 2107 * writeback_inodes_sb() while holding a write lock on sb->s_umount. 2177 * waste a bunch of time on lock contentio [all...] |
H A D | root-tree.c | 545 spin_lock(&rsv->lock); 547 spin_unlock(&rsv->lock);
|
H A D | delayed-inode.c | 47 spin_lock_init(&delayed_root->lock); 179 spin_lock(&root->lock); 192 spin_unlock(&root->lock); 199 spin_lock(&root->lock); 208 spin_unlock(&root->lock); 217 spin_lock(&delayed_root->lock); 225 spin_unlock(&delayed_root->lock); 238 spin_lock(&delayed_root->lock); 252 spin_unlock(&delayed_root->lock); 301 spin_lock(&delayed_root->lock); [all...] |
/linux-master/kernel/time/ |
H A D | timer.c | 192 * If multiple bases need to be locked, use the base ordering for lock 208 * @lock: Lock protecting the timer_base 209 * @running_timer: When expiring timers, the lock is dropped. To make 220 * prevents a life lock, when the task which tries to 252 raw_spinlock_t lock; member in struct:timer_base 358 * same lock or cachelines, so we skew each extra cpu with an extra 406 * to lock contention or spurious cache line bouncing. 432 * to lock contention or spurious cache line bouncing. 642 * the base lock: 914 * @key: lockdep class key of the fake lock use [all...] |
H A D | timer_migration.c | 24 * lock contention won't happen across nodes. Depending on the number of 46 * which would result in massive lock bouncing. 69 * time. This spares locking in active path as the lock protects (after 98 * For setting up new groups and handling events it's required to lock both 99 * child and parent group. The lock ordering is always bottom up. This also 102 * the per CPU tmigr_cpu->lock is held. 107 * When @timer_base->lock as well as tmigr related locks are required, the lock 108 * ordering is: first @timer_base->lock, afterwards tmigr related locks. 214 * while holding the lock i [all...] |
/linux-master/arch/x86/events/intel/ |
H A D | ds.c | 242 static inline void pebs_set_tlb_lock(u64 *val, bool tlb, bool lock) argument 255 if (lock) 261 u8 dse, bool tlb, bool lock, bool blk) 270 pebs_set_tlb_lock(&val, tlb, lock); 260 __adl_latency_data_small(struct perf_event *event, u64 status, u8 dse, bool tlb, bool lock, bool blk) argument
|
/linux-master/sound/soc/codecs/ |
H A D | es8326.c | 30 /* The lock protects the situation that an irq is generated 33 struct mutex lock; member in struct:es8326_priv 748 mutex_lock(&es8326->lock); 799 mutex_unlock(&es8326->lock); 809 mutex_lock(&es8326->lock); 908 mutex_unlock(&es8326->lock); 1130 mutex_lock(&es8326->lock); 1136 mutex_unlock(&es8326->lock); 1149 mutex_lock(&es8326->lock); 1155 mutex_unlock(&es8326->lock); [all...] |
/linux-master/net/ipv4/ |
H A D | inet_connection_sock.c | 286 * Caller must hold bhash hashbucket lock with local bh disabled, to protect 308 spin_lock(&head2->lock); 321 spin_unlock(&head2->lock); 379 spin_lock_bh(&head->lock); 386 spin_lock(&head2->lock); 393 spin_unlock(&head2->lock); 399 spin_unlock_bh(&head->lock); 543 spin_lock_bh(&head->lock); 571 spin_lock(&head2->lock); 606 spin_unlock(&head2->lock); 982 spinlock_t *lock = inet_ehash_lockp(hashinfo, req->rsk_hash); local [all...] |
/linux-master/net/core/ |
H A D | sock_map.c | 21 spinlock_t lock; member in struct:bpf_stab 47 spin_lock_init(&stab->lock); 417 spin_lock_bh(&stab->lock); 427 spin_unlock_bh(&stab->lock); 493 spin_lock_bh(&stab->lock); 507 spin_unlock_bh(&stab->lock); 510 spin_unlock_bh(&stab->lock); 845 spinlock_t lock; member in struct:bpf_shtab_bucket 920 spin_lock_bh(&bucket->lock); 928 spin_unlock_bh(&bucket->lock); [all...] |
/linux-master/net/bridge/netfilter/ |
H A D | ebtables.c | 211 read_lock_bh(&table->lock); 233 read_unlock_bh(&table->lock); 254 read_unlock_bh(&table->lock); 258 read_unlock_bh(&table->lock); 282 read_unlock_bh(&table->lock); 294 read_unlock_bh(&table->lock); 312 read_unlock_bh(&table->lock); 315 read_unlock_bh(&table->lock); 1018 * the check on the size is done later, when we have the lock 1053 /* we have the mutex lock, s [all...] |
/linux-master/net/bluetooth/ |
H A D | hci_core.c | 1649 /* This function requires the caller holds hdev->lock */ 1662 /* This function requires the caller holds hdev->lock */ 1679 /* This function requires the caller holds hdev->lock */ 1716 /* This function requires the caller holds hdev->lock */ 1746 /* This function requires the caller holds hdev->lock */ 1804 /* This function requires the caller holds hdev->lock */ 1826 /* This function requires the caller holds hdev->lock */ 1861 /* This function requires the caller holds hdev->lock */ 1915 /* This function requires the caller holds hdev->lock */ 1928 * This function requires the caller holds hdev->lock [all...] |
/linux-master/net/9p/ |
H A D | client.c | 313 spin_lock_irq(&c->lock); 320 spin_unlock_irq(&c->lock); 394 spin_lock_irqsave(&c->lock, flags); 396 spin_unlock_irqrestore(&c->lock, flags); 863 spin_lock_irq(&clnt->lock); 866 spin_unlock_irq(&clnt->lock); 885 spin_lock_irqsave(&clnt->lock, flags); 887 spin_unlock_irqrestore(&clnt->lock, flags); 991 spin_lock_init(&clnt->lock);
|
/linux-master/mm/ |
H A D | vmalloc.c | 841 * rb-tree are part of one entity protected by the lock. Nodes are 848 spinlock_t lock; member in struct:rb_list 1057 spin_lock(&vn->busy.lock); 1063 spin_unlock(&vn->busy.lock); 1074 spin_lock(&vn->busy.lock); 1080 spin_unlock(&vn->busy.lock); 1796 spin_lock(&vn->busy.lock); 1798 spin_unlock(&vn->busy.lock); 1809 preload_this_cpu_lock(spinlock_t *lock, gfp_t gfp_mask, int node) argument 1825 spin_lock(lock); 2457 spinlock_t lock; member in struct:vmap_block_queue 2469 spinlock_t lock; member in struct:vmap_block [all...] |
/linux-master/kernel/bpf/ |
H A D | syscall.c | 256 /* lock 'ptr' and copy everything but lock */ 260 /* mask lock and timer, since value wasn't zero inited */ 843 spin_lock(&map->owner.lock); 846 spin_unlock(&map->owner.lock); 1336 spin_lock_init(&map->owner.lock); 1468 * protected by rcu read lock. 4280 spinlock_t *lock) 4292 spin_lock_bh(lock); 4295 spin_unlock_bh(lock); 4277 bpf_obj_get_next_id(const union bpf_attr *attr, union bpf_attr __user *uattr, struct idr *idr, spinlock_t *lock) argument [all...] |
/linux-master/include/linux/ |
H A D | fs.h | 454 * @nrpages: Number of page entries, protected by the i_pages lock. 768 * inode->i_mutex nesting subclasses for the lock validator: 777 * I_MUTEX_NONDIR2 is for certain operations (such as rename) which lock two 881 * cmpxchg8b without the need of the lock prefix). For SMP compiles 946 rwlock_t lock; /* protects pid, uid, euid fields */ member in struct:fown_struct 986 * f_{lock,count,pos_lock} members can be highly contended and share 987 * the same cacheline. f_{lock,mode} are very frequently used together 1340 struct mutex s_sync_lock; /* sync serialisation lock */ 1790 * Since freeze protection behaves as a lock, users have to preserve 1792 * freeze protection should be the outermost lock 2026 int (*lock) (struct file *, int, struct file_lock *); member in struct:file_operations [all...] |
H A D | bpf.h | 295 spinlock_t lock; member in struct:bpf_map::__anon205 2075 RCU_LOCKDEP_WARN(!rcu_read_lock_held(), "no rcu lock held"); 2153 * these events can happen inside a region which holds a map bucket lock
|
/linux-master/fs/smb/client/ |
H A D | smb1ops.c | 944 __u64 length, __u32 type, int lock, int unlock, bool wait) 947 current->tgid, length, offset, unlock, lock, 943 cifs_mand_lock(const unsigned int xid, struct cifsFileInfo *cfile, __u64 offset, __u64 length, __u32 type, int lock, int unlock, bool wait) argument
|
H A D | file.c | 573 * Delete any outstanding lock records. We'll lose them when the file 1295 /* BB can we lock the filestruct while this is going on? */ 1303 struct cifsLockInfo *lock = local 1305 if (!lock) 1306 return lock; 1307 lock->offset = offset; 1308 lock->length = length; 1309 lock->type = type; 1310 lock->pid = current->tgid; 1311 lock 1318 cifs_del_lock_waiters(struct cifsLockInfo *lock) argument 1428 cifs_lock_add(struct cifsFileInfo *cfile, struct cifsLockInfo *lock) argument 1443 cifs_lock_add_if(struct cifsFileInfo *cfile, struct cifsLockInfo *lock, bool wait) argument 1758 cifs_read_flock(struct file_lock *flock, __u32 *type, int *lock, int *unlock, bool *wait_flag, struct TCP_Server_Info *server) argument 2011 cifs_setlk(struct file *file, struct file_lock *flock, __u32 type, bool wait_flag, bool posix_lck, int lock, int unlock, unsigned int xid) argument 2046 struct cifsLockInfo *lock; local 2108 int lock = 0, unlock = 0; local 2157 int lock = 0, unlock = 0; local [all...] |
/linux-master/fs/reiserfs/ |
H A D | journal.c | 528 /* lock the current transaction */ 690 spinlock_t * lock, void (fn) (struct buffer_chunk *)) 697 if (lock) { 698 spin_unlock(lock); 700 spin_lock(lock); 789 static int write_ordered_buffers(spinlock_t * lock, argument 802 spin_lock(lock); 812 spin_unlock(lock); 817 spin_lock(lock); 831 add_to_chunk(&chunk, bh, lock, write_ordered_chun 689 add_to_chunk(struct buffer_chunk *chunk, struct buffer_head *bh, spinlock_t * lock, void (fn) (struct buffer_chunk *)) argument [all...] |
/linux-master/fs/nfsd/ |
H A D | nfs4state.c | 704 /* hash tables for lock and open owners */ 760 /* ignore lock stateid */ 1098 * is used to manage concurrent access. Testing does not need the lock 1715 * All nfs4 states (open, lock, delegation, layout) held by the server instance 2817 seq_puts(s, ": { type: lock, "); 2823 * Note: a lock stateid isn't really the same thing as a lock, 2825 * may be multiple (or no) lock ranges associated with it. 4749 /* ignore lock owners */ 5910 * open or lock stat 7482 nfs4_transform_lock_offset(struct file_lock *lock) argument 7648 alloc_init_lock_stateowner(unsigned int strhashval, struct nfs4_client *clp, struct nfs4_ol_stateid *open_stp, struct nfsd4_lock *lock) argument 7793 lookup_or_create_lock_state(struct nfsd4_compound_state *cstate, struct nfs4_ol_stateid *ost, struct nfsd4_lock *lock, struct nfs4_ol_stateid **plst, bool *new) argument 7841 struct nfsd4_lock *lock = &u->lock; local 8076 struct nfsd4_lock *lock = &u->lock; local 8087 nfsd_test_lock(struct svc_rqst *rqstp, struct svc_fh *fhp, struct file_lock *lock) argument [all...] |