Searched refs:lock (Results 276 - 300 of 6515) sorted by last modified time

<<11121314151617181920>>

/linux-master/drivers/cxl/
H A Dcxlmem.h271 * @lock: Protect reads of the poison list
282 struct mutex lock; /* Protect reads of poison list */ member in struct:cxl_poison_state
/linux-master/drivers/accel/ivpu/
H A Divpu_mmu.c682 mutex_lock(&mmu->lock);
692 mutex_unlock(&mmu->lock);
743 mutex_lock(&mmu->lock);
753 mutex_unlock(&mmu->lock);
793 ret = drmm_mutex_init(&vdev->drm, &mmu->lock);
829 mutex_lock(&mmu->lock);
851 mutex_unlock(&mmu->lock);
856 mutex_unlock(&mmu->lock);
864 mutex_lock(&mmu->lock);
866 mutex_unlock(&mmu->lock);
[all...]
H A Divpu_ipc.c219 mutex_lock(&ipc->lock);
233 mutex_unlock(&ipc->lock);
504 ret = drmm_mutex_init(&vdev->drm, &ipc->lock);
506 ivpu_err(vdev, "Failed to initialize ipc->lock, ret %d\n", ret);
535 mutex_lock(&ipc->lock);
537 mutex_unlock(&ipc->lock);
548 mutex_lock(&ipc->lock);
550 mutex_unlock(&ipc->lock);
571 mutex_lock(&ipc->lock);
578 mutex_unlock(&ipc->lock);
[all...]
H A Divpu_drv.c72 mutex_lock(&file_priv->lock);
83 mutex_unlock(&file_priv->lock);
100 mutex_destroy(&file_priv->lock);
234 mutex_init(&file_priv->lock);
263 mutex_destroy(&file_priv->lock);
H A Divpu_drv.h151 struct mutex lock; /* Protects cmdq */ member in struct:ivpu_file_priv
/linux-master/fs/btrfs/
H A Dtransaction.c368 spin_lock_init(&cur_trans->delayed_refs.lock);
435 * take the relocation lock in btrfs_record_root_in_trans
450 * lock. smp_wmb() makes sure that all the writes above are
2097 * Currently are holding the fs freeze lock, if we do an async flush
2099 * wait for the fs freeze lock. Using the direct flushing we benefit
2101 * have to re-take the fs freeze lock.
2104 * if it can read lock sb->s_umount. It will always be able to lock it,
2107 * writeback_inodes_sb() while holding a write lock on sb->s_umount.
2177 * waste a bunch of time on lock contentio
[all...]
H A Droot-tree.c545 spin_lock(&rsv->lock);
547 spin_unlock(&rsv->lock);
H A Ddelayed-inode.c47 spin_lock_init(&delayed_root->lock);
179 spin_lock(&root->lock);
192 spin_unlock(&root->lock);
199 spin_lock(&root->lock);
208 spin_unlock(&root->lock);
217 spin_lock(&delayed_root->lock);
225 spin_unlock(&delayed_root->lock);
238 spin_lock(&delayed_root->lock);
252 spin_unlock(&delayed_root->lock);
301 spin_lock(&delayed_root->lock);
[all...]
/linux-master/kernel/time/
H A Dtimer.c192 * If multiple bases need to be locked, use the base ordering for lock
208 * @lock: Lock protecting the timer_base
209 * @running_timer: When expiring timers, the lock is dropped. To make
220 * prevents a life lock, when the task which tries to
252 raw_spinlock_t lock; member in struct:timer_base
358 * same lock or cachelines, so we skew each extra cpu with an extra
406 * to lock contention or spurious cache line bouncing.
432 * to lock contention or spurious cache line bouncing.
642 * the base lock:
914 * @key: lockdep class key of the fake lock use
[all...]
H A Dtimer_migration.c24 * lock contention won't happen across nodes. Depending on the number of
46 * which would result in massive lock bouncing.
69 * time. This spares locking in active path as the lock protects (after
98 * For setting up new groups and handling events it's required to lock both
99 * child and parent group. The lock ordering is always bottom up. This also
102 * the per CPU tmigr_cpu->lock is held.
107 * When @timer_base->lock as well as tmigr related locks are required, the lock
108 * ordering is: first @timer_base->lock, afterwards tmigr related locks.
214 * while holding the lock i
[all...]
/linux-master/arch/x86/events/intel/
H A Dds.c242 static inline void pebs_set_tlb_lock(u64 *val, bool tlb, bool lock) argument
255 if (lock)
261 u8 dse, bool tlb, bool lock, bool blk)
270 pebs_set_tlb_lock(&val, tlb, lock);
260 __adl_latency_data_small(struct perf_event *event, u64 status, u8 dse, bool tlb, bool lock, bool blk) argument
/linux-master/sound/soc/codecs/
H A Des8326.c30 /* The lock protects the situation that an irq is generated
33 struct mutex lock; member in struct:es8326_priv
748 mutex_lock(&es8326->lock);
799 mutex_unlock(&es8326->lock);
809 mutex_lock(&es8326->lock);
908 mutex_unlock(&es8326->lock);
1130 mutex_lock(&es8326->lock);
1136 mutex_unlock(&es8326->lock);
1149 mutex_lock(&es8326->lock);
1155 mutex_unlock(&es8326->lock);
[all...]
/linux-master/net/ipv4/
H A Dinet_connection_sock.c286 * Caller must hold bhash hashbucket lock with local bh disabled, to protect
308 spin_lock(&head2->lock);
321 spin_unlock(&head2->lock);
379 spin_lock_bh(&head->lock);
386 spin_lock(&head2->lock);
393 spin_unlock(&head2->lock);
399 spin_unlock_bh(&head->lock);
543 spin_lock_bh(&head->lock);
571 spin_lock(&head2->lock);
606 spin_unlock(&head2->lock);
982 spinlock_t *lock = inet_ehash_lockp(hashinfo, req->rsk_hash); local
[all...]
/linux-master/net/core/
H A Dsock_map.c21 spinlock_t lock; member in struct:bpf_stab
47 spin_lock_init(&stab->lock);
417 spin_lock_bh(&stab->lock);
427 spin_unlock_bh(&stab->lock);
493 spin_lock_bh(&stab->lock);
507 spin_unlock_bh(&stab->lock);
510 spin_unlock_bh(&stab->lock);
845 spinlock_t lock; member in struct:bpf_shtab_bucket
920 spin_lock_bh(&bucket->lock);
928 spin_unlock_bh(&bucket->lock);
[all...]
/linux-master/net/bridge/netfilter/
H A Debtables.c211 read_lock_bh(&table->lock);
233 read_unlock_bh(&table->lock);
254 read_unlock_bh(&table->lock);
258 read_unlock_bh(&table->lock);
282 read_unlock_bh(&table->lock);
294 read_unlock_bh(&table->lock);
312 read_unlock_bh(&table->lock);
315 read_unlock_bh(&table->lock);
1018 * the check on the size is done later, when we have the lock
1053 /* we have the mutex lock, s
[all...]
/linux-master/net/bluetooth/
H A Dhci_core.c1649 /* This function requires the caller holds hdev->lock */
1662 /* This function requires the caller holds hdev->lock */
1679 /* This function requires the caller holds hdev->lock */
1716 /* This function requires the caller holds hdev->lock */
1746 /* This function requires the caller holds hdev->lock */
1804 /* This function requires the caller holds hdev->lock */
1826 /* This function requires the caller holds hdev->lock */
1861 /* This function requires the caller holds hdev->lock */
1915 /* This function requires the caller holds hdev->lock */
1928 * This function requires the caller holds hdev->lock
[all...]
/linux-master/net/9p/
H A Dclient.c313 spin_lock_irq(&c->lock);
320 spin_unlock_irq(&c->lock);
394 spin_lock_irqsave(&c->lock, flags);
396 spin_unlock_irqrestore(&c->lock, flags);
863 spin_lock_irq(&clnt->lock);
866 spin_unlock_irq(&clnt->lock);
885 spin_lock_irqsave(&clnt->lock, flags);
887 spin_unlock_irqrestore(&clnt->lock, flags);
991 spin_lock_init(&clnt->lock);
/linux-master/mm/
H A Dvmalloc.c841 * rb-tree are part of one entity protected by the lock. Nodes are
848 spinlock_t lock; member in struct:rb_list
1057 spin_lock(&vn->busy.lock);
1063 spin_unlock(&vn->busy.lock);
1074 spin_lock(&vn->busy.lock);
1080 spin_unlock(&vn->busy.lock);
1796 spin_lock(&vn->busy.lock);
1798 spin_unlock(&vn->busy.lock);
1809 preload_this_cpu_lock(spinlock_t *lock, gfp_t gfp_mask, int node) argument
1825 spin_lock(lock);
2457 spinlock_t lock; member in struct:vmap_block_queue
2469 spinlock_t lock; member in struct:vmap_block
[all...]
/linux-master/kernel/bpf/
H A Dsyscall.c256 /* lock 'ptr' and copy everything but lock */
260 /* mask lock and timer, since value wasn't zero inited */
843 spin_lock(&map->owner.lock);
846 spin_unlock(&map->owner.lock);
1336 spin_lock_init(&map->owner.lock);
1468 * protected by rcu read lock.
4280 spinlock_t *lock)
4292 spin_lock_bh(lock);
4295 spin_unlock_bh(lock);
4277 bpf_obj_get_next_id(const union bpf_attr *attr, union bpf_attr __user *uattr, struct idr *idr, spinlock_t *lock) argument
[all...]
/linux-master/include/linux/
H A Dfs.h454 * @nrpages: Number of page entries, protected by the i_pages lock.
768 * inode->i_mutex nesting subclasses for the lock validator:
777 * I_MUTEX_NONDIR2 is for certain operations (such as rename) which lock two
881 * cmpxchg8b without the need of the lock prefix). For SMP compiles
946 rwlock_t lock; /* protects pid, uid, euid fields */ member in struct:fown_struct
986 * f_{lock,count,pos_lock} members can be highly contended and share
987 * the same cacheline. f_{lock,mode} are very frequently used together
1340 struct mutex s_sync_lock; /* sync serialisation lock */
1790 * Since freeze protection behaves as a lock, users have to preserve
1792 * freeze protection should be the outermost lock
2026 int (*lock) (struct file *, int, struct file_lock *); member in struct:file_operations
[all...]
H A Dbpf.h295 spinlock_t lock; member in struct:bpf_map::__anon205
2075 RCU_LOCKDEP_WARN(!rcu_read_lock_held(), "no rcu lock held");
2153 * these events can happen inside a region which holds a map bucket lock
/linux-master/fs/smb/client/
H A Dsmb1ops.c944 __u64 length, __u32 type, int lock, int unlock, bool wait)
947 current->tgid, length, offset, unlock, lock,
943 cifs_mand_lock(const unsigned int xid, struct cifsFileInfo *cfile, __u64 offset, __u64 length, __u32 type, int lock, int unlock, bool wait) argument
H A Dfile.c573 * Delete any outstanding lock records. We'll lose them when the file
1295 /* BB can we lock the filestruct while this is going on? */
1303 struct cifsLockInfo *lock = local
1305 if (!lock)
1306 return lock;
1307 lock->offset = offset;
1308 lock->length = length;
1309 lock->type = type;
1310 lock->pid = current->tgid;
1311 lock
1318 cifs_del_lock_waiters(struct cifsLockInfo *lock) argument
1428 cifs_lock_add(struct cifsFileInfo *cfile, struct cifsLockInfo *lock) argument
1443 cifs_lock_add_if(struct cifsFileInfo *cfile, struct cifsLockInfo *lock, bool wait) argument
1758 cifs_read_flock(struct file_lock *flock, __u32 *type, int *lock, int *unlock, bool *wait_flag, struct TCP_Server_Info *server) argument
2011 cifs_setlk(struct file *file, struct file_lock *flock, __u32 type, bool wait_flag, bool posix_lck, int lock, int unlock, unsigned int xid) argument
2046 struct cifsLockInfo *lock; local
2108 int lock = 0, unlock = 0; local
2157 int lock = 0, unlock = 0; local
[all...]
/linux-master/fs/reiserfs/
H A Djournal.c528 /* lock the current transaction */
690 spinlock_t * lock, void (fn) (struct buffer_chunk *))
697 if (lock) {
698 spin_unlock(lock);
700 spin_lock(lock);
789 static int write_ordered_buffers(spinlock_t * lock, argument
802 spin_lock(lock);
812 spin_unlock(lock);
817 spin_lock(lock);
831 add_to_chunk(&chunk, bh, lock, write_ordered_chun
689 add_to_chunk(struct buffer_chunk *chunk, struct buffer_head *bh, spinlock_t * lock, void (fn) (struct buffer_chunk *)) argument
[all...]
/linux-master/fs/nfsd/
H A Dnfs4state.c704 /* hash tables for lock and open owners */
760 /* ignore lock stateid */
1098 * is used to manage concurrent access. Testing does not need the lock
1715 * All nfs4 states (open, lock, delegation, layout) held by the server instance
2817 seq_puts(s, ": { type: lock, ");
2823 * Note: a lock stateid isn't really the same thing as a lock,
2825 * may be multiple (or no) lock ranges associated with it.
4749 /* ignore lock owners */
5910 * open or lock stat
7482 nfs4_transform_lock_offset(struct file_lock *lock) argument
7648 alloc_init_lock_stateowner(unsigned int strhashval, struct nfs4_client *clp, struct nfs4_ol_stateid *open_stp, struct nfsd4_lock *lock) argument
7793 lookup_or_create_lock_state(struct nfsd4_compound_state *cstate, struct nfs4_ol_stateid *ost, struct nfsd4_lock *lock, struct nfs4_ol_stateid **plst, bool *new) argument
7841 struct nfsd4_lock *lock = &u->lock; local
8076 struct nfsd4_lock *lock = &u->lock; local
8087 nfsd_test_lock(struct svc_rqst *rqstp, struct svc_fh *fhp, struct file_lock *lock) argument
[all...]

Completed in 672 milliseconds

<<11121314151617181920>>