/linux-master/tools/virtio/ringtest/ |
H A D | ptr_ring.c | 64 static void spin_lock_init(spinlock_t *lock) argument 66 int r = pthread_spin_init(lock, 0); 70 static void spin_lock(spinlock_t *lock) argument 72 int ret = pthread_spin_lock(lock); 76 static void spin_unlock(spinlock_t *lock) argument 78 int ret = pthread_spin_unlock(lock); 82 static void spin_lock_bh(spinlock_t *lock) argument 84 spin_lock(lock); 87 static void spin_unlock_bh(spinlock_t *lock) argument 89 spin_unlock(lock); 92 spin_lock_irq(spinlock_t *lock) argument 97 spin_unlock_irq(spinlock_t *lock) argument 102 spin_lock_irqsave(spinlock_t *lock, unsigned long f) argument 107 spin_unlock_irqrestore(spinlock_t *lock, unsigned long f) argument [all...] |
/linux-master/drivers/clk/berlin/ |
H A D | berlin2-div.c | 58 spinlock_t *lock; member in struct:berlin2_div 71 if (div->lock) 72 spin_lock(div->lock); 77 if (div->lock) 78 spin_unlock(div->lock); 89 if (div->lock) 90 spin_lock(div->lock); 96 if (div->lock) 97 spin_unlock(div->lock); 108 if (div->lock) 229 berlin2_div_register(const struct berlin2_div_map *map, void __iomem *base, const char *name, u8 div_flags, const char **parent_names, int num_parents, unsigned long flags, spinlock_t *lock) argument [all...] |
/linux-master/kernel/ |
H A D | cpu_pm.c | 23 raw_spinlock_t lock; member in struct:__anon4563 26 .lock = __RAW_SPIN_LOCK_UNLOCKED(cpu_pm_notifier.lock), 45 raw_spin_lock_irqsave(&cpu_pm_notifier.lock, flags); 47 raw_spin_unlock_irqrestore(&cpu_pm_notifier.lock, flags); 66 raw_spin_lock_irqsave(&cpu_pm_notifier.lock, flags); 68 raw_spin_unlock_irqrestore(&cpu_pm_notifier.lock, flags); 86 raw_spin_lock_irqsave(&cpu_pm_notifier.lock, flags); 88 raw_spin_unlock_irqrestore(&cpu_pm_notifier.lock, flags);
|
/linux-master/tools/perf/tests/shell/ |
H A D | lock_contention.sh | 2 # kernel lock contention analysis test 30 if ! perf list | grep -q lock:contention_begin; then 31 echo "[Skip] No lock contention tracepoints" 38 echo "[Skip] Low number of CPUs (`nproc`), lock event cannot be triggered certainly" 46 echo "Testing perf lock record and perf lock contention" 47 perf lock record -o ${perfdata} -- perf bench sched messaging > /dev/null 2>&1 49 perf lock contention -i ${perfdata} -E 1 -q 2> ${result} 59 echo "Testing perf lock contention --use-bpf" 61 if ! perf lock co [all...] |
/linux-master/include/linux/ |
H A D | blockgroup_lock.h | 20 spinlock_t lock; member in struct:bgl_lock 32 spin_lock_init(&bgl->locks[i].lock); 38 return &bgl->locks[block_group & (NR_BG_LOCKS-1)].lock;
|
H A D | kref.h | 73 struct mutex *lock) 75 if (refcount_dec_and_mutex_lock(&kref->refcount, lock)) { 84 spinlock_t *lock) 86 if (refcount_dec_and_lock(&kref->refcount, lock)) { 102 * Operations on such objects require at least a read lock around 103 * lookup + kref_get, and a write lock around kref_put + remove from lookup 71 kref_put_mutex(struct kref *kref, void (*release)(struct kref *kref), struct mutex *lock) argument 82 kref_put_lock(struct kref *kref, void (*release)(struct kref *kref), spinlock_t *lock) argument
|
/linux-master/drivers/clk/hisilicon/ |
H A D | clkgate-separated.c | 29 spinlock_t *lock; member in struct:clkgate_separated 39 if (sclk->lock) 40 spin_lock_irqsave(sclk->lock, flags); 44 if (sclk->lock) 45 spin_unlock_irqrestore(sclk->lock, flags); 56 if (sclk->lock) 57 spin_lock_irqsave(sclk->lock, flags); 61 if (sclk->lock) 62 spin_unlock_irqrestore(sclk->lock, flags); 87 u8 clk_gate_flags, spinlock_t *lock) 83 hisi_register_clkgate_sep(struct device *dev, const char *name, const char *parent_name, unsigned long flags, void __iomem *reg, u8 bit_idx, u8 clk_gate_flags, spinlock_t *lock) argument [all...] |
/linux-master/drivers/clk/mmp/ |
H A D | pwr-island.c | 19 spinlock_t *lock; member in struct:mmp_pm_domain 32 if (pm_domain->lock) 33 spin_lock_irqsave(pm_domain->lock, flags); 61 if (pm_domain->lock) 62 spin_unlock_irqrestore(pm_domain->lock, flags); 76 if (pm_domain->lock) 77 spin_lock_irqsave(pm_domain->lock, flags); 85 if (pm_domain->lock) 86 spin_unlock_irqrestore(pm_domain->lock, flags); 94 unsigned int flags, spinlock_t *lock) 91 mmp_pm_domain_register(const char *name, void __iomem *reg, u32 power_on, u32 reset, u32 clock_enable, unsigned int flags, spinlock_t *lock) argument [all...] |
/linux-master/lib/ |
H A D | lockref.c | 16 while (likely(arch_spin_value_unlocked(old.lock.rlock.raw_lock))) { \ 50 spin_lock(&lockref->lock); 52 spin_unlock(&lockref->lock); 73 spin_lock(&lockref->lock); 79 spin_unlock(&lockref->lock); 101 spin_lock(&lockref->lock); 107 spin_unlock(&lockref->lock); 135 * Return: 1 if count updated successfully or 0 if count <= 1 and lock taken 147 spin_lock(&lockref->lock); 151 spin_unlock(&lockref->lock); [all...] |
/linux-master/drivers/s390/scsi/ |
H A D | zfcp_reqlist.h | 21 * @lock: Spinlock for protecting the hash list 25 spinlock_t lock; member in struct:zfcp_reqlist 49 spin_lock_init(&rl->lock); 112 spin_lock_irqsave(&rl->lock, flags); 114 spin_unlock_irqrestore(&rl->lock, flags); 125 * id and then removes it from the reqlist. The reqlist lock is held 137 spin_lock_irqsave(&rl->lock, flags); 141 spin_unlock_irqrestore(&rl->lock, flags); 164 spin_lock_irqsave(&rl->lock, flags); 166 spin_unlock_irqrestore(&rl->lock, flag [all...] |
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | freplace_attach_probe.c | 12 struct bpf_spin_lock lock; member in struct:hmap_elem 33 bpf_spin_lock(&val->lock); 35 bpf_spin_unlock(&val->lock);
|
H A D | test_spin_lock_fail.c | 8 struct bpf_spin_lock lock; member in struct:foo 113 CHECK(kptr_kptr, &f1->lock, &f2->lock); 114 CHECK(kptr_global, &f1->lock, &lockA); 115 CHECK(kptr_mapval, &f1->lock, &v->lock); 116 CHECK(kptr_innermapval, &f1->lock, &iv->lock); 119 CHECK(global_kptr, &lockA, &f1->lock); 120 CHECK(global_mapval, &lockA, &v->lock); [all...] |
/linux-master/arch/s390/include/asm/ |
H A D | mmu.h | 10 spinlock_t lock; member in struct:__anon4 24 * written once, they can be read without a lock. 40 .context.lock = __SPIN_LOCK_UNLOCKED(name.context.lock), \
|
/linux-master/fs/ocfs2/dlm/ |
H A D | dlmunlock.c | 43 struct dlm_lock *lock, 48 struct dlm_lock *lock, 54 struct dlm_lock *lock, 67 * So to unlock a converting lock, you must first cancel the 76 * taken: res->spinlock and lock->spinlock taken and dropped 79 * all callers should have taken an extra ref on lock coming in 83 struct dlm_lock *lock, 103 /* We want to be sure that we're not freeing a lock 105 in_use = !list_empty(&lock->ast_list); 125 spin_lock(&lock 81 dlmunlock_common(struct dlm_ctxt *dlm, struct dlm_lock_resource *res, struct dlm_lock *lock, struct dlm_lockstatus *lksb, int flags, int *call_ast, int master_node) argument 265 dlm_commit_pending_unlock(struct dlm_lock_resource *res, struct dlm_lock *lock) argument 273 dlm_commit_pending_cancel(struct dlm_lock_resource *res, struct dlm_lock *lock) argument 281 dlmunlock_master(struct dlm_ctxt *dlm, struct dlm_lock_resource *res, struct dlm_lock *lock, struct dlm_lockstatus *lksb, int flags, int *call_ast) argument 291 dlmunlock_remote(struct dlm_ctxt *dlm, struct dlm_lock_resource *res, struct dlm_lock *lock, struct dlm_lockstatus *lksb, int flags, int *call_ast) argument 307 dlm_send_remote_unlock_request(struct dlm_ctxt *dlm, struct dlm_lock_resource *res, struct dlm_lock *lock, struct dlm_lockstatus *lksb, int flags, u8 owner) argument 395 struct dlm_lock *lock = NULL, *iter; local 525 dlm_get_cancel_actions(struct dlm_ctxt *dlm, struct dlm_lock_resource *res, struct dlm_lock *lock, struct dlm_lockstatus *lksb, int *actions) argument 557 dlm_get_unlock_actions(struct dlm_ctxt *dlm, struct dlm_lock_resource *res, struct dlm_lock *lock, struct dlm_lockstatus *lksb, int *actions) argument 589 struct dlm_lock *lock = NULL; local [all...] |
/linux-master/drivers/base/ |
H A D | map.c | 26 int (*lock)(dev_t, void *); member in struct:kobj_map::probe 29 struct mutex *lock; member in struct:kobj_map 34 int (*lock)(dev_t, void *), void *data) 51 p->lock = lock; 56 mutex_lock(domain->lock); 64 mutex_unlock(domain->lock); 78 mutex_lock(domain->lock); 91 mutex_unlock(domain->lock); 102 mutex_lock(domain->lock); 32 kobj_map(struct kobj_map *domain, dev_t dev, unsigned long range, struct module *module, kobj_probe_t *probe, int (*lock)(dev_t, void *), void *data) argument 135 kobj_map_init(kobj_probe_t *base_probe, struct mutex *lock) argument [all...] |
/linux-master/fs/ |
H A D | fs_struct.c | 20 spin_lock(&fs->lock); 25 spin_unlock(&fs->lock); 39 spin_lock(&fs->lock); 44 spin_unlock(&fs->lock); 70 spin_lock(&fs->lock); 79 spin_unlock(&fs->lock); 102 spin_lock(&fs->lock); 105 spin_unlock(&fs->lock); 115 /* We don't need to lock fs - think why ;-) */ 119 spin_lock_init(&fs->lock); [all...] |
/linux-master/fs/lockd/ |
H A D | xdr.c | 5 * XDR support for lockd and the lock client. 72 svcxdr_decode_lock(struct xdr_stream *xdr, struct nlm_lock *lock) argument 74 struct file_lock *fl = &lock->fl; 77 if (!svcxdr_decode_string(xdr, &lock->caller, &lock->len)) 79 if (!svcxdr_decode_fhandle(xdr, &lock->fh)) 81 if (!svcxdr_decode_owner(xdr, &lock->oh)) 83 if (xdr_stream_decode_u32(xdr, &lock->svid) < 0) 104 svcxdr_encode_holder(struct xdr_stream *xdr, const struct nlm_lock *lock) argument 106 const struct file_lock *fl = &lock 274 struct nlm_lock *lock = &argp->lock; local 301 struct nlm_lock *lock = &argp->lock; local [all...] |
/linux-master/include/linux/clk/ |
H A D | zynq.h | 16 spinlock_t *lock);
|
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nv10_fence.h | 16 spinlock_t lock; member in struct:nv10_fence_priv
|
/linux-master/drivers/usb/gadget/function/ |
H A D | u_printer.h | 29 struct mutex lock; member in struct:f_printer_opts
|
H A D | u_gether.h | 29 struct mutex lock; member in struct:f_gether_opts
|
H A D | u_eem.h | 29 struct mutex lock; member in struct:f_eem_opts
|
H A D | u_ecm.h | 29 struct mutex lock; member in struct:f_ecm_opts
|
/linux-master/drivers/staging/media/atomisp/pci/runtime/tagger/interface/ |
H A D | ia_css_tagger_common.h | 36 u8 lock; /* the lock on the element */ member in struct:__anon222
|
/linux-master/kernel/locking/ |
H A D | mcs_spinlock.h | 3 * MCS lock defines 5 * This file contains the main data structure and API definitions of MCS lock. 7 * The MCS lock (proposed by Mellor-Crummey and Scott) is a simple spin-lock 9 * to acquire the lock spinning on a local variable. 10 * It avoids expensive cache bounces that common test-and-set spin-lock 20 int locked; /* 1 if lock acquired */ 28 * lock is acquired. Additionally, some architectures such as 58 * In order to acquire the lock, the caller should declare a local node and 59 * pass a reference of the node to this function in addition to the lock 65 mcs_spin_lock(struct mcs_spinlock **lock, struct mcs_spinlock *node) argument 102 mcs_spin_unlock(struct mcs_spinlock **lock, struct mcs_spinlock *node) argument [all...] |