/linux-master/kernel/ |
H A D | audit_tree.c | 53 * association between mark and chunk is protected by hash_lock and 68 * tree.chunks anchors chunk.owners[].list hash_lock 70 * chunk.trees anchors tree.same_root hash_lock 72 * a hash function. RCU, hash_lock 210 static __cacheline_aligned_in_smp DEFINE_SPINLOCK(hash_lock); 225 /* hash_lock & mark->group->mark_mutex is held by caller */ 284 assert_spin_locked(&hash_lock); 357 * whether it didn't change while we've dropped hash_lock. 365 spin_lock(&hash_lock); 369 spin_unlock(&hash_lock); [all...] |
/linux-master/drivers/md/dm-vdo/ |
H A D | dedupe.c | 11 * A hash_lock controls and coordinates writing, index access, and dedupe among groups of data_vios 15 * index query is needed for each hash_lock, instead of one for every data_vio. 20 * more fine-grained locking for the hash_lock structures. 22 * A hash_lock acts like a state machine perhaps more than as a lock. Other than the starting and 49 * lock, releasing the hash_lock itself back to the hash zone (BYPASSING). 92 * When a hash_lock needs to query the index, it attempts to acquire an unused dedupe_context from 94 * hash_lock's agent, added to the list of pending contexts, and then sent to the index. The 113 * If a hash_lock needs a dedupe context, and the available list is empty, the timed_out list will 223 struct hash_lock { struct 347 static void return_hash_lock_to_pool(struct hash_zone *zone, struct hash_lock *loc 415 set_duplicate_lock(struct hash_lock *hash_lock, struct pbn_lock *pbn_lock) argument 1963 struct hash_lock *hash_lock = data_vio->hash_lock; local [all...] |
H A D | dedupe.h | 30 struct hash_lock; 55 struct hash_lock *lock_array;
|
H A D | statistics.h | 240 struct hash_lock_statistics hash_lock; member in struct:vdo_statistics
|
H A D | data-vio.h | 224 struct hash_lock *hash_lock; member in struct:data_vio
|
H A D | data-vio.c | 1279 VDO_ASSERT_LOG_ONLY(data_vio->hash_lock == NULL, 1307 if (data_vio->hash_lock != NULL) { 1638 if (data_vio->hash_lock != NULL) 1799 VDO_ASSERT_LOG_ONLY(data_vio->hash_lock != NULL, 1800 "data_vio to compress has a hash_lock");
|
/linux-master/net/netfilter/ |
H A D | xt_RATEEST.c | 23 struct mutex hash_lock; member in struct:xt_rateest_net 68 mutex_lock(&xn->hash_lock); 70 mutex_unlock(&xn->hash_lock); 79 mutex_lock(&xn->hash_lock); 89 mutex_unlock(&xn->hash_lock); 123 mutex_lock(&xn->hash_lock); 126 mutex_unlock(&xn->hash_lock); 165 mutex_unlock(&xn->hash_lock); 171 mutex_unlock(&xn->hash_lock); 199 mutex_init(&xn->hash_lock); [all...] |
/linux-master/drivers/md/ |
H A D | dm-region-hash.c | 49 * A rw spin lock 'hash_lock' protects just the hash table, 68 rwlock_t hash_lock; member in struct:dm_region_hash 202 rwlock_init(&rh->hash_lock); 307 write_lock_irq(&rh->hash_lock); 322 write_unlock_irq(&rh->hash_lock); 333 read_unlock(&rh->hash_lock); 335 read_lock(&rh->hash_lock); 346 read_lock(&rh->hash_lock); 348 read_unlock(&rh->hash_lock); 417 read_lock(&rh->hash_lock); [all...] |
/linux-master/fs/hfs/ |
H A D | bnode.c | 269 spin_lock(&tree->hash_lock); 278 spin_unlock(&tree->hash_lock); 283 spin_unlock(&tree->hash_lock); 324 spin_lock(&tree->hash_lock); 328 spin_unlock(&tree->hash_lock); 334 spin_unlock(&tree->hash_lock); 417 spin_lock(&tree->hash_lock); 419 spin_unlock(&tree->hash_lock); 468 if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock)) 478 spin_unlock(&tree->hash_lock); [all...] |
H A D | btree.h | 47 spinlock_t hash_lock; member in struct:hfs_btree
|
H A D | inode.c | 94 spin_lock(&tree->hash_lock); 104 spin_unlock(&tree->hash_lock); 108 spin_lock(&tree->hash_lock); 120 spin_unlock(&tree->hash_lock);
|
H A D | btree.c | 32 spin_lock_init(&tree->hash_lock);
|
/linux-master/net/bridge/ |
H A D | br_fdb.c | 226 /* requires bridge hash_lock */ 233 lockdep_assert_held_once(&br->hash_lock); 381 spin_lock_bh(&br->hash_lock); 386 spin_unlock_bh(&br->hash_lock); 470 spin_lock_bh(&br->hash_lock); 502 spin_unlock_bh(&br->hash_lock); 511 spin_lock_bh(&br->hash_lock); 537 spin_unlock_bh(&br->hash_lock); 573 spin_lock_bh(&br->hash_lock); 576 spin_unlock_bh(&br->hash_lock); [all...] |
/linux-master/drivers/net/ppp/ |
H A D | pppoe.c | 104 rwlock_t hash_lock; member in struct:pppoe_net 228 read_lock_bh(&pn->hash_lock); 232 read_unlock_bh(&pn->hash_lock); 261 write_lock_bh(&pn->hash_lock); 263 write_unlock_bh(&pn->hash_lock); 279 write_lock_bh(&pn->hash_lock); 295 * hash_lock, in that order. Since we should hold the 303 write_unlock_bh(&pn->hash_lock); 323 write_lock_bh(&pn->hash_lock); 327 write_unlock_bh(&pn->hash_lock); [all...] |
/linux-master/net/sunrpc/ |
H A D | cache.c | 86 /* Must be called under cd->hash_lock */ 116 spin_lock(&detail->hash_lock); 120 lockdep_is_held(&detail->hash_lock)) { 131 spin_unlock(&detail->hash_lock); 139 spin_unlock(&detail->hash_lock); 212 spin_lock(&detail->hash_lock); 216 spin_unlock(&detail->hash_lock); 220 spin_unlock(&detail->hash_lock); 231 spin_lock(&detail->hash_lock); 238 spin_unlock(&detail->hash_lock); [all...] |
/linux-master/fs/hfsplus/ |
H A D | bnode.c | 419 spin_lock(&tree->hash_lock); 427 spin_unlock(&tree->hash_lock); 433 spin_unlock(&tree->hash_lock); 474 spin_lock(&tree->hash_lock); 478 spin_unlock(&tree->hash_lock); 485 spin_unlock(&tree->hash_lock); 568 spin_lock(&tree->hash_lock); 570 spin_unlock(&tree->hash_lock); 619 if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock)) 629 spin_unlock(&tree->hash_lock); [all...] |
H A D | inode.c | 90 spin_lock(&tree->hash_lock); 100 spin_unlock(&tree->hash_lock); 105 spin_lock(&tree->hash_lock); 117 spin_unlock(&tree->hash_lock);
|
H A D | btree.c | 147 spin_lock_init(&tree->hash_lock);
|
H A D | hfsplus_fs.h | 95 spinlock_t hash_lock; member in struct:hfs_btree
|
/linux-master/drivers/net/vxlan/ |
H A D | vxlan_core.c | 444 /* caller should hold vxlan->hash_lock */ 524 spin_lock_bh(&vxlan->hash_lock[h]); 536 spin_unlock_bh(&vxlan->hash_lock[h]); 541 spin_unlock_bh(&vxlan->hash_lock[h]); 558 spin_lock_bh(&vxlan->hash_lock[h]); 563 spin_unlock_bh(&vxlan->hash_lock[h]); 1277 spin_lock_bh(&vxlan->hash_lock[hash_index]); 1282 spin_unlock_bh(&vxlan->hash_lock[hash_index]); 1340 spin_lock_bh(&vxlan->hash_lock[hash_index]); 1343 spin_unlock_bh(&vxlan->hash_lock[hash_inde [all...] |
H A D | vxlan_vnifilter.c | 485 spin_lock_bh(&vxlan->hash_lock[hash_index]); 497 spin_unlock_bh(&vxlan->hash_lock[hash_index]); 510 spin_unlock_bh(&vxlan->hash_lock[hash_index]);
|
/linux-master/kernel/time/ |
H A D | posix-timers.c | 50 static DEFINE_SPINLOCK(hash_lock); 81 hlist_for_each_entry_rcu(timer, head, t_hash, lockdep_is_held(&hash_lock)) { 108 spin_lock(&hash_lock); 117 spin_unlock(&hash_lock); 120 spin_unlock(&hash_lock); 431 spin_lock(&hash_lock); 433 spin_unlock(&hash_lock); 581 * 3) Remove from the hash under hash_lock
|
/linux-master/include/linux/sunrpc/ |
H A D | cache.h | 74 spinlock_t hash_lock; member in struct:cache_detail
|
/linux-master/drivers/crypto/ |
H A D | s5p-sss.c | 285 * @hash_lock: Lock for protecting hash_req, hash_queue and hash_flags 321 spinlock_t hash_lock; /* protect hash_ vars */ member in struct:s5p_aes_dev 1347 spin_lock_irqsave(&dd->hash_lock, flags); 1351 spin_unlock_irqrestore(&dd->hash_lock, flags); 1376 spin_lock_irqsave(&dd->hash_lock, flags); 1381 spin_unlock_irqrestore(&dd->hash_lock, flags); 1390 spin_unlock_irqrestore(&dd->hash_lock, flags); 2229 spin_lock_init(&pdata->hash_lock);
|
/linux-master/include/net/ |
H A D | vxlan.h | 298 spinlock_t hash_lock[FDB_HASH_SIZE]; member in struct:vxlan_dev
|