/linux-master/include/linux/ |
H A D | hashtable.h | 122 * @bkt: integer to use as bucket loop cursor 126 #define hash_for_each(name, bkt, obj, member) \ 127 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ 128 (bkt)++)\ 129 hlist_for_each_entry(obj, &name[bkt], member) 134 * @bkt: integer to use as bucket loop cursor 138 #define hash_for_each_rcu(name, bkt, obj, member) \ 139 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZ [all...] |
H A D | rhashtable.h | 327 struct rhash_lock_head __rcu **bkt) 332 bit_spin_lock(0, (unsigned long *)bkt); 350 struct rhash_lock_head __rcu **bkt, 354 bit_spin_unlock(0, (unsigned long *)bkt); 359 struct rhash_lock_head *p, struct rhash_lock_head __rcu *const *bkt) 363 (unsigned long)RHT_NULLS_MARKER(bkt)); 367 * Where 'bkt' is a bucket and might be locked: 374 struct rhash_lock_head __rcu *const *bkt) 376 return __rht_ptr(rcu_dereference(*bkt), bkt); 326 rht_lock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt) argument 349 rht_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt, unsigned long flags) argument 358 __rht_ptr( struct rhash_lock_head *p, struct rhash_lock_head __rcu *const *bkt) argument 373 rht_ptr_rcu( struct rhash_lock_head __rcu *const *bkt) argument 379 rht_ptr( struct rhash_lock_head __rcu *const *bkt, struct bucket_table *tbl, unsigned int hash) argument 387 rht_ptr_exclusive( struct rhash_lock_head __rcu *const *bkt) argument 393 rht_assign_locked(struct rhash_lock_head __rcu **bkt, struct rhash_head *obj) argument 401 rht_assign_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt, struct rhash_head *obj, unsigned long flags) argument 597 struct rhash_lock_head __rcu *const *bkt; local 713 struct rhash_lock_head __rcu **bkt; local 1000 struct rhash_lock_head __rcu **bkt; local 1153 struct rhash_lock_head __rcu **bkt; local [all...] |
/linux-master/tools/include/linux/ |
H A D | hashtable.h | 101 * @bkt: integer to use as bucket loop cursor 105 #define hash_for_each(name, bkt, obj, member) \ 106 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ 107 (bkt)++)\ 108 hlist_for_each_entry(obj, &name[bkt], member) 114 * @bkt: integer to use as bucket loop cursor 119 #define hash_for_each_safe(name, bkt, tmp, obj, member) \ 120 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZ [all...] |
/linux-master/tools/lib/bpf/ |
H A D | hashmap.h | 166 * @bkt: integer used as a bucket loop cursor 168 #define hashmap__for_each_entry(map, cur, bkt) \ 169 for (bkt = 0; bkt < map->cap; bkt++) \ 170 for (cur = map->buckets[bkt]; cur; cur = cur->next) 178 * @bkt: integer used as a bucket loop cursor 180 #define hashmap__for_each_entry_safe(map, cur, tmp, bkt) \ 181 for (bkt = 0; bkt < ma [all...] |
H A D | hashmap.c | 66 size_t bkt; local 68 hashmap__for_each_entry_safe(map, cur, tmp, bkt) { 106 size_t h, bkt; local 117 hashmap__for_each_entry_safe(map, cur, tmp, bkt) {
|
/linux-master/tools/perf/util/ |
H A D | hashmap.h | 166 * @bkt: integer used as a bucket loop cursor 168 #define hashmap__for_each_entry(map, cur, bkt) \ 169 for (bkt = 0; bkt < map->cap; bkt++) \ 170 for (cur = map->buckets[bkt]; cur; cur = cur->next) 178 * @bkt: integer used as a bucket loop cursor 180 #define hashmap__for_each_entry_safe(map, cur, tmp, bkt) \ 181 for (bkt = 0; bkt < ma [all...] |
H A D | expr.c | 86 size_t bkt; local 91 hashmap__for_each_entry(ids, cur, bkt) { 115 size_t bkt; local 133 hashmap__for_each_entry(ids2, cur, bkt) { 232 size_t bkt; local 235 hashmap__for_each_entry(needles->ids, cur, bkt) { 314 size_t bkt; local 316 hashmap__for_each_entry(ctx->ids, cur, bkt) { 326 size_t bkt; local 332 hashmap__for_each_entry(ctx->ids, cur, bkt) { [all...] |
H A D | hashmap.c | 66 size_t bkt; local 68 hashmap__for_each_entry_safe(map, cur, tmp, bkt) { 106 size_t h, bkt; local 117 hashmap__for_each_entry_safe(map, cur, tmp, bkt) {
|
H A D | threads.c | 140 size_t bkt; local 144 hashmap__for_each_entry_safe((&table->shard), cur, tmp, bkt) { 175 size_t bkt; local 178 hashmap__for_each_entry((&table->shard), cur, bkt) {
|
/linux-master/drivers/infiniband/ulp/opa_vnic/ |
H A D | opa_vnic_internal.h | 280 #define vnic_hash_for_each_safe(name, bkt, tmp, obj, member) \ 281 for ((bkt) = 0, obj = NULL; \ 282 !obj && (bkt) < OPA_VNIC_MAC_TBL_SIZE; (bkt)++) \ 283 hlist_for_each_entry_safe(obj, tmp, &name[bkt], member) 289 #define vnic_hash_for_each(name, bkt, obj, member) \ 290 for ((bkt) = 0, obj = NULL; \ 291 !obj && (bkt) < OPA_VNIC_MAC_TBL_SIZE; (bkt)++) \ 292 hlist_for_each_entry(obj, &name[bkt], membe [all...] |
H A D | opa_vnic_encap.c | 107 int bkt; local 112 vnic_hash_for_each_safe(mactbl, bkt, tmp, node, hlist) { 157 int bkt; local 168 vnic_hash_for_each(mactbl, bkt, node, hlist) { 209 int i, bkt, rc = 0; local 263 vnic_hash_for_each(old_mactbl, bkt, node, hlist) {
|
/linux-master/lib/ |
H A D | hashtable_test.c | 68 int bkt; local 80 hash_for_each(hash, bkt, x, node) { 127 int bkt, i, j, count; local 139 hash_for_each(hash, bkt, x, node) { 157 int bkt, i, j, count; local 169 hash_for_each_safe(hash, bkt, tmp, x, node) { 190 int bkt, i, j, count; local 220 hash_for_each(hash, bkt, y, node) { 223 buckets[y->key] = bkt; 245 int bkt, local [all...] |
H A D | rhashtable.c | 225 struct rhash_lock_head __rcu **bkt, 241 rht_for_each_from(entry, rht_ptr(bkt, old_tbl, old_hash), 270 rht_assign_locked(bkt, next); 280 struct rhash_lock_head __rcu **bkt = rht_bucket_var(old_tbl, old_hash); local 284 if (!bkt) 286 flags = rht_lock(old_tbl, bkt); 288 while (!(err = rhashtable_rehash_one(ht, bkt, old_hash))) 293 rht_unlock(old_tbl, bkt, flags); 491 struct rhash_lock_head __rcu **bkt, 504 rht_for_each_from(head, rht_ptr(bkt, tb 224 rhashtable_rehash_one(struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, unsigned int old_hash) argument 490 rhashtable_lookup_one(struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, struct bucket_table *tbl, unsigned int hash, const void *key, struct rhash_head *obj) argument 541 rhashtable_insert_one( struct rhashtable *ht, struct rhash_lock_head __rcu **bkt, struct bucket_table *tbl, unsigned int hash, struct rhash_head *obj, void *data) argument 595 struct rhash_lock_head __rcu **bkt; local [all...] |
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | hashmap.c | 47 int err, bkt, found_cnt, i; local 92 hashmap__for_each_entry(map, entry, bkt) { 140 hashmap__for_each_entry_safe(map, entry, tmp, bkt) { 197 hashmap__for_each_entry_safe(map, entry, tmp, bkt) { 230 hashmap__for_each_entry(map, entry, bkt) { 238 hashmap__for_each_entry(map, entry, bkt) { 265 int err, i, bkt; local 319 hashmap__for_each_entry(map, cur, bkt) { 345 int err, bkt; local 382 hashmap__for_each_entry(map, entry, bkt) { 414 int bkt; local [all...] |
/linux-master/drivers/s390/crypto/ |
H A D | ap_card.c | 80 int bkt; local 85 hash_for_each(ap_queues, bkt, aq, hnode) 99 int bkt; local 106 hash_for_each(ap_queues, bkt, aq, hnode) 118 int bkt; local 125 hash_for_each(ap_queues, bkt, aq, hnode)
|
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_vf_lib.h | 179 * @bkt: bucket index used for iteration 182 * The bkt variable is an unsigned integer iterator used to traverse the VF 189 #define ice_for_each_vf(pf, bkt, vf) \ 190 hash_for_each((pf)->vfs.table, (bkt), (vf), entry) 195 * @bkt: bucket index used for iteration 198 * The bkt variable is an unsigned integer iterator used to traverse the VF 206 #define ice_for_each_vf_rcu(pf, bkt, vf) \ 207 hash_for_each_rcu((pf)->vfs.table, (bkt), (vf), entry)
|
H A D | ice_sriov.c | 29 unsigned int bkt; local 37 hash_for_each_safe(vfs->table, bkt, tmp, vf, entry) { 156 unsigned int bkt; local 177 ice_for_each_vf(pf, bkt, vf) { 585 unsigned int bkt, it_cnt; local 592 ice_for_each_vf(pf, bkt, vf) { 620 ice_for_each_vf(pf, bkt, vf) { 1017 int to_remap = 0, bkt; local 1022 ice_for_each_vf(pf, bkt, tmp_vf) { 1213 unsigned int bkt; local 1250 unsigned int bkt; local 1586 unsigned int bkt; local 1882 unsigned int bkt; local 1926 u32 bkt; local [all...] |
H A D | ice_vf_lib.c | 108 unsigned int bkt; local 112 ice_for_each_vf_rcu(pf, bkt, vf) 574 unsigned int bkt; local 577 ice_for_each_vf_rcu(pf, bkt, vf) { 730 unsigned int bkt; local 739 ice_for_each_vf(pf, bkt, vf) 749 ice_for_each_vf(pf, bkt, vf) 756 ice_for_each_vf(pf, bkt, vf) { 768 ice_for_each_vf(pf, bkt, vf) { 1355 unsigned int bkt; local [all...] |
/linux-master/net/ipv6/ |
H A D | calipso.c | 203 u32 bkt; local 212 bkt = hash & (CALIPSO_CACHE_BUCKETS - 1); 213 spin_lock_bh(&calipso_cache[bkt].lock); 214 list_for_each_entry(entry, &calipso_cache[bkt].list, list) { 224 spin_unlock_bh(&calipso_cache[bkt].lock); 239 spin_unlock_bh(&calipso_cache[bkt].lock); 244 spin_unlock_bh(&calipso_cache[bkt].lock); 268 u32 bkt; local 291 bkt = entry->hash & (CALIPSO_CACHE_BUCKETS - 1); 292 spin_lock_bh(&calipso_cache[bkt] [all...] |
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
H A D | vxlan.c | 183 int bkt; local 188 hash_for_each_safe(vxlan->htable, bkt, tmp, vxlanp, hlist) {
|
/linux-master/arch/x86/kvm/ |
H A D | debugfs.c | 115 int bkt; local 118 kvm_for_each_memslot(slot, bkt, slots)
|
/linux-master/net/ipv4/ |
H A D | cipso_ipv4.c | 237 u32 bkt; local 246 bkt = hash & (CIPSO_V4_CACHE_BUCKETS - 1); 247 spin_lock_bh(&cipso_v4_cache[bkt].lock); 248 list_for_each_entry(entry, &cipso_v4_cache[bkt].list, list) { 258 spin_unlock_bh(&cipso_v4_cache[bkt].lock); 273 spin_unlock_bh(&cipso_v4_cache[bkt].lock); 278 spin_unlock_bh(&cipso_v4_cache[bkt].lock); 301 u32 bkt; local 324 bkt = entry->hash & (CIPSO_V4_CACHE_BUCKETS - 1); 325 spin_lock_bh(&cipso_v4_cache[bkt] [all...] |
/linux-master/net/sched/ |
H A D | cls_route.c | 55 struct route4_bucket *bkt; member in struct:route4_filter 328 b = f->bkt; 465 f->bkt = b; 524 f->bkt = fold->bkt; 534 fp = &f->bkt->ht[h];
|
/linux-master/tools/bpf/bpftool/ |
H A D | pids.c | 180 size_t bkt; local 185 hashmap__for_each_entry(map, entry, bkt) {
|
/linux-master/arch/x86/kvm/mmu/ |
H A D | page_track.c | 173 int r = 0, i, bkt; local 187 kvm_for_each_memslot(slot, bkt, slots) {
|