/linux-master/drivers/md/dm-vdo/ |
H A D | int-map.c | 16 * probing, all the entries that hash to a given bucket are stored within a fixed neighborhood 17 * starting at that bucket. Chaining is effectively represented as a bit vector relative to each 18 * bucket instead of as pointers or explicit offsets. 20 * When an empty bucket cannot be found within a given neighborhood, subsequent neighborhoods are 22 * an empty bucket will move into the desired neighborhood, allowing the entry to be added. When 36 * entries of the hash bucket at the start of the neighborhood, a pair of small offset fields each 41 * in the list is always the bucket closest to the start of the neighborhood. 61 #define MAX_PROBES 1024 /* limit on the number of probes for a free bucket */ 66 * struct bucket - hash bucket 73 struct __packed bucket { struct 276 struct bucket *bucket = dereference_hop(neighborhood, next_hop); local 324 search_hop_list(struct int_map *map __always_unused, struct bucket *bucket, u64 key, struct bucket **previous_ptr) argument 426 find_empty_bucket(struct int_map *map, struct bucket *bucket, unsigned int max_probes) argument 466 struct bucket *bucket; local 531 struct bucket *bucket = search_hop_list(map, neighborhood, key, NULL); local 613 struct bucket *neighborhood, *bucket; local 680 struct bucket *bucket = select_bucket(map, key); local [all...] |
H A D | priority-table.c | 20 * All the entries with the same priority are queued in a circular list in a bucket for that 23 struct bucket { struct 28 /* The priority of all the entries in this bucket */ 34 * of the queue in the appropriate bucket. The dequeue operation finds the highest-priority 35 * non-empty bucket by searching a bit vector represented as a single 8-byte word, which is very 44 struct bucket buckets[]; 64 struct bucket, __func__, &table); 69 struct bucket *bucket = &table->buckets[priority]; local 71 bucket 140 mark_bucket_empty(struct priority_table *table, struct bucket *bucket) argument 157 struct bucket *bucket; local [all...] |
/linux-master/net/mptcp/ |
H A D | token.c | 53 /* called with bucket lock held */ 66 /* called with bucket lock held */ 111 struct token_bucket *bucket; local 122 bucket = token_bucket(token); 123 spin_lock_bh(&bucket->lock); 124 if (__token_bucket_busy(bucket, token)) { 125 spin_unlock_bh(&bucket->lock); 129 hlist_nulls_add_head_rcu(&subflow_req->token_node, &bucket->req_chain); 130 bucket->chain_len++; 131 spin_unlock_bh(&bucket 157 struct token_bucket *bucket; local 196 struct token_bucket *bucket; local 213 struct token_bucket *bucket; local 249 struct token_bucket *bucket; local 305 struct token_bucket *bucket = &token_hash[slot]; local 354 struct token_bucket *bucket; local 378 struct token_bucket *bucket; local [all...] |
/linux-master/fs/bcachefs/ |
H A D | nocow_locking.h | 28 struct bpos bucket, int flags) 30 u64 dev_bucket = bucket_to_u64(bucket); 37 struct bpos bucket, int flags) 39 u64 dev_bucket = bucket_to_u64(bucket); 27 bch2_bucket_nocow_lock(struct bucket_nocow_lock_table *t, struct bpos bucket, int flags) argument 36 bch2_bucket_nocow_trylock(struct bucket_nocow_lock_table *t, struct bpos bucket, int flags) argument
|
H A D | move_types.h | 20 struct bpos bucket; member in struct:move_bucket_key 32 struct move_bucket bucket; member in struct:move_bucket_in_flight
|
H A D | backpointers.h | 36 * Convert from pos in backpointer btree to pos of corresponding bucket in alloc 49 * Convert from pos in alloc btree + bucket offset to pos in backpointer btree: 52 struct bpos bucket, 55 struct bch_dev *ca = bch_dev_bkey_exists(c, bucket.inode); 56 struct bpos ret = POS(bucket.inode, 57 (bucket_to_sector(ca, bucket.offset) << 60 EBUG_ON(!bkey_eq(bucket, bp_pos_to_bucket(c, ret))); 64 int bch2_bucket_backpointer_mod_nowritebuffer(struct btree_trans *, struct bpos bucket, 68 struct bpos bucket, 74 return bch2_bucket_backpointer_mod_nowritebuffer(trans, bucket, b 51 bucket_pos_to_bp(const struct bch_fs *c, struct bpos bucket, u64 bucket_offset) argument 67 bch2_bucket_backpointer_mod(struct btree_trans *trans, struct bpos bucket, struct bch_backpointer bp, struct bkey_s_c orig_k, bool insert) argument [all...] |
H A D | alloc_foreground.h | 115 unsigned dev, u64 bucket) 118 (jhash_3words(dev, bucket, bucket >> 32, 0) & 122 static inline bool bch2_bucket_is_open(struct bch_fs *c, unsigned dev, u64 bucket) argument 124 open_bucket_idx_t slot = *open_bucket_hashslot(c, dev, bucket); 129 if (ob->dev == dev && ob->bucket == bucket) 138 static inline bool bch2_bucket_is_open_safe(struct bch_fs *c, unsigned dev, u64 bucket) argument 142 if (bch2_bucket_is_open(c, dev, bucket)) 146 ret = bch2_bucket_is_open(c, dev, bucket); 114 open_bucket_hashslot(struct bch_fs *c, unsigned dev, u64 bucket) argument [all...] |
H A D | buckets_types.h | 10 struct bucket { struct 25 struct bucket b[];
|
H A D | backpointers.c | 19 struct bpos bucket, 34 if (bpos_eq(bucket, bucket2) && 53 struct bpos bucket = bp_pos_to_bucket(c, bp.k->p); local 57 !bpos_eq(bp.k->p, bucket_pos_to_bp(c, bucket, bp.v->bucket_offset)), 79 prt_str(out, "bucket="); 148 struct bpos bucket, 164 bp_k->k.p = bucket_pos_to_bp(trans->c, bucket, bp.bucket_offset); 200 struct bpos bucket, int gen, 206 struct bpos bp_end_pos = bucket_pos_to_bp(c, bpos_nosnap_successor(bucket), 0); 216 bucket, BTREE_ITER_CACHE 16 extent_matches_bp(struct bch_fs *c, enum btree_id btree_id, unsigned level, struct bkey_s_c k, struct bpos bucket, struct bch_backpointer bp) argument 147 bch2_bucket_backpointer_mod_nowritebuffer(struct btree_trans *trans, struct bpos bucket, struct bch_backpointer bp, struct bkey_s_c orig_k, bool insert) argument 199 bch2_get_next_backpointer(struct btree_trans *trans, struct bpos bucket, int gen, struct bpos *bp_pos, struct bch_backpointer *bp, unsigned iter_flags) argument 252 struct bpos bucket = bp_pos_to_bucket(c, bp_pos); local 291 struct bpos bucket = bp_pos_to_bucket(c, bp_pos); local 328 struct bpos bucket = bp_pos_to_bucket(c, bp_pos); local 502 check_bp_exists(struct btree_trans *trans, struct extents_to_bp_state *s, struct bpos bucket, struct bch_backpointer bp, struct bkey_s_c orig_k) argument [all...] |
/linux-master/drivers/infiniband/sw/rdmavt/ |
H A D | trace_qp.h | 18 TP_PROTO(struct rvt_qp *qp, u32 bucket), 19 TP_ARGS(qp, bucket), 23 __field(u32, bucket) 28 __entry->bucket = bucket; 31 "[%s] qpn 0x%x bucket %u", 34 __entry->bucket 39 TP_PROTO(struct rvt_qp *qp, u32 bucket), 40 TP_ARGS(qp, bucket)); 43 TP_PROTO(struct rvt_qp *qp, u32 bucket), [all...] |
/linux-master/drivers/interconnect/qcom/ |
H A D | bcm-voter.c | 65 int bucket, i; local 67 for (bucket = 0; bucket < QCOM_ICC_NUM_BUCKETS; bucket++) { 68 bcm->vote_x[bucket] = 0; 69 bcm->vote_y[bucket] = 0; 74 /* If any vote in this bucket exists, keep the BCM enabled */ 75 if (node->sum_avg[bucket] || node->max_peak[bucket]) { 76 bcm->vote_x[bucket] 94 size_t i, bucket; local 156 tcs_list_gen(struct bcm_voter *voter, int bucket, struct tcs_cmd tcs_list[MAX_VCD], int n[MAX_VCD + 1]) argument [all...] |
/linux-master/block/ |
H A D | blk-stat.c | 55 int bucket, cpu; local 69 bucket = cb->bucket_fn(rq); 70 if (bucket < 0) 73 stat = &per_cpu_ptr(cb->cpu_stat, cpu)[bucket]; 83 unsigned int bucket; local 86 for (bucket = 0; bucket < cb->buckets; bucket++) 87 blk_rq_stat_init(&cb->stat[bucket]); 93 for (bucket 139 unsigned int bucket; local [all...] |
/linux-master/net/9p/ |
H A D | error.c | 179 int bucket; local 182 for (bucket = 0; bucket < ERRHASHSZ; bucket++) 183 INIT_HLIST_HEAD(&hash_errmap[bucket]); 188 bucket = jhash(c->name, c->namelen, 0) % ERRHASHSZ; 190 hlist_add_head(&c->list, &hash_errmap[bucket]); 208 int bucket; local 212 bucket = jhash(errstr, len, 0) % ERRHASHSZ; 213 hlist_for_each_entry(c, &hash_errmap[bucket], lis [all...] |
/linux-master/net/vmw_vsock/ |
H A D | diag.c | 52 unsigned int bucket; local 63 bucket = cb->args[1]; 72 while (bucket < ARRAY_SIZE(vsock_bind_table)) { 73 struct list_head *head = &vsock_bind_table[bucket]; 94 bucket++; 98 bucket = 0; 102 while (bucket < ARRAY_SIZE(vsock_connected_table)) { 103 struct list_head *head = &vsock_connected_table[bucket]; 128 bucket++; 135 cb->args[1] = bucket; [all...] |
/linux-master/net/ceph/crush/ |
H A D | mapper.c | 58 * bucket choose methods 60 * For each bucket algorithm, we have a "choose" method that, given a 62 * will produce an item in the bucket. 66 * Choose based on a random permutation of the bucket. 70 * calculate an actual random permutation of the bucket members. 74 static int bucket_perm_choose(const struct crush_bucket *bucket, argument 78 unsigned int pr = r % bucket->size; 83 dprintk("bucket %d new x=%d\n", bucket->id, x); 88 s = crush_hash32_3(bucket 135 bucket_uniform_choose(const struct crush_bucket_uniform *bucket, struct crush_work_bucket *work, int x, int r) argument 142 bucket_list_choose(const struct crush_bucket_list *bucket, int x, int r) argument 196 bucket_tree_choose(const struct crush_bucket_tree *bucket, int x, int r) argument 228 bucket_straw_choose(const struct crush_bucket_straw *bucket, int x, int r) argument 305 get_choose_arg_weights(const struct crush_bucket_straw2 *bucket, const struct crush_choose_arg *arg, int position) argument 317 get_choose_arg_ids(const struct crush_bucket_straw2 *bucket, const struct crush_choose_arg *arg) argument 326 bucket_straw2_choose(const struct crush_bucket_straw2 *bucket, int x, int r, const struct crush_choose_arg *arg, int position) argument 449 crush_choose_firstn(const struct crush_map *map, struct crush_work *work, const struct crush_bucket *bucket, const __u32 *weight, int weight_max, int x, int numrep, int type, int *out, int outpos, int out_size, unsigned int tries, unsigned int recurse_tries, unsigned int local_retries, unsigned int local_fallback_retries, int recurse_to_leaf, unsigned int vary_r, unsigned int stable, int *out2, int parent_r, const struct crush_choose_arg *choose_args) argument 643 crush_choose_indep(const struct crush_map *map, struct crush_work *work, const struct crush_bucket *bucket, const __u32 *weight, int weight_max, int x, int left, int numrep, int type, int *out, int outpos, unsigned int tries, unsigned int recurse_tries, int recurse_to_leaf, int *out2, int parent_r, const struct crush_choose_arg *choose_args) argument [all...] |
/linux-master/net/sched/ |
H A D | sch_hhf.c | 21 * as heavy-hitter, it is immediately switched to the heavy-hitter bucket. 23 * in which the heavy-hitter bucket is served with less weight. 61 * dispatched to the heavy-hitter bucket accordingly. 68 * bucket. 71 * to the non-heavy-hitter bucket. 74 * send p to the heavy-hitter bucket. 105 WDRR_BUCKET_FOR_HH = 0, /* bucket id for heavy-hitters */ 106 WDRR_BUCKET_FOR_NON_HH = 1 /* bucket id for non-heavy-hitters */ 328 /* Removes one skb from head of bucket. */ 329 static struct sk_buff *dequeue_head(struct wdrr_bucket *bucket) argument 339 bucket_add(struct wdrr_bucket *bucket, struct sk_buff *skb) argument 352 struct wdrr_bucket *bucket; local 376 struct wdrr_bucket *bucket; local 422 struct wdrr_bucket *bucket; local 642 struct wdrr_bucket *bucket = q->buckets + i; local [all...] |
/linux-master/tools/testing/selftests/net/forwarding/ |
H A D | ethtool_rmon.sh | 34 local bucket=$1; shift 48 jq -r ".[0].rmon[\"${set}-pktsNtoM\"][$bucket].val") 56 jq -r ".[0].rmon[\"${set}-pktsNtoM\"][$bucket].val") 76 while read -r -a bucket; do 77 step="$set-pkts${bucket[0]}to${bucket[1]} on $iface" 80 if ! ensure_mtu $if ${bucket[0]}; then 86 if ! bucket_test $iface $neigh $set $nbuckets ${bucket[0]}; then
|
/linux-master/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
H A D | pno.h | 56 * brcmf_pno_find_reqid_by_bucket - find request id for given bucket index. 59 * @bucket: index of firmware bucket. 61 u64 brcmf_pno_find_reqid_by_bucket(struct brcmf_pno_info *pi, u32 bucket); 64 * brcmf_pno_get_bucket_map - determine bucket map for given netinfo. 67 * @netinfo: netinfo to compare with bucket configuration.
|
/linux-master/include/net/ |
H A D | transp_v6.h | 48 __u16 srcp, __u16 destp, int rqueue, int bucket); 51 __u16 destp, int bucket) 54 bucket); 50 ip6_dgram_sock_seq_show(struct seq_file *seq, struct sock *sp, __u16 srcp, __u16 destp, int bucket) argument
|
/linux-master/drivers/cpuidle/governors/ |
H A D | menu.c | 116 unsigned int bucket; member in struct:menu_device 124 int bucket = 0; local 133 bucket = BUCKETS/2; 136 return bucket; 138 return bucket + 1; 140 return bucket + 2; 142 return bucket + 3; 144 return bucket + 4; 145 return bucket + 5; 293 data->bucket [all...] |
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | sock_iter_batch.c | 21 unsigned int bucket[2]; variable 48 /* bucket selection as in inet_lhash2_bucket_sk() */ 53 bucket[idx] = hash & hinfo->lhash2_mask; 83 /* bucket selection as in udp_hashslot2() */ 85 bucket[idx] = udp_sk(sk)->udp_portaddr_hash & udptable->mask;
|
/linux-master/fs/nfs/ |
H A D | nfs42xattr.c | 87 struct nfs4_xattr_bucket *bucket; member in struct:nfs4_xattr_entry 120 * 1. inode i_lock or bucket lock 238 entry->bucket = NULL; 388 struct nfs4_xattr_bucket *bucket; local 394 bucket = &cache->buckets[i]; 396 spin_lock(&bucket->lock); 397 bucket->draining = true; 398 hlist_for_each_entry_safe(entry, n, &bucket->hlist, hnode) { 403 spin_unlock(&bucket->lock); 511 nfs4_xattr_get_entry(struct nfs4_xattr_bucket *bucket, cons argument 529 struct nfs4_xattr_bucket *bucket; local 566 struct nfs4_xattr_bucket *bucket; local 589 struct nfs4_xattr_bucket *bucket; local 873 struct nfs4_xattr_bucket *bucket; local [all...] |
H A D | pnfs_nfs.c | 63 pnfs_free_bucket_lseg(struct pnfs_commit_bucket *bucket) argument 65 if (list_empty(&bucket->committing) && list_empty(&bucket->written)) { 66 struct pnfs_layout_segment *freeme = bucket->lseg; 67 bucket->lseg = NULL; 74 * If this will make the bucket empty, it will need to put the lseg reference. 81 struct pnfs_commit_bucket *bucket = NULL; local 87 bucket = list_first_entry(&req->wb_list, 91 if (bucket) 92 pnfs_put_lseg(pnfs_free_bucket_lseg(bucket)); 241 pnfs_bucket_scan_ds_commit_list(struct pnfs_commit_bucket *bucket, struct nfs_commit_info *cinfo, int max) argument 402 pnfs_bucket_get_committing(struct list_head *head, struct pnfs_commit_bucket *bucket, struct nfs_commit_info *cinfo) argument 419 pnfs_bucket_fetch_commitdata(struct pnfs_commit_bucket *bucket, struct nfs_commit_info *cinfo) argument 435 struct pnfs_commit_bucket *bucket; local 456 struct pnfs_commit_bucket *bucket; local 1207 struct pnfs_commit_bucket *bucket; local [all...] |
/linux-master/fs/dlm/ |
H A D | debug_fs.c | 417 unsigned bucket; member in struct:rsbtbl_iter 484 unsigned bucket, entry; local 487 bucket = n >> 32; 490 if (bucket >= ls->ls_rsbtbl_size) 509 tree = toss ? &ls->ls_rsbtbl[bucket].toss : &ls->ls_rsbtbl[bucket].keep; 511 spin_lock(&ls->ls_rsbtbl[bucket].lock); 518 ri->bucket = bucket; 519 spin_unlock(&ls->ls_rsbtbl[bucket] 566 unsigned bucket; local [all...] |
/linux-master/kernel/dma/ |
H A D | debug.c | 248 * Request exclusive access to a hash bucket for a given dma_debug_entry. 263 * Give up exclusive access to the hash bucket 265 static void put_hash_bucket(struct hash_bucket *bucket, 267 __releases(&bucket->lock) 269 spin_unlock_irqrestore(&bucket->lock, flags); 292 * Search a given entry in the hash bucket list 294 static struct dma_debug_entry *__hash_bucket_find(struct hash_bucket *bucket, argument 301 list_for_each_entry(entry, &bucket->list, list) { 344 static struct dma_debug_entry *bucket_find_exact(struct hash_bucket *bucket, argument 347 return __hash_bucket_find(bucket, re 350 bucket_find_contain(struct hash_bucket **bucket, struct dma_debug_entry *ref, unsigned long *flags) argument 378 hash_bucket_add(struct hash_bucket *bucket, struct dma_debug_entry *entry) argument 529 struct hash_bucket *bucket = &dma_entry_hash[idx]; local 561 struct hash_bucket *bucket = &dma_entry_hash[idx]; local 590 struct hash_bucket *bucket; local 964 struct hash_bucket *bucket; local 1097 struct hash_bucket *bucket; local 1249 struct hash_bucket *bucket; local 1341 struct hash_bucket *bucket; local [all...] |