/linux-master/fs/bcachefs/ |
H A D | buckets.c | 14 #include "buckets.h" 69 usage->b.hidden += (dev.d[BCH_DATA_sb].buckets + 70 dev.d[BCH_DATA_journal].buckets) * 272 ca->usage_base->d[BCH_DATA_free].buckets = ca->mi.nbuckets - ca->mi.first_bucket; 278 prt_str(out, "buckets"); 289 prt_u64(out, usage->d[i].buckets); 317 u->d[old->data_type].buckets--; 318 u->d[new->data_type].buckets++; 1236 ca->journal.buckets[i], 1336 struct bucket_gens *buckets local [all...] |
H A D | super.c | 1971 bch_err_msg(ca, ret, "resizing buckets"); 1995 ca->usage_base->d[BCH_DATA_free].buckets += nbuckets - old_nbuckets;
|
H A D | journal_io.c | 8 #include "buckets.h" 711 prt_printf(out, ": buckets=%llu sectors=%llu fragmented=%llu", 712 le64_to_cpu(u->d[i].buckets), 990 u64 offset = bucket_to_sector(ca, ja->buckets[bucket]), 1094 bucket_to_sector(ca, ja->buckets[bucket]), 1141 pr_debug("%u journal buckets", ja->nr); 1475 ja->buckets[ja->cur_idx]) + 1657 * more buckets:
|
H A D | recovery.c | 11 #include "buckets.h" 368 ca->usage_base->d[i].buckets = le64_to_cpu(u->d[i].buckets); 970 * Write out the superblock and journal buckets, now that we can do
|
H A D | sb-clean.c | 5 #include "buckets.h" 250 u->d[i].buckets = cpu_to_le64(ca->usage_base->d[i].buckets);
|
H A D | chardev.c | 6 #include "buckets.h" 613 arg.d[i].buckets = src.d[i].buckets; 660 .buckets = src.d[i].buckets,
|
H A D | btree_gc.c | 20 #include "buckets.h" 1155 b = ca->journal.buckets[i]; 1242 d[i].buckets, "%s buckets", bch2_data_type_str(i)); 1329 this_cpu_write(ca->usage_gc->d[BCH_DATA_free].buckets, 1493 struct bucket_array *buckets = kvmalloc(sizeof(struct bucket_array) + local 1496 if (!buckets) { 1498 bch_err(c, "error allocating ca->buckets[gc]"); 1502 buckets->first_bucket = ca->mi.first_bucket; 1503 buckets 1539 struct bucket_array *buckets = gc_bucket_array(ca); local [all...] |
H A D | bcachefs_format.h | 535 __le64 buckets[]; member in struct:bch_sb_field_journal 1392 __le64 buckets; member in struct:jset_entry_dev_usage_type
|
H A D | ec.c | 14 #include "buckets.h" 1356 * data buckets of new stripe all written: create the stripe 1382 bch_err(c, "error creating stripe: error writing data buckets"); 1415 bch_err(c, "error creating stripe: error writing redundancy buckets"); 1743 struct open_buckets buckets; local 1763 buckets.nr = 0; 1765 ret = bch2_bucket_alloc_set_trans(trans, &buckets, 1775 open_bucket_for_each(c, &buckets, ob, i) { 1781 h->s->blocks[j] = buckets.v[i]; 1790 buckets [all...] |
H A D | buckets.h | 95 struct bucket_array *buckets = gc_bucket_array(ca); local 97 BUG_ON(b < buckets->first_bucket || b >= buckets->nbuckets); 98 return buckets->b + b; 241 usage.d[BCH_DATA_free].buckets - 251 usage.d[BCH_DATA_free].buckets 252 + usage.d[BCH_DATA_cached].buckets 253 + usage.d[BCH_DATA_need_gc_gens].buckets 254 + usage.d[BCH_DATA_need_discard].buckets
|
H A D | journal_types.h | 302 * buckets, in bch_sb. 322 u64 *buckets; member in struct:journal_device
|
H A D | journal_reclaim.c | 7 #include "buckets.h" 82 unsigned sectors, buckets, unwritten; local 91 buckets = bch2_journal_dev_buckets_available(j, ja, from); 111 if (!buckets) { 116 buckets--; 123 if (sectors < ca->mi.bucket_size && buckets) { 124 buckets--; 130 .total = sectors + buckets * ca->mi.bucket_size, 255 * Advance ja->discard_idx as long as it points to buckets that are no longer 273 ja->buckets[j [all...] |
H A D | alloc_background.c | 13 #include "buckets.h" 1178 " device %llu buckets %llu-%llu", 1330 * valid for buckets that exist; this just checks for keys for nonexistent 1331 * buckets. 1367 "bucket_gens key for invalid buckets:\n %s", 1656 bch2_dev_usage_read(s->ca).d[BCH_DATA_free].buckets) 2056 * Scan the alloc btree for every bucket on @ca, and add buckets to the 2225 * We need to reserve buckets (from the number 2226 * of currently available buckets) against 2234 * available buckets) [all...] |
H A D | alloc_foreground.c | 5 * Foreground allocator code: allocate buckets from freelist, and allocate in 10 * bch2_bucket_alloc_set() allocates one or more buckets from different devices 21 #include "buckets.h" 55 * Open buckets represent a bucket that's currently being allocated from. They 58 * - They track buckets that have been partially allocated, allowing for 61 * - They provide a reference to the buckets they own that mark and sweep GC 379 * journal buckets - journal buckets will be < ca->new_fs_bucket_idx 519 * @cl: if not NULL, closure to be used to wait if buckets not available 540 if (usage->d[BCH_DATA_need_discard].buckets > avai [all...] |
H A D | Makefile | 25 buckets.o \
|
H A D | movinggc.c | 14 #include "buckets.h" 144 move_buckets *buckets) 179 ret2 = darray_push(buckets, b); 185 ret2 = buckets->nr >= nr_to_get; 190 pr_debug("have: %zu (%zu) saw %zu in flight %zu not movable %zu got %zu (%zu)/%zu buckets ret %i", 192 saw, in_flight, not_movable, buckets->nr, sectors, nr_to_get, ret); 207 move_buckets buckets = { 0 }; local 212 ret = bch2_copygc_get_buckets(ctxt, buckets_in_flight, &buckets); 216 darray_for_each(buckets, i) { 239 darray_exit(&buckets); 142 bch2_copygc_get_buckets(struct moving_context *ctxt, struct buckets_in_flight *buckets_in_flight, move_buckets *buckets) argument 312 struct buckets_in_flight *buckets; local [all...] |
H A D | journal.c | 14 #include "buckets.h" 143 * btree transactions, to discard journal buckets, nor acquire priority 954 bch_err_msg(c, ret, "marking new journal buckets"); 965 /* Don't return an error if we successfully allocated some buckets: */ 974 memcpy(new_buckets, ja->buckets, ja->nr * sizeof(u64)); 1006 swap(new_buckets, ja->buckets); 1059 /* don't handle reducing nr of buckets yet: */ 1067 * note: journal buckets aren't really counted as _sectors_ used yet, so 1068 * we don't need the disk reservation to avoid the BUG_ON() in buckets.c 1112 * clamp journal size to 8192 buckets o [all...] |
/linux-master/net/unix/ |
H A D | af_unix.c | 322 sk_add_node(sk, &net->unx.table.buckets[sk->sk_hash]); 373 sk_for_each(s, &net->unx.table.buckets[hash]) { 3268 for (sk = sk_head(&seq_file_net(seq)->unx.table.buckets[bucket]); 3618 net->unx.table.buckets = kvmalloc_array(UNIX_HASH_SIZE, 3621 if (!net->unx.table.buckets) 3626 INIT_HLIST_HEAD(&net->unx.table.buckets[i]); 3645 kvfree(net->unx.table.buckets);
|
/linux-master/net/netfilter/ |
H A D | nft_set_hash.c | 457 /* Number of buckets is stored in u32, so cap our result to 1U<<31 */ 482 u32 buckets; member in struct:nft_hash 502 hash = reciprocal_scale(hash, priv->buckets); 523 hash = reciprocal_scale(hash, priv->buckets); 544 hash = reciprocal_scale(hash, priv->buckets); 568 hash = reciprocal_scale(hash, priv->buckets); 649 for (i = 0; i < priv->buckets; i++) { 676 priv->buckets = nft_hash_buckets(desc->size); 690 for (i = 0; i < priv->buckets; i++) {
|
/linux-master/net/core/ |
H A D | dev.c | 530 * MUST BE last in hash buckets and checking protocol handlers 4727 if (likely(fl->buckets[old_flow])) 4728 fl->buckets[old_flow]--; 4730 if (++fl->buckets[new_flow] > (FLOW_LIMIT_HISTORY >> 1)) {
|
H A D | sock_map.c | 850 struct bpf_shtab_bucket *buckets; member in struct:bpf_shtab 865 return &htab->buckets[hash & (htab->buckets_num - 1)]; 1118 htab->buckets = bpf_map_area_alloc(htab->buckets_num * 1121 if (!htab->buckets) { 1127 INIT_HLIST_HEAD(&htab->buckets[i].head); 1128 spin_lock_init(&htab->buckets[i].lock); 1185 bpf_map_area_free(htab->buckets); 1324 bucket = &htab->buckets[info->bucket_id];
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/lag/ |
H A D | lag.c | 207 for (j = 0; j < ldev->buckets; j++) { 208 idx = i * ldev->buckets + j; 279 ldev->buckets = 1; 307 * As we have ldev->buckets slots per port first assume the native 314 u8 buckets, 334 /* Use native mapping by default where each port's buckets 338 for (j = 0; j < buckets; j++) { 339 idx = i * buckets + j; 350 for (j = 0; j < buckets; j++) { 352 ports[disabled[i] * buckets 312 mlx5_infer_tx_affinity_mapping(struct lag_tracker *tracker, u8 num_ports, u8 buckets, u8 *ports) argument [all...] |
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | arena_htab.c | 20 printf("htab %p buckets %p n_buckets %d\n", htab, htab->buckets, htab->n_buckets); 21 ASSERT_OK_PTR(htab->buckets, "htab->buckets shouldn't be NULL"); 22 for (i = 0; htab->buckets && i < 16; i += 4) { 24 * Walk htab buckets and link lists since all pointers are correct,
|
/linux-master/kernel/trace/ |
H A D | trace.h | 839 struct hlist_head *buckets; member in struct:ftrace_hash
|
H A D | ftrace.c | 1060 .buckets = (struct hlist_head *)empty_buckets, 1146 hhd = &hash->buckets[key]; 1181 hhd = &hash->buckets[key]; 1230 hhd = &hash->buckets[i]; 1264 kfree(hash->buckets); 1305 hash->buckets = kcalloc(size, sizeof(*hash->buckets), GFP_KERNEL); 1307 if (!hash->buckets) { 1368 hlist_for_each_entry(entry, &hash->buckets[i], hlist) { 1418 hhd = &src->buckets[ [all...] |