/linux-master/fs/btrfs/ |
H A D | backref.c | 46 bool cached; local 65 cached = ctx->cache_lookup(eb->start, ctx->user_ctx, &root_ids, 67 if (!cached) 1281 * We cached a false result, but the last snapshot generation of the 1289 * If we cached a true result and the last generation used for dropping 1545 * cached result for the leaf is valid and only if there's only 1552 bool cached; local 1555 cached = lookup_backref_shared_cache(sc->ctx, sc->root, 1558 if (cached) { 1984 bool cached; local 2469 bool cached; local [all...] |
H A D | extent_io.c | 529 struct extent_state *cached = NULL; local 558 unlock_extent(tree, processed->start, processed->end, &cached); 2504 /* Fields for the cached extent (unsubmitted, not ready, extent). */ 2509 bool cached; member in struct:fiemap_cache 2538 * @len and @flags with cached one. 2539 * And only when we fails to merge, cached one will be submitted as 2554 if (!cache->cached) 2577 * to user space, we trim the length of the previously cached extent and 2593 * We cached a dealloc range (found in the io tree) for 2620 * is behind the cached offse [all...] |
H A D | extent-tree.c | 3588 * the block group isn't cached. 3794 if (ffe_ctl->cached) { 4422 ffe_ctl->cached = 0; 4456 * allocation bits, or if its not cached. 4459 * picked out then we don't care that the block group is cached. 4462 block_group->cached != BTRFS_CACHE_NO) { 4540 ffe_ctl->cached = btrfs_block_group_done(block_group); 4541 if (unlikely(!ffe_ctl->cached)) { 4561 if (unlikely(block_group->cached == BTRFS_CACHE_ERROR)) { 4618 if (!ffe_ctl->cached [all...] |
/linux-master/fs/bcachefs/ |
H A D | btree_types.h | 69 bool cached; member in struct:btree_bkey_cached_common 242 bool cached:1; member in struct:btree_path 340 return !b->cached 350 bool cached:1; member in struct:btree_insert_entry
|
H A D | btree_key_cache.c | 288 ck->c.cached = true;
|
H A D | btree_gc.c | 619 if (!p.ptr.cached) { 636 if (!p.ptr.cached) { 659 if (fsck_err_on(!p.ptr.cached && gen_cmp(p.ptr.gen, g->gen) < 0, 754 if ((p.ptr.cached && 756 (!p.ptr.cached && 1265 b.cached, "cached");
|
H A D | bcachefs_format.h | 691 x(cached, 5) \
|
H A D | backpointers.c | 30 if (p.ptr.cached) 667 if (p.ptr.cached)
|
H A D | ec.c | 1234 * It doesn't generally make sense to erasure code cached ptrs: 1237 if (!ptr_c || ptr_c->cached)
|
H A D | extents.c | 132 if (!ret && !p.ptr.cached) 135 if (p.ptr.cached && ptr_stale(ca, &p.ptr)) 605 ret += !p.ptr.cached && !crc_is_compressed(p.crc); 619 if (!p.ptr.cached && crc_is_compressed(p.crc)) 645 if (p.ptr.cached) 660 if (p->ptr.cached) 815 * If we deleted all the dirty pointers and there's still cached 816 * pointers, we could set the cached pointers to dirty if they're not 866 (!ptr->cached || 947 ptr->cached [all...] |
H A D | buckets.h | 153 return ptr->cached ? BCH_DATA_cached : BCH_DATA_user;
|
H A D | btree_trans_commit.c | 106 if (!i->cached) 313 BUG_ON(i->cached != path->cached); 586 * XXX: synchronization of cached update triggers with gc 589 BUG_ON(i->cached || i->level); 631 ret = !i->cached 743 if (!i->cached) { 832 if (i->cached)
|
H A D | data_update.c | 56 !ptr->cached) 165 !ptr->cached) { 186 !ptr_c->cached) { 209 if (!p.ptr.cached && 477 * KEY_TYPE_error key, or just a discard if it was a cached extent) 552 BUG_ON(p.ptr.cached); 559 } else if (!p.ptr.cached && 673 if ((opts->rewrite_ptrs & (1U << i)) && ptr->cached) {
|
H A D | btree_locking.c | 665 if (!path->cached && !trans->in_traverse_all) { 671 linked->cached == path->cached &&
|
H A D | replicas.c | 123 if (p.ptr.cached)
|
H A D | extents.h | 581 if (!ptr->cached) 593 if (ptr->cached)
|
H A D | alloc_foreground.c | 398 * uncached iter will return a cached key if one exists, but if not 400 * slot. To avoid racing bucket allocations, look up the cached key slot 422 /* now check the cached key to serialize concurrent allocs of the bucket */ 1461 bool cached) 1463 bch2_alloc_sectors_append_ptrs_inlined(c, wp, k, sectors, cached); 1459 bch2_alloc_sectors_append_ptrs(struct bch_fs *c, struct write_point *wp, struct bkey_i *k, unsigned sectors, bool cached) argument
|
H A D | buckets.c | 41 fs_usage->cached += sectors; 194 prt_printf(out, "cached:\t\t\t\t%llu\n", 195 fs_usage->u.b.cached); 600 if (b_gen != ptr->gen && !ptr->cached) { 739 dst->cached += src->cached; 798 u32 *dst_sectors = !ptr->cached 846 if (!p.ptr.cached) { 990 if (p.ptr.cached) { 995 bch2_fs_fatal_err_on(ret && gc, c, "%s: no replicas entry while updating cached sector [all...] |
H A D | btree_update.c | 21 cmp_int(l->cached, r->cached) ?: 403 .cached = path->cached, 430 i->cached = n.cached; 457 * a btree that's cached in the key cache, the btree iter code checks 461 if (path->cached && !i->old_btree_u64s) 528 * Ensure that updates to cached btrees go to the key cache: 532 !path->cached [all...] |
H A D | btree_iter.c | 50 cmp_int((int) l->cached, (int) r_cached) ?: 58 return __btree_path_cmp(l, r->btree_id, r->cached, r->pos, r->level); 157 if (path->cached) { 228 for (i = 0; i < (!path->cached ? BTREE_MAX_DEPTH : 1); i++) { 230 BUG_ON(!path->cached && 256 BUG_ON(!!(iter->flags & BTREE_ITER_CACHED) != btree_iter_path(trans, iter)->cached); 343 cmp_int(path->cached, key_cache); 369 key_cache ? " cached" : ""); 641 BUG_ON(path->cached); 657 if (!i->cached 1611 bool cached = flags & BTREE_ITER_CACHED; local [all...] |
/linux-master/fs/fuse/ |
H A D | fuse_i.h | 151 /* true if fully cached */ 152 bool cached; member in struct:fuse_inode::__anon39::__anon41 779 /** Do multi-page cached writes */
|
H A D | dir.c | 2090 * the filesystem, so forget cached acls for the inode. 2194 fi->rdc.cached = false;
|
/linux-master/net/openvswitch/ |
H A D | conntrack.c | 539 * connection was not confirmed, it is not cached and needs to be run 710 bool cached = skb_nfct_cached(net, key, info, skb); local 714 if (!cached) { 787 * - nf_conntrack_in() was executed above ("!cached") or a 792 if ((nf_ct_is_confirmed(ct) ? !cached || add_helper :
|
/linux-master/net/ipv4/ |
H A D | route.c | 682 /* Update all cached dsts too */ 714 /* Exception created; mark the cached routes for the nexthop 1574 bool cached = false; local 1601 cached = rt_bind_exception(rt, fnhe, daddr, do_cache); 1603 cached = rt_cache_route(nhc, rt); 1604 if (unlikely(!cached)) {
|
/linux-master/drivers/ata/ |
H A D | sata_mv.c | 528 struct mv_cached_regs cached; member in struct:mv_port_priv 906 * mv_save_cached_regs - (re-)initialize cached port registers 920 pp->cached.fiscfg = readl(port_mmio + FISCFG); 921 pp->cached.ltmode = readl(port_mmio + LTMODE); 922 pp->cached.haltcond = readl(port_mmio + EDMA_HALTCOND); 923 pp->cached.unknown_rsvd = readl(port_mmio + EDMA_UNKNOWN_RSVD); 927 * mv_write_cached_reg - write to a cached port register 929 * @old: pointer to cached value of the register 932 * Write a new value to a cached register, 1448 u32 fiscfg, *old_fiscfg = &pp->cached [all...] |