/linux-master/fs/btrfs/ |
H A D | extent-tree.h | 112 u64 offset, int metadata, u64 *refs, u64 *flags,
|
H A D | extent-io-tree.c | 45 pr_err("BTRFS: state leak: start %llu end %llu state %u in tree %d refs %d\n", 48 refcount_read(&state->refs)); 176 refcount_set(&state->refs, 1); 194 if (refcount_dec_and_test(&state->refs)) { 676 refcount_dec(&cached->refs); 819 refcount_inc(&state->refs); 855 refcount_inc(&state->refs); 1023 refcount_inc(&state->refs); 1716 refcount_inc(&state->refs);
|
H A D | extent_map.c | 51 refcount_set(&em->refs, 1); 64 if (refcount_dec_and_test(&em->refs)) { 239 if (refcount_read(&em->refs) > 2) 363 refcount_inc(&em->refs); 426 refcount_inc(&em->refs); 732 refcount_inc(&next_em->refs);
|
H A D | volumes.h | 466 refcount_t refs; member in struct:btrfs_io_context 576 refcount_t refs; member in struct:btrfs_chunk_map 593 if (map && refcount_dec_and_test(&map->refs)) {
|
H A D | delayed-ref.c | 60 * @nr_refs: number of delayed refs to drop 63 * Drops the delayed ref head's count from the delayed refs rsv and free any 82 * Adjust the size of the delayed refs rsv. 106 * reserve doesn't have enough reserved space, the delayed refs reserve 129 * Adjust the size of the delayed refs block reserve for 1 block group item 148 * Adjust the size of the delayed refs block reserve to release space for 1 164 * Adjust the size of the delayed refs block reserve for 1 block group item 183 * Adjust the size of the delayed refs block reserve to release space for 1 199 * Transfer bytes to our delayed refs rsv. 241 * Refill based on our delayed refs usag [all...] |
H A D | extent-io-tree.h | 122 refcount_t refs; member in struct:extent_state
|
H A D | block-group.h | 179 refcount_t refs; member in struct:btrfs_block_group
|
/linux-master/drivers/block/ |
H A D | nbd.c | 120 refcount_t refs; member in struct:nbd_device 274 if (!refcount_dec_and_test(&nbd->refs)) 1618 if (!refcount_inc_not_zero(&nbd->refs)) { 1636 refcount_inc(&nbd->refs); 1804 static struct nbd_device *nbd_dev_add(int index, unsigned int refs) argument 1879 refcount_set(&nbd->refs, 0); 1894 refcount_set(&nbd->refs, refs); 1925 if (refcount_inc_not_zero(&nbd->refs)) 2017 !refcount_inc_not_zero(&nbd->refs)) { [all...] |
/linux-master/io_uring/ |
H A D | kbuf.h | 28 atomic_t refs; member in struct:io_buffer_list
|
H A D | io_uring.c | 85 #include "refs.h" 245 struct io_ring_ctx *ctx = container_of(ref, struct io_ring_ctx, refs); 258 percpu_ref_get(&ctx->refs); 266 percpu_ref_put(&ctx->refs); 305 if (percpu_ref_init(&ctx->refs, io_ring_ctx_ref_free, 427 /* linked timeouts should have two refs once prep'ed */ 554 if (atomic_dec_and_test(&ev_fd->refs)) { 586 atomic_inc(&ev_fd->refs); 590 atomic_dec(&ev_fd->refs); 771 unsigned int refs local [all...] |
H A D | register.c | 67 atomic_set(&ev_fd->refs, 1); 331 refcount_inc(&sqd->refs); 398 * We don't quiesce the refs for register anymore and so it can't be 401 if (WARN_ON_ONCE(percpu_ref_is_dying(&ctx->refs)))
|
H A D | kbuf.c | 64 atomic_set(&bl->refs, 1); 271 if (atomic_dec_and_test(&bl->refs)) { 744 ret = atomic_inc_not_zero(&bl->refs);
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmm.c | 150 if (--pgd->refs[0]) { 155 if (!pgt->refs[!type]) { 191 if (!pgt->refs[!type]) 216 if (!pgt->refs[0]) 277 pgt->refs[type] -= ptes; 280 if (desc->type == SPT && (pgt->refs[0] || pgt->refs[1])) 284 if (!pgt->refs[type]) { 315 if (!pgt->refs[0]) 371 pgt->refs[typ [all...] |
/linux-master/net/mctp/test/ |
H A D | route-test.c | 31 refcount_set(&rt->rt.refs, 1); 66 unsigned int refs; local 76 refs = refcount_read(&rt->rt.refs); 77 KUNIT_ASSERT_EQ_MSG(test, refs, 1, "route ref imbalance"); 873 refcount_inc(&rt->rt.refs); 1010 refcount_inc(&rt->rt.refs);
|
/linux-master/tools/perf/arch/x86/util/ |
H A D | intel-pt.c | 911 struct intel_pt_snapshot_ref *refs; local 919 refs = calloc(new_cnt, sz); 920 if (!refs) 923 memcpy(refs, ptr->snapshot_refs, cnt * sz); 925 ptr->snapshot_refs = refs;
|
/linux-master/mm/ |
H A D | swap_state.c | 315 unsigned int refs[PAGEVEC_SIZE]; local 323 refs[folios.nr] = 1; 326 refs[folios.nr] = encoded_nr_pages(pages[++i]); 329 folios_put_refs(&folios, refs); 332 folios_put_refs(&folios, refs);
|
H A D | memremap.c | 513 bool __put_devmap_managed_page_refs(struct page *page, int refs) argument 523 if (page_ref_sub_return(page, refs) == 1)
|
/linux-master/include/uapi/linux/ |
H A D | btrfs_tree.h | 204 * root refs make a fast index for listing all of the snapshots and 229 * top of the existing extent item. However, unlike the other inline refs, 232 * Because of this, it goes at the front of the list of inline refs, and thus 234 * disk format rule that inline refs have non-decreasing type). 787 __le64 refs; member in struct:btrfs_extent_item 793 __le32 refs; member in struct:btrfs_extent_item_v0 933 __le32 refs; member in struct:btrfs_root_item 979 * this is used for both forward and backward root refs
|
/linux-master/include/trace/events/ |
H A D | btrfs.h | 301 __field( int, refs ) 313 __entry->refs = refcount_read(&map->refs); 318 "block_len=%llu flags=%s refs=%u", 327 __entry->refs) 528 __field( int, refs ) 542 __entry->refs = refcount_read(&ordered->refs); 551 "refs=%d", 561 __entry->compress_type, __entry->refs) [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
H A D | chan.c | 79 if (refcount_dec_and_mutex_lock(&cctx->refs, &chan->cgrp->mutex)) { 105 refcount_inc(&cctx->refs); 127 refcount_set(&cctx->refs, 1);
|
/linux-master/fs/ |
H A D | eventpoll.c | 220 struct hlist_head refs; member in struct:eventpoll 1461 static int reverse_path_check_proc(struct hlist_head *refs, int depth) argument 1470 hlist_for_each_entry_rcu(epi, refs, fllink) { 1471 struct hlist_head *refs = &epi->ep->refs; local 1472 if (hlist_empty(refs)) 1475 error = reverse_path_check_proc(refs, depth + 1); 1556 head = &ep->refs;
|
/linux-master/tools/bpf/bpftool/ |
H A D | main.h | 108 struct obj_ref *refs; member in struct:obj_refs
|
/linux-master/drivers/android/ |
H A D | binder_internal.h | 183 * @refs: list of references on this node 189 * @local_weak_refs: weak user refs from local process 192 * @local_strong_refs: strong user refs from local process 195 * @tmp_refs: temporary kernel refs 239 struct hlist_head refs; member in struct:binder_node 302 * @node_entry: list entry for node->refs list in target node 318 /* node => refs + procs (proc exit) */ 336 * @refs_by_desc: rbtree of refs ordered by ref->desc 338 * @refs_by_node: rbtree of refs ordered by ref->node 548 * @fbo: binder object (nodes and refs) [all...] |
/linux-master/kernel/ |
H A D | audit_tree.c | 31 atomic_long_t refs; member in struct:audit_chunk 77 * chunk is refcounted by embedded .refs. Mark associated with the chunk holds 141 if (atomic_long_dec_and_test(&chunk->refs)) 200 atomic_long_set(&chunk->refs, 1); 254 atomic_long_inc(&p->refs);
|
/linux-master/kernel/bpf/ |
H A D | log.c | 845 if (state->acquired_refs && state->refs[0].id) { 846 verbose(env, " refs=%d", state->refs[0].id); 848 if (state->refs[i].id) 849 verbose(env, ",%d", state->refs[i].id);
|