/linux-master/kernel/bpf/ |
H A D | verifier.c | 1266 dst->refs = copy_array(dst->refs, src->refs, src->acquired_refs, 1268 if (!dst->refs) 1290 state->refs = realloc_array(state->refs, state->acquired_refs, n, 1292 if (!state->refs) 1326 /* Acquire a pointer id from the env and update the state->refs to include 1341 state->refs[new_ofs].id = id; 1342 state->refs[new_of [all...] |
/linux-master/kernel/sched/ |
H A D | core.c | 2546 * @refs: number of wait_for_completion() 2550 refcount_t refs; member in struct:set_affinity_pending 3013 refcount_set(&my_pending.refs, 1); 3024 refcount_inc(&pending->refs); 3096 if (refcount_dec_and_test(&pending->refs)) 3097 wake_up_var(&pending->refs); /* No UaF, just an address */ 3103 wait_var_event(&my_pending.refs, !refcount_read(&my_pending.refs));
|
/linux-master/kernel/ |
H A D | seccomp.c | 193 * @refs: Reference count to manage the object lifetime. 196 * requested for the user notifier. When @refs reaches zero, 204 * or equal to @refs. Hence, reaching 0 for @users does not mean 224 * to a task_struct (other than @refs). 227 refcount_t refs; member in struct:seccomp_filter 543 while (orig && refcount_dec_and_test(&orig->refs)) { 687 refcount_set(&sfilter->refs, 1); 930 refcount_inc(&filter->refs);
|
/linux-master/kernel/trace/ |
H A D | trace_events_user.c | 165 struct user_event_refs *refs; member in struct:user_event_file_info 821 * Put for mm must be done after RCU delay to handle new refs in 822 * between the list_del_rcu() and now. This ensures any get refs 2171 struct user_event_refs *refs; local 2185 refs = rcu_dereference_sched(info->refs); 2188 * The refs->events array is protected by RCU, and new items may be 2192 if (likely(refs && idx < refs->count)) 2193 user = refs 2289 struct user_event_refs *refs, *new_refs; local 2641 struct user_event_refs *refs; local [all...] |
/linux-master/lib/ |
H A D | refcount.c | 59 return atomic_try_cmpxchg_release(&r->refs, &val, 0); 76 unsigned int new, val = atomic_read(&r->refs); 91 } while (!atomic_try_cmpxchg_release(&r->refs, &val, new));
|
H A D | test_objpool.c | 121 unsigned long refs; member in struct:ot_node 339 on->refs++; 490 on->refs++; 527 on->refs++;
|
/linux-master/mm/ |
H A D | filemap.c | 238 int refs = 1; local 245 refs = folio_nr_pages(folio); 246 folio_put_refs(folio, refs);
|
H A D | gup.c | 71 static inline struct folio *try_get_folio(struct page *page, int refs) argument 79 if (unlikely(!folio_ref_try_add_rcu(folio, refs))) 92 if (!put_devmap_managed_page_refs(&folio->page, refs)) 93 folio_put_refs(folio, refs); 103 * @refs: the value to (effectively) add to the folio's refcount 113 * FOLL_GET: folio's refcount will be incremented by @refs. 116 * @refs, and its pincount will be incremented by @refs. 119 * @refs * GUP_PIN_COUNTING_BIAS. 126 struct folio *try_grab_folio(struct page *page, int refs, unsigne argument 189 gup_put_folio(struct folio *folio, int refs, unsigned int flags) argument 2809 int refs; local 2882 int refs; local 2926 int refs; local 2969 int refs; local [all...] |
H A D | hugetlb.c | 350 struct hugetlb_vma_lock, refs); 367 kref_put(&vma_lock->refs, hugetlb_vma_lock_release); 428 kref_init(&vma_lock->refs); 1094 kref_init(&resv_map->refs); 1117 struct resv_map *resv_map = container_of(ref, struct resv_map, refs); 1242 kref_put(&reservations->refs, resv_map_release); 5188 kref_get(&resv->refs); 5239 kref_put(&resv->refs, resv_map_release); 7249 kref_put(&resv_map->refs, resv_map_release);
|
H A D | internal.h | 1099 struct folio *try_grab_folio(struct page *page, int refs, unsigned int flags);
|
H A D | memremap.c | 513 bool __put_devmap_managed_page_refs(struct page *page, int refs) argument 523 if (page_ref_sub_return(page, refs) == 1)
|
H A D | migrate.c | 384 int refs = 1; local 386 return refs; 388 refs += folio_nr_pages(folio); 390 refs++; 392 return refs; 455 /* Move dirty while page refs frozen and newpage not yet exposed */ 745 /* Check whether page does not have extra refs before we do more work */
|
H A D | page_owner.c | 217 if (atomic_try_cmpxchg_relaxed(&stack_record->count.refs, &old, 1))
|
H A D | swap.c | 966 * @refs: The number of refs to subtract from each folio. 972 * to reinitialise it. If @refs is NULL, we subtract one from each 978 void folios_put_refs(struct folio_batch *folios, unsigned int *refs) argument 986 unsigned int nr_refs = refs ? refs[i] : 1; 1053 int refs[PAGEVEC_SIZE]; local 1063 refs[fbatch.nr] = 1; 1066 refs[fbatch.nr] = encoded_nr_pages(encoded[++i]); 1070 folios_put_refs(&fbatch, refs); [all...] |
H A D | swap_state.c | 315 unsigned int refs[PAGEVEC_SIZE]; local 323 refs[folios.nr] = 1; 326 refs[folios.nr] = encoded_nr_pages(pages[++i]); 329 folios_put_refs(&folios, refs); 332 folios_put_refs(&folios, refs);
|
H A D | vmscan.c | 4236 int refs = folio_lru_refs(folio); local 4237 int tier = lru_tier_from_refs(refs); 4268 if (tier > tier_idx || refs == BIT(LRU_REFS_WIDTH)) {
|
H A D | workingset.c | 241 int refs = folio_lru_refs(folio); local 242 int tier = lru_tier_from_refs(refs); 251 token = (min_seq << LRU_REFS_WIDTH) | max(refs - 1, 0); 256 return pack_shadow(mem_cgroup_id(memcg), pgdat, token, refs); 283 int hist, tier, refs; local 306 refs = (token & (BIT(LRU_REFS_WIDTH) - 1)) + workingset; 307 tier = lru_tier_from_refs(refs); 319 if (lru_gen_in_fault() || refs >= BIT(LRU_REFS_WIDTH) - 1) {
|
/linux-master/net/core/ |
H A D | datagram.c | 627 int refs, order, n = 0; local 657 for (refs = 0; copied != 0; start = 0) { 680 refs++; 684 if (refs) { 685 page_ref_sub(last_head, refs); 686 refs = 0; 690 if (refs) 691 page_ref_sub(last_head, refs);
|
H A D | skbuff.c | 6855 refcount_inc(&flow->key->refs);
|
/linux-master/net/mctp/ |
H A D | device.c | 39 if (!refcount_inc_not_zero(&mdev->refs)) 310 refcount_inc(&mdev->refs); 315 if (mdev && refcount_dec_and_test(&mdev->refs)) { 355 refcount_set(&mdev->refs, 1);
|
H A D | route.c | 153 refcount_inc(&key->refs); 188 refcount_set(&key->refs, 1); 198 if (!refcount_dec_and_test(&key->refs)) 201 /* even though no refs exist here, the lock allows us to stay 239 refcount_inc(&key->refs); 301 refcount_inc(&key->refs); 590 if (refcount_dec_and_test(&rt->refs)) { 606 refcount_set(&rt->refs, 1); 641 refcount_inc(&key->refs); 762 refcount_inc(&key->refs); [all...] |
/linux-master/net/mctp/test/ |
H A D | route-test.c | 31 refcount_set(&rt->rt.refs, 1); 66 unsigned int refs; local 76 refs = refcount_read(&rt->rt.refs); 77 KUNIT_ASSERT_EQ_MSG(test, refs, 1, "route ref imbalance"); 873 refcount_inc(&rt->rt.refs); 1010 refcount_inc(&rt->rt.refs);
|
/linux-master/net/netfilter/ |
H A D | nf_tables_api.c | 5252 refcount_set(&set->refs, 1); 5322 if (refcount_dec_and_test(&set->refs)) { 9839 refcount_inc(&set->refs);
|
/linux-master/net/xdp/ |
H A D | xsk_diag.c | 67 du.refs = refcount_read(&umem->users);
|
/linux-master/sound/core/ |
H A D | pcm_native.c | 82 refcount_set(&group->refs, 1); 1313 do_free = refcount_dec_and_test(&group->refs); 1335 refcount_inc(&group->refs); 2280 refcount_inc(&target_group->refs); 2308 refcount_dec(&group->refs); 2315 do_free = refcount_dec_and_test(&group->refs);
|