Searched refs:refs (Results 126 - 150 of 170) sorted by path

1234567

/linux-master/kernel/bpf/
H A Dverifier.c1266 dst->refs = copy_array(dst->refs, src->refs, src->acquired_refs,
1268 if (!dst->refs)
1290 state->refs = realloc_array(state->refs, state->acquired_refs, n,
1292 if (!state->refs)
1326 /* Acquire a pointer id from the env and update the state->refs to include
1341 state->refs[new_ofs].id = id;
1342 state->refs[new_of
[all...]
/linux-master/kernel/sched/
H A Dcore.c2546 * @refs: number of wait_for_completion()
2550 refcount_t refs; member in struct:set_affinity_pending
3013 refcount_set(&my_pending.refs, 1);
3024 refcount_inc(&pending->refs);
3096 if (refcount_dec_and_test(&pending->refs))
3097 wake_up_var(&pending->refs); /* No UaF, just an address */
3103 wait_var_event(&my_pending.refs, !refcount_read(&my_pending.refs));
/linux-master/kernel/
H A Dseccomp.c193 * @refs: Reference count to manage the object lifetime.
196 * requested for the user notifier. When @refs reaches zero,
204 * or equal to @refs. Hence, reaching 0 for @users does not mean
224 * to a task_struct (other than @refs).
227 refcount_t refs; member in struct:seccomp_filter
543 while (orig && refcount_dec_and_test(&orig->refs)) {
687 refcount_set(&sfilter->refs, 1);
930 refcount_inc(&filter->refs);
/linux-master/kernel/trace/
H A Dtrace_events_user.c165 struct user_event_refs *refs; member in struct:user_event_file_info
821 * Put for mm must be done after RCU delay to handle new refs in
822 * between the list_del_rcu() and now. This ensures any get refs
2171 struct user_event_refs *refs; local
2185 refs = rcu_dereference_sched(info->refs);
2188 * The refs->events array is protected by RCU, and new items may be
2192 if (likely(refs && idx < refs->count))
2193 user = refs
2289 struct user_event_refs *refs, *new_refs; local
2641 struct user_event_refs *refs; local
[all...]
/linux-master/lib/
H A Drefcount.c59 return atomic_try_cmpxchg_release(&r->refs, &val, 0);
76 unsigned int new, val = atomic_read(&r->refs);
91 } while (!atomic_try_cmpxchg_release(&r->refs, &val, new));
H A Dtest_objpool.c121 unsigned long refs; member in struct:ot_node
339 on->refs++;
490 on->refs++;
527 on->refs++;
/linux-master/mm/
H A Dfilemap.c238 int refs = 1; local
245 refs = folio_nr_pages(folio);
246 folio_put_refs(folio, refs);
H A Dgup.c71 static inline struct folio *try_get_folio(struct page *page, int refs) argument
79 if (unlikely(!folio_ref_try_add_rcu(folio, refs)))
92 if (!put_devmap_managed_page_refs(&folio->page, refs))
93 folio_put_refs(folio, refs);
103 * @refs: the value to (effectively) add to the folio's refcount
113 * FOLL_GET: folio's refcount will be incremented by @refs.
116 * @refs, and its pincount will be incremented by @refs.
119 * @refs * GUP_PIN_COUNTING_BIAS.
126 struct folio *try_grab_folio(struct page *page, int refs, unsigne argument
189 gup_put_folio(struct folio *folio, int refs, unsigned int flags) argument
2809 int refs; local
2882 int refs; local
2926 int refs; local
2969 int refs; local
[all...]
H A Dhugetlb.c350 struct hugetlb_vma_lock, refs);
367 kref_put(&vma_lock->refs, hugetlb_vma_lock_release);
428 kref_init(&vma_lock->refs);
1094 kref_init(&resv_map->refs);
1117 struct resv_map *resv_map = container_of(ref, struct resv_map, refs);
1242 kref_put(&reservations->refs, resv_map_release);
5188 kref_get(&resv->refs);
5239 kref_put(&resv->refs, resv_map_release);
7249 kref_put(&resv_map->refs, resv_map_release);
H A Dinternal.h1099 struct folio *try_grab_folio(struct page *page, int refs, unsigned int flags);
H A Dmemremap.c513 bool __put_devmap_managed_page_refs(struct page *page, int refs) argument
523 if (page_ref_sub_return(page, refs) == 1)
H A Dmigrate.c384 int refs = 1; local
386 return refs;
388 refs += folio_nr_pages(folio);
390 refs++;
392 return refs;
455 /* Move dirty while page refs frozen and newpage not yet exposed */
745 /* Check whether page does not have extra refs before we do more work */
H A Dpage_owner.c217 if (atomic_try_cmpxchg_relaxed(&stack_record->count.refs, &old, 1))
H A Dswap.c966 * @refs: The number of refs to subtract from each folio.
972 * to reinitialise it. If @refs is NULL, we subtract one from each
978 void folios_put_refs(struct folio_batch *folios, unsigned int *refs) argument
986 unsigned int nr_refs = refs ? refs[i] : 1;
1053 int refs[PAGEVEC_SIZE]; local
1063 refs[fbatch.nr] = 1;
1066 refs[fbatch.nr] = encoded_nr_pages(encoded[++i]);
1070 folios_put_refs(&fbatch, refs);
[all...]
H A Dswap_state.c315 unsigned int refs[PAGEVEC_SIZE]; local
323 refs[folios.nr] = 1;
326 refs[folios.nr] = encoded_nr_pages(pages[++i]);
329 folios_put_refs(&folios, refs);
332 folios_put_refs(&folios, refs);
H A Dvmscan.c4236 int refs = folio_lru_refs(folio); local
4237 int tier = lru_tier_from_refs(refs);
4268 if (tier > tier_idx || refs == BIT(LRU_REFS_WIDTH)) {
H A Dworkingset.c241 int refs = folio_lru_refs(folio); local
242 int tier = lru_tier_from_refs(refs);
251 token = (min_seq << LRU_REFS_WIDTH) | max(refs - 1, 0);
256 return pack_shadow(mem_cgroup_id(memcg), pgdat, token, refs);
283 int hist, tier, refs; local
306 refs = (token & (BIT(LRU_REFS_WIDTH) - 1)) + workingset;
307 tier = lru_tier_from_refs(refs);
319 if (lru_gen_in_fault() || refs >= BIT(LRU_REFS_WIDTH) - 1) {
/linux-master/net/core/
H A Ddatagram.c627 int refs, order, n = 0; local
657 for (refs = 0; copied != 0; start = 0) {
680 refs++;
684 if (refs) {
685 page_ref_sub(last_head, refs);
686 refs = 0;
690 if (refs)
691 page_ref_sub(last_head, refs);
H A Dskbuff.c6855 refcount_inc(&flow->key->refs);
/linux-master/net/mctp/
H A Ddevice.c39 if (!refcount_inc_not_zero(&mdev->refs))
310 refcount_inc(&mdev->refs);
315 if (mdev && refcount_dec_and_test(&mdev->refs)) {
355 refcount_set(&mdev->refs, 1);
H A Droute.c153 refcount_inc(&key->refs);
188 refcount_set(&key->refs, 1);
198 if (!refcount_dec_and_test(&key->refs))
201 /* even though no refs exist here, the lock allows us to stay
239 refcount_inc(&key->refs);
301 refcount_inc(&key->refs);
590 if (refcount_dec_and_test(&rt->refs)) {
606 refcount_set(&rt->refs, 1);
641 refcount_inc(&key->refs);
762 refcount_inc(&key->refs);
[all...]
/linux-master/net/mctp/test/
H A Droute-test.c31 refcount_set(&rt->rt.refs, 1);
66 unsigned int refs; local
76 refs = refcount_read(&rt->rt.refs);
77 KUNIT_ASSERT_EQ_MSG(test, refs, 1, "route ref imbalance");
873 refcount_inc(&rt->rt.refs);
1010 refcount_inc(&rt->rt.refs);
/linux-master/net/netfilter/
H A Dnf_tables_api.c5252 refcount_set(&set->refs, 1);
5322 if (refcount_dec_and_test(&set->refs)) {
9839 refcount_inc(&set->refs);
/linux-master/net/xdp/
H A Dxsk_diag.c67 du.refs = refcount_read(&umem->users);
/linux-master/sound/core/
H A Dpcm_native.c82 refcount_set(&group->refs, 1);
1313 do_free = refcount_dec_and_test(&group->refs);
1335 refcount_inc(&group->refs);
2280 refcount_inc(&target_group->refs);
2308 refcount_dec(&group->refs);
2315 do_free = refcount_dec_and_test(&group->refs);

Completed in 586 milliseconds

1234567