Searched refs:memcg (Results 1 - 25 of 49) sorted by relevance

12

/linux-master/mm/
H A Dmemcontrol.c24 * Per memcg lru locking
138 * memcg which the event belongs to.
140 struct mem_cgroup *memcg; member in struct:mem_cgroup_event
154 int (*register_event)(struct mem_cgroup *memcg,
161 void (*unregister_event)(struct mem_cgroup *memcg,
173 static void mem_cgroup_threshold(struct mem_cgroup *memcg);
174 static void mem_cgroup_oom_notify(struct mem_cgroup *memcg);
242 struct vmpressure *memcg_to_vmpressure(struct mem_cgroup *memcg) argument
244 if (!memcg)
245 memcg
329 memcg_reparent_objcgs(struct mem_cgroup *memcg, struct mem_cgroup *parent) argument
377 struct mem_cgroup *memcg = folio_memcg(folio); local
400 struct mem_cgroup *memcg; local
473 soft_limit_excess(struct mem_cgroup *memcg) argument
485 mem_cgroup_update_tree(struct mem_cgroup *memcg, int nid) argument
528 mem_cgroup_remove_from_trees(struct mem_cgroup *memcg) argument
714 memcg_rstat_updated(struct mem_cgroup *memcg, int val) argument
740 do_flush_stats(struct mem_cgroup *memcg) argument
757 mem_cgroup_flush_stats(struct mem_cgroup *memcg) argument
769 mem_cgroup_flush_stats_ratelimited(struct mem_cgroup *memcg) argument
786 memcg_page_state(struct mem_cgroup *memcg, int idx) argument
818 __mod_memcg_state(struct mem_cgroup *memcg, int idx, int val) argument
828 memcg_page_state_local(struct mem_cgroup *memcg, int idx) argument
843 struct mem_cgroup *memcg; local
903 struct mem_cgroup *memcg; local
925 struct mem_cgroup *memcg; local
952 __count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, unsigned long count) argument
966 memcg_events(struct mem_cgroup *memcg, int event) argument
975 memcg_events_local(struct mem_cgroup *memcg, int event) argument
985 mem_cgroup_charge_statistics(struct mem_cgroup *memcg, int nr_pages) argument
999 mem_cgroup_event_ratelimit(struct mem_cgroup *memcg, enum mem_cgroup_events_target target) argument
1028 memcg_check_events(struct mem_cgroup *memcg, int nid) argument
1081 struct mem_cgroup *memcg; local
1123 struct mem_cgroup *memcg; local
1162 struct mem_cgroup *memcg = NULL; local
1287 struct mem_cgroup *memcg = dead_memcg; local
1319 mem_cgroup_scan_tasks(struct mem_cgroup *memcg, int (*fn)(struct task_struct *, void *), void *arg) argument
1345 struct mem_cgroup *memcg; local
1474 mem_cgroup_margin(struct mem_cgroup *memcg) argument
1504 mem_cgroup_under_move(struct mem_cgroup *memcg) argument
1526 mem_cgroup_wait_acct_move(struct mem_cgroup *memcg) argument
1628 memcg_page_state_output(struct mem_cgroup *memcg, int item) argument
1635 memcg_page_state_local_output( struct mem_cgroup *memcg, int item) argument
1642 memcg_stat_format(struct mem_cgroup *memcg, struct seq_buf *s) argument
1697 memory_stat_format(struct mem_cgroup *memcg, struct seq_buf *s) argument
1715 mem_cgroup_print_oom_context(struct mem_cgroup *memcg, struct task_struct *p) argument
1736 mem_cgroup_print_oom_meminfo(struct mem_cgroup *memcg) argument
1771 mem_cgroup_get_max(struct mem_cgroup *memcg) argument
1790 mem_cgroup_size(struct mem_cgroup *memcg) argument
1795 mem_cgroup_out_of_memory(struct mem_cgroup *memcg, gfp_t gfp_mask, int order) argument
1886 mem_cgroup_oom_trylock(struct mem_cgroup *memcg) argument
1925 mem_cgroup_oom_unlock(struct mem_cgroup *memcg) argument
1936 mem_cgroup_mark_under_oom(struct mem_cgroup *memcg) argument
1946 mem_cgroup_unmark_under_oom(struct mem_cgroup *memcg) argument
1964 struct mem_cgroup *memcg; member in struct:oom_wait_info
1984 memcg_oom_recover(struct mem_cgroup *memcg) argument
2002 mem_cgroup_oom(struct mem_cgroup *memcg, gfp_t mask, int order) argument
2074 struct mem_cgroup *memcg = current->memcg_in_oom; local
2125 struct mem_cgroup *memcg; local
2168 mem_cgroup_print_oom_group(struct mem_cgroup *memcg) argument
2187 struct mem_cgroup *memcg; local
2229 __folio_memcg_unlock(struct mem_cgroup *memcg) argument
2294 memcg_account_kmem(struct mem_cgroup *memcg, int nr_pages) argument
2310 consume_stock(struct mem_cgroup *memcg, unsigned int nr_pages) argument
2380 __refill_stock(struct mem_cgroup *memcg, unsigned int nr_pages) argument
2396 refill_stock(struct mem_cgroup *memcg, unsigned int nr_pages) argument
2426 struct mem_cgroup *memcg; local
2460 reclaim_high(struct mem_cgroup *memcg, unsigned int nr_pages, gfp_t gfp_mask) argument
2488 struct mem_cgroup *memcg; local
2565 mem_find_max_overage(struct mem_cgroup *memcg) argument
2579 swap_find_max_overage(struct mem_cgroup *memcg) argument
2599 calculate_high_delay(struct mem_cgroup *memcg, unsigned int nr_pages, u64 max_overage) argument
2643 struct mem_cgroup *memcg; local
2729 try_charge_memcg(struct mem_cgroup *memcg, gfp_t gfp_mask, unsigned int nr_pages) argument
2925 try_charge(struct mem_cgroup *memcg, gfp_t gfp_mask, unsigned int nr_pages) argument
2939 mem_cgroup_cancel_charge(struct mem_cgroup *memcg, unsigned int nr_pages) argument
2949 commit_charge(struct folio *folio, struct mem_cgroup *memcg) argument
2969 mem_cgroup_commit_charge(struct folio *folio, struct mem_cgroup *memcg) argument
2997 struct mem_cgroup *memcg; local
3124 __get_obj_cgroup_from_memcg(struct mem_cgroup *memcg) argument
3139 struct mem_cgroup *memcg; local
3191 struct mem_cgroup *memcg; local
3243 struct mem_cgroup *memcg; local
3256 memcg_account_kmem(struct mem_cgroup *memcg, int nr_pages) argument
3276 struct mem_cgroup *memcg; local
3297 struct mem_cgroup *memcg; local
3456 struct mem_cgroup *memcg; local
3511 struct mem_cgroup *memcg; local
3613 struct mem_cgroup *memcg = folio_memcg(folio); local
3670 mem_cgroup_resize_max(struct mem_cgroup *memcg, unsigned long max, bool memsw) argument
3811 mem_cgroup_force_empty(struct mem_cgroup *memcg) argument
3837 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
3863 mem_cgroup_usage(struct mem_cgroup *memcg, bool swap) argument
3896 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
3947 memcg_online_kmem(struct mem_cgroup *memcg) argument
3973 memcg_offline_kmem(struct mem_cgroup *memcg) argument
3998 memcg_online_kmem(struct mem_cgroup *memcg) argument
4002 memcg_offline_kmem(struct mem_cgroup *memcg) argument
4007 memcg_update_tcp_max(struct mem_cgroup *memcg, unsigned long max) argument
4049 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
4098 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
4142 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
4174 mem_cgroup_node_nr_lru_pages(struct mem_cgroup *memcg, int nid, unsigned int lru_mask, bool tree) argument
4194 mem_cgroup_nr_lru_pages(struct mem_cgroup *memcg, unsigned int lru_mask, bool tree) argument
4227 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
4302 memcg1_stat_format(struct mem_cgroup *memcg, struct seq_buf *s) argument
4379 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
4387 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
4400 __mem_cgroup_threshold(struct mem_cgroup *memcg, bool swap) argument
4451 mem_cgroup_threshold(struct mem_cgroup *memcg) argument
4476 mem_cgroup_oom_notify_cb(struct mem_cgroup *memcg) argument
4489 mem_cgroup_oom_notify(struct mem_cgroup *memcg) argument
4497 __mem_cgroup_usage_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args, enum res_type type) argument
4577 mem_cgroup_usage_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args) argument
4583 memsw_cgroup_usage_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args) argument
4589 __mem_cgroup_usage_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, enum res_type type) argument
4674 mem_cgroup_usage_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd) argument
4680 memsw_cgroup_usage_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd) argument
4686 mem_cgroup_oom_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args) argument
4708 mem_cgroup_oom_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd) argument
4727 struct mem_cgroup *memcg = mem_cgroup_from_seq(sf); local
4739 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
4756 memcg_wb_domain_init(struct mem_cgroup *memcg, gfp_t gfp) argument
4761 memcg_wb_domain_exit(struct mem_cgroup *memcg) argument
4766 memcg_wb_domain_size_changed(struct mem_cgroup *memcg) argument
4773 struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css); local
4803 struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css); local
4871 struct mem_cgroup *memcg = folio_memcg(folio); local
4923 struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css); local
4950 memcg_wb_domain_init(struct mem_cgroup *memcg, gfp_t gfp) argument
4955 memcg_wb_domain_exit(struct mem_cgroup *memcg) argument
4959 memcg_wb_domain_size_changed(struct mem_cgroup *memcg) argument
4987 struct mem_cgroup *memcg = event->memcg; local
5011 struct mem_cgroup *memcg = event->memcg; local
5061 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
5370 mem_cgroup_id_remove(struct mem_cgroup *memcg) argument
5378 mem_cgroup_id_get_many(struct mem_cgroup *memcg, unsigned int n) argument
5384 mem_cgroup_id_put_many(struct mem_cgroup *memcg, unsigned int n) argument
5394 mem_cgroup_id_put(struct mem_cgroup *memcg) argument
5416 struct mem_cgroup *memcg; local
5434 alloc_mem_cgroup_per_node_info(struct mem_cgroup *memcg, int node) argument
5456 free_mem_cgroup_per_node_info(struct mem_cgroup *memcg, int node) argument
5467 __mem_cgroup_free(struct mem_cgroup *memcg) argument
5481 mem_cgroup_free(struct mem_cgroup *memcg) argument
5491 struct mem_cgroup *memcg; local
5566 struct mem_cgroup *memcg, *old_memcg; local
5614 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
5658 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
5690 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
5698 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
5738 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
5754 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
6358 struct mem_cgroup *memcg = NULL; /* unneeded init to make gcc happy */ local
6694 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
6702 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
6716 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
6739 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
6762 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
6811 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
6873 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
6881 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
6889 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
6913 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
6942 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
6952 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
6974 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
7232 mem_cgroup_calculate_protection(struct mem_cgroup *root, struct mem_cgroup *memcg) argument
7279 charge_memcg(struct folio *folio, struct mem_cgroup *memcg, gfp_t gfp) argument
7295 struct mem_cgroup *memcg; local
7322 mem_cgroup_hugetlb_try_charge(struct mem_cgroup *memcg, gfp_t gfp, long nr_pages) argument
7355 struct mem_cgroup *memcg; local
7409 struct mem_cgroup *memcg; member in struct:uncharge_gather
7447 struct mem_cgroup *memcg; local
7541 struct mem_cgroup *memcg; local
7591 struct mem_cgroup *memcg; local
7632 struct mem_cgroup *memcg; local
7668 mem_cgroup_charge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages, gfp_t gfp_mask) argument
7699 mem_cgroup_uncharge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages) argument
7772 mem_cgroup_id_get_online(struct mem_cgroup *memcg) argument
7799 struct mem_cgroup *memcg, *swap_memcg; local
7871 struct mem_cgroup *memcg; local
7915 struct mem_cgroup *memcg; local
7934 mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg) argument
7949 struct mem_cgroup *memcg; local
7989 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
7997 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
8011 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
8034 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
8050 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
8136 struct mem_cgroup *memcg, *original_memcg; local
8180 struct mem_cgroup *memcg; local
8207 struct mem_cgroup *memcg; local
8221 mem_cgroup_zswap_writeback_enabled(struct mem_cgroup *memcg) argument
8230 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local
8245 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
8261 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local
8270 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local
[all...]
H A Dshrinker_debug.c19 struct mem_cgroup *memcg,
30 .memcg = memcg,
51 struct mem_cgroup *memcg; local
64 memcg = mem_cgroup_iter(NULL, NULL, NULL);
66 if (memcg && !mem_cgroup_online(memcg))
70 memcg_aware ? memcg : NULL,
73 seq_printf(m, "%lu", mem_cgroup_ino(memcg));
80 mem_cgroup_iter_break(NULL, memcg);
18 shrinker_count_objects(struct shrinker *shrinker, struct mem_cgroup *memcg, unsigned long *count_per_node) argument
113 struct mem_cgroup *memcg = NULL; local
[all...]
H A Dshrinker.c62 void free_shrinker_info(struct mem_cgroup *memcg) argument
69 pn = memcg->nodeinfo[nid];
77 int alloc_shrinker_info(struct mem_cgroup *memcg) argument
92 rcu_assign_pointer(memcg->nodeinfo[nid]->shrinker_info, info);
100 free_shrinker_info(memcg);
104 static struct shrinker_info *shrinker_info_protected(struct mem_cgroup *memcg, argument
107 return rcu_dereference_protected(memcg->nodeinfo[nid]->shrinker_info,
111 static int expand_one_shrinker_info(struct mem_cgroup *memcg, int new_size, argument
119 pn = memcg->nodeinfo[nid];
120 old = shrinker_info_protected(memcg, ni
153 struct mem_cgroup *memcg; local
194 set_shrinker_bit(struct mem_cgroup *memcg, int nid, int shrinker_id) argument
250 xchg_nr_deferred_memcg(int nid, struct shrinker *shrinker, struct mem_cgroup *memcg) argument
266 add_nr_deferred_memcg(long nr, int nid, struct shrinker *shrinker, struct mem_cgroup *memcg) argument
283 reparent_shrinker_deferred(struct mem_cgroup *memcg) argument
321 xchg_nr_deferred_memcg(int nid, struct shrinker *shrinker, struct mem_cgroup *memcg) argument
327 add_nr_deferred_memcg(long nr, int nid, struct shrinker *shrinker, struct mem_cgroup *memcg) argument
467 shrink_slab_memcg(gfp_t gfp_mask, int nid, struct mem_cgroup *memcg, int priority) argument
585 shrink_slab_memcg(gfp_t gfp_mask, int nid, struct mem_cgroup *memcg, int priority) argument
612 shrink_slab(gfp_t gfp_mask, int nid, struct mem_cgroup *memcg, int priority) argument
[all...]
H A Dvmpressure.c77 struct mem_cgroup *memcg = vmpressure_to_memcg(vmpr); local
79 memcg = parent_mem_cgroup(memcg);
80 if (!memcg)
82 return memcg_to_vmpressure(memcg);
221 * @memcg: cgroup memory controller handle
231 * mode: @memcg is considered the pressure root and userspace is
234 * If @tree is not set, reclaim efficiency is recorded for @memcg, and
239 void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, argument
249 * for this @memcg rathe
335 vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio) argument
374 vmpressure_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args) argument
435 vmpressure_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd) argument
[all...]
H A Dlist_lru.c89 struct mem_cgroup *memcg)
96 l = list_lru_from_memcg_idx(lru, nid, memcg_kmem_id(memcg));
100 set_shrinker_bit(memcg, nid, lru_shrinker_id(lru));
113 struct mem_cgroup *memcg = list_lru_memcg_aware(lru) ? local
116 return list_lru_add(lru, item, nid, memcg);
121 struct mem_cgroup *memcg)
128 l = list_lru_from_memcg_idx(lru, nid, memcg_kmem_id(memcg));
143 struct mem_cgroup *memcg = list_lru_memcg_aware(lru) ? local
146 return list_lru_del(lru, item, nid, memcg);
166 int nid, struct mem_cgroup *memcg)
88 list_lru_add(struct list_lru *lru, struct list_head *item, int nid, struct mem_cgroup *memcg) argument
120 list_lru_del(struct list_lru *lru, struct list_head *item, int nid, struct mem_cgroup *memcg) argument
165 list_lru_count_one(struct list_lru *lru, int nid, struct mem_cgroup *memcg) argument
258 list_lru_walk_one(struct list_lru *lru, int nid, struct mem_cgroup *memcg, list_lru_walk_cb isolate, void *cb_arg, unsigned long *nr_to_walk) argument
274 list_lru_walk_one_irq(struct list_lru *lru, int nid, struct mem_cgroup *memcg, list_lru_walk_cb isolate, void *cb_arg, unsigned long *nr_to_walk) argument
420 memcg_reparent_list_lrus(struct mem_cgroup *memcg, struct mem_cgroup *parent) argument
454 memcg_list_lru_allocated(struct mem_cgroup *memcg, struct list_lru *lru) argument
462 memcg_list_lru_alloc(struct mem_cgroup *memcg, struct list_lru *lru, gfp_t gfp) argument
469 struct mem_cgroup *memcg; member in struct:list_lru_memcg_table
[all...]
H A Dworkingset.c243 struct mem_cgroup *memcg = folio_memcg(folio); local
248 lruvec = mem_cgroup_lruvec(memcg, pgdat);
256 return pack_shadow(mem_cgroup_id(memcg), pgdat, token, refs);
268 struct mem_cgroup *memcg; local
273 memcg = mem_cgroup_from_id(memcg_id);
274 *lruvec = mem_cgroup_lruvec(memcg, pgdat);
444 * Look up the memcg associated with the stored ID. It might
470 * XXX: With per-memcg flushing and thresholding, is ratelimiting
527 * evicted folio in the context of the node and the memcg whose memory
534 struct mem_cgroup *memcg; local
586 struct mem_cgroup *memcg; local
[all...]
H A Dmmap_lock.c32 * Size of the buffer for memcg path names. Ignoring stack trace support,
188 * Write the given mm_struct's memcg path to a percpu buffer, and return a
202 struct mem_cgroup *memcg = get_mem_cgroup_from_mm(mm); local
204 if (memcg == NULL)
206 if (unlikely(memcg->css.cgroup == NULL))
213 cgroup_path(memcg->css.cgroup, buf, MEMCG_PATH_BUF_SIZE);
216 css_put(&memcg->css);
H A Dzswap.c738 struct mem_cgroup *memcg; local
743 * concurrent memcg offlining. Thanks to the memcg->kmemcg_id indirection
746 * 1. list_lru_add() is called before memcg->kmemcg_id is updated. The
747 * new entry will be reparented to memcg's parent's list_lru.
748 * 2. list_lru_add() is called after memcg->kmemcg_id is updated. The
749 * new entry will be added directly to memcg's parent's list_lru.
754 memcg = mem_cgroup_from_entry(entry);
756 list_lru_add(list_lru, &entry->lru, nid, memcg);
759 lru_size = list_lru_count_one(list_lru, nid, memcg);
776 struct mem_cgroup *memcg; local
800 zswap_memcg_offline_cleanup(struct mem_cgroup *memcg) argument
1319 struct mem_cgroup *memcg = sc->memcg; local
1381 shrink_memcg(struct mem_cgroup *memcg) argument
1406 struct mem_cgroup *memcg; local
1498 struct mem_cgroup *memcg = NULL; local
[all...]
H A Dvmscan.c218 * completely broken with the legacy memcg and direct stalling in
279 * single memcg. For example, a memcg-aware shrinker can free one object
280 * charged to the target memcg, causing an entire page to be freed.
281 * If we count the entire page as reclaimed from the memcg, we end up
285 * from the target memcg; preventing unnecessary retries during memcg
289 * charged to the target memcg, we end up underestimating the reclaimed
296 * memcg reclaim, to make reporting more accurate and reduce
317 static inline bool can_reclaim_anon_pages(struct mem_cgroup *memcg, argument
389 struct mem_cgroup *memcg = NULL; local
2338 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
2589 get_lruvec(struct mem_cgroup *memcg, int nid) argument
2611 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
2737 get_mm_list(struct mem_cgroup *memcg) argument
2779 struct mem_cgroup *memcg = get_mem_cgroup_from_mm(mm); local
2807 struct mem_cgroup *memcg = NULL; local
2845 struct mem_cgroup *memcg; local
2874 get_mm_list(struct mem_cgroup *memcg) argument
2922 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
2983 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
3299 get_pfn_folio(unsigned long pfn, struct mem_cgroup *memcg, struct pglist_data *pgdat, bool can_swap) argument
3340 struct mem_cgroup *memcg = lruvec_memcg(walk->lruvec); local
3408 struct mem_cgroup *memcg = lruvec_memcg(walk->lruvec); local
3614 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
3893 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
3917 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
3940 struct mem_cgroup *memcg; local
4000 struct mem_cgroup *memcg = folio_memcg(folio); local
4151 lru_gen_online_memcg(struct mem_cgroup *memcg) argument
4176 lru_gen_offline_memcg(struct mem_cgroup *memcg) argument
4187 lru_gen_release_memcg(struct mem_cgroup *memcg) argument
4213 lru_gen_soft_reclaim(struct mem_cgroup *memcg, int nid) argument
4343 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
4520 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
4669 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
4756 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
4801 struct mem_cgroup *memcg; local
5039 struct mem_cgroup *memcg; local
5174 struct mem_cgroup *memcg; local
5206 struct mem_cgroup *memcg = lruvec_memcg(v); local
5285 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local
5394 struct mem_cgroup *memcg = NULL; local
5567 lru_gen_init_memcg(struct mem_cgroup *memcg) argument
5578 lru_gen_exit_memcg(struct mem_cgroup *memcg) argument
5832 struct mem_cgroup *memcg; local
6460 mem_cgroup_shrink_node(struct mem_cgroup *memcg, gfp_t gfp_mask, bool noswap, pg_data_t *pgdat, unsigned long *nr_scanned) argument
6499 try_to_free_mem_cgroup_pages(struct mem_cgroup *memcg, unsigned long nr_pages, gfp_t gfp_mask, unsigned int reclaim_options) argument
6541 struct mem_cgroup *memcg; local
[all...]
H A Doom_kill.c74 return oc->memcg != NULL;
261 oc->totalpages = mem_cgroup_get_max(oc->memcg) ?: 1;
370 mem_cgroup_scan_tasks(oc->memcg, oom_evaluate_task, oc);
420 * memcg, not in the same cpuset, or bound to a disjoint set of mempolicy nodes
431 mem_cgroup_scan_tasks(oc->memcg, dump_task, oc);
449 mem_cgroup_print_oom_context(oc->memcg, victim);
464 mem_cgroup_print_oom_meminfo(oc->memcg);
762 /* OOM killer might race with memcg OOM */
1045 oom_group = mem_cgroup_get_oom_group(victim, oc->memcg);
1071 * does not panic for cpuset, mempolicy, or memcg allocatio
[all...]
/linux-master/include/linux/
H A Dmemcontrol.h73 * Per memcg event counter is incremented at every pagein/pageout. With THP,
76 * than using jiffies etc. to handle periodic memcg event.
131 struct mem_cgroup *memcg; /* Back pointer, we cannot */ member in struct:mem_cgroup_per_node
173 int memcg_id; /* memcg->css.id of foreign inode */
186 struct mem_cgroup *memcg; member in struct:obj_cgroup
203 /* Private memcg ID. Used to ID objects that outlive the cgroup */
225 * Prevent pages from this memcg from being written back from zswap to
300 * memcg->objcg is wiped out as a part of the objcg repaprenting
301 * process. memcg->orig_objcg preserves a pointer (and a reference)
302 * to the original objcg until the end of live of memcg
531 struct mem_cgroup *memcg; local
573 mem_cgroup_is_root(struct mem_cgroup *memcg) argument
583 mem_cgroup_protection(struct mem_cgroup *root, struct mem_cgroup *memcg, unsigned long *min, unsigned long *low) argument
636 mem_cgroup_unprotected(struct mem_cgroup *target, struct mem_cgroup *memcg) argument
648 mem_cgroup_below_low(struct mem_cgroup *target, struct mem_cgroup *memcg) argument
658 mem_cgroup_below_min(struct mem_cgroup *target, struct mem_cgroup *memcg) argument
737 mem_cgroup_lruvec(struct mem_cgroup *memcg, struct pglist_data *pgdat) argument
772 struct mem_cgroup *memcg = folio_memcg(folio); local
824 mem_cgroup_tryget(struct mem_cgroup *memcg) argument
829 mem_cgroup_tryget_online(struct mem_cgroup *memcg) argument
834 mem_cgroup_put(struct mem_cgroup *memcg) argument
850 mem_cgroup_id(struct mem_cgroup *memcg) argument
860 mem_cgroup_ino(struct mem_cgroup *memcg) argument
890 parent_mem_cgroup(struct mem_cgroup *memcg) argument
895 mem_cgroup_is_descendant(struct mem_cgroup *memcg, struct mem_cgroup *root) argument
903 mm_match_cgroup(struct mm_struct *mm, struct mem_cgroup *memcg) argument
920 mem_cgroup_online(struct mem_cgroup *memcg) argument
979 mem_cgroup_trylock_pages(struct mem_cgroup *memcg) argument
996 mod_memcg_state(struct mem_cgroup *memcg, int idx, int val) argument
1009 struct mem_cgroup *memcg; local
1089 count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, unsigned long count) argument
1103 struct mem_cgroup *memcg = folio_memcg(folio); local
1112 struct mem_cgroup *memcg; local
1124 memcg_memory_event(struct mem_cgroup *memcg, enum memcg_memory_event event) argument
1152 struct mem_cgroup *memcg; local
1215 mem_cgroup_is_root(struct mem_cgroup *memcg) argument
1225 memcg_memory_event(struct mem_cgroup *memcg, enum memcg_memory_event event) argument
1235 mem_cgroup_protection(struct mem_cgroup *root, struct mem_cgroup *memcg, unsigned long *min, unsigned long *low) argument
1243 mem_cgroup_calculate_protection(struct mem_cgroup *root, struct mem_cgroup *memcg) argument
1248 mem_cgroup_unprotected(struct mem_cgroup *target, struct mem_cgroup *memcg) argument
1253 mem_cgroup_below_low(struct mem_cgroup *target, struct mem_cgroup *memcg) argument
1259 mem_cgroup_below_min(struct mem_cgroup *target, struct mem_cgroup *memcg) argument
1265 mem_cgroup_commit_charge(struct folio *folio, struct mem_cgroup *memcg) argument
1276 mem_cgroup_hugetlb_try_charge(struct mem_cgroup *memcg, gfp_t gfp, long nr_pages) argument
1300 mem_cgroup_cancel_charge(struct mem_cgroup *memcg, unsigned int nr_pages) argument
1314 mem_cgroup_lruvec(struct mem_cgroup *memcg, struct pglist_data *pgdat) argument
1331 parent_mem_cgroup(struct mem_cgroup *memcg) argument
1336 mm_match_cgroup(struct mm_struct *mm, struct mem_cgroup *memcg) argument
1362 mem_cgroup_tryget(struct mem_cgroup *memcg) argument
1367 mem_cgroup_tryget_online(struct mem_cgroup *memcg) argument
1372 mem_cgroup_put(struct mem_cgroup *memcg) argument
1414 mem_cgroup_scan_tasks(struct mem_cgroup *memcg, int (*fn)(struct task_struct *, void *), void *arg) argument
1419 mem_cgroup_id(struct mem_cgroup *memcg) argument
1432 mem_cgroup_ino(struct mem_cgroup *memcg) argument
1453 mem_cgroup_online(struct mem_cgroup *memcg) argument
1465 mem_cgroup_get_max(struct mem_cgroup *memcg) argument
1470 mem_cgroup_size(struct mem_cgroup *memcg) argument
1476 mem_cgroup_print_oom_context(struct mem_cgroup *memcg, struct task_struct *p) argument
1481 mem_cgroup_print_oom_meminfo(struct mem_cgroup *memcg) argument
1493 mem_cgroup_trylock_pages(struct mem_cgroup *memcg) argument
1533 mem_cgroup_print_oom_group(struct mem_cgroup *memcg) argument
1537 __mod_memcg_state(struct mem_cgroup *memcg, int idx, int nr) argument
1543 mod_memcg_state(struct mem_cgroup *memcg, int idx, int nr) argument
1554 memcg_page_state(struct mem_cgroup *memcg, int idx) argument
1571 mem_cgroup_flush_stats(struct mem_cgroup *memcg) argument
1575 mem_cgroup_flush_stats_ratelimited(struct mem_cgroup *memcg) argument
1600 count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, unsigned long count) argument
1606 __count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, unsigned long count) argument
1647 struct mem_cgroup *memcg; local
1723 struct mem_cgroup *memcg; local
1770 mem_cgroup_under_socket_pressure(struct mem_cgroup *memcg) argument
1789 mem_cgroup_under_socket_pressure(struct mem_cgroup *memcg) argument
1794 set_shrinker_bit(struct mem_cgroup *memcg, int nid, int shrinker_id) argument
1860 memcg_kmem_id(struct mem_cgroup *memcg) argument
1871 struct mem_cgroup *memcg; local
1923 memcg_kmem_id(struct mem_cgroup *memcg) argument
1963 mem_cgroup_zswap_writeback_enabled(struct mem_cgroup *memcg) argument
[all...]
H A Dvmpressure.h33 extern void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree,
35 extern void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio);
39 extern struct vmpressure *memcg_to_vmpressure(struct mem_cgroup *memcg);
41 extern int vmpressure_register_event(struct mem_cgroup *memcg,
44 extern void vmpressure_unregister_event(struct mem_cgroup *memcg,
47 static inline void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, argument
49 static inline void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, argument
H A Dlist_lru.h33 /* may become negative during memcg reparenting */
70 int memcg_list_lru_alloc(struct mem_cgroup *memcg, struct list_lru *lru,
72 void memcg_reparent_list_lrus(struct mem_cgroup *memcg, struct mem_cgroup *parent);
79 * @memcg: the cgroup of the sublist to add the item to.
93 struct mem_cgroup *memcg);
101 * memcg of the sublist is determined by @item list_head. This assumption is
113 * @memcg: the cgroup of the sublist to delete the item from.
122 struct mem_cgroup *memcg);
130 * memcg of the sublist is determined by @item list_head. This assumption is
141 * @memcg
[all...]
H A Dzswap.h35 void zswap_memcg_offline_cleanup(struct mem_cgroup *memcg);
59 static inline void zswap_memcg_offline_cleanup(struct mem_cgroup *memcg) {} argument
H A Doom.h37 struct mem_cgroup *memcg; member in struct:oom_control
H A Dshrinker.h13 * Bitmap and deferred work of shrinker::id corresponding to memcg-aware
14 * shrinkers, which have elements charged to the memcg.
54 /* current memcg being shrunk (for memcg aware shrinkers) */
55 struct mem_cgroup *memcg; member in struct:shrink_control
H A Dswap.h397 extern unsigned long try_to_free_mem_cgroup_pages(struct mem_cgroup *memcg,
600 static inline int mem_cgroup_swappiness(struct mem_cgroup *memcg) argument
607 if (mem_cgroup_disabled() || mem_cgroup_is_root(memcg))
610 return READ_ONCE(memcg->swappiness);
652 extern long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg);
670 static inline long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg) argument
H A Dmmzone.h252 * Per-memcg and per-lruvec counters track memory, consumed
448 /* the memcg generation this lru_gen_folio belongs to */
505 * An onlining memcg is added to the tail of a random bin in the old generation.
507 * per-node memcg generation counter, whose reminder (mod MEMCG_NR_GENS) indexes
511 * 1. MEMCG_LRU_HEAD, which moves a memcg to the head of a random bin in its
513 * 2. MEMCG_LRU_TAIL, which moves a memcg to the tail of a random bin in its
515 * 3. MEMCG_LRU_OLD, which moves a memcg to the head of a random bin in the old
517 * 4. MEMCG_LRU_YOUNG, which moves a memcg to the tail of a random bin in the
523 * 2. The first attempt to reclaim a memcg below low, which triggers
525 * 3. The first attempt to reclaim a memcg offline
580 lru_gen_init_memcg(struct mem_cgroup *memcg) argument
584 lru_gen_exit_memcg(struct mem_cgroup *memcg) argument
588 lru_gen_online_memcg(struct mem_cgroup *memcg) argument
592 lru_gen_offline_memcg(struct mem_cgroup *memcg) argument
596 lru_gen_release_memcg(struct mem_cgroup *memcg) argument
600 lru_gen_soft_reclaim(struct mem_cgroup *memcg, int nid) argument
[all...]
H A Dbacking-dev.h158 void wb_memcg_offline(struct mem_cgroup *memcg);
165 * Cgroup writeback requires support from the filesystem. Also, both memcg and
170 * depending on how memcg and iocg are configured.
186 * Find the wb of @bdi which matches both the memcg and blkcg of %current.
202 * %current's blkcg equals the effective blkcg of its memcg. No
215 * Equivalent to wb_get_create() on %current's memcg. This function is
362 static inline void wb_memcg_offline(struct mem_cgroup *memcg) argument
/linux-master/tools/testing/selftests/cgroup/
H A Dtest_memcontrol.c170 char *memcg; local
172 memcg = cg_name(root, "memcg_test");
173 if (!memcg)
176 if (cg_create(memcg))
179 current = cg_read_long(memcg, "memory.current");
183 if (cg_run(memcg, alloc_anon_50M_check, NULL))
186 if (cg_run(memcg, alloc_pagecache_50M_check, NULL))
192 cg_destroy(memcg);
193 free(memcg);
251 static bool reclaim_until(const char *memcg, lon
503 char *memcg; local
566 char *memcg; local
628 char *memcg; local
686 reclaim_until(const char *memcg, long goal) argument
720 char *memcg; local
823 char *memcg; local
887 char *memcg; local
1048 char *memcg; local
1246 char *memcg; local
[all...]
/linux-master/tools/cgroup/
H A Dmemcg_slabinfo.py42 memcg = container_of(css, 'struct mem_cgroup', 'css')
43 MEMCGS[css.cgroup.kn.id.value_()] = memcg
171 memcg = MEMCGS[cgroup_id]
186 # find memcg pointers belonging to the specified cgroup
187 obj_cgroups.add(memcg.objcg.value_())
189 memcg.objcg_list.address_of_(),
221 memcg.kmem_caches.address_of_(),
/linux-master/mm/damon/
H A Dpaddr.c191 struct mem_cgroup *memcg; local
199 memcg = folio_memcg_check(folio);
200 if (!memcg)
203 matched = filter->memcg_id == mem_cgroup_id(memcg);
/linux-master/include/linux/sched/
H A Dmm.h468 * set_active_memcg - Starts the remote memcg charging scope.
469 * @memcg: memcg to charge.
471 * This function marks the beginning of the remote memcg charging scope. All the
473 * given memcg.
475 * Please, make sure that caller has a reference to the passed memcg structure,
483 set_active_memcg(struct mem_cgroup *memcg) argument
489 this_cpu_write(int_active_memcg, memcg);
492 current->active_memcg = memcg;
499 set_active_memcg(struct mem_cgroup *memcg) argument
[all...]
/linux-master/fs/notify/
H A Dgroup.c27 mem_cgroup_put(group->memcg);
/linux-master/kernel/bpf/
H A Dmemalloc.c208 struct mem_cgroup *memcg = NULL, *old_memcg; local
239 memcg = get_memcg(c);
240 old_memcg = set_active_memcg(memcg);
253 mem_cgroup_put(memcg);
1000 struct mem_cgroup *memcg, *old_memcg; local
1002 memcg = get_memcg(c);
1003 old_memcg = set_active_memcg(memcg);
1008 mem_cgroup_put(memcg);

Completed in 307 milliseconds

12