/linux-master/mm/ |
H A D | memcontrol.c | 24 * Per memcg lru locking 138 * memcg which the event belongs to. 140 struct mem_cgroup *memcg; member in struct:mem_cgroup_event 154 int (*register_event)(struct mem_cgroup *memcg, 161 void (*unregister_event)(struct mem_cgroup *memcg, 173 static void mem_cgroup_threshold(struct mem_cgroup *memcg); 174 static void mem_cgroup_oom_notify(struct mem_cgroup *memcg); 242 struct vmpressure *memcg_to_vmpressure(struct mem_cgroup *memcg) argument 244 if (!memcg) 245 memcg 329 memcg_reparent_objcgs(struct mem_cgroup *memcg, struct mem_cgroup *parent) argument 377 struct mem_cgroup *memcg = folio_memcg(folio); local 400 struct mem_cgroup *memcg; local 473 soft_limit_excess(struct mem_cgroup *memcg) argument 485 mem_cgroup_update_tree(struct mem_cgroup *memcg, int nid) argument 528 mem_cgroup_remove_from_trees(struct mem_cgroup *memcg) argument 714 memcg_rstat_updated(struct mem_cgroup *memcg, int val) argument 740 do_flush_stats(struct mem_cgroup *memcg) argument 757 mem_cgroup_flush_stats(struct mem_cgroup *memcg) argument 769 mem_cgroup_flush_stats_ratelimited(struct mem_cgroup *memcg) argument 786 memcg_page_state(struct mem_cgroup *memcg, int idx) argument 818 __mod_memcg_state(struct mem_cgroup *memcg, int idx, int val) argument 828 memcg_page_state_local(struct mem_cgroup *memcg, int idx) argument 843 struct mem_cgroup *memcg; local 903 struct mem_cgroup *memcg; local 925 struct mem_cgroup *memcg; local 952 __count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, unsigned long count) argument 966 memcg_events(struct mem_cgroup *memcg, int event) argument 975 memcg_events_local(struct mem_cgroup *memcg, int event) argument 985 mem_cgroup_charge_statistics(struct mem_cgroup *memcg, int nr_pages) argument 999 mem_cgroup_event_ratelimit(struct mem_cgroup *memcg, enum mem_cgroup_events_target target) argument 1028 memcg_check_events(struct mem_cgroup *memcg, int nid) argument 1081 struct mem_cgroup *memcg; local 1123 struct mem_cgroup *memcg; local 1162 struct mem_cgroup *memcg = NULL; local 1287 struct mem_cgroup *memcg = dead_memcg; local 1319 mem_cgroup_scan_tasks(struct mem_cgroup *memcg, int (*fn)(struct task_struct *, void *), void *arg) argument 1345 struct mem_cgroup *memcg; local 1474 mem_cgroup_margin(struct mem_cgroup *memcg) argument 1504 mem_cgroup_under_move(struct mem_cgroup *memcg) argument 1526 mem_cgroup_wait_acct_move(struct mem_cgroup *memcg) argument 1628 memcg_page_state_output(struct mem_cgroup *memcg, int item) argument 1635 memcg_page_state_local_output( struct mem_cgroup *memcg, int item) argument 1642 memcg_stat_format(struct mem_cgroup *memcg, struct seq_buf *s) argument 1697 memory_stat_format(struct mem_cgroup *memcg, struct seq_buf *s) argument 1715 mem_cgroup_print_oom_context(struct mem_cgroup *memcg, struct task_struct *p) argument 1736 mem_cgroup_print_oom_meminfo(struct mem_cgroup *memcg) argument 1771 mem_cgroup_get_max(struct mem_cgroup *memcg) argument 1790 mem_cgroup_size(struct mem_cgroup *memcg) argument 1795 mem_cgroup_out_of_memory(struct mem_cgroup *memcg, gfp_t gfp_mask, int order) argument 1886 mem_cgroup_oom_trylock(struct mem_cgroup *memcg) argument 1925 mem_cgroup_oom_unlock(struct mem_cgroup *memcg) argument 1936 mem_cgroup_mark_under_oom(struct mem_cgroup *memcg) argument 1946 mem_cgroup_unmark_under_oom(struct mem_cgroup *memcg) argument 1964 struct mem_cgroup *memcg; member in struct:oom_wait_info 1984 memcg_oom_recover(struct mem_cgroup *memcg) argument 2002 mem_cgroup_oom(struct mem_cgroup *memcg, gfp_t mask, int order) argument 2074 struct mem_cgroup *memcg = current->memcg_in_oom; local 2125 struct mem_cgroup *memcg; local 2168 mem_cgroup_print_oom_group(struct mem_cgroup *memcg) argument 2187 struct mem_cgroup *memcg; local 2229 __folio_memcg_unlock(struct mem_cgroup *memcg) argument 2294 memcg_account_kmem(struct mem_cgroup *memcg, int nr_pages) argument 2310 consume_stock(struct mem_cgroup *memcg, unsigned int nr_pages) argument 2380 __refill_stock(struct mem_cgroup *memcg, unsigned int nr_pages) argument 2396 refill_stock(struct mem_cgroup *memcg, unsigned int nr_pages) argument 2426 struct mem_cgroup *memcg; local 2460 reclaim_high(struct mem_cgroup *memcg, unsigned int nr_pages, gfp_t gfp_mask) argument 2488 struct mem_cgroup *memcg; local 2565 mem_find_max_overage(struct mem_cgroup *memcg) argument 2579 swap_find_max_overage(struct mem_cgroup *memcg) argument 2599 calculate_high_delay(struct mem_cgroup *memcg, unsigned int nr_pages, u64 max_overage) argument 2643 struct mem_cgroup *memcg; local 2729 try_charge_memcg(struct mem_cgroup *memcg, gfp_t gfp_mask, unsigned int nr_pages) argument 2925 try_charge(struct mem_cgroup *memcg, gfp_t gfp_mask, unsigned int nr_pages) argument 2939 mem_cgroup_cancel_charge(struct mem_cgroup *memcg, unsigned int nr_pages) argument 2949 commit_charge(struct folio *folio, struct mem_cgroup *memcg) argument 2969 mem_cgroup_commit_charge(struct folio *folio, struct mem_cgroup *memcg) argument 2997 struct mem_cgroup *memcg; local 3124 __get_obj_cgroup_from_memcg(struct mem_cgroup *memcg) argument 3139 struct mem_cgroup *memcg; local 3191 struct mem_cgroup *memcg; local 3243 struct mem_cgroup *memcg; local 3256 memcg_account_kmem(struct mem_cgroup *memcg, int nr_pages) argument 3276 struct mem_cgroup *memcg; local 3297 struct mem_cgroup *memcg; local 3456 struct mem_cgroup *memcg; local 3511 struct mem_cgroup *memcg; local 3613 struct mem_cgroup *memcg = folio_memcg(folio); local 3670 mem_cgroup_resize_max(struct mem_cgroup *memcg, unsigned long max, bool memsw) argument 3811 mem_cgroup_force_empty(struct mem_cgroup *memcg) argument 3837 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local 3863 mem_cgroup_usage(struct mem_cgroup *memcg, bool swap) argument 3896 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 3947 memcg_online_kmem(struct mem_cgroup *memcg) argument 3973 memcg_offline_kmem(struct mem_cgroup *memcg) argument 3998 memcg_online_kmem(struct mem_cgroup *memcg) argument 4002 memcg_offline_kmem(struct mem_cgroup *memcg) argument 4007 memcg_update_tcp_max(struct mem_cgroup *memcg, unsigned long max) argument 4049 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local 4098 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local 4142 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 4174 mem_cgroup_node_nr_lru_pages(struct mem_cgroup *memcg, int nid, unsigned int lru_mask, bool tree) argument 4194 mem_cgroup_nr_lru_pages(struct mem_cgroup *memcg, unsigned int lru_mask, bool tree) argument 4227 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local 4302 memcg1_stat_format(struct mem_cgroup *memcg, struct seq_buf *s) argument 4379 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 4387 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 4400 __mem_cgroup_threshold(struct mem_cgroup *memcg, bool swap) argument 4451 mem_cgroup_threshold(struct mem_cgroup *memcg) argument 4476 mem_cgroup_oom_notify_cb(struct mem_cgroup *memcg) argument 4489 mem_cgroup_oom_notify(struct mem_cgroup *memcg) argument 4497 __mem_cgroup_usage_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args, enum res_type type) argument 4577 mem_cgroup_usage_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args) argument 4583 memsw_cgroup_usage_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args) argument 4589 __mem_cgroup_usage_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, enum res_type type) argument 4674 mem_cgroup_usage_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd) argument 4680 memsw_cgroup_usage_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd) argument 4686 mem_cgroup_oom_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args) argument 4708 mem_cgroup_oom_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd) argument 4727 struct mem_cgroup *memcg = mem_cgroup_from_seq(sf); local 4739 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 4756 memcg_wb_domain_init(struct mem_cgroup *memcg, gfp_t gfp) argument 4761 memcg_wb_domain_exit(struct mem_cgroup *memcg) argument 4766 memcg_wb_domain_size_changed(struct mem_cgroup *memcg) argument 4773 struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css); local 4803 struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css); local 4871 struct mem_cgroup *memcg = folio_memcg(folio); local 4923 struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css); local 4950 memcg_wb_domain_init(struct mem_cgroup *memcg, gfp_t gfp) argument 4955 memcg_wb_domain_exit(struct mem_cgroup *memcg) argument 4959 memcg_wb_domain_size_changed(struct mem_cgroup *memcg) argument 4987 struct mem_cgroup *memcg = event->memcg; local 5011 struct mem_cgroup *memcg = event->memcg; local 5061 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 5370 mem_cgroup_id_remove(struct mem_cgroup *memcg) argument 5378 mem_cgroup_id_get_many(struct mem_cgroup *memcg, unsigned int n) argument 5384 mem_cgroup_id_put_many(struct mem_cgroup *memcg, unsigned int n) argument 5394 mem_cgroup_id_put(struct mem_cgroup *memcg) argument 5416 struct mem_cgroup *memcg; local 5434 alloc_mem_cgroup_per_node_info(struct mem_cgroup *memcg, int node) argument 5456 free_mem_cgroup_per_node_info(struct mem_cgroup *memcg, int node) argument 5467 __mem_cgroup_free(struct mem_cgroup *memcg) argument 5481 mem_cgroup_free(struct mem_cgroup *memcg) argument 5491 struct mem_cgroup *memcg; local 5566 struct mem_cgroup *memcg, *old_memcg; local 5614 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 5658 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 5690 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 5698 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 5738 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 5754 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 6358 struct mem_cgroup *memcg = NULL; /* unneeded init to make gcc happy */ local 6694 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 6702 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 6716 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local 6739 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local 6762 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local 6811 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local 6873 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local 6881 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local 6889 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local 6913 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local 6942 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local 6952 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local 6974 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local 7232 mem_cgroup_calculate_protection(struct mem_cgroup *root, struct mem_cgroup *memcg) argument 7279 charge_memcg(struct folio *folio, struct mem_cgroup *memcg, gfp_t gfp) argument 7295 struct mem_cgroup *memcg; local 7322 mem_cgroup_hugetlb_try_charge(struct mem_cgroup *memcg, gfp_t gfp, long nr_pages) argument 7355 struct mem_cgroup *memcg; local 7409 struct mem_cgroup *memcg; member in struct:uncharge_gather 7447 struct mem_cgroup *memcg; local 7541 struct mem_cgroup *memcg; local 7591 struct mem_cgroup *memcg; local 7632 struct mem_cgroup *memcg; local 7668 mem_cgroup_charge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages, gfp_t gfp_mask) argument 7699 mem_cgroup_uncharge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages) argument 7772 mem_cgroup_id_get_online(struct mem_cgroup *memcg) argument 7799 struct mem_cgroup *memcg, *swap_memcg; local 7871 struct mem_cgroup *memcg; local 7915 struct mem_cgroup *memcg; local 7934 mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg) argument 7949 struct mem_cgroup *memcg; local 7989 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 7997 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 8011 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local 8034 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local 8050 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local 8136 struct mem_cgroup *memcg, *original_memcg; local 8180 struct mem_cgroup *memcg; local 8207 struct mem_cgroup *memcg; local 8221 mem_cgroup_zswap_writeback_enabled(struct mem_cgroup *memcg) argument 8230 struct mem_cgroup *memcg = mem_cgroup_from_css(css); local 8245 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local 8261 struct mem_cgroup *memcg = mem_cgroup_from_seq(m); local 8270 struct mem_cgroup *memcg = mem_cgroup_from_css(of_css(of)); local [all...] |
H A D | shrinker_debug.c | 19 struct mem_cgroup *memcg, 30 .memcg = memcg, 51 struct mem_cgroup *memcg; local 64 memcg = mem_cgroup_iter(NULL, NULL, NULL); 66 if (memcg && !mem_cgroup_online(memcg)) 70 memcg_aware ? memcg : NULL, 73 seq_printf(m, "%lu", mem_cgroup_ino(memcg)); 80 mem_cgroup_iter_break(NULL, memcg); 18 shrinker_count_objects(struct shrinker *shrinker, struct mem_cgroup *memcg, unsigned long *count_per_node) argument 113 struct mem_cgroup *memcg = NULL; local [all...] |
H A D | shrinker.c | 62 void free_shrinker_info(struct mem_cgroup *memcg) argument 69 pn = memcg->nodeinfo[nid]; 77 int alloc_shrinker_info(struct mem_cgroup *memcg) argument 92 rcu_assign_pointer(memcg->nodeinfo[nid]->shrinker_info, info); 100 free_shrinker_info(memcg); 104 static struct shrinker_info *shrinker_info_protected(struct mem_cgroup *memcg, argument 107 return rcu_dereference_protected(memcg->nodeinfo[nid]->shrinker_info, 111 static int expand_one_shrinker_info(struct mem_cgroup *memcg, int new_size, argument 119 pn = memcg->nodeinfo[nid]; 120 old = shrinker_info_protected(memcg, ni 153 struct mem_cgroup *memcg; local 194 set_shrinker_bit(struct mem_cgroup *memcg, int nid, int shrinker_id) argument 250 xchg_nr_deferred_memcg(int nid, struct shrinker *shrinker, struct mem_cgroup *memcg) argument 266 add_nr_deferred_memcg(long nr, int nid, struct shrinker *shrinker, struct mem_cgroup *memcg) argument 283 reparent_shrinker_deferred(struct mem_cgroup *memcg) argument 321 xchg_nr_deferred_memcg(int nid, struct shrinker *shrinker, struct mem_cgroup *memcg) argument 327 add_nr_deferred_memcg(long nr, int nid, struct shrinker *shrinker, struct mem_cgroup *memcg) argument 467 shrink_slab_memcg(gfp_t gfp_mask, int nid, struct mem_cgroup *memcg, int priority) argument 585 shrink_slab_memcg(gfp_t gfp_mask, int nid, struct mem_cgroup *memcg, int priority) argument 612 shrink_slab(gfp_t gfp_mask, int nid, struct mem_cgroup *memcg, int priority) argument [all...] |
H A D | vmpressure.c | 77 struct mem_cgroup *memcg = vmpressure_to_memcg(vmpr); local 79 memcg = parent_mem_cgroup(memcg); 80 if (!memcg) 82 return memcg_to_vmpressure(memcg); 221 * @memcg: cgroup memory controller handle 231 * mode: @memcg is considered the pressure root and userspace is 234 * If @tree is not set, reclaim efficiency is recorded for @memcg, and 239 void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, argument 249 * for this @memcg rathe 335 vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio) argument 374 vmpressure_register_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd, const char *args) argument 435 vmpressure_unregister_event(struct mem_cgroup *memcg, struct eventfd_ctx *eventfd) argument [all...] |
H A D | list_lru.c | 89 struct mem_cgroup *memcg) 96 l = list_lru_from_memcg_idx(lru, nid, memcg_kmem_id(memcg)); 100 set_shrinker_bit(memcg, nid, lru_shrinker_id(lru)); 113 struct mem_cgroup *memcg = list_lru_memcg_aware(lru) ? local 116 return list_lru_add(lru, item, nid, memcg); 121 struct mem_cgroup *memcg) 128 l = list_lru_from_memcg_idx(lru, nid, memcg_kmem_id(memcg)); 143 struct mem_cgroup *memcg = list_lru_memcg_aware(lru) ? local 146 return list_lru_del(lru, item, nid, memcg); 166 int nid, struct mem_cgroup *memcg) 88 list_lru_add(struct list_lru *lru, struct list_head *item, int nid, struct mem_cgroup *memcg) argument 120 list_lru_del(struct list_lru *lru, struct list_head *item, int nid, struct mem_cgroup *memcg) argument 165 list_lru_count_one(struct list_lru *lru, int nid, struct mem_cgroup *memcg) argument 258 list_lru_walk_one(struct list_lru *lru, int nid, struct mem_cgroup *memcg, list_lru_walk_cb isolate, void *cb_arg, unsigned long *nr_to_walk) argument 274 list_lru_walk_one_irq(struct list_lru *lru, int nid, struct mem_cgroup *memcg, list_lru_walk_cb isolate, void *cb_arg, unsigned long *nr_to_walk) argument 420 memcg_reparent_list_lrus(struct mem_cgroup *memcg, struct mem_cgroup *parent) argument 454 memcg_list_lru_allocated(struct mem_cgroup *memcg, struct list_lru *lru) argument 462 memcg_list_lru_alloc(struct mem_cgroup *memcg, struct list_lru *lru, gfp_t gfp) argument 469 struct mem_cgroup *memcg; member in struct:list_lru_memcg_table [all...] |
H A D | workingset.c | 243 struct mem_cgroup *memcg = folio_memcg(folio); local 248 lruvec = mem_cgroup_lruvec(memcg, pgdat); 256 return pack_shadow(mem_cgroup_id(memcg), pgdat, token, refs); 268 struct mem_cgroup *memcg; local 273 memcg = mem_cgroup_from_id(memcg_id); 274 *lruvec = mem_cgroup_lruvec(memcg, pgdat); 444 * Look up the memcg associated with the stored ID. It might 470 * XXX: With per-memcg flushing and thresholding, is ratelimiting 527 * evicted folio in the context of the node and the memcg whose memory 534 struct mem_cgroup *memcg; local 586 struct mem_cgroup *memcg; local [all...] |
H A D | mmap_lock.c | 32 * Size of the buffer for memcg path names. Ignoring stack trace support, 188 * Write the given mm_struct's memcg path to a percpu buffer, and return a 202 struct mem_cgroup *memcg = get_mem_cgroup_from_mm(mm); local 204 if (memcg == NULL) 206 if (unlikely(memcg->css.cgroup == NULL)) 213 cgroup_path(memcg->css.cgroup, buf, MEMCG_PATH_BUF_SIZE); 216 css_put(&memcg->css);
|
H A D | zswap.c | 738 struct mem_cgroup *memcg; local 743 * concurrent memcg offlining. Thanks to the memcg->kmemcg_id indirection 746 * 1. list_lru_add() is called before memcg->kmemcg_id is updated. The 747 * new entry will be reparented to memcg's parent's list_lru. 748 * 2. list_lru_add() is called after memcg->kmemcg_id is updated. The 749 * new entry will be added directly to memcg's parent's list_lru. 754 memcg = mem_cgroup_from_entry(entry); 756 list_lru_add(list_lru, &entry->lru, nid, memcg); 759 lru_size = list_lru_count_one(list_lru, nid, memcg); 776 struct mem_cgroup *memcg; local 800 zswap_memcg_offline_cleanup(struct mem_cgroup *memcg) argument 1319 struct mem_cgroup *memcg = sc->memcg; local 1381 shrink_memcg(struct mem_cgroup *memcg) argument 1406 struct mem_cgroup *memcg; local 1498 struct mem_cgroup *memcg = NULL; local [all...] |
H A D | vmscan.c | 218 * completely broken with the legacy memcg and direct stalling in 279 * single memcg. For example, a memcg-aware shrinker can free one object 280 * charged to the target memcg, causing an entire page to be freed. 281 * If we count the entire page as reclaimed from the memcg, we end up 285 * from the target memcg; preventing unnecessary retries during memcg 289 * charged to the target memcg, we end up underestimating the reclaimed 296 * memcg reclaim, to make reporting more accurate and reduce 317 static inline bool can_reclaim_anon_pages(struct mem_cgroup *memcg, argument 389 struct mem_cgroup *memcg = NULL; local 2338 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local 2589 get_lruvec(struct mem_cgroup *memcg, int nid) argument 2611 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local 2737 get_mm_list(struct mem_cgroup *memcg) argument 2779 struct mem_cgroup *memcg = get_mem_cgroup_from_mm(mm); local 2807 struct mem_cgroup *memcg = NULL; local 2845 struct mem_cgroup *memcg; local 2874 get_mm_list(struct mem_cgroup *memcg) argument 2922 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local 2983 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local 3299 get_pfn_folio(unsigned long pfn, struct mem_cgroup *memcg, struct pglist_data *pgdat, bool can_swap) argument 3340 struct mem_cgroup *memcg = lruvec_memcg(walk->lruvec); local 3408 struct mem_cgroup *memcg = lruvec_memcg(walk->lruvec); local 3614 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local 3893 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local 3917 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local 3940 struct mem_cgroup *memcg; local 4000 struct mem_cgroup *memcg = folio_memcg(folio); local 4151 lru_gen_online_memcg(struct mem_cgroup *memcg) argument 4176 lru_gen_offline_memcg(struct mem_cgroup *memcg) argument 4187 lru_gen_release_memcg(struct mem_cgroup *memcg) argument 4213 lru_gen_soft_reclaim(struct mem_cgroup *memcg, int nid) argument 4343 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local 4520 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local 4669 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local 4756 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local 4801 struct mem_cgroup *memcg; local 5039 struct mem_cgroup *memcg; local 5174 struct mem_cgroup *memcg; local 5206 struct mem_cgroup *memcg = lruvec_memcg(v); local 5285 struct mem_cgroup *memcg = lruvec_memcg(lruvec); local 5394 struct mem_cgroup *memcg = NULL; local 5567 lru_gen_init_memcg(struct mem_cgroup *memcg) argument 5578 lru_gen_exit_memcg(struct mem_cgroup *memcg) argument 5832 struct mem_cgroup *memcg; local 6460 mem_cgroup_shrink_node(struct mem_cgroup *memcg, gfp_t gfp_mask, bool noswap, pg_data_t *pgdat, unsigned long *nr_scanned) argument 6499 try_to_free_mem_cgroup_pages(struct mem_cgroup *memcg, unsigned long nr_pages, gfp_t gfp_mask, unsigned int reclaim_options) argument 6541 struct mem_cgroup *memcg; local [all...] |
H A D | oom_kill.c | 74 return oc->memcg != NULL; 261 oc->totalpages = mem_cgroup_get_max(oc->memcg) ?: 1; 370 mem_cgroup_scan_tasks(oc->memcg, oom_evaluate_task, oc); 420 * memcg, not in the same cpuset, or bound to a disjoint set of mempolicy nodes 431 mem_cgroup_scan_tasks(oc->memcg, dump_task, oc); 449 mem_cgroup_print_oom_context(oc->memcg, victim); 464 mem_cgroup_print_oom_meminfo(oc->memcg); 762 /* OOM killer might race with memcg OOM */ 1045 oom_group = mem_cgroup_get_oom_group(victim, oc->memcg); 1071 * does not panic for cpuset, mempolicy, or memcg allocatio [all...] |
/linux-master/include/linux/ |
H A D | memcontrol.h | 73 * Per memcg event counter is incremented at every pagein/pageout. With THP, 76 * than using jiffies etc. to handle periodic memcg event. 131 struct mem_cgroup *memcg; /* Back pointer, we cannot */ member in struct:mem_cgroup_per_node 173 int memcg_id; /* memcg->css.id of foreign inode */ 186 struct mem_cgroup *memcg; member in struct:obj_cgroup 203 /* Private memcg ID. Used to ID objects that outlive the cgroup */ 225 * Prevent pages from this memcg from being written back from zswap to 300 * memcg->objcg is wiped out as a part of the objcg repaprenting 301 * process. memcg->orig_objcg preserves a pointer (and a reference) 302 * to the original objcg until the end of live of memcg 531 struct mem_cgroup *memcg; local 573 mem_cgroup_is_root(struct mem_cgroup *memcg) argument 583 mem_cgroup_protection(struct mem_cgroup *root, struct mem_cgroup *memcg, unsigned long *min, unsigned long *low) argument 636 mem_cgroup_unprotected(struct mem_cgroup *target, struct mem_cgroup *memcg) argument 648 mem_cgroup_below_low(struct mem_cgroup *target, struct mem_cgroup *memcg) argument 658 mem_cgroup_below_min(struct mem_cgroup *target, struct mem_cgroup *memcg) argument 737 mem_cgroup_lruvec(struct mem_cgroup *memcg, struct pglist_data *pgdat) argument 772 struct mem_cgroup *memcg = folio_memcg(folio); local 824 mem_cgroup_tryget(struct mem_cgroup *memcg) argument 829 mem_cgroup_tryget_online(struct mem_cgroup *memcg) argument 834 mem_cgroup_put(struct mem_cgroup *memcg) argument 850 mem_cgroup_id(struct mem_cgroup *memcg) argument 860 mem_cgroup_ino(struct mem_cgroup *memcg) argument 890 parent_mem_cgroup(struct mem_cgroup *memcg) argument 895 mem_cgroup_is_descendant(struct mem_cgroup *memcg, struct mem_cgroup *root) argument 903 mm_match_cgroup(struct mm_struct *mm, struct mem_cgroup *memcg) argument 920 mem_cgroup_online(struct mem_cgroup *memcg) argument 979 mem_cgroup_trylock_pages(struct mem_cgroup *memcg) argument 996 mod_memcg_state(struct mem_cgroup *memcg, int idx, int val) argument 1009 struct mem_cgroup *memcg; local 1089 count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, unsigned long count) argument 1103 struct mem_cgroup *memcg = folio_memcg(folio); local 1112 struct mem_cgroup *memcg; local 1124 memcg_memory_event(struct mem_cgroup *memcg, enum memcg_memory_event event) argument 1152 struct mem_cgroup *memcg; local 1215 mem_cgroup_is_root(struct mem_cgroup *memcg) argument 1225 memcg_memory_event(struct mem_cgroup *memcg, enum memcg_memory_event event) argument 1235 mem_cgroup_protection(struct mem_cgroup *root, struct mem_cgroup *memcg, unsigned long *min, unsigned long *low) argument 1243 mem_cgroup_calculate_protection(struct mem_cgroup *root, struct mem_cgroup *memcg) argument 1248 mem_cgroup_unprotected(struct mem_cgroup *target, struct mem_cgroup *memcg) argument 1253 mem_cgroup_below_low(struct mem_cgroup *target, struct mem_cgroup *memcg) argument 1259 mem_cgroup_below_min(struct mem_cgroup *target, struct mem_cgroup *memcg) argument 1265 mem_cgroup_commit_charge(struct folio *folio, struct mem_cgroup *memcg) argument 1276 mem_cgroup_hugetlb_try_charge(struct mem_cgroup *memcg, gfp_t gfp, long nr_pages) argument 1300 mem_cgroup_cancel_charge(struct mem_cgroup *memcg, unsigned int nr_pages) argument 1314 mem_cgroup_lruvec(struct mem_cgroup *memcg, struct pglist_data *pgdat) argument 1331 parent_mem_cgroup(struct mem_cgroup *memcg) argument 1336 mm_match_cgroup(struct mm_struct *mm, struct mem_cgroup *memcg) argument 1362 mem_cgroup_tryget(struct mem_cgroup *memcg) argument 1367 mem_cgroup_tryget_online(struct mem_cgroup *memcg) argument 1372 mem_cgroup_put(struct mem_cgroup *memcg) argument 1414 mem_cgroup_scan_tasks(struct mem_cgroup *memcg, int (*fn)(struct task_struct *, void *), void *arg) argument 1419 mem_cgroup_id(struct mem_cgroup *memcg) argument 1432 mem_cgroup_ino(struct mem_cgroup *memcg) argument 1453 mem_cgroup_online(struct mem_cgroup *memcg) argument 1465 mem_cgroup_get_max(struct mem_cgroup *memcg) argument 1470 mem_cgroup_size(struct mem_cgroup *memcg) argument 1476 mem_cgroup_print_oom_context(struct mem_cgroup *memcg, struct task_struct *p) argument 1481 mem_cgroup_print_oom_meminfo(struct mem_cgroup *memcg) argument 1493 mem_cgroup_trylock_pages(struct mem_cgroup *memcg) argument 1533 mem_cgroup_print_oom_group(struct mem_cgroup *memcg) argument 1537 __mod_memcg_state(struct mem_cgroup *memcg, int idx, int nr) argument 1543 mod_memcg_state(struct mem_cgroup *memcg, int idx, int nr) argument 1554 memcg_page_state(struct mem_cgroup *memcg, int idx) argument 1571 mem_cgroup_flush_stats(struct mem_cgroup *memcg) argument 1575 mem_cgroup_flush_stats_ratelimited(struct mem_cgroup *memcg) argument 1600 count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, unsigned long count) argument 1606 __count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, unsigned long count) argument 1647 struct mem_cgroup *memcg; local 1723 struct mem_cgroup *memcg; local 1770 mem_cgroup_under_socket_pressure(struct mem_cgroup *memcg) argument 1789 mem_cgroup_under_socket_pressure(struct mem_cgroup *memcg) argument 1794 set_shrinker_bit(struct mem_cgroup *memcg, int nid, int shrinker_id) argument 1860 memcg_kmem_id(struct mem_cgroup *memcg) argument 1871 struct mem_cgroup *memcg; local 1923 memcg_kmem_id(struct mem_cgroup *memcg) argument 1963 mem_cgroup_zswap_writeback_enabled(struct mem_cgroup *memcg) argument [all...] |
H A D | vmpressure.h | 33 extern void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, 35 extern void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio); 39 extern struct vmpressure *memcg_to_vmpressure(struct mem_cgroup *memcg); 41 extern int vmpressure_register_event(struct mem_cgroup *memcg, 44 extern void vmpressure_unregister_event(struct mem_cgroup *memcg, 47 static inline void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, argument 49 static inline void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, argument
|
H A D | list_lru.h | 33 /* may become negative during memcg reparenting */ 70 int memcg_list_lru_alloc(struct mem_cgroup *memcg, struct list_lru *lru, 72 void memcg_reparent_list_lrus(struct mem_cgroup *memcg, struct mem_cgroup *parent); 79 * @memcg: the cgroup of the sublist to add the item to. 93 struct mem_cgroup *memcg); 101 * memcg of the sublist is determined by @item list_head. This assumption is 113 * @memcg: the cgroup of the sublist to delete the item from. 122 struct mem_cgroup *memcg); 130 * memcg of the sublist is determined by @item list_head. This assumption is 141 * @memcg [all...] |
H A D | zswap.h | 35 void zswap_memcg_offline_cleanup(struct mem_cgroup *memcg); 59 static inline void zswap_memcg_offline_cleanup(struct mem_cgroup *memcg) {} argument
|
H A D | oom.h | 37 struct mem_cgroup *memcg; member in struct:oom_control
|
H A D | shrinker.h | 13 * Bitmap and deferred work of shrinker::id corresponding to memcg-aware 14 * shrinkers, which have elements charged to the memcg. 54 /* current memcg being shrunk (for memcg aware shrinkers) */ 55 struct mem_cgroup *memcg; member in struct:shrink_control
|
H A D | swap.h | 397 extern unsigned long try_to_free_mem_cgroup_pages(struct mem_cgroup *memcg, 600 static inline int mem_cgroup_swappiness(struct mem_cgroup *memcg) argument 607 if (mem_cgroup_disabled() || mem_cgroup_is_root(memcg)) 610 return READ_ONCE(memcg->swappiness); 652 extern long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg); 670 static inline long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg) argument
|
H A D | mmzone.h | 252 * Per-memcg and per-lruvec counters track memory, consumed 448 /* the memcg generation this lru_gen_folio belongs to */ 505 * An onlining memcg is added to the tail of a random bin in the old generation. 507 * per-node memcg generation counter, whose reminder (mod MEMCG_NR_GENS) indexes 511 * 1. MEMCG_LRU_HEAD, which moves a memcg to the head of a random bin in its 513 * 2. MEMCG_LRU_TAIL, which moves a memcg to the tail of a random bin in its 515 * 3. MEMCG_LRU_OLD, which moves a memcg to the head of a random bin in the old 517 * 4. MEMCG_LRU_YOUNG, which moves a memcg to the tail of a random bin in the 523 * 2. The first attempt to reclaim a memcg below low, which triggers 525 * 3. The first attempt to reclaim a memcg offline 580 lru_gen_init_memcg(struct mem_cgroup *memcg) argument 584 lru_gen_exit_memcg(struct mem_cgroup *memcg) argument 588 lru_gen_online_memcg(struct mem_cgroup *memcg) argument 592 lru_gen_offline_memcg(struct mem_cgroup *memcg) argument 596 lru_gen_release_memcg(struct mem_cgroup *memcg) argument 600 lru_gen_soft_reclaim(struct mem_cgroup *memcg, int nid) argument [all...] |
H A D | backing-dev.h | 158 void wb_memcg_offline(struct mem_cgroup *memcg); 165 * Cgroup writeback requires support from the filesystem. Also, both memcg and 170 * depending on how memcg and iocg are configured. 186 * Find the wb of @bdi which matches both the memcg and blkcg of %current. 202 * %current's blkcg equals the effective blkcg of its memcg. No 215 * Equivalent to wb_get_create() on %current's memcg. This function is 362 static inline void wb_memcg_offline(struct mem_cgroup *memcg) argument
|
/linux-master/tools/testing/selftests/cgroup/ |
H A D | test_memcontrol.c | 170 char *memcg; local 172 memcg = cg_name(root, "memcg_test"); 173 if (!memcg) 176 if (cg_create(memcg)) 179 current = cg_read_long(memcg, "memory.current"); 183 if (cg_run(memcg, alloc_anon_50M_check, NULL)) 186 if (cg_run(memcg, alloc_pagecache_50M_check, NULL)) 192 cg_destroy(memcg); 193 free(memcg); 251 static bool reclaim_until(const char *memcg, lon 503 char *memcg; local 566 char *memcg; local 628 char *memcg; local 686 reclaim_until(const char *memcg, long goal) argument 720 char *memcg; local 823 char *memcg; local 887 char *memcg; local 1048 char *memcg; local 1246 char *memcg; local [all...] |
/linux-master/tools/cgroup/ |
H A D | memcg_slabinfo.py | 42 memcg = container_of(css, 'struct mem_cgroup', 'css') 43 MEMCGS[css.cgroup.kn.id.value_()] = memcg 171 memcg = MEMCGS[cgroup_id] 186 # find memcg pointers belonging to the specified cgroup 187 obj_cgroups.add(memcg.objcg.value_()) 189 memcg.objcg_list.address_of_(), 221 memcg.kmem_caches.address_of_(),
|
/linux-master/mm/damon/ |
H A D | paddr.c | 191 struct mem_cgroup *memcg; local 199 memcg = folio_memcg_check(folio); 200 if (!memcg) 203 matched = filter->memcg_id == mem_cgroup_id(memcg);
|
/linux-master/include/linux/sched/ |
H A D | mm.h | 468 * set_active_memcg - Starts the remote memcg charging scope. 469 * @memcg: memcg to charge. 471 * This function marks the beginning of the remote memcg charging scope. All the 473 * given memcg. 475 * Please, make sure that caller has a reference to the passed memcg structure, 483 set_active_memcg(struct mem_cgroup *memcg) argument 489 this_cpu_write(int_active_memcg, memcg); 492 current->active_memcg = memcg; 499 set_active_memcg(struct mem_cgroup *memcg) argument [all...] |
/linux-master/fs/notify/ |
H A D | group.c | 27 mem_cgroup_put(group->memcg);
|
/linux-master/kernel/bpf/ |
H A D | memalloc.c | 208 struct mem_cgroup *memcg = NULL, *old_memcg; local 239 memcg = get_memcg(c); 240 old_memcg = set_active_memcg(memcg); 253 mem_cgroup_put(memcg); 1000 struct mem_cgroup *memcg, *old_memcg; local 1002 memcg = get_memcg(c); 1003 old_memcg = set_active_memcg(memcg); 1008 mem_cgroup_put(memcg);
|