/linux-master/include/linux/ |
H A D | vmpressure.h | 30 struct mem_cgroup; 33 extern void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, 35 extern void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio); 39 extern struct vmpressure *memcg_to_vmpressure(struct mem_cgroup *memcg); 40 extern struct mem_cgroup *vmpressure_to_memcg(struct vmpressure *vmpr); 41 extern int vmpressure_register_event(struct mem_cgroup *memcg, 44 extern void vmpressure_unregister_event(struct mem_cgroup *memcg, 47 static inline void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, 49 static inline void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg,
|
H A D | memcontrol.h | 27 struct mem_cgroup; 88 struct mem_cgroup *position; 131 struct mem_cgroup *memcg; /* Back pointer, we cannot */ 186 struct mem_cgroup *memcg; 200 struct mem_cgroup { struct 270 * mem_cgroup ? And what type of charges should we move ? 349 extern struct mem_cgroup *root_mem_cgroup; 371 static inline struct mem_cgroup *obj_cgroup_memcg(struct obj_cgroup *objcg) 386 static inline struct mem_cgroup *__folio_memcg(struct folio *folio) 394 return (struct mem_cgroup *)(memcg_dat [all...] |
H A D | list_lru.h | 16 struct mem_cgroup; 70 int memcg_list_lru_alloc(struct mem_cgroup *memcg, struct list_lru *lru, 72 void memcg_reparent_list_lrus(struct mem_cgroup *memcg, struct mem_cgroup *parent); 93 struct mem_cgroup *memcg); 122 struct mem_cgroup *memcg); 150 int nid, struct mem_cgroup *memcg); 200 int nid, struct mem_cgroup *memcg, 217 int nid, struct mem_cgroup *memcg,
|
H A D | zswap.h | 35 void zswap_memcg_offline_cleanup(struct mem_cgroup *memcg); 59 static inline void zswap_memcg_offline_cleanup(struct mem_cgroup *memcg) {}
|
H A D | oom.h | 15 struct mem_cgroup; 37 struct mem_cgroup *memcg;
|
H A D | shrinker.h | 55 struct mem_cgroup *memcg;
|
H A D | swap.h | 349 void *workingset_eviction(struct folio *folio, struct mem_cgroup *target_memcg); 397 extern unsigned long try_to_free_mem_cgroup_pages(struct mem_cgroup *memcg, 401 extern unsigned long mem_cgroup_shrink_node(struct mem_cgroup *mem, 600 static inline int mem_cgroup_swappiness(struct mem_cgroup *memcg) 613 static inline int mem_cgroup_swappiness(struct mem_cgroup *mem) 652 extern long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg); 670 static inline long mem_cgroup_get_nr_swap_pages(struct mem_cgroup *memcg)
|
H A D | mmzone.h | 559 void lru_gen_init_memcg(struct mem_cgroup *memcg); 560 void lru_gen_exit_memcg(struct mem_cgroup *memcg); 561 void lru_gen_online_memcg(struct mem_cgroup *memcg); 562 void lru_gen_offline_memcg(struct mem_cgroup *memcg); 563 void lru_gen_release_memcg(struct mem_cgroup *memcg); 564 void lru_gen_soft_reclaim(struct mem_cgroup *memcg, int nid); 580 static inline void lru_gen_init_memcg(struct mem_cgroup *memcg) 584 static inline void lru_gen_exit_memcg(struct mem_cgroup *memcg) 588 static inline void lru_gen_online_memcg(struct mem_cgroup *memcg) 592 static inline void lru_gen_offline_memcg(struct mem_cgroup *memc [all...] |
H A D | backing-dev.h | 158 void wb_memcg_offline(struct mem_cgroup *memcg); 362 static inline void wb_memcg_offline(struct mem_cgroup *memcg)
|
H A D | rmap.h | 626 struct mem_cgroup *memcg, unsigned long *vm_flags); 741 struct mem_cgroup *memcg,
|
H A D | fsnotify_backend.h | 119 struct mem_cgroup; 232 struct mem_cgroup *memcg; /* memcg to charge allocations */
|
H A D | mm_types.h | 33 struct mem_cgroup; 1005 struct mem_cgroup *memcg;
|
/linux-master/mm/ |
H A D | memcontrol.c | 82 struct mem_cgroup *root_mem_cgroup __read_mostly; 85 DEFINE_PER_CPU(struct mem_cgroup *, int_active_memcg); 140 struct mem_cgroup *memcg; 154 int (*register_event)(struct mem_cgroup *memcg, 161 void (*unregister_event)(struct mem_cgroup *memcg, 173 static void mem_cgroup_threshold(struct mem_cgroup *memcg); 174 static void mem_cgroup_oom_notify(struct mem_cgroup *memcg); 188 struct mem_cgroup *from; 189 struct mem_cgroup *to; 242 struct vmpressure *memcg_to_vmpressure(struct mem_cgroup *memc [all...] |
H A D | shrinker.c | 62 void free_shrinker_info(struct mem_cgroup *memcg) 77 int alloc_shrinker_info(struct mem_cgroup *memcg) 104 static struct shrinker_info *shrinker_info_protected(struct mem_cgroup *memcg, 111 static int expand_one_shrinker_info(struct mem_cgroup *memcg, int new_size, 153 struct mem_cgroup *memcg; 194 void set_shrinker_bit(struct mem_cgroup *memcg, int nid, int shrinker_id) 251 struct mem_cgroup *memcg) 267 struct mem_cgroup *memcg) 283 void reparent_shrinker_deferred(struct mem_cgroup *memcg) 287 struct mem_cgroup *paren [all...] |
H A D | list_lru.c | 89 struct mem_cgroup *memcg) 113 struct mem_cgroup *memcg = list_lru_memcg_aware(lru) ? 121 struct mem_cgroup *memcg) 143 struct mem_cgroup *memcg = list_lru_memcg_aware(lru) ? 166 int nid, struct mem_cgroup *memcg) 258 list_lru_walk_one(struct list_lru *lru, int nid, struct mem_cgroup *memcg, 274 list_lru_walk_one_irq(struct list_lru *lru, int nid, struct mem_cgroup *memcg, 381 int src_idx, struct mem_cgroup *dst_memcg) 410 int src_idx, struct mem_cgroup *dst_memcg) 420 void memcg_reparent_list_lrus(struct mem_cgroup *memc [all...] |
H A D | vmpressure.c | 77 struct mem_cgroup *memcg = vmpressure_to_memcg(vmpr); 239 void vmpressure(gfp_t gfp, struct mem_cgroup *memcg, bool tree, 335 void vmpressure_prio(gfp_t gfp, struct mem_cgroup *memcg, int prio) 374 int vmpressure_register_event(struct mem_cgroup *memcg, 435 void vmpressure_unregister_event(struct mem_cgroup *memcg,
|
H A D | workingset.c | 243 struct mem_cgroup *memcg = folio_memcg(folio); 268 struct mem_cgroup *memcg; 382 void *workingset_eviction(struct folio *folio, struct mem_cgroup *target_memcg) 420 struct mem_cgroup *eviction_memcg; 534 struct mem_cgroup *memcg; 586 struct mem_cgroup *memcg;
|
H A D | shrinker_debug.c | 19 struct mem_cgroup *memcg, 51 struct mem_cgroup *memcg; 113 struct mem_cgroup *memcg = NULL;
|
H A D | vmscan.c | 87 struct mem_cgroup *target_mem_cgroup; 317 static inline bool can_reclaim_anon_pages(struct mem_cgroup *memcg, 389 struct mem_cgroup *memcg = NULL; 685 bool reclaimed, struct mem_cgroup *target_memcg) 2338 struct mem_cgroup *memcg = lruvec_memcg(lruvec); 2589 static struct lruvec *get_lruvec(struct mem_cgroup *memcg, int nid) 2611 struct mem_cgroup *memcg = lruvec_memcg(lruvec); 2737 static struct lru_gen_mm_list *get_mm_list(struct mem_cgroup *memcg) 2779 struct mem_cgroup *memcg = get_mem_cgroup_from_mm(mm); 2807 struct mem_cgroup *memc [all...] |
H A D | zswap.c | 191 static struct mem_cgroup *zswap_next_shrink; 717 static inline struct mem_cgroup *mem_cgroup_from_entry(struct zswap_entry *entry) 722 static inline struct mem_cgroup *mem_cgroup_from_entry(struct zswap_entry *entry) 738 struct mem_cgroup *memcg; 776 struct mem_cgroup *memcg; 800 void zswap_memcg_offline_cleanup(struct mem_cgroup *memcg) 1319 struct mem_cgroup *memcg = sc->memcg; 1388 static int shrink_memcg(struct mem_cgroup *memcg) 1413 struct mem_cgroup *memcg; 1505 struct mem_cgroup *memc [all...] |
H A D | mmap_lock.c | 202 struct mem_cgroup *memcg = get_mem_cgroup_from_mm(mm);
|
/linux-master/include/linux/sched/ |
H A D | mm.h | 466 DECLARE_PER_CPU(struct mem_cgroup *, int_active_memcg); 482 static inline struct mem_cgroup * 483 set_active_memcg(struct mem_cgroup *memcg) 485 struct mem_cgroup *old; 498 static inline struct mem_cgroup * 499 set_active_memcg(struct mem_cgroup *memcg)
|
/linux-master/fs/notify/inotify/ |
H A D | inotify_fsnotify.c | 70 struct mem_cgroup *old_memcg;
|
/linux-master/arch/x86/kernel/cpu/sgx/ |
H A D | encl.c | 996 * When called from ksgxd, returns the mem_cgroup of a struct mm stored 998 * the mem_cgroup of the current task. 1000 static struct mem_cgroup *sgx_encl_get_mem_cgroup(struct sgx_encl *encl) 1002 struct mem_cgroup *memcg = NULL; 1007 * If called from normal task context, return the mem_cgroup 1035 * the enclave, set memcg to the current active mem_cgroup. 1036 * This will be the root mem_cgroup if there is no active 1037 * mem_cgroup. 1065 struct mem_cgroup *encl_memcg = sgx_encl_get_mem_cgroup(encl); 1066 struct mem_cgroup *memc [all...] |
/linux-master/mm/damon/ |
H A D | paddr.c | 191 struct mem_cgroup *memcg;
|