/linux-master/include/linux/ |
H A D | bpf_mem_alloc.h | 14 struct obj_cgroup *objcg; member in struct:bpf_mem_alloc 31 int bpf_mem_alloc_percpu_init(struct bpf_mem_alloc *ma, struct obj_cgroup *objcg);
|
H A D | memcontrol.h | 300 * memcg->objcg is wiped out as a part of the objcg repaprenting 302 * to the original objcg until the end of live of memcg. 304 struct obj_cgroup __rcu *objcg; member in struct:mem_cgroup 365 * After the initialization objcg->memcg is always pointing at 371 static inline struct mem_cgroup *obj_cgroup_memcg(struct obj_cgroup *objcg) argument 373 return READ_ONCE(objcg->memcg); 470 struct obj_cgroup *objcg; local 472 objcg = (void *)(memcg_data & ~MEMCG_DATA_FLAGS_MASK); 473 return obj_cgroup_memcg(objcg); 513 struct obj_cgroup *objcg; local 529 get_mem_cgroup_from_objcg(struct obj_cgroup *objcg) argument 803 obj_cgroup_tryget(struct obj_cgroup *objcg) argument 808 obj_cgroup_get(struct obj_cgroup *objcg) argument 813 obj_cgroup_get_many(struct obj_cgroup *objcg, unsigned long nr) argument 819 obj_cgroup_put(struct obj_cgroup *objcg) argument 1200 get_mem_cgroup_from_objcg(struct obj_cgroup *objcg) argument 1358 obj_cgroup_put(struct obj_cgroup *objcg) argument 1818 struct obj_cgroup *objcg = current_obj_cgroup(); local 1868 count_objcg_event(struct obj_cgroup *objcg, enum vm_event_item idx) argument 1938 count_objcg_event(struct obj_cgroup *objcg, enum vm_event_item idx) argument 1951 obj_cgroup_may_zswap(struct obj_cgroup *objcg) argument 1955 obj_cgroup_charge_zswap(struct obj_cgroup *objcg, size_t size) argument 1959 obj_cgroup_uncharge_zswap(struct obj_cgroup *objcg, size_t size) argument [all...] |
H A D | sched.h | 1457 struct obj_cgroup *objcg; member in struct:task_struct
|
H A D | bpf.h | 277 struct obj_cgroup *objcg; member in struct:bpf_map
|
/linux-master/kernel/bpf/ |
H A D | memalloc.c | 95 struct obj_cgroup *objcg; member in struct:bpf_mem_cache 159 if (c->objcg) 160 return get_mem_cgroup_from_objcg(c->objcg); 517 struct obj_cgroup *objcg = NULL; local 539 objcg = get_obj_cgroup_from_current(); 541 ma->objcg = objcg; 546 c->objcg = objcg; 560 objcg 581 bpf_mem_alloc_percpu_init(struct bpf_mem_alloc *ma, struct obj_cgroup *objcg) argument 599 struct obj_cgroup *objcg; local [all...] |
H A D | syscall.c | 393 * So we have to check map->objcg for being NULL each time it's 397 map->objcg = get_obj_cgroup_from_current(); 402 if (map->objcg) 403 obj_cgroup_put(map->objcg); 408 if (map->objcg) 409 return get_mem_cgroup_from_objcg(map->objcg);
|
/linux-master/mm/ |
H A D | memcontrol.c | 265 static void obj_cgroup_uncharge_pages(struct obj_cgroup *objcg, 270 struct obj_cgroup *objcg = container_of(ref, struct obj_cgroup, refcnt); local 277 * objcg->nr_charged_bytes can't have an arbitrary byte value. 281 * 1) CPU0: objcg == stock->cached_objcg 286 * objcg->nr_charged_bytes = PAGE_SIZE - 92 290 * 92 bytes are added to objcg->nr_charged_bytes 295 nr_bytes = atomic_read(&objcg->nr_charged_bytes); 300 obj_cgroup_uncharge_pages(objcg, nr_pages); 303 list_del(&objcg->list); 307 kfree_rcu(objcg, rc 312 struct obj_cgroup *objcg; local 332 struct obj_cgroup *objcg, *iter; local 2993 mod_objcg_mlstate(struct obj_cgroup *objcg, struct pglist_data *pgdat, enum node_stat_item idx, int nr) argument 3126 struct obj_cgroup *objcg = NULL; local 3140 struct obj_cgroup *old, *objcg = NULL; local 3192 struct obj_cgroup *objcg; local 3234 struct obj_cgroup *objcg; local 3273 obj_cgroup_uncharge_pages(struct obj_cgroup *objcg, unsigned int nr_pages) argument 3294 obj_cgroup_charge_pages(struct obj_cgroup *objcg, gfp_t gfp, unsigned int nr_pages) argument 3323 struct obj_cgroup *objcg; local 3347 struct obj_cgroup *objcg; local 3359 mod_objcg_state(struct obj_cgroup *objcg, struct pglist_data *pgdat, enum node_stat_item idx, int nr) argument 3425 consume_obj_stock(struct obj_cgroup *objcg, unsigned int nr_bytes) argument 3510 struct obj_cgroup *objcg = READ_ONCE(stock->cached_objcg); local 3522 refill_obj_stock(struct obj_cgroup *objcg, unsigned int nr_bytes, bool allow_uncharge) argument 3556 obj_cgroup_charge(struct obj_cgroup *objcg, gfp_t gfp, size_t size) argument 3600 obj_cgroup_uncharge(struct obj_cgroup *objcg, size_t size) argument 3949 struct obj_cgroup *objcg; local 6619 struct obj_cgroup *objcg = task->objcg; local 7448 struct obj_cgroup *objcg; local 8134 obj_cgroup_may_zswap(struct obj_cgroup *objcg) argument 8178 obj_cgroup_charge_zswap(struct obj_cgroup *objcg, size_t size) argument 8205 obj_cgroup_uncharge_zswap(struct obj_cgroup *objcg, size_t size) argument [all...] |
H A D | zswap.c | 209 * objcg - the obj_cgroup that the compressed memory is charged to 221 struct obj_cgroup *objcg; member in struct:zswap_entry 719 return entry->objcg ? obj_cgroup_memcg(entry->objcg) : NULL; 910 if (entry->objcg) { 911 obj_cgroup_uncharge_zswap(entry->objcg, entry->length); 912 obj_cgroup_put(entry->objcg); 1182 if (entry->objcg) 1183 count_objcg_event(entry->objcg, ZSWPWB); 1497 struct obj_cgroup *objcg local [all...] |
H A D | percpu.c | 1626 struct obj_cgroup *objcg; local 1631 objcg = current_obj_cgroup(); 1632 if (!objcg) 1635 if (obj_cgroup_charge(objcg, gfp, pcpu_obj_full_size(size))) 1638 *objcgp = objcg; 1642 static void pcpu_memcg_post_alloc_hook(struct obj_cgroup *objcg, argument 1646 if (!objcg) 1650 obj_cgroup_get(objcg); 1651 chunk->obj_cgroups[off >> PCPU_MIN_ALLOC_SHIFT] = objcg; 1654 mod_memcg_state(obj_cgroup_memcg(objcg), MEMCG_PERCPU_ 1664 struct obj_cgroup *objcg; local 1691 pcpu_memcg_post_alloc_hook(struct obj_cgroup *objcg, struct pcpu_chunk *chunk, int off, size_t size) argument 1723 struct obj_cgroup *objcg = NULL; local [all...] |
H A D | slub.c | 1899 * The obtained objcg pointer is safe to use within the current scope, 1903 struct obj_cgroup *objcg = current_obj_cgroup(); local 1904 if (!objcg) 1911 memcg = get_mem_cgroup_from_objcg(objcg); 1919 if (obj_cgroup_charge(objcg, flags, objects * obj_full_size(s))) 1922 *objcgp = objcg; 1945 struct obj_cgroup *objcg, 1961 obj_cgroup_uncharge(objcg, obj_full_size(s)); 1966 obj_cgroup_get(objcg); 1967 slab_objcgs(slab)[off] = objcg; 1944 __memcg_slab_post_alloc_hook(struct kmem_cache *s, struct obj_cgroup *objcg, gfp_t flags, size_t size, void **p) argument 1977 memcg_slab_post_alloc_hook(struct kmem_cache *s, struct obj_cgroup *objcg, gfp_t flags, size_t size, void **p) argument 1991 struct obj_cgroup *objcg; local 2024 memcg_slab_alloc_error_hook(struct kmem_cache *s, int objects, struct obj_cgroup *objcg) argument 2043 memcg_slab_post_alloc_hook(struct kmem_cache *s, struct obj_cgroup *objcg, gfp_t flags, size_t size, void **p) argument 2056 memcg_slab_alloc_error_hook(struct kmem_cache *s, int objects, struct obj_cgroup *objcg) argument 3758 slab_post_alloc_hook(struct kmem_cache *s, struct obj_cgroup *objcg, gfp_t flags, size_t size, void **p, bool init, unsigned int orig_size) argument 3824 struct obj_cgroup *objcg = NULL; local 4619 struct obj_cgroup *objcg = NULL; local [all...] |
H A D | slab.h | 560 void mod_objcg_state(struct obj_cgroup *objcg, struct pglist_data *pgdat,
|
/linux-master/mm/kfence/ |
H A D | kfence.h | 101 struct obj_cgroup *objcg; member in struct:kfence_metadata
|
H A D | core.c | 599 slab->memcg_data = (unsigned long)&kfence_metadata_init[i / 2 - 1].objcg | 1143 KFENCE_WARN_ON(meta->objcg);
|
/linux-master/tools/cgroup/ |
H A D | memcg_slabinfo.py | 187 obj_cgroups.add(memcg.objcg.value_())
|