/linux-master/tools/cgroup/ |
H A D | memcg_shrinker.py | 11 cgroups = {} 17 cgroups[ino] = path 20 return cgroups 44 cgroups = scan_cgroups("/sys/fs/cgroup/") 58 cg = cgroups[ino]
|
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | cgroup_hierarchical_stats.c | 12 * propagates the changes to the ancestor cgroups. 18 * processes in the leaf cgroups and makes sure all the counters are aggregated 54 } cgroups[] = { variable in typeref:struct:__anon3580 64 #define N_CGROUPS ARRAY_SIZE(cgroups) 119 /* sets up cgroups, returns 0 on success. */ 133 fd = create_and_get_cgroup(cgroups[i].path); 137 cgroups[i].fd = fd; 138 cgroups[i].id = get_cgroup_id(cgroups[i].path); 147 close(cgroups[ [all...] |
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | percpu_alloc_cgrp_local_storage.c | 30 e = bpf_cgrp_storage_get(&cgrp, task->cgroups->dfl_cgrp, 0, 56 e = bpf_cgrp_storage_get(&cgrp, task->cgroups->dfl_cgrp, 0, 0); 89 e = bpf_cgrp_storage_get(&cgrp, task->cgroups->dfl_cgrp, 0, 0);
|
H A D | cgrp_ls_sleepable.c | 86 /* task->cgroups is untrusted in sleepable prog outside of RCU CS */ 87 __no_rcu_lock(task->cgroups->dfl_cgrp); 119 cgrp = task->cgroups->dfl_cgrp;
|
H A D | cgrp_ls_recursion.c | 59 __on_update(task->cgroups->dfl_cgrp); 92 __on_enter(regs, id, task->cgroups->dfl_cgrp);
|
H A D | cgrp_ls_tp_btf.c | 86 __on_enter(regs, id, task->cgroups->dfl_cgrp); 124 __on_exit(regs, id, task->cgroups->dfl_cgrp);
|
H A D | rcu_read_lock.c | 33 struct css_set *cgroups; local 41 cgroups = task->cgroups; 42 if (!cgroups) 44 cgroup_id = cgroups->dfl_cgrp->kn->id;
|
H A D | profiler.inc.h | 258 struct kernfs_node* proc_kernfs = BPF_CORE_READ(task, cgroups, dfl_cgrp, kn); 269 BPF_CORE_READ(task, cgroups, subsys[i]); 630 struct kernfs_node* proc_kernfs = BPF_CORE_READ(task, cgroups, dfl_cgrp, kn);
|
/linux-master/tools/perf/util/ |
H A D | cgroup.c | 259 /* collect given cgroups only */ 299 /* collect all cgroups first and then match with the pattern */ 315 /* collect all cgroups in the cgroup_list */ 323 /* allow empty cgroups, i.e., skip */ 370 fprintf(stderr, "must define events before cgroups\n"); 378 /* allow empty cgroups, i.e., skip */ 389 /* nr_cgroups is increased een for empty cgroups */ 428 fprintf(stderr, "must define events before cgroups\n"); 567 down_write(&env->cgroups.lock); 568 cgrp = __cgroup__findnew(&env->cgroups [all...] |
H A D | cgroup.h | 20 extern int nr_cgroups; /* number of explicit cgroups defined */ 31 int evlist__expand_cgroup(struct evlist *evlist, const char *cgroups, 54 /* read all cgroups in the system and save them in the rbtree */
|
H A D | lock-contention.h | 141 struct rb_root cgroups; member in struct:lock_contention
|
H A D | bpf_lock_contention.c | 173 read_all_cgroups(&con->cgroups); 367 struct cgroup *cgrp = __cgroup__find(&con->cgroups, cgrp_id); 528 while (!RB_EMPTY_ROOT(&con->cgroups)) { 529 struct rb_node *node = rb_first(&con->cgroups); 532 rb_erase(node, &con->cgroups);
|
H A D | env.h | 125 } cgroups; member in struct:perf_env
|
/linux-master/include/linux/ |
H A D | psi.h | 63 rcu_assign_pointer(p->cgroups, to);
|
H A D | cgroup.h | 391 rcu_dereference_check((task)->cgroups, \ 398 rcu_dereference((task)->cgroups)
|
H A D | sched.h | 1234 struct css_set __rcu *cgroups; member in struct:task_struct
|
/linux-master/tools/perf/util/bpf_skel/vmlinux/ |
H A D | vmlinux.h | 106 struct css_set *cgroups; member in struct:task_struct
|
/linux-master/tools/perf/util/bpf_skel/ |
H A D | off_cpu.bpf.c | 125 return BPF_CORE_READ(t, cgroups, dfl_cgrp, kn, id); 136 cgrp = BPF_CORE_READ(t, cgroups, subsys[perf_subsys_id], cgroup);
|
H A D | bperf_cgroup.bpf.c | 97 cgrp = BPF_CORE_READ(p, cgroups, subsys[perf_subsys_id], cgroup);
|
H A D | lock_contention.bpf.c | 166 cgrp = BPF_CORE_READ(task, cgroups, subsys[perf_subsys_id], cgroup);
|
/linux-master/tools/testing/selftests/mm/ |
H A D | charge_reserved_hugetlb.sh | 506 echo Test normal case, multiple cgroups. 552 echo Test normal case with write, multiple cgroups.
|
/linux-master/kernel/sched/ |
H A D | psi.c | 926 * Set TSK_ONCPU on @next's cgroups. If @next shares any 1144 * from the outside, so we move cgroups from inside sched/. 1146 rcu_assign_pointer(task->cgroups, to); 1168 * task->cgroups = to 1182 rcu_assign_pointer(task->cgroups, to);
|
/linux-master/tools/perf/ |
H A D | builtin-lock.c | 2098 .cgroups = RB_ROOT, 2638 "Filter specific cgroups", parse_cgroup_filter),
|
/linux-master/kernel/cgroup/ |
H A D | cgroup.c | 84 * css_set_lock protects task->cgroups pointer, the list of css_set 194 * cgroups with bigger numbers are newer than those with smaller numbers. 447 /* threaded cgroups can only have threaded controllers */ 467 /* threaded cgroups can only have threaded controllers */ 725 * reference-counted, to improve performance when child cgroups 828 * populated counters of all associated cgroups accordingly. 907 * account cgroups in empty hierarchies. 982 * associated with the same set of cgroups but different csses. 1000 * different cgroups in hierarchies. As different cgroups ma [all...] |
/linux-master/kernel/bpf/ |
H A D | verifier.c | 6338 struct css_set __rcu *cgroups; local 6529 * 'cgroups' pointer is untrusted if task->cgroups dereference
|