/linux-master/kernel/sched/ |
H A D | fair.c | 72 * Minimal preemption granularity for CPU-bound tasks: 437 * both tasks until we find their ancestors who are siblings of common 873 * 2) from those tasks that meet 1), we select the one 1025 * Tasks are initialized with full load to be seen as heavy tasks until 1037 * With new tasks being created, their initial util_avgs are extrapolated 1046 * To solve this problem, we also cap the util_avg of successive tasks to 1072 * For !fair tasks do: 1259 * Are we enqueueing a waiting task? (for current tasks 1339 * threshold. Above this threshold, individual tasks may be contending 1341 * approximation as the number of running tasks ma 8812 struct list_head tasks; member in struct:lb_env 9061 struct list_head *tasks = &env->src_rq->cfs_tasks; local 9227 struct list_head *tasks = &env->tasks; local [all...] |
H A D | sched.h | 232 * tasks, but still be able to sleep. We need this on platforms that cannot 284 * To keep the bandwidth of -deadline tasks under control 325 * server when it runs out of tasks to run. 334 * dl_server_stop() -- start/stop the server when it has (no) tasks. 645 * leaf cfs_rqs are those that hold tasks (lowest schedulable entity in 752 * an rb-tree, ordered by tasks' deadlines, with caching 767 * Utilization of the tasks "assigned" to this runqueue (including 768 * the tasks that are in runqueue and the tasks that executed on this 780 * tasks o 939 unsigned long tasks : BITS_PER_LONG - bits_per(SCHED_CAPACITY_SCALE); member in struct:uclamp_bucket [all...] |
H A D | core.c | 148 * Number of tasks to iterate in a single balance run. 523 * [ The astute reader will observe that it is possible for two tasks on one 531 * the CPU assignment of blocked tasks isn't required to be valid. 1234 /* Deadline tasks, even if single, need the tick */ 1239 * If there are more than one RR tasks, we need the tick to affect the 1250 * If there's no RR tasks, but FIFO tasks, we can skip the tick, no 1251 * forced preemption between FIFO tasks. 1258 * If there are no DL,RR/FIFO tasks, there must only be CFS tasks lef [all...] |
/linux-master/drivers/dma/ |
H A D | pl330.c | 493 struct tasklet_struct tasks; member in struct:pl330_dmac 1582 struct pl330_dmac *pl330 = from_tasklet(pl330, t, tasks); 1735 tasklet_schedule(&pl330->tasks); 1986 tasklet_setup(&pl330->tasks, pl330_dotask); 2014 tasklet_kill(&pl330->tasks);
|
/linux-master/kernel/ |
H A D | fork.c | 307 * stacks to tasks. Drop __GFP_ACCOUNT. 2586 list_add_tail_rcu(&p->tasks, &init_task.tasks);
|
H A D | pid.c | 9 * pid-structures are backing objects for tasks sharing a given ID to chain 11 * parking tasks using given ID's on a list. 51 .tasks = { 270 INIT_HLIST_HEAD(&pid->tasks[type]); 343 hlist_add_head_rcu(&task->pid_links[type], &pid->tasks[type]); 386 struct hlist_head *head1 = &pid1->tasks[PIDTYPE_PID]; 387 struct hlist_head *head2 = &pid2->tasks[PIDTYPE_PID]; 414 first = rcu_dereference_check(hlist_first_rcu(&pid->tasks[type]),
|
H A D | exit.c | 136 list_del_rcu(&p->tasks); 1247 * Traditionally we see ptrace'd stopped tasks regardless of options.
|
/linux-master/fs/ |
H A D | exec.c | 1151 list_replace_rcu(&leader->tasks, &tsk->tasks);
|
/linux-master/kernel/trace/ |
H A D | trace_events_user.c | 102 * updated properly for each task. As tasks are forked, we use 425 /* Ensure MM has tasks, cannot use after exit_mm() */ 426 if (refcount_read(&mm->tasks) == 0) { 541 /* Ensure MM has tasks, cannot use after exit_mm() */ 542 if (refcount_read(&mm->tasks) == 0) 680 * updated, so we use RCU to walk the current tasks that have used 717 refcount_set(&user_mm->tasks, 1); 797 /* Clone will increment the tasks, only remove if last clone */ 798 if (!refcount_dec_and_test(&mm->tasks)) 924 * be forcefully deleted from the system until all tasks usin [all...] |
/linux-master/include/linux/sunrpc/ |
H A D | sched.h | 36 struct list_head links; /* Links to related tasks */ 57 struct list_head tk_task; /* global list of tasks */ 61 * action next procedure for async tasks 93 pid_t tk_owner; /* Process id for batching tasks */ 134 #define RPC_TASK_MOVEABLE 0x0004 /* nfs4.1+ rpc tasks */ 196 struct list_head tasks[RPC_NR_PRIORITY]; /* task queue for each priority level */ member in struct:rpc_wait_queue 199 unsigned char nr; /* # tasks remaining for cookie */ 200 unsigned int qlen; /* total # tasks waiting in queue */
|
/linux-master/include/linux/ |
H A D | sched.h | 783 * push tasks around a CPU where each wakeup moves to the next one. 876 struct list_head tasks; member in struct:task_struct 1004 * 'ptraced' is the list of tasks this task is using ptrace() on. 1281 * pagefault context (and for tasks being destroyed), so it can be read 1533 * to kill tasks that are not running on SMT disabled 1655 * tasks can access tsk->flags in readonly mode for example
|
H A D | pid.h | 16 * It refers to individual tasks, process groups, and sessions. While 62 /* lists of tasks that use this pid */ 63 struct hlist_head tasks[PIDTYPE_MAX]; member in struct:pid 92 return !hlist_empty(&pid->tasks[type]); 190 &(pid)->tasks[type], pid_links[type]) {
|
H A D | cgroup-defs.h | 243 * Lists running through all tasks using this cgroup group. 244 * mg_tasks lists tasks which belong to this cset but are in the 246 * css_set_lock, but, during migration, once tasks are moved to 249 struct list_head tasks; member in struct:css_set 293 * migration. mg_dst_cset is the destination cset the target tasks 391 * Number of tasks, which are counted as frozen: 432 * doesn't have any tasks. 438 * of the type in the subtree proper don't have any tasks. 471 * List of cgrp_cset_links pointing at css_sets with tasks in this 520 * for tasks); create [all...] |
/linux-master/tools/testing/selftests/mm/ |
H A D | charge_reserved_hugetlb.sh | 51 echo $$ >$cgroup_path/tasks
|
/linux-master/tools/perf/ |
H A D | builtin-sched.c | 168 struct task_desc **tasks; member in struct:perf_sched 465 sched->tasks = realloc(sched->tasks, sched->nr_tasks * sizeof(struct task_desc *)); 466 BUG_ON(!sched->tasks); 467 sched->tasks[task->nr] = task; 482 task = sched->tasks[i]; 494 task1 = sched->tasks[i]; 498 task2 = sched->tasks[j]; 663 parms->task = task = sched->tasks[i]; 687 task = sched->tasks[ [all...] |
H A D | builtin-report.c | 1009 LIST_HEAD(tasks); 1012 ret = machine__thread_list(machine, &tasks); 1016 list_sort(machine, &tasks, task_list_cmp); 1020 list_for_each_entry(task, &tasks, list) 1023 thread_list__delete(&tasks); 1318 OPT_BOOLEAN(0, "tasks", &report.tasks_mode, "Display recorded tasks"), 1319 OPT_BOOLEAN(0, "mmaps", &report.mmaps_mode, "Display recorded tasks memory maps"), 1671 pr_err("Error: --tasks and --mmaps can't be used together with --stats\n");
|
/linux-master/tools/testing/selftests/resctrl/ |
H A D | resctrlfs.c | 501 static int write_pid_to_tasks(char *tasks, pid_t pid) argument 505 fp = fopen(tasks, "w"); 507 ksft_perror("Failed to open tasks file"); 512 ksft_print_msg("Failed to write pid to tasks file\n"); 541 char tasks[1024]; local 553 sprintf(tasks, "%s/tasks", controlgroup); 554 ret = write_pid_to_tasks(tasks, bm_pid); 568 sprintf(tasks, "%s/mon_groups/%s/tasks", [all...] |
/linux-master/tools/testing/selftests/bpf/ |
H A D | test_maps.c | 658 static void test_sockmap(unsigned int tasks, void *data) argument 671 pid_t pid[tasks]; 1077 /* Test tasks number of forked operations */ 1078 for (i = 0; i < tasks; i++) { 1096 for (i = 0; i < tasks; i++) { 1352 printf("Fork %u tasks to '" #FN "'\n", N); \ 1355 static void __run_parallel(unsigned int tasks, argument 1359 pid_t pid[tasks]; 1364 for (i = 0; i < tasks; i++) { 1375 for (i = 0; i < tasks; [all...] |
/linux-master/samples/bpf/ |
H A D | map_perf_test_user.c | 94 static int pre_test_lru_hash_lookup(int tasks) argument 295 typedef int (*pre_test_func)(int tasks); 315 static int pre_test(int tasks) argument 321 int ret = pre_test_funcs[i](tasks); 346 static void run_perf_test(int tasks) argument 348 pid_t pid[tasks]; 351 assert(!pre_test(tasks)); 353 for (i = 0; i < tasks; i++) { 363 for (i = 0; i < tasks; i++) {
|
/linux-master/init/ |
H A D | init_task.c | 94 .tasks = LIST_HEAD_INIT(init_task.tasks),
|
/linux-master/include/linux/sched/ |
H A D | signal.h | 632 list_empty(&init_task.tasks) 635 list_entry_rcu((p)->tasks.next, struct task_struct, tasks)
|
/linux-master/fs/proc/ |
H A D | base.c | 1748 /* Are we allowed to snoop on the tasks file descriptors? */ 1791 /* Are we allowed to snoop on the tasks file descriptors? */ 1830 /* Default to the tasks effective ownership */ 1849 /* Make non-dumpable tasks owned by some root */ 2026 return !proc_pid(inode)->tasks[PIDTYPE_PID].first; 3465 /* Limit procfs to only ptraceable tasks */
|
/linux-master/kernel/cgroup/ |
H A D | cgroup.c | 85 * objects, and the chain of tasks off each css_set. 294 * - "tasks" is removed. Everything should be at process granularity. Use 310 * - cpuset: tasks will be kept in empty cpusets when hotplug happens and 410 * A domain which has tasks and explicit threaded controllers 604 * __cgroup_task_count - count the number of tasks in a cgroup. The caller 622 * cgroup_task_count - count the number of tasks in a cgroup. 731 .tasks = LIST_HEAD_INIT(init_css_set.tasks), 758 * css_set_populated - does a css_set contain any tasks? 770 return !list_empty(&cset->tasks) || !list_empt [all...] |
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | bpf_iter.c | 444 if (!ASSERT_NEQ(bss->tasks, 0, "no task iteration, did BPF program run?"))
|
/linux-master/drivers/gpu/drm/ |
H A D | drm_flip_work.c | 104 struct list_head tasks; local 110 INIT_LIST_HEAD(&tasks); 112 list_splice_tail(&work->commited, &tasks); 116 if (list_empty(&tasks)) 119 list_for_each_entry_safe(task, tmp, &tasks, node) {
|