/linux-master/sound/pci/cs46xx/ |
H A D | dsp_spos.c | 555 snd_iprintf(buffer,"\n%04x %s:\n",ins->tasks[i].address,ins->tasks[i].task_name); 557 for (col = 0,j = 0;j < ins->tasks[i].size; j++,col++) { 563 val = readl(dst + (ins->tasks[i].address + j) * sizeof(u32)); 956 strcpy(ins->tasks[ins->ntask].task_name, name); 958 strcpy(ins->tasks[ins->ntask].task_name, "(NULL)"); 959 ins->tasks[ins->ntask].address = dest; 960 ins->tasks[ins->ntask].size = size; 963 ins->tasks[ins->ntask].index = ins->ntask; 964 desc = (ins->tasks [all...] |
/linux-master/drivers/net/ethernet/qlogic/qed/ |
H A D | qed_nvmetcp.c | 193 struct qed_nvmetcp_tid *tasks, 218 if (!tasks) 238 tasks->size = tid_info->tid_size; 239 tasks->num_tids_per_block = tid_info->num_tids_per_block; 240 memcpy(tasks->blocks, tid_info->blocks, 192 qed_nvmetcp_start(struct qed_dev *cdev, struct qed_nvmetcp_tid *tasks, void *event_context, nvmetcp_event_cb_t async_event_cb) argument
|
H A D | qed_fcoe.c | 800 static int qed_fcoe_start(struct qed_dev *cdev, struct qed_fcoe_tid *tasks) argument 819 if (tasks) { 825 "Failed to allocate tasks information\n"); 839 tasks->size = tid_info->tid_size; 840 tasks->num_tids_per_block = tid_info->num_tids_per_block; 841 memcpy(tasks->blocks, tid_info->blocks,
|
H A D | qed_iscsi.c | 1103 struct qed_iscsi_tid *tasks, 1125 if (!tasks) 1144 tasks->size = tid_info->tid_size; 1145 tasks->num_tids_per_block = tid_info->num_tids_per_block; 1146 memcpy(tasks->blocks, tid_info->blocks, 1102 qed_iscsi_start(struct qed_dev *cdev, struct qed_iscsi_tid *tasks, void *event_context, iscsi_event_cb_t async_event_cb) argument
|
/linux-master/include/linux/sched/ |
H A D | signal.h | 632 list_empty(&init_task.tasks) 635 list_entry_rcu((p)->tasks.next, struct task_struct, tasks)
|
/linux-master/tools/testing/selftests/bpf/ |
H A D | test_maps.c | 658 static void test_sockmap(unsigned int tasks, void *data) argument 671 pid_t pid[tasks]; 1077 /* Test tasks number of forked operations */ 1078 for (i = 0; i < tasks; i++) { 1096 for (i = 0; i < tasks; i++) { 1352 printf("Fork %u tasks to '" #FN "'\n", N); \ 1355 static void __run_parallel(unsigned int tasks, argument 1359 pid_t pid[tasks]; 1364 for (i = 0; i < tasks; i++) { 1375 for (i = 0; i < tasks; [all...] |
/linux-master/kernel/cgroup/ |
H A D | debug.c | 31 * debug_taskcount_read - return the number of tasks in a cgroup. 164 list_for_each_entry(task, &cset->tasks, cg_list) { 175 /* show # of overflowed tasks */
|
H A D | cgroup.c | 85 * objects, and the chain of tasks off each css_set. 294 * - "tasks" is removed. Everything should be at process granularity. Use 310 * - cpuset: tasks will be kept in empty cpusets when hotplug happens and 410 * A domain which has tasks and explicit threaded controllers 604 * __cgroup_task_count - count the number of tasks in a cgroup. The caller 622 * cgroup_task_count - count the number of tasks in a cgroup. 731 .tasks = LIST_HEAD_INIT(init_css_set.tasks), 758 * css_set_populated - does a css_set contain any tasks? 770 return !list_empty(&cset->tasks) || !list_empt [all...] |
/linux-master/drivers/scsi/qedi/ |
H A D | qedi.h | 322 struct qed_iscsi_tid tasks; member in struct:qedi_ctx
|
H A D | qedi_fw.c | 95 task_ctx = qedi_get_task_mem(&qedi->tasks, cmd->task_id); 273 task_ctx = qedi_get_task_mem(&qedi->tasks, cmd->task_id); 986 (struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks, 1087 (struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks, 1448 (struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks, 1564 (struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks, 1663 (struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks, 2008 (struct iscsi_task_context *)qedi_get_task_mem(&qedi->tasks,
|
H A D | qedi_main.c | 2693 rc = qedi_ops->start(qedi->cdev, &qedi->tasks, 2701 task_start = qedi_get_task_mem(&qedi->tasks, 0); 2702 task_end = qedi_get_task_mem(&qedi->tasks, MAX_TID_BLOCKS_ISCSI - 1); 2705 task_start, task_end, qedi->tasks.size);
|
/linux-master/include/linux/sunrpc/ |
H A D | sched.h | 36 struct list_head links; /* Links to related tasks */ 57 struct list_head tk_task; /* global list of tasks */ 61 * action next procedure for async tasks 93 pid_t tk_owner; /* Process id for batching tasks */ 134 #define RPC_TASK_MOVEABLE 0x0004 /* nfs4.1+ rpc tasks */ 196 struct list_head tasks[RPC_NR_PRIORITY]; /* task queue for each priority level */ member in struct:rpc_wait_queue 199 unsigned char nr; /* # tasks remaining for cookie */ 200 unsigned int qlen; /* total # tasks waiting in queue */
|
/linux-master/tools/perf/ |
H A D | builtin-sched.c | 168 struct task_desc **tasks; member in struct:perf_sched 465 sched->tasks = realloc(sched->tasks, sched->nr_tasks * sizeof(struct task_desc *)); 466 BUG_ON(!sched->tasks); 467 sched->tasks[task->nr] = task; 482 task = sched->tasks[i]; 494 task1 = sched->tasks[i]; 498 task2 = sched->tasks[j]; 663 parms->task = task = sched->tasks[i]; 687 task = sched->tasks[ [all...] |
H A D | builtin-report.c | 1009 LIST_HEAD(tasks); 1012 ret = machine__thread_list(machine, &tasks); 1016 list_sort(machine, &tasks, task_list_cmp); 1020 list_for_each_entry(task, &tasks, list) 1023 thread_list__delete(&tasks); 1318 OPT_BOOLEAN(0, "tasks", &report.tasks_mode, "Display recorded tasks"), 1319 OPT_BOOLEAN(0, "mmaps", &report.mmaps_mode, "Display recorded tasks memory maps"), 1671 pr_err("Error: --tasks and --mmaps can't be used together with --stats\n");
|
/linux-master/include/linux/ |
H A D | cgroup-defs.h | 243 * Lists running through all tasks using this cgroup group. 244 * mg_tasks lists tasks which belong to this cset but are in the 246 * css_set_lock, but, during migration, once tasks are moved to 249 struct list_head tasks; member in struct:css_set 293 * migration. mg_dst_cset is the destination cset the target tasks 391 * Number of tasks, which are counted as frozen: 432 * doesn't have any tasks. 438 * of the type in the subtree proper don't have any tasks. 471 * List of cgrp_cset_links pointing at css_sets with tasks in this 520 * for tasks); create [all...] |
/linux-master/drivers/net/ethernet/freescale/fman/ |
H A D | fman_port.c | 443 struct fman_port_rsrc tasks; member in struct:fman_port 1350 port->tasks.num = 1353 port->tasks.extra = 1461 params.num_of_tasks = (u8)port->tasks.num; 1462 params.num_of_extra_tasks = (u8)port->tasks.extra; 1552 * Gracefully disable an FM port. The port will not start new tasks after all 1553 * tasks associated with the port are terminated. 1557 * or tasks which were already began.
|
H A D | fman.c | 1569 u8 tasks = *num_of_tasks; local 1577 /* check that there are enough uncommitted tasks */ 1578 if ((fman->state->accumulated_num_of_tasks + tasks) > 1581 dev_err(fman->dev, "%s: Requested num_of_tasks and extra tasks pool for fm%d exceed total num_of_tasks.\n", 1586 fman->state->accumulated_num_of_tasks += tasks; 1591 tmp |= ((u32)((tasks - 1) << BMI_NUM_OF_TASKS_SHIFT) |
|
/linux-master/drivers/scsi/qedf/ |
H A D | qedf.h | 388 struct qed_fcoe_tid tasks; member in struct:qedf_ctx 455 * ptr = qedf_get_task_mem(&qedf->tasks, 128);
|
H A D | qedf_els.c | 124 task = qedf_get_task_mem(&qedf->tasks, xid);
|
/linux-master/tools/testing/selftests/mm/ |
H A D | charge_reserved_hugetlb.sh | 51 echo $$ >$cgroup_path/tasks
|
/linux-master/kernel/trace/ |
H A D | trace_events_user.c | 102 * updated properly for each task. As tasks are forked, we use 425 /* Ensure MM has tasks, cannot use after exit_mm() */ 426 if (refcount_read(&mm->tasks) == 0) { 541 /* Ensure MM has tasks, cannot use after exit_mm() */ 542 if (refcount_read(&mm->tasks) == 0) 680 * updated, so we use RCU to walk the current tasks that have used 717 refcount_set(&user_mm->tasks, 1); 797 /* Clone will increment the tasks, only remove if last clone */ 798 if (!refcount_dec_and_test(&mm->tasks)) 924 * be forcefully deleted from the system until all tasks usin [all...] |
/linux-master/drivers/gpu/drm/amd/pm/powerplay/smumgr/ |
H A D | smu8_smumgr.c | 334 struct SMU_Task *task = &toc->tasks[smu8_smu->toc_entry_used_count++]; 371 struct SMU_Task *task = &toc->tasks[smu8_smu->toc_entry_used_count++];
|
/linux-master/drivers/dma/ |
H A D | pl330.c | 493 struct tasklet_struct tasks; member in struct:pl330_dmac 1582 struct pl330_dmac *pl330 = from_tasklet(pl330, t, tasks); 1735 tasklet_schedule(&pl330->tasks); 1986 tasklet_setup(&pl330->tasks, pl330_dotask); 2014 tasklet_kill(&pl330->tasks);
|
/linux-master/kernel/sched/ |
H A D | fair.c | 72 * Minimal preemption granularity for CPU-bound tasks: 437 * both tasks until we find their ancestors who are siblings of common 873 * 2) from those tasks that meet 1), we select the one 1025 * Tasks are initialized with full load to be seen as heavy tasks until 1037 * With new tasks being created, their initial util_avgs are extrapolated 1046 * To solve this problem, we also cap the util_avg of successive tasks to 1072 * For !fair tasks do: 1259 * Are we enqueueing a waiting task? (for current tasks 1339 * threshold. Above this threshold, individual tasks may be contending 1341 * approximation as the number of running tasks ma 8812 struct list_head tasks; member in struct:lb_env 9061 struct list_head *tasks = &env->src_rq->cfs_tasks; local 9227 struct list_head *tasks = &env->tasks; local [all...] |
/linux-master/fs/ |
H A D | exec.c | 1151 list_replace_rcu(&leader->tasks, &tsk->tasks);
|