/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/include/linux/ |
H A D | kthread.h | 45 struct task_struct *task; local 47 task = arg; 48 task_struct_set(curthread, task); 49 if (task->should_stop == 0) 50 task->task_ret = task->task_fn(task->task_data); 51 PROC_LOCK(task->task_thread->td_proc); 52 task->should_stop = TASK_STOPPED; 53 wakeup(task); 61 struct task_struct *task; local 92 kthread_stop(struct task_struct *task) argument [all...] |
/barrelfish-2018-10-04/usr/eclipseclp/CPViz/viz/src/ie/ucc/cccc/viz/ |
H A D | CumulativeProfile.java | 35 public static int[] computeProfile(Tuple[] task,int width){
argument 37 for(int i=1;i<task.length;i++) {
38 int startMin = task[i].getField("start").getMin();
39 int startMax = task[i].getField("start").getMax();
40 int durMin = task[i].getField("dur").getMin();
41 // int durMax = task[i].getField("dur").getMax();
42 int resMin = task[i].getField("res").getMin();
43 // int resMax = task[i].getField("red").getMax();
52 public static int[] computeOuterProfile(Tuple[] task,int width){
argument 54 for(int i=1;i<task [all...] |
H A D | VisualizerCumulativeGantt.java | 41 Tuple[] task = visualState.argumentTupleArray("tasks");
45 int n = task.length;
49 setWidth(computeLastEnd(task));
51 setWidth(Math.max(computeLastEnd(task),end.getMax()));
55 drawTask(out,task[i],i);
57 int[] profile = CumulativeProfile.computeProfile(task, width());
67 public int computeLastEnd(Tuple[] task) {
argument 69 for(int i=1;i<task.length;i++) {
70 int startMax = task[i].getField("start").getMax();
71 int durMax = task[ 76 drawTask(PrintWriter out, Tuple task,int i) argument [all...] |
H A D | VisualizerCumulative.java | 41 Tuple[] task = visualState.argumentTupleArray("tasks");
45 int n = task.length;
49 int startMax = task[i].getField("start").getMax();
50 int durMax = task[i].getField("dur").getMax();
57 int[] profile = CumulativeProfile.computeProfile(task,width());
67 Tuple[] task = visualState.argumentTupleArray("tasks");
71 int n = task.length;
74 int minStart = task[1].getField("start").getMin();
76 int startMax = task[i].getField("start").getMax();
77 int startMin = task[ [all...] |
H A D | VisualizerCumulativeCost.java | 37 Tuple[] task = visualState.argumentTupleArray("tasks");
42 int n = task.length;
47 int startMax = task[i].getField("start").getMax();
48 int durMax = task[i].getField("dur").getMax();
66 int[] profile = CumulativeProfile.computeProfile(task,width());
67 int[] outerProfile = CumulativeProfile.computeOuterProfile(task,width());
98 Tuple[] task = visualState.argumentTupleArray("tasks");
102 int n = task.length;
105 int minStart = task[1].getField("start").getMin();
107 int startMax = task[ [all...] |
/barrelfish-2018-10-04/lib/phoenix/ |
H A D | taskQ.h | 47 int tq_enqueue (taskQ_t* tq, task_t *task, int lgrp, int tid); 48 int tq_enqueue_seq (taskQ_t* tq, task_t *task, int lgrp); 49 int tq_dequeue (taskQ_t* tq, task_t *task, int lgrp, int tid);
|
H A D | taskQ.c | 42 task_t task; member in struct:__anon1385 69 taskQ_t* tq, task_t* task, int lgrp, int tid); 71 taskQ_t* tq, task_t* task, int lgrp, int tid); 73 taskQ_t* tq, task_t* task, int lgrp, int tid, dequeue_fn dequeue_fn); 91 * Initialize task queue for a normal machine 152 * Destroys an initialized queue (i.e. free queue and alloc queue) in task queue 183 * Initialize a queue for a given index in the task queue 252 * Empties out a queue in the task queue by dequeuing and freeing 253 * every task. 298 /* Queue TASK at LGRP task queu 302 tq_enqueue(taskQ_t* tq, task_t *task, int lgrp, int tid) argument 331 tq_enqueue_seq(taskQ_t* tq, task_t *task, int lgrp) argument 394 tq_dequeue_normal_seq( taskQ_t* tq, task_t* task, int lgrp, int tid) argument 401 tq_dequeue_normal( taskQ_t* tq, task_t* task, int lgrp, int tid) argument 408 tq_dequeue_normal_internal( taskQ_t* tq, task_t* task, int lgrp, int tid, dequeue_fn dequeue_fn) argument 445 tq_dequeue(taskQ_t* tq, task_t *task, int lgrp, int tid) argument [all...] |
H A D | map_reduce.c | 124 task_t task; member in struct:__anon1375 134 int chunk_size; /* # of units of data for each map task. */ 142 bool oneOutputQueuePerMapTask; /* One output queue per map task? */ 143 bool oneOutputQueuePerReduceTask; /* One output queue per reduce task? */ 161 /* Array to send to reduce task. */ 163 keyval_arr_t *final_vals; /* Array to send to merge task. */ 189 TASK_TYPE_T task_type; /* Assigned task type. */ 396 /* Determine the number of threads to schedule for each type of task. */ 673 * Dequeue the latest task and run it 674 * @return true if ran a task, fals 1091 task_queued *task = NULL; local 1155 task_queued *task; local 1192 task_queued *task; local [all...] |
/barrelfish-2018-10-04/lib/bomp/backends/ |
H A D | bomp_exclusive.c | 120 struct xomp_task *task = calloc(1, sizeof(struct xomp_task)); local 121 assert(task); 122 task->arg = data; 123 task->barrier = barrier; 124 task->total_threads = nthreads; 125 task->fn = fn; 126 task->done = 1; // set the main thread to done 130 err = xomp_master_do_work(task);
|
H A D | xomp_master.c | 341 struct xomp_task *task = (struct xomp_task *) tid; local 369 xbomp_barrier_enter_no_wait(task->barrier); 371 /* if the last worker returns, free up the task data structure */ 372 task->done++; 373 if (task->done == task->total_threads) { 374 free(task); 1098 * \param task information about the task 1103 errval_t xomp_master_do_work(struct xomp_task *task) argument [all...] |
/barrelfish-2018-10-04/lib/bomp/include/ |
H A D | bomp_internal.h | 47 * \brief this struct stores thread local data such as the team / task 55 struct bomp_task *task; member in struct:bomp_thread 95 void *task = NULL; local 96 if (task) {
|
/barrelfish-2018-10-04/lib/bomp_new/include/ |
H A D | omp_icv.h | 80 * task region on the current device. 89 * the number of nested parallel regions enclosing the current task such that 90 * all of the parallel regions are enclosed by the outermost initial task region 176 * task region on the current device. 210 struct omp_icv_task *task; ///< task specific control variables member in struct:omp_icv 237 (bomp_icv_get()->task)->_var=(_val) 239 ((bomp_icv_get()->task)->_var)
|
H A D | bomp_internal.h | 156 static inline void bomp_icv_set_task(struct omp_icv_task *task) argument 159 tls->icv.task = task; 165 return tls->icv.task; 172 * \brief this struct stores thread local data such as the team / task 180 struct bomp_task *task;
|
/barrelfish-2018-10-04/usr/skb/programs/ |
H A D | globalthreadalloc.pl | 119 % task(functionaddress, clientsocket, minblocksize(sz)). 120 %task(t1, 1, minblocksize(1)). 121 %task(t2, 2, minblocksize(1)). 122 %task(t3, 3, minblocksize(1)). 123 %task(t4, 4, minblocksize(1)). 132 % data entries to define whether the task is compute or memory bound 154 findall(_, (task(Name,ClientID,_),task_status(Name,ClientID,_,_,running)), TaskList), 158 findall(taskname(Name,ClientID), ( task(Name,ClientID,_), task_running_status(Name, ClientID, running)), TaskNames), 164 Task = task(I, N) 175 get_max_parallel(task( [all...] |
H A D | db_partitions.pl | 20 member(task(taskname(987654321,987654321), CoreList), AllocatedCores), 44 member(task(taskname(987654321,987654321), CoreList), AllocatedCores),
|
/barrelfish-2018-10-04/usr/eclipseclp/icparc_solvers/ |
H A D | ic_make_overlap_bivs.pl | 38 task, if the second task is in progress at the start of the first task. 40 records the amount of resource in use at the time the first task starts. 63 /* Check the minimum resource required at a certain task start time is 100 second task. 105 Task1 = task with resource:R, 108 Task1 = task with start:S1, 109 Task2 = task with [start:S2,duration:D2], 113 Task1 = task wit [all...] |
H A D | make_overlap_bivs.pl | 38 task, if the second task is in progress at the start of the first task. 40 records the amount of resource in use at the time the first task starts. 63 /* Check the minimum resource required at a certain task start time is 100 second task. 105 Task1 = task with resource:R, 108 Task1 = task with start:S1, 109 Task2 = task with [start:S2,duration:D2], 113 Task1 = task wit [all...] |
H A D | ic_probe_search.pl | 42 binaries associated with a given task start time 48 Based on the tentative assignments, find_bottleneck finds a task 50 tentative assignment feasible. In case a bottleneck task has a 101 task to start after the second has finished. On backtracking once it 104 to force the second task to start after the first. 114 Task1 = task with start:S1, 115 Task2 = task with [start:S2,duration:D2,resource:R2], 128 a single task 183 Otherwise, for each bottleneck task, it returns the tightest 265 These binaries represent the overlaps for a particular task [all...] |
/barrelfish-2018-10-04/include/tweed/ |
H A D | tweed.h | 45 /* Main tweed task declaration */ 60 struct generic_task_desc task; \ 66 struct name ## _task_desc * task = (struct name ## _task_desc *)task_arg;\ 67 task->ret = _INLINE_FUNC_ ## arg_count (_tweed_top_, name, task); \ 80 struct generic_task_desc task; \ 86 struct name ## _task_desc * task = (struct name ## _task_desc *)task_arg;\ 87 _INLINE_FUNC_ ## arg_count (_tweed_top_, name, task); \ 108 /* move top pointer past last task */ \ 112 /* Allocate space for new task */ \ 466 struct generic_task_desc task; member in struct:main_task_desc [all...] |
/barrelfish-2018-10-04/include/xomp/ |
H A D | xomp_master.h | 80 * \param task information about the task 85 errval_t xomp_master_do_work(struct xomp_task *task);
|
/barrelfish-2018-10-04/lib/bomp_new/ |
H A D | omp.c | 43 * the nthreads-var ICV of the current task to num_threads. 72 participating in working in the task 75 if (bomp_icv_get()->task) { 85 * \brief the maximum number of threads that can be used for a new parallel task 94 * the nthreads-var ICV of the current task. This value is also an upper bound on 101 if (bomp_icv_get()->task) { 114 if (bomp_icv_get()->task) { 141 * the current task is enclosed by an active parallel region, and the parallel 142 * region is enclosed by the outermost initial task region on the device; 147 if (bomp_icv_get()->task) { [all...] |
H A D | bomp.c | 60 struct omp_icv_task *icv = bomp_icv_get()->task; 141 free(tls->icv.task); 142 tls->icv.task = NULL;
|
/barrelfish-2018-10-04/lib/tweed/ |
H A D | tweed.c | 27 /** Space used to store task for each worker (accessed as a stack) */ 49 /** Start the main task */ 146 // alloc task stack space for all workers, leave space for alignment 196 static inline tweed_task_func_t steal_task(struct generic_task_desc * task, argument 199 tweed_task_func_t func = task->f.func; 200 int success = cmpxchg128((uint64_t *)&(task->f.func), 205 task->balarm = TWEED_TASK_STOLEN; 207 tweed_task_func_t func = task->f.func; 208 task->thief = thief; 216 /** Initializes _tweed_top_ to start of this worker's task bloc [all...] |
/barrelfish-2018-10-04/lib/cxx/unwind/ |
H A D | libunwind.cpp | 105 static bool is64bit(task_t task) { argument 109 /// Create an address_space object for use in examining another task. 110 _LIBUNWIND_EXPORT unw_addr_space_t unw_create_addr_space_for_task(task_t task) { argument 112 if (is64bit(task)) { 113 unw_addr_space_x86_64 *as = new unw_addr_space_x86_64(task); 114 as->taskPort = task; 118 unw_addr_space_i386 *as = new unw_addr_space_i386(task); 119 as->taskPort = task;
|
H A D | AddressSpace.hpp | 409 OtherAddressSpace(task_t task) : fTask(task) {} argument 493 unw_addr_space_i386(task_t task) : oas(task) {} argument 501 unw_addr_space_x86_64(task_t task) : oas(task) {} argument 509 unw_addr_space_ppc(task_t task) : oas(task) {} argument
|