/barrelfish-2018-10-04/lib/dist/ |
H A D | skb.h | 20 errval_t get_cores_skb(coreid_t **cores, int *n_cores);
|
H A D | skb.c | 26 static errval_t parse_core_id_list(char *s, coreid_t *cores, int n) argument 28 assert(cores != NULL); 38 cores[i] = strtol(tok, NULL, 10); 51 errval_t get_cores_skb(coreid_t **cores, int *n_cores) argument 53 assert(cores != NULL); 74 // parse #cores 88 *cores = (coreid_t*)malloc(sizeof(coreid_t) * *n_cores); 89 if (*cores == NULL) { 93 err = parse_core_id_list(result, *cores, *n_cores); 96 free(*cores); [all...] |
H A D | args.c | 38 static void make_core_list(coreid_t *cores, int cores_len, argument 43 assert(cores != NULL); 48 // nothing to exclude, simply copy cores to res_cores 52 if (memcpy(*res_cores, cores, cores_len*sizeof(coreid_t)) == NULL) { 72 if (!is_excluded(cores[i], exclude, exclude_len)) { 73 (*res_cores)[r_i] = cores[i]; 192 .cores = NULL, 219 // use all available cores 223 // use the cores given in the list argument 224 parse_list(optarg, &res.cores, 269 coreid_t *cores; local [all...] |
H A D | start.c | 37 debug_printf("rmw: spawning on %d cores\n", d_args->cores_len); 56 new_argv[3] = list_to_string(d_args->cores, d_args->cores_len); 71 err = spawn_program(d_args->cores[i], new_argv[0], new_argv, 87 err = nsb_master_l(d_args->cores, d_args->cores_len, name); 95 err = run_master(core, d_args->cores, d_args->cores_len, m_args, name); 117 err = run_worker(core, d_args->cores, d_args->cores_len, m_args, name);
|
H A D | barrier.c | 131 errval_t nsb_master_l(coreid_t *cores, int n_cores, char *name) argument 133 assert(cores != NULL); 140 snprintf(serv_name, name_len, "%s.%d.%s", name, cores[i], READY_SUFFIX);
|
/barrelfish-2018-10-04/include/dist/ |
H A D | start.h | 23 * worker dispatchers on the given cores (d_args->cores) and then calls 61 * @param cores a list of the cores that the workers run on. 67 errval_t run_master(coreid_t core, coreid_t *cores, int cores_len, 79 * @param cores a list of the cores that the workers run on. 85 errval_t run_worker(coreid_t core, coreid_t *cores, int cores_len,
|
H A D | args.h | 24 coreid_t *cores; /*< list of cores to run on */ member in struct:dist_args 28 int num_cores; /*< total number of cores to use */ 29 // bool all_cores; /*< run on all available cores */ 39 * creates and returns a list of cores to start the service on based on the
|
H A D | barrier.h | 30 errval_t nsb_master_l(coreid_t *cores, int n_cores, char *name);
|
/barrelfish-2018-10-04/usr/mem_serv_dist/ |
H A D | skb.h | 21 errval_t get_cores_skb(coreid_t **cores, int *n_cores);
|
H A D | args.c | 38 static void make_core_list(coreid_t *cores, int cores_len, argument 43 assert(cores != NULL); 48 // nothing to exclude, simply copy cores to res_cores 52 if (memcpy(*res_cores, cores, cores_len*sizeof(coreid_t)) == NULL) { 72 if (!is_excluded(cores[i], exclude, exclude_len)) { 73 (*res_cores)[r_i] = cores[i]; 190 .cores = NULL, 214 // use all available cores 218 // use the cores given in the list argument 219 parse_list(optarg, &res.cores, 262 coreid_t *cores; local [all...] |
H A D | args.h | 20 coreid_t *cores; member in struct:args
|
H A D | steal.h | 26 errval_t init_peers(coreid_t core, int len_cores, coreid_t *cores);
|
H A D | skb.c | 133 static errval_t parse_core_id_list(char *s, coreid_t *cores, int n) argument 135 assert(cores != NULL); 145 cores[i] = strtol(tok, NULL, 10); 158 errval_t get_cores_skb(coreid_t **cores, int *n_cores) argument 160 assert(cores != NULL); 181 // parse #cores 195 *cores = (coreid_t*)malloc(sizeof(coreid_t) * *n_cores); 196 if (*cores == NULL) { 200 err = parse_core_id_list(result, *cores, *n_cores); 203 free(*cores); [all...] |
H A D | no_steal.c | 44 errval_t init_peers(coreid_t core, int len_cores, coreid_t *cores) argument
|
H A D | mem_serv.h | 113 coreid_t *cores, 118 errval_t percore_mem_serv(coreid_t core, coreid_t *cores,
|
/barrelfish-2018-10-04/tools/harness/tests/ |
H A D | bomp_sidebyside.py | 20 def get_modules(self, build, machine, cores=0): 22 modules.add_module("bomp_sync_progress", ["core=0", str(cores)]) 27 for cores in range(2, machine.get_ncores()): 28 modules = self.get_modules(build, machine, cores)
|
H A D | channel_cost.py | 22 cores = machine.get_ncores() 25 ["core=%d" % (cores - 1), "client"])
|
/barrelfish-2018-10-04/usr/tests/spantest/ |
H A D | xcorejoin.c | 3 * \brief Test spanning of domains across cores 47 int cores = strtol(argv[1], NULL, 10) + 1; local 49 for(int i = 1; i < cores; i++) { 58 while (ndispatchers < cores) { 62 struct thread *threads[cores]; 63 for(int i = 1; i < cores; i++) { 69 for(int i = 1; i < cores; i++) {
|
H A D | spantest.c | 3 * \brief Test spanning of domains across cores 339 int cores = strtol(argv[1], NULL, 10) + 1; local 341 NPROC = cores -1; 348 for (int i = 1; i < cores; i++) { 357 while (ndispatchers < cores) { 369 for(int i = 1; i < cores; i++) {
|
/barrelfish-2018-10-04/usr/bench/net_latency/ |
H A D | main.c | 22 errval_t get_cores_skb(coreid_t **cores, int *n_cores); 78 coreid_t *cores; local 95 // Get information about available cores 96 get_cores_skb(&cores, &core_count); 107 start_run(cores[c_idx], memory, payloadsz, nocache,
|
/barrelfish-2018-10-04/lib/numa/ |
H A D | numa_internal.h | 47 coreid_t num_cores; ///< number of cores in the system 54 struct numa_core **cores; ///< cores in the system (sorted by core id) member in struct:numa_topology 63 coreid_t num_cores; ///< number of cores within the 64 struct numa_core *cores; ///< pointer to the cores array member in struct:numa_node 65 struct bitmask *coresbm; ///< bitmask for the cores 142 NUMA_WARNING("Core ID exceeds number of available cores: %"PRIuCOREID \
|
H A D | utilities.c | 51 struct numa_core *core = &node->cores[coreid]; 59 struct numa_core *core = topology->cores[coreid]; 108 "writeln(num(nodes(Nnodes),cores(Ncores)))," 117 err = skb_read_output("num(nodes(%d), cores(%d))", &node, &core); 123 " cores\n", node, core); 126 NUMA_ERROR("invalid number of cores %" PRIu32 " or nodes %" PRIu32 ".", 132 NUMA_ERROR("too large number of cores %" PRIu32 " or nodes %" PRIu32 ".", 153 topology->cores = (struct numa_core **) (cores_array + core); 154 topology->distances = (uint32_t*)(topology->cores + core); 182 topology->nodes[parsed].cores [all...] |
H A D | numa.c | 85 bitmap_set_bit(numa_all_cpus_ptr, numa_topology.cores[i]->id); 161 * \brief returns the highest ID of the present cores 163 * \returns the maximum number of cores in the system 183 return numa_topology.cores[disp_get_current_core_id()]->node->id; 242 // XXX we assume that we can schedule all cores 255 // XXX: we do not have any restrictions yet, return all cores 537 bitmap_set_bit(mask, nnode->cores[i].id); 562 return numa_topology.cores[cpu]->node->id; 567 * \brief gets the number of cores for the given numa node 569 * \param node NUMA node to get the number of cores [all...] |
/barrelfish-2018-10-04/lib/bomp/backends/ |
H A D | linux.c | 81 void backend_create_time(int cores) argument
|
/barrelfish-2018-10-04/usr/skb/programs/ |
H A D | plat_VE_A15x4_A7x4.pl | 11 % Platform with 4 A15 cores and 4 A7
|