/linux-master/lib/ |
H A D | lwq.c | 14 * dedicated threads. 105 struct task_struct *threads[8]; local 109 printk(KERN_INFO " lwq: run some threads\n"); 110 for (i = 0; i < ARRAY_SIZE(threads); i++) 111 threads[i] = kthread_run(lwq_exercise, &q, "lwq-test-%d", i); 121 /* wait for threads to exit */ 122 for (i = 0; i < ARRAY_SIZE(threads); i++) 123 if (!IS_ERR_OR_NULL(threads[i])) 124 kthread_stop(threads[i]);
|
/linux-master/tools/perf/util/ |
H A D | values.c | 19 pr_debug("failed to allocate read_values threads arrays"); 22 values->threads = 0; 54 for (i = 0; i < values->threads; i++) 84 pr_debug("failed to enlarge read_values threads arrays"); 93 for (i = 0; i < values->threads; i++) 97 if (values->threads == values->threads_max) { 103 i = values->threads; 112 values->threads = i + 1; 134 for (i = 0; i < values->threads; i++) { 217 for (i = 0; i < values->threads; [all...] |
H A D | values.h | 8 int threads; member in struct:perf_read_values
|
H A D | pfm.c | 128 struct perf_thread_map *threads) 148 ret = evsel__open(evsel, cpus, threads); 158 ret = evsel__open(evsel, cpus, threads); 194 struct perf_thread_map *threads = thread_map__new_by_tid(0); local 235 if (is_libpfm_event_supported(name, cpus, threads)) { 266 if (!is_libpfm_event_supported(name, cpus, threads)) 281 perf_thread_map__put(threads); 127 is_libpfm_event_supported(const char *name, struct perf_cpu_map *cpus, struct perf_thread_map *threads) argument
|
/linux-master/arch/powerpc/include/asm/nohash/32/ |
H A D | pgalloc.h | 5 #include <linux/threads.h>
|
/linux-master/tools/perf/scripts/python/ |
H A D | stat-cpi.py | 7 threads = [] variable 20 if (thread not in threads): 21 threads.append(thread) 55 for thread in threads: 72 # for thread in threads:
|
/linux-master/arch/microblaze/include/asm/ |
H A D | fixmap.h | 22 #include <linux/threads.h>
|
/linux-master/arch/sh/include/asm/ |
H A D | fixmap.h | 14 #include <linux/threads.h>
|
/linux-master/tools/perf/bench/ |
H A D | synthesize.c | 34 OPT_UINTEGER('m', "min-threads", &min_threads, 35 "Minimum number of threads in multithreaded bench"), 36 OPT_UINTEGER('M', "max-threads", &max_threads, 37 "Maximum number of threads in multithreaded bench"), 62 struct perf_thread_map *threads, 81 target, threads, 116 struct perf_thread_map *threads; local 125 threads = thread_map__new_by_pid(getpid()); 126 if (!threads) { 136 err = do_run_single_threaded(session, threads, 61 do_run_single_threaded(struct perf_session *session, struct perf_thread_map *threads, struct target *target, bool data_mmap) argument [all...] |
H A D | sched-pipe.c | 85 OPT_BOOLEAN('T', "threaded", &threaded, "Specify threads/process based task setup"), 125 fd = openat(cgrp->fd, "cgroup.threads", O_WRONLY); 199 struct thread_data threads[2] = {}; local 223 td = threads + t; 238 td = threads + t; 245 td = threads + t; 255 worker_thread(threads + 0); 258 worker_thread(threads + 1); 271 if (threads[0].cgroup_failed || threads[ [all...] |
/linux-master/tools/lib/perf/include/perf/ |
H A D | threadmap.h | 15 LIBPERF_API int perf_thread_map__nr(struct perf_thread_map *threads);
|
/linux-master/arch/parisc/include/asm/ |
H A D | smp.h | 17 #include <linux/threads.h> /* for NR_CPUS */
|
/linux-master/arch/x86/include/asm/ |
H A D | highmem.h | 25 #include <linux/threads.h>
|
/linux-master/arch/powerpc/kernel/ |
H A D | dbell.c | 11 #include <linux/threads.h>
|
/linux-master/tools/testing/selftests/mm/ |
H A D | migration.c | 25 pthread_t *threads; local 51 self->threads = malloc(self->nthreads * sizeof(*self->threads)); 52 ASSERT_NE(self->threads, NULL); 59 free(self->threads); 115 * between nodes whilst other threads try and access them triggering the 124 SKIP(return, "Not enough threads or NUMA nodes available"); 132 if (pthread_create(&self->threads[i], NULL, access_mem, ptr)) 137 ASSERT_EQ(pthread_cancel(self->threads[i]), 0); 150 SKIP(return, "Not enough threads o [all...] |
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | selftest_slpc.c | 484 struct slpc_thread *threads; local 487 threads = kcalloc(I915_MAX_GT, sizeof(*threads), GFP_KERNEL); 488 if (!threads) 492 threads[i].worker = kthread_create_worker(0, "igt/slpc_parallel:%d", gt->info.id); 494 if (IS_ERR(threads[i].worker)) { 495 ret = PTR_ERR(threads[i].worker); 499 threads[i].gt = gt; 500 kthread_init_work(&threads[i].work, slpc_spinner_thread); 501 kthread_queue_work(threads[ [all...] |
/linux-master/tools/perf/tests/ |
H A D | openat-syscall-all-cpus.c | 31 struct perf_thread_map *threads = thread_map__new(-1, getpid(), UINT_MAX); local 35 if (threads == NULL) { 56 if (evsel__open(evsel, cpus, threads) < 0) { 123 perf_thread_map__put(threads);
|
H A D | sigtrap.c | 28 int tids_want_signal; /* Which threads still want a signal. */ 189 static int run_test_threads(pthread_t *threads, pthread_barrier_t *barrier) argument 195 TEST_ASSERT_EQUAL("pthread_join() failed", pthread_join(threads[i], NULL), 0); 200 static int run_stress_test(int fd, pthread_t *threads, pthread_barrier_t *barrier) argument 208 ret = run_test_threads(threads, barrier); 238 pthread_t threads[NUM_THREADS]; local 271 if (pthread_create(&threads[i], NULL, test_thread, &barrier)) { 277 ret = run_stress_test(fd, threads, &barrier);
|
/linux-master/tools/testing/selftests/kvm/ |
H A D | max_guest_memory_test.c | 94 pthread_t *threads; local 97 threads = malloc(nr_vcpus * sizeof(*threads)); 98 TEST_ASSERT(threads, "Failed to allocate vCPU threads"); 111 pthread_create(&threads[i], NULL, vcpu_worker, &info[i]); 113 return threads; 171 pthread_t *threads; local 251 threads = spawn_workers(vm, vcpus, start_gpa, gpa); 285 pthread_join(threads[ [all...] |
/linux-master/tools/perf/python/ |
H A D | tracepoint.py | 20 threads = perf.thread_map(-1) 22 evlist = perf.evlist(cpus, threads)
|
/linux-master/tools/lib/perf/ |
H A D | evlist.c | 71 perf_thread_map__put(evsel->threads); 72 evsel->threads = perf_thread_map__new_dummy(); 74 perf_thread_map__put(evsel->threads); 75 evsel->threads = perf_thread_map__get(evlist->threads); 155 perf_thread_map__put(evlist->threads); 158 evlist->threads = NULL; 176 struct perf_thread_map *threads) 190 if (threads != evlist->threads) { 174 perf_evlist__set_maps(struct perf_evlist *evlist, struct perf_cpu_map *cpus, struct perf_thread_map *threads) argument [all...] |
/linux-master/tools/testing/selftests/bpf/benchs/ |
H A D | bench_local_storage_create.c | 19 static struct thread *threads; variable in typeref:struct:thread 109 threads = calloc(env.producer_cnt, sizeof(*threads)); 111 if (!threads) { 117 struct thread *t = &threads[i]; 148 struct thread *t = &threads[(long)(input)]; 175 struct thread *t = &threads[(long)(input)];
|
/linux-master/drivers/firmware/psci/ |
H A D | psci_checker.c | 285 /* Set maximum priority to preempt all other threads on this CPU. */ 370 struct task_struct **threads; local 373 threads = kmalloc_array(nb_available_cpus, sizeof(*threads), 375 if (!threads) 380 * mode, as it might interfere with the suspend threads on other CPUs. 381 * This does not prevent the suspend threads from using cpuidle (only 405 threads[nb_threads++] = thread; 416 * Wake up the suspend threads. To avoid the main thread being preempted 417 * before all the threads hav [all...] |
/linux-master/kernel/dma/ |
H A D | map_benchmark.c | 102 int threads = map->bparam.threads; local 109 tsk = kmalloc_array(threads, sizeof(*tsk), GFP_KERNEL); 115 for (i = 0; i < threads; i++) { 135 for (i = 0; i < threads; i++) { 142 /* wait for the completion of benchmark threads */ 143 for (i = 0; i < threads; i++) { 173 for (i = 0; i < threads; i++) 193 if (map->bparam.threads == 0 || 194 map->bparam.threads > DMA_MAP_MAX_THREAD [all...] |
/linux-master/arch/alpha/include/asm/ |
H A D | smp.h | 5 #include <linux/threads.h>
|