/linux/tools/perf/util/ |
A D | thread_map.c | 58 return threads; in thread_map__new_by_pid() 71 return threads; in thread_map__new_by_tid() 130 perf_thread_map__set_pid(threads, threads->nr + i, in __thread_map__new_all_cpus() 144 return threads; in __thread_map__new_all_cpus() 147 free(threads); in __thread_map__new_all_cpus() 217 threads = nt; in thread_map__new_by_pid_str() 229 if (threads) in thread_map__new_by_pid_str() 282 if (threads) in thread_map__new_by_tid_str() 311 threads->nr, threads->nr > 1 ? "s" : ""); in thread_map__fprintf() 391 if (threads) in thread_map__new_event() [all …]
|
A D | values.c | 22 values->threads = 0; in perf_read_values_init() 54 for (i = 0; i < values->threads; i++) in perf_read_values_destroy() 93 for (i = 0; i < values->threads; i++) in perf_read_values__findnew_thread() 103 i = values->threads; in perf_read_values__findnew_thread() 112 values->threads = i + 1; in perf_read_values__findnew_thread() 134 for (i = 0; i < values->threads; i++) { in perf_read_values__enlarge_counters() 217 for (i = 0; i < values->threads; i++) { in perf_read_values__display_pretty() 238 for (i = 0; i < values->threads; i++) { in perf_read_values__display_pretty() 261 for (i = 0; i < values->threads; i++) { in perf_read_values__display_raw() 277 for (i = 0; i < values->threads; i++) { in perf_read_values__display_raw() [all …]
|
A D | machine.c | 63 struct threads *threads = &machine->threads[i]; in machine__threads_init() local 66 threads->nr = 0; in machine__threads_init() 200 struct threads *threads = &machine->threads[i]; in machine__delete_threads() local 229 struct threads *threads = &machine->threads[i]; in machine__exit() local 509 struct threads *threads, in ____machine__findnew_thread() argument 566 ++threads->nr; in ____machine__findnew_thread() 580 struct threads *threads = machine__threads(machine, tid); in machine__findnew_thread() local 592 struct threads *threads = machine__threads(machine, tid); in machine__find_thread() local 985 struct threads *threads = &machine->threads[i]; in machine__fprintf() local 2988 struct threads *threads; in machine__for_each_thread() local [all …]
|
/linux/tools/perf/tests/ |
A D | thread-map.c | 68 struct perf_thread_map *threads; in process_event() local 85 perf_thread_map__put(threads); in process_event() 91 struct perf_thread_map *threads; in test__thread_map_synthesize() local 100 thread_map__read_comms(threads); in test__thread_map_synthesize() 105 perf_thread_map__put(threads); in test__thread_map_synthesize() 111 struct perf_thread_map *threads; in test__thread_map_remove() local 121 threads); in test__thread_map_remove() 127 !thread_map__remove(threads, 0)); in test__thread_map_remove() 135 !thread_map__remove(threads, 0)); in test__thread_map_remove() 143 thread_map__remove(threads, 0)); in test__thread_map_remove() [all …]
|
A D | event-times.c | 62 struct perf_thread_map *threads; in attach__current_disabled() local 67 threads = thread_map__new(-1, getpid(), UINT_MAX); in attach__current_disabled() 68 if (threads == NULL) { in attach__current_disabled() 75 err = evsel__open_per_thread(evsel, threads); in attach__current_disabled() 81 perf_thread_map__put(threads); in attach__current_disabled() 88 struct perf_thread_map *threads; in attach__current_enabled() local 93 threads = thread_map__new(-1, getpid(), UINT_MAX); in attach__current_enabled() 94 if (threads == NULL) { in attach__current_enabled() 99 err = evsel__open_per_thread(evsel, threads); in attach__current_enabled() 101 perf_thread_map__put(threads); in attach__current_enabled()
|
/linux/tools/lib/perf/ |
A D | threadmap.c | 49 if (threads != NULL) { in perf_thread_map__new_dummy() 50 perf_thread_map__set_pid(threads, 0, -1); in perf_thread_map__new_dummy() 51 threads->nr = 1; in perf_thread_map__new_dummy() 52 refcount_set(&threads->refcnt, 1); in perf_thread_map__new_dummy() 54 return threads; in perf_thread_map__new_dummy() 59 if (threads) { in perf_thread_map__delete() 62 WARN_ONCE(refcount_read(&threads->refcnt) != 0, in perf_thread_map__delete() 64 for (i = 0; i < threads->nr; i++) in perf_thread_map__delete() 65 free(perf_thread_map__comm(threads, i)); in perf_thread_map__delete() 66 free(threads); in perf_thread_map__delete() [all …]
|
A D | evlist.c | 53 perf_thread_map__put(evsel->threads); in __perf_evlist__propagate_maps() 54 evsel->threads = perf_thread_map__get(evlist->threads); in __perf_evlist__propagate_maps() 128 perf_thread_map__put(evlist->threads); in perf_evlist__exit() 131 evlist->threads = NULL; in perf_evlist__exit() 149 struct perf_thread_map *threads) in perf_evlist__set_maps() argument 163 if (threads != evlist->threads) { in perf_evlist__set_maps() 164 perf_thread_map__put(evlist->threads); in perf_evlist__set_maps() 165 evlist->threads = perf_thread_map__get(threads); in perf_evlist__set_maps() 298 int nr_threads = perf_thread_map__nr(evlist->threads); in perf_evlist__alloc_pollfd() 569 nr_mmaps = perf_thread_map__nr(evlist->threads); in perf_evlist__nr_mmaps() [all …]
|
/linux/tools/lib/perf/tests/ |
A D | test-evsel.c | 53 struct perf_thread_map *threads; in test_stat_thread() local 61 threads = perf_thread_map__new_dummy(); in test_stat_thread() 62 __T("failed to create threads", threads); in test_stat_thread() 64 perf_thread_map__set_pid(threads, 0, 0); in test_stat_thread() 78 perf_thread_map__put(threads); in test_stat_thread() 85 struct perf_thread_map *threads; in test_stat_thread_enable() local 94 threads = perf_thread_map__new_dummy(); in test_stat_thread_enable() 120 perf_thread_map__put(threads); in test_stat_thread_enable() 127 struct perf_thread_map *threads; in test_stat_user_read() local 136 threads = perf_thread_map__new_dummy(); in test_stat_user_read() [all …]
|
A D | test-evlist.c | 91 struct perf_thread_map *threads; in test_stat_thread() local 104 threads = perf_thread_map__new_dummy(); in test_stat_thread() 107 perf_thread_map__set_pid(threads, 0, 0); in test_stat_thread() 139 perf_thread_map__put(threads); in test_stat_thread() 146 struct perf_thread_map *threads; in test_stat_thread_enable() local 161 threads = perf_thread_map__new_dummy(); in test_stat_thread_enable() 164 perf_thread_map__set_pid(threads, 0, 0); in test_stat_thread_enable() 205 perf_thread_map__put(threads); in test_stat_thread_enable() 215 struct perf_thread_map *threads; in test_mmap_thread() local 257 threads = perf_thread_map__new_dummy(); in test_mmap_thread() [all …]
|
A D | test-threadmap.c | 16 struct perf_thread_map *threads; in test_threadmap() local 22 threads = perf_thread_map__new_dummy(); in test_threadmap() 23 if (!threads) in test_threadmap() 26 perf_thread_map__get(threads); in test_threadmap() 27 perf_thread_map__put(threads); in test_threadmap() 28 perf_thread_map__put(threads); in test_threadmap()
|
/linux/tools/testing/selftests/dma/ |
A D | dma_map_benchmark.c | 37 __u32 threads; /* how many threads will do map/unmap in parallel */ member 52 int threads = 1, seconds = 20, node = -1; in main() local 64 threads = atoi(optarg); in main() 89 if (threads <= 0 || threads > DMA_MAP_MAX_THREADS) { in main() 132 map.threads = threads; in main() 145 threads, seconds, node, dir[directions], granule); in main()
|
/linux/tools/testing/selftests/powerpc/math/ |
A D | vmx_signal.c | 95 int i, j, rc, threads; in test_signal_vmx() local 102 threads = sysconf(_SC_NPROCESSORS_ONLN) * THREAD_FACTOR; in test_signal_vmx() 103 tids = malloc(threads * sizeof(pthread_t)); in test_signal_vmx() 107 threads_starting = threads; in test_signal_vmx() 108 for (i = 0; i < threads; i++) { in test_signal_vmx() 114 printf("\tWaiting for %d workers to start... %d", threads, threads_starting); in test_signal_vmx() 124 for (j = 0; j < threads; j++) { in test_signal_vmx() 133 for (i = 0; i < threads; i++) { in test_signal_vmx()
|
A D | fpu_signal.c | 77 int i, j, rc, threads; in test_signal_fpu() local 81 threads = sysconf(_SC_NPROCESSORS_ONLN) * THREAD_FACTOR; in test_signal_fpu() 82 tids = malloc(threads * sizeof(pthread_t)); in test_signal_fpu() 86 threads_starting = threads; in test_signal_fpu() 87 for (i = 0; i < threads; i++) { in test_signal_fpu() 100 for (j = 0; j < threads; j++) { in test_signal_fpu() 109 for (i = 0; i < threads; i++) { in test_signal_fpu()
|
A D | fpu_preempt.c | 57 int i, rc, threads; in test_preempt_fpu() local 60 threads = sysconf(_SC_NPROCESSORS_ONLN) * THREAD_FACTOR; in test_preempt_fpu() 61 tids = malloc((threads) * sizeof(pthread_t)); in test_preempt_fpu() 65 threads_starting = threads; in test_preempt_fpu() 66 for (i = 0; i < threads; i++) { in test_preempt_fpu() 88 for (i = 0; i < threads; i++) { in test_preempt_fpu()
|
A D | vmx_preempt.c | 57 int i, rc, threads; in test_preempt_vmx() local 63 threads = sysconf(_SC_NPROCESSORS_ONLN) * THREAD_FACTOR; in test_preempt_vmx() 64 tids = malloc(threads * sizeof(pthread_t)); in test_preempt_vmx() 68 threads_starting = threads; in test_preempt_vmx() 69 for (i = 0; i < threads; i++) { in test_preempt_vmx() 91 for (i = 0; i < threads; i++) { in test_preempt_vmx()
|
A D | vsx_preempt.c | 92 int i, rc, threads; in test_preempt_vsx() local 97 threads = sysconf(_SC_NPROCESSORS_ONLN) * THREAD_FACTOR; in test_preempt_vsx() 98 tids = malloc(threads * sizeof(pthread_t)); in test_preempt_vsx() 102 threads_starting = threads; in test_preempt_vsx() 103 for (i = 0; i < threads; i++) { in test_preempt_vsx() 125 for (i = 0; i < threads; i++) { in test_preempt_vsx()
|
/linux/Documentation/driver-api/dmaengine/ |
A D | dmatest.rst | 160 dmatest: Added 1 threads using dma0chan2 168 dmatest: Added 1 threads using dma0chan1 170 dmatest: Added 1 threads using dma0chan2 180 dmatest: Added 1 threads using dma0chan0 194 dmatest: 1 threads using dma0chan0 195 dmatest: 1 threads using dma0chan3 196 dmatest: 1 threads using dma0chan4 197 dmatest: 1 threads using dma0chan5 198 dmatest: 1 threads using dma0chan6 199 dmatest: 1 threads using dma0chan7 [all …]
|
/linux/tools/perf/bench/ |
A D | synthesize.c | 62 struct perf_thread_map *threads, in do_run_single_threaded() argument 81 target, threads, in do_run_single_threaded() 116 struct perf_thread_map *threads; in run_single_threaded() local 125 threads = thread_map__new_by_pid(getpid()); in run_single_threaded() 126 if (!threads) { in run_single_threaded() 136 err = do_run_single_threaded(session, threads, &target, false); in run_single_threaded() 140 err = do_run_single_threaded(session, threads, &target, true); in run_single_threaded() 143 if (threads) in run_single_threaded() 144 perf_thread_map__put(threads); in run_single_threaded()
|
A D | sched-pipe.c | 79 struct thread_data threads[2], *td; in bench_sched_pipe() local 102 td = threads + t; in bench_sched_pipe() 119 td = threads + t; in bench_sched_pipe() 126 td = threads + t; in bench_sched_pipe() 137 worker_thread(threads + 0); in bench_sched_pipe() 140 worker_thread(threads + 1); in bench_sched_pipe()
|
/linux/tools/lib/perf/Documentation/examples/ |
A D | counting.c | 23 struct perf_thread_map *threads; in main() local 40 threads = perf_thread_map__new_dummy(); in main() 41 if (!threads) { in main() 45 perf_thread_map__set_pid(threads, 0, 0); in main() 63 perf_evlist__set_maps(evlist, NULL, threads); in main() 81 perf_thread_map__put(threads); in main()
|
/linux/Documentation/x86/ |
A D | topology.rst | 24 threads, cores, packages, etc. 37 - threads 95 A core consists of 1 or more threads. It does not matter whether the threads 96 are SMT- or CMT-type threads. 105 The number of threads in a core. The number of threads in a package can be 116 AMDs nomenclature for CMT threads is "Compute Unit Core". The kernel always 123 The cpumask contains all online threads in the package to which a thread 130 The cpumask contains all online threads in the core to which a thread 153 threads. Many BIOSes enumerate all threads 0 first and then all threads 1. 169 b) Two threads per core:: [all …]
|
/linux/tools/testing/radix-tree/ |
A D | iteration_check.c | 15 static pthread_t threads[NUM_THREADS]; variable 178 if (pthread_create(&threads[0], NULL, tagged_iteration_fn, NULL)) { in iteration_test() 182 if (pthread_create(&threads[1], NULL, untagged_iteration_fn, NULL)) { in iteration_test() 186 if (pthread_create(&threads[2], NULL, add_entries_fn, NULL)) { in iteration_test() 190 if (pthread_create(&threads[3], NULL, remove_entries_fn, NULL)) { in iteration_test() 194 if (pthread_create(&threads[4], NULL, tag_entries_fn, NULL)) { in iteration_test() 203 if (pthread_join(threads[i], NULL)) { in iteration_test()
|
/linux/tools/testing/selftests/powerpc/tm/ |
A D | tm-vmx-unavail.c | 90 int threads; in tm_vmx_unavail_test() local 98 threads = sysconf(_SC_NPROCESSORS_ONLN) * 4; in tm_vmx_unavail_test() 99 thread = malloc(sizeof(pthread_t)*threads); in tm_vmx_unavail_test() 103 for (uint64_t i = 0; i < threads; i++) in tm_vmx_unavail_test() 106 for (uint64_t i = 0; i < threads; i++) in tm_vmx_unavail_test()
|
/linux/tools/testing/selftests/sync/ |
A D | sync_stress_consumer.c | 56 int threads; member 112 int n = test_data_mpsc.threads; in mpcs_consumer_thread() 160 pthread_t threads[n]; in test_consumer_stress_multi_producer_single_consumer() local 169 test_data_mpsc.threads = n; in test_consumer_stress_multi_producer_single_consumer() 174 pthread_create(&threads[i], NULL, (void * (*)(void *)) in test_consumer_stress_multi_producer_single_consumer() 182 pthread_join(threads[i], NULL); in test_consumer_stress_multi_producer_single_consumer()
|
/linux/kernel/dma/ |
A D | map_benchmark.c | 35 __u32 threads; /* how many threads will do map/unmap in parallel */ member 125 int threads = map->bparam.threads; in do_map_benchmark() local 132 tsk = kmalloc_array(threads, sizeof(*tsk), GFP_KERNEL); in do_map_benchmark() 138 for (i = 0; i < threads; i++) { in do_map_benchmark() 158 for (i = 0; i < threads; i++) { in do_map_benchmark() 166 for (i = 0; i < threads; i++) { in do_map_benchmark() 196 for (i = 0; i < threads; i++) in do_map_benchmark() 216 if (map->bparam.threads == 0 || in map_benchmark_ioctl() 217 map->bparam.threads > DMA_MAP_MAX_THREADS) { in map_benchmark_ioctl()
|