/linux/tools/lib/perf/ |
A D | cpumap.c | 51 int nr_cpus; in cpu_map__default_new() local 54 if (nr_cpus < 0) in cpu_map__default_new() 61 for (i = 0; i < nr_cpus; ++i) in cpu_map__default_new() 64 cpus->nr = nr_cpus; in cpu_map__default_new() 97 assert(j <= nr_cpus); in cpu_map__trim_new() 107 int nr_cpus = 0; in perf_cpu_map__read() local 134 tmp_cpus[nr_cpus++] = prev; in perf_cpu_map__read() 144 tmp_cpus[nr_cpus++] = cpu; in perf_cpu_map__read() 153 if (nr_cpus > 0) in perf_cpu_map__read() 181 int i, nr_cpus = 0; in perf_cpu_map__new() local [all …]
|
/linux/tools/testing/selftests/bpf/prog_tests/ |
A D | map_init.c | 10 static int nr_cpus; variable 22 pcpu_map_value_t value[nr_cpus]; in map_populate() 26 for (i = 0; i < nr_cpus; i++) in map_populate() 105 for (i = 0; i < nr_cpus; i++) { in check_values_one_cpu() 131 pcpu_map_value_t value[nr_cpus]; in test_pcpu_map_init() 171 pcpu_map_value_t value[nr_cpus]; in test_pcpu_lru_map_init() 203 nr_cpus = bpf_num_possible_cpus(); in test_map_init() 204 if (nr_cpus <= 1) { in test_map_init()
|
A D | lookup_and_delete.c | 11 static int nr_cpus; variable 29 __u64 key, value[nr_cpus]; in fill_values_percpu() 32 for (i = 0; i < nr_cpus; i++) in fill_values_percpu() 136 __u64 key, val, value[nr_cpus]; in test_lookup_and_delete_percpu_hash() 154 for (i = 0; i < nr_cpus; i++) { in test_lookup_and_delete_percpu_hash() 220 __u64 key, val, value[nr_cpus]; in test_lookup_and_delete_lru_percpu_hash() 239 for (i = 0; i < nr_cpus; i++) in test_lookup_and_delete_lru_percpu_hash() 249 for (i = 0; i < nr_cpus; i++) { in test_lookup_and_delete_lru_percpu_hash() 278 nr_cpus = bpf_num_possible_cpus(); in test_lookup_and_delete()
|
A D | perf_buffer.c | 48 int err, on_len, nr_on_cpus = 0, nr_cpus, i, j; in serial_test_perf_buffer() local 57 nr_cpus = libbpf_num_possible_cpus(); in serial_test_perf_buffer() 58 if (CHECK(nr_cpus < 0, "nr_cpus", "err %d\n", nr_cpus)) in serial_test_perf_buffer() 96 for (i = 0; i < nr_cpus; i++) { in serial_test_perf_buffer() 119 for (i = 0, j = 0; i < nr_cpus; i++) { in serial_test_perf_buffer()
|
A D | xdp_noinline.c | 8 unsigned int nr_cpus = bpf_num_possible_cpus(); in test_xdp_noinline() local 19 } stats[nr_cpus]; in test_xdp_noinline() 59 for (i = 0; i < nr_cpus; i++) { in test_xdp_noinline()
|
/linux/tools/testing/selftests/bpf/map_tests/ |
A D | array_map_batch_ops.c | 12 static int nr_cpus; variable 27 cpu_offset = i * nr_cpus; in map_batch_update() 28 for (j = 0; j < nr_cpus; j++) in map_batch_update() 48 cpu_offset = i * nr_cpus; in map_batch_verify() 49 for (j = 0; j < nr_cpus; j++) { in map_batch_verify() 96 value_size *= nr_cpus; in __test_map_lookup_and_update_batch() 163 nr_cpus = libbpf_num_possible_cpus(); in test_array_map_batch_ops() 165 CHECK(nr_cpus < 0, "nr_cpus checking", in test_array_map_batch_ops()
|
/linux/tools/testing/selftests/bpf/ |
A D | test_lru_map.c | 27 static int nr_cpus; variable 107 unsigned long long value0[nr_cpus], value1[nr_cpus]; in map_subset() 138 while (next < nr_cpus) { in sched_next_online() 161 unsigned long long key, value[nr_cpus]; in test_lru_sanity0() 334 unsigned long long key, value[nr_cpus]; in test_lru_sanity2() 507 unsigned long long key, value[nr_cpus]; in test_lru_sanity4() 518 3 * tgt_free * nr_cpus); in test_lru_sanity4() 565 unsigned long long key, value[nr_cpus]; in do_test_lru_sanity5() 887 nr_cpus = bpf_num_possible_cpus(); in main() 888 assert(nr_cpus != -1); in main() [all …]
|
/linux/samples/bpf/ |
A D | test_lru_dist.c | 32 static int nr_cpus; variable 225 if (next_to_try == nr_cpus) in sched_next_online() 228 while (next_to_try < nr_cpus) { in sched_next_online() 321 nr_cpus * lru_size); in test_parallel_lru_dist() 337 unsigned long long key, value[nr_cpus]; in test_lru_loss0() 391 unsigned long long key, value[nr_cpus]; in test_lru_loss1() 478 nr_cpus * (1000 + 200)); in test_parallel_lru_loss() 512 nr_cpus = bpf_num_possible_cpus(); in main() 513 assert(nr_cpus != -1); in main() 514 printf("nr_cpus:%d\n\n", nr_cpus); in main() [all …]
|
A D | xdp_sample_user.c | 324 nr_cpus); in alloc_record_per_cpu() 357 for (i = 0; i < nr_cpus; i++) { in map_collect_percpu() 415 arr = &values[i * nr_cpus]; in map_collect_percpu_devmap() 661 for (i = 0; i < nr_cpus; i++) { in stats_get_rx_cnt() 694 unsigned int nr_cpus) in stats_get_cpumap_enqueue() argument 748 unsigned int nr_cpus) in stats_get_cpumap_remote() argument 787 unsigned int nr_cpus) in stats_get_cpumap_kthread() argument 825 unsigned int nr_cpus, in stats_get_redirect_cnt() argument 906 unsigned int nr_cpus, in stats_get_exception_cnt() argument 951 unsigned int nr_cpus, in stats_get_devmap_xmit() argument [all …]
|
A D | tracex3_user.c | 20 unsigned int nr_cpus = bpf_num_possible_cpus(); in clear_stats() local 21 __u64 values[nr_cpus]; in clear_stats() 77 unsigned int nr_cpus = bpf_num_possible_cpus(); in print_hist() local 79 long values[nr_cpus]; in print_hist() 89 for (i = 0; i < nr_cpus; i++) in print_hist()
|
A D | sampleip_user.c | 27 static int nr_cpus; variable 49 for (i = 0; i < nr_cpus; i++) { in sampling_start() 72 for (i = 0; i < nr_cpus; i++) in sampling_end() 173 nr_cpus = sysconf(_SC_NPROCESSORS_ONLN); in main() 174 links = calloc(nr_cpus, sizeof(struct bpf_link *)); in main()
|
A D | xdp_rxq_info_user.c | 199 unsigned int nr_cpus = bpf_num_possible_cpus(); in alloc_record_per_cpu() local 202 array = calloc(nr_cpus, sizeof(struct datarec)); in alloc_record_per_cpu() 204 fprintf(stderr, "Mem alloc error (nr_cpus:%u)\n", nr_cpus); in alloc_record_per_cpu() 258 unsigned int nr_cpus = bpf_num_possible_cpus(); in map_collect_percpu() local 259 struct datarec values[nr_cpus]; in map_collect_percpu() 273 for (i = 0; i < nr_cpus; i++) { in map_collect_percpu() 339 unsigned int nr_cpus = bpf_num_possible_cpus(); in stats_print() local 362 for (i = 0; i < nr_cpus; i++) { in stats_print() 396 for (i = 0; i < nr_cpus; i++) { in stats_print()
|
A D | lwt_len_hist_user.c | 31 unsigned int nr_cpus = bpf_num_possible_cpus(); in main() local 33 uint64_t values[nr_cpus], sum, max_value = 0, data[MAX_INDEX] = {}; in main() 54 for (i = 0; i < nr_cpus; i++) in main()
|
A D | tracex6_user.c | 77 int i, status, nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in test_perf_event_array() local 78 pid_t pid[nr_cpus]; in test_perf_event_array() 83 for (i = 0; i < nr_cpus; i++) { in test_perf_event_array() 92 for (i = 0; i < nr_cpus; i++) { in test_perf_event_array()
|
/linux/tools/testing/selftests/kvm/lib/aarch64/ |
A D | gic.c | 26 gic_dist_init(enum gic_type type, unsigned int nr_cpus, void *dist_base) in gic_dist_init() argument 43 gic_ops->gic_init(nr_cpus, dist_base); in gic_dist_init() 52 void gic_init(enum gic_type type, unsigned int nr_cpus, in gic_init() argument 60 GUEST_ASSERT(nr_cpus); in gic_init() 62 gic_dist_init(type, nr_cpus, dist_base); in gic_init()
|
A D | gic_v3.c | 18 unsigned int nr_cpus; member 94 GUEST_ASSERT(cpu < gicv3_data.nr_cpus); in gicv3_config_irq() 146 GUEST_ASSERT(cpu < gicv3_data.nr_cpus); in gicv3_cpu_init() 214 static void gicv3_init(unsigned int nr_cpus, void *dist_base) in gicv3_init() argument 216 GUEST_ASSERT(nr_cpus <= GICV3_MAX_CPUS); in gicv3_init() 218 gicv3_data.nr_cpus = nr_cpus; in gicv3_init()
|
/linux/drivers/powercap/ |
A D | dtpm_cpu.c | 48 int i, nr_cpus; in set_pd_power_limit() local 51 nr_cpus = cpumask_weight(&cpus); in set_pd_power_limit() 55 power = pd->table[i].power * MICROWATT_PER_MILLIWATT * nr_cpus; in set_pd_power_limit() 66 MICROWATT_PER_MILLIWATT * nr_cpus; in set_pd_power_limit() 134 int nr_cpus; in update_pd_power_uw() local 137 nr_cpus = cpumask_weight(&cpus); in update_pd_power_uw() 141 dtpm->power_min *= nr_cpus; in update_pd_power_uw() 145 dtpm->power_max *= nr_cpus; in update_pd_power_uw()
|
/linux/arch/mips/loongson64/ |
A D | env.c | 139 loongson_sysconf.nr_cpus = ecpu->nr_cpus; in prom_lefi_init_env() 142 if (ecpu->nr_cpus > NR_CPUS || ecpu->nr_cpus == 0) in prom_lefi_init_env() 143 loongson_sysconf.nr_cpus = NR_CPUS; in prom_lefi_init_env() 144 loongson_sysconf.nr_nodes = (loongson_sysconf.nr_cpus + in prom_lefi_init_env()
|
/linux/tools/perf/util/ |
A D | bpf_counter_cgroup.c | 50 __u32 nr_cpus = evlist->core.all_cpus->nr; in bperf_load_program() local 96 for (i = 0; i < nr_cpus; i++) { in bperf_load_program() 125 for (cpu = 0; cpu < nr_cpus; cpu++) { in bperf_load_program() 211 int nr_cpus = evlist->core.all_cpus->nr; in bperf_cgrp__sync_counters() local 214 for (i = 0; i < nr_cpus; i++) { in bperf_cgrp__sync_counters() 247 int i, cpu, nr_cpus = evlist->core.all_cpus->nr; in bperf_cgrp__read() local 274 for (i = 0; i < nr_cpus; i++) { in bperf_cgrp__read()
|
A D | svghelper.c | 700 int *pos, int nr_cpus) in scan_thread_topology() argument 723 scan_thread_topology(map, t, cpu, &pos, nr_cpus); in scan_core_topology() 726 static int str_to_bitmap(char *s, cpumask_t *b, int nr_cpus) in str_to_bitmap() argument 739 if (c >= nr_cpus) { in str_to_bitmap() 754 int i, nr_cpus; in svg_build_topology_map() local 758 nr_cpus = min(env->nr_cpus_online, MAX_NR_CPUS); in svg_build_topology_map() 774 if (str_to_bitmap(sib_core, &t.sib_core[i], nr_cpus)) { in svg_build_topology_map() 783 if (str_to_bitmap(sib_thr, &t.sib_thr[i], nr_cpus)) { in svg_build_topology_map() 791 topology_map = malloc(sizeof(int) * nr_cpus); in svg_build_topology_map() 797 for (i = 0; i < nr_cpus; i++) in svg_build_topology_map() [all …]
|
/linux/arch/arm64/kvm/vgic/ |
A D | vgic-debug.c | 26 int nr_cpus; member 45 ++iter->vcpu_id < iter->nr_cpus) in iter_next() 58 int nr_cpus = atomic_read(&kvm->online_vcpus); in iter_init() local 62 iter->nr_cpus = nr_cpus; in iter_init() 78 iter->vcpu_id == iter->nr_cpus && in end_of_vgic() 246 if (iter->vcpu_id < iter->nr_cpus) in vgic_debug_show()
|
/linux/drivers/pci/controller/ |
A D | pcie-iproc-msi.c | 98 int nr_cpus; member 197 return (hwirq % msi->nr_cpus); in hwirq_to_cpu() 253 if (msi->nr_cpus > 1 && nr_irqs > 1) in iproc_msi_irq_domain_alloc() 263 order_base_2(msi->nr_cpus * nr_irqs)); in iproc_msi_irq_domain_alloc() 291 order_base_2(msi->nr_cpus * nr_irqs)); in iproc_msi_irq_domain_free() 478 for (i = cpu; i < msi->nr_irqs; i += msi->nr_cpus) { in iproc_msi_irq_free() 542 msi->nr_cpus = num_possible_cpus(); in iproc_msi_init() 544 if (msi->nr_cpus == 1) in iproc_msi_init() 559 if (msi->nr_irqs < msi->nr_cpus) { in iproc_msi_init() 565 if (msi->nr_irqs % msi->nr_cpus != 0) { in iproc_msi_init() [all …]
|
/linux/tools/testing/selftests/vm/ |
A D | userfaultfd.c | 411 for (i = 0; i < nr_cpus * 2; ++i) { in uffd_test_ctx_clear() 489 for (cpu = 0; cpu < nr_cpus; cpu++) in uffd_test_ctx_init() 820 pthread_t locking_threads[nr_cpus]; in stress() 821 pthread_t uffd_threads[nr_cpus]; in stress() 822 pthread_t background_threads[nr_cpus]; in stress() 825 for (cpu = 0; cpu < nr_cpus; cpu++) { in stress() 845 for (cpu = 0; cpu < nr_cpus; cpu++) in stress() 861 for (cpu = 0; cpu < nr_cpus; cpu++) in stress() 865 for (cpu = 0; cpu < nr_cpus; cpu++) { in stress() 1423 struct uffd_stats uffd_stats[nr_cpus]; in userfaultfd_stress() [all …]
|
/linux/tools/perf/arch/arm/util/ |
A D | auxtrace.c | 21 int ret, i, nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in find_all_arm_spe_pmus() local 25 arm_spe_pmus = zalloc(sizeof(struct perf_pmu *) * nr_cpus); in find_all_arm_spe_pmus() 32 for (i = 0; i < nr_cpus; i++) { in find_all_arm_spe_pmus()
|
/linux/tools/virtio/virtio-trace/ |
A D | trace-agent.c | 31 int nr_cpus = (int)sysconf(_SC_NPROCESSORS_CONF); in get_total_cpus() local 33 if (nr_cpus <= 0) { in get_total_cpus() 36 } else if (nr_cpus > MAX_CPUS) { in get_total_cpus() 41 return nr_cpus; in get_total_cpus()
|