/linux/arch/sh/kernel/ |
A D | perf_event.c | 39 static atomic_t num_events; variable 65 if (!atomic_add_unless(&num_events, -1, 1)) { in hw_perf_event_destroy() 67 if (atomic_dec_return(&num_events) == 0) in hw_perf_event_destroy() 118 if (!atomic_inc_not_zero(&num_events)) { in __hw_perf_event_init() 120 if (atomic_read(&num_events) == 0 && in __hw_perf_event_init() 124 atomic_inc(&num_events); in __hw_perf_event_init() 254 idx = find_first_zero_bit(cpuc->used_mask, sh_pmu->num_events); in sh_pmu_add() 255 if (idx == sh_pmu->num_events) in sh_pmu_add() 357 WARN_ON(_pmu->num_events > MAX_HWEVENTS); in register_sh_pmu()
|
A D | hw_breakpoint.c | 35 static struct sh_ubc ubc_dummy = { .num_events = 0 }; 52 for (i = 0; i < sh_ubc->num_events; i++) { in arch_install_hw_breakpoint() 61 if (WARN_ONCE(i == sh_ubc->num_events, "Can't find any breakpoint slot")) in arch_install_hw_breakpoint() 84 for (i = 0; i < sh_ubc->num_events; i++) { in arch_uninstall_hw_breakpoint() 93 if (WARN_ONCE(i == sh_ubc->num_events, "Can't find any breakpoint slot")) in arch_uninstall_hw_breakpoint() 267 for (i = 0; i < sh_ubc->num_events; i++) { in flush_ptrace_hw_breakpoint() 297 for (i = 0; i < sh_ubc->num_events; i++) { in hw_breakpoint_handler() 405 WARN_ON(ubc->num_events > HBP_NUM); in register_sh_ubc()
|
/linux/arch/sh/kernel/cpu/sh4a/ |
A D | ubc.c | 48 for (i = 0; i < sh4a_ubc.num_events; i++) in sh4a_ubc_enable_all() 58 for (i = 0; i < sh4a_ubc.num_events; i++) in sh4a_ubc_disable_all() 68 for (i = 0; i < sh4a_ubc.num_events; i++) in sh4a_ubc_active_mask() 87 .num_events = 2, 114 for (i = 0; i < sh4a_ubc.num_events; i++) { in sh4a_ubc_init()
|
A D | perf_event.c | 261 for (i = 0; i < sh4a_pmu.num_events; i++) in sh4a_pmu_disable_all() 269 for (i = 0; i < sh4a_pmu.num_events; i++) in sh4a_pmu_enable_all() 275 .num_events = 2,
|
/linux/drivers/gpu/drm/amd/amdkfd/ |
A D | kfd_events.c | 531 event_waiters = kmalloc_array(num_events, in alloc_event_waiters() 586 for (i = 0; i < num_events; i++) { in test_event_condition() 598 return activated_count == num_events ? in test_event_condition() 616 for (i = 0; i < num_events; i++) { in copy_signaled_event_data() 656 for (i = 0; i < num_events; i++) in free_waiters() 665 uint32_t num_events, void __user *data, in kfd_wait_on_events() argument 685 for (i = 0; i < num_events; i++) { in kfd_wait_on_events() 703 ret = copy_signaled_event_data(num_events, in kfd_wait_on_events() 714 for (i = 0; i < num_events; i++) in kfd_wait_on_events() 765 ret = copy_signaled_event_data(num_events, in kfd_wait_on_events() [all …]
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
A D | amdgpu_pmu.c | 92 unsigned int num_events; member 138 .num_events = ARRAY_SIZE(vega20_events), 173 .num_events = ARRAY_SIZE(df_vega20_events), 203 .num_events = ARRAY_SIZE(arcturus_events), 417 int num_events) in amdgpu_pmu_create_attrs() argument 420 num_events, AMDGPU_PMU_EVENT_CONFIG_TYPE_NONE); in amdgpu_pmu_create_attrs() 443 *evt_attr = kcalloc(config->num_events, sizeof(**evt_attr), GFP_KERNEL); in amdgpu_pmu_alloc_pmu_attrs() 448 evt_attr_group->attrs = kcalloc(config->num_events + 1, in amdgpu_pmu_alloc_pmu_attrs() 515 config->events, config->num_events); in init_pmu_entry_by_type_and_add() 516 total_num_events = config->num_events; in init_pmu_entry_by_type_and_add()
|
/linux/arch/powerpc/perf/ |
A D | e500-pmu.c | 83 static int num_events = 128; variable 91 if (event_low >= num_events) in e500_xlate_event() 125 num_events = 256; in init_e500_pmu()
|
A D | core-fsl-emb.c | 30 static atomic_t num_events; variable 209 if (atomic_read(&num_events)) { in fsl_emb_pmu_disable() 431 if (!atomic_add_unless(&num_events, -1, 1)) { in hw_perf_event_destroy() 433 if (atomic_dec_return(&num_events) == 0) in hw_perf_event_destroy() 556 if (!atomic_inc_not_zero(&num_events)) { in fsl_emb_pmu_event_init() 558 if (atomic_read(&num_events) == 0 && in fsl_emb_pmu_event_init() 562 atomic_inc(&num_events); in fsl_emb_pmu_event_init()
|
A D | e6500-pmu.c | 85 static int num_events = 512; variable 91 if (event_low >= num_events || in e6500_xlate_event()
|
/linux/tools/perf/util/bpf_skel/ |
A D | bperf_cgroup.bpf.c | 46 const volatile __u32 num_events = 1; variable 126 if (idx == num_events) in bperf_cgroup_count() 160 key = cgrp * num_events + idx; in bperf_cgroup_count()
|
/linux/arch/sh/kernel/cpu/sh4/ |
A D | perf_event.c | 227 for (i = 0; i < sh7750_pmu.num_events; i++) in sh7750_pmu_disable_all() 235 for (i = 0; i < sh7750_pmu.num_events; i++) in sh7750_pmu_enable_all() 241 .num_events = 2,
|
/linux/drivers/firmware/arm_scmi/ |
A D | notify.c | 162 if (pd_ && eid_ < pd_->num_events) \ 303 int num_events; member 693 int num_events, in scmi_allocate_registered_events_desc() argument 720 pd->registered_events = devm_kcalloc(ni->handle->dev, num_events, in scmi_allocate_registered_events_desc() 724 pd->num_events = num_events; in scmi_allocate_registered_events_desc() 779 for (i = 0; i < ee->num_events; i++) in scmi_register_protocol_events() 784 payld_sz, ee->num_events, in scmi_register_protocol_events() 790 for (i = 0; i < ee->num_events; i++, evt++) { in scmi_register_protocol_events()
|
A D | notify.h | 76 unsigned int num_events; member
|
/linux/drivers/dma/ |
A D | imx-sdma.c | 481 int num_events; member 523 .num_events = 32, 544 .num_events = 48, 550 .num_events = 48, 568 .num_events = 48, 588 .num_events = 48, 607 .num_events = 48, 613 .num_events = 48, 632 .num_events = 48, 638 .num_events = 48, [all …]
|
/linux/drivers/bus/mhi/ |
A D | pci_generic.c | 256 .num_events = ARRAY_SIZE(modem_qcom_v1_mhi_events), 320 .num_events = ARRAY_SIZE(mhi_quectel_em1xx_events), 358 .num_events = ARRAY_SIZE(mhi_foxconn_sdx55_events), 395 .num_events = ARRAY_SIZE(mhi_mv31_events), 560 mhi_cntrl->nr_irqs = 1 + mhi_cntrl_config->num_events; in mhi_pci_get_irqs() 573 for (i = 0; i < mhi_cntrl_config->num_events; i++) in mhi_pci_get_irqs()
|
/linux/arch/sh/include/asm/ |
A D | perf_event.h | 11 unsigned int num_events; member
|
A D | hw_breakpoint.h | 31 unsigned int num_events; member
|
/linux/arch/arm/kernel/ |
A D | perf_event_xscale.c | 173 for (idx = 0; idx < cpu_pmu->num_events; ++idx) { in xscale1pmu_handle_irq() 383 cpu_pmu->num_events = 3; in xscale1pmu_init() 519 for (idx = 0; idx < cpu_pmu->num_events; ++idx) { in xscale2pmu_handle_irq() 752 cpu_pmu->num_events = 5; in xscale2pmu_init()
|
A D | perf_event_v6.c | 326 for (idx = 0; idx < cpu_pmu->num_events; ++idx) { in armv6pmu_handle_irq() 504 cpu_pmu->num_events = 3; in armv6pmu_init() 555 cpu_pmu->num_events = 3; in armv6mpcore_pmu_init()
|
/linux/drivers/devfreq/event/ |
A D | exynos-ppmu.c | 36 unsigned int num_events; member 524 info->num_events = count; in of_get_devfreq_events() 670 size = sizeof(struct devfreq_event_dev *) * info->num_events; in exynos_ppmu_probe() 678 for (i = 0; i < info->num_events; i++) { in exynos_ppmu_probe()
|
/linux/drivers/perf/ |
A D | arm_pmu.c | 527 int enabled = bitmap_weight(hw_events->used_mask, armpmu->num_events); in armpmu_enable() 749 for (idx = 0; idx < armpmu->num_events; idx++) { in cpu_pm_pmu_setup() 788 int enabled = bitmap_weight(hw_events->used_mask, armpmu->num_events); in cpu_pm_pmu_notify() 952 pmu->name, pmu->num_events, in armpmu_register()
|
/linux/drivers/infiniband/hw/mlx5/ |
A D | devx.c | 1902 for (i = 0; i < num_events; i++) { in is_valid_events_legacy() 1917 int num_events, u16 *event_type_num_list, in is_valid_events() argument 1932 return is_valid_events_legacy(num_events, event_type_num_list, in is_valid_events() 1936 for (i = 0; i < num_events; i++) { in is_valid_events() 1982 int num_events; in UVERBS_HANDLER() local 2028 num_events = uverbs_attr_ptr_get_array_size( in UVERBS_HANDLER() 2032 if (num_events < 0) in UVERBS_HANDLER() 2033 return num_events; in UVERBS_HANDLER() 2035 if (num_events > MAX_NUM_EVENTS) in UVERBS_HANDLER() 2041 if (!is_valid_events(dev->mdev, num_events, event_type_num_list, obj)) in UVERBS_HANDLER() [all …]
|
/linux/arch/s390/kernel/ |
A D | perf_cpum_sf.c | 535 static atomic_t num_events; variable 605 if (!atomic_add_unless(&num_events, -1, 1)) { in hw_perf_event_destroy() 607 if (atomic_dec_return(&num_events) == 0) in hw_perf_event_destroy() 774 if (!atomic_inc_not_zero(&num_events)) { in __hw_perf_event_init() 776 if (atomic_read(&num_events) == 0 && reserve_pmc_hardware()) in __hw_perf_event_init() 779 atomic_inc(&num_events); in __hw_perf_event_init() 2097 if (!atomic_read(&num_events)) in cpusf_pmu_setup()
|
/linux/drivers/iio/accel/ |
A D | mma9553.c | 182 int num_events; member 225 data->num_events = MMA9553_EVENTS_INFO_SIZE; in mma9553_init_events() 226 for (i = 0; i < data->num_events; i++) { in mma9553_init_events() 239 for (i = 0; i < data->num_events; i++) in mma9553_get_event() 254 for (i = 0; i < data->num_events; i++) in mma9553_is_any_event_enabled()
|
/linux/arch/nds32/kernel/ |
A D | perf_event_cpu.c | 250 for (idx = 0; idx < cpu_pmu->num_events; ++idx) { in nds32_pmu_handle_irq() 289 return ((idx >= 0) && (idx < cpu_pmu->num_events)); in nds32_pfm_counter_valid() 670 cpu_pmu->num_events = nds32_read_num_pfm_events(); in device_pmu_init() 699 nds32_pmu->num_events); in nds32_pmu_enable() 1062 nds32_pmu->name, nds32_pmu->num_events); in nds32_pmu_register()
|