| /linux/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/ |
| A D | base.c | 32 struct nvkm_pmu *pmu = device->pmu; in nvkm_pmu_fan_controlled() local 37 if (pmu && pmu->func->code.size) in nvkm_pmu_fan_controlled() 50 if (pmu && pmu->func->pgob) in nvkm_pmu_pgob() 51 pmu->func->pgob(pmu, enable); in nvkm_pmu_pgob() 58 return pmu->func->recv(pmu); in nvkm_pmu_recv() 65 if (!pmu || !pmu->func->send) in nvkm_pmu_send() 76 pmu->func->intr(pmu); in nvkm_pmu_intr() 85 pmu->func->fini(pmu); in nvkm_pmu_fini() 102 if (!pmu->func->enabled(pmu)) in nvkm_pmu_reset() 114 pmu->func->reset(pmu); in nvkm_pmu_reset() [all …]
|
| A D | gt215.c | 37 mutex_lock(&pmu->send.mutex); in gt215_pmu_send() 54 pmu->recv.message = message; in gt215_pmu_send() 55 pmu->recv.process = process; in gt215_pmu_send() 65 pmu->send.base)); in gt215_pmu_send() 77 wait_event(pmu->recv.wait, (pmu->recv.process == 0)); in gt215_pmu_send() 105 pmu->recv.base)); in gt215_pmu_recv() 116 if (pmu->recv.process) { in gt215_pmu_recv() 119 pmu->recv.data[0] = data0; in gt215_pmu_recv() 120 pmu->recv.data[1] = data1; in gt215_pmu_recv() 121 pmu->recv.process = 0; in gt215_pmu_recv() [all …]
|
| A D | gk20a.c | 116 struct gk20a_pmu *pmu = in gk20a_pmu_dvfs_work() local 149 gk20a_pmu_dvfs_target(pmu, &state); in gk20a_pmu_dvfs_work() 153 gk20a_pmu_dvfs_reset_dev_status(pmu); in gk20a_pmu_dvfs_work() 158 gk20a_pmu_fini(struct nvkm_pmu *pmu) in gk20a_pmu_fini() argument 160 struct gk20a_pmu *gpmu = gk20a_pmu(pmu); in gk20a_pmu_fini() 163 nvkm_falcon_put(&pmu->falcon, &pmu->subdev); in gk20a_pmu_fini() 167 gk20a_pmu_init(struct nvkm_pmu *pmu) in gk20a_pmu_init() argument 216 struct gk20a_pmu *pmu; in gk20a_pmu_new() local 219 if (!(pmu = kzalloc(sizeof(*pmu), GFP_KERNEL))) in gk20a_pmu_new() 221 *ppmu = &pmu->base; in gk20a_pmu_new() [all …]
|
| A D | gm20b.c | 42 struct nvkm_pmu *pmu = container_of(falcon, typeof(*pmu), falcon); in gm20b_pmu_acr_bootstrap_falcon() local 140 struct nvkm_pmu *pmu = priv; in gm20b_pmu_acr_init_wpr_callback() local 150 complete_all(&pmu->wpr_ready); in gm20b_pmu_acr_init_wpr_callback() 192 return gm20b_pmu_acr_init_wpr(pmu); in gm20b_pmu_initmsg() 196 gm20b_pmu_recv(struct nvkm_pmu *pmu) in gm20b_pmu_recv() argument 198 if (!pmu->initmsg_received) { in gm20b_pmu_recv() 199 int ret = pmu->func->initmsg(pmu); in gm20b_pmu_recv() 201 nvkm_error(&pmu->subdev, in gm20b_pmu_recv() 206 pmu->initmsg_received = true; in gm20b_pmu_recv() 209 nvkm_falcon_msgq_recv(pmu->msgq); in gm20b_pmu_recv() [all …]
|
| /linux/drivers/gpu/drm/i915/ |
| A D | i915_pmu.c | 171 struct i915_pmu *pmu = &i915->pmu; in get_rc6() local 223 struct i915_pmu *pmu = &i915->pmu; in park_rc6() local 242 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_gt_parked() local 262 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_gt_unparked() local 376 struct i915_pmu *pmu = &i915->pmu; in frequency_sample() local 421 struct i915_pmu *pmu = &i915->pmu; in i915_sample() local 524 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_event_init() local 566 struct i915_pmu *pmu = &i915->pmu; in __i915_pmu_event_read() local 620 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_event_read() local 641 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_enable() local [all …]
|
| /linux/drivers/soc/dove/ |
| A D | pmu.c | 108 pmu->reset.of_node = pmu->of_node; in pmu_reset_init() 146 struct pmu_data *pmu = pmu_dom->pmu; in pmu_domain_power_off() local 180 struct pmu_data *pmu = pmu_dom->pmu; in pmu_domain_power_on() local 315 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in dove_init_pmu_legacy() 316 if (!pmu) in dove_init_pmu_legacy() 330 domain->pmu = pmu; in dove_init_pmu_legacy() 390 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in dove_init_pmu() 396 pmu->pmc_base = of_iomap(pmu->of_node, 0); in dove_init_pmu() 397 pmu->pmu_base = of_iomap(pmu->of_node, 1); in dove_init_pmu() 398 if (!pmu->pmc_base || !pmu->pmu_base) { in dove_init_pmu() [all …]
|
| /linux/drivers/perf/ |
| A D | fsl_imx8_ddr_perf.c | 94 struct pmu pmu; member 317 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_is_enhanced_filtered() local 370 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_init() local 390 if (event->group_leader->pmu != event->pmu && in ddr_perf_event_init() 404 if (sibling->pmu != event->pmu && in ddr_perf_event_init() 463 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_update() local 491 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_start() local 504 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_add() local 546 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_stop() local 596 .pmu = (struct pmu) { in ddr_perf_init() [all …]
|
| A D | arm_pmu_platform.c | 38 ret = info->init(pmu); in probe_current_pmu() 113 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in pmu_parse_irqs() 193 struct arm_pmu *pmu; in arm_pmu_device_probe() local 196 pmu = armpmu_alloc(); in arm_pmu_device_probe() 197 if (!pmu) in arm_pmu_device_probe() 200 pmu->plat_device = pdev; in arm_pmu_device_probe() 202 ret = pmu_parse_irqs(pmu); in arm_pmu_device_probe() 217 ret = init_fn(pmu); in arm_pmu_device_probe() 232 ret = armpmu_register(pmu); in arm_pmu_device_probe() 241 armpmu_free_irqs(pmu); in arm_pmu_device_probe() [all …]
|
| A D | arm_pmu_acpi.c | 191 struct arm_pmu *pmu; in arm_pmu_acpi_find_alloc_pmu() local 196 if (!pmu || pmu->acpi_cpuid != cpuid) in arm_pmu_acpi_find_alloc_pmu() 199 return pmu; in arm_pmu_acpi_find_alloc_pmu() 203 if (!pmu) { in arm_pmu_acpi_find_alloc_pmu() 211 return pmu; in arm_pmu_acpi_find_alloc_pmu() 254 struct arm_pmu *pmu; in arm_pmu_acpi_cpu_starting() local 265 if (!pmu) in arm_pmu_acpi_cpu_starting() 309 if (!pmu || pmu->name) in arm_pmu_acpi_probe() 312 ret = init_fn(pmu); in arm_pmu_acpi_probe() 323 if (!pmu->name) { in arm_pmu_acpi_probe() [all …]
|
| A D | arm_pmu.c | 373 if (event->pmu != pmu) in validate_event() 523 static void armpmu_enable(struct pmu *pmu) in armpmu_enable() argument 537 static void armpmu_disable(struct pmu *pmu) in armpmu_disable() argument 714 pmu->reset(pmu); in arm_perf_starting_cpu() 869 pmu = kzalloc(sizeof(*pmu), flags); in __armpmu_alloc() 870 if (!pmu) in __armpmu_alloc() 879 pmu->pmu = (struct pmu) { in __armpmu_alloc() 911 return pmu; in __armpmu_alloc() 945 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_EXCLUDE; in armpmu_register() 947 ret = perf_pmu_register(&pmu->pmu, pmu->name, -1); in armpmu_register() [all …]
|
| /linux/arch/x86/kvm/vmx/ |
| A D | pmu_intel.c | 56 pmu->fixed_ctr_ctrl = data; in reprogram_fixed_counters() 65 pmu->global_ctrl = data; in global_ctrl_changed() 220 ret = pmu->version > 1; in intel_is_valid_msr() 225 get_fixed_pmc(pmu, msr) || get_fw_gp_pmc(pmu, msr) || in intel_is_valid_msr() 303 pmu->event_count++; in intel_pmu_create_guest_lbr_event() 476 pmu->version = 0; in intel_pmu_refresh() 486 if (!pmu->version) in intel_pmu_refresh() 511 pmu->global_ctrl = ((1ull << pmu->nr_arch_gp_counters) - 1) | in intel_pmu_refresh() 513 pmu->global_ctrl_mask = ~pmu->global_ctrl; in intel_pmu_refresh() 514 pmu->global_ovf_ctrl_mask = pmu->global_ctrl_mask in intel_pmu_refresh() [all …]
|
| /linux/arch/x86/kvm/svm/ |
| A D | pmu.c | 49 struct kvm_vcpu *vcpu = pmu_to_vcpu(pmu); in get_msr_base() 190 return idx < pmu->nr_arch_gp_counters; in amd_is_valid_rdpmc_ecx() 201 if (idx >= pmu->nr_arch_gp_counters) in amd_rdpmc_ecx_to_pmc() 203 counters = pmu->gp_counters; in amd_rdpmc_ecx_to_pmc() 265 if (!(data & pmu->reserved_bits)) { in amd_pmu_set_msr() 285 pmu->version = 1; in amd_pmu_refresh() 288 pmu->nr_arch_fixed_counters = 0; in amd_pmu_refresh() 289 pmu->global_status = 0; in amd_pmu_refresh() 290 bitmap_set(pmu->all_valid_pmc_idx, 0, pmu->nr_arch_gp_counters); in amd_pmu_refresh() 302 pmu->gp_counters[i].vcpu = vcpu; in amd_pmu_init() [all …]
|
| /linux/Documentation/devicetree/bindings/arm/samsung/ |
| A D | pmu.yaml | 18 - samsung,exynos3250-pmu 19 - samsung,exynos4210-pmu 20 - samsung,exynos4412-pmu 21 - samsung,exynos5250-pmu 22 - samsung,exynos5260-pmu 23 - samsung,exynos5410-pmu 24 - samsung,exynos5420-pmu 26 - samsung,exynos7-pmu 27 - samsung-s5pv210-pmu 43 - samsung,exynos7-pmu [all …]
|
| /linux/arch/x86/events/intel/ |
| A D | uncore.h | 85 struct pmu *pmu; /* for custom pmu ops */ member 119 struct pmu pmu; member 148 struct intel_uncore_pmu *pmu; member 254 box->pmu->type->mmio_offset * box->pmu->pmu_idx; in uncore_mmio_box_ctl() 289 struct intel_uncore_pmu *pmu = box->pmu; in uncore_msr_box_offset() local 291 pmu->type->msr_offsets[pmu->pmu_idx] : in uncore_msr_box_offset() 292 pmu->type->msr_offset * pmu->pmu_idx; in uncore_msr_box_offset() 361 struct intel_uncore_pmu *pmu = box->pmu; in uncore_freerunning_counter() local 366 pmu->type->freerunning[type].box_offsets[pmu->pmu_idx] : in uncore_freerunning_counter() 367 pmu->type->freerunning[type].box_offset * pmu->pmu_idx); in uncore_freerunning_counter() [all …]
|
| A D | uncore.c | 370 return &box->pmu->pmu == event->pmu; in is_box_event() 693 fake_box->pmu = pmu; in uncore_validate_group() 888 if (!pmu->type->pmu) { in uncore_pmu_register() 889 pmu->pmu = (struct pmu) { in uncore_pmu_register() 905 pmu->pmu = *pmu->type->pmu; in uncore_pmu_register() 906 pmu->pmu.attr_groups = pmu->type->attr_groups; in uncore_pmu_register() 907 pmu->pmu.attr_update = pmu->type->attr_update; in uncore_pmu_register() 912 ret = perf_pmu_register(&pmu->pmu, pmu->name, -1); in uncore_pmu_register() 1144 box->pmu = pmu; in uncore_pci_pmu_register() 1247 pmu = box->pmu; in uncore_pci_remove() [all …]
|
| /linux/arch/x86/events/ |
| A D | rapl.c | 110 struct pmu *pmu; member 116 struct pmu pmu; member 212 hrtimer_start(&pmu->hrtimer, pmu->timer_interval, in rapl_start_hrtimer() 257 pmu->n_active++; in __rapl_pmu_event_start() 364 if (!pmu) in rapl_pmu_event_init() 553 pmu->cpu = -1; in rapl_cpu_offline() 561 perf_pmu_migrate_context(pmu->pmu, cpu, target); in rapl_cpu_offline() 571 if (!pmu) { in rapl_cpu_online() 573 if (!pmu) in rapl_cpu_online() 578 pmu->pmu = &rapl_pmus->pmu; in rapl_cpu_online() [all …]
|
| /linux/arch/powerpc/perf/ |
| A D | imc-pmu.c | 103 struct pmu *pmu = dev_get_drvdata(dev); in imc_pmu_cpumask_get_attr() local 1469 pmu->pmu.add = imc_event_add; in update_pmu_ops() 1470 pmu->pmu.del = imc_event_stop; in update_pmu_ops() 1471 pmu->pmu.start = imc_event_start; in update_pmu_ops() 1472 pmu->pmu.stop = imc_event_stop; in update_pmu_ops() 1473 pmu->pmu.read = imc_event_update; in update_pmu_ops() 1474 pmu->pmu.attr_groups = pmu->attr_groups; in update_pmu_ops() 1489 pmu->pmu.add = thread_imc_event_add; in update_pmu_ops() 1490 pmu->pmu.del = thread_imc_event_del; in update_pmu_ops() 1497 pmu->pmu.add = trace_imc_event_add; in update_pmu_ops() [all …]
|
| /linux/tools/perf/util/ |
| A D | pmu.c | 906 struct perf_pmu *pmu = idata->pmu; in pmu_add_sys_aliases_iter_fn() local 933 .pmu = pmu, in pmu_add_sys_aliases() 1006 pmu = zalloc(sizeof(*pmu)); in pmu_lookup() 1083 pmu = list_prepare_entry(pmu, &pmus, list); in perf_pmu__scan() 1094 while ((pmu = perf_pmu__scan(pmu)) != NULL) { in evsel__find_pmu() 1106 return pmu && pmu->auxtrace; in evsel__is_aux_event() 1629 if (as->pmu && bs->pmu) in cmp_sevent() 1630 return strcmp(as->pmu, bs->pmu); in cmp_sevent() 1693 while ((pmu = perf_pmu__scan(pmu)) != NULL) { in print_pmu_events() 1704 while ((pmu = perf_pmu__scan(pmu)) != NULL) { in print_pmu_events() [all …]
|
| A D | parse-events-hybrid.c | 44 struct perf_pmu *pmu) in create_event_hybrid() argument 52 pmu, config_terms); in create_event_hybrid() 64 struct perf_pmu *pmu) in pmu_cmp() argument 77 struct perf_pmu *pmu; in add_hw_hybrid() local 83 if (pmu_cmp(parse_state, pmu)) in add_hw_hybrid() 89 metric_id, &terms, pmu); in add_hw_hybrid() 103 struct perf_pmu *pmu) in create_raw_event_hybrid() argument 107 attr->type = pmu->type; in create_raw_event_hybrid() 109 pmu, config_terms); in create_raw_event_hybrid() 123 struct perf_pmu *pmu; in add_raw_hybrid() local [all …]
|
| /linux/Documentation/devicetree/bindings/pinctrl/ |
| A D | marvell,dove-pinctrl.txt | 18 mpp0 0 gpio, pmu, uart2(rts), sdio0(cd), lcd0(pwm), pmu* 19 mpp1 1 gpio, pmu, uart2(cts), sdio0(wp), lcd1(pwm), pmu* 24 mpp4 4 gpio, pmu, uart3(rts), sdio1(cd), spi1(miso), pmu* 25 mpp5 5 gpio, pmu, uart3(cts), sdio1(wp), spi1(cs), pmu* 26 mpp6 6 gpio, pmu, uart3(txd), sdio1(buspwr), spi1(mosi), pmu* 27 mpp7 7 gpio, pmu, uart3(rxd), sdio1(ledctrl), spi1(sck), pmu* 28 mpp8 8 gpio, pmu, watchdog(rstout), pmu* 29 mpp9 9 gpio, pmu, pex1(clkreq), pmu* 30 mpp10 10 gpio, pmu, ssp(sclk), pmu* 37 mpp14 14 gpio, pmu, uart2(txd), sdio1(buspwr), ssp(rxd), pmu* [all …]
|
| /linux/Documentation/devicetree/bindings/arm/ |
| A D | pmu.yaml | 22 - apm,potenza-pmu 24 - arm,arm1136-pmu 25 - arm,arm1176-pmu 27 - arm,cortex-a5-pmu 28 - arm,cortex-a7-pmu 29 - arm,cortex-a8-pmu 30 - arm,cortex-a9-pmu 31 - arm,cortex-a12-pmu 32 - arm,cortex-a15-pmu 49 - brcm,vulcan-pmu [all …]
|
| /linux/arch/arm64/kvm/ |
| A D | pmu-emul.c | 166 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_get_counter_value() local 252 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_vcpu_init() local 266 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_vcpu_reset() local 283 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_vcpu_destroy() local 311 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_enable_counter_mask() local 346 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_disable_counter_mask() local 383 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_update_state() local 397 pmu->irq_num, overflow, pmu); in kvm_pmu_update_state() 404 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_should_notify_user() local 516 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_software_increment() local [all …]
|
| /linux/drivers/soc/rockchip/ |
| A D | pm_domains.c | 144 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_is_idle() local 156 regmap_read(pmu->regmap, pmu->info->ack_offset, &val); in rockchip_pmu_read_ack() 165 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_set_idle_request() local 174 regmap_write(pmu->regmap, pmu->info->req_offset, in rockchip_pmu_set_idle_request() 258 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_is_on() local 274 struct rockchip_pmu *pmu = pd->pmu; in rockchip_do_pmu_set_power_domain() local 281 regmap_write(pmu->regmap, pmu->info->pwr_offset, in rockchip_do_pmu_set_power_domain() 301 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pd_power() local 425 pd->pmu = pmu; in rockchip_pm_add_one_domain() 657 if (!pmu) in rockchip_pm_domain_probe() [all …]
|
| /linux/drivers/regulator/ |
| A D | bcm590xx-regulator.c | 278 struct bcm590xx_reg *pmu; in bcm590xx_probe() local 284 pmu = devm_kzalloc(&pdev->dev, sizeof(*pmu), GFP_KERNEL); in bcm590xx_probe() 285 if (!pmu) in bcm590xx_probe() 288 pmu->mfd = bcm590xx; in bcm590xx_probe() 290 platform_set_drvdata(pdev, pmu); in bcm590xx_probe() 292 pmu->desc = devm_kcalloc(&pdev->dev, in bcm590xx_probe() 296 if (!pmu->desc) in bcm590xx_probe() 303 pmu->desc[i].name = info->name; in bcm590xx_probe() 307 pmu->desc[i].id = i; in bcm590xx_probe() 332 pmu->desc[i].owner = THIS_MODULE; in bcm590xx_probe() [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ |
| A D | gm200.c | 41 nvkm_wr32(device, 0x10a188, (pmu + i) >> 8); in pmu_code() 87 struct nvbios_pmuR pmu; in pmu_load() local 89 if (!nvbios_pmuRm(bios, type, &pmu)) in pmu_load() 95 pmu_code(init, pmu.boot_addr_pmu, pmu.boot_addr, pmu.boot_size, false); in pmu_load() 96 pmu_code(init, pmu.code_addr_pmu, pmu.code_addr, pmu.code_size, true); in pmu_load() 97 pmu_data(init, pmu.data_addr_pmu, pmu.data_addr, pmu.data_size); in pmu_load() 100 *init_addr_pmu = pmu.init_addr_pmu; in pmu_load() 101 *args_addr_pmu = pmu.args_addr_pmu; in pmu_load() 105 return pmu_exec(init, pmu.init_addr_pmu), 0; in pmu_load() 146 pmu_data(init, pmu, img, len); in gm200_devinit_post() [all …]
|