Home
last modified time | relevance | path

Searched refs:pmu (Results 1 – 25 of 553) sorted by relevance

12345678910>>...23

/linux/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/
A Dbase.c32 struct nvkm_pmu *pmu = device->pmu; in nvkm_pmu_fan_controlled() local
37 if (pmu && pmu->func->code.size) in nvkm_pmu_fan_controlled()
50 if (pmu && pmu->func->pgob) in nvkm_pmu_pgob()
51 pmu->func->pgob(pmu, enable); in nvkm_pmu_pgob()
58 return pmu->func->recv(pmu); in nvkm_pmu_recv()
65 if (!pmu || !pmu->func->send) in nvkm_pmu_send()
76 pmu->func->intr(pmu); in nvkm_pmu_intr()
85 pmu->func->fini(pmu); in nvkm_pmu_fini()
102 if (!pmu->func->enabled(pmu)) in nvkm_pmu_reset()
114 pmu->func->reset(pmu); in nvkm_pmu_reset()
[all …]
A Dgt215.c37 mutex_lock(&pmu->send.mutex); in gt215_pmu_send()
54 pmu->recv.message = message; in gt215_pmu_send()
55 pmu->recv.process = process; in gt215_pmu_send()
65 pmu->send.base)); in gt215_pmu_send()
77 wait_event(pmu->recv.wait, (pmu->recv.process == 0)); in gt215_pmu_send()
105 pmu->recv.base)); in gt215_pmu_recv()
116 if (pmu->recv.process) { in gt215_pmu_recv()
119 pmu->recv.data[0] = data0; in gt215_pmu_recv()
120 pmu->recv.data[1] = data1; in gt215_pmu_recv()
121 pmu->recv.process = 0; in gt215_pmu_recv()
[all …]
A Dgk20a.c116 struct gk20a_pmu *pmu = in gk20a_pmu_dvfs_work() local
149 gk20a_pmu_dvfs_target(pmu, &state); in gk20a_pmu_dvfs_work()
153 gk20a_pmu_dvfs_reset_dev_status(pmu); in gk20a_pmu_dvfs_work()
158 gk20a_pmu_fini(struct nvkm_pmu *pmu) in gk20a_pmu_fini() argument
160 struct gk20a_pmu *gpmu = gk20a_pmu(pmu); in gk20a_pmu_fini()
163 nvkm_falcon_put(&pmu->falcon, &pmu->subdev); in gk20a_pmu_fini()
167 gk20a_pmu_init(struct nvkm_pmu *pmu) in gk20a_pmu_init() argument
216 struct gk20a_pmu *pmu; in gk20a_pmu_new() local
219 if (!(pmu = kzalloc(sizeof(*pmu), GFP_KERNEL))) in gk20a_pmu_new()
221 *ppmu = &pmu->base; in gk20a_pmu_new()
[all …]
A Dgm20b.c42 struct nvkm_pmu *pmu = container_of(falcon, typeof(*pmu), falcon); in gm20b_pmu_acr_bootstrap_falcon() local
140 struct nvkm_pmu *pmu = priv; in gm20b_pmu_acr_init_wpr_callback() local
150 complete_all(&pmu->wpr_ready); in gm20b_pmu_acr_init_wpr_callback()
192 return gm20b_pmu_acr_init_wpr(pmu); in gm20b_pmu_initmsg()
196 gm20b_pmu_recv(struct nvkm_pmu *pmu) in gm20b_pmu_recv() argument
198 if (!pmu->initmsg_received) { in gm20b_pmu_recv()
199 int ret = pmu->func->initmsg(pmu); in gm20b_pmu_recv()
201 nvkm_error(&pmu->subdev, in gm20b_pmu_recv()
206 pmu->initmsg_received = true; in gm20b_pmu_recv()
209 nvkm_falcon_msgq_recv(pmu->msgq); in gm20b_pmu_recv()
[all …]
/linux/drivers/gpu/drm/i915/
A Di915_pmu.c171 struct i915_pmu *pmu = &i915->pmu; in get_rc6() local
223 struct i915_pmu *pmu = &i915->pmu; in park_rc6() local
242 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_gt_parked() local
262 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_gt_unparked() local
376 struct i915_pmu *pmu = &i915->pmu; in frequency_sample() local
421 struct i915_pmu *pmu = &i915->pmu; in i915_sample() local
524 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_event_init() local
566 struct i915_pmu *pmu = &i915->pmu; in __i915_pmu_event_read() local
620 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_event_read() local
641 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_enable() local
[all …]
/linux/drivers/soc/dove/
A Dpmu.c108 pmu->reset.of_node = pmu->of_node; in pmu_reset_init()
146 struct pmu_data *pmu = pmu_dom->pmu; in pmu_domain_power_off() local
180 struct pmu_data *pmu = pmu_dom->pmu; in pmu_domain_power_on() local
315 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in dove_init_pmu_legacy()
316 if (!pmu) in dove_init_pmu_legacy()
330 domain->pmu = pmu; in dove_init_pmu_legacy()
390 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in dove_init_pmu()
396 pmu->pmc_base = of_iomap(pmu->of_node, 0); in dove_init_pmu()
397 pmu->pmu_base = of_iomap(pmu->of_node, 1); in dove_init_pmu()
398 if (!pmu->pmc_base || !pmu->pmu_base) { in dove_init_pmu()
[all …]
/linux/drivers/perf/
A Dfsl_imx8_ddr_perf.c94 struct pmu pmu; member
317 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_is_enhanced_filtered() local
370 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_init() local
390 if (event->group_leader->pmu != event->pmu && in ddr_perf_event_init()
404 if (sibling->pmu != event->pmu && in ddr_perf_event_init()
463 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_update() local
491 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_start() local
504 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_add() local
546 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_stop() local
596 .pmu = (struct pmu) { in ddr_perf_init()
[all …]
A Darm_pmu_platform.c38 ret = info->init(pmu); in probe_current_pmu()
113 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in pmu_parse_irqs()
193 struct arm_pmu *pmu; in arm_pmu_device_probe() local
196 pmu = armpmu_alloc(); in arm_pmu_device_probe()
197 if (!pmu) in arm_pmu_device_probe()
200 pmu->plat_device = pdev; in arm_pmu_device_probe()
202 ret = pmu_parse_irqs(pmu); in arm_pmu_device_probe()
217 ret = init_fn(pmu); in arm_pmu_device_probe()
232 ret = armpmu_register(pmu); in arm_pmu_device_probe()
241 armpmu_free_irqs(pmu); in arm_pmu_device_probe()
[all …]
A Darm_pmu_acpi.c191 struct arm_pmu *pmu; in arm_pmu_acpi_find_alloc_pmu() local
196 if (!pmu || pmu->acpi_cpuid != cpuid) in arm_pmu_acpi_find_alloc_pmu()
199 return pmu; in arm_pmu_acpi_find_alloc_pmu()
203 if (!pmu) { in arm_pmu_acpi_find_alloc_pmu()
211 return pmu; in arm_pmu_acpi_find_alloc_pmu()
254 struct arm_pmu *pmu; in arm_pmu_acpi_cpu_starting() local
265 if (!pmu) in arm_pmu_acpi_cpu_starting()
309 if (!pmu || pmu->name) in arm_pmu_acpi_probe()
312 ret = init_fn(pmu); in arm_pmu_acpi_probe()
323 if (!pmu->name) { in arm_pmu_acpi_probe()
[all …]
A Darm_pmu.c373 if (event->pmu != pmu) in validate_event()
523 static void armpmu_enable(struct pmu *pmu) in armpmu_enable() argument
537 static void armpmu_disable(struct pmu *pmu) in armpmu_disable() argument
714 pmu->reset(pmu); in arm_perf_starting_cpu()
869 pmu = kzalloc(sizeof(*pmu), flags); in __armpmu_alloc()
870 if (!pmu) in __armpmu_alloc()
879 pmu->pmu = (struct pmu) { in __armpmu_alloc()
911 return pmu; in __armpmu_alloc()
945 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_EXCLUDE; in armpmu_register()
947 ret = perf_pmu_register(&pmu->pmu, pmu->name, -1); in armpmu_register()
[all …]
/linux/arch/x86/kvm/vmx/
A Dpmu_intel.c56 pmu->fixed_ctr_ctrl = data; in reprogram_fixed_counters()
65 pmu->global_ctrl = data; in global_ctrl_changed()
220 ret = pmu->version > 1; in intel_is_valid_msr()
225 get_fixed_pmc(pmu, msr) || get_fw_gp_pmc(pmu, msr) || in intel_is_valid_msr()
303 pmu->event_count++; in intel_pmu_create_guest_lbr_event()
476 pmu->version = 0; in intel_pmu_refresh()
486 if (!pmu->version) in intel_pmu_refresh()
511 pmu->global_ctrl = ((1ull << pmu->nr_arch_gp_counters) - 1) | in intel_pmu_refresh()
513 pmu->global_ctrl_mask = ~pmu->global_ctrl; in intel_pmu_refresh()
514 pmu->global_ovf_ctrl_mask = pmu->global_ctrl_mask in intel_pmu_refresh()
[all …]
/linux/arch/x86/kvm/svm/
A Dpmu.c49 struct kvm_vcpu *vcpu = pmu_to_vcpu(pmu); in get_msr_base()
190 return idx < pmu->nr_arch_gp_counters; in amd_is_valid_rdpmc_ecx()
201 if (idx >= pmu->nr_arch_gp_counters) in amd_rdpmc_ecx_to_pmc()
203 counters = pmu->gp_counters; in amd_rdpmc_ecx_to_pmc()
265 if (!(data & pmu->reserved_bits)) { in amd_pmu_set_msr()
285 pmu->version = 1; in amd_pmu_refresh()
288 pmu->nr_arch_fixed_counters = 0; in amd_pmu_refresh()
289 pmu->global_status = 0; in amd_pmu_refresh()
290 bitmap_set(pmu->all_valid_pmc_idx, 0, pmu->nr_arch_gp_counters); in amd_pmu_refresh()
302 pmu->gp_counters[i].vcpu = vcpu; in amd_pmu_init()
[all …]
/linux/Documentation/devicetree/bindings/arm/samsung/
A Dpmu.yaml18 - samsung,exynos3250-pmu
19 - samsung,exynos4210-pmu
20 - samsung,exynos4412-pmu
21 - samsung,exynos5250-pmu
22 - samsung,exynos5260-pmu
23 - samsung,exynos5410-pmu
24 - samsung,exynos5420-pmu
26 - samsung,exynos7-pmu
27 - samsung-s5pv210-pmu
43 - samsung,exynos7-pmu
[all …]
/linux/arch/x86/events/intel/
A Duncore.h85 struct pmu *pmu; /* for custom pmu ops */ member
119 struct pmu pmu; member
148 struct intel_uncore_pmu *pmu; member
254 box->pmu->type->mmio_offset * box->pmu->pmu_idx; in uncore_mmio_box_ctl()
289 struct intel_uncore_pmu *pmu = box->pmu; in uncore_msr_box_offset() local
291 pmu->type->msr_offsets[pmu->pmu_idx] : in uncore_msr_box_offset()
292 pmu->type->msr_offset * pmu->pmu_idx; in uncore_msr_box_offset()
361 struct intel_uncore_pmu *pmu = box->pmu; in uncore_freerunning_counter() local
366 pmu->type->freerunning[type].box_offsets[pmu->pmu_idx] : in uncore_freerunning_counter()
367 pmu->type->freerunning[type].box_offset * pmu->pmu_idx); in uncore_freerunning_counter()
[all …]
A Duncore.c370 return &box->pmu->pmu == event->pmu; in is_box_event()
693 fake_box->pmu = pmu; in uncore_validate_group()
888 if (!pmu->type->pmu) { in uncore_pmu_register()
889 pmu->pmu = (struct pmu) { in uncore_pmu_register()
905 pmu->pmu = *pmu->type->pmu; in uncore_pmu_register()
906 pmu->pmu.attr_groups = pmu->type->attr_groups; in uncore_pmu_register()
907 pmu->pmu.attr_update = pmu->type->attr_update; in uncore_pmu_register()
912 ret = perf_pmu_register(&pmu->pmu, pmu->name, -1); in uncore_pmu_register()
1144 box->pmu = pmu; in uncore_pci_pmu_register()
1247 pmu = box->pmu; in uncore_pci_remove()
[all …]
/linux/arch/x86/events/
A Drapl.c110 struct pmu *pmu; member
116 struct pmu pmu; member
212 hrtimer_start(&pmu->hrtimer, pmu->timer_interval, in rapl_start_hrtimer()
257 pmu->n_active++; in __rapl_pmu_event_start()
364 if (!pmu) in rapl_pmu_event_init()
553 pmu->cpu = -1; in rapl_cpu_offline()
561 perf_pmu_migrate_context(pmu->pmu, cpu, target); in rapl_cpu_offline()
571 if (!pmu) { in rapl_cpu_online()
573 if (!pmu) in rapl_cpu_online()
578 pmu->pmu = &rapl_pmus->pmu; in rapl_cpu_online()
[all …]
/linux/arch/powerpc/perf/
A Dimc-pmu.c103 struct pmu *pmu = dev_get_drvdata(dev); in imc_pmu_cpumask_get_attr() local
1469 pmu->pmu.add = imc_event_add; in update_pmu_ops()
1470 pmu->pmu.del = imc_event_stop; in update_pmu_ops()
1471 pmu->pmu.start = imc_event_start; in update_pmu_ops()
1472 pmu->pmu.stop = imc_event_stop; in update_pmu_ops()
1473 pmu->pmu.read = imc_event_update; in update_pmu_ops()
1474 pmu->pmu.attr_groups = pmu->attr_groups; in update_pmu_ops()
1489 pmu->pmu.add = thread_imc_event_add; in update_pmu_ops()
1490 pmu->pmu.del = thread_imc_event_del; in update_pmu_ops()
1497 pmu->pmu.add = trace_imc_event_add; in update_pmu_ops()
[all …]
/linux/tools/perf/util/
A Dpmu.c906 struct perf_pmu *pmu = idata->pmu; in pmu_add_sys_aliases_iter_fn() local
933 .pmu = pmu, in pmu_add_sys_aliases()
1006 pmu = zalloc(sizeof(*pmu)); in pmu_lookup()
1083 pmu = list_prepare_entry(pmu, &pmus, list); in perf_pmu__scan()
1094 while ((pmu = perf_pmu__scan(pmu)) != NULL) { in evsel__find_pmu()
1106 return pmu && pmu->auxtrace; in evsel__is_aux_event()
1629 if (as->pmu && bs->pmu) in cmp_sevent()
1630 return strcmp(as->pmu, bs->pmu); in cmp_sevent()
1693 while ((pmu = perf_pmu__scan(pmu)) != NULL) { in print_pmu_events()
1704 while ((pmu = perf_pmu__scan(pmu)) != NULL) { in print_pmu_events()
[all …]
A Dparse-events-hybrid.c44 struct perf_pmu *pmu) in create_event_hybrid() argument
52 pmu, config_terms); in create_event_hybrid()
64 struct perf_pmu *pmu) in pmu_cmp() argument
77 struct perf_pmu *pmu; in add_hw_hybrid() local
83 if (pmu_cmp(parse_state, pmu)) in add_hw_hybrid()
89 metric_id, &terms, pmu); in add_hw_hybrid()
103 struct perf_pmu *pmu) in create_raw_event_hybrid() argument
107 attr->type = pmu->type; in create_raw_event_hybrid()
109 pmu, config_terms); in create_raw_event_hybrid()
123 struct perf_pmu *pmu; in add_raw_hybrid() local
[all …]
/linux/Documentation/devicetree/bindings/pinctrl/
A Dmarvell,dove-pinctrl.txt18 mpp0 0 gpio, pmu, uart2(rts), sdio0(cd), lcd0(pwm), pmu*
19 mpp1 1 gpio, pmu, uart2(cts), sdio0(wp), lcd1(pwm), pmu*
24 mpp4 4 gpio, pmu, uart3(rts), sdio1(cd), spi1(miso), pmu*
25 mpp5 5 gpio, pmu, uart3(cts), sdio1(wp), spi1(cs), pmu*
26 mpp6 6 gpio, pmu, uart3(txd), sdio1(buspwr), spi1(mosi), pmu*
27 mpp7 7 gpio, pmu, uart3(rxd), sdio1(ledctrl), spi1(sck), pmu*
28 mpp8 8 gpio, pmu, watchdog(rstout), pmu*
29 mpp9 9 gpio, pmu, pex1(clkreq), pmu*
30 mpp10 10 gpio, pmu, ssp(sclk), pmu*
37 mpp14 14 gpio, pmu, uart2(txd), sdio1(buspwr), ssp(rxd), pmu*
[all …]
/linux/Documentation/devicetree/bindings/arm/
A Dpmu.yaml22 - apm,potenza-pmu
24 - arm,arm1136-pmu
25 - arm,arm1176-pmu
27 - arm,cortex-a5-pmu
28 - arm,cortex-a7-pmu
29 - arm,cortex-a8-pmu
30 - arm,cortex-a9-pmu
31 - arm,cortex-a12-pmu
32 - arm,cortex-a15-pmu
49 - brcm,vulcan-pmu
[all …]
/linux/arch/arm64/kvm/
A Dpmu-emul.c166 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_get_counter_value() local
252 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_vcpu_init() local
266 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_vcpu_reset() local
283 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_vcpu_destroy() local
311 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_enable_counter_mask() local
346 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_disable_counter_mask() local
383 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_update_state() local
397 pmu->irq_num, overflow, pmu); in kvm_pmu_update_state()
404 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_should_notify_user() local
516 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_software_increment() local
[all …]
/linux/drivers/soc/rockchip/
A Dpm_domains.c144 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_is_idle() local
156 regmap_read(pmu->regmap, pmu->info->ack_offset, &val); in rockchip_pmu_read_ack()
165 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_set_idle_request() local
174 regmap_write(pmu->regmap, pmu->info->req_offset, in rockchip_pmu_set_idle_request()
258 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pmu_domain_is_on() local
274 struct rockchip_pmu *pmu = pd->pmu; in rockchip_do_pmu_set_power_domain() local
281 regmap_write(pmu->regmap, pmu->info->pwr_offset, in rockchip_do_pmu_set_power_domain()
301 struct rockchip_pmu *pmu = pd->pmu; in rockchip_pd_power() local
425 pd->pmu = pmu; in rockchip_pm_add_one_domain()
657 if (!pmu) in rockchip_pm_domain_probe()
[all …]
/linux/drivers/regulator/
A Dbcm590xx-regulator.c278 struct bcm590xx_reg *pmu; in bcm590xx_probe() local
284 pmu = devm_kzalloc(&pdev->dev, sizeof(*pmu), GFP_KERNEL); in bcm590xx_probe()
285 if (!pmu) in bcm590xx_probe()
288 pmu->mfd = bcm590xx; in bcm590xx_probe()
290 platform_set_drvdata(pdev, pmu); in bcm590xx_probe()
292 pmu->desc = devm_kcalloc(&pdev->dev, in bcm590xx_probe()
296 if (!pmu->desc) in bcm590xx_probe()
303 pmu->desc[i].name = info->name; in bcm590xx_probe()
307 pmu->desc[i].id = i; in bcm590xx_probe()
332 pmu->desc[i].owner = THIS_MODULE; in bcm590xx_probe()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/devinit/
A Dgm200.c41 nvkm_wr32(device, 0x10a188, (pmu + i) >> 8); in pmu_code()
87 struct nvbios_pmuR pmu; in pmu_load() local
89 if (!nvbios_pmuRm(bios, type, &pmu)) in pmu_load()
95 pmu_code(init, pmu.boot_addr_pmu, pmu.boot_addr, pmu.boot_size, false); in pmu_load()
96 pmu_code(init, pmu.code_addr_pmu, pmu.code_addr, pmu.code_size, true); in pmu_load()
97 pmu_data(init, pmu.data_addr_pmu, pmu.data_addr, pmu.data_size); in pmu_load()
100 *init_addr_pmu = pmu.init_addr_pmu; in pmu_load()
101 *args_addr_pmu = pmu.args_addr_pmu; in pmu_load()
105 return pmu_exec(init, pmu.init_addr_pmu), 0; in pmu_load()
146 pmu_data(init, pmu, img, len); in gm200_devinit_post()
[all …]

Completed in 68 milliseconds

12345678910>>...23