Home
last modified time | relevance | path

Searched refs:pmu (Results 1 – 25 of 602) sorted by relevance

12345678910>>...25

/linux-6.1.9/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/
Dbase.c32 struct nvkm_pmu *pmu = device->pmu; in nvkm_pmu_fan_controlled() local
37 if (pmu && pmu->func->code.size) in nvkm_pmu_fan_controlled()
48 nvkm_pmu_pgob(struct nvkm_pmu *pmu, bool enable) in nvkm_pmu_pgob() argument
50 if (pmu && pmu->func->pgob) in nvkm_pmu_pgob()
51 pmu->func->pgob(pmu, enable); in nvkm_pmu_pgob()
57 struct nvkm_pmu *pmu = container_of(work, typeof(*pmu), recv.work); in nvkm_pmu_recv() local
58 return pmu->func->recv(pmu); in nvkm_pmu_recv()
62 nvkm_pmu_send(struct nvkm_pmu *pmu, u32 reply[2], in nvkm_pmu_send() argument
65 if (!pmu || !pmu->func->send) in nvkm_pmu_send()
67 return pmu->func->send(pmu, reply, process, message, data0, data1); in nvkm_pmu_send()
[all …]
Dgk20a.c51 gk20a_pmu_dvfs_target(struct gk20a_pmu *pmu, int *state) in gk20a_pmu_dvfs_target() argument
53 struct nvkm_clk *clk = pmu->base.subdev.device->clk; in gk20a_pmu_dvfs_target()
59 gk20a_pmu_dvfs_get_cur_state(struct gk20a_pmu *pmu, int *state) in gk20a_pmu_dvfs_get_cur_state() argument
61 struct nvkm_clk *clk = pmu->base.subdev.device->clk; in gk20a_pmu_dvfs_get_cur_state()
67 gk20a_pmu_dvfs_get_target_state(struct gk20a_pmu *pmu, in gk20a_pmu_dvfs_get_target_state() argument
70 struct gk20a_pmu_dvfs_data *data = pmu->data; in gk20a_pmu_dvfs_get_target_state()
71 struct nvkm_clk *clk = pmu->base.subdev.device->clk; in gk20a_pmu_dvfs_get_target_state()
86 nvkm_trace(&pmu->base.subdev, "cur level = %d, new level = %d\n", in gk20a_pmu_dvfs_get_target_state()
95 gk20a_pmu_dvfs_get_dev_status(struct gk20a_pmu *pmu, in gk20a_pmu_dvfs_get_dev_status() argument
98 struct nvkm_falcon *falcon = &pmu->base.falcon; in gk20a_pmu_dvfs_get_dev_status()
[all …]
Dgt215.c30 gt215_pmu_send(struct nvkm_pmu *pmu, u32 reply[2], in gt215_pmu_send() argument
33 struct nvkm_subdev *subdev = &pmu->subdev; in gt215_pmu_send()
37 mutex_lock(&pmu->send.mutex); in gt215_pmu_send()
45 mutex_unlock(&pmu->send.mutex); in gt215_pmu_send()
54 pmu->recv.message = message; in gt215_pmu_send()
55 pmu->recv.process = process; in gt215_pmu_send()
65 pmu->send.base)); in gt215_pmu_send()
77 wait_event(pmu->recv.wait, (pmu->recv.process == 0)); in gt215_pmu_send()
78 reply[0] = pmu->recv.data[0]; in gt215_pmu_send()
79 reply[1] = pmu->recv.data[1]; in gt215_pmu_send()
[all …]
DKbuild2 nvkm-y += nvkm/subdev/pmu/base.o
3 nvkm-y += nvkm/subdev/pmu/memx.o
4 nvkm-y += nvkm/subdev/pmu/gt215.o
5 nvkm-y += nvkm/subdev/pmu/gf100.o
6 nvkm-y += nvkm/subdev/pmu/gf119.o
7 nvkm-y += nvkm/subdev/pmu/gk104.o
8 nvkm-y += nvkm/subdev/pmu/gk110.o
9 nvkm-y += nvkm/subdev/pmu/gk208.o
10 nvkm-y += nvkm/subdev/pmu/gk20a.o
11 nvkm-y += nvkm/subdev/pmu/gm107.o
[all …]
/linux-6.1.9/drivers/gpu/drm/i915/
Di915_pmu.c107 static bool pmu_needs_timer(struct i915_pmu *pmu, bool gpu_active) in pmu_needs_timer() argument
109 struct drm_i915_private *i915 = container_of(pmu, typeof(*i915), pmu); in pmu_needs_timer()
117 enable = pmu->enable; in pmu_needs_timer()
170 struct i915_pmu *pmu = &i915->pmu; in get_rc6() local
181 spin_lock_irqsave(&pmu->lock, flags); in get_rc6()
184 pmu->sample[__I915_SAMPLE_RC6].cur = val; in get_rc6()
193 val = ktime_since_raw(pmu->sleep_last); in get_rc6()
194 val += pmu->sample[__I915_SAMPLE_RC6].cur; in get_rc6()
197 if (val < pmu->sample[__I915_SAMPLE_RC6_LAST_REPORTED].cur) in get_rc6()
198 val = pmu->sample[__I915_SAMPLE_RC6_LAST_REPORTED].cur; in get_rc6()
[all …]
/linux-6.1.9/arch/x86/kvm/vmx/
Dpmu_intel.c38 static void reprogram_fixed_counters(struct kvm_pmu *pmu, u64 data) in reprogram_fixed_counters() argument
41 u8 old_fixed_ctr_ctrl = pmu->fixed_ctr_ctrl; in reprogram_fixed_counters()
44 pmu->fixed_ctr_ctrl = data; in reprogram_fixed_counters()
45 for (i = 0; i < pmu->nr_arch_fixed_counters; i++) { in reprogram_fixed_counters()
52 pmc = get_fixed_pmc(pmu, MSR_CORE_PERF_FIXED_CTR0 + i); in reprogram_fixed_counters()
54 __set_bit(INTEL_PMC_IDX_FIXED + i, pmu->pmc_in_use); in reprogram_fixed_counters()
59 static struct kvm_pmc *intel_pmc_idx_to_pmc(struct kvm_pmu *pmu, int pmc_idx) in intel_pmc_idx_to_pmc() argument
62 return get_gp_pmc(pmu, MSR_P6_EVNTSEL0 + pmc_idx, in intel_pmc_idx_to_pmc()
67 return get_fixed_pmc(pmu, idx + MSR_CORE_PERF_FIXED_CTR0); in intel_pmc_idx_to_pmc()
71 static void reprogram_counters(struct kvm_pmu *pmu, u64 diff) in reprogram_counters() argument
[all …]
/linux-6.1.9/drivers/soc/dove/
Dpmu.c50 struct pmu_data *pmu = rcdev_to_pmu(rc); in pmu_reset_reset() local
54 spin_lock_irqsave(&pmu->lock, flags); in pmu_reset_reset()
55 val = readl_relaxed(pmu->pmc_base + PMC_SW_RST); in pmu_reset_reset()
56 writel_relaxed(val & ~BIT(id), pmu->pmc_base + PMC_SW_RST); in pmu_reset_reset()
57 writel_relaxed(val | BIT(id), pmu->pmc_base + PMC_SW_RST); in pmu_reset_reset()
58 spin_unlock_irqrestore(&pmu->lock, flags); in pmu_reset_reset()
65 struct pmu_data *pmu = rcdev_to_pmu(rc); in pmu_reset_assert() local
69 spin_lock_irqsave(&pmu->lock, flags); in pmu_reset_assert()
70 val &= readl_relaxed(pmu->pmc_base + PMC_SW_RST); in pmu_reset_assert()
71 writel_relaxed(val, pmu->pmc_base + PMC_SW_RST); in pmu_reset_assert()
[all …]
/linux-6.1.9/drivers/perf/
Dfsl_imx8_ddr_perf.c40 #define to_ddr_pmu(p) container_of(p, struct ddr_pmu, pmu)
94 struct pmu pmu; member
111 struct ddr_pmu *pmu = dev_get_drvdata(dev); in ddr_perf_identifier_show() local
113 return sysfs_emit(page, "%s\n", pmu->devtype_data->identifier); in ddr_perf_identifier_show()
121 struct ddr_pmu *pmu = dev_get_drvdata(dev); in ddr_perf_identifier_attr_visible() local
123 if (!pmu->devtype_data->identifier) in ddr_perf_identifier_attr_visible()
147 static u32 ddr_perf_filter_cap_get(struct ddr_pmu *pmu, int cap) in ddr_perf_filter_cap_get() argument
149 u32 quirks = pmu->devtype_data->quirks; in ddr_perf_filter_cap_get()
168 struct ddr_pmu *pmu = dev_get_drvdata(dev); in ddr_perf_filter_cap_show() local
173 return sysfs_emit(buf, "%u\n", ddr_perf_filter_cap_get(pmu, cap)); in ddr_perf_filter_cap_show()
[all …]
Dmarvell_cn10k_ddr_pmu.c125 struct pmu pmu; member
135 #define to_cn10k_ddr_pmu(p) container_of(p, struct cn10k_ddr_pmu, pmu)
233 struct cn10k_ddr_pmu *pmu = dev_get_drvdata(dev); in cn10k_ddr_perf_cpumask_show() local
235 return cpumap_print_to_pagebuf(true, buf, cpumask_of(pmu->cpu)); in cn10k_ddr_perf_cpumask_show()
289 static int cn10k_ddr_perf_alloc_counter(struct cn10k_ddr_pmu *pmu, in cn10k_ddr_perf_alloc_counter() argument
297 pmu->events[DDRC_PERF_READ_COUNTER_IDX] = event; in cn10k_ddr_perf_alloc_counter()
303 pmu->events[DDRC_PERF_WRITE_COUNTER_IDX] = event; in cn10k_ddr_perf_alloc_counter()
309 if (pmu->events[i] == NULL) { in cn10k_ddr_perf_alloc_counter()
310 pmu->events[i] = event; in cn10k_ddr_perf_alloc_counter()
318 static void cn10k_ddr_perf_free_counter(struct cn10k_ddr_pmu *pmu, int counter) in cn10k_ddr_perf_free_counter() argument
[all …]
Darm_pmu_platform.c26 static int probe_current_pmu(struct arm_pmu *pmu, in probe_current_pmu() argument
38 ret = info->init(pmu); in probe_current_pmu()
46 static int pmu_parse_percpu_irq(struct arm_pmu *pmu, int irq) in pmu_parse_percpu_irq() argument
49 struct pmu_hw_events __percpu *hw_events = pmu->hw_events; in pmu_parse_percpu_irq()
51 ret = irq_get_percpu_devid_partition(irq, &pmu->supported_cpus); in pmu_parse_percpu_irq()
55 for_each_cpu(cpu, &pmu->supported_cpus) in pmu_parse_percpu_irq()
96 static int pmu_parse_irqs(struct arm_pmu *pmu) in pmu_parse_irqs() argument
99 struct platform_device *pdev = pmu->plat_device; in pmu_parse_irqs()
100 struct pmu_hw_events __percpu *hw_events = pmu->hw_events; in pmu_parse_irqs()
113 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in pmu_parse_irqs()
[all …]
Darm_pmu_acpi.c193 struct arm_pmu *pmu; in arm_pmu_acpi_find_alloc_pmu() local
197 pmu = per_cpu(probed_pmus, cpu); in arm_pmu_acpi_find_alloc_pmu()
198 if (!pmu || pmu->acpi_cpuid != cpuid) in arm_pmu_acpi_find_alloc_pmu()
201 return pmu; in arm_pmu_acpi_find_alloc_pmu()
204 pmu = armpmu_alloc_atomic(); in arm_pmu_acpi_find_alloc_pmu()
205 if (!pmu) { in arm_pmu_acpi_find_alloc_pmu()
211 pmu->acpi_cpuid = cpuid; in arm_pmu_acpi_find_alloc_pmu()
213 return pmu; in arm_pmu_acpi_find_alloc_pmu()
220 static bool pmu_irq_matches(struct arm_pmu *pmu, int irq) in pmu_irq_matches() argument
222 struct pmu_hw_events __percpu *hw_events = pmu->hw_events; in pmu_irq_matches()
[all …]
Darm_pmu.c183 if (type == event->pmu->type) in armpmu_map_event()
200 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_set_period()
242 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_event_update()
272 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_stop()
288 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_start()
313 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_del()
329 struct arm_pmu *armpmu = to_arm_pmu(event->pmu); in armpmu_add()
362 validate_event(struct pmu *pmu, struct pmu_hw_events *hw_events, in validate_event() argument
375 if (event->pmu != pmu) in validate_event()
384 armpmu = to_arm_pmu(event->pmu); in validate_event()
[all …]
/linux-6.1.9/arch/x86/kvm/svm/
Dpmu.c26 static struct kvm_pmc *amd_pmc_idx_to_pmc(struct kvm_pmu *pmu, int pmc_idx) in amd_pmc_idx_to_pmc() argument
28 unsigned int num_counters = pmu->nr_arch_gp_counters; in amd_pmc_idx_to_pmc()
33 return &pmu->gp_counters[array_index_nospec(pmc_idx, num_counters)]; in amd_pmc_idx_to_pmc()
36 static inline struct kvm_pmc *get_gp_pmc_amd(struct kvm_pmu *pmu, u32 msr, in get_gp_pmc_amd() argument
39 struct kvm_vcpu *vcpu = pmu_to_vcpu(pmu); in get_gp_pmc_amd()
71 return amd_pmc_idx_to_pmc(pmu, idx); in get_gp_pmc_amd()
89 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_is_valid_rdpmc_ecx() local
93 return idx < pmu->nr_arch_gp_counters; in amd_is_valid_rdpmc_ecx()
111 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_msr_idx_to_pmc() local
114 pmc = get_gp_pmc_amd(pmu, msr, PMU_TYPE_COUNTER); in amd_msr_idx_to_pmc()
[all …]
/linux-6.1.9/Documentation/devicetree/bindings/arm/
Dpmu.yaml4 $id: http://devicetree.org/schemas/arm/pmu.yaml#
22 - apm,potenza-pmu
23 - apple,firestorm-pmu
24 - apple,icestorm-pmu
26 - arm,arm1136-pmu
27 - arm,arm1176-pmu
28 - arm,arm11mpcore-pmu
29 - arm,cortex-a5-pmu
30 - arm,cortex-a7-pmu
31 - arm,cortex-a8-pmu
[all …]
/linux-6.1.9/Documentation/devicetree/bindings/soc/samsung/
Dexynos-pmu.yaml4 $id: http://devicetree.org/schemas/soc/samsung/exynos-pmu.yaml#
18 - samsung,exynos3250-pmu
19 - samsung,exynos4210-pmu
20 - samsung,exynos4412-pmu
21 - samsung,exynos5250-pmu
22 - samsung,exynos5260-pmu
23 - samsung,exynos5410-pmu
24 - samsung,exynos5420-pmu
25 - samsung,exynos5433-pmu
26 - samsung,exynos7-pmu
[all …]
/linux-6.1.9/arch/x86/events/intel/
Duncore.h86 struct pmu *pmu; /* for custom pmu ops */ member
120 struct pmu pmu; member
149 struct intel_uncore_pmu *pmu; member
202 return container_of(dev_get_drvdata(dev), struct intel_uncore_pmu, pmu); in dev_to_uncore_pmu()
242 if (offset < box->pmu->type->mmio_map_size) in uncore_mmio_is_valid_offset()
246 offset, box->pmu->type->name); in uncore_mmio_is_valid_offset()
254 return box->pmu->type->box_ctl + in uncore_mmio_box_ctl()
255 box->pmu->type->mmio_offset * box->pmu->pmu_idx; in uncore_mmio_box_ctl()
260 return box->pmu->type->box_ctl; in uncore_pci_box_ctl()
265 return box->pmu->type->fixed_ctl; in uncore_pci_fixed_ctl()
[all …]
Duncore.c122 struct intel_uncore_box *uncore_pmu_to_box(struct intel_uncore_pmu *pmu, int cpu) in uncore_pmu_to_box() argument
130 return dieid < uncore_max_dies() ? pmu->boxes[dieid] : NULL; in uncore_pmu_to_box()
370 return &box->pmu->pmu == event->pmu; in is_box_event()
380 max_count = box->pmu->type->num_counters; in uncore_collect_events()
381 if (box->pmu->type->fixed_ctl) in uncore_collect_events()
414 struct intel_uncore_type *type = box->pmu->type; in uncore_get_event_constraint()
439 if (box->pmu->type->ops->put_constraint) in uncore_put_event_constraint()
440 box->pmu->type->ops->put_constraint(box, event); in uncore_put_event_constraint()
678 static int uncore_validate_group(struct intel_uncore_pmu *pmu, in uncore_validate_group() argument
689 fake_box = uncore_alloc_box(pmu->type, NUMA_NO_NODE); in uncore_validate_group()
[all …]
/linux-6.1.9/arch/x86/events/
Drapl.c110 struct pmu *pmu; member
116 struct pmu pmu; member
210 static void rapl_start_hrtimer(struct rapl_pmu *pmu) in rapl_start_hrtimer() argument
212 hrtimer_start(&pmu->hrtimer, pmu->timer_interval, in rapl_start_hrtimer()
218 struct rapl_pmu *pmu = container_of(hrtimer, struct rapl_pmu, hrtimer); in rapl_hrtimer_handle() local
222 if (!pmu->n_active) in rapl_hrtimer_handle()
225 raw_spin_lock_irqsave(&pmu->lock, flags); in rapl_hrtimer_handle()
227 list_for_each_entry(event, &pmu->active_list, active_entry) in rapl_hrtimer_handle()
230 raw_spin_unlock_irqrestore(&pmu->lock, flags); in rapl_hrtimer_handle()
232 hrtimer_forward_now(hrtimer, pmu->timer_interval); in rapl_hrtimer_handle()
[all …]
/linux-6.1.9/tools/perf/util/
Dpmu.c337 pmu_name = (char *)pe->pmu; in __perf_pmu__new_alias()
693 char *perf_pmu__getcpuid(struct perf_pmu *pmu) in perf_pmu__getcpuid() argument
702 cpuid = get_cpuid_str(pmu); in perf_pmu__getcpuid()
798 struct perf_pmu *pmu; member
806 const char *pname = pe->pmu ? pe->pmu : data->cpu_name; in pmu_add_cpu_aliases_map_callback()
811 if (data->pmu->is_uncore && pmu_uncore_alias_match(pname, data->name)) in pmu_add_cpu_aliases_map_callback()
829 void pmu_add_cpu_aliases_table(struct list_head *head, struct perf_pmu *pmu, in pmu_add_cpu_aliases_table() argument
834 .name = pmu->name, in pmu_add_cpu_aliases_table()
835 .cpu_name = is_arm_pmu_core(pmu->name) ? pmu->name : "cpu", in pmu_add_cpu_aliases_table()
836 .pmu = pmu, in pmu_add_cpu_aliases_table()
[all …]
Dparse-events-hybrid.c45 struct perf_pmu *pmu) in create_event_hybrid() argument
51 config_hybrid_attr(attr, config_type, pmu->type); in create_event_hybrid()
64 pmu, config_terms); in create_event_hybrid()
66 evsel->pmu_name = strdup(pmu->name); in create_event_hybrid()
77 struct perf_pmu *pmu) in pmu_cmp() argument
80 return strcmp(parse_state->evlist->hybrid_pmu_name, pmu->name); in pmu_cmp()
83 return strcmp(parse_state->hybrid_pmu_name, pmu->name); in pmu_cmp()
93 struct perf_pmu *pmu; in add_hw_hybrid() local
96 perf_pmu__for_each_hybrid_pmu(pmu) { in add_hw_hybrid()
99 if (pmu_cmp(parse_state, pmu)) in add_hw_hybrid()
[all …]
/linux-6.1.9/Documentation/devicetree/bindings/arm/rockchip/
Dpmu.yaml4 $id: http://devicetree.org/schemas/arm/rockchip/pmu.yaml#
22 - rockchip,px30-pmu
23 - rockchip,rk3066-pmu
24 - rockchip,rk3128-pmu
25 - rockchip,rk3288-pmu
26 - rockchip,rk3368-pmu
27 - rockchip,rk3399-pmu
28 - rockchip,rk3568-pmu
29 - rockchip,rk3588-pmu
38 - rockchip,px30-pmu
[all …]
/linux-6.1.9/Documentation/devicetree/bindings/pinctrl/
Dmarvell,dove-pinctrl.txt14 Note: pmu* also allows for Power Management functions listed below
18 mpp0 0 gpio, pmu, uart2(rts), sdio0(cd), lcd0(pwm), pmu*
19 mpp1 1 gpio, pmu, uart2(cts), sdio0(wp), lcd1(pwm), pmu*
20 mpp2 2 gpio, pmu, uart2(txd), sdio0(buspwr), sata(prsnt),
21 uart1(rts), pmu*
22 mpp3 3 gpio, pmu, uart2(rxd), sdio0(ledctrl), sata(act),
23 uart1(cts), lcd-spi(cs1), pmu*
24 mpp4 4 gpio, pmu, uart3(rts), sdio1(cd), spi1(miso), pmu*
25 mpp5 5 gpio, pmu, uart3(cts), sdio1(wp), spi1(cs), pmu*
26 mpp6 6 gpio, pmu, uart3(txd), sdio1(buspwr), spi1(mosi), pmu*
[all …]
/linux-6.1.9/arch/powerpc/perf/
Dimc-pmu.c61 return container_of(event->pmu, struct imc_pmu, pmu); in imc_event_to_pmu()
105 struct pmu *pmu = dev_get_drvdata(dev); in imc_pmu_cpumask_get_attr() local
106 struct imc_pmu *imc_pmu = container_of(pmu, struct imc_pmu, pmu); in imc_pmu_cpumask_get_attr()
221 static int update_events_in_group(struct device_node *node, struct imc_pmu *pmu) in update_events_in_group() argument
260 pmu->events = kcalloc(ct, sizeof(struct imc_events), GFP_KERNEL); in update_events_in_group()
261 if (!pmu->events) { in update_events_in_group()
269 ret = imc_parse_event(np, g_scale, g_unit, prefix, base_reg, &pmu->events[ct]); in update_events_in_group()
279 imc_free_events(pmu->events, ct); in update_events_in_group()
294 imc_free_events(pmu->events, ct); in update_events_in_group()
301 ev_val_str = kasprintf(GFP_KERNEL, "event=0x%x", pmu->events[i].value); in update_events_in_group()
[all …]
DMakefile7 obj64-$(CONFIG_PPC_PERF_CTRS) += ppc970-pmu.o power5-pmu.o \
8 power5+-pmu.o power6-pmu.o power7-pmu.o \
9 isa207-common.o power8-pmu.o power9-pmu.o \
10 generic-compat-pmu.o power10-pmu.o bhrb.o
11 obj32-$(CONFIG_PPC_PERF_CTRS) += mpc7450-pmu.o
13 obj-$(CONFIG_PPC_POWERNV) += imc-pmu.o
15 obj-$(CONFIG_FSL_EMB_PERF_EVENT_E500) += e500-pmu.o e6500-pmu.o
19 obj-$(CONFIG_PPC_8xx) += 8xx-pmu.o
/linux-6.1.9/arch/arm64/kvm/
Dpmu-emul.c62 struct kvm_pmu *pmu; in kvm_pmc_to_vcpu() local
66 pmu = container_of(pmc, struct kvm_pmu, pmc[0]); in kvm_pmc_to_vcpu()
67 vcpu_arch = container_of(pmu, struct kvm_vcpu_arch, pmu); in kvm_pmc_to_vcpu()
79 return test_bit(pmc->idx >> 1, vcpu->arch.pmu.chained); in kvm_pmu_pmc_is_chained()
177 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_get_counter_value() local
178 struct kvm_pmc *pmc = &pmu->pmc[select_idx]; in kvm_pmu_get_counter_value()
269 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_vcpu_init() local
272 pmu->pmc[i].idx = i; in kvm_pmu_vcpu_init()
283 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_vcpu_reset() local
287 kvm_pmu_stop_counter(vcpu, &pmu->pmc[i]); in kvm_pmu_vcpu_reset()
[all …]

12345678910>>...25