Searched refs:hybrid_pmu (Results 1 – 3 of 3) sorted by relevance
755 if (event->pmu == &x86_pmu.hybrid_pmu[i].pmu) in is_x86_event()2171 struct x86_hybrid_pmu *hybrid_pmu; in init_hw_perf_events() local2178 hybrid_pmu = &x86_pmu.hybrid_pmu[i]; in init_hw_perf_events()2180 hybrid_pmu->pmu = pmu; in init_hw_perf_events()2181 hybrid_pmu->pmu.type = -1; in init_hw_perf_events()2186 err = perf_pmu_register(&hybrid_pmu->pmu, hybrid_pmu->name, in init_hw_perf_events()2191 if (cpu_type == hybrid_pmu->cpu_type) in init_hw_perf_events()2199 kfree(x86_pmu.hybrid_pmu); in init_hw_perf_events()2200 x86_pmu.hybrid_pmu = NULL; in init_hw_perf_events()2342 h_pmu = hybrid_pmu(event_pmu); in allocate_fake_cpuc()[all …]
666 static __always_inline struct x86_hybrid_pmu *hybrid_pmu(struct pmu *pmu) in hybrid_pmu() function679 __Fp = &hybrid_pmu(_pmu)->_field; \689 __Fp = &hybrid_pmu(_pmu)->_var; \699 __Fp = hybrid_pmu(_pmu)->_field; \915 struct x86_hybrid_pmu *hybrid_pmu; member
2494 struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu); in adl_set_topdown_event_period()2644 struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu); in adl_update_topdown_event()4167 struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu); in adl_get_event_constraints()4180 struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu); in adl_hw_config()4364 if (x86_pmu.hybrid_pmu[i].cpu_type == cpu_type) { in init_hybrid_pmu()4365 pmu = &x86_pmu.hybrid_pmu[i]; in init_hybrid_pmu()4566 struct x86_hybrid_pmu *pmu = hybrid_pmu(event->pmu); in intel_pmu_filter_match()5543 pmu = &x86_pmu.hybrid_pmu[i]; in intel_pmu_check_hybrid_pmus()6189 x86_pmu.hybrid_pmu = kcalloc(X86_HYBRID_NUM_PMUS, in intel_pmu_init()6192 if (!x86_pmu.hybrid_pmu) in intel_pmu_init()[all …]
Completed in 26 milliseconds