1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * Performance counter support for POWER10 processors.
5 * Copyright 2020 Madhavan Srinivasan, IBM Corporation.
6 * Copyright 2020 Athira Rajeev, IBM Corporation.
9 #define pr_fmt(fmt) "power10-pmu: " fmt
11 #include "isa207-common.h"
14 * Raw event encoding for Power10:
16 * 60 56 52 48 44 40 36 32
17 * | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - |
18 * | | [ ] [ src_match ] [ src_mask ] | [ ] [ l2l3_sel ] [ thresh_ctl ]
20 * | | *- IFM (Linux) | | thresh start/stop -*
21 * | *- BHRB (Linux) | src_sel
22 * *- EBB (Linux) *invert_bit
24 * 28 24 20 16 12 8 4 0
25 * | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - | - - - - |
26 * [ ] [ sample ] [ ] [ ] [ pmc ] [unit ] [ ] | m [ pmcxsel ]
29 * | | | *- L1/L2/L3 cache_sel | |*-radix_scope_qual
31 * | *- sampling mode for marked events *- combine
35 * Below uses IBM bit numbering.
37 * MMCR1[x:y] = unit (PMCxUNIT)
38 * MMCR1[24] = pmc1combine[0]
39 * MMCR1[25] = pmc1combine[1]
40 * MMCR1[26] = pmc2combine[0]
41 * MMCR1[27] = pmc2combine[1]
42 * MMCR1[28] = pmc3combine[0]
43 * MMCR1[29] = pmc3combine[1]
44 * MMCR1[30] = pmc4combine[0]
45 * MMCR1[31] = pmc4combine[1]
47 * if pmc == 3 and unit == 0 and pmcxsel[0:6] == 0b0101011
48 * MMCR1[20:27] = thresh_ctl
49 * else if pmc == 4 and unit == 0xf and pmcxsel[0:6] == 0b0101001
50 * MMCR1[20:27] = thresh_ctl
52 * MMCRA[48:55] = thresh_ctl (THRESH START/END)
55 * MMCRA[45:47] = thresh_sel
58 * MMCR2[56:60] = l2l3_sel[0:4]
60 * MMCR1[16] = cache_sel[0]
61 * MMCR1[17] = cache_sel[1]
62 * MMCR1[18] = radix_scope_qual
65 * MMCRA[63] = 1 (SAMPLE_ENABLE)
66 * MMCRA[57:59] = sample[0:2] (RAND_SAMP_ELIG)
67 * MMCRA[61:62] = sample[3:4] (RAND_SAMP_MODE)
72 * MMCRA[SDAR_MODE] = sdar_mode[0:1]
76 * Some power10 event codes.
78 #define EVENT(_name, _code) enum{_name = _code}
80 #include "power10-events-list.h"
84 /* MMCRA IFM bits - POWER10 */
85 #define POWER10_MMCRA_IFM1 0x0000000040000000UL
86 #define POWER10_MMCRA_IFM2 0x0000000080000000UL
87 #define POWER10_MMCRA_IFM3 0x00000000C0000000UL
88 #define POWER10_MMCRA_BHRB_MASK 0x00000000C0000000UL
90 extern u64 PERF_REG_EXTENDED_MASK
;
92 /* Table of alternatives, sorted by column 0 */
93 static const unsigned int power10_event_alternatives
[][MAX_ALT
] = {
94 { PM_CYC_ALT
, PM_CYC
},
95 { PM_INST_CMPL_ALT
, PM_INST_CMPL
},
98 static int power10_get_alternatives(u64 event
, unsigned int flags
, u64 alt
[])
102 num_alt
= isa207_get_alternatives(event
, alt
,
103 ARRAY_SIZE(power10_event_alternatives
), flags
,
104 power10_event_alternatives
);
109 static int power10_check_attr_config(struct perf_event
*ev
)
112 u64 event
= ev
->attr
.config
;
114 val
= (event
>> EVENT_SAMPLE_SHIFT
) & EVENT_SAMPLE_MASK
;
115 if (val
== 0x10 || isa3XX_check_attr_config(ev
))
121 GENERIC_EVENT_ATTR(cpu
-cycles
, PM_CYC
);
122 GENERIC_EVENT_ATTR(instructions
, PM_INST_CMPL
);
123 GENERIC_EVENT_ATTR(branch
-instructions
, PM_BR_CMPL
);
124 GENERIC_EVENT_ATTR(branch
-misses
, PM_BR_MPRED_CMPL
);
125 GENERIC_EVENT_ATTR(cache
-references
, PM_LD_REF_L1
);
126 GENERIC_EVENT_ATTR(cache
-misses
, PM_LD_MISS_L1
);
127 GENERIC_EVENT_ATTR(mem
-loads
, MEM_LOADS
);
128 GENERIC_EVENT_ATTR(mem
-stores
, MEM_STORES
);
129 GENERIC_EVENT_ATTR(branch
-instructions
, PM_BR_FIN
);
130 GENERIC_EVENT_ATTR(branch
-misses
, PM_MPRED_BR_FIN
);
131 GENERIC_EVENT_ATTR(cache
-misses
, PM_LD_DEMAND_MISS_L1_FIN
);
133 CACHE_EVENT_ATTR(L1
-dcache
-load
-misses
, PM_LD_MISS_L1
);
134 CACHE_EVENT_ATTR(L1
-dcache
-loads
, PM_LD_REF_L1
);
135 CACHE_EVENT_ATTR(L1
-dcache
-prefetches
, PM_LD_PREFETCH_CACHE_LINE_MISS
);
136 CACHE_EVENT_ATTR(L1
-dcache
-store
-misses
, PM_ST_MISS_L1
);
137 CACHE_EVENT_ATTR(L1
-icache
-load
-misses
, PM_L1_ICACHE_MISS
);
138 CACHE_EVENT_ATTR(L1
-icache
-loads
, PM_INST_FROM_L1
);
139 CACHE_EVENT_ATTR(L1
-icache
-prefetches
, PM_IC_PREF_REQ
);
140 CACHE_EVENT_ATTR(LLC
-load
-misses
, PM_DATA_FROM_L3MISS
);
141 CACHE_EVENT_ATTR(LLC
-loads
, PM_DATA_FROM_L3
);
142 CACHE_EVENT_ATTR(LLC
-prefetches
, PM_L3_PF_MISS_L3
);
143 CACHE_EVENT_ATTR(LLC
-store
-misses
, PM_L2_ST_MISS
);
144 CACHE_EVENT_ATTR(LLC
-stores
, PM_L2_ST
);
145 CACHE_EVENT_ATTR(branch
-load
-misses
, PM_BR_MPRED_CMPL
);
146 CACHE_EVENT_ATTR(branch
-loads
, PM_BR_CMPL
);
147 CACHE_EVENT_ATTR(dTLB
-load
-misses
, PM_DTLB_MISS
);
148 CACHE_EVENT_ATTR(iTLB
-load
-misses
, PM_ITLB_MISS
);
150 static struct attribute
*power10_events_attr_dd1
[] = {
151 GENERIC_EVENT_PTR(PM_CYC
),
152 GENERIC_EVENT_PTR(PM_INST_CMPL
),
153 GENERIC_EVENT_PTR(PM_BR_CMPL
),
154 GENERIC_EVENT_PTR(PM_BR_MPRED_CMPL
),
155 GENERIC_EVENT_PTR(PM_LD_REF_L1
),
156 GENERIC_EVENT_PTR(PM_LD_MISS_L1
),
157 GENERIC_EVENT_PTR(MEM_LOADS
),
158 GENERIC_EVENT_PTR(MEM_STORES
),
159 CACHE_EVENT_PTR(PM_LD_MISS_L1
),
160 CACHE_EVENT_PTR(PM_LD_REF_L1
),
161 CACHE_EVENT_PTR(PM_LD_PREFETCH_CACHE_LINE_MISS
),
162 CACHE_EVENT_PTR(PM_ST_MISS_L1
),
163 CACHE_EVENT_PTR(PM_L1_ICACHE_MISS
),
164 CACHE_EVENT_PTR(PM_INST_FROM_L1
),
165 CACHE_EVENT_PTR(PM_IC_PREF_REQ
),
166 CACHE_EVENT_PTR(PM_DATA_FROM_L3MISS
),
167 CACHE_EVENT_PTR(PM_DATA_FROM_L3
),
168 CACHE_EVENT_PTR(PM_BR_MPRED_CMPL
),
169 CACHE_EVENT_PTR(PM_BR_CMPL
),
170 CACHE_EVENT_PTR(PM_DTLB_MISS
),
171 CACHE_EVENT_PTR(PM_ITLB_MISS
),
175 static struct attribute
*power10_events_attr
[] = {
176 GENERIC_EVENT_PTR(PM_CYC
),
177 GENERIC_EVENT_PTR(PM_INST_CMPL
),
178 GENERIC_EVENT_PTR(PM_BR_FIN
),
179 GENERIC_EVENT_PTR(PM_MPRED_BR_FIN
),
180 GENERIC_EVENT_PTR(PM_LD_REF_L1
),
181 GENERIC_EVENT_PTR(PM_LD_DEMAND_MISS_L1_FIN
),
182 GENERIC_EVENT_PTR(MEM_LOADS
),
183 GENERIC_EVENT_PTR(MEM_STORES
),
184 CACHE_EVENT_PTR(PM_LD_MISS_L1
),
185 CACHE_EVENT_PTR(PM_LD_REF_L1
),
186 CACHE_EVENT_PTR(PM_LD_PREFETCH_CACHE_LINE_MISS
),
187 CACHE_EVENT_PTR(PM_ST_MISS_L1
),
188 CACHE_EVENT_PTR(PM_L1_ICACHE_MISS
),
189 CACHE_EVENT_PTR(PM_INST_FROM_L1
),
190 CACHE_EVENT_PTR(PM_IC_PREF_REQ
),
191 CACHE_EVENT_PTR(PM_DATA_FROM_L3MISS
),
192 CACHE_EVENT_PTR(PM_DATA_FROM_L3
),
193 CACHE_EVENT_PTR(PM_L3_PF_MISS_L3
),
194 CACHE_EVENT_PTR(PM_L2_ST_MISS
),
195 CACHE_EVENT_PTR(PM_L2_ST
),
196 CACHE_EVENT_PTR(PM_BR_MPRED_CMPL
),
197 CACHE_EVENT_PTR(PM_BR_CMPL
),
198 CACHE_EVENT_PTR(PM_DTLB_MISS
),
199 CACHE_EVENT_PTR(PM_ITLB_MISS
),
203 static struct attribute_group power10_pmu_events_group_dd1
= {
205 .attrs
= power10_events_attr_dd1
,
208 static struct attribute_group power10_pmu_events_group
= {
210 .attrs
= power10_events_attr
,
213 PMU_FORMAT_ATTR(event
, "config:0-59");
214 PMU_FORMAT_ATTR(pmcxsel
, "config:0-7");
215 PMU_FORMAT_ATTR(mark
, "config:8");
216 PMU_FORMAT_ATTR(combine
, "config:10-11");
217 PMU_FORMAT_ATTR(unit
, "config:12-15");
218 PMU_FORMAT_ATTR(pmc
, "config:16-19");
219 PMU_FORMAT_ATTR(cache_sel
, "config:20-21");
220 PMU_FORMAT_ATTR(sdar_mode
, "config:22-23");
221 PMU_FORMAT_ATTR(sample_mode
, "config:24-28");
222 PMU_FORMAT_ATTR(thresh_sel
, "config:29-31");
223 PMU_FORMAT_ATTR(thresh_stop
, "config:32-35");
224 PMU_FORMAT_ATTR(thresh_start
, "config:36-39");
225 PMU_FORMAT_ATTR(l2l3_sel
, "config:40-44");
226 PMU_FORMAT_ATTR(src_sel
, "config:45-46");
227 PMU_FORMAT_ATTR(invert_bit
, "config:47");
228 PMU_FORMAT_ATTR(src_mask
, "config:48-53");
229 PMU_FORMAT_ATTR(src_match
, "config:54-59");
230 PMU_FORMAT_ATTR(radix_scope
, "config:9");
231 PMU_FORMAT_ATTR(thresh_cmp
, "config1:0-17");
233 static struct attribute
*power10_pmu_format_attr
[] = {
234 &format_attr_event
.attr
,
235 &format_attr_pmcxsel
.attr
,
236 &format_attr_mark
.attr
,
237 &format_attr_combine
.attr
,
238 &format_attr_unit
.attr
,
239 &format_attr_pmc
.attr
,
240 &format_attr_cache_sel
.attr
,
241 &format_attr_sdar_mode
.attr
,
242 &format_attr_sample_mode
.attr
,
243 &format_attr_thresh_sel
.attr
,
244 &format_attr_thresh_stop
.attr
,
245 &format_attr_thresh_start
.attr
,
246 &format_attr_l2l3_sel
.attr
,
247 &format_attr_src_sel
.attr
,
248 &format_attr_invert_bit
.attr
,
249 &format_attr_src_mask
.attr
,
250 &format_attr_src_match
.attr
,
251 &format_attr_radix_scope
.attr
,
252 &format_attr_thresh_cmp
.attr
,
256 static struct attribute_group power10_pmu_format_group
= {
258 .attrs
= power10_pmu_format_attr
,
261 static const struct attribute_group
*power10_pmu_attr_groups_dd1
[] = {
262 &power10_pmu_format_group
,
263 &power10_pmu_events_group_dd1
,
267 static const struct attribute_group
*power10_pmu_attr_groups
[] = {
268 &power10_pmu_format_group
,
269 &power10_pmu_events_group
,
273 static int power10_generic_events_dd1
[] = {
274 [PERF_COUNT_HW_CPU_CYCLES
] = PM_CYC
,
275 [PERF_COUNT_HW_INSTRUCTIONS
] = PM_INST_CMPL
,
276 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS
] = PM_BR_CMPL
,
277 [PERF_COUNT_HW_BRANCH_MISSES
] = PM_BR_MPRED_CMPL
,
278 [PERF_COUNT_HW_CACHE_REFERENCES
] = PM_LD_REF_L1
,
279 [PERF_COUNT_HW_CACHE_MISSES
] = PM_LD_MISS_L1
,
282 static int power10_generic_events
[] = {
283 [PERF_COUNT_HW_CPU_CYCLES
] = PM_CYC
,
284 [PERF_COUNT_HW_INSTRUCTIONS
] = PM_INST_CMPL
,
285 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS
] = PM_BR_FIN
,
286 [PERF_COUNT_HW_BRANCH_MISSES
] = PM_MPRED_BR_FIN
,
287 [PERF_COUNT_HW_CACHE_REFERENCES
] = PM_LD_REF_L1
,
288 [PERF_COUNT_HW_CACHE_MISSES
] = PM_LD_DEMAND_MISS_L1_FIN
,
291 static u64
power10_bhrb_filter_map(u64 branch_sample_type
)
293 u64 pmu_bhrb_filter
= 0;
295 /* BHRB and regular PMU events share the same privilege state
296 * filter configuration. BHRB is always recorded along with a
297 * regular PMU event. As the privilege state filter is handled
298 * in the basic PMC configuration of the accompanying regular
299 * PMU event, we ignore any separate BHRB specific request.
302 /* No branch filter requested */
303 if (branch_sample_type
& PERF_SAMPLE_BRANCH_ANY
)
304 return pmu_bhrb_filter
;
306 /* Invalid branch filter options - HW does not support */
307 if (branch_sample_type
& PERF_SAMPLE_BRANCH_ANY_RETURN
)
310 if (branch_sample_type
& PERF_SAMPLE_BRANCH_IND_CALL
) {
311 pmu_bhrb_filter
|= POWER10_MMCRA_IFM2
;
312 return pmu_bhrb_filter
;
315 if (branch_sample_type
& PERF_SAMPLE_BRANCH_COND
) {
316 pmu_bhrb_filter
|= POWER10_MMCRA_IFM3
;
317 return pmu_bhrb_filter
;
320 if (branch_sample_type
& PERF_SAMPLE_BRANCH_CALL
)
323 if (branch_sample_type
& PERF_SAMPLE_BRANCH_ANY_CALL
) {
324 pmu_bhrb_filter
|= POWER10_MMCRA_IFM1
;
325 return pmu_bhrb_filter
;
328 /* Every thing else is unsupported */
332 static void power10_config_bhrb(u64 pmu_bhrb_filter
)
334 pmu_bhrb_filter
&= POWER10_MMCRA_BHRB_MASK
;
336 /* Enable BHRB filter in PMU */
337 mtspr(SPRN_MMCRA
, (mfspr(SPRN_MMCRA
) | pmu_bhrb_filter
));
340 #define C(x) PERF_COUNT_HW_CACHE_##x
343 * Table of generalized cache-related events.
344 * 0 means not supported, -1 means nonsensical, other values
347 static u64 power10_cache_events_dd1
[C(MAX
)][C(OP_MAX
)][C(RESULT_MAX
)] = {
350 [C(RESULT_ACCESS
)] = PM_LD_REF_L1
,
351 [C(RESULT_MISS
)] = PM_LD_MISS_L1
,
354 [C(RESULT_ACCESS
)] = 0,
355 [C(RESULT_MISS
)] = PM_ST_MISS_L1
,
358 [C(RESULT_ACCESS
)] = PM_LD_PREFETCH_CACHE_LINE_MISS
,
359 [C(RESULT_MISS
)] = 0,
364 [C(RESULT_ACCESS
)] = PM_INST_FROM_L1
,
365 [C(RESULT_MISS
)] = PM_L1_ICACHE_MISS
,
368 [C(RESULT_ACCESS
)] = PM_INST_FROM_L1MISS
,
369 [C(RESULT_MISS
)] = -1,
372 [C(RESULT_ACCESS
)] = PM_IC_PREF_REQ
,
373 [C(RESULT_MISS
)] = 0,
378 [C(RESULT_ACCESS
)] = PM_DATA_FROM_L3
,
379 [C(RESULT_MISS
)] = PM_DATA_FROM_L3MISS
,
382 [C(RESULT_ACCESS
)] = -1,
383 [C(RESULT_MISS
)] = -1,
386 [C(RESULT_ACCESS
)] = -1,
387 [C(RESULT_MISS
)] = 0,
392 [C(RESULT_ACCESS
)] = 0,
393 [C(RESULT_MISS
)] = PM_DTLB_MISS
,
396 [C(RESULT_ACCESS
)] = -1,
397 [C(RESULT_MISS
)] = -1,
400 [C(RESULT_ACCESS
)] = -1,
401 [C(RESULT_MISS
)] = -1,
406 [C(RESULT_ACCESS
)] = 0,
407 [C(RESULT_MISS
)] = PM_ITLB_MISS
,
410 [C(RESULT_ACCESS
)] = -1,
411 [C(RESULT_MISS
)] = -1,
414 [C(RESULT_ACCESS
)] = -1,
415 [C(RESULT_MISS
)] = -1,
420 [C(RESULT_ACCESS
)] = PM_BR_CMPL
,
421 [C(RESULT_MISS
)] = PM_BR_MPRED_CMPL
,
424 [C(RESULT_ACCESS
)] = -1,
425 [C(RESULT_MISS
)] = -1,
428 [C(RESULT_ACCESS
)] = -1,
429 [C(RESULT_MISS
)] = -1,
434 [C(RESULT_ACCESS
)] = -1,
435 [C(RESULT_MISS
)] = -1,
438 [C(RESULT_ACCESS
)] = -1,
439 [C(RESULT_MISS
)] = -1,
442 [C(RESULT_ACCESS
)] = -1,
443 [C(RESULT_MISS
)] = -1,
448 static u64 power10_cache_events
[C(MAX
)][C(OP_MAX
)][C(RESULT_MAX
)] = {
451 [C(RESULT_ACCESS
)] = PM_LD_REF_L1
,
452 [C(RESULT_MISS
)] = PM_LD_MISS_L1
,
455 [C(RESULT_ACCESS
)] = 0,
456 [C(RESULT_MISS
)] = PM_ST_MISS_L1
,
459 [C(RESULT_ACCESS
)] = PM_LD_PREFETCH_CACHE_LINE_MISS
,
460 [C(RESULT_MISS
)] = 0,
465 [C(RESULT_ACCESS
)] = PM_INST_FROM_L1
,
466 [C(RESULT_MISS
)] = PM_L1_ICACHE_MISS
,
469 [C(RESULT_ACCESS
)] = PM_INST_FROM_L1MISS
,
470 [C(RESULT_MISS
)] = -1,
473 [C(RESULT_ACCESS
)] = PM_IC_PREF_REQ
,
474 [C(RESULT_MISS
)] = 0,
479 [C(RESULT_ACCESS
)] = PM_DATA_FROM_L3
,
480 [C(RESULT_MISS
)] = PM_DATA_FROM_L3MISS
,
483 [C(RESULT_ACCESS
)] = PM_L2_ST
,
484 [C(RESULT_MISS
)] = PM_L2_ST_MISS
,
487 [C(RESULT_ACCESS
)] = PM_L3_PF_MISS_L3
,
488 [C(RESULT_MISS
)] = 0,
493 [C(RESULT_ACCESS
)] = 0,
494 [C(RESULT_MISS
)] = PM_DTLB_MISS
,
497 [C(RESULT_ACCESS
)] = -1,
498 [C(RESULT_MISS
)] = -1,
501 [C(RESULT_ACCESS
)] = -1,
502 [C(RESULT_MISS
)] = -1,
507 [C(RESULT_ACCESS
)] = 0,
508 [C(RESULT_MISS
)] = PM_ITLB_MISS
,
511 [C(RESULT_ACCESS
)] = -1,
512 [C(RESULT_MISS
)] = -1,
515 [C(RESULT_ACCESS
)] = -1,
516 [C(RESULT_MISS
)] = -1,
521 [C(RESULT_ACCESS
)] = PM_BR_CMPL
,
522 [C(RESULT_MISS
)] = PM_BR_MPRED_CMPL
,
525 [C(RESULT_ACCESS
)] = -1,
526 [C(RESULT_MISS
)] = -1,
529 [C(RESULT_ACCESS
)] = -1,
530 [C(RESULT_MISS
)] = -1,
535 [C(RESULT_ACCESS
)] = -1,
536 [C(RESULT_MISS
)] = -1,
539 [C(RESULT_ACCESS
)] = -1,
540 [C(RESULT_MISS
)] = -1,
543 [C(RESULT_ACCESS
)] = -1,
544 [C(RESULT_MISS
)] = -1,
552 * Set the MMCR0[CC56RUN] bit to enable counting for
553 * PMC5 and PMC6 regardless of the state of CTRL[RUN],
554 * so that we can use counters 5 and 6 as PM_INST_CMPL and
557 static int power10_compute_mmcr(u64 event
[], int n_ev
,
558 unsigned int hwc
[], struct mmcr_regs
*mmcr
,
559 struct perf_event
*pevents
[], u32 flags
)
563 ret
= isa207_compute_mmcr(event
, n_ev
, hwc
, mmcr
, pevents
, flags
);
565 mmcr
->mmcr0
|= MMCR0_C56RUN
;
569 static struct power_pmu power10_pmu
= {
571 .n_counter
= MAX_PMU_COUNTERS
,
572 .add_fields
= ISA207_ADD_FIELDS
,
573 .test_adder
= ISA207_TEST_ADDER
,
574 .group_constraint_mask
= CNST_CACHE_PMC4_MASK
,
575 .group_constraint_val
= CNST_CACHE_PMC4_VAL
,
576 .compute_mmcr
= power10_compute_mmcr
,
577 .config_bhrb
= power10_config_bhrb
,
578 .bhrb_filter_map
= power10_bhrb_filter_map
,
579 .get_constraint
= isa207_get_constraint
,
580 .get_alternatives
= power10_get_alternatives
,
581 .get_mem_data_src
= isa207_get_mem_data_src
,
582 .get_mem_weight
= isa207_get_mem_weight
,
583 .disable_pmc
= isa207_disable_pmc
,
584 .flags
= PPMU_HAS_SIER
| PPMU_ARCH_207S
|
585 PPMU_ARCH_31
| PPMU_HAS_ATTR_CONFIG1
,
586 .n_generic
= ARRAY_SIZE(power10_generic_events
),
587 .generic_events
= power10_generic_events
,
588 .cache_events
= &power10_cache_events
,
589 .attr_groups
= power10_pmu_attr_groups
,
591 .capabilities
= PERF_PMU_CAP_EXTENDED_REGS
,
592 .check_attr_config
= power10_check_attr_config
,
595 int init_power10_pmu(void)
600 /* Comes from cpu_specs[] */
601 if (!cur_cpu_spec
->oprofile_cpu_type
||
602 strcmp(cur_cpu_spec
->oprofile_cpu_type
, "ppc64/power10"))
605 pvr
= mfspr(SPRN_PVR
);
606 /* Add the ppmu flag for power10 DD1 */
607 if ((PVR_CFG(pvr
) == 1))
608 power10_pmu
.flags
|= PPMU_P10_DD1
;
610 /* Set the PERF_REG_EXTENDED_MASK here */
611 PERF_REG_EXTENDED_MASK
= PERF_REG_PMU_MASK_31
;
613 if ((PVR_CFG(pvr
) == 1)) {
614 power10_pmu
.generic_events
= power10_generic_events_dd1
;
615 power10_pmu
.attr_groups
= power10_pmu_attr_groups_dd1
;
616 power10_pmu
.cache_events
= &power10_cache_events_dd1
;
619 rc
= register_power_pmu(&power10_pmu
);
623 /* Tell userspace that EBB is supported */
624 cur_cpu_spec
->cpu_user_features2
|= PPC_FEATURE2_EBB
;