@@ -299,7 +299,7 @@ static struct extra_reg intel_icl_extra_regs[] __read_mostly = {
299
299
EVENT_EXTRA_END
300
300
};
301
301
302
- static struct extra_reg intel_spr_extra_regs [] __read_mostly = {
302
+ static struct extra_reg intel_glc_extra_regs [] __read_mostly = {
303
303
INTEL_UEVENT_EXTRA_REG (0x012a , MSR_OFFCORE_RSP_0 , 0x3fffffffffull , RSP_0 ),
304
304
INTEL_UEVENT_EXTRA_REG (0x012b , MSR_OFFCORE_RSP_1 , 0x3fffffffffull , RSP_1 ),
305
305
INTEL_UEVENT_PEBS_LDLAT_EXTRA_REG (0x01cd ),
@@ -309,7 +309,7 @@ static struct extra_reg intel_spr_extra_regs[] __read_mostly = {
309
309
EVENT_EXTRA_END
310
310
};
311
311
312
- static struct event_constraint intel_spr_event_constraints [] = {
312
+ static struct event_constraint intel_glc_event_constraints [] = {
313
313
FIXED_EVENT_CONSTRAINT (0x00c0 , 0 ), /* INST_RETIRED.ANY */
314
314
FIXED_EVENT_CONSTRAINT (0x0100 , 0 ), /* INST_RETIRED.PREC_DIST */
315
315
FIXED_EVENT_CONSTRAINT (0x003c , 1 ), /* CPU_CLK_UNHALTED.CORE */
@@ -349,7 +349,7 @@ static struct event_constraint intel_spr_event_constraints[] = {
349
349
EVENT_CONSTRAINT_END
350
350
};
351
351
352
- static struct extra_reg intel_gnr_extra_regs [] __read_mostly = {
352
+ static struct extra_reg intel_rwc_extra_regs [] __read_mostly = {
353
353
INTEL_UEVENT_EXTRA_REG (0x012a , MSR_OFFCORE_RSP_0 , 0x3fffffffffull , RSP_0 ),
354
354
INTEL_UEVENT_EXTRA_REG (0x012b , MSR_OFFCORE_RSP_1 , 0x3fffffffffull , RSP_1 ),
355
355
INTEL_UEVENT_PEBS_LDLAT_EXTRA_REG (0x01cd ),
@@ -473,7 +473,7 @@ static u64 intel_pmu_event_map(int hw_event)
473
473
return intel_perfmon_event_map [hw_event ];
474
474
}
475
475
476
- static __initconst const u64 spr_hw_cache_event_ids
476
+ static __initconst const u64 glc_hw_cache_event_ids
477
477
[PERF_COUNT_HW_CACHE_MAX ]
478
478
[PERF_COUNT_HW_CACHE_OP_MAX ]
479
479
[PERF_COUNT_HW_CACHE_RESULT_MAX ] =
@@ -552,7 +552,7 @@ static __initconst const u64 spr_hw_cache_event_ids
552
552
},
553
553
};
554
554
555
- static __initconst const u64 spr_hw_cache_extra_regs
555
+ static __initconst const u64 glc_hw_cache_extra_regs
556
556
[PERF_COUNT_HW_CACHE_MAX ]
557
557
[PERF_COUNT_HW_CACHE_OP_MAX ]
558
558
[PERF_COUNT_HW_CACHE_RESULT_MAX ] =
@@ -4273,7 +4273,7 @@ icl_get_event_constraints(struct cpu_hw_events *cpuc, int idx,
4273
4273
}
4274
4274
4275
4275
static struct event_constraint *
4276
- spr_get_event_constraints (struct cpu_hw_events * cpuc , int idx ,
4276
+ glc_get_event_constraints (struct cpu_hw_events * cpuc , int idx ,
4277
4277
struct perf_event * event )
4278
4278
{
4279
4279
struct event_constraint * c ;
@@ -4362,7 +4362,7 @@ adl_get_event_constraints(struct cpu_hw_events *cpuc, int idx,
4362
4362
struct x86_hybrid_pmu * pmu = hybrid_pmu (event -> pmu );
4363
4363
4364
4364
if (pmu -> cpu_type == hybrid_big )
4365
- return spr_get_event_constraints (cpuc , idx , event );
4365
+ return glc_get_event_constraints (cpuc , idx , event );
4366
4366
else if (pmu -> cpu_type == hybrid_small )
4367
4367
return tnt_get_event_constraints (cpuc , idx , event );
4368
4368
@@ -4409,7 +4409,7 @@ rwc_get_event_constraints(struct cpu_hw_events *cpuc, int idx,
4409
4409
{
4410
4410
struct event_constraint * c ;
4411
4411
4412
- c = spr_get_event_constraints (cpuc , idx , event );
4412
+ c = glc_get_event_constraints (cpuc , idx , event );
4413
4413
4414
4414
/* The Retire Latency is not supported by the fixed counter 0. */
4415
4415
if (event -> attr .precise_ip &&
@@ -4490,7 +4490,7 @@ static void nhm_limit_period(struct perf_event *event, s64 *left)
4490
4490
* left = max (* left , 32LL );
4491
4491
}
4492
4492
4493
- static void spr_limit_period (struct perf_event * event , s64 * left )
4493
+ static void glc_limit_period (struct perf_event * event , s64 * left )
4494
4494
{
4495
4495
if (event -> attr .precise_ip == 3 )
4496
4496
* left = max (* left , 128LL );
@@ -5337,14 +5337,14 @@ static struct attribute *icl_tsx_events_attrs[] = {
5337
5337
EVENT_ATTR_STR (mem - stores , mem_st_spr , "event=0xcd,umask=0x2" );
5338
5338
EVENT_ATTR_STR (mem - loads - aux , mem_ld_aux , "event=0x03,umask=0x82" );
5339
5339
5340
- static struct attribute * spr_events_attrs [] = {
5340
+ static struct attribute * glc_events_attrs [] = {
5341
5341
EVENT_PTR (mem_ld_hsw ),
5342
5342
EVENT_PTR (mem_st_spr ),
5343
5343
EVENT_PTR (mem_ld_aux ),
5344
5344
NULL ,
5345
5345
};
5346
5346
5347
- static struct attribute * spr_td_events_attrs [] = {
5347
+ static struct attribute * glc_td_events_attrs [] = {
5348
5348
EVENT_PTR (slots ),
5349
5349
EVENT_PTR (td_retiring ),
5350
5350
EVENT_PTR (td_bad_spec ),
@@ -5357,7 +5357,7 @@ static struct attribute *spr_td_events_attrs[] = {
5357
5357
NULL ,
5358
5358
};
5359
5359
5360
- static struct attribute * spr_tsx_events_attrs [] = {
5360
+ static struct attribute * glc_tsx_events_attrs [] = {
5361
5361
EVENT_PTR (tx_start ),
5362
5362
EVENT_PTR (tx_abort ),
5363
5363
EVENT_PTR (tx_commit ),
@@ -6215,7 +6215,7 @@ __init int intel_pmu_init(void)
6215
6215
intel_pmu_pebs_data_source_grt ();
6216
6216
x86_pmu .pebs_latency_data = adl_latency_data_small ;
6217
6217
x86_pmu .get_event_constraints = tnt_get_event_constraints ;
6218
- x86_pmu .limit_period = spr_limit_period ;
6218
+ x86_pmu .limit_period = glc_limit_period ;
6219
6219
td_attr = tnt_events_attrs ;
6220
6220
mem_attr = grt_mem_attrs ;
6221
6221
extra_attr = nhm_format_attr ;
@@ -6246,7 +6246,7 @@ __init int intel_pmu_init(void)
6246
6246
intel_pmu_pebs_data_source_cmt ();
6247
6247
x86_pmu .pebs_latency_data = mtl_latency_data_small ;
6248
6248
x86_pmu .get_event_constraints = cmt_get_event_constraints ;
6249
- x86_pmu .limit_period = spr_limit_period ;
6249
+ x86_pmu .limit_period = glc_limit_period ;
6250
6250
td_attr = cmt_events_attrs ;
6251
6251
mem_attr = grt_mem_attrs ;
6252
6252
extra_attr = cmt_format_attr ;
@@ -6563,20 +6563,20 @@ __init int intel_pmu_init(void)
6563
6563
case INTEL_FAM6_SAPPHIRERAPIDS_X :
6564
6564
case INTEL_FAM6_EMERALDRAPIDS_X :
6565
6565
x86_pmu .flags |= PMU_FL_MEM_LOADS_AUX ;
6566
- x86_pmu .extra_regs = intel_spr_extra_regs ;
6566
+ x86_pmu .extra_regs = intel_glc_extra_regs ;
6567
6567
fallthrough ;
6568
6568
case INTEL_FAM6_GRANITERAPIDS_X :
6569
6569
case INTEL_FAM6_GRANITERAPIDS_D :
6570
6570
pmem = true;
6571
6571
x86_pmu .late_ack = true;
6572
- memcpy (hw_cache_event_ids , spr_hw_cache_event_ids , sizeof (hw_cache_event_ids ));
6573
- memcpy (hw_cache_extra_regs , spr_hw_cache_extra_regs , sizeof (hw_cache_extra_regs ));
6572
+ memcpy (hw_cache_event_ids , glc_hw_cache_event_ids , sizeof (hw_cache_event_ids ));
6573
+ memcpy (hw_cache_extra_regs , glc_hw_cache_extra_regs , sizeof (hw_cache_extra_regs ));
6574
6574
6575
- x86_pmu .event_constraints = intel_spr_event_constraints ;
6576
- x86_pmu .pebs_constraints = intel_spr_pebs_event_constraints ;
6575
+ x86_pmu .event_constraints = intel_glc_event_constraints ;
6576
+ x86_pmu .pebs_constraints = intel_glc_pebs_event_constraints ;
6577
6577
if (!x86_pmu .extra_regs )
6578
- x86_pmu .extra_regs = intel_gnr_extra_regs ;
6579
- x86_pmu .limit_period = spr_limit_period ;
6578
+ x86_pmu .extra_regs = intel_rwc_extra_regs ;
6579
+ x86_pmu .limit_period = glc_limit_period ;
6580
6580
x86_pmu .pebs_ept = 1 ;
6581
6581
x86_pmu .pebs_aliases = NULL ;
6582
6582
x86_pmu .pebs_prec_dist = true;
@@ -6586,13 +6586,13 @@ __init int intel_pmu_init(void)
6586
6586
x86_pmu .flags |= PMU_FL_INSTR_LATENCY ;
6587
6587
6588
6588
x86_pmu .hw_config = hsw_hw_config ;
6589
- x86_pmu .get_event_constraints = spr_get_event_constraints ;
6589
+ x86_pmu .get_event_constraints = glc_get_event_constraints ;
6590
6590
extra_attr = boot_cpu_has (X86_FEATURE_RTM ) ?
6591
6591
hsw_format_attr : nhm_format_attr ;
6592
6592
extra_skl_attr = skl_format_attr ;
6593
- mem_attr = spr_events_attrs ;
6594
- td_attr = spr_td_events_attrs ;
6595
- tsx_attr = spr_tsx_events_attrs ;
6593
+ mem_attr = glc_events_attrs ;
6594
+ td_attr = glc_td_events_attrs ;
6595
+ tsx_attr = glc_tsx_events_attrs ;
6596
6596
x86_pmu .rtm_abort_event = X86_CONFIG (.event = 0xc9 , .umask = 0x04 );
6597
6597
x86_pmu .lbr_pt_coexist = true;
6598
6598
intel_pmu_pebs_data_source_skl (pmem );
@@ -6642,7 +6642,7 @@ __init int intel_pmu_init(void)
6642
6642
x86_pmu .filter = intel_pmu_filter ;
6643
6643
x86_pmu .get_event_constraints = adl_get_event_constraints ;
6644
6644
x86_pmu .hw_config = adl_hw_config ;
6645
- x86_pmu .limit_period = spr_limit_period ;
6645
+ x86_pmu .limit_period = glc_limit_period ;
6646
6646
x86_pmu .get_hybrid_cpu_type = adl_get_hybrid_cpu_type ;
6647
6647
/*
6648
6648
* The rtm_abort_event is used to check whether to enable GPRs
@@ -6691,11 +6691,11 @@ __init int intel_pmu_init(void)
6691
6691
pmu -> intel_cap .perf_metrics = 1 ;
6692
6692
pmu -> intel_cap .pebs_output_pt_available = 0 ;
6693
6693
6694
- memcpy (pmu -> hw_cache_event_ids , spr_hw_cache_event_ids , sizeof (pmu -> hw_cache_event_ids ));
6695
- memcpy (pmu -> hw_cache_extra_regs , spr_hw_cache_extra_regs , sizeof (pmu -> hw_cache_extra_regs ));
6696
- pmu -> event_constraints = intel_spr_event_constraints ;
6697
- pmu -> pebs_constraints = intel_spr_pebs_event_constraints ;
6698
- pmu -> extra_regs = intel_spr_extra_regs ;
6694
+ memcpy (pmu -> hw_cache_event_ids , glc_hw_cache_event_ids , sizeof (pmu -> hw_cache_event_ids ));
6695
+ memcpy (pmu -> hw_cache_extra_regs , glc_hw_cache_extra_regs , sizeof (pmu -> hw_cache_extra_regs ));
6696
+ pmu -> event_constraints = intel_glc_event_constraints ;
6697
+ pmu -> pebs_constraints = intel_glc_pebs_event_constraints ;
6698
+ pmu -> extra_regs = intel_glc_extra_regs ;
6699
6699
6700
6700
/* Initialize Atom core specific PerfMon capabilities.*/
6701
6701
pmu = & x86_pmu .hybrid_pmu [X86_HYBRID_PMU_ATOM_IDX ];
@@ -6719,7 +6719,7 @@ __init int intel_pmu_init(void)
6719
6719
pmu -> pebs_constraints = intel_grt_pebs_event_constraints ;
6720
6720
pmu -> extra_regs = intel_grt_extra_regs ;
6721
6721
if (is_mtl (boot_cpu_data .x86_model )) {
6722
- x86_pmu .hybrid_pmu [X86_HYBRID_PMU_CORE_IDX ].extra_regs = intel_gnr_extra_regs ;
6722
+ x86_pmu .hybrid_pmu [X86_HYBRID_PMU_CORE_IDX ].extra_regs = intel_rwc_extra_regs ;
6723
6723
x86_pmu .pebs_latency_data = mtl_latency_data_small ;
6724
6724
extra_attr = boot_cpu_has (X86_FEATURE_RTM ) ?
6725
6725
mtl_hybrid_extra_attr_rtm : mtl_hybrid_extra_attr ;
0 commit comments