1 /* SPDX-License-Identifier: GPL-2.0-or-later */
3 * Performance event support - PowerPC classic/server specific definitions.
5 * Copyright 2008-2009 Paul Mackerras, IBM Corporation.
8 #include <linux/types.h>
9 #include <asm/hw_irq.h>
10 #include <linux/device.h>
11 #include <uapi/asm/perf_event.h>
13 /* Update perf_event_print_debug() if this changes */
14 #define MAX_HWEVENTS 8
15 #define MAX_EVENT_ALTERNATIVES 8
16 #define MAX_LIMITED_HWCOUNTERS 2
28 * This struct provides the constants and functions needed to
29 * describe the PMU on a particular POWER-family CPU.
35 unsigned long add_fields
;
36 unsigned long test_adder
;
37 int (*compute_mmcr
)(u64 events
[], int n_ev
,
38 unsigned int hwc
[], struct mmcr_regs
*mmcr
,
39 struct perf_event
*pevents
[]);
40 int (*get_constraint
)(u64 event_id
, unsigned long *mskp
,
42 int (*get_alternatives
)(u64 event_id
, unsigned int flags
,
44 void (*get_mem_data_src
)(union perf_mem_data_src
*dsrc
,
45 u32 flags
, struct pt_regs
*regs
);
46 void (*get_mem_weight
)(u64
*weight
);
47 unsigned long group_constraint_mask
;
48 unsigned long group_constraint_val
;
49 u64 (*bhrb_filter_map
)(u64 branch_sample_type
);
50 void (*config_bhrb
)(u64 pmu_bhrb_filter
);
51 void (*disable_pmc
)(unsigned int pmc
, struct mmcr_regs
*mmcr
);
52 int (*limited_pmc_event
)(u64 event_id
);
54 const struct attribute_group
**attr_groups
;
57 u64 (*cache_events
)[PERF_COUNT_HW_CACHE_MAX
]
58 [PERF_COUNT_HW_CACHE_OP_MAX
]
59 [PERF_COUNT_HW_CACHE_RESULT_MAX
];
63 /* BHRB entries in the PMU */
66 * set this flag with `PERF_PMU_CAP_EXTENDED_REGS` if
67 * the pmu supports extended perf regs capability
73 * Values for power_pmu.flags
75 #define PPMU_LIMITED_PMC5_6 0x00000001 /* PMC5/6 have limited function */
76 #define PPMU_ALT_SIPR 0x00000002 /* uses alternate posn for SIPR/HV */
77 #define PPMU_NO_SIPR 0x00000004 /* no SIPR/HV in MMCRA at all */
78 #define PPMU_NO_CONT_SAMPLING 0x00000008 /* no continuous sampling */
79 #define PPMU_SIAR_VALID 0x00000010 /* Processor has SIAR Valid bit */
80 #define PPMU_HAS_SSLOT 0x00000020 /* Has sampled slot in MMCRA */
81 #define PPMU_HAS_SIER 0x00000040 /* Has SIER */
82 #define PPMU_ARCH_207S 0x00000080 /* PMC is architecture v2.07S */
83 #define PPMU_NO_SIAR 0x00000100 /* Do not use SIAR */
84 #define PPMU_ARCH_31 0x00000200 /* Has MMCR3, SIER2 and SIER3 */
85 #define PPMU_P10_DD1 0x00000400 /* Is power10 DD1 processor version */
88 * Values for flags to get_alternatives()
90 #define PPMU_LIMITED_PMC_OK 1 /* can put this on a limited PMC */
91 #define PPMU_LIMITED_PMC_REQD 2 /* have to put this on a limited PMC */
92 #define PPMU_ONLY_COUNT_RUN 4 /* only counting in run state */
94 extern int register_power_pmu(struct power_pmu
*);
97 extern unsigned long perf_misc_flags(struct pt_regs
*regs
);
98 extern unsigned long perf_instruction_pointer(struct pt_regs
*regs
);
99 extern unsigned long int read_bhrb(int n
);
102 * Only override the default definitions in include/linux/perf_event.h
103 * if we have hardware PMU support.
105 #ifdef CONFIG_PPC_PERF_CTRS
106 #define perf_misc_flags(regs) perf_misc_flags(regs)
110 * The power_pmu.get_constraint function returns a 32/64-bit value and
111 * a 32/64-bit mask that express the constraints between this event_id and
114 * The value and mask are divided up into (non-overlapping) bitfields
115 * of three different types:
117 * Select field: this expresses the constraint that some set of bits
118 * in MMCR* needs to be set to a specific value for this event_id. For a
119 * select field, the mask contains 1s in every bit of the field, and
120 * the value contains a unique value for each possible setting of the
121 * MMCR* bits. The constraint checking code will ensure that two events
122 * that set the same field in their masks have the same value in their
125 * Add field: this expresses the constraint that there can be at most
126 * N events in a particular class. A field of k bits can be used for
127 * N <= 2^(k-1) - 1. The mask has the most significant bit of the field
128 * set (and the other bits 0), and the value has only the least significant
129 * bit of the field set. In addition, the 'add_fields' and 'test_adder'
130 * in the struct power_pmu for this processor come into play. The
131 * add_fields value contains 1 in the LSB of the field, and the
132 * test_adder contains 2^(k-1) - 1 - N in the field.
134 * NAND field: this expresses the constraint that you may not have events
135 * in all of a set of classes. (For example, on PPC970, you can't select
136 * events from the FPU, ISU and IDU simultaneously, although any two are
137 * possible.) For N classes, the field is N+1 bits wide, and each class
138 * is assigned one bit from the least-significant N bits. The mask has
139 * only the most-significant bit set, and the value has only the bit
140 * for the event_id's class set. The test_adder has the least significant
141 * bit set in the field.
143 * If an event_id is not subject to the constraint expressed by a particular
144 * field, then it will have 0 in both the mask and value for that field.
147 extern ssize_t
power_events_sysfs_show(struct device
*dev
,
148 struct device_attribute
*attr
, char *page
);
151 * EVENT_VAR() is same as PMU_EVENT_VAR with a suffix.
153 * Having a suffix allows us to have aliases in sysfs - eg: the generic
154 * event 'cpu-cycles' can have two entries in sysfs: 'cpu-cycles' and
155 * 'PM_CYC' where the latter is the name by which the event is known in
156 * POWER CPU specification.
158 * Similarly, some hardware and cache events use the same event code. Eg.
159 * on POWER8, both "cache-references" and "L1-dcache-loads" events refer
160 * to the same event, PM_LD_REF_L1. The suffix, allows us to have two
161 * sysfs objects for the same event and thus two entries/aliases in sysfs.
163 #define EVENT_VAR(_id, _suffix) event_attr_##_id##_suffix
164 #define EVENT_PTR(_id, _suffix) &EVENT_VAR(_id, _suffix).attr.attr
166 #define EVENT_ATTR(_name, _id, _suffix) \
167 PMU_EVENT_ATTR(_name, EVENT_VAR(_id, _suffix), _id, \
168 power_events_sysfs_show)
170 #define GENERIC_EVENT_ATTR(_name, _id) EVENT_ATTR(_name, _id, _g)
171 #define GENERIC_EVENT_PTR(_id) EVENT_PTR(_id, _g)
173 #define CACHE_EVENT_ATTR(_name, _id) EVENT_ATTR(_name, _id, _c)
174 #define CACHE_EVENT_PTR(_id) EVENT_PTR(_id, _c)
176 #define POWER_EVENT_ATTR(_name, _id) EVENT_ATTR(_name, _id, _p)
177 #define POWER_EVENT_PTR(_id) EVENT_PTR(_id, _p)