1 // SPDX-License-Identifier: GPL-2.0
2 /* Marvell CN10K DRAM Subsystem (DSS) Performance Monitor Driver
4 * Copyright (C) 2021 Marvell.
7 #include <linux/init.h>
9 #include <linux/module.h>
11 #include <linux/perf_event.h>
12 #include <linux/hrtimer.h>
13 #include <linux/acpi.h>
14 #include <linux/platform_device.h>
16 /* Performance Counters Operating Mode Control Registers */
17 #define DDRC_PERF_CNT_OP_MODE_CTRL 0x8020
18 #define OP_MODE_CTRL_VAL_MANNUAL 0x1
20 /* Performance Counters Start Operation Control Registers */
21 #define DDRC_PERF_CNT_START_OP_CTRL 0x8028
22 #define START_OP_CTRL_VAL_START 0x1ULL
23 #define START_OP_CTRL_VAL_ACTIVE 0x2
25 /* Performance Counters End Operation Control Registers */
26 #define DDRC_PERF_CNT_END_OP_CTRL 0x8030
27 #define END_OP_CTRL_VAL_END 0x1ULL
29 /* Performance Counters End Status Registers */
30 #define DDRC_PERF_CNT_END_STATUS 0x8038
31 #define END_STATUS_VAL_END_TIMER_MODE_END 0x1
33 /* Performance Counters Configuration Registers */
34 #define DDRC_PERF_CFG_BASE 0x8040
36 /* 8 Generic event counter + 2 fixed event counters */
37 #define DDRC_PERF_NUM_GEN_COUNTERS 8
38 #define DDRC_PERF_NUM_FIX_COUNTERS 2
39 #define DDRC_PERF_READ_COUNTER_IDX DDRC_PERF_NUM_GEN_COUNTERS
40 #define DDRC_PERF_WRITE_COUNTER_IDX (DDRC_PERF_NUM_GEN_COUNTERS + 1)
41 #define DDRC_PERF_NUM_COUNTERS (DDRC_PERF_NUM_GEN_COUNTERS + \
42 DDRC_PERF_NUM_FIX_COUNTERS)
44 /* Generic event counter registers */
45 #define DDRC_PERF_CFG(n) (DDRC_PERF_CFG_BASE + 8 * (n))
46 #define EVENT_ENABLE BIT_ULL(63)
48 /* Two dedicated event counters for DDR reads and writes */
49 #define EVENT_DDR_READS 101
50 #define EVENT_DDR_WRITES 100
53 * programmable events IDs in programmable event counters.
54 * DO NOT change these event-id numbers, they are used to
55 * program event bitmap in h/w.
57 #define EVENT_OP_IS_ZQLATCH 55
58 #define EVENT_OP_IS_ZQSTART 54
59 #define EVENT_OP_IS_TCR_MRR 53
60 #define EVENT_OP_IS_DQSOSC_MRR 52
61 #define EVENT_OP_IS_DQSOSC_MPC 51
62 #define EVENT_VISIBLE_WIN_LIMIT_REACHED_WR 50
63 #define EVENT_VISIBLE_WIN_LIMIT_REACHED_RD 49
64 #define EVENT_BSM_STARVATION 48
65 #define EVENT_BSM_ALLOC 47
66 #define EVENT_LPR_REQ_WITH_NOCREDIT 46
67 #define EVENT_HPR_REQ_WITH_NOCREDIT 45
68 #define EVENT_OP_IS_ZQCS 44
69 #define EVENT_OP_IS_ZQCL 43
70 #define EVENT_OP_IS_LOAD_MODE 42
71 #define EVENT_OP_IS_SPEC_REF 41
72 #define EVENT_OP_IS_CRIT_REF 40
73 #define EVENT_OP_IS_REFRESH 39
74 #define EVENT_OP_IS_ENTER_MPSM 35
75 #define EVENT_OP_IS_ENTER_POWERDOWN 31
76 #define EVENT_OP_IS_ENTER_SELFREF 27
77 #define EVENT_WAW_HAZARD 26
78 #define EVENT_RAW_HAZARD 25
79 #define EVENT_WAR_HAZARD 24
80 #define EVENT_WRITE_COMBINE 23
81 #define EVENT_RDWR_TRANSITIONS 22
82 #define EVENT_PRECHARGE_FOR_OTHER 21
83 #define EVENT_PRECHARGE_FOR_RDWR 20
84 #define EVENT_OP_IS_PRECHARGE 19
85 #define EVENT_OP_IS_MWR 18
86 #define EVENT_OP_IS_WR 17
87 #define EVENT_OP_IS_RD 16
88 #define EVENT_OP_IS_RD_ACTIVATE 15
89 #define EVENT_OP_IS_RD_OR_WR 14
90 #define EVENT_OP_IS_ACTIVATE 13
91 #define EVENT_WR_XACT_WHEN_CRITICAL 12
92 #define EVENT_LPR_XACT_WHEN_CRITICAL 11
93 #define EVENT_HPR_XACT_WHEN_CRITICAL 10
94 #define EVENT_DFI_RD_DATA_CYCLES 9
95 #define EVENT_DFI_WR_DATA_CYCLES 8
96 #define EVENT_ACT_BYPASS 7
97 #define EVENT_READ_BYPASS 6
98 #define EVENT_HIF_HI_PRI_RD 5
99 #define EVENT_HIF_RMW 4
100 #define EVENT_HIF_RD 3
101 #define EVENT_HIF_WR 2
102 #define EVENT_HIF_RD_OR_WR 1
104 /* Event counter value registers */
105 #define DDRC_PERF_CNT_VALUE_BASE 0x8080
106 #define DDRC_PERF_CNT_VALUE(n) (DDRC_PERF_CNT_VALUE_BASE + 8 * (n))
108 /* Fixed event counter enable/disable register */
109 #define DDRC_PERF_CNT_FREERUN_EN 0x80C0
110 #define DDRC_PERF_FREERUN_WRITE_EN 0x1
111 #define DDRC_PERF_FREERUN_READ_EN 0x2
113 /* Fixed event counter control register */
114 #define DDRC_PERF_CNT_FREERUN_CTRL 0x80C8
115 #define DDRC_FREERUN_WRITE_CNT_CLR 0x1
116 #define DDRC_FREERUN_READ_CNT_CLR 0x2
118 /* Fixed event counter value register */
119 #define DDRC_PERF_CNT_VALUE_WR_OP 0x80D0
120 #define DDRC_PERF_CNT_VALUE_RD_OP 0x80D8
121 #define DDRC_PERF_CNT_VALUE_OVERFLOW BIT_ULL(48)
122 #define DDRC_PERF_CNT_MAX_VALUE GENMASK_ULL(48, 0)
124 struct cn10k_ddr_pmu
{
130 struct perf_event
*events
[DDRC_PERF_NUM_COUNTERS
];
131 struct hrtimer hrtimer
;
132 struct hlist_node node
;
135 #define to_cn10k_ddr_pmu(p) container_of(p, struct cn10k_ddr_pmu, pmu)
137 static ssize_t
cn10k_ddr_pmu_event_show(struct device
*dev
,
138 struct device_attribute
*attr
,
141 struct perf_pmu_events_attr
*pmu_attr
;
143 pmu_attr
= container_of(attr
, struct perf_pmu_events_attr
, attr
);
144 return sysfs_emit(page
, "event=0x%02llx\n", pmu_attr
->id
);
148 #define CN10K_DDR_PMU_EVENT_ATTR(_name, _id) \
149 PMU_EVENT_ATTR_ID(_name, cn10k_ddr_pmu_event_show, _id)
151 static struct attribute
*cn10k_ddr_perf_events_attrs
[] = {
152 CN10K_DDR_PMU_EVENT_ATTR(ddr_hif_rd_or_wr_access
, EVENT_HIF_RD_OR_WR
),
153 CN10K_DDR_PMU_EVENT_ATTR(ddr_hif_wr_access
, EVENT_HIF_WR
),
154 CN10K_DDR_PMU_EVENT_ATTR(ddr_hif_rd_access
, EVENT_HIF_RD
),
155 CN10K_DDR_PMU_EVENT_ATTR(ddr_hif_rmw_access
, EVENT_HIF_RMW
),
156 CN10K_DDR_PMU_EVENT_ATTR(ddr_hif_pri_rdaccess
, EVENT_HIF_HI_PRI_RD
),
157 CN10K_DDR_PMU_EVENT_ATTR(ddr_rd_bypass_access
, EVENT_READ_BYPASS
),
158 CN10K_DDR_PMU_EVENT_ATTR(ddr_act_bypass_access
, EVENT_ACT_BYPASS
),
159 CN10K_DDR_PMU_EVENT_ATTR(ddr_dif_wr_data_access
, EVENT_DFI_WR_DATA_CYCLES
),
160 CN10K_DDR_PMU_EVENT_ATTR(ddr_dif_rd_data_access
, EVENT_DFI_RD_DATA_CYCLES
),
161 CN10K_DDR_PMU_EVENT_ATTR(ddr_hpri_sched_rd_crit_access
,
162 EVENT_HPR_XACT_WHEN_CRITICAL
),
163 CN10K_DDR_PMU_EVENT_ATTR(ddr_lpri_sched_rd_crit_access
,
164 EVENT_LPR_XACT_WHEN_CRITICAL
),
165 CN10K_DDR_PMU_EVENT_ATTR(ddr_wr_trxn_crit_access
,
166 EVENT_WR_XACT_WHEN_CRITICAL
),
167 CN10K_DDR_PMU_EVENT_ATTR(ddr_cam_active_access
, EVENT_OP_IS_ACTIVATE
),
168 CN10K_DDR_PMU_EVENT_ATTR(ddr_cam_rd_or_wr_access
, EVENT_OP_IS_RD_OR_WR
),
169 CN10K_DDR_PMU_EVENT_ATTR(ddr_cam_rd_active_access
, EVENT_OP_IS_RD_ACTIVATE
),
170 CN10K_DDR_PMU_EVENT_ATTR(ddr_cam_read
, EVENT_OP_IS_RD
),
171 CN10K_DDR_PMU_EVENT_ATTR(ddr_cam_write
, EVENT_OP_IS_WR
),
172 CN10K_DDR_PMU_EVENT_ATTR(ddr_cam_mwr
, EVENT_OP_IS_MWR
),
173 CN10K_DDR_PMU_EVENT_ATTR(ddr_precharge
, EVENT_OP_IS_PRECHARGE
),
174 CN10K_DDR_PMU_EVENT_ATTR(ddr_precharge_for_rdwr
, EVENT_PRECHARGE_FOR_RDWR
),
175 CN10K_DDR_PMU_EVENT_ATTR(ddr_precharge_for_other
,
176 EVENT_PRECHARGE_FOR_OTHER
),
177 CN10K_DDR_PMU_EVENT_ATTR(ddr_rdwr_transitions
, EVENT_RDWR_TRANSITIONS
),
178 CN10K_DDR_PMU_EVENT_ATTR(ddr_write_combine
, EVENT_WRITE_COMBINE
),
179 CN10K_DDR_PMU_EVENT_ATTR(ddr_war_hazard
, EVENT_WAR_HAZARD
),
180 CN10K_DDR_PMU_EVENT_ATTR(ddr_raw_hazard
, EVENT_RAW_HAZARD
),
181 CN10K_DDR_PMU_EVENT_ATTR(ddr_waw_hazard
, EVENT_WAW_HAZARD
),
182 CN10K_DDR_PMU_EVENT_ATTR(ddr_enter_selfref
, EVENT_OP_IS_ENTER_SELFREF
),
183 CN10K_DDR_PMU_EVENT_ATTR(ddr_enter_powerdown
, EVENT_OP_IS_ENTER_POWERDOWN
),
184 CN10K_DDR_PMU_EVENT_ATTR(ddr_enter_mpsm
, EVENT_OP_IS_ENTER_MPSM
),
185 CN10K_DDR_PMU_EVENT_ATTR(ddr_refresh
, EVENT_OP_IS_REFRESH
),
186 CN10K_DDR_PMU_EVENT_ATTR(ddr_crit_ref
, EVENT_OP_IS_CRIT_REF
),
187 CN10K_DDR_PMU_EVENT_ATTR(ddr_spec_ref
, EVENT_OP_IS_SPEC_REF
),
188 CN10K_DDR_PMU_EVENT_ATTR(ddr_load_mode
, EVENT_OP_IS_LOAD_MODE
),
189 CN10K_DDR_PMU_EVENT_ATTR(ddr_zqcl
, EVENT_OP_IS_ZQCL
),
190 CN10K_DDR_PMU_EVENT_ATTR(ddr_cam_wr_access
, EVENT_OP_IS_ZQCS
),
191 CN10K_DDR_PMU_EVENT_ATTR(ddr_hpr_req_with_nocredit
,
192 EVENT_HPR_REQ_WITH_NOCREDIT
),
193 CN10K_DDR_PMU_EVENT_ATTR(ddr_lpr_req_with_nocredit
,
194 EVENT_LPR_REQ_WITH_NOCREDIT
),
195 CN10K_DDR_PMU_EVENT_ATTR(ddr_bsm_alloc
, EVENT_BSM_ALLOC
),
196 CN10K_DDR_PMU_EVENT_ATTR(ddr_bsm_starvation
, EVENT_BSM_STARVATION
),
197 CN10K_DDR_PMU_EVENT_ATTR(ddr_win_limit_reached_rd
,
198 EVENT_VISIBLE_WIN_LIMIT_REACHED_RD
),
199 CN10K_DDR_PMU_EVENT_ATTR(ddr_win_limit_reached_wr
,
200 EVENT_VISIBLE_WIN_LIMIT_REACHED_WR
),
201 CN10K_DDR_PMU_EVENT_ATTR(ddr_dqsosc_mpc
, EVENT_OP_IS_DQSOSC_MPC
),
202 CN10K_DDR_PMU_EVENT_ATTR(ddr_dqsosc_mrr
, EVENT_OP_IS_DQSOSC_MRR
),
203 CN10K_DDR_PMU_EVENT_ATTR(ddr_tcr_mrr
, EVENT_OP_IS_TCR_MRR
),
204 CN10K_DDR_PMU_EVENT_ATTR(ddr_zqstart
, EVENT_OP_IS_ZQSTART
),
205 CN10K_DDR_PMU_EVENT_ATTR(ddr_zqlatch
, EVENT_OP_IS_ZQLATCH
),
206 /* Free run event counters */
207 CN10K_DDR_PMU_EVENT_ATTR(ddr_ddr_reads
, EVENT_DDR_READS
),
208 CN10K_DDR_PMU_EVENT_ATTR(ddr_ddr_writes
, EVENT_DDR_WRITES
),
212 static struct attribute_group cn10k_ddr_perf_events_attr_group
= {
214 .attrs
= cn10k_ddr_perf_events_attrs
,
217 PMU_FORMAT_ATTR(event
, "config:0-8");
219 static struct attribute
*cn10k_ddr_perf_format_attrs
[] = {
220 &format_attr_event
.attr
,
224 static struct attribute_group cn10k_ddr_perf_format_attr_group
= {
226 .attrs
= cn10k_ddr_perf_format_attrs
,
229 static ssize_t
cn10k_ddr_perf_cpumask_show(struct device
*dev
,
230 struct device_attribute
*attr
,
233 struct cn10k_ddr_pmu
*pmu
= dev_get_drvdata(dev
);
235 return cpumap_print_to_pagebuf(true, buf
, cpumask_of(pmu
->cpu
));
238 static struct device_attribute cn10k_ddr_perf_cpumask_attr
=
239 __ATTR(cpumask
, 0444, cn10k_ddr_perf_cpumask_show
, NULL
);
241 static struct attribute
*cn10k_ddr_perf_cpumask_attrs
[] = {
242 &cn10k_ddr_perf_cpumask_attr
.attr
,
246 static struct attribute_group cn10k_ddr_perf_cpumask_attr_group
= {
247 .attrs
= cn10k_ddr_perf_cpumask_attrs
,
250 static const struct attribute_group
*cn10k_attr_groups
[] = {
251 &cn10k_ddr_perf_events_attr_group
,
252 &cn10k_ddr_perf_format_attr_group
,
253 &cn10k_ddr_perf_cpumask_attr_group
,
257 /* Default poll timeout is 100 sec, which is very sufficient for
258 * 48 bit counter incremented max at 5.6 GT/s, which may take many
261 static unsigned long cn10k_ddr_pmu_poll_period_sec
= 100;
262 module_param_named(poll_period_sec
, cn10k_ddr_pmu_poll_period_sec
, ulong
, 0644);
264 static ktime_t
cn10k_ddr_pmu_timer_period(void)
266 return ms_to_ktime((u64
)cn10k_ddr_pmu_poll_period_sec
* USEC_PER_SEC
);
269 static int ddr_perf_get_event_bitmap(int eventid
, u64
*event_bitmap
)
272 case EVENT_HIF_RD_OR_WR
... EVENT_WAW_HAZARD
:
273 case EVENT_OP_IS_REFRESH
... EVENT_OP_IS_ZQLATCH
:
274 *event_bitmap
= (1ULL << (eventid
- 1));
276 case EVENT_OP_IS_ENTER_SELFREF
:
277 case EVENT_OP_IS_ENTER_POWERDOWN
:
278 case EVENT_OP_IS_ENTER_MPSM
:
279 *event_bitmap
= (0xFULL
<< (eventid
- 1));
282 pr_err("%s Invalid eventid %d\n", __func__
, eventid
);
289 static int cn10k_ddr_perf_alloc_counter(struct cn10k_ddr_pmu
*pmu
,
290 struct perf_event
*event
)
292 u8 config
= event
->attr
.config
;
295 /* DDR read free-run counter index */
296 if (config
== EVENT_DDR_READS
) {
297 pmu
->events
[DDRC_PERF_READ_COUNTER_IDX
] = event
;
298 return DDRC_PERF_READ_COUNTER_IDX
;
301 /* DDR write free-run counter index */
302 if (config
== EVENT_DDR_WRITES
) {
303 pmu
->events
[DDRC_PERF_WRITE_COUNTER_IDX
] = event
;
304 return DDRC_PERF_WRITE_COUNTER_IDX
;
307 /* Allocate DDR generic counters */
308 for (i
= 0; i
< DDRC_PERF_NUM_GEN_COUNTERS
; i
++) {
309 if (pmu
->events
[i
] == NULL
) {
310 pmu
->events
[i
] = event
;
318 static void cn10k_ddr_perf_free_counter(struct cn10k_ddr_pmu
*pmu
, int counter
)
320 pmu
->events
[counter
] = NULL
;
323 static int cn10k_ddr_perf_event_init(struct perf_event
*event
)
325 struct cn10k_ddr_pmu
*pmu
= to_cn10k_ddr_pmu(event
->pmu
);
326 struct hw_perf_event
*hwc
= &event
->hw
;
328 if (event
->attr
.type
!= event
->pmu
->type
)
331 if (is_sampling_event(event
)) {
332 dev_info(pmu
->dev
, "Sampling not supported!\n");
336 if (event
->cpu
< 0) {
337 dev_warn(pmu
->dev
, "Can't provide per-task data!\n");
341 /* We must NOT create groups containing mixed PMUs */
342 if (event
->group_leader
->pmu
!= event
->pmu
&&
343 !is_software_event(event
->group_leader
))
346 /* Set ownership of event to one CPU, same event can not be observed
347 * on multiple cpus at same time.
349 event
->cpu
= pmu
->cpu
;
354 static void cn10k_ddr_perf_counter_enable(struct cn10k_ddr_pmu
*pmu
,
355 int counter
, bool enable
)
360 if (counter
> DDRC_PERF_NUM_COUNTERS
) {
361 pr_err("Error: unsupported counter %d\n", counter
);
365 if (counter
< DDRC_PERF_NUM_GEN_COUNTERS
) {
366 reg
= DDRC_PERF_CFG(counter
);
367 val
= readq_relaxed(pmu
->base
+ reg
);
372 val
&= ~EVENT_ENABLE
;
374 writeq_relaxed(val
, pmu
->base
+ reg
);
376 val
= readq_relaxed(pmu
->base
+ DDRC_PERF_CNT_FREERUN_EN
);
378 if (counter
== DDRC_PERF_READ_COUNTER_IDX
)
379 val
|= DDRC_PERF_FREERUN_READ_EN
;
381 val
|= DDRC_PERF_FREERUN_WRITE_EN
;
383 if (counter
== DDRC_PERF_READ_COUNTER_IDX
)
384 val
&= ~DDRC_PERF_FREERUN_READ_EN
;
386 val
&= ~DDRC_PERF_FREERUN_WRITE_EN
;
388 writeq_relaxed(val
, pmu
->base
+ DDRC_PERF_CNT_FREERUN_EN
);
392 static u64
cn10k_ddr_perf_read_counter(struct cn10k_ddr_pmu
*pmu
, int counter
)
396 if (counter
== DDRC_PERF_READ_COUNTER_IDX
)
397 return readq_relaxed(pmu
->base
+ DDRC_PERF_CNT_VALUE_RD_OP
);
399 if (counter
== DDRC_PERF_WRITE_COUNTER_IDX
)
400 return readq_relaxed(pmu
->base
+ DDRC_PERF_CNT_VALUE_WR_OP
);
402 val
= readq_relaxed(pmu
->base
+ DDRC_PERF_CNT_VALUE(counter
));
406 static void cn10k_ddr_perf_event_update(struct perf_event
*event
)
408 struct cn10k_ddr_pmu
*pmu
= to_cn10k_ddr_pmu(event
->pmu
);
409 struct hw_perf_event
*hwc
= &event
->hw
;
410 u64 prev_count
, new_count
, mask
;
413 prev_count
= local64_read(&hwc
->prev_count
);
414 new_count
= cn10k_ddr_perf_read_counter(pmu
, hwc
->idx
);
415 } while (local64_xchg(&hwc
->prev_count
, new_count
) != prev_count
);
417 mask
= DDRC_PERF_CNT_MAX_VALUE
;
419 local64_add((new_count
- prev_count
) & mask
, &event
->count
);
422 static void cn10k_ddr_perf_event_start(struct perf_event
*event
, int flags
)
424 struct cn10k_ddr_pmu
*pmu
= to_cn10k_ddr_pmu(event
->pmu
);
425 struct hw_perf_event
*hwc
= &event
->hw
;
426 int counter
= hwc
->idx
;
428 local64_set(&hwc
->prev_count
, 0);
430 cn10k_ddr_perf_counter_enable(pmu
, counter
, true);
435 static int cn10k_ddr_perf_event_add(struct perf_event
*event
, int flags
)
437 struct cn10k_ddr_pmu
*pmu
= to_cn10k_ddr_pmu(event
->pmu
);
438 struct hw_perf_event
*hwc
= &event
->hw
;
439 u8 config
= event
->attr
.config
;
444 counter
= cn10k_ddr_perf_alloc_counter(pmu
, event
);
448 pmu
->active_events
++;
451 if (pmu
->active_events
== 1)
452 hrtimer_start(&pmu
->hrtimer
, cn10k_ddr_pmu_timer_period(),
453 HRTIMER_MODE_REL_PINNED
);
455 if (counter
< DDRC_PERF_NUM_GEN_COUNTERS
) {
456 /* Generic counters, configure event id */
457 reg_offset
= DDRC_PERF_CFG(counter
);
458 ret
= ddr_perf_get_event_bitmap(config
, &val
);
462 writeq_relaxed(val
, pmu
->base
+ reg_offset
);
464 /* fixed event counter, clear counter value */
465 if (counter
== DDRC_PERF_READ_COUNTER_IDX
)
466 val
= DDRC_FREERUN_READ_CNT_CLR
;
468 val
= DDRC_FREERUN_WRITE_CNT_CLR
;
470 writeq_relaxed(val
, pmu
->base
+ DDRC_PERF_CNT_FREERUN_CTRL
);
473 hwc
->state
|= PERF_HES_STOPPED
;
475 if (flags
& PERF_EF_START
)
476 cn10k_ddr_perf_event_start(event
, flags
);
481 static void cn10k_ddr_perf_event_stop(struct perf_event
*event
, int flags
)
483 struct cn10k_ddr_pmu
*pmu
= to_cn10k_ddr_pmu(event
->pmu
);
484 struct hw_perf_event
*hwc
= &event
->hw
;
485 int counter
= hwc
->idx
;
487 cn10k_ddr_perf_counter_enable(pmu
, counter
, false);
489 if (flags
& PERF_EF_UPDATE
)
490 cn10k_ddr_perf_event_update(event
);
492 hwc
->state
|= PERF_HES_STOPPED
;
495 static void cn10k_ddr_perf_event_del(struct perf_event
*event
, int flags
)
497 struct cn10k_ddr_pmu
*pmu
= to_cn10k_ddr_pmu(event
->pmu
);
498 struct hw_perf_event
*hwc
= &event
->hw
;
499 int counter
= hwc
->idx
;
501 cn10k_ddr_perf_event_stop(event
, PERF_EF_UPDATE
);
503 cn10k_ddr_perf_free_counter(pmu
, counter
);
504 pmu
->active_events
--;
507 /* Cancel timer when no events to capture */
508 if (pmu
->active_events
== 0)
509 hrtimer_cancel(&pmu
->hrtimer
);
512 static void cn10k_ddr_perf_pmu_enable(struct pmu
*pmu
)
514 struct cn10k_ddr_pmu
*ddr_pmu
= to_cn10k_ddr_pmu(pmu
);
516 writeq_relaxed(START_OP_CTRL_VAL_START
, ddr_pmu
->base
+
517 DDRC_PERF_CNT_START_OP_CTRL
);
520 static void cn10k_ddr_perf_pmu_disable(struct pmu
*pmu
)
522 struct cn10k_ddr_pmu
*ddr_pmu
= to_cn10k_ddr_pmu(pmu
);
524 writeq_relaxed(END_OP_CTRL_VAL_END
, ddr_pmu
->base
+
525 DDRC_PERF_CNT_END_OP_CTRL
);
528 static void cn10k_ddr_perf_event_update_all(struct cn10k_ddr_pmu
*pmu
)
530 struct hw_perf_event
*hwc
;
533 for (i
= 0; i
< DDRC_PERF_NUM_GEN_COUNTERS
; i
++) {
534 if (pmu
->events
[i
] == NULL
)
537 cn10k_ddr_perf_event_update(pmu
->events
[i
]);
540 /* Reset previous count as h/w counter are reset */
541 for (i
= 0; i
< DDRC_PERF_NUM_GEN_COUNTERS
; i
++) {
542 if (pmu
->events
[i
] == NULL
)
545 hwc
= &pmu
->events
[i
]->hw
;
546 local64_set(&hwc
->prev_count
, 0);
550 static irqreturn_t
cn10k_ddr_pmu_overflow_handler(struct cn10k_ddr_pmu
*pmu
)
552 struct perf_event
*event
;
553 struct hw_perf_event
*hwc
;
554 u64 prev_count
, new_count
;
558 event
= pmu
->events
[DDRC_PERF_READ_COUNTER_IDX
];
561 prev_count
= local64_read(&hwc
->prev_count
);
562 new_count
= cn10k_ddr_perf_read_counter(pmu
, hwc
->idx
);
564 /* Overflow condition is when new count less than
567 if (new_count
< prev_count
)
568 cn10k_ddr_perf_event_update(event
);
571 event
= pmu
->events
[DDRC_PERF_WRITE_COUNTER_IDX
];
574 prev_count
= local64_read(&hwc
->prev_count
);
575 new_count
= cn10k_ddr_perf_read_counter(pmu
, hwc
->idx
);
577 /* Overflow condition is when new count less than
580 if (new_count
< prev_count
)
581 cn10k_ddr_perf_event_update(event
);
584 for (i
= 0; i
< DDRC_PERF_NUM_GEN_COUNTERS
; i
++) {
585 if (pmu
->events
[i
] == NULL
)
588 value
= cn10k_ddr_perf_read_counter(pmu
, i
);
589 if (value
== DDRC_PERF_CNT_MAX_VALUE
) {
590 pr_info("Counter-(%d) reached max value\n", i
);
591 cn10k_ddr_perf_event_update_all(pmu
);
592 cn10k_ddr_perf_pmu_disable(&pmu
->pmu
);
593 cn10k_ddr_perf_pmu_enable(&pmu
->pmu
);
600 static enum hrtimer_restart
cn10k_ddr_pmu_timer_handler(struct hrtimer
*hrtimer
)
602 struct cn10k_ddr_pmu
*pmu
= container_of(hrtimer
, struct cn10k_ddr_pmu
,
606 local_irq_save(flags
);
607 cn10k_ddr_pmu_overflow_handler(pmu
);
608 local_irq_restore(flags
);
610 hrtimer_forward_now(hrtimer
, cn10k_ddr_pmu_timer_period());
611 return HRTIMER_RESTART
;
614 static int cn10k_ddr_pmu_offline_cpu(unsigned int cpu
, struct hlist_node
*node
)
616 struct cn10k_ddr_pmu
*pmu
= hlist_entry_safe(node
, struct cn10k_ddr_pmu
,
623 target
= cpumask_any_but(cpu_online_mask
, cpu
);
624 if (target
>= nr_cpu_ids
)
627 perf_pmu_migrate_context(&pmu
->pmu
, cpu
, target
);
632 static int cn10k_ddr_perf_probe(struct platform_device
*pdev
)
634 struct cn10k_ddr_pmu
*ddr_pmu
;
635 struct resource
*res
;
640 ddr_pmu
= devm_kzalloc(&pdev
->dev
, sizeof(*ddr_pmu
), GFP_KERNEL
);
644 ddr_pmu
->dev
= &pdev
->dev
;
645 platform_set_drvdata(pdev
, ddr_pmu
);
647 base
= devm_platform_get_and_ioremap_resource(pdev
, 0, &res
);
649 return PTR_ERR(base
);
651 ddr_pmu
->base
= base
;
653 /* Setup the PMU counter to work in manual mode */
654 writeq_relaxed(OP_MODE_CTRL_VAL_MANNUAL
, ddr_pmu
->base
+
655 DDRC_PERF_CNT_OP_MODE_CTRL
);
657 ddr_pmu
->pmu
= (struct pmu
) {
658 .module
= THIS_MODULE
,
659 .capabilities
= PERF_PMU_CAP_NO_EXCLUDE
,
660 .task_ctx_nr
= perf_invalid_context
,
661 .attr_groups
= cn10k_attr_groups
,
662 .event_init
= cn10k_ddr_perf_event_init
,
663 .add
= cn10k_ddr_perf_event_add
,
664 .del
= cn10k_ddr_perf_event_del
,
665 .start
= cn10k_ddr_perf_event_start
,
666 .stop
= cn10k_ddr_perf_event_stop
,
667 .read
= cn10k_ddr_perf_event_update
,
668 .pmu_enable
= cn10k_ddr_perf_pmu_enable
,
669 .pmu_disable
= cn10k_ddr_perf_pmu_disable
,
672 /* Choose this cpu to collect perf data */
673 ddr_pmu
->cpu
= raw_smp_processor_id();
675 name
= devm_kasprintf(ddr_pmu
->dev
, GFP_KERNEL
, "mrvl_ddr_pmu_%llx",
680 hrtimer_init(&ddr_pmu
->hrtimer
, CLOCK_MONOTONIC
, HRTIMER_MODE_REL
);
681 ddr_pmu
->hrtimer
.function
= cn10k_ddr_pmu_timer_handler
;
683 cpuhp_state_add_instance_nocalls(
684 CPUHP_AP_PERF_ARM_MARVELL_CN10K_DDR_ONLINE
,
687 ret
= perf_pmu_register(&ddr_pmu
->pmu
, name
, -1);
691 pr_info("CN10K DDR PMU Driver for ddrc@%llx\n", res
->start
);
694 cpuhp_state_remove_instance_nocalls(
695 CPUHP_AP_PERF_ARM_MARVELL_CN10K_DDR_ONLINE
,
700 static void cn10k_ddr_perf_remove(struct platform_device
*pdev
)
702 struct cn10k_ddr_pmu
*ddr_pmu
= platform_get_drvdata(pdev
);
704 cpuhp_state_remove_instance_nocalls(
705 CPUHP_AP_PERF_ARM_MARVELL_CN10K_DDR_ONLINE
,
708 perf_pmu_unregister(&ddr_pmu
->pmu
);
712 static const struct of_device_id cn10k_ddr_pmu_of_match
[] = {
713 { .compatible
= "marvell,cn10k-ddr-pmu", },
716 MODULE_DEVICE_TABLE(of
, cn10k_ddr_pmu_of_match
);
720 static const struct acpi_device_id cn10k_ddr_pmu_acpi_match
[] = {
724 MODULE_DEVICE_TABLE(acpi
, cn10k_ddr_pmu_acpi_match
);
727 static struct platform_driver cn10k_ddr_pmu_driver
= {
729 .name
= "cn10k-ddr-pmu",
730 .of_match_table
= of_match_ptr(cn10k_ddr_pmu_of_match
),
731 .acpi_match_table
= ACPI_PTR(cn10k_ddr_pmu_acpi_match
),
732 .suppress_bind_attrs
= true,
734 .probe
= cn10k_ddr_perf_probe
,
735 .remove
= cn10k_ddr_perf_remove
,
738 static int __init
cn10k_ddr_pmu_init(void)
742 ret
= cpuhp_setup_state_multi(
743 CPUHP_AP_PERF_ARM_MARVELL_CN10K_DDR_ONLINE
,
744 "perf/marvell/cn10k/ddr:online", NULL
,
745 cn10k_ddr_pmu_offline_cpu
);
749 ret
= platform_driver_register(&cn10k_ddr_pmu_driver
);
751 cpuhp_remove_multi_state(
752 CPUHP_AP_PERF_ARM_MARVELL_CN10K_DDR_ONLINE
);
756 static void __exit
cn10k_ddr_pmu_exit(void)
758 platform_driver_unregister(&cn10k_ddr_pmu_driver
);
759 cpuhp_remove_multi_state(CPUHP_AP_PERF_ARM_MARVELL_CN10K_DDR_ONLINE
);
762 module_init(cn10k_ddr_pmu_init
);
763 module_exit(cn10k_ddr_pmu_exit
);
765 MODULE_AUTHOR("Bharat Bhushan <bbhushan2@marvell.com>");
766 MODULE_DESCRIPTION("Marvell CN10K DRAM Subsystem (DSS) Performance Monitor Driver");
767 MODULE_LICENSE("GPL v2");