1 // SPDX-License-Identifier: GPL-2.0-only
3 * OMAP4+ CPU idle Routines
5 * Copyright (C) 2011-2013 Texas Instruments, Inc.
6 * Santosh Shilimkar <santosh.shilimkar@ti.com>
7 * Rajendra Nayak <rnayak@ti.com>
10 #include <linux/sched.h>
11 #include <linux/cpuidle.h>
12 #include <linux/cpu_pm.h>
13 #include <linux/export.h>
14 #include <linux/tick.h>
16 #include <asm/cpuidle.h>
22 #include "clockdomain.h"
26 /* Machine specific information */
27 struct idle_statedata
{
34 static struct idle_statedata omap4_idle_data
[] = {
36 .cpu_state
= PWRDM_POWER_ON
,
37 .mpu_state
= PWRDM_POWER_ON
,
38 .mpu_logic_state
= PWRDM_POWER_RET
,
41 .cpu_state
= PWRDM_POWER_OFF
,
42 .mpu_state
= PWRDM_POWER_RET
,
43 .mpu_logic_state
= PWRDM_POWER_RET
,
46 .cpu_state
= PWRDM_POWER_OFF
,
47 .mpu_state
= PWRDM_POWER_RET
,
48 .mpu_logic_state
= PWRDM_POWER_OFF
,
52 static struct idle_statedata omap5_idle_data
[] = {
54 .cpu_state
= PWRDM_POWER_ON
,
55 .mpu_state
= PWRDM_POWER_ON
,
56 .mpu_logic_state
= PWRDM_POWER_ON
,
59 .cpu_state
= PWRDM_POWER_RET
,
60 .mpu_state
= PWRDM_POWER_RET
,
61 .mpu_logic_state
= PWRDM_POWER_RET
,
65 static struct powerdomain
*mpu_pd
, *cpu_pd
[MAX_CPUS
];
66 static struct clockdomain
*cpu_clkdm
[MAX_CPUS
];
68 static atomic_t abort_barrier
;
69 static bool cpu_done
[MAX_CPUS
];
70 static struct idle_statedata
*state_ptr
= &omap4_idle_data
[0];
71 static DEFINE_RAW_SPINLOCK(mpu_lock
);
73 /* Private functions */
76 * omap_enter_idle_[simple/coupled] - OMAP4PLUS cpuidle entry functions
77 * @dev: cpuidle device
78 * @drv: cpuidle driver
79 * @index: the index of state to be entered
81 * Called from the CPUidle framework to program the device to the
82 * specified low power state selected by the governor.
83 * Returns the amount of time spent in the low power state.
85 static int omap_enter_idle_simple(struct cpuidle_device
*dev
,
86 struct cpuidle_driver
*drv
,
93 static int omap_enter_idle_smp(struct cpuidle_device
*dev
,
94 struct cpuidle_driver
*drv
,
97 struct idle_statedata
*cx
= state_ptr
+ index
;
100 raw_spin_lock_irqsave(&mpu_lock
, flag
);
101 cx
->mpu_state_vote
++;
102 if (cx
->mpu_state_vote
== num_online_cpus()) {
103 pwrdm_set_logic_retst(mpu_pd
, cx
->mpu_logic_state
);
104 omap_set_pwrdm_state(mpu_pd
, cx
->mpu_state
);
106 raw_spin_unlock_irqrestore(&mpu_lock
, flag
);
108 omap4_enter_lowpower(dev
->cpu
, cx
->cpu_state
, true);
110 raw_spin_lock_irqsave(&mpu_lock
, flag
);
111 if (cx
->mpu_state_vote
== num_online_cpus())
112 omap_set_pwrdm_state(mpu_pd
, PWRDM_POWER_ON
);
113 cx
->mpu_state_vote
--;
114 raw_spin_unlock_irqrestore(&mpu_lock
, flag
);
119 static int omap_enter_idle_coupled(struct cpuidle_device
*dev
,
120 struct cpuidle_driver
*drv
,
123 struct idle_statedata
*cx
= state_ptr
+ index
;
124 u32 mpuss_can_lose_context
= 0;
128 * CPU0 has to wait and stay ON until CPU1 is OFF state.
129 * This is necessary to honour hardware recommondation
130 * of triggeing all the possible low power modes once CPU1 is
131 * out of coherency and in OFF mode.
133 if (dev
->cpu
== 0 && cpumask_test_cpu(1, cpu_online_mask
)) {
134 while (pwrdm_read_pwrst(cpu_pd
[1]) != PWRDM_POWER_OFF
) {
138 * CPU1 could have already entered & exited idle
139 * without hitting off because of a wakeup
140 * or a failed attempt to hit off mode. Check for
141 * that here, otherwise we could spin forever
142 * waiting for CPU1 off.
150 mpuss_can_lose_context
= (cx
->mpu_state
== PWRDM_POWER_RET
) &&
151 (cx
->mpu_logic_state
== PWRDM_POWER_OFF
);
153 /* Enter broadcast mode for periodic timers */
154 tick_broadcast_enable();
156 /* Enter broadcast mode for one-shot timers */
157 tick_broadcast_enter();
160 * Call idle CPU PM enter notifier chain so that
161 * VFP and per CPU interrupt context is saved.
163 error
= cpu_pm_enter();
168 pwrdm_set_logic_retst(mpu_pd
, cx
->mpu_logic_state
);
169 omap_set_pwrdm_state(mpu_pd
, cx
->mpu_state
);
172 * Call idle CPU cluster PM enter notifier chain
173 * to save GIC and wakeupgen context.
175 if (mpuss_can_lose_context
) {
176 error
= cpu_cluster_pm_enter();
179 cx
= state_ptr
+ index
;
180 pwrdm_set_logic_retst(mpu_pd
, cx
->mpu_logic_state
);
181 omap_set_pwrdm_state(mpu_pd
, cx
->mpu_state
);
182 mpuss_can_lose_context
= 0;
187 omap4_enter_lowpower(dev
->cpu
, cx
->cpu_state
, true);
188 cpu_done
[dev
->cpu
] = true;
190 /* Wakeup CPU1 only if it is not offlined */
191 if (dev
->cpu
== 0 && cpumask_test_cpu(1, cpu_online_mask
)) {
193 if (IS_PM44XX_ERRATUM(PM_OMAP4_ROM_SMP_BOOT_ERRATUM_GICD
) &&
194 mpuss_can_lose_context
)
197 clkdm_deny_idle(cpu_clkdm
[1]);
198 omap_set_pwrdm_state(cpu_pd
[1], PWRDM_POWER_ON
);
199 clkdm_allow_idle(cpu_clkdm
[1]);
201 if (IS_PM44XX_ERRATUM(PM_OMAP4_ROM_SMP_BOOT_ERRATUM_GICD
) &&
202 mpuss_can_lose_context
) {
203 while (gic_dist_disabled()) {
207 gic_timer_retrigger();
212 * Call idle CPU cluster PM exit notifier chain
213 * to restore GIC and wakeupgen context.
215 if (dev
->cpu
== 0 && mpuss_can_lose_context
)
216 cpu_cluster_pm_exit();
219 * Call idle CPU PM exit notifier chain to restore
220 * VFP and per CPU IRQ context.
225 tick_broadcast_exit();
228 cpuidle_coupled_parallel_barrier(dev
, &abort_barrier
);
229 cpu_done
[dev
->cpu
] = false;
234 static struct cpuidle_driver omap4_idle_driver
= {
235 .name
= "omap4_idle",
236 .owner
= THIS_MODULE
,
239 /* C1 - CPU0 ON + CPU1 ON + MPU ON */
240 .exit_latency
= 2 + 2,
241 .target_residency
= 5,
242 .enter
= omap_enter_idle_simple
,
244 .desc
= "CPUx ON, MPUSS ON"
247 /* C2 - CPU0 OFF + CPU1 OFF + MPU CSWR */
248 .exit_latency
= 328 + 440,
249 .target_residency
= 960,
250 .flags
= CPUIDLE_FLAG_COUPLED
|
251 CPUIDLE_FLAG_RCU_IDLE
,
252 .enter
= omap_enter_idle_coupled
,
254 .desc
= "CPUx OFF, MPUSS CSWR",
257 /* C3 - CPU0 OFF + CPU1 OFF + MPU OSWR */
258 .exit_latency
= 460 + 518,
259 .target_residency
= 1100,
260 .flags
= CPUIDLE_FLAG_COUPLED
|
261 CPUIDLE_FLAG_RCU_IDLE
,
262 .enter
= omap_enter_idle_coupled
,
264 .desc
= "CPUx OFF, MPUSS OSWR",
267 .state_count
= ARRAY_SIZE(omap4_idle_data
),
268 .safe_state_index
= 0,
271 static struct cpuidle_driver omap5_idle_driver
= {
272 .name
= "omap5_idle",
273 .owner
= THIS_MODULE
,
276 /* C1 - CPU0 ON + CPU1 ON + MPU ON */
277 .exit_latency
= 2 + 2,
278 .target_residency
= 5,
279 .enter
= omap_enter_idle_simple
,
281 .desc
= "CPUx WFI, MPUSS ON"
284 /* C2 - CPU0 RET + CPU1 RET + MPU CSWR */
285 .exit_latency
= 48 + 60,
286 .target_residency
= 100,
287 .flags
= CPUIDLE_FLAG_TIMER_STOP
|
288 CPUIDLE_FLAG_RCU_IDLE
,
289 .enter
= omap_enter_idle_smp
,
291 .desc
= "CPUx CSWR, MPUSS CSWR",
294 .state_count
= ARRAY_SIZE(omap5_idle_data
),
295 .safe_state_index
= 0,
298 /* Public functions */
301 * omap4_idle_init - Init routine for OMAP4+ idle
303 * Registers the OMAP4+ specific cpuidle driver to the cpuidle
304 * framework with the valid set of states.
306 int __init
omap4_idle_init(void)
308 struct cpuidle_driver
*idle_driver
;
310 if (soc_is_omap54xx()) {
311 state_ptr
= &omap5_idle_data
[0];
312 idle_driver
= &omap5_idle_driver
;
314 state_ptr
= &omap4_idle_data
[0];
315 idle_driver
= &omap4_idle_driver
;
318 mpu_pd
= pwrdm_lookup("mpu_pwrdm");
319 cpu_pd
[0] = pwrdm_lookup("cpu0_pwrdm");
320 cpu_pd
[1] = pwrdm_lookup("cpu1_pwrdm");
321 if ((!mpu_pd
) || (!cpu_pd
[0]) || (!cpu_pd
[1]))
324 cpu_clkdm
[0] = clkdm_lookup("mpu0_clkdm");
325 cpu_clkdm
[1] = clkdm_lookup("mpu1_clkdm");
326 if (!cpu_clkdm
[0] || !cpu_clkdm
[1])
329 return cpuidle_register(idle_driver
, cpu_online_mask
);