2 * OMAP4+ CPU idle Routines
4 * Copyright (C) 2011-2013 Texas Instruments, Inc.
5 * Santosh Shilimkar <santosh.shilimkar@ti.com>
6 * Rajendra Nayak <rnayak@ti.com>
8 * This program is free software; you can redistribute it and/or modify
9 * it under the terms of the GNU General Public License version 2 as
10 * published by the Free Software Foundation.
13 #include <linux/sched.h>
14 #include <linux/cpuidle.h>
15 #include <linux/cpu_pm.h>
16 #include <linux/export.h>
17 #include <linux/tick.h>
19 #include <asm/cpuidle.h>
25 #include "clockdomain.h"
29 /* Machine specific information */
30 struct idle_statedata
{
37 static struct idle_statedata omap4_idle_data
[] = {
39 .cpu_state
= PWRDM_POWER_ON
,
40 .mpu_state
= PWRDM_POWER_ON
,
41 .mpu_logic_state
= PWRDM_POWER_RET
,
44 .cpu_state
= PWRDM_POWER_OFF
,
45 .mpu_state
= PWRDM_POWER_RET
,
46 .mpu_logic_state
= PWRDM_POWER_RET
,
49 .cpu_state
= PWRDM_POWER_OFF
,
50 .mpu_state
= PWRDM_POWER_RET
,
51 .mpu_logic_state
= PWRDM_POWER_OFF
,
55 static struct idle_statedata omap5_idle_data
[] = {
57 .cpu_state
= PWRDM_POWER_ON
,
58 .mpu_state
= PWRDM_POWER_ON
,
59 .mpu_logic_state
= PWRDM_POWER_ON
,
62 .cpu_state
= PWRDM_POWER_RET
,
63 .mpu_state
= PWRDM_POWER_RET
,
64 .mpu_logic_state
= PWRDM_POWER_RET
,
68 static struct powerdomain
*mpu_pd
, *cpu_pd
[MAX_CPUS
];
69 static struct clockdomain
*cpu_clkdm
[MAX_CPUS
];
71 static atomic_t abort_barrier
;
72 static bool cpu_done
[MAX_CPUS
];
73 static struct idle_statedata
*state_ptr
= &omap4_idle_data
[0];
74 static DEFINE_RAW_SPINLOCK(mpu_lock
);
76 /* Private functions */
79 * omap_enter_idle_[simple/coupled] - OMAP4PLUS cpuidle entry functions
80 * @dev: cpuidle device
81 * @drv: cpuidle driver
82 * @index: the index of state to be entered
84 * Called from the CPUidle framework to program the device to the
85 * specified low power state selected by the governor.
86 * Returns the amount of time spent in the low power state.
88 static int omap_enter_idle_simple(struct cpuidle_device
*dev
,
89 struct cpuidle_driver
*drv
,
96 static int omap_enter_idle_smp(struct cpuidle_device
*dev
,
97 struct cpuidle_driver
*drv
,
100 struct idle_statedata
*cx
= state_ptr
+ index
;
103 raw_spin_lock_irqsave(&mpu_lock
, flag
);
104 cx
->mpu_state_vote
++;
105 if (cx
->mpu_state_vote
== num_online_cpus()) {
106 pwrdm_set_logic_retst(mpu_pd
, cx
->mpu_logic_state
);
107 omap_set_pwrdm_state(mpu_pd
, cx
->mpu_state
);
109 raw_spin_unlock_irqrestore(&mpu_lock
, flag
);
111 omap4_enter_lowpower(dev
->cpu
, cx
->cpu_state
);
113 raw_spin_lock_irqsave(&mpu_lock
, flag
);
114 if (cx
->mpu_state_vote
== num_online_cpus())
115 omap_set_pwrdm_state(mpu_pd
, PWRDM_POWER_ON
);
116 cx
->mpu_state_vote
--;
117 raw_spin_unlock_irqrestore(&mpu_lock
, flag
);
122 static int omap_enter_idle_coupled(struct cpuidle_device
*dev
,
123 struct cpuidle_driver
*drv
,
126 struct idle_statedata
*cx
= state_ptr
+ index
;
127 u32 mpuss_can_lose_context
= 0;
130 * CPU0 has to wait and stay ON until CPU1 is OFF state.
131 * This is necessary to honour hardware recommondation
132 * of triggeing all the possible low power modes once CPU1 is
133 * out of coherency and in OFF mode.
135 if (dev
->cpu
== 0 && cpumask_test_cpu(1, cpu_online_mask
)) {
136 while (pwrdm_read_pwrst(cpu_pd
[1]) != PWRDM_POWER_OFF
) {
140 * CPU1 could have already entered & exited idle
141 * without hitting off because of a wakeup
142 * or a failed attempt to hit off mode. Check for
143 * that here, otherwise we could spin forever
144 * waiting for CPU1 off.
152 mpuss_can_lose_context
= (cx
->mpu_state
== PWRDM_POWER_RET
) &&
153 (cx
->mpu_logic_state
== PWRDM_POWER_OFF
);
155 tick_broadcast_enter();
158 * Call idle CPU PM enter notifier chain so that
159 * VFP and per CPU interrupt context is saved.
164 pwrdm_set_logic_retst(mpu_pd
, cx
->mpu_logic_state
);
165 omap_set_pwrdm_state(mpu_pd
, cx
->mpu_state
);
168 * Call idle CPU cluster PM enter notifier chain
169 * to save GIC and wakeupgen context.
171 if (mpuss_can_lose_context
)
172 cpu_cluster_pm_enter();
175 omap4_enter_lowpower(dev
->cpu
, cx
->cpu_state
);
176 cpu_done
[dev
->cpu
] = true;
178 /* Wakeup CPU1 only if it is not offlined */
179 if (dev
->cpu
== 0 && cpumask_test_cpu(1, cpu_online_mask
)) {
181 if (IS_PM44XX_ERRATUM(PM_OMAP4_ROM_SMP_BOOT_ERRATUM_GICD
) &&
182 mpuss_can_lose_context
)
185 clkdm_deny_idle(cpu_clkdm
[1]);
186 omap_set_pwrdm_state(cpu_pd
[1], PWRDM_POWER_ON
);
187 clkdm_allow_idle(cpu_clkdm
[1]);
189 if (IS_PM44XX_ERRATUM(PM_OMAP4_ROM_SMP_BOOT_ERRATUM_GICD
) &&
190 mpuss_can_lose_context
) {
191 while (gic_dist_disabled()) {
195 gic_timer_retrigger();
200 * Call idle CPU PM exit notifier chain to restore
201 * VFP and per CPU IRQ context.
206 * Call idle CPU cluster PM exit notifier chain
207 * to restore GIC and wakeupgen context.
209 if (dev
->cpu
== 0 && mpuss_can_lose_context
)
210 cpu_cluster_pm_exit();
212 tick_broadcast_exit();
215 cpuidle_coupled_parallel_barrier(dev
, &abort_barrier
);
216 cpu_done
[dev
->cpu
] = false;
222 * For each cpu, setup the broadcast timer because local timers
223 * stops for the states above C1.
225 static void omap_setup_broadcast_timer(void *arg
)
227 tick_broadcast_enable();
230 static struct cpuidle_driver omap4_idle_driver
= {
231 .name
= "omap4_idle",
232 .owner
= THIS_MODULE
,
235 /* C1 - CPU0 ON + CPU1 ON + MPU ON */
236 .exit_latency
= 2 + 2,
237 .target_residency
= 5,
238 .enter
= omap_enter_idle_simple
,
240 .desc
= "CPUx ON, MPUSS ON"
243 /* C2 - CPU0 OFF + CPU1 OFF + MPU CSWR */
244 .exit_latency
= 328 + 440,
245 .target_residency
= 960,
246 .flags
= CPUIDLE_FLAG_COUPLED
,
247 .enter
= omap_enter_idle_coupled
,
249 .desc
= "CPUx OFF, MPUSS CSWR",
252 /* C3 - CPU0 OFF + CPU1 OFF + MPU OSWR */
253 .exit_latency
= 460 + 518,
254 .target_residency
= 1100,
255 .flags
= CPUIDLE_FLAG_COUPLED
,
256 .enter
= omap_enter_idle_coupled
,
258 .desc
= "CPUx OFF, MPUSS OSWR",
261 .state_count
= ARRAY_SIZE(omap4_idle_data
),
262 .safe_state_index
= 0,
265 static struct cpuidle_driver omap5_idle_driver
= {
266 .name
= "omap5_idle",
267 .owner
= THIS_MODULE
,
270 /* C1 - CPU0 ON + CPU1 ON + MPU ON */
271 .exit_latency
= 2 + 2,
272 .target_residency
= 5,
273 .enter
= omap_enter_idle_simple
,
275 .desc
= "CPUx WFI, MPUSS ON"
278 /* C2 - CPU0 RET + CPU1 RET + MPU CSWR */
279 .exit_latency
= 48 + 60,
280 .target_residency
= 100,
281 .flags
= CPUIDLE_FLAG_TIMER_STOP
,
282 .enter
= omap_enter_idle_smp
,
284 .desc
= "CPUx CSWR, MPUSS CSWR",
287 .state_count
= ARRAY_SIZE(omap5_idle_data
),
288 .safe_state_index
= 0,
291 /* Public functions */
294 * omap4_idle_init - Init routine for OMAP4+ idle
296 * Registers the OMAP4+ specific cpuidle driver to the cpuidle
297 * framework with the valid set of states.
299 int __init
omap4_idle_init(void)
301 struct cpuidle_driver
*idle_driver
;
303 if (soc_is_omap54xx()) {
304 state_ptr
= &omap5_idle_data
[0];
305 idle_driver
= &omap5_idle_driver
;
307 state_ptr
= &omap4_idle_data
[0];
308 idle_driver
= &omap4_idle_driver
;
311 mpu_pd
= pwrdm_lookup("mpu_pwrdm");
312 cpu_pd
[0] = pwrdm_lookup("cpu0_pwrdm");
313 cpu_pd
[1] = pwrdm_lookup("cpu1_pwrdm");
314 if ((!mpu_pd
) || (!cpu_pd
[0]) || (!cpu_pd
[1]))
317 cpu_clkdm
[0] = clkdm_lookup("mpu0_clkdm");
318 cpu_clkdm
[1] = clkdm_lookup("mpu1_clkdm");
319 if (!cpu_clkdm
[0] || !cpu_clkdm
[1])
322 /* Configure the broadcast timer on each cpu */
323 on_each_cpu(omap_setup_broadcast_timer
, NULL
, 1);
325 return cpuidle_register(idle_driver
, cpu_online_mask
);