]> git.kernelconcepts.de Git - karo-tx-linux.git/blob - arch/arm/mach-omap2/cpuidle44xx.c
Merge branch 'for_3.6/pm/coupled-cpuidle' of git://git.kernel.org/pub/scm/linux/kerne...
[karo-tx-linux.git] / arch / arm / mach-omap2 / cpuidle44xx.c
1 /*
2  * OMAP4 CPU idle Routines
3  *
4  * Copyright (C) 2011 Texas Instruments, Inc.
5  * Santosh Shilimkar <santosh.shilimkar@ti.com>
6  * Rajendra Nayak <rnayak@ti.com>
7  *
8  * This program is free software; you can redistribute it and/or modify
9  * it under the terms of the GNU General Public License version 2 as
10  * published by the Free Software Foundation.
11  */
12
13 #include <linux/sched.h>
14 #include <linux/cpuidle.h>
15 #include <linux/cpu_pm.h>
16 #include <linux/export.h>
17 #include <linux/clockchips.h>
18
19 #include <asm/proc-fns.h>
20
21 #include "common.h"
22 #include "pm.h"
23 #include "prm.h"
24 #include "clockdomain.h"
25
26 #ifdef CONFIG_CPU_IDLE
27
28 /* Machine specific information */
29 struct omap4_idle_statedata {
30         u32 cpu_state;
31         u32 mpu_logic_state;
32         u32 mpu_state;
33 };
34
35 static struct omap4_idle_statedata omap4_idle_data[] = {
36         {
37                 .cpu_state = PWRDM_POWER_ON,
38                 .mpu_state = PWRDM_POWER_ON,
39                 .mpu_logic_state = PWRDM_POWER_RET,
40         },
41         {
42                 .cpu_state = PWRDM_POWER_OFF,
43                 .mpu_state = PWRDM_POWER_RET,
44                 .mpu_logic_state = PWRDM_POWER_RET,
45         },
46         {
47                 .cpu_state = PWRDM_POWER_OFF,
48                 .mpu_state = PWRDM_POWER_RET,
49                 .mpu_logic_state = PWRDM_POWER_OFF,
50         },
51 };
52
53 static struct powerdomain *mpu_pd, *cpu_pd[NR_CPUS];
54 static struct clockdomain *cpu_clkdm[NR_CPUS];
55
56 static atomic_t abort_barrier;
57 static bool cpu_done[NR_CPUS];
58
59 /**
60  * omap4_enter_idle_coupled_[simple/coupled] - OMAP4 cpuidle entry functions
61  * @dev: cpuidle device
62  * @drv: cpuidle driver
63  * @index: the index of state to be entered
64  *
65  * Called from the CPUidle framework to program the device to the
66  * specified low power state selected by the governor.
67  * Returns the amount of time spent in the low power state.
68  */
69 static int omap4_enter_idle_simple(struct cpuidle_device *dev,
70                         struct cpuidle_driver *drv,
71                         int index)
72 {
73         local_fiq_disable();
74         omap_do_wfi();
75         local_fiq_enable();
76
77         return index;
78 }
79
80 static int omap4_enter_idle_coupled(struct cpuidle_device *dev,
81                         struct cpuidle_driver *drv,
82                         int index)
83 {
84         struct omap4_idle_statedata *cx = &omap4_idle_data[index];
85         int cpu_id = smp_processor_id();
86
87         local_fiq_disable();
88
89         /*
90          * CPU0 has to wait and stay ON until CPU1 is OFF state.
91          * This is necessary to honour hardware recommondation
92          * of triggeing all the possible low power modes once CPU1 is
93          * out of coherency and in OFF mode.
94          */
95         if (dev->cpu == 0 && cpumask_test_cpu(1, cpu_online_mask)) {
96                 while (pwrdm_read_pwrst(cpu_pd[1]) != PWRDM_POWER_OFF) {
97                         cpu_relax();
98
99                         /*
100                          * CPU1 could have already entered & exited idle
101                          * without hitting off because of a wakeup
102                          * or a failed attempt to hit off mode.  Check for
103                          * that here, otherwise we could spin forever
104                          * waiting for CPU1 off.
105                          */
106                         if (cpu_done[1])
107                             goto fail;
108
109                 }
110         }
111
112         clockevents_notify(CLOCK_EVT_NOTIFY_BROADCAST_ENTER, &cpu_id);
113
114         /*
115          * Call idle CPU PM enter notifier chain so that
116          * VFP and per CPU interrupt context is saved.
117          */
118         cpu_pm_enter();
119
120         if (dev->cpu == 0) {
121                 pwrdm_set_logic_retst(mpu_pd, cx->mpu_logic_state);
122                 omap_set_pwrdm_state(mpu_pd, cx->mpu_state);
123
124                 /*
125                  * Call idle CPU cluster PM enter notifier chain
126                  * to save GIC and wakeupgen context.
127                  */
128                 if ((cx->mpu_state == PWRDM_POWER_RET) &&
129                         (cx->mpu_logic_state == PWRDM_POWER_OFF))
130                                 cpu_cluster_pm_enter();
131         }
132
133         omap4_enter_lowpower(dev->cpu, cx->cpu_state);
134         cpu_done[dev->cpu] = true;
135
136         /* Wakeup CPU1 only if it is not offlined */
137         if (dev->cpu == 0 && cpumask_test_cpu(1, cpu_online_mask)) {
138                 clkdm_wakeup(cpu_clkdm[1]);
139                 clkdm_allow_idle(cpu_clkdm[1]);
140         }
141
142         /*
143          * Call idle CPU PM exit notifier chain to restore
144          * VFP and per CPU IRQ context.
145          */
146         cpu_pm_exit();
147
148         /*
149          * Call idle CPU cluster PM exit notifier chain
150          * to restore GIC and wakeupgen context.
151          */
152         if (omap4_mpuss_read_prev_context_state())
153                 cpu_cluster_pm_exit();
154
155         clockevents_notify(CLOCK_EVT_NOTIFY_BROADCAST_EXIT, &cpu_id);
156
157 fail:
158         cpuidle_coupled_parallel_barrier(dev, &abort_barrier);
159         cpu_done[dev->cpu] = false;
160
161         local_fiq_enable();
162
163         return index;
164 }
165
166 DEFINE_PER_CPU(struct cpuidle_device, omap4_idle_dev);
167
168 struct cpuidle_driver omap4_idle_driver = {
169         .name                           = "omap4_idle",
170         .owner                          = THIS_MODULE,
171         .en_core_tk_irqen               = 1,
172         .states = {
173                 {
174                         /* C1 - CPU0 ON + CPU1 ON + MPU ON */
175                         .exit_latency = 2 + 2,
176                         .target_residency = 5,
177                         .flags = CPUIDLE_FLAG_TIME_VALID,
178                         .enter = omap4_enter_idle_simple,
179                         .name = "C1",
180                         .desc = "MPUSS ON"
181                 },
182                 {
183                         /* C2 - CPU0 OFF + CPU1 OFF + MPU CSWR */
184                         .exit_latency = 328 + 440,
185                         .target_residency = 960,
186                         .flags = CPUIDLE_FLAG_TIME_VALID | CPUIDLE_FLAG_COUPLED,
187                         .enter = omap4_enter_idle_coupled,
188                         .name = "C2",
189                         .desc = "MPUSS CSWR",
190                 },
191                 {
192                         /* C3 - CPU0 OFF + CPU1 OFF + MPU OSWR */
193                         .exit_latency = 460 + 518,
194                         .target_residency = 1100,
195                         .flags = CPUIDLE_FLAG_TIME_VALID | CPUIDLE_FLAG_COUPLED,
196                         .enter = omap4_enter_idle_coupled,
197                         .name = "C3",
198                         .desc = "MPUSS OSWR",
199                 },
200         },
201         .state_count = ARRAY_SIZE(omap4_idle_data),
202         .safe_state_index = 0,
203 };
204
205 /*
206  * For each cpu, setup the broadcast timer because local timers
207  * stops for the states above C1.
208  */
209 static void omap_setup_broadcast_timer(void *arg)
210 {
211         int cpu = smp_processor_id();
212         clockevents_notify(CLOCK_EVT_NOTIFY_BROADCAST_ON, &cpu);
213 }
214
215 /**
216  * omap4_idle_init - Init routine for OMAP4 idle
217  *
218  * Registers the OMAP4 specific cpuidle driver to the cpuidle
219  * framework with the valid set of states.
220  */
221 int __init omap4_idle_init(void)
222 {
223         struct cpuidle_device *dev;
224         unsigned int cpu_id = 0;
225
226         mpu_pd = pwrdm_lookup("mpu_pwrdm");
227         cpu_pd[0] = pwrdm_lookup("cpu0_pwrdm");
228         cpu_pd[1] = pwrdm_lookup("cpu1_pwrdm");
229         if ((!mpu_pd) || (!cpu_pd[0]) || (!cpu_pd[1]))
230                 return -ENODEV;
231
232         cpu_clkdm[0] = clkdm_lookup("mpu0_clkdm");
233         cpu_clkdm[1] = clkdm_lookup("mpu1_clkdm");
234         if (!cpu_clkdm[0] || !cpu_clkdm[1])
235                 return -ENODEV;
236
237         /* Configure the broadcast timer on each cpu */
238         on_each_cpu(omap_setup_broadcast_timer, NULL, 1);
239
240         for_each_cpu(cpu_id, cpu_online_mask) {
241                 dev = &per_cpu(omap4_idle_dev, cpu_id);
242                 dev->cpu = cpu_id;
243                 dev->coupled_cpus = *cpu_online_mask;
244
245                 cpuidle_register_driver(&omap4_idle_driver);
246
247                 if (cpuidle_register_device(dev)) {
248                         pr_err("%s: CPUidle register failed\n", __func__);
249                         return -EIO;
250                 }
251         }
252
253         return 0;
254 }
255 #else
256 int __init omap4_idle_init(void)
257 {
258         return 0;
259 }
260 #endif /* CONFIG_CPU_IDLE */