2 * processor_thermal.c - Passive cooling submodule of the ACPI processor driver
4 * Copyright (C) 2001, 2002 Andy Grover <andrew.grover@intel.com>
5 * Copyright (C) 2001, 2002 Paul Diefenbaugh <paul.s.diefenbaugh@intel.com>
6 * Copyright (C) 2004 Dominik Brodowski <linux@brodo.de>
7 * Copyright (C) 2004 Anil S Keshavamurthy <anil.s.keshavamurthy@intel.com>
8 * - Added processor hotplug support
10 * ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
12 * This program is free software; you can redistribute it and/or modify
13 * it under the terms of the GNU General Public License as published by
14 * the Free Software Foundation; either version 2 of the License, or (at
15 * your option) any later version.
17 * This program is distributed in the hope that it will be useful, but
18 * WITHOUT ANY WARRANTY; without even the implied warranty of
19 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
20 * General Public License for more details.
22 * ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
25 #include <linux/kernel.h>
26 #include <linux/module.h>
27 #include <linux/init.h>
28 #include <linux/cpufreq.h>
29 #include <linux/acpi.h>
30 #include <acpi/processor.h>
31 #include <linux/uaccess.h>
33 #define PREFIX "ACPI: "
35 #define ACPI_PROCESSOR_CLASS "processor"
36 #define _COMPONENT ACPI_PROCESSOR_COMPONENT
37 ACPI_MODULE_NAME("processor_thermal");
39 #ifdef CONFIG_CPU_FREQ
41 /* If a passive cooling situation is detected, primarily CPUfreq is used, as it
42 * offers (in most cases) voltage scaling in addition to frequency scaling, and
43 * thus a cubic (instead of linear) reduction of energy. Also, we allow for
44 * _any_ cpufreq driver and not only the acpi-cpufreq driver.
47 #define CPUFREQ_THERMAL_MIN_STEP 0
48 #define CPUFREQ_THERMAL_MAX_STEP 3
50 static DEFINE_PER_CPU(unsigned int, cpufreq_thermal_reduction_pctg
);
51 static unsigned int acpi_thermal_cpufreq_is_init
= 0;
53 #define reduction_pctg(cpu) \
54 per_cpu(cpufreq_thermal_reduction_pctg, phys_package_first_cpu(cpu))
57 * Emulate "per package data" using per cpu data (which should really be
60 * Note we can lose a CPU on cpu hotunplug, in this case we forget the state
61 * temporarily. Fortunately that's not a big issue here (I hope)
63 static int phys_package_first_cpu(int cpu
)
66 int id
= topology_physical_package_id(cpu
);
68 for_each_online_cpu(i
)
69 if (topology_physical_package_id(i
) == id
)
74 static int cpu_has_cpufreq(unsigned int cpu
)
76 struct cpufreq_policy policy
;
77 if (!acpi_thermal_cpufreq_is_init
|| cpufreq_get_policy(&policy
, cpu
))
82 static int acpi_thermal_cpufreq_notifier(struct notifier_block
*nb
,
83 unsigned long event
, void *data
)
85 struct cpufreq_policy
*policy
= data
;
86 unsigned long max_freq
= 0;
88 if (event
!= CPUFREQ_ADJUST
)
92 policy
->cpuinfo
.max_freq
*
93 (100 - reduction_pctg(policy
->cpu
) * 20)
96 cpufreq_verify_within_limits(policy
, 0, max_freq
);
102 static struct notifier_block acpi_thermal_cpufreq_notifier_block
= {
103 .notifier_call
= acpi_thermal_cpufreq_notifier
,
106 static int cpufreq_get_max_state(unsigned int cpu
)
108 if (!cpu_has_cpufreq(cpu
))
111 return CPUFREQ_THERMAL_MAX_STEP
;
114 static int cpufreq_get_cur_state(unsigned int cpu
)
116 if (!cpu_has_cpufreq(cpu
))
119 return reduction_pctg(cpu
);
122 static int cpufreq_set_cur_state(unsigned int cpu
, int state
)
126 if (!cpu_has_cpufreq(cpu
))
129 reduction_pctg(cpu
) = state
;
132 * Update all the CPUs in the same package because they all
133 * contribute to the temperature and often share the same
136 for_each_online_cpu(i
) {
137 if (topology_physical_package_id(i
) ==
138 topology_physical_package_id(cpu
))
139 cpufreq_update_policy(i
);
144 void acpi_thermal_cpufreq_init(void)
148 i
= cpufreq_register_notifier(&acpi_thermal_cpufreq_notifier_block
,
149 CPUFREQ_POLICY_NOTIFIER
);
151 acpi_thermal_cpufreq_is_init
= 1;
154 void acpi_thermal_cpufreq_exit(void)
156 if (acpi_thermal_cpufreq_is_init
)
157 cpufreq_unregister_notifier
158 (&acpi_thermal_cpufreq_notifier_block
,
159 CPUFREQ_POLICY_NOTIFIER
);
161 acpi_thermal_cpufreq_is_init
= 0;
164 #else /* ! CONFIG_CPU_FREQ */
165 static int cpufreq_get_max_state(unsigned int cpu
)
170 static int cpufreq_get_cur_state(unsigned int cpu
)
175 static int cpufreq_set_cur_state(unsigned int cpu
, int state
)
182 /* thermal cooling device callbacks */
183 static int acpi_processor_max_state(struct acpi_processor
*pr
)
188 * There exists four states according to
189 * cpufreq_thermal_reduction_pctg. 0, 1, 2, 3
191 max_state
+= cpufreq_get_max_state(pr
->id
);
192 if (pr
->flags
.throttling
)
193 max_state
+= (pr
->throttling
.state_count
-1);
198 processor_get_max_state(struct thermal_cooling_device
*cdev
,
199 unsigned long *state
)
201 struct acpi_device
*device
= cdev
->devdata
;
202 struct acpi_processor
*pr
;
207 pr
= acpi_driver_data(device
);
211 *state
= acpi_processor_max_state(pr
);
216 processor_get_cur_state(struct thermal_cooling_device
*cdev
,
217 unsigned long *cur_state
)
219 struct acpi_device
*device
= cdev
->devdata
;
220 struct acpi_processor
*pr
;
225 pr
= acpi_driver_data(device
);
229 *cur_state
= cpufreq_get_cur_state(pr
->id
);
230 if (pr
->flags
.throttling
)
231 *cur_state
+= pr
->throttling
.state
;
236 processor_set_cur_state(struct thermal_cooling_device
*cdev
,
239 struct acpi_device
*device
= cdev
->devdata
;
240 struct acpi_processor
*pr
;
247 pr
= acpi_driver_data(device
);
251 max_pstate
= cpufreq_get_max_state(pr
->id
);
253 if (state
> acpi_processor_max_state(pr
))
256 if (state
<= max_pstate
) {
257 if (pr
->flags
.throttling
&& pr
->throttling
.state
)
258 result
= acpi_processor_set_throttling(pr
, 0, false);
259 cpufreq_set_cur_state(pr
->id
, state
);
261 cpufreq_set_cur_state(pr
->id
, max_pstate
);
262 result
= acpi_processor_set_throttling(pr
,
263 state
- max_pstate
, false);
268 const struct thermal_cooling_device_ops processor_cooling_ops
= {
269 .get_max_state
= processor_get_max_state
,
270 .get_cur_state
= processor_get_cur_state
,
271 .set_cur_state
= processor_set_cur_state
,