blob: d93eee2f101b0aeede2049f515154687df9428db [file] [log] [blame]
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +00001/*
2 * x86_pkg_temp_thermal driver
3 * Copyright (c) 2013, Intel Corporation.
4 *
5 * This program is free software; you can redistribute it and/or modify it
6 * under the terms and conditions of the GNU General Public License,
7 * version 2, as published by the Free Software Foundation.
8 *
9 * This program is distributed in the hope it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
12 * more details.
13 *
14 * You should have received a copy of the GNU General Public License along with
15 * this program; if not, write to the Free Software Foundation, Inc.
16 *
17 */
18#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
19
20#include <linux/module.h>
21#include <linux/init.h>
22#include <linux/err.h>
23#include <linux/param.h>
24#include <linux/device.h>
25#include <linux/platform_device.h>
26#include <linux/cpu.h>
27#include <linux/smp.h>
28#include <linux/slab.h>
29#include <linux/pm.h>
30#include <linux/thermal.h>
31#include <linux/debugfs.h>
32#include <asm/cpu_device_id.h>
33#include <asm/mce.h>
34
35/*
36* Rate control delay: Idea is to introduce denounce effect
37* This should be long enough to avoid reduce events, when
38* threshold is set to a temperature, which is constantly
39* violated, but at the short enough to take any action.
40* The action can be remove threshold or change it to next
41* interesting setting. Based on experiments, in around
42* every 5 seconds under load will give us a significant
43* temperature change.
44*/
45#define PKG_TEMP_THERMAL_NOTIFY_DELAY 5000
46static int notify_delay_ms = PKG_TEMP_THERMAL_NOTIFY_DELAY;
47module_param(notify_delay_ms, int, 0644);
48MODULE_PARM_DESC(notify_delay_ms,
49 "User space notification delay in milli seconds.");
50
51/* Number of trip points in thermal zone. Currently it can't
52* be more than 2. MSR can allow setting and getting notifications
53* for only 2 thresholds. This define enforces this, if there
54* is some wrong values returned by cpuid for number of thresholds.
55*/
56#define MAX_NUMBER_OF_TRIPS 2
57
Thomas Gleixner3883a642016-11-22 17:57:08 +000058struct pkg_device {
Thomas Gleixner556238e2016-11-22 17:57:14 +000059 int cpu;
Thomas Gleixner64ca7382016-11-22 17:57:12 +000060 bool work_scheduled;
Thomas Gleixner3883a642016-11-22 17:57:08 +000061 u32 tj_max;
62 u32 msr_pkg_therm_low;
63 u32 msr_pkg_therm_high;
Thomas Gleixner411bb382016-11-22 17:57:13 +000064 struct delayed_work work;
Thomas Gleixner3883a642016-11-22 17:57:08 +000065 struct thermal_zone_device *tzone;
Thomas Gleixnerab47bd92016-11-22 17:57:10 +000066 struct cpumask cpumask;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +000067};
68
Javi Merino4abe6022015-03-03 15:30:50 +000069static struct thermal_zone_params pkg_temp_tz_params = {
Jean Delvare79786882014-03-02 15:33:35 +010070 .no_hwmon = true,
71};
72
Thomas Gleixner556238e2016-11-22 17:57:14 +000073/* Keep track of how many package pointers we allocated in init() */
74static int max_packages __read_mostly;
75/* Array of package pointers */
76static struct pkg_device **packages;
Thomas Gleixnerab47bd92016-11-22 17:57:10 +000077/* Serializes interrupt notification, work and hotplug */
78static DEFINE_SPINLOCK(pkg_temp_lock);
79/* Protects zone operation in the work function against hotplug removal */
80static DEFINE_MUTEX(thermal_zone_mutex);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +000081
Sebastian Andrzej Siewior7646ff22016-11-22 17:57:15 +000082/* The dynamically assigned cpu hotplug state for module_exit() */
83static enum cpuhp_state pkg_thermal_hp_state __read_mostly;
84
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +000085/* Debug counters to show using debugfs */
86static struct dentry *debugfs;
87static unsigned int pkg_interrupt_cnt;
88static unsigned int pkg_work_cnt;
89
90static int pkg_temp_debugfs_init(void)
91{
92 struct dentry *d;
93
94 debugfs = debugfs_create_dir("pkg_temp_thermal", NULL);
95 if (!debugfs)
96 return -ENOENT;
97
98 d = debugfs_create_u32("pkg_thres_interrupt", S_IRUGO, debugfs,
99 (u32 *)&pkg_interrupt_cnt);
100 if (!d)
101 goto err_out;
102
103 d = debugfs_create_u32("pkg_thres_work", S_IRUGO, debugfs,
104 (u32 *)&pkg_work_cnt);
105 if (!d)
106 goto err_out;
107
108 return 0;
109
110err_out:
111 debugfs_remove_recursive(debugfs);
112 return -ENOENT;
113}
114
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000115/*
116 * Protection:
117 *
118 * - cpu hotplug: Read serialized by cpu hotplug lock
119 * Write must hold pkg_temp_lock
120 *
121 * - Other callsites: Must hold pkg_temp_lock
122 */
Thomas Gleixner3883a642016-11-22 17:57:08 +0000123static struct pkg_device *pkg_temp_thermal_get_dev(unsigned int cpu)
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000124{
Thomas Gleixner556238e2016-11-22 17:57:14 +0000125 int pkgid = topology_logical_package_id(cpu);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000126
Thomas Gleixner556238e2016-11-22 17:57:14 +0000127 if (pkgid >= 0 && pkgid < max_packages)
128 return packages[pkgid];
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000129 return NULL;
130}
131
132/*
133* tj-max is is interesting because threshold is set relative to this
134* temperature.
135*/
136static int get_tj_max(int cpu, u32 *tj_max)
137{
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000138 u32 eax, edx, val;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000139 int err;
140
141 err = rdmsr_safe_on_cpu(cpu, MSR_IA32_TEMPERATURE_TARGET, &eax, &edx);
142 if (err)
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000143 return err;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000144
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000145 val = (eax >> 16) & 0xff;
146 *tj_max = val * 1000;
147
148 return val ? 0 : -EINVAL;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000149}
150
Sascha Hauer17e83512015-07-24 08:12:54 +0200151static int sys_get_curr_temp(struct thermal_zone_device *tzd, int *temp)
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000152{
Thomas Gleixner3883a642016-11-22 17:57:08 +0000153 struct pkg_device *pkgdev = tzd->devdata;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000154 u32 eax, edx;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000155
Thomas Gleixner3883a642016-11-22 17:57:08 +0000156 rdmsr_on_cpu(pkgdev->cpu, MSR_IA32_PACKAGE_THERM_STATUS, &eax, &edx);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000157 if (eax & 0x80000000) {
Thomas Gleixner3883a642016-11-22 17:57:08 +0000158 *temp = pkgdev->tj_max - ((eax >> 16) & 0x7f) * 1000;
Sascha Hauer17e83512015-07-24 08:12:54 +0200159 pr_debug("sys_get_curr_temp %d\n", *temp);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000160 return 0;
161 }
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000162 return -EINVAL;
163}
164
165static int sys_get_trip_temp(struct thermal_zone_device *tzd,
Thomas Gleixner3883a642016-11-22 17:57:08 +0000166 int trip, int *temp)
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000167{
Thomas Gleixner3883a642016-11-22 17:57:08 +0000168 struct pkg_device *pkgdev = tzd->devdata;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000169 unsigned long thres_reg_value;
Thomas Gleixner3883a642016-11-22 17:57:08 +0000170 u32 mask, shift, eax, edx;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000171 int ret;
172
173 if (trip >= MAX_NUMBER_OF_TRIPS)
174 return -EINVAL;
175
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000176 if (trip) {
177 mask = THERM_MASK_THRESHOLD1;
178 shift = THERM_SHIFT_THRESHOLD1;
179 } else {
180 mask = THERM_MASK_THRESHOLD0;
181 shift = THERM_SHIFT_THRESHOLD0;
182 }
183
Thomas Gleixner3883a642016-11-22 17:57:08 +0000184 ret = rdmsr_on_cpu(pkgdev->cpu, MSR_IA32_PACKAGE_THERM_INTERRUPT,
Thomas Gleixnerb6badbe2016-11-22 17:57:07 +0000185 &eax, &edx);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000186 if (ret < 0)
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000187 return ret;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000188
189 thres_reg_value = (eax & mask) >> shift;
190 if (thres_reg_value)
Thomas Gleixner3883a642016-11-22 17:57:08 +0000191 *temp = pkgdev->tj_max - thres_reg_value * 1000;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000192 else
193 *temp = 0;
Sascha Hauer17e83512015-07-24 08:12:54 +0200194 pr_debug("sys_get_trip_temp %d\n", *temp);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000195
196 return 0;
197}
198
Thomas Gleixner3883a642016-11-22 17:57:08 +0000199static int
200sys_set_trip_temp(struct thermal_zone_device *tzd, int trip, int temp)
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000201{
Thomas Gleixner3883a642016-11-22 17:57:08 +0000202 struct pkg_device *pkgdev = tzd->devdata;
203 u32 l, h, mask, shift, intr;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000204 int ret;
205
Thomas Gleixner3883a642016-11-22 17:57:08 +0000206 if (trip >= MAX_NUMBER_OF_TRIPS || temp >= pkgdev->tj_max)
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000207 return -EINVAL;
208
Thomas Gleixner3883a642016-11-22 17:57:08 +0000209 ret = rdmsr_on_cpu(pkgdev->cpu, MSR_IA32_PACKAGE_THERM_INTERRUPT,
Thomas Gleixnerb6badbe2016-11-22 17:57:07 +0000210 &l, &h);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000211 if (ret < 0)
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000212 return ret;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000213
214 if (trip) {
215 mask = THERM_MASK_THRESHOLD1;
216 shift = THERM_SHIFT_THRESHOLD1;
217 intr = THERM_INT_THRESHOLD1_ENABLE;
218 } else {
219 mask = THERM_MASK_THRESHOLD0;
220 shift = THERM_SHIFT_THRESHOLD0;
221 intr = THERM_INT_THRESHOLD0_ENABLE;
222 }
223 l &= ~mask;
224 /*
225 * When users space sets a trip temperature == 0, which is indication
226 * that, it is no longer interested in receiving notifications.
227 */
Thomas Gleixner3883a642016-11-22 17:57:08 +0000228 if (!temp) {
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000229 l &= ~intr;
Thomas Gleixner3883a642016-11-22 17:57:08 +0000230 } else {
231 l |= (pkgdev->tj_max - temp)/1000 << shift;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000232 l |= intr;
233 }
234
Thomas Gleixner3883a642016-11-22 17:57:08 +0000235 return wrmsr_on_cpu(pkgdev->cpu, MSR_IA32_PACKAGE_THERM_INTERRUPT, l, h);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000236}
237
Thomas Gleixner3883a642016-11-22 17:57:08 +0000238static int sys_get_trip_type(struct thermal_zone_device *thermal, int trip,
239 enum thermal_trip_type *type)
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000240{
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000241 *type = THERMAL_TRIP_PASSIVE;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000242 return 0;
243}
244
245/* Thermal zone callback registry */
246static struct thermal_zone_device_ops tzone_ops = {
247 .get_temp = sys_get_curr_temp,
248 .get_trip_temp = sys_get_trip_temp,
249 .get_trip_type = sys_get_trip_type,
250 .set_trip_temp = sys_set_trip_temp,
251};
252
Thomas Gleixner09a674c2016-11-22 17:57:06 +0000253static bool pkg_thermal_rate_control(void)
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000254{
255 return true;
256}
257
258/* Enable threshold interrupt on local package/cpu */
259static inline void enable_pkg_thres_interrupt(void)
260{
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000261 u8 thres_0, thres_1;
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000262 u32 l, h;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000263
264 rdmsr(MSR_IA32_PACKAGE_THERM_INTERRUPT, l, h);
265 /* only enable/disable if it had valid threshold value */
266 thres_0 = (l & THERM_MASK_THRESHOLD0) >> THERM_SHIFT_THRESHOLD0;
267 thres_1 = (l & THERM_MASK_THRESHOLD1) >> THERM_SHIFT_THRESHOLD1;
268 if (thres_0)
269 l |= THERM_INT_THRESHOLD0_ENABLE;
270 if (thres_1)
271 l |= THERM_INT_THRESHOLD1_ENABLE;
272 wrmsr(MSR_IA32_PACKAGE_THERM_INTERRUPT, l, h);
273}
274
275/* Disable threshold interrupt on local package/cpu */
276static inline void disable_pkg_thres_interrupt(void)
277{
278 u32 l, h;
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000279
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000280 rdmsr(MSR_IA32_PACKAGE_THERM_INTERRUPT, l, h);
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000281
282 l &= ~(THERM_INT_THRESHOLD0_ENABLE | THERM_INT_THRESHOLD1_ENABLE);
283 wrmsr(MSR_IA32_PACKAGE_THERM_INTERRUPT, l, h);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000284}
285
286static void pkg_temp_thermal_threshold_work_fn(struct work_struct *work)
287{
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000288 struct thermal_zone_device *tzone = NULL;
Thomas Gleixner64ca7382016-11-22 17:57:12 +0000289 int cpu = smp_processor_id();
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000290 struct pkg_device *pkgdev;
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000291 u64 msr_val, wr_val;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000292
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000293 mutex_lock(&thermal_zone_mutex);
294 spin_lock_irq(&pkg_temp_lock);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000295 ++pkg_work_cnt;
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000296
297 pkgdev = pkg_temp_thermal_get_dev(cpu);
298 if (!pkgdev) {
299 spin_unlock_irq(&pkg_temp_lock);
300 mutex_unlock(&thermal_zone_mutex);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000301 return;
302 }
Thomas Gleixner64ca7382016-11-22 17:57:12 +0000303 pkgdev->work_scheduled = false;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000304
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000305 rdmsrl(MSR_IA32_PACKAGE_THERM_STATUS, msr_val);
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000306 wr_val = msr_val & ~(THERM_LOG_THRESHOLD0 | THERM_LOG_THRESHOLD1);
307 if (wr_val != msr_val) {
308 wrmsrl(MSR_IA32_PACKAGE_THERM_STATUS, wr_val);
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000309 tzone = pkgdev->tzone;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000310 }
Thomas Gleixner768bd132016-11-22 17:57:04 +0000311
312 enable_pkg_thres_interrupt();
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000313 spin_unlock_irq(&pkg_temp_lock);
Thomas Gleixner768bd132016-11-22 17:57:04 +0000314
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000315 /*
316 * If tzone is not NULL, then thermal_zone_mutex will prevent the
317 * concurrent removal in the cpu offline callback.
318 */
319 if (tzone)
320 thermal_zone_device_update(tzone, THERMAL_EVENT_UNSPECIFIED);
321
322 mutex_unlock(&thermal_zone_mutex);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000323}
324
Thomas Gleixner411bb382016-11-22 17:57:13 +0000325static void pkg_thermal_schedule_work(int cpu, struct delayed_work *work)
326{
327 unsigned long ms = msecs_to_jiffies(notify_delay_ms);
328
329 schedule_delayed_work_on(cpu, work, ms);
330}
331
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000332static int pkg_thermal_notify(u64 msr_val)
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000333{
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000334 int cpu = smp_processor_id();
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000335 struct pkg_device *pkgdev;
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000336 unsigned long flags;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000337
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000338 spin_lock_irqsave(&pkg_temp_lock, flags);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000339 ++pkg_interrupt_cnt;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000340
341 disable_pkg_thres_interrupt();
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000342
343 /* Work is per package, so scheduling it once is enough. */
344 pkgdev = pkg_temp_thermal_get_dev(cpu);
Thomas Gleixner64ca7382016-11-22 17:57:12 +0000345 if (pkgdev && !pkgdev->work_scheduled) {
346 pkgdev->work_scheduled = true;
Thomas Gleixner411bb382016-11-22 17:57:13 +0000347 pkg_thermal_schedule_work(pkgdev->cpu, &pkgdev->work);
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000348 }
349
350 spin_unlock_irqrestore(&pkg_temp_lock, flags);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000351 return 0;
352}
353
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000354static int pkg_temp_thermal_device_add(unsigned int cpu)
355{
Thomas Gleixner556238e2016-11-22 17:57:14 +0000356 int pkgid = topology_logical_package_id(cpu);
Thomas Gleixner3883a642016-11-22 17:57:08 +0000357 u32 tj_max, eax, ebx, ecx, edx;
358 struct pkg_device *pkgdev;
359 int thres_count, err;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000360
Thomas Gleixner556238e2016-11-22 17:57:14 +0000361 if (pkgid >= max_packages)
362 return -ENOMEM;
363
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000364 cpuid(6, &eax, &ebx, &ecx, &edx);
365 thres_count = ebx & 0x07;
366 if (!thres_count)
367 return -ENODEV;
368
369 thres_count = clamp_val(thres_count, 0, MAX_NUMBER_OF_TRIPS);
370
371 err = get_tj_max(cpu, &tj_max);
372 if (err)
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000373 return err;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000374
Thomas Gleixner3883a642016-11-22 17:57:08 +0000375 pkgdev = kzalloc(sizeof(*pkgdev), GFP_KERNEL);
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000376 if (!pkgdev)
377 return -ENOMEM;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000378
Thomas Gleixner411bb382016-11-22 17:57:13 +0000379 INIT_DELAYED_WORK(&pkgdev->work, pkg_temp_thermal_threshold_work_fn);
Thomas Gleixner3883a642016-11-22 17:57:08 +0000380 pkgdev->cpu = cpu;
381 pkgdev->tj_max = tj_max;
382 pkgdev->tzone = thermal_zone_device_register("x86_pkg_temp",
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000383 thres_count,
Thomas Gleixner3883a642016-11-22 17:57:08 +0000384 (thres_count == MAX_NUMBER_OF_TRIPS) ? 0x03 : 0x01,
385 pkgdev, &tzone_ops, &pkg_temp_tz_params, 0, 0);
386 if (IS_ERR(pkgdev->tzone)) {
387 err = PTR_ERR(pkgdev->tzone);
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000388 kfree(pkgdev);
389 return err;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000390 }
391 /* Store MSR value for package thermal interrupt, to restore at exit */
Sebastian Andrzej Siewior7646ff22016-11-22 17:57:15 +0000392 rdmsr(MSR_IA32_PACKAGE_THERM_INTERRUPT, pkgdev->msr_pkg_therm_low,
393 pkgdev->msr_pkg_therm_high);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000394
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000395 cpumask_set_cpu(cpu, &pkgdev->cpumask);
396 spin_lock_irq(&pkg_temp_lock);
Thomas Gleixner556238e2016-11-22 17:57:14 +0000397 packages[pkgid] = pkgdev;
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000398 spin_unlock_irq(&pkg_temp_lock);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000399 return 0;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000400}
401
Sebastian Andrzej Siewior7646ff22016-11-22 17:57:15 +0000402static int pkg_thermal_cpu_offline(unsigned int cpu)
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000403{
Thomas Gleixner3883a642016-11-22 17:57:08 +0000404 struct pkg_device *pkgdev = pkg_temp_thermal_get_dev(cpu);
Thomas Gleixner411bb382016-11-22 17:57:13 +0000405 bool lastcpu, was_target;
Thomas Gleixnerb6badbe2016-11-22 17:57:07 +0000406 int target;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000407
Thomas Gleixner3883a642016-11-22 17:57:08 +0000408 if (!pkgdev)
Sebastian Andrzej Siewior7646ff22016-11-22 17:57:15 +0000409 return 0;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000410
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000411 target = cpumask_any_but(&pkgdev->cpumask, cpu);
412 cpumask_clear_cpu(cpu, &pkgdev->cpumask);
413 lastcpu = target >= nr_cpu_ids;
414 /*
415 * Remove the sysfs files, if this is the last cpu in the package
416 * before doing further cleanups.
417 */
418 if (lastcpu) {
419 struct thermal_zone_device *tzone = pkgdev->tzone;
Thomas Gleixner21a3d3d2016-11-22 17:57:06 +0000420
Thomas Gleixner89baa562016-11-22 17:57:05 +0000421 /*
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000422 * We must protect against a work function calling
423 * thermal_zone_update, after/while unregister. We null out
424 * the pointer under the zone mutex, so the worker function
425 * won't try to call.
Thomas Gleixner89baa562016-11-22 17:57:05 +0000426 */
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000427 mutex_lock(&thermal_zone_mutex);
428 pkgdev->tzone = NULL;
429 mutex_unlock(&thermal_zone_mutex);
430
431 thermal_zone_device_unregister(tzone);
432 }
433
Thomas Gleixner411bb382016-11-22 17:57:13 +0000434 /* Protect against work and interrupts */
435 spin_lock_irq(&pkg_temp_lock);
436
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000437 /*
Thomas Gleixner411bb382016-11-22 17:57:13 +0000438 * Check whether this cpu was the current target and store the new
439 * one. When we drop the lock, then the interrupt notify function
440 * will see the new target.
441 */
442 was_target = pkgdev->cpu == cpu;
443 pkgdev->cpu = target;
444
445 /*
446 * If this is the last CPU in the package remove the package
Thomas Gleixner556238e2016-11-22 17:57:14 +0000447 * reference from the array and restore the interrupt MSR. When we
Thomas Gleixner411bb382016-11-22 17:57:13 +0000448 * drop the lock neither the interrupt notify function nor the
449 * worker will see the package anymore.
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000450 */
451 if (lastcpu) {
Thomas Gleixner556238e2016-11-22 17:57:14 +0000452 packages[topology_logical_package_id(cpu)] = NULL;
Sebastian Andrzej Siewior7646ff22016-11-22 17:57:15 +0000453 /* After this point nothing touches the MSR anymore. */
454 wrmsr(MSR_IA32_PACKAGE_THERM_INTERRUPT,
455 pkgdev->msr_pkg_therm_low, pkgdev->msr_pkg_therm_high);
Thomas Gleixner89baa562016-11-22 17:57:05 +0000456 }
Thomas Gleixnerb6badbe2016-11-22 17:57:07 +0000457
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000458 /*
Thomas Gleixner411bb382016-11-22 17:57:13 +0000459 * Check whether there is work scheduled and whether the work is
460 * targeted at the outgoing CPU.
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000461 */
Thomas Gleixner411bb382016-11-22 17:57:13 +0000462 if (pkgdev->work_scheduled && was_target) {
463 /*
464 * To cancel the work we need to drop the lock, otherwise
465 * we might deadlock if the work needs to be flushed.
466 */
467 spin_unlock_irq(&pkg_temp_lock);
468 cancel_delayed_work_sync(&pkgdev->work);
469 spin_lock_irq(&pkg_temp_lock);
470 /*
471 * If this is not the last cpu in the package and the work
472 * did not run after we dropped the lock above, then we
473 * need to reschedule the work, otherwise the interrupt
474 * stays disabled forever.
475 */
476 if (!lastcpu && pkgdev->work_scheduled)
477 pkg_thermal_schedule_work(target, &pkgdev->work);
478 }
479
480 spin_unlock_irq(&pkg_temp_lock);
481
482 /* Final cleanup if this is the last cpu */
483 if (lastcpu)
484 kfree(pkgdev);
Sebastian Andrzej Siewior7646ff22016-11-22 17:57:15 +0000485 return 0;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000486}
487
Sebastian Andrzej Siewior7646ff22016-11-22 17:57:15 +0000488static int pkg_thermal_cpu_online(unsigned int cpu)
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000489{
Thomas Gleixner3883a642016-11-22 17:57:08 +0000490 struct pkg_device *pkgdev = pkg_temp_thermal_get_dev(cpu);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000491 struct cpuinfo_x86 *c = &cpu_data(cpu);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000492
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000493 /* Paranoia check */
494 if (!cpu_has(c, X86_FEATURE_DTHERM) || !cpu_has(c, X86_FEATURE_PTS))
495 return -ENODEV;
Thomas Gleixner3883a642016-11-22 17:57:08 +0000496
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000497 /* If the package exists, nothing to do */
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000498 if (pkgdev) {
499 cpumask_set_cpu(cpu, &pkgdev->cpumask);
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000500 return 0;
Thomas Gleixnerab47bd92016-11-22 17:57:10 +0000501 }
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000502 return pkg_temp_thermal_device_add(cpu);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000503}
504
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000505static const struct x86_cpu_id __initconst pkg_temp_thermal_ids[] = {
Srinivas Pandruvadaf3ed0a12013-07-11 09:50:30 -0700506 { X86_VENDOR_INTEL, X86_FAMILY_ANY, X86_MODEL_ANY, X86_FEATURE_PTS },
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000507 {}
508};
509MODULE_DEVICE_TABLE(x86cpu, pkg_temp_thermal_ids);
510
511static int __init pkg_temp_thermal_init(void)
512{
Sebastian Andrzej Siewior7646ff22016-11-22 17:57:15 +0000513 int ret;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000514
515 if (!x86_match_cpu(pkg_temp_thermal_ids))
516 return -ENODEV;
517
Thomas Gleixner556238e2016-11-22 17:57:14 +0000518 max_packages = topology_max_packages();
519 packages = kzalloc(max_packages * sizeof(struct pkg_device *), GFP_KERNEL);
520 if (!packages)
521 return -ENOMEM;
522
Sebastian Andrzej Siewior7646ff22016-11-22 17:57:15 +0000523 ret = cpuhp_setup_state(CPUHP_AP_ONLINE_DYN, "thermal/x86_pkg:online",
524 pkg_thermal_cpu_online, pkg_thermal_cpu_offline);
525 if (ret < 0)
526 goto err;
527
528 /* Store the state for module exit */
529 pkg_thermal_hp_state = ret;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000530
Thomas Gleixner09a674c2016-11-22 17:57:06 +0000531 platform_thermal_package_notify = pkg_thermal_notify;
532 platform_thermal_package_rate_control = pkg_thermal_rate_control;
533
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000534 /* Don't care if it fails */
535 pkg_temp_debugfs_init();
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000536 return 0;
537
Sebastian Andrzej Siewior7646ff22016-11-22 17:57:15 +0000538err:
Thomas Gleixner556238e2016-11-22 17:57:14 +0000539 kfree(packages);
Sebastian Andrzej Siewior7646ff22016-11-22 17:57:15 +0000540 return ret;
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000541}
Thomas Gleixner8079a4b2016-11-22 17:57:09 +0000542module_init(pkg_temp_thermal_init)
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000543
544static void __exit pkg_temp_thermal_exit(void)
545{
Thomas Gleixner09a674c2016-11-22 17:57:06 +0000546 platform_thermal_package_notify = NULL;
547 platform_thermal_package_rate_control = NULL;
548
Sebastian Andrzej Siewior7646ff22016-11-22 17:57:15 +0000549 cpuhp_remove_state(pkg_thermal_hp_state);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000550 debugfs_remove_recursive(debugfs);
Thomas Gleixner556238e2016-11-22 17:57:14 +0000551 kfree(packages);
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000552}
Srinivas Pandruvadaf1a18a12013-05-17 23:42:02 +0000553module_exit(pkg_temp_thermal_exit)
554
555MODULE_DESCRIPTION("X86 PKG TEMP Thermal Driver");
556MODULE_AUTHOR("Srinivas Pandruvada <srinivas.pandruvada@linux.intel.com>");
557MODULE_LICENSE("GPL v2");