blob: c831d4b62d3ad814ecd360e939462a9618fbc301 [file] [log] [blame]
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001/* Copyright (c) 2010-2011, Code Aurora Forum. All rights reserved.
2 *
3 * This program is free software; you can redistribute it and/or modify
4 * it under the terms of the GNU General Public License version 2 and
5 * only version 2 as published by the Free Software Foundation.
6 *
7 * This program is distributed in the hope that it will be useful,
8 * but WITHOUT ANY WARRANTY; without even the implied warranty of
9 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
10 * GNU General Public License for more details.
11 *
12 */
13
14#include <linux/module.h>
15#include <linux/kernel.h>
16#include <linux/init.h>
17#include <linux/completion.h>
18#include <linux/cpuidle.h>
19#include <linux/interrupt.h>
20#include <linux/io.h>
21#include <linux/ktime.h>
22#include <linux/pm.h>
23#include <linux/pm_qos_params.h>
24#include <linux/proc_fs.h>
25#include <linux/smp.h>
26#include <linux/suspend.h>
27#include <linux/tick.h>
28#include <linux/uaccess.h>
29#include <linux/wakelock.h>
30#include <mach/msm_iomap.h>
31#include <mach/system.h>
32#include <asm/cacheflush.h>
33#include <asm/hardware/gic.h>
34#include <asm/pgtable.h>
35#include <asm/pgalloc.h>
36#ifdef CONFIG_VFP
37#include <asm/vfp.h>
38#endif
39
40#include "acpuclock.h"
41#include "clock.h"
42#include "avs.h"
43#include "cpuidle.h"
44#include "idle.h"
45#include "pm.h"
46#include "rpm_resources.h"
47#include "scm-boot.h"
48#include "spm.h"
49#include "timer.h"
Pratik Patele5771792011-09-17 18:33:54 -070050#include "qdss.h"
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070051
52/******************************************************************************
53 * Debug Definitions
54 *****************************************************************************/
55
56enum {
57 MSM_PM_DEBUG_SUSPEND = BIT(0),
58 MSM_PM_DEBUG_POWER_COLLAPSE = BIT(1),
59 MSM_PM_DEBUG_SUSPEND_LIMITS = BIT(2),
60 MSM_PM_DEBUG_CLOCK = BIT(3),
61 MSM_PM_DEBUG_RESET_VECTOR = BIT(4),
62 MSM_PM_DEBUG_IDLE = BIT(6),
63 MSM_PM_DEBUG_IDLE_LIMITS = BIT(7),
64 MSM_PM_DEBUG_HOTPLUG = BIT(8),
65};
66
67static int msm_pm_debug_mask = 1;
68module_param_named(
69 debug_mask, msm_pm_debug_mask, int, S_IRUGO | S_IWUSR | S_IWGRP
70);
71
72
73/******************************************************************************
74 * Sleep Modes and Parameters
75 *****************************************************************************/
76
77static struct msm_pm_platform_data *msm_pm_modes;
78static int rpm_cpu0_wakeup_irq;
79
80void __init msm_pm_set_platform_data(
81 struct msm_pm_platform_data *data, int count)
82{
83 BUG_ON(MSM_PM_SLEEP_MODE_NR * num_possible_cpus() > count);
84 msm_pm_modes = data;
85}
86
87void __init msm_pm_set_rpm_wakeup_irq(unsigned int irq)
88{
89 rpm_cpu0_wakeup_irq = irq;
90}
91
92enum {
93 MSM_PM_MODE_ATTR_SUSPEND,
94 MSM_PM_MODE_ATTR_IDLE,
95 MSM_PM_MODE_ATTR_NR,
96};
97
98static char *msm_pm_mode_attr_labels[MSM_PM_MODE_ATTR_NR] = {
99 [MSM_PM_MODE_ATTR_SUSPEND] = "suspend_enabled",
100 [MSM_PM_MODE_ATTR_IDLE] = "idle_enabled",
101};
102
103struct msm_pm_kobj_attribute {
104 unsigned int cpu;
105 struct kobj_attribute ka;
106};
107
108#define GET_CPU_OF_ATTR(attr) \
109 (container_of(attr, struct msm_pm_kobj_attribute, ka)->cpu)
110
111struct msm_pm_sysfs_sleep_mode {
112 struct kobject *kobj;
113 struct attribute_group attr_group;
114 struct attribute *attrs[MSM_PM_MODE_ATTR_NR + 1];
115 struct msm_pm_kobj_attribute kas[MSM_PM_MODE_ATTR_NR];
116};
117
118static char *msm_pm_sleep_mode_labels[MSM_PM_SLEEP_MODE_NR] = {
119 [MSM_PM_SLEEP_MODE_POWER_COLLAPSE] = "power_collapse",
120 [MSM_PM_SLEEP_MODE_WAIT_FOR_INTERRUPT] = "wfi",
121 [MSM_PM_SLEEP_MODE_POWER_COLLAPSE_STANDALONE] =
122 "standalone_power_collapse",
123};
124
125/*
126 * Write out the attribute.
127 */
128static ssize_t msm_pm_mode_attr_show(
129 struct kobject *kobj, struct kobj_attribute *attr, char *buf)
130{
131 int ret = -EINVAL;
132 int i;
133
134 for (i = 0; i < MSM_PM_SLEEP_MODE_NR; i++) {
135 struct kernel_param kp;
136 unsigned int cpu;
137 struct msm_pm_platform_data *mode;
138
139 if (msm_pm_sleep_mode_labels[i] == NULL)
140 continue;
141
142 if (strcmp(kobj->name, msm_pm_sleep_mode_labels[i]))
143 continue;
144
145 cpu = GET_CPU_OF_ATTR(attr);
146 mode = &msm_pm_modes[MSM_PM_MODE(cpu, i)];
147
148 if (!strcmp(attr->attr.name,
149 msm_pm_mode_attr_labels[MSM_PM_MODE_ATTR_SUSPEND])) {
150 u32 arg = mode->suspend_enabled;
151 kp.arg = &arg;
152 ret = param_get_ulong(buf, &kp);
153 } else if (!strcmp(attr->attr.name,
154 msm_pm_mode_attr_labels[MSM_PM_MODE_ATTR_IDLE])) {
155 u32 arg = mode->idle_enabled;
156 kp.arg = &arg;
157 ret = param_get_ulong(buf, &kp);
158 }
159
160 break;
161 }
162
163 if (ret > 0) {
164 strcat(buf, "\n");
165 ret++;
166 }
167
168 return ret;
169}
170
171/*
172 * Read in the new attribute value.
173 */
174static ssize_t msm_pm_mode_attr_store(struct kobject *kobj,
175 struct kobj_attribute *attr, const char *buf, size_t count)
176{
177 int ret = -EINVAL;
178 int i;
179
180 for (i = 0; i < MSM_PM_SLEEP_MODE_NR; i++) {
181 struct kernel_param kp;
182 unsigned int cpu;
183 struct msm_pm_platform_data *mode;
184
185 if (msm_pm_sleep_mode_labels[i] == NULL)
186 continue;
187
188 if (strcmp(kobj->name, msm_pm_sleep_mode_labels[i]))
189 continue;
190
191 cpu = GET_CPU_OF_ATTR(attr);
192 mode = &msm_pm_modes[MSM_PM_MODE(cpu, i)];
193
194 if (!strcmp(attr->attr.name,
195 msm_pm_mode_attr_labels[MSM_PM_MODE_ATTR_SUSPEND])) {
196 kp.arg = &mode->suspend_enabled;
197 ret = param_set_byte(buf, &kp);
198 } else if (!strcmp(attr->attr.name,
199 msm_pm_mode_attr_labels[MSM_PM_MODE_ATTR_IDLE])) {
200 kp.arg = &mode->idle_enabled;
201 ret = param_set_byte(buf, &kp);
202 }
203
204 break;
205 }
206
207 return ret ? ret : count;
208}
209
210/*
211 * Add sysfs entries for one cpu.
212 */
213static int __init msm_pm_mode_sysfs_add_cpu(
214 unsigned int cpu, struct kobject *modes_kobj)
215{
216 char cpu_name[8];
217 struct kobject *cpu_kobj;
218 struct msm_pm_sysfs_sleep_mode *mode;
219 int i, j, k;
220 int ret;
221
222 snprintf(cpu_name, sizeof(cpu_name), "cpu%u", cpu);
223 cpu_kobj = kobject_create_and_add(cpu_name, modes_kobj);
224 if (!cpu_kobj) {
225 pr_err("%s: cannot create %s kobject\n", __func__, cpu_name);
226 ret = -ENOMEM;
227 goto mode_sysfs_add_cpu_exit;
228 }
229
230 for (i = 0; i < MSM_PM_SLEEP_MODE_NR; i++) {
231 int idx = MSM_PM_MODE(cpu, i);
232
233 if ((!msm_pm_modes[idx].suspend_supported)
234 && (!msm_pm_modes[idx].idle_supported))
235 continue;
236
237 mode = kzalloc(sizeof(*mode), GFP_KERNEL);
238 if (!mode) {
239 pr_err("%s: cannot allocate memory for attributes\n",
240 __func__);
241 ret = -ENOMEM;
242 goto mode_sysfs_add_cpu_exit;
243 }
244
245 mode->kobj = kobject_create_and_add(
246 msm_pm_sleep_mode_labels[i], cpu_kobj);
247 if (!mode->kobj) {
248 pr_err("%s: cannot create kobject\n", __func__);
249 ret = -ENOMEM;
250 goto mode_sysfs_add_cpu_exit;
251 }
252
253 for (k = 0, j = 0; k < MSM_PM_MODE_ATTR_NR; k++) {
254 if ((k == MSM_PM_MODE_ATTR_IDLE) &&
255 !msm_pm_modes[idx].idle_supported)
256 continue;
257 if ((k == MSM_PM_MODE_ATTR_SUSPEND) &&
258 !msm_pm_modes[idx].suspend_supported)
259 continue;
260 mode->kas[j].cpu = cpu;
261 mode->kas[j].ka.attr.mode = 0644;
262 mode->kas[j].ka.show = msm_pm_mode_attr_show;
263 mode->kas[j].ka.store = msm_pm_mode_attr_store;
264 mode->kas[j].ka.attr.name = msm_pm_mode_attr_labels[k];
265 mode->attrs[j] = &mode->kas[j].ka.attr;
266 j++;
267 }
268 mode->attrs[j] = NULL;
269
270 mode->attr_group.attrs = mode->attrs;
271 ret = sysfs_create_group(mode->kobj, &mode->attr_group);
272 if (ret) {
273 pr_err("%s: cannot create kobject attribute group\n",
274 __func__);
275 goto mode_sysfs_add_cpu_exit;
276 }
277 }
278
279 ret = 0;
280
281mode_sysfs_add_cpu_exit:
Praveen Chidambaram2cfda632011-10-11 16:58:09 -0600282 if (!ret) {
283 if (mode && mode->kobj)
284 kobject_del(mode->kobj);
285 kfree(mode);
286 }
287
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700288 return ret;
289}
290
291/*
292 * Add sysfs entries for the sleep modes.
293 */
294static int __init msm_pm_mode_sysfs_add(void)
295{
296 struct kobject *module_kobj;
297 struct kobject *modes_kobj;
298 unsigned int cpu;
299 int ret;
300
301 module_kobj = kset_find_obj(module_kset, KBUILD_MODNAME);
302 if (!module_kobj) {
303 pr_err("%s: cannot find kobject for module %s\n",
304 __func__, KBUILD_MODNAME);
305 ret = -ENOENT;
306 goto mode_sysfs_add_exit;
307 }
308
309 modes_kobj = kobject_create_and_add("modes", module_kobj);
310 if (!modes_kobj) {
311 pr_err("%s: cannot create modes kobject\n", __func__);
312 ret = -ENOMEM;
313 goto mode_sysfs_add_exit;
314 }
315
316 for_each_possible_cpu(cpu) {
317 ret = msm_pm_mode_sysfs_add_cpu(cpu, modes_kobj);
318 if (ret)
319 goto mode_sysfs_add_exit;
320 }
321
322 ret = 0;
323
324mode_sysfs_add_exit:
325 return ret;
326}
327
328/******************************************************************************
329 * CONFIG_MSM_IDLE_STATS
330 *****************************************************************************/
331
332#ifdef CONFIG_MSM_IDLE_STATS
333enum msm_pm_time_stats_id {
334 MSM_PM_STAT_REQUESTED_IDLE,
335 MSM_PM_STAT_IDLE_WFI,
336 MSM_PM_STAT_IDLE_STANDALONE_POWER_COLLAPSE,
337 MSM_PM_STAT_IDLE_POWER_COLLAPSE,
338 MSM_PM_STAT_SUSPEND,
339 MSM_PM_STAT_COUNT
340};
341
342struct msm_pm_time_stats {
343 const char *name;
344 int64_t first_bucket_time;
345 int bucket[CONFIG_MSM_IDLE_STATS_BUCKET_COUNT];
346 int64_t min_time[CONFIG_MSM_IDLE_STATS_BUCKET_COUNT];
347 int64_t max_time[CONFIG_MSM_IDLE_STATS_BUCKET_COUNT];
348 int count;
349 int64_t total_time;
350};
351
352struct msm_pm_cpu_time_stats {
353 struct msm_pm_time_stats stats[MSM_PM_STAT_COUNT];
354};
355
356static DEFINE_SPINLOCK(msm_pm_stats_lock);
357static DEFINE_PER_CPU_SHARED_ALIGNED(
358 struct msm_pm_cpu_time_stats, msm_pm_stats);
359
360/*
361 * Add the given time data to the statistics collection.
362 */
363static void msm_pm_add_stat(enum msm_pm_time_stats_id id, int64_t t)
364{
365 unsigned long flags;
366 struct msm_pm_time_stats *stats;
367 int64_t bt;
368 int i;
369
370 spin_lock_irqsave(&msm_pm_stats_lock, flags);
371 stats = __get_cpu_var(msm_pm_stats).stats;
372
373 stats[id].total_time += t;
374 stats[id].count++;
375
376 bt = t;
377 do_div(bt, stats[id].first_bucket_time);
378
379 if (bt < 1ULL << (CONFIG_MSM_IDLE_STATS_BUCKET_SHIFT *
380 (CONFIG_MSM_IDLE_STATS_BUCKET_COUNT - 1)))
381 i = DIV_ROUND_UP(fls((uint32_t)bt),
382 CONFIG_MSM_IDLE_STATS_BUCKET_SHIFT);
383 else
384 i = CONFIG_MSM_IDLE_STATS_BUCKET_COUNT - 1;
385
386 stats[id].bucket[i]++;
387
388 if (t < stats[id].min_time[i] || !stats[id].max_time[i])
389 stats[id].min_time[i] = t;
390 if (t > stats[id].max_time[i])
391 stats[id].max_time[i] = t;
392
393 spin_unlock_irqrestore(&msm_pm_stats_lock, flags);
394}
395
396/*
397 * Helper function of snprintf where buf is auto-incremented, size is auto-
398 * decremented, and there is no return value.
399 *
400 * NOTE: buf and size must be l-values (e.g. variables)
401 */
402#define SNPRINTF(buf, size, format, ...) \
403 do { \
404 if (size > 0) { \
405 int ret; \
406 ret = snprintf(buf, size, format, ## __VA_ARGS__); \
407 if (ret > size) { \
408 buf += size; \
409 size = 0; \
410 } else { \
411 buf += ret; \
412 size -= ret; \
413 } \
414 } \
415 } while (0)
416
417/*
418 * Write out the power management statistics.
419 */
420static int msm_pm_read_proc
421 (char *page, char **start, off_t off, int count, int *eof, void *data)
422{
423 unsigned int cpu = off / MSM_PM_STAT_COUNT;
424 int id = off % MSM_PM_STAT_COUNT;
425 char *p = page;
426
427 if (count < 1024) {
428 *start = (char *) 0;
429 *eof = 0;
430 return 0;
431 }
432
433 if (cpu < num_possible_cpus()) {
434 unsigned long flags;
435 struct msm_pm_time_stats *stats;
436 int i;
437 int64_t bucket_time;
438 int64_t s;
439 uint32_t ns;
440
441 spin_lock_irqsave(&msm_pm_stats_lock, flags);
442 stats = per_cpu(msm_pm_stats, cpu).stats;
443
444 s = stats[id].total_time;
445 ns = do_div(s, NSEC_PER_SEC);
446 SNPRINTF(p, count,
447 "[cpu %u] %s:\n"
448 " count: %7d\n"
449 " total_time: %lld.%09u\n",
450 cpu, stats[id].name,
451 stats[id].count,
452 s, ns);
453
454 bucket_time = stats[id].first_bucket_time;
455 for (i = 0; i < CONFIG_MSM_IDLE_STATS_BUCKET_COUNT - 1; i++) {
456 s = bucket_time;
457 ns = do_div(s, NSEC_PER_SEC);
458 SNPRINTF(p, count,
459 " <%6lld.%09u: %7d (%lld-%lld)\n",
460 s, ns, stats[id].bucket[i],
461 stats[id].min_time[i],
462 stats[id].max_time[i]);
463
464 bucket_time <<= CONFIG_MSM_IDLE_STATS_BUCKET_SHIFT;
465 }
466
467 SNPRINTF(p, count, " >=%6lld.%09u: %7d (%lld-%lld)\n",
468 s, ns, stats[id].bucket[i],
469 stats[id].min_time[i],
470 stats[id].max_time[i]);
471
472 *start = (char *) 1;
473 *eof = (off + 1 >= MSM_PM_STAT_COUNT * num_possible_cpus());
474
475 spin_unlock_irqrestore(&msm_pm_stats_lock, flags);
476 }
477
478 return p - page;
479}
480#undef SNPRINTF
481
482#define MSM_PM_STATS_RESET "reset"
483
484/*
485 * Reset the power management statistics values.
486 */
487static int msm_pm_write_proc(struct file *file, const char __user *buffer,
488 unsigned long count, void *data)
489{
490 char buf[sizeof(MSM_PM_STATS_RESET)];
491 int ret;
492 unsigned long flags;
493 unsigned int cpu;
494
495 if (count < strlen(MSM_PM_STATS_RESET)) {
496 ret = -EINVAL;
497 goto write_proc_failed;
498 }
499
500 if (copy_from_user(buf, buffer, strlen(MSM_PM_STATS_RESET))) {
501 ret = -EFAULT;
502 goto write_proc_failed;
503 }
504
505 if (memcmp(buf, MSM_PM_STATS_RESET, strlen(MSM_PM_STATS_RESET))) {
506 ret = -EINVAL;
507 goto write_proc_failed;
508 }
509
510 spin_lock_irqsave(&msm_pm_stats_lock, flags);
511 for_each_possible_cpu(cpu) {
512 struct msm_pm_time_stats *stats;
513 int i;
514
515 stats = per_cpu(msm_pm_stats, cpu).stats;
516 for (i = 0; i < MSM_PM_STAT_COUNT; i++) {
517 memset(stats[i].bucket,
518 0, sizeof(stats[i].bucket));
519 memset(stats[i].min_time,
520 0, sizeof(stats[i].min_time));
521 memset(stats[i].max_time,
522 0, sizeof(stats[i].max_time));
523 stats[i].count = 0;
524 stats[i].total_time = 0;
525 }
526 }
527
528 spin_unlock_irqrestore(&msm_pm_stats_lock, flags);
529 return count;
530
531write_proc_failed:
532 return ret;
533}
534#undef MSM_PM_STATS_RESET
535#endif /* CONFIG_MSM_IDLE_STATS */
536
537
538/******************************************************************************
539 * Configure Hardware before/after Low Power Mode
540 *****************************************************************************/
541
542/*
543 * Configure hardware registers in preparation for Apps power down.
544 */
545static void msm_pm_config_hw_before_power_down(void)
546{
547 return;
548}
549
550/*
551 * Clear hardware registers after Apps powers up.
552 */
553static void msm_pm_config_hw_after_power_up(void)
554{
555 return;
556}
557
558/*
559 * Configure hardware registers in preparation for SWFI.
560 */
561static void msm_pm_config_hw_before_swfi(void)
562{
563 return;
564}
565
566
567/******************************************************************************
568 * Suspend Max Sleep Time
569 *****************************************************************************/
570
571#ifdef CONFIG_MSM_SLEEP_TIME_OVERRIDE
572static int msm_pm_sleep_time_override;
573module_param_named(sleep_time_override,
574 msm_pm_sleep_time_override, int, S_IRUGO | S_IWUSR | S_IWGRP);
575#endif
576
577#define SCLK_HZ (32768)
578#define MSM_PM_SLEEP_TICK_LIMIT (0x6DDD000)
579
580static uint32_t msm_pm_max_sleep_time;
581
582/*
583 * Convert time from nanoseconds to slow clock ticks, then cap it to the
584 * specified limit
585 */
586static int64_t msm_pm_convert_and_cap_time(int64_t time_ns, int64_t limit)
587{
588 do_div(time_ns, NSEC_PER_SEC / SCLK_HZ);
589 return (time_ns > limit) ? limit : time_ns;
590}
591
592/*
593 * Set the sleep time for suspend. 0 means infinite sleep time.
594 */
595void msm_pm_set_max_sleep_time(int64_t max_sleep_time_ns)
596{
597 if (max_sleep_time_ns == 0) {
598 msm_pm_max_sleep_time = 0;
599 } else {
600 msm_pm_max_sleep_time = (uint32_t)msm_pm_convert_and_cap_time(
601 max_sleep_time_ns, MSM_PM_SLEEP_TICK_LIMIT);
602
603 if (msm_pm_max_sleep_time == 0)
604 msm_pm_max_sleep_time = 1;
605 }
606
607 if (msm_pm_debug_mask & MSM_PM_DEBUG_SUSPEND)
608 pr_info("%s: Requested %lld ns Giving %u sclk ticks\n",
609 __func__, max_sleep_time_ns, msm_pm_max_sleep_time);
610}
611EXPORT_SYMBOL(msm_pm_set_max_sleep_time);
612
613
614/******************************************************************************
615 *
616 *****************************************************************************/
617
618struct msm_pm_device {
619 unsigned int cpu;
620#ifdef CONFIG_HOTPLUG_CPU
621 struct completion cpu_killed;
622 unsigned int warm_boot;
623#endif
624};
625
626static DEFINE_PER_CPU_SHARED_ALIGNED(struct msm_pm_device, msm_pm_devices);
627static struct msm_rpmrs_limits *msm_pm_idle_rs_limits;
628
629static void msm_pm_swfi(void)
630{
631 msm_pm_config_hw_before_swfi();
632 msm_arch_idle();
633}
634
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -0600635static bool msm_pm_spm_power_collapse(
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700636 struct msm_pm_device *dev, bool from_idle, bool notify_rpm)
637{
638 void *entry;
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -0600639 bool collapsed = 0;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700640 int ret;
641
642 if (MSM_PM_DEBUG_POWER_COLLAPSE & msm_pm_debug_mask)
643 pr_info("CPU%u: %s: notify_rpm %d\n",
644 dev->cpu, __func__, (int) notify_rpm);
645
646 ret = msm_spm_set_low_power_mode(
647 MSM_SPM_MODE_POWER_COLLAPSE, notify_rpm);
648 WARN_ON(ret);
649
650 entry = (!dev->cpu || from_idle) ?
651 msm_pm_collapse_exit : msm_secondary_startup;
652 msm_pm_write_boot_vector(dev->cpu, virt_to_phys(entry));
653
654 if (MSM_PM_DEBUG_RESET_VECTOR & msm_pm_debug_mask)
655 pr_info("CPU%u: %s: program vector to %p\n",
656 dev->cpu, __func__, entry);
657
658#ifdef CONFIG_VFP
659 vfp_flush_context();
660#endif
661
662 collapsed = msm_pm_collapse();
663
664 if (collapsed) {
665#ifdef CONFIG_VFP
666 vfp_reinit();
667#endif
668 cpu_init();
669 writel(0xF0, MSM_QGIC_CPU_BASE + GIC_CPU_PRIMASK);
670 writel(1, MSM_QGIC_CPU_BASE + GIC_CPU_CTRL);
671 local_fiq_enable();
672 }
673
674 if (MSM_PM_DEBUG_POWER_COLLAPSE & msm_pm_debug_mask)
675 pr_info("CPU%u: %s: msm_pm_collapse returned, collapsed %d\n",
676 dev->cpu, __func__, collapsed);
677
678 ret = msm_spm_set_low_power_mode(MSM_SPM_MODE_CLOCK_GATING, false);
679 WARN_ON(ret);
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -0600680 return collapsed;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700681}
682
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -0600683static bool msm_pm_power_collapse_standalone(bool from_idle)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700684{
685 struct msm_pm_device *dev = &__get_cpu_var(msm_pm_devices);
686 unsigned int avsdscr_setting;
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -0600687 bool collapsed;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700688
689 avsdscr_setting = avs_get_avsdscr();
690 avs_disable();
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -0600691 collapsed = msm_pm_spm_power_collapse(dev, from_idle, false);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700692 avs_reset_delays(avsdscr_setting);
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -0600693 return collapsed;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700694}
695
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -0600696static bool msm_pm_power_collapse(bool from_idle)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700697{
698 struct msm_pm_device *dev = &__get_cpu_var(msm_pm_devices);
699 unsigned long saved_acpuclk_rate;
700 unsigned int avsdscr_setting;
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -0600701 bool collapsed;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700702
703 if (MSM_PM_DEBUG_POWER_COLLAPSE & msm_pm_debug_mask)
704 pr_info("CPU%u: %s: idle %d\n",
705 dev->cpu, __func__, (int)from_idle);
706
707 msm_pm_config_hw_before_power_down();
708 if (MSM_PM_DEBUG_POWER_COLLAPSE & msm_pm_debug_mask)
709 pr_info("CPU%u: %s: pre power down\n", dev->cpu, __func__);
710
711 avsdscr_setting = avs_get_avsdscr();
712 avs_disable();
713
714 if (cpu_online(dev->cpu))
715 saved_acpuclk_rate = acpuclk_power_collapse();
716 else
717 saved_acpuclk_rate = 0;
718
719 if (MSM_PM_DEBUG_CLOCK & msm_pm_debug_mask)
720 pr_info("CPU%u: %s: change clock rate (old rate = %lu)\n",
721 dev->cpu, __func__, saved_acpuclk_rate);
722
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -0600723 collapsed = msm_pm_spm_power_collapse(dev, from_idle, true);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700724
725 if (MSM_PM_DEBUG_CLOCK & msm_pm_debug_mask)
726 pr_info("CPU%u: %s: restore clock rate to %lu\n",
727 dev->cpu, __func__, saved_acpuclk_rate);
728 if (acpuclk_set_rate(dev->cpu, saved_acpuclk_rate, SETRATE_PC) < 0)
729 pr_err("CPU%u: %s: failed to restore clock rate(%lu)\n",
730 dev->cpu, __func__, saved_acpuclk_rate);
731
732 avs_reset_delays(avsdscr_setting);
733 msm_pm_config_hw_after_power_up();
734 if (MSM_PM_DEBUG_POWER_COLLAPSE & msm_pm_debug_mask)
735 pr_info("CPU%u: %s: post power up\n", dev->cpu, __func__);
736
737 if (MSM_PM_DEBUG_POWER_COLLAPSE & msm_pm_debug_mask)
738 pr_info("CPU%u: %s: return\n", dev->cpu, __func__);
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -0600739 return collapsed;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700740}
741
742static irqreturn_t msm_pm_rpm_wakeup_interrupt(int irq, void *dev_id)
743{
744 if (dev_id != &msm_pm_rpm_wakeup_interrupt)
745 return IRQ_NONE;
746
747 return IRQ_HANDLED;
748}
749
750
751/******************************************************************************
752 * External Idle/Suspend Functions
753 *****************************************************************************/
754
755void arch_idle(void)
756{
757 return;
758}
759
760int msm_pm_idle_prepare(struct cpuidle_device *dev)
761{
762 uint32_t latency_us;
763 uint32_t sleep_us;
764 int i;
765
766 latency_us = (uint32_t) pm_qos_request(PM_QOS_CPU_DMA_LATENCY);
767 sleep_us = (uint32_t) ktime_to_ns(tick_nohz_get_sleep_length());
768 sleep_us = DIV_ROUND_UP(sleep_us, 1000);
769
770 for (i = 0; i < dev->state_count; i++) {
771 struct cpuidle_state *state = &dev->states[i];
772 enum msm_pm_sleep_mode mode;
773 bool allow;
774 struct msm_rpmrs_limits *rs_limits = NULL;
775 int idx;
776
777 mode = (enum msm_pm_sleep_mode) state->driver_data;
778 idx = MSM_PM_MODE(dev->cpu, mode);
779
780 allow = msm_pm_modes[idx].idle_enabled &&
781 msm_pm_modes[idx].idle_supported;
782
783 switch (mode) {
784 case MSM_PM_SLEEP_MODE_POWER_COLLAPSE:
785 if (!allow)
786 break;
787
788 if (num_online_cpus() > 1) {
789 allow = false;
790 break;
791 }
792#ifdef CONFIG_HAS_WAKELOCK
793 if (has_wake_lock(WAKE_LOCK_IDLE)) {
794 allow = false;
795 break;
796 }
797#endif
798 /* fall through */
799
800 case MSM_PM_SLEEP_MODE_POWER_COLLAPSE_STANDALONE:
801 if (!allow)
802 break;
803
804 if (!dev->cpu &&
805 msm_rpm_local_request_is_outstanding()) {
806 allow = false;
807 break;
808 }
809 /* fall through */
810
811 case MSM_PM_SLEEP_MODE_WAIT_FOR_INTERRUPT:
812 if (!allow)
813 break;
814
815 rs_limits = msm_rpmrs_lowest_limits(true,
816 mode, latency_us, sleep_us);
817
818 if (MSM_PM_DEBUG_IDLE & msm_pm_debug_mask)
819 pr_info("CPU%u: %s: %s, latency %uus, "
820 "sleep %uus, limit %p\n",
821 dev->cpu, __func__, state->desc,
822 latency_us, sleep_us, rs_limits);
823
824 if ((MSM_PM_DEBUG_IDLE_LIMITS & msm_pm_debug_mask) &&
825 rs_limits)
826 pr_info("CPU%u: %s: limit %p: "
827 "pxo %d, l2_cache %d, "
828 "vdd_mem %d, vdd_dig %d\n",
829 dev->cpu, __func__, rs_limits,
830 rs_limits->pxo,
831 rs_limits->l2_cache,
832 rs_limits->vdd_mem,
833 rs_limits->vdd_dig);
834
835 if (!rs_limits)
836 allow = false;
837 break;
838
839 default:
840 allow = false;
841 break;
842 }
843
844 if (MSM_PM_DEBUG_IDLE & msm_pm_debug_mask)
845 pr_info("CPU%u: %s: allow %s: %d\n",
846 dev->cpu, __func__, state->desc, (int)allow);
847
848 if (allow) {
849 state->flags &= ~CPUIDLE_FLAG_IGNORE;
850 state->target_residency = 0;
851 state->exit_latency = 0;
852 state->power_usage = rs_limits->power[dev->cpu];
853
854 if (MSM_PM_SLEEP_MODE_POWER_COLLAPSE == mode)
855 msm_pm_idle_rs_limits = rs_limits;
856 } else {
857 state->flags |= CPUIDLE_FLAG_IGNORE;
858 }
859 }
860
861 return 0;
862}
863
864int msm_pm_idle_enter(enum msm_pm_sleep_mode sleep_mode)
865{
866 int64_t time;
867#ifdef CONFIG_MSM_IDLE_STATS
868 int exit_stat;
869#endif
870
871 if (MSM_PM_DEBUG_IDLE & msm_pm_debug_mask)
872 pr_info("CPU%u: %s: mode %d\n",
873 smp_processor_id(), __func__, sleep_mode);
874
875 time = ktime_to_ns(ktime_get());
876
877 switch (sleep_mode) {
878 case MSM_PM_SLEEP_MODE_WAIT_FOR_INTERRUPT:
879 msm_pm_swfi();
880#ifdef CONFIG_MSM_IDLE_STATS
881 exit_stat = MSM_PM_STAT_IDLE_WFI;
882#endif
883 break;
884
885 case MSM_PM_SLEEP_MODE_POWER_COLLAPSE_STANDALONE:
886 msm_pm_power_collapse_standalone(true);
887#ifdef CONFIG_MSM_IDLE_STATS
888 exit_stat = MSM_PM_STAT_IDLE_STANDALONE_POWER_COLLAPSE;
889#endif
890 break;
891
892 case MSM_PM_SLEEP_MODE_POWER_COLLAPSE: {
893 int64_t timer_expiration = msm_timer_enter_idle();
894 bool timer_halted = false;
895 uint32_t sleep_delay;
896 int ret;
897 int notify_rpm =
898 (sleep_mode == MSM_PM_SLEEP_MODE_POWER_COLLAPSE);
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -0600899 int collapsed;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700900
901 sleep_delay = (uint32_t) msm_pm_convert_and_cap_time(
902 timer_expiration, MSM_PM_SLEEP_TICK_LIMIT);
903 if (sleep_delay == 0) /* 0 would mean infinite time */
904 sleep_delay = 1;
905
906 ret = msm_rpmrs_enter_sleep(
907 sleep_delay, msm_pm_idle_rs_limits, true, notify_rpm);
908 if (!ret) {
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -0600909 collapsed = msm_pm_power_collapse(true);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700910 timer_halted = true;
911
912 msm_rpmrs_exit_sleep(msm_pm_idle_rs_limits, true,
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -0600913 notify_rpm, collapsed);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700914 }
915
916 msm_timer_exit_idle((int) timer_halted);
917#ifdef CONFIG_MSM_IDLE_STATS
918 exit_stat = MSM_PM_STAT_IDLE_POWER_COLLAPSE;
919#endif
920 break;
921 }
922
923 default:
924 __WARN();
925 goto cpuidle_enter_bail;
926 }
927
928 time = ktime_to_ns(ktime_get()) - time;
929#ifdef CONFIG_MSM_IDLE_STATS
930 msm_pm_add_stat(exit_stat, time);
931#endif
932
933 do_div(time, 1000);
934 return (int) time;
935
936cpuidle_enter_bail:
937 return 0;
938}
939
940static int msm_pm_enter(suspend_state_t state)
941{
942 bool allow[MSM_PM_SLEEP_MODE_NR];
943 int i;
944
945#ifdef CONFIG_MSM_IDLE_STATS
946 int64_t period = 0;
947 int64_t time = msm_timer_get_sclk_time(&period);
948#endif
949
950 if (MSM_PM_DEBUG_SUSPEND & msm_pm_debug_mask)
951 pr_info("%s\n", __func__);
952
953 if (smp_processor_id()) {
954 __WARN();
955 goto enter_exit;
956 }
957
958
959 for (i = 0; i < MSM_PM_SLEEP_MODE_NR; i++) {
960 struct msm_pm_platform_data *mode;
961
962 mode = &msm_pm_modes[MSM_PM_MODE(0, i)];
963 allow[i] = mode->suspend_supported && mode->suspend_enabled;
964 }
965
966 if (allow[MSM_PM_SLEEP_MODE_POWER_COLLAPSE]) {
967 struct msm_rpmrs_limits *rs_limits;
968 int ret;
969
970 if (MSM_PM_DEBUG_SUSPEND & msm_pm_debug_mask)
971 pr_info("%s: power collapse\n", __func__);
972
973 clock_debug_print_enabled();
974
975#ifdef CONFIG_MSM_SLEEP_TIME_OVERRIDE
976 if (msm_pm_sleep_time_override > 0) {
977 int64_t ns = NSEC_PER_SEC *
978 (int64_t) msm_pm_sleep_time_override;
979 msm_pm_set_max_sleep_time(ns);
980 msm_pm_sleep_time_override = 0;
981 }
982#endif /* CONFIG_MSM_SLEEP_TIME_OVERRIDE */
983
984 if (MSM_PM_DEBUG_SUSPEND_LIMITS & msm_pm_debug_mask)
985 msm_rpmrs_show_resources();
986
987 rs_limits = msm_rpmrs_lowest_limits(false,
988 MSM_PM_SLEEP_MODE_POWER_COLLAPSE, -1, -1);
989
990 if ((MSM_PM_DEBUG_SUSPEND_LIMITS & msm_pm_debug_mask) &&
991 rs_limits)
992 pr_info("%s: limit %p: pxo %d, l2_cache %d, "
993 "vdd_mem %d, vdd_dig %d\n",
994 __func__, rs_limits,
995 rs_limits->pxo, rs_limits->l2_cache,
996 rs_limits->vdd_mem, rs_limits->vdd_dig);
997
998 if (rs_limits) {
999 ret = msm_rpmrs_enter_sleep(
1000 msm_pm_max_sleep_time, rs_limits, false, true);
1001 if (!ret) {
Maheshkumar Sivasubramaniandd93ecf2011-09-15 19:39:14 -06001002 int collapsed = msm_pm_power_collapse(false);
1003 msm_rpmrs_exit_sleep(rs_limits, false, true,
1004 collapsed);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001005 }
1006 } else {
1007 pr_err("%s: cannot find the lowest power limit\n",
1008 __func__);
1009 }
1010
1011#ifdef CONFIG_MSM_IDLE_STATS
1012 if (time != 0) {
1013 int64_t end_time = msm_timer_get_sclk_time(NULL);
1014 if (end_time != 0) {
1015 time = end_time - time;
1016 if (time < 0)
1017 time += period;
1018 } else
1019 time = 0;
1020 }
1021
1022 msm_pm_add_stat(MSM_PM_STAT_SUSPEND, time);
1023#endif /* CONFIG_MSM_IDLE_STATS */
1024 } else if (allow[MSM_PM_SLEEP_MODE_POWER_COLLAPSE_STANDALONE]) {
1025 if (MSM_PM_DEBUG_SUSPEND & msm_pm_debug_mask)
1026 pr_info("%s: standalone power collapse\n", __func__);
1027 msm_pm_power_collapse_standalone(false);
1028 } else if (allow[MSM_PM_SLEEP_MODE_WAIT_FOR_INTERRUPT]) {
1029 if (MSM_PM_DEBUG_SUSPEND & msm_pm_debug_mask)
1030 pr_info("%s: swfi\n", __func__);
1031 msm_pm_swfi();
1032 }
1033
1034
1035enter_exit:
1036 if (MSM_PM_DEBUG_SUSPEND & msm_pm_debug_mask)
1037 pr_info("%s: return\n", __func__);
1038
1039 return 0;
1040}
1041
1042static struct platform_suspend_ops msm_pm_ops = {
1043 .enter = msm_pm_enter,
1044 .valid = suspend_valid_only_mem,
1045};
1046
1047#ifdef CONFIG_HOTPLUG_CPU
1048int platform_cpu_disable(unsigned int cpu)
1049{
1050 return cpu == 0 ? -EPERM : 0;
1051}
1052
1053int platform_cpu_kill(unsigned int cpu)
1054{
1055 struct completion *killed = &per_cpu(msm_pm_devices, cpu).cpu_killed;
1056 return wait_for_completion_timeout(killed, HZ * 5);
1057}
1058
1059void platform_cpu_die(unsigned int cpu)
1060{
1061 bool allow[MSM_PM_SLEEP_MODE_NR];
1062 int i;
1063
1064 if (unlikely(cpu != smp_processor_id())) {
1065 pr_crit("%s: running on %u, should be %u\n",
1066 __func__, smp_processor_id(), cpu);
1067 BUG();
1068 }
1069
1070 for (i = 0; i < MSM_PM_SLEEP_MODE_NR; i++) {
1071 struct msm_pm_platform_data *mode;
1072
1073 mode = &msm_pm_modes[MSM_PM_MODE(cpu, i)];
1074 allow[i] = mode->suspend_supported && mode->suspend_enabled;
1075 }
1076
1077 if (MSM_PM_DEBUG_HOTPLUG & msm_pm_debug_mask)
1078 pr_notice("CPU%u: %s: shutting down cpu\n", cpu, __func__);
1079 complete(&__get_cpu_var(msm_pm_devices).cpu_killed);
1080
1081 flush_cache_all();
1082
1083 for (;;) {
1084 if (allow[MSM_PM_SLEEP_MODE_POWER_COLLAPSE])
1085 msm_pm_power_collapse(false);
1086 else if (allow[MSM_PM_SLEEP_MODE_POWER_COLLAPSE_STANDALONE])
1087 msm_pm_power_collapse_standalone(false);
1088 else if (allow[MSM_PM_SLEEP_MODE_WAIT_FOR_INTERRUPT])
1089 msm_pm_swfi();
1090
1091 if (pen_release == cpu) {
1092 /* OK, proper wakeup, we're done */
1093 break;
1094 }
1095 }
1096
1097 pen_release = -1;
1098 pr_notice("CPU%u: %s: normal wakeup\n", cpu, __func__);
1099}
1100
1101int msm_pm_platform_secondary_init(unsigned int cpu)
1102{
1103 int ret;
1104 struct msm_pm_device *dev = &__get_cpu_var(msm_pm_devices);
1105
1106 if (!dev->warm_boot) {
1107 dev->warm_boot = 1;
1108 return 0;
1109 }
Pratik Patele5771792011-09-17 18:33:54 -07001110 etm_restore_reg_check();
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001111#ifdef CONFIG_VFP
1112 vfp_reinit();
1113#endif
1114 ret = msm_spm_set_low_power_mode(MSM_SPM_MODE_CLOCK_GATING, false);
1115
1116 return ret;
1117}
1118#endif /* CONFIG_HOTPLUG_CPU */
1119
1120/******************************************************************************
1121 * Initialization routine
1122 *****************************************************************************/
1123
1124static int __init msm_pm_init(void)
1125{
1126 pgd_t *pc_pgd;
1127 pmd_t *pmd;
1128 unsigned long pmdval;
1129 unsigned int cpu;
1130#ifdef CONFIG_MSM_IDLE_STATS
1131 struct proc_dir_entry *d_entry;
1132#endif
1133 int ret;
1134
1135 /* Page table for cores to come back up safely. */
1136 pc_pgd = pgd_alloc(&init_mm);
1137 if (!pc_pgd)
1138 return -ENOMEM;
1139
1140 pmd = pmd_offset(pc_pgd +
1141 pgd_index(virt_to_phys(msm_pm_collapse_exit)),
1142 virt_to_phys(msm_pm_collapse_exit));
1143 pmdval = (virt_to_phys(msm_pm_collapse_exit) & PGDIR_MASK) |
1144 PMD_TYPE_SECT | PMD_SECT_AP_WRITE;
1145 pmd[0] = __pmd(pmdval);
1146 pmd[1] = __pmd(pmdval + (1 << (PGDIR_SHIFT - 1)));
1147
1148 /* It is remotely possible that the code in msm_pm_collapse_exit()
1149 * which turns on the MMU with this mapping is in the
1150 * next even-numbered megabyte beyond the
1151 * start of msm_pm_collapse_exit().
1152 * Map this megabyte in as well.
1153 */
1154 pmd[2] = __pmd(pmdval + (2 << (PGDIR_SHIFT - 1)));
1155 flush_pmd_entry(pmd);
1156 msm_pm_pc_pgd = virt_to_phys(pc_pgd);
1157
1158 ret = request_irq(rpm_cpu0_wakeup_irq,
1159 msm_pm_rpm_wakeup_interrupt, IRQF_TRIGGER_RISING,
1160 "pm_drv", msm_pm_rpm_wakeup_interrupt);
1161 if (ret) {
1162 pr_err("%s: failed to request irq %u: %d\n",
1163 __func__, rpm_cpu0_wakeup_irq, ret);
1164 return ret;
1165 }
1166
1167 ret = irq_set_irq_wake(rpm_cpu0_wakeup_irq, 1);
1168 if (ret) {
1169 pr_err("%s: failed to set wakeup irq %u: %d\n",
1170 __func__, rpm_cpu0_wakeup_irq, ret);
1171 return ret;
1172 }
1173
1174 for_each_possible_cpu(cpu) {
1175 struct msm_pm_device *dev = &per_cpu(msm_pm_devices, cpu);
1176
1177 dev->cpu = cpu;
1178#ifdef CONFIG_HOTPLUG_CPU
1179 init_completion(&dev->cpu_killed);
1180#endif
1181 }
Maheshkumar Sivasubramaniane2fc3f62011-09-15 09:22:42 -06001182#ifdef CONFIG_MSM_SCM
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001183 ret = scm_set_boot_addr((void *)virt_to_phys(msm_pm_boot_entry),
1184 SCM_FLAG_WARMBOOT_CPU0 | SCM_FLAG_WARMBOOT_CPU1);
1185 if (ret) {
1186 pr_err("%s: failed to set up scm boot addr: %d\n",
1187 __func__, ret);
1188 return ret;
1189 }
Maheshkumar Sivasubramaniane2fc3f62011-09-15 09:22:42 -06001190#endif
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001191
1192#ifdef CONFIG_MSM_IDLE_STATS
1193 for_each_possible_cpu(cpu) {
1194 struct msm_pm_time_stats *stats =
1195 per_cpu(msm_pm_stats, cpu).stats;
1196
1197 stats[MSM_PM_STAT_REQUESTED_IDLE].name = "idle-request";
1198 stats[MSM_PM_STAT_REQUESTED_IDLE].first_bucket_time =
1199 CONFIG_MSM_IDLE_STATS_FIRST_BUCKET;
1200
1201 stats[MSM_PM_STAT_IDLE_WFI].name = "idle-wfi";
1202 stats[MSM_PM_STAT_IDLE_WFI].first_bucket_time =
1203 CONFIG_MSM_IDLE_STATS_FIRST_BUCKET;
1204
1205 stats[MSM_PM_STAT_IDLE_STANDALONE_POWER_COLLAPSE].name =
1206 "idle-standalone-power-collapse";
1207 stats[MSM_PM_STAT_IDLE_STANDALONE_POWER_COLLAPSE].
1208 first_bucket_time = CONFIG_MSM_IDLE_STATS_FIRST_BUCKET;
1209
1210 stats[MSM_PM_STAT_IDLE_POWER_COLLAPSE].name =
1211 "idle-power-collapse";
1212 stats[MSM_PM_STAT_IDLE_POWER_COLLAPSE].first_bucket_time =
1213 CONFIG_MSM_IDLE_STATS_FIRST_BUCKET;
1214
1215 stats[MSM_PM_STAT_SUSPEND].name = "suspend";
1216 stats[MSM_PM_STAT_SUSPEND].first_bucket_time =
1217 CONFIG_MSM_SUSPEND_STATS_FIRST_BUCKET;
1218 }
1219
1220 d_entry = create_proc_entry("msm_pm_stats",
1221 S_IRUGO | S_IWUSR | S_IWGRP, NULL);
1222 if (d_entry) {
1223 d_entry->read_proc = msm_pm_read_proc;
1224 d_entry->write_proc = msm_pm_write_proc;
1225 d_entry->data = NULL;
1226 }
1227#endif /* CONFIG_MSM_IDLE_STATS */
1228
1229 msm_pm_mode_sysfs_add();
1230 msm_spm_allow_x_cpu_set_vdd(false);
1231
1232 suspend_set_ops(&msm_pm_ops);
1233 msm_cpuidle_init();
1234
1235 return 0;
1236}
1237
1238late_initcall(msm_pm_init);