]> git.proxmox.com Git - mirror_ubuntu-artful-kernel.git/blob - drivers/base/power/wakeup.c
Merge branch 'pm-wakeirq'
[mirror_ubuntu-artful-kernel.git] / drivers / base / power / wakeup.c
1 /*
2 * drivers/base/power/wakeup.c - System wakeup events framework
3 *
4 * Copyright (c) 2010 Rafael J. Wysocki <rjw@sisk.pl>, Novell Inc.
5 *
6 * This file is released under the GPLv2.
7 */
8
9 #include <linux/device.h>
10 #include <linux/slab.h>
11 #include <linux/sched.h>
12 #include <linux/capability.h>
13 #include <linux/export.h>
14 #include <linux/suspend.h>
15 #include <linux/seq_file.h>
16 #include <linux/debugfs.h>
17 #include <linux/pm_wakeirq.h>
18 #include <trace/events/power.h>
19
20 #include "power.h"
21
22 /*
23 * If set, the suspend/hibernate code will abort transitions to a sleep state
24 * if wakeup events are registered during or immediately before the transition.
25 */
26 bool events_check_enabled __read_mostly;
27
28 /* If set and the system is suspending, terminate the suspend. */
29 static bool pm_abort_suspend __read_mostly;
30
31 /*
32 * Combined counters of registered wakeup events and wakeup events in progress.
33 * They need to be modified together atomically, so it's better to use one
34 * atomic variable to hold them both.
35 */
36 static atomic_t combined_event_count = ATOMIC_INIT(0);
37
38 #define IN_PROGRESS_BITS (sizeof(int) * 4)
39 #define MAX_IN_PROGRESS ((1 << IN_PROGRESS_BITS) - 1)
40
41 static void split_counters(unsigned int *cnt, unsigned int *inpr)
42 {
43 unsigned int comb = atomic_read(&combined_event_count);
44
45 *cnt = (comb >> IN_PROGRESS_BITS);
46 *inpr = comb & MAX_IN_PROGRESS;
47 }
48
49 /* A preserved old value of the events counter. */
50 static unsigned int saved_count;
51
52 static DEFINE_SPINLOCK(events_lock);
53
54 static void pm_wakeup_timer_fn(unsigned long data);
55
56 static LIST_HEAD(wakeup_sources);
57
58 static DECLARE_WAIT_QUEUE_HEAD(wakeup_count_wait_queue);
59
60 static struct wakeup_source deleted_ws = {
61 .name = "deleted",
62 .lock = __SPIN_LOCK_UNLOCKED(deleted_ws.lock),
63 };
64
65 /**
66 * wakeup_source_prepare - Prepare a new wakeup source for initialization.
67 * @ws: Wakeup source to prepare.
68 * @name: Pointer to the name of the new wakeup source.
69 *
70 * Callers must ensure that the @name string won't be freed when @ws is still in
71 * use.
72 */
73 void wakeup_source_prepare(struct wakeup_source *ws, const char *name)
74 {
75 if (ws) {
76 memset(ws, 0, sizeof(*ws));
77 ws->name = name;
78 }
79 }
80 EXPORT_SYMBOL_GPL(wakeup_source_prepare);
81
82 /**
83 * wakeup_source_create - Create a struct wakeup_source object.
84 * @name: Name of the new wakeup source.
85 */
86 struct wakeup_source *wakeup_source_create(const char *name)
87 {
88 struct wakeup_source *ws;
89
90 ws = kmalloc(sizeof(*ws), GFP_KERNEL);
91 if (!ws)
92 return NULL;
93
94 wakeup_source_prepare(ws, name ? kstrdup(name, GFP_KERNEL) : NULL);
95 return ws;
96 }
97 EXPORT_SYMBOL_GPL(wakeup_source_create);
98
99 /**
100 * wakeup_source_drop - Prepare a struct wakeup_source object for destruction.
101 * @ws: Wakeup source to prepare for destruction.
102 *
103 * Callers must ensure that __pm_stay_awake() or __pm_wakeup_event() will never
104 * be run in parallel with this function for the same wakeup source object.
105 */
106 void wakeup_source_drop(struct wakeup_source *ws)
107 {
108 if (!ws)
109 return;
110
111 del_timer_sync(&ws->timer);
112 __pm_relax(ws);
113 }
114 EXPORT_SYMBOL_GPL(wakeup_source_drop);
115
116 /*
117 * Record wakeup_source statistics being deleted into a dummy wakeup_source.
118 */
119 static void wakeup_source_record(struct wakeup_source *ws)
120 {
121 unsigned long flags;
122
123 spin_lock_irqsave(&deleted_ws.lock, flags);
124
125 if (ws->event_count) {
126 deleted_ws.total_time =
127 ktime_add(deleted_ws.total_time, ws->total_time);
128 deleted_ws.prevent_sleep_time =
129 ktime_add(deleted_ws.prevent_sleep_time,
130 ws->prevent_sleep_time);
131 deleted_ws.max_time =
132 ktime_compare(deleted_ws.max_time, ws->max_time) > 0 ?
133 deleted_ws.max_time : ws->max_time;
134 deleted_ws.event_count += ws->event_count;
135 deleted_ws.active_count += ws->active_count;
136 deleted_ws.relax_count += ws->relax_count;
137 deleted_ws.expire_count += ws->expire_count;
138 deleted_ws.wakeup_count += ws->wakeup_count;
139 }
140
141 spin_unlock_irqrestore(&deleted_ws.lock, flags);
142 }
143
144 /**
145 * wakeup_source_destroy - Destroy a struct wakeup_source object.
146 * @ws: Wakeup source to destroy.
147 *
148 * Use only for wakeup source objects created with wakeup_source_create().
149 */
150 void wakeup_source_destroy(struct wakeup_source *ws)
151 {
152 if (!ws)
153 return;
154
155 wakeup_source_drop(ws);
156 wakeup_source_record(ws);
157 kfree(ws->name);
158 kfree(ws);
159 }
160 EXPORT_SYMBOL_GPL(wakeup_source_destroy);
161
162 /**
163 * wakeup_source_add - Add given object to the list of wakeup sources.
164 * @ws: Wakeup source object to add to the list.
165 */
166 void wakeup_source_add(struct wakeup_source *ws)
167 {
168 unsigned long flags;
169
170 if (WARN_ON(!ws))
171 return;
172
173 spin_lock_init(&ws->lock);
174 setup_timer(&ws->timer, pm_wakeup_timer_fn, (unsigned long)ws);
175 ws->active = false;
176 ws->last_time = ktime_get();
177
178 spin_lock_irqsave(&events_lock, flags);
179 list_add_rcu(&ws->entry, &wakeup_sources);
180 spin_unlock_irqrestore(&events_lock, flags);
181 }
182 EXPORT_SYMBOL_GPL(wakeup_source_add);
183
184 /**
185 * wakeup_source_remove - Remove given object from the wakeup sources list.
186 * @ws: Wakeup source object to remove from the list.
187 */
188 void wakeup_source_remove(struct wakeup_source *ws)
189 {
190 unsigned long flags;
191
192 if (WARN_ON(!ws))
193 return;
194
195 spin_lock_irqsave(&events_lock, flags);
196 list_del_rcu(&ws->entry);
197 spin_unlock_irqrestore(&events_lock, flags);
198 synchronize_rcu();
199 }
200 EXPORT_SYMBOL_GPL(wakeup_source_remove);
201
202 /**
203 * wakeup_source_register - Create wakeup source and add it to the list.
204 * @name: Name of the wakeup source to register.
205 */
206 struct wakeup_source *wakeup_source_register(const char *name)
207 {
208 struct wakeup_source *ws;
209
210 ws = wakeup_source_create(name);
211 if (ws)
212 wakeup_source_add(ws);
213
214 return ws;
215 }
216 EXPORT_SYMBOL_GPL(wakeup_source_register);
217
218 /**
219 * wakeup_source_unregister - Remove wakeup source from the list and remove it.
220 * @ws: Wakeup source object to unregister.
221 */
222 void wakeup_source_unregister(struct wakeup_source *ws)
223 {
224 if (ws) {
225 wakeup_source_remove(ws);
226 wakeup_source_destroy(ws);
227 }
228 }
229 EXPORT_SYMBOL_GPL(wakeup_source_unregister);
230
231 /**
232 * device_wakeup_attach - Attach a wakeup source object to a device object.
233 * @dev: Device to handle.
234 * @ws: Wakeup source object to attach to @dev.
235 *
236 * This causes @dev to be treated as a wakeup device.
237 */
238 static int device_wakeup_attach(struct device *dev, struct wakeup_source *ws)
239 {
240 spin_lock_irq(&dev->power.lock);
241 if (dev->power.wakeup) {
242 spin_unlock_irq(&dev->power.lock);
243 return -EEXIST;
244 }
245 dev->power.wakeup = ws;
246 spin_unlock_irq(&dev->power.lock);
247 return 0;
248 }
249
250 /**
251 * device_wakeup_enable - Enable given device to be a wakeup source.
252 * @dev: Device to handle.
253 *
254 * Create a wakeup source object, register it and attach it to @dev.
255 */
256 int device_wakeup_enable(struct device *dev)
257 {
258 struct wakeup_source *ws;
259 int ret;
260
261 if (!dev || !dev->power.can_wakeup)
262 return -EINVAL;
263
264 ws = wakeup_source_register(dev_name(dev));
265 if (!ws)
266 return -ENOMEM;
267
268 ret = device_wakeup_attach(dev, ws);
269 if (ret)
270 wakeup_source_unregister(ws);
271
272 return ret;
273 }
274 EXPORT_SYMBOL_GPL(device_wakeup_enable);
275
276 /**
277 * device_wakeup_attach_irq - Attach a wakeirq to a wakeup source
278 * @dev: Device to handle
279 * @wakeirq: Device specific wakeirq entry
280 *
281 * Attach a device wakeirq to the wakeup source so the device
282 * wake IRQ can be configured automatically for suspend and
283 * resume.
284 */
285 int device_wakeup_attach_irq(struct device *dev,
286 struct wake_irq *wakeirq)
287 {
288 struct wakeup_source *ws;
289 int ret = 0;
290
291 spin_lock_irq(&dev->power.lock);
292 ws = dev->power.wakeup;
293 if (!ws) {
294 dev_err(dev, "forgot to call call device_init_wakeup?\n");
295 ret = -EINVAL;
296 goto unlock;
297 }
298
299 if (ws->wakeirq) {
300 ret = -EEXIST;
301 goto unlock;
302 }
303
304 ws->wakeirq = wakeirq;
305
306 unlock:
307 spin_unlock_irq(&dev->power.lock);
308
309 return ret;
310 }
311
312 /**
313 * device_wakeup_detach_irq - Detach a wakeirq from a wakeup source
314 * @dev: Device to handle
315 *
316 * Removes a device wakeirq from the wakeup source.
317 */
318 void device_wakeup_detach_irq(struct device *dev)
319 {
320 struct wakeup_source *ws;
321
322 spin_lock_irq(&dev->power.lock);
323 ws = dev->power.wakeup;
324 if (!ws)
325 goto unlock;
326
327 ws->wakeirq = NULL;
328
329 unlock:
330 spin_unlock_irq(&dev->power.lock);
331 }
332
333 /**
334 * device_wakeup_arm_wake_irqs(void)
335 *
336 * Itereates over the list of device wakeirqs to arm them.
337 */
338 void device_wakeup_arm_wake_irqs(void)
339 {
340 struct wakeup_source *ws;
341
342 rcu_read_lock();
343 list_for_each_entry_rcu(ws, &wakeup_sources, entry) {
344 if (ws->wakeirq)
345 dev_pm_arm_wake_irq(ws->wakeirq);
346 }
347 rcu_read_unlock();
348 }
349
350 /**
351 * device_wakeup_disarm_wake_irqs(void)
352 *
353 * Itereates over the list of device wakeirqs to disarm them.
354 */
355 void device_wakeup_disarm_wake_irqs(void)
356 {
357 struct wakeup_source *ws;
358
359 rcu_read_lock();
360 list_for_each_entry_rcu(ws, &wakeup_sources, entry) {
361 if (ws->wakeirq)
362 dev_pm_disarm_wake_irq(ws->wakeirq);
363 }
364 rcu_read_unlock();
365 }
366
367 /**
368 * device_wakeup_detach - Detach a device's wakeup source object from it.
369 * @dev: Device to detach the wakeup source object from.
370 *
371 * After it returns, @dev will not be treated as a wakeup device any more.
372 */
373 static struct wakeup_source *device_wakeup_detach(struct device *dev)
374 {
375 struct wakeup_source *ws;
376
377 spin_lock_irq(&dev->power.lock);
378 ws = dev->power.wakeup;
379 dev->power.wakeup = NULL;
380 spin_unlock_irq(&dev->power.lock);
381 return ws;
382 }
383
384 /**
385 * device_wakeup_disable - Do not regard a device as a wakeup source any more.
386 * @dev: Device to handle.
387 *
388 * Detach the @dev's wakeup source object from it, unregister this wakeup source
389 * object and destroy it.
390 */
391 int device_wakeup_disable(struct device *dev)
392 {
393 struct wakeup_source *ws;
394
395 if (!dev || !dev->power.can_wakeup)
396 return -EINVAL;
397
398 ws = device_wakeup_detach(dev);
399 if (ws)
400 wakeup_source_unregister(ws);
401
402 return 0;
403 }
404 EXPORT_SYMBOL_GPL(device_wakeup_disable);
405
406 /**
407 * device_set_wakeup_capable - Set/reset device wakeup capability flag.
408 * @dev: Device to handle.
409 * @capable: Whether or not @dev is capable of waking up the system from sleep.
410 *
411 * If @capable is set, set the @dev's power.can_wakeup flag and add its
412 * wakeup-related attributes to sysfs. Otherwise, unset the @dev's
413 * power.can_wakeup flag and remove its wakeup-related attributes from sysfs.
414 *
415 * This function may sleep and it can't be called from any context where
416 * sleeping is not allowed.
417 */
418 void device_set_wakeup_capable(struct device *dev, bool capable)
419 {
420 if (!!dev->power.can_wakeup == !!capable)
421 return;
422
423 if (device_is_registered(dev) && !list_empty(&dev->power.entry)) {
424 if (capable) {
425 if (wakeup_sysfs_add(dev))
426 return;
427 } else {
428 wakeup_sysfs_remove(dev);
429 }
430 }
431 dev->power.can_wakeup = capable;
432 }
433 EXPORT_SYMBOL_GPL(device_set_wakeup_capable);
434
435 /**
436 * device_init_wakeup - Device wakeup initialization.
437 * @dev: Device to handle.
438 * @enable: Whether or not to enable @dev as a wakeup device.
439 *
440 * By default, most devices should leave wakeup disabled. The exceptions are
441 * devices that everyone expects to be wakeup sources: keyboards, power buttons,
442 * possibly network interfaces, etc. Also, devices that don't generate their
443 * own wakeup requests but merely forward requests from one bus to another
444 * (like PCI bridges) should have wakeup enabled by default.
445 */
446 int device_init_wakeup(struct device *dev, bool enable)
447 {
448 int ret = 0;
449
450 if (!dev)
451 return -EINVAL;
452
453 if (enable) {
454 device_set_wakeup_capable(dev, true);
455 ret = device_wakeup_enable(dev);
456 } else {
457 if (dev->power.can_wakeup)
458 device_wakeup_disable(dev);
459
460 device_set_wakeup_capable(dev, false);
461 }
462
463 return ret;
464 }
465 EXPORT_SYMBOL_GPL(device_init_wakeup);
466
467 /**
468 * device_set_wakeup_enable - Enable or disable a device to wake up the system.
469 * @dev: Device to handle.
470 */
471 int device_set_wakeup_enable(struct device *dev, bool enable)
472 {
473 if (!dev || !dev->power.can_wakeup)
474 return -EINVAL;
475
476 return enable ? device_wakeup_enable(dev) : device_wakeup_disable(dev);
477 }
478 EXPORT_SYMBOL_GPL(device_set_wakeup_enable);
479
480 /**
481 * wakeup_source_not_registered - validate the given wakeup source.
482 * @ws: Wakeup source to be validated.
483 */
484 static bool wakeup_source_not_registered(struct wakeup_source *ws)
485 {
486 /*
487 * Use timer struct to check if the given source is initialized
488 * by wakeup_source_add.
489 */
490 return ws->timer.function != pm_wakeup_timer_fn ||
491 ws->timer.data != (unsigned long)ws;
492 }
493
494 /*
495 * The functions below use the observation that each wakeup event starts a
496 * period in which the system should not be suspended. The moment this period
497 * will end depends on how the wakeup event is going to be processed after being
498 * detected and all of the possible cases can be divided into two distinct
499 * groups.
500 *
501 * First, a wakeup event may be detected by the same functional unit that will
502 * carry out the entire processing of it and possibly will pass it to user space
503 * for further processing. In that case the functional unit that has detected
504 * the event may later "close" the "no suspend" period associated with it
505 * directly as soon as it has been dealt with. The pair of pm_stay_awake() and
506 * pm_relax(), balanced with each other, is supposed to be used in such
507 * situations.
508 *
509 * Second, a wakeup event may be detected by one functional unit and processed
510 * by another one. In that case the unit that has detected it cannot really
511 * "close" the "no suspend" period associated with it, unless it knows in
512 * advance what's going to happen to the event during processing. This
513 * knowledge, however, may not be available to it, so it can simply specify time
514 * to wait before the system can be suspended and pass it as the second
515 * argument of pm_wakeup_event().
516 *
517 * It is valid to call pm_relax() after pm_wakeup_event(), in which case the
518 * "no suspend" period will be ended either by the pm_relax(), or by the timer
519 * function executed when the timer expires, whichever comes first.
520 */
521
522 /**
523 * wakup_source_activate - Mark given wakeup source as active.
524 * @ws: Wakeup source to handle.
525 *
526 * Update the @ws' statistics and, if @ws has just been activated, notify the PM
527 * core of the event by incrementing the counter of of wakeup events being
528 * processed.
529 */
530 static void wakeup_source_activate(struct wakeup_source *ws)
531 {
532 unsigned int cec;
533
534 if (WARN_ONCE(wakeup_source_not_registered(ws),
535 "unregistered wakeup source\n"))
536 return;
537
538 /*
539 * active wakeup source should bring the system
540 * out of PM_SUSPEND_FREEZE state
541 */
542 freeze_wake();
543
544 ws->active = true;
545 ws->active_count++;
546 ws->last_time = ktime_get();
547 if (ws->autosleep_enabled)
548 ws->start_prevent_time = ws->last_time;
549
550 /* Increment the counter of events in progress. */
551 cec = atomic_inc_return(&combined_event_count);
552
553 trace_wakeup_source_activate(ws->name, cec);
554 }
555
556 /**
557 * wakeup_source_report_event - Report wakeup event using the given source.
558 * @ws: Wakeup source to report the event for.
559 */
560 static void wakeup_source_report_event(struct wakeup_source *ws)
561 {
562 ws->event_count++;
563 /* This is racy, but the counter is approximate anyway. */
564 if (events_check_enabled)
565 ws->wakeup_count++;
566
567 if (!ws->active)
568 wakeup_source_activate(ws);
569 }
570
571 /**
572 * __pm_stay_awake - Notify the PM core of a wakeup event.
573 * @ws: Wakeup source object associated with the source of the event.
574 *
575 * It is safe to call this function from interrupt context.
576 */
577 void __pm_stay_awake(struct wakeup_source *ws)
578 {
579 unsigned long flags;
580
581 if (!ws)
582 return;
583
584 spin_lock_irqsave(&ws->lock, flags);
585
586 wakeup_source_report_event(ws);
587 del_timer(&ws->timer);
588 ws->timer_expires = 0;
589
590 spin_unlock_irqrestore(&ws->lock, flags);
591 }
592 EXPORT_SYMBOL_GPL(__pm_stay_awake);
593
594 /**
595 * pm_stay_awake - Notify the PM core that a wakeup event is being processed.
596 * @dev: Device the wakeup event is related to.
597 *
598 * Notify the PM core of a wakeup event (signaled by @dev) by calling
599 * __pm_stay_awake for the @dev's wakeup source object.
600 *
601 * Call this function after detecting of a wakeup event if pm_relax() is going
602 * to be called directly after processing the event (and possibly passing it to
603 * user space for further processing).
604 */
605 void pm_stay_awake(struct device *dev)
606 {
607 unsigned long flags;
608
609 if (!dev)
610 return;
611
612 spin_lock_irqsave(&dev->power.lock, flags);
613 __pm_stay_awake(dev->power.wakeup);
614 spin_unlock_irqrestore(&dev->power.lock, flags);
615 }
616 EXPORT_SYMBOL_GPL(pm_stay_awake);
617
618 #ifdef CONFIG_PM_AUTOSLEEP
619 static void update_prevent_sleep_time(struct wakeup_source *ws, ktime_t now)
620 {
621 ktime_t delta = ktime_sub(now, ws->start_prevent_time);
622 ws->prevent_sleep_time = ktime_add(ws->prevent_sleep_time, delta);
623 }
624 #else
625 static inline void update_prevent_sleep_time(struct wakeup_source *ws,
626 ktime_t now) {}
627 #endif
628
629 /**
630 * wakup_source_deactivate - Mark given wakeup source as inactive.
631 * @ws: Wakeup source to handle.
632 *
633 * Update the @ws' statistics and notify the PM core that the wakeup source has
634 * become inactive by decrementing the counter of wakeup events being processed
635 * and incrementing the counter of registered wakeup events.
636 */
637 static void wakeup_source_deactivate(struct wakeup_source *ws)
638 {
639 unsigned int cnt, inpr, cec;
640 ktime_t duration;
641 ktime_t now;
642
643 ws->relax_count++;
644 /*
645 * __pm_relax() may be called directly or from a timer function.
646 * If it is called directly right after the timer function has been
647 * started, but before the timer function calls __pm_relax(), it is
648 * possible that __pm_stay_awake() will be called in the meantime and
649 * will set ws->active. Then, ws->active may be cleared immediately
650 * by the __pm_relax() called from the timer function, but in such a
651 * case ws->relax_count will be different from ws->active_count.
652 */
653 if (ws->relax_count != ws->active_count) {
654 ws->relax_count--;
655 return;
656 }
657
658 ws->active = false;
659
660 now = ktime_get();
661 duration = ktime_sub(now, ws->last_time);
662 ws->total_time = ktime_add(ws->total_time, duration);
663 if (ktime_to_ns(duration) > ktime_to_ns(ws->max_time))
664 ws->max_time = duration;
665
666 ws->last_time = now;
667 del_timer(&ws->timer);
668 ws->timer_expires = 0;
669
670 if (ws->autosleep_enabled)
671 update_prevent_sleep_time(ws, now);
672
673 /*
674 * Increment the counter of registered wakeup events and decrement the
675 * couter of wakeup events in progress simultaneously.
676 */
677 cec = atomic_add_return(MAX_IN_PROGRESS, &combined_event_count);
678 trace_wakeup_source_deactivate(ws->name, cec);
679
680 split_counters(&cnt, &inpr);
681 if (!inpr && waitqueue_active(&wakeup_count_wait_queue))
682 wake_up(&wakeup_count_wait_queue);
683 }
684
685 /**
686 * __pm_relax - Notify the PM core that processing of a wakeup event has ended.
687 * @ws: Wakeup source object associated with the source of the event.
688 *
689 * Call this function for wakeup events whose processing started with calling
690 * __pm_stay_awake().
691 *
692 * It is safe to call it from interrupt context.
693 */
694 void __pm_relax(struct wakeup_source *ws)
695 {
696 unsigned long flags;
697
698 if (!ws)
699 return;
700
701 spin_lock_irqsave(&ws->lock, flags);
702 if (ws->active)
703 wakeup_source_deactivate(ws);
704 spin_unlock_irqrestore(&ws->lock, flags);
705 }
706 EXPORT_SYMBOL_GPL(__pm_relax);
707
708 /**
709 * pm_relax - Notify the PM core that processing of a wakeup event has ended.
710 * @dev: Device that signaled the event.
711 *
712 * Execute __pm_relax() for the @dev's wakeup source object.
713 */
714 void pm_relax(struct device *dev)
715 {
716 unsigned long flags;
717
718 if (!dev)
719 return;
720
721 spin_lock_irqsave(&dev->power.lock, flags);
722 __pm_relax(dev->power.wakeup);
723 spin_unlock_irqrestore(&dev->power.lock, flags);
724 }
725 EXPORT_SYMBOL_GPL(pm_relax);
726
727 /**
728 * pm_wakeup_timer_fn - Delayed finalization of a wakeup event.
729 * @data: Address of the wakeup source object associated with the event source.
730 *
731 * Call wakeup_source_deactivate() for the wakeup source whose address is stored
732 * in @data if it is currently active and its timer has not been canceled and
733 * the expiration time of the timer is not in future.
734 */
735 static void pm_wakeup_timer_fn(unsigned long data)
736 {
737 struct wakeup_source *ws = (struct wakeup_source *)data;
738 unsigned long flags;
739
740 spin_lock_irqsave(&ws->lock, flags);
741
742 if (ws->active && ws->timer_expires
743 && time_after_eq(jiffies, ws->timer_expires)) {
744 wakeup_source_deactivate(ws);
745 ws->expire_count++;
746 }
747
748 spin_unlock_irqrestore(&ws->lock, flags);
749 }
750
751 /**
752 * __pm_wakeup_event - Notify the PM core of a wakeup event.
753 * @ws: Wakeup source object associated with the event source.
754 * @msec: Anticipated event processing time (in milliseconds).
755 *
756 * Notify the PM core of a wakeup event whose source is @ws that will take
757 * approximately @msec milliseconds to be processed by the kernel. If @ws is
758 * not active, activate it. If @msec is nonzero, set up the @ws' timer to
759 * execute pm_wakeup_timer_fn() in future.
760 *
761 * It is safe to call this function from interrupt context.
762 */
763 void __pm_wakeup_event(struct wakeup_source *ws, unsigned int msec)
764 {
765 unsigned long flags;
766 unsigned long expires;
767
768 if (!ws)
769 return;
770
771 spin_lock_irqsave(&ws->lock, flags);
772
773 wakeup_source_report_event(ws);
774
775 if (!msec) {
776 wakeup_source_deactivate(ws);
777 goto unlock;
778 }
779
780 expires = jiffies + msecs_to_jiffies(msec);
781 if (!expires)
782 expires = 1;
783
784 if (!ws->timer_expires || time_after(expires, ws->timer_expires)) {
785 mod_timer(&ws->timer, expires);
786 ws->timer_expires = expires;
787 }
788
789 unlock:
790 spin_unlock_irqrestore(&ws->lock, flags);
791 }
792 EXPORT_SYMBOL_GPL(__pm_wakeup_event);
793
794
795 /**
796 * pm_wakeup_event - Notify the PM core of a wakeup event.
797 * @dev: Device the wakeup event is related to.
798 * @msec: Anticipated event processing time (in milliseconds).
799 *
800 * Call __pm_wakeup_event() for the @dev's wakeup source object.
801 */
802 void pm_wakeup_event(struct device *dev, unsigned int msec)
803 {
804 unsigned long flags;
805
806 if (!dev)
807 return;
808
809 spin_lock_irqsave(&dev->power.lock, flags);
810 __pm_wakeup_event(dev->power.wakeup, msec);
811 spin_unlock_irqrestore(&dev->power.lock, flags);
812 }
813 EXPORT_SYMBOL_GPL(pm_wakeup_event);
814
815 void pm_print_active_wakeup_sources(void)
816 {
817 struct wakeup_source *ws;
818 int active = 0;
819 struct wakeup_source *last_activity_ws = NULL;
820
821 rcu_read_lock();
822 list_for_each_entry_rcu(ws, &wakeup_sources, entry) {
823 if (ws->active) {
824 pr_info("active wakeup source: %s\n", ws->name);
825 active = 1;
826 } else if (!active &&
827 (!last_activity_ws ||
828 ktime_to_ns(ws->last_time) >
829 ktime_to_ns(last_activity_ws->last_time))) {
830 last_activity_ws = ws;
831 }
832 }
833
834 if (!active && last_activity_ws)
835 pr_info("last active wakeup source: %s\n",
836 last_activity_ws->name);
837 rcu_read_unlock();
838 }
839 EXPORT_SYMBOL_GPL(pm_print_active_wakeup_sources);
840
841 /**
842 * pm_wakeup_pending - Check if power transition in progress should be aborted.
843 *
844 * Compare the current number of registered wakeup events with its preserved
845 * value from the past and return true if new wakeup events have been registered
846 * since the old value was stored. Also return true if the current number of
847 * wakeup events being processed is different from zero.
848 */
849 bool pm_wakeup_pending(void)
850 {
851 unsigned long flags;
852 bool ret = false;
853
854 spin_lock_irqsave(&events_lock, flags);
855 if (events_check_enabled) {
856 unsigned int cnt, inpr;
857
858 split_counters(&cnt, &inpr);
859 ret = (cnt != saved_count || inpr > 0);
860 events_check_enabled = !ret;
861 }
862 spin_unlock_irqrestore(&events_lock, flags);
863
864 if (ret) {
865 pr_info("PM: Wakeup pending, aborting suspend\n");
866 pm_print_active_wakeup_sources();
867 }
868
869 return ret || pm_abort_suspend;
870 }
871
872 void pm_system_wakeup(void)
873 {
874 pm_abort_suspend = true;
875 freeze_wake();
876 }
877 EXPORT_SYMBOL_GPL(pm_system_wakeup);
878
879 void pm_wakeup_clear(void)
880 {
881 pm_abort_suspend = false;
882 }
883
884 /**
885 * pm_get_wakeup_count - Read the number of registered wakeup events.
886 * @count: Address to store the value at.
887 * @block: Whether or not to block.
888 *
889 * Store the number of registered wakeup events at the address in @count. If
890 * @block is set, block until the current number of wakeup events being
891 * processed is zero.
892 *
893 * Return 'false' if the current number of wakeup events being processed is
894 * nonzero. Otherwise return 'true'.
895 */
896 bool pm_get_wakeup_count(unsigned int *count, bool block)
897 {
898 unsigned int cnt, inpr;
899
900 if (block) {
901 DEFINE_WAIT(wait);
902
903 for (;;) {
904 prepare_to_wait(&wakeup_count_wait_queue, &wait,
905 TASK_INTERRUPTIBLE);
906 split_counters(&cnt, &inpr);
907 if (inpr == 0 || signal_pending(current))
908 break;
909
910 schedule();
911 }
912 finish_wait(&wakeup_count_wait_queue, &wait);
913 }
914
915 split_counters(&cnt, &inpr);
916 *count = cnt;
917 return !inpr;
918 }
919
920 /**
921 * pm_save_wakeup_count - Save the current number of registered wakeup events.
922 * @count: Value to compare with the current number of registered wakeup events.
923 *
924 * If @count is equal to the current number of registered wakeup events and the
925 * current number of wakeup events being processed is zero, store @count as the
926 * old number of registered wakeup events for pm_check_wakeup_events(), enable
927 * wakeup events detection and return 'true'. Otherwise disable wakeup events
928 * detection and return 'false'.
929 */
930 bool pm_save_wakeup_count(unsigned int count)
931 {
932 unsigned int cnt, inpr;
933 unsigned long flags;
934
935 events_check_enabled = false;
936 spin_lock_irqsave(&events_lock, flags);
937 split_counters(&cnt, &inpr);
938 if (cnt == count && inpr == 0) {
939 saved_count = count;
940 events_check_enabled = true;
941 }
942 spin_unlock_irqrestore(&events_lock, flags);
943 return events_check_enabled;
944 }
945
946 #ifdef CONFIG_PM_AUTOSLEEP
947 /**
948 * pm_wakep_autosleep_enabled - Modify autosleep_enabled for all wakeup sources.
949 * @enabled: Whether to set or to clear the autosleep_enabled flags.
950 */
951 void pm_wakep_autosleep_enabled(bool set)
952 {
953 struct wakeup_source *ws;
954 ktime_t now = ktime_get();
955
956 rcu_read_lock();
957 list_for_each_entry_rcu(ws, &wakeup_sources, entry) {
958 spin_lock_irq(&ws->lock);
959 if (ws->autosleep_enabled != set) {
960 ws->autosleep_enabled = set;
961 if (ws->active) {
962 if (set)
963 ws->start_prevent_time = now;
964 else
965 update_prevent_sleep_time(ws, now);
966 }
967 }
968 spin_unlock_irq(&ws->lock);
969 }
970 rcu_read_unlock();
971 }
972 #endif /* CONFIG_PM_AUTOSLEEP */
973
974 static struct dentry *wakeup_sources_stats_dentry;
975
976 /**
977 * print_wakeup_source_stats - Print wakeup source statistics information.
978 * @m: seq_file to print the statistics into.
979 * @ws: Wakeup source object to print the statistics for.
980 */
981 static int print_wakeup_source_stats(struct seq_file *m,
982 struct wakeup_source *ws)
983 {
984 unsigned long flags;
985 ktime_t total_time;
986 ktime_t max_time;
987 unsigned long active_count;
988 ktime_t active_time;
989 ktime_t prevent_sleep_time;
990
991 spin_lock_irqsave(&ws->lock, flags);
992
993 total_time = ws->total_time;
994 max_time = ws->max_time;
995 prevent_sleep_time = ws->prevent_sleep_time;
996 active_count = ws->active_count;
997 if (ws->active) {
998 ktime_t now = ktime_get();
999
1000 active_time = ktime_sub(now, ws->last_time);
1001 total_time = ktime_add(total_time, active_time);
1002 if (active_time.tv64 > max_time.tv64)
1003 max_time = active_time;
1004
1005 if (ws->autosleep_enabled)
1006 prevent_sleep_time = ktime_add(prevent_sleep_time,
1007 ktime_sub(now, ws->start_prevent_time));
1008 } else {
1009 active_time = ktime_set(0, 0);
1010 }
1011
1012 seq_printf(m, "%-12s\t%lu\t\t%lu\t\t%lu\t\t%lu\t\t%lld\t\t%lld\t\t%lld\t\t%lld\t\t%lld\n",
1013 ws->name, active_count, ws->event_count,
1014 ws->wakeup_count, ws->expire_count,
1015 ktime_to_ms(active_time), ktime_to_ms(total_time),
1016 ktime_to_ms(max_time), ktime_to_ms(ws->last_time),
1017 ktime_to_ms(prevent_sleep_time));
1018
1019 spin_unlock_irqrestore(&ws->lock, flags);
1020
1021 return 0;
1022 }
1023
1024 /**
1025 * wakeup_sources_stats_show - Print wakeup sources statistics information.
1026 * @m: seq_file to print the statistics into.
1027 */
1028 static int wakeup_sources_stats_show(struct seq_file *m, void *unused)
1029 {
1030 struct wakeup_source *ws;
1031
1032 seq_puts(m, "name\t\tactive_count\tevent_count\twakeup_count\t"
1033 "expire_count\tactive_since\ttotal_time\tmax_time\t"
1034 "last_change\tprevent_suspend_time\n");
1035
1036 rcu_read_lock();
1037 list_for_each_entry_rcu(ws, &wakeup_sources, entry)
1038 print_wakeup_source_stats(m, ws);
1039 rcu_read_unlock();
1040
1041 print_wakeup_source_stats(m, &deleted_ws);
1042
1043 return 0;
1044 }
1045
1046 static int wakeup_sources_stats_open(struct inode *inode, struct file *file)
1047 {
1048 return single_open(file, wakeup_sources_stats_show, NULL);
1049 }
1050
1051 static const struct file_operations wakeup_sources_stats_fops = {
1052 .owner = THIS_MODULE,
1053 .open = wakeup_sources_stats_open,
1054 .read = seq_read,
1055 .llseek = seq_lseek,
1056 .release = single_release,
1057 };
1058
1059 static int __init wakeup_sources_debugfs_init(void)
1060 {
1061 wakeup_sources_stats_dentry = debugfs_create_file("wakeup_sources",
1062 S_IRUGO, NULL, NULL, &wakeup_sources_stats_fops);
1063 return 0;
1064 }
1065
1066 postcore_initcall(wakeup_sources_debugfs_init);