2009-08-19 05:38:32 +08:00
|
|
|
/*
|
|
|
|
* pm_runtime.h - Device run-time power management helper functions.
|
|
|
|
*
|
|
|
|
* Copyright (C) 2009 Rafael J. Wysocki <rjw@sisk.pl>
|
|
|
|
*
|
|
|
|
* This file is released under the GPLv2.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef _LINUX_PM_RUNTIME_H
|
|
|
|
#define _LINUX_PM_RUNTIME_H
|
|
|
|
|
|
|
|
#include <linux/device.h>
|
2011-05-27 19:08:41 +08:00
|
|
|
#include <linux/notifier.h>
|
2009-08-19 05:38:32 +08:00
|
|
|
#include <linux/pm.h>
|
|
|
|
|
2010-09-26 05:35:21 +08:00
|
|
|
#include <linux/jiffies.h>
|
|
|
|
|
2010-09-26 05:34:54 +08:00
|
|
|
/* Runtime PM flag argument bits */
|
|
|
|
#define RPM_ASYNC 0x01 /* Request is asynchronous */
|
|
|
|
#define RPM_NOWAIT 0x02 /* Don't wait for concurrent
|
|
|
|
state change */
|
2010-09-26 05:35:07 +08:00
|
|
|
#define RPM_GET_PUT 0x04 /* Increment/decrement the
|
|
|
|
usage_count */
|
2010-09-26 05:35:21 +08:00
|
|
|
#define RPM_AUTO 0x08 /* Use autosuspend_delay */
|
2010-09-26 05:34:54 +08:00
|
|
|
|
2013-12-10 21:37:41 +08:00
|
|
|
#ifdef CONFIG_PM
|
2014-07-23 07:00:36 +08:00
|
|
|
extern struct workqueue_struct *pm_wq;
|
|
|
|
|
|
|
|
static inline bool queue_pm_work(struct work_struct *work)
|
|
|
|
{
|
|
|
|
return queue_work(pm_wq, work);
|
|
|
|
}
|
|
|
|
|
2013-12-10 21:37:41 +08:00
|
|
|
extern int pm_generic_runtime_suspend(struct device *dev);
|
|
|
|
extern int pm_generic_runtime_resume(struct device *dev);
|
2014-03-01 18:56:05 +08:00
|
|
|
extern int pm_runtime_force_suspend(struct device *dev);
|
|
|
|
extern int pm_runtime_force_resume(struct device *dev);
|
2009-08-19 05:38:32 +08:00
|
|
|
|
2010-09-26 05:35:07 +08:00
|
|
|
extern int __pm_runtime_idle(struct device *dev, int rpmflags);
|
|
|
|
extern int __pm_runtime_suspend(struct device *dev, int rpmflags);
|
|
|
|
extern int __pm_runtime_resume(struct device *dev, int rpmflags);
|
2009-08-19 05:38:32 +08:00
|
|
|
extern int pm_schedule_suspend(struct device *dev, unsigned int delay);
|
|
|
|
extern int __pm_runtime_set_status(struct device *dev, unsigned int status);
|
|
|
|
extern int pm_runtime_barrier(struct device *dev);
|
|
|
|
extern void pm_runtime_enable(struct device *dev);
|
|
|
|
extern void __pm_runtime_disable(struct device *dev, bool check_resume);
|
2010-01-24 05:02:51 +08:00
|
|
|
extern void pm_runtime_allow(struct device *dev);
|
|
|
|
extern void pm_runtime_forbid(struct device *dev);
|
2010-09-26 05:35:15 +08:00
|
|
|
extern void pm_runtime_no_callbacks(struct device *dev);
|
2010-12-01 07:14:42 +08:00
|
|
|
extern void pm_runtime_irq_safe(struct device *dev);
|
2010-09-26 05:35:21 +08:00
|
|
|
extern void __pm_runtime_use_autosuspend(struct device *dev, bool use);
|
|
|
|
extern void pm_runtime_set_autosuspend_delay(struct device *dev, int delay);
|
|
|
|
extern unsigned long pm_runtime_autosuspend_expiration(struct device *dev);
|
2011-12-01 07:01:31 +08:00
|
|
|
extern void pm_runtime_update_max_time_suspended(struct device *dev,
|
|
|
|
s64 delta_ns);
|
2013-02-23 08:34:11 +08:00
|
|
|
extern void pm_runtime_set_memalloc_noio(struct device *dev, bool enable);
|
2009-08-19 05:38:32 +08:00
|
|
|
|
|
|
|
static inline bool pm_children_suspended(struct device *dev)
|
|
|
|
{
|
|
|
|
return dev->power.ignore_children
|
|
|
|
|| !atomic_read(&dev->power.child_count);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void pm_runtime_get_noresume(struct device *dev)
|
|
|
|
{
|
|
|
|
atomic_inc(&dev->power.usage_count);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void pm_runtime_put_noidle(struct device *dev)
|
|
|
|
{
|
|
|
|
atomic_add_unless(&dev->power.usage_count, -1, 0);
|
|
|
|
}
|
|
|
|
|
2009-12-04 04:19:18 +08:00
|
|
|
static inline bool device_run_wake(struct device *dev)
|
|
|
|
{
|
|
|
|
return dev->power.run_wake;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void device_set_run_wake(struct device *dev, bool enable)
|
|
|
|
{
|
|
|
|
dev->power.run_wake = enable;
|
|
|
|
}
|
|
|
|
|
2010-03-07 04:28:37 +08:00
|
|
|
static inline bool pm_runtime_suspended(struct device *dev)
|
|
|
|
{
|
2010-12-17 00:11:58 +08:00
|
|
|
return dev->power.runtime_status == RPM_SUSPENDED
|
|
|
|
&& !dev->power.disable_depth;
|
2010-03-07 04:28:37 +08:00
|
|
|
}
|
|
|
|
|
2013-01-24 04:49:37 +08:00
|
|
|
static inline bool pm_runtime_active(struct device *dev)
|
|
|
|
{
|
|
|
|
return dev->power.runtime_status == RPM_ACTIVE
|
|
|
|
|| dev->power.disable_depth;
|
|
|
|
}
|
|
|
|
|
2011-07-12 17:17:09 +08:00
|
|
|
static inline bool pm_runtime_status_suspended(struct device *dev)
|
|
|
|
{
|
|
|
|
return dev->power.runtime_status == RPM_SUSPENDED;
|
|
|
|
}
|
|
|
|
|
PM / sleep: Mechanism to avoid resuming runtime-suspended devices unnecessarily
Currently, some subsystems (e.g. PCI and the ACPI PM domain) have to
resume all runtime-suspended devices during system suspend, mostly
because those devices may need to be reprogrammed due to different
wakeup settings for system sleep and for runtime PM.
For some devices, though, it's OK to remain in runtime suspend
throughout a complete system suspend/resume cycle (if the device was in
runtime suspend at the start of the cycle). We would like to do this
whenever possible, to avoid the overhead of extra power-up and power-down
events.
However, problems may arise because the device's descendants may require
it to be at full power at various points during the cycle. Therefore the
most straightforward way to do this safely is if the device and all its
descendants can remain runtime suspended until the complete stage of
system resume.
To this end, introduce a new device PM flag, power.direct_complete
and modify the PM core to use that flag as follows.
If the ->prepare() callback of a device returns a positive number,
the PM core will regard that as an indication that it may leave the
device runtime-suspended. It will then check if the system power
transition in progress is a suspend (and not hibernation in particular)
and if the device is, indeed, runtime-suspended. In that case, the PM
core will set the device's power.direct_complete flag. Otherwise it
will clear power.direct_complete for the device and it also will later
clear it for the device's parent (if there's one).
Next, the PM core will not invoke the ->suspend() ->suspend_late(),
->suspend_irq(), ->resume_irq(), ->resume_early(), or ->resume()
callbacks for all devices having power.direct_complete set. It
will invoke their ->complete() callbacks, however, and those
callbacks are then responsible for resuming the devices as
appropriate, if necessary. For example, in some cases they may
need to queue up runtime resume requests for the devices using
pm_request_resume().
Changelog partly based on an Alan Stern's description of the idea
(http://marc.info/?l=linux-pm&m=139940466625569&w=2).
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
Acked-by: Alan Stern <stern@rowland.harvard.edu>
2014-05-16 08:46:50 +08:00
|
|
|
static inline bool pm_runtime_suspended_if_enabled(struct device *dev)
|
|
|
|
{
|
|
|
|
return pm_runtime_status_suspended(dev) && dev->power.disable_depth == 1;
|
|
|
|
}
|
|
|
|
|
2010-12-24 22:04:06 +08:00
|
|
|
static inline bool pm_runtime_enabled(struct device *dev)
|
|
|
|
{
|
|
|
|
return !dev->power.disable_depth;
|
|
|
|
}
|
|
|
|
|
2011-02-09 06:26:02 +08:00
|
|
|
static inline bool pm_runtime_callbacks_present(struct device *dev)
|
|
|
|
{
|
|
|
|
return !dev->power.no_callbacks;
|
|
|
|
}
|
|
|
|
|
2010-09-26 05:35:21 +08:00
|
|
|
static inline void pm_runtime_mark_last_busy(struct device *dev)
|
|
|
|
{
|
|
|
|
ACCESS_ONCE(dev->power.last_busy) = jiffies;
|
|
|
|
}
|
|
|
|
|
2014-11-14 16:47:25 +08:00
|
|
|
static inline bool pm_runtime_is_irq_safe(struct device *dev)
|
|
|
|
{
|
|
|
|
return dev->power.irq_safe;
|
|
|
|
}
|
|
|
|
|
2014-11-28 05:38:05 +08:00
|
|
|
#else /* !CONFIG_PM */
|
|
|
|
|
|
|
|
static inline bool queue_pm_work(struct work_struct *work) { return false; }
|
|
|
|
|
|
|
|
static inline int pm_generic_runtime_suspend(struct device *dev) { return 0; }
|
|
|
|
static inline int pm_generic_runtime_resume(struct device *dev) { return 0; }
|
|
|
|
static inline int pm_runtime_force_suspend(struct device *dev) { return 0; }
|
|
|
|
static inline int pm_runtime_force_resume(struct device *dev) { return 0; }
|
2009-08-19 05:38:32 +08:00
|
|
|
|
2010-09-26 05:35:07 +08:00
|
|
|
static inline int __pm_runtime_idle(struct device *dev, int rpmflags)
|
|
|
|
{
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
|
|
|
static inline int __pm_runtime_suspend(struct device *dev, int rpmflags)
|
|
|
|
{
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
|
|
|
static inline int __pm_runtime_resume(struct device *dev, int rpmflags)
|
|
|
|
{
|
|
|
|
return 1;
|
|
|
|
}
|
2009-08-19 05:38:32 +08:00
|
|
|
static inline int pm_schedule_suspend(struct device *dev, unsigned int delay)
|
|
|
|
{
|
|
|
|
return -ENOSYS;
|
|
|
|
}
|
|
|
|
static inline int __pm_runtime_set_status(struct device *dev,
|
|
|
|
unsigned int status) { return 0; }
|
|
|
|
static inline int pm_runtime_barrier(struct device *dev) { return 0; }
|
|
|
|
static inline void pm_runtime_enable(struct device *dev) {}
|
|
|
|
static inline void __pm_runtime_disable(struct device *dev, bool c) {}
|
2010-01-24 05:02:51 +08:00
|
|
|
static inline void pm_runtime_allow(struct device *dev) {}
|
|
|
|
static inline void pm_runtime_forbid(struct device *dev) {}
|
2009-08-19 05:38:32 +08:00
|
|
|
|
|
|
|
static inline bool pm_children_suspended(struct device *dev) { return false; }
|
|
|
|
static inline void pm_runtime_get_noresume(struct device *dev) {}
|
|
|
|
static inline void pm_runtime_put_noidle(struct device *dev) {}
|
2009-12-04 04:19:18 +08:00
|
|
|
static inline bool device_run_wake(struct device *dev) { return false; }
|
|
|
|
static inline void device_set_run_wake(struct device *dev, bool enable) {}
|
2010-03-07 04:28:37 +08:00
|
|
|
static inline bool pm_runtime_suspended(struct device *dev) { return false; }
|
2013-01-24 04:49:37 +08:00
|
|
|
static inline bool pm_runtime_active(struct device *dev) { return true; }
|
2011-07-12 17:17:09 +08:00
|
|
|
static inline bool pm_runtime_status_suspended(struct device *dev) { return false; }
|
PM / sleep: Mechanism to avoid resuming runtime-suspended devices unnecessarily
Currently, some subsystems (e.g. PCI and the ACPI PM domain) have to
resume all runtime-suspended devices during system suspend, mostly
because those devices may need to be reprogrammed due to different
wakeup settings for system sleep and for runtime PM.
For some devices, though, it's OK to remain in runtime suspend
throughout a complete system suspend/resume cycle (if the device was in
runtime suspend at the start of the cycle). We would like to do this
whenever possible, to avoid the overhead of extra power-up and power-down
events.
However, problems may arise because the device's descendants may require
it to be at full power at various points during the cycle. Therefore the
most straightforward way to do this safely is if the device and all its
descendants can remain runtime suspended until the complete stage of
system resume.
To this end, introduce a new device PM flag, power.direct_complete
and modify the PM core to use that flag as follows.
If the ->prepare() callback of a device returns a positive number,
the PM core will regard that as an indication that it may leave the
device runtime-suspended. It will then check if the system power
transition in progress is a suspend (and not hibernation in particular)
and if the device is, indeed, runtime-suspended. In that case, the PM
core will set the device's power.direct_complete flag. Otherwise it
will clear power.direct_complete for the device and it also will later
clear it for the device's parent (if there's one).
Next, the PM core will not invoke the ->suspend() ->suspend_late(),
->suspend_irq(), ->resume_irq(), ->resume_early(), or ->resume()
callbacks for all devices having power.direct_complete set. It
will invoke their ->complete() callbacks, however, and those
callbacks are then responsible for resuming the devices as
appropriate, if necessary. For example, in some cases they may
need to queue up runtime resume requests for the devices using
pm_request_resume().
Changelog partly based on an Alan Stern's description of the idea
(http://marc.info/?l=linux-pm&m=139940466625569&w=2).
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
Acked-by: Alan Stern <stern@rowland.harvard.edu>
2014-05-16 08:46:50 +08:00
|
|
|
static inline bool pm_runtime_suspended_if_enabled(struct device *dev) { return false; }
|
2010-12-24 22:04:06 +08:00
|
|
|
static inline bool pm_runtime_enabled(struct device *dev) { return false; }
|
2009-08-19 05:38:32 +08:00
|
|
|
|
2010-09-26 05:35:15 +08:00
|
|
|
static inline void pm_runtime_no_callbacks(struct device *dev) {}
|
2010-12-01 07:14:42 +08:00
|
|
|
static inline void pm_runtime_irq_safe(struct device *dev) {}
|
2014-11-14 16:47:25 +08:00
|
|
|
static inline bool pm_runtime_is_irq_safe(struct device *dev) { return false; }
|
2010-05-11 05:09:30 +08:00
|
|
|
|
2011-02-09 06:26:02 +08:00
|
|
|
static inline bool pm_runtime_callbacks_present(struct device *dev) { return false; }
|
2010-09-26 05:35:21 +08:00
|
|
|
static inline void pm_runtime_mark_last_busy(struct device *dev) {}
|
|
|
|
static inline void __pm_runtime_use_autosuspend(struct device *dev,
|
|
|
|
bool use) {}
|
|
|
|
static inline void pm_runtime_set_autosuspend_delay(struct device *dev,
|
|
|
|
int delay) {}
|
|
|
|
static inline unsigned long pm_runtime_autosuspend_expiration(
|
|
|
|
struct device *dev) { return 0; }
|
2013-02-23 08:34:11 +08:00
|
|
|
static inline void pm_runtime_set_memalloc_noio(struct device *dev,
|
|
|
|
bool enable){}
|
2010-09-26 05:35:21 +08:00
|
|
|
|
2014-11-28 05:38:05 +08:00
|
|
|
#endif /* !CONFIG_PM */
|
2009-08-19 05:38:32 +08:00
|
|
|
|
2010-09-26 05:35:07 +08:00
|
|
|
static inline int pm_runtime_idle(struct device *dev)
|
|
|
|
{
|
|
|
|
return __pm_runtime_idle(dev, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int pm_runtime_suspend(struct device *dev)
|
|
|
|
{
|
|
|
|
return __pm_runtime_suspend(dev, 0);
|
|
|
|
}
|
|
|
|
|
2010-09-26 05:35:21 +08:00
|
|
|
static inline int pm_runtime_autosuspend(struct device *dev)
|
|
|
|
{
|
|
|
|
return __pm_runtime_suspend(dev, RPM_AUTO);
|
|
|
|
}
|
|
|
|
|
2010-09-26 05:35:07 +08:00
|
|
|
static inline int pm_runtime_resume(struct device *dev)
|
|
|
|
{
|
|
|
|
return __pm_runtime_resume(dev, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int pm_request_idle(struct device *dev)
|
|
|
|
{
|
|
|
|
return __pm_runtime_idle(dev, RPM_ASYNC);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int pm_request_resume(struct device *dev)
|
|
|
|
{
|
|
|
|
return __pm_runtime_resume(dev, RPM_ASYNC);
|
|
|
|
}
|
|
|
|
|
2010-09-29 06:12:22 +08:00
|
|
|
static inline int pm_request_autosuspend(struct device *dev)
|
|
|
|
{
|
|
|
|
return __pm_runtime_suspend(dev, RPM_ASYNC | RPM_AUTO);
|
|
|
|
}
|
|
|
|
|
2009-08-19 05:38:32 +08:00
|
|
|
static inline int pm_runtime_get(struct device *dev)
|
|
|
|
{
|
2010-09-26 05:35:07 +08:00
|
|
|
return __pm_runtime_resume(dev, RPM_GET_PUT | RPM_ASYNC);
|
2009-08-19 05:38:32 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline int pm_runtime_get_sync(struct device *dev)
|
|
|
|
{
|
2010-09-26 05:35:07 +08:00
|
|
|
return __pm_runtime_resume(dev, RPM_GET_PUT);
|
2009-08-19 05:38:32 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline int pm_runtime_put(struct device *dev)
|
|
|
|
{
|
2010-09-26 05:35:07 +08:00
|
|
|
return __pm_runtime_idle(dev, RPM_GET_PUT | RPM_ASYNC);
|
2009-08-19 05:38:32 +08:00
|
|
|
}
|
|
|
|
|
2010-09-26 05:35:21 +08:00
|
|
|
static inline int pm_runtime_put_autosuspend(struct device *dev)
|
|
|
|
{
|
|
|
|
return __pm_runtime_suspend(dev,
|
|
|
|
RPM_GET_PUT | RPM_ASYNC | RPM_AUTO);
|
|
|
|
}
|
|
|
|
|
2009-08-19 05:38:32 +08:00
|
|
|
static inline int pm_runtime_put_sync(struct device *dev)
|
|
|
|
{
|
2010-09-26 05:35:07 +08:00
|
|
|
return __pm_runtime_idle(dev, RPM_GET_PUT);
|
2009-08-19 05:38:32 +08:00
|
|
|
}
|
|
|
|
|
2010-12-01 07:14:42 +08:00
|
|
|
static inline int pm_runtime_put_sync_suspend(struct device *dev)
|
|
|
|
{
|
|
|
|
return __pm_runtime_suspend(dev, RPM_GET_PUT);
|
|
|
|
}
|
|
|
|
|
2010-09-26 05:35:21 +08:00
|
|
|
static inline int pm_runtime_put_sync_autosuspend(struct device *dev)
|
|
|
|
{
|
|
|
|
return __pm_runtime_suspend(dev, RPM_GET_PUT | RPM_AUTO);
|
|
|
|
}
|
|
|
|
|
2009-08-19 05:38:32 +08:00
|
|
|
static inline int pm_runtime_set_active(struct device *dev)
|
|
|
|
{
|
|
|
|
return __pm_runtime_set_status(dev, RPM_ACTIVE);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void pm_runtime_set_suspended(struct device *dev)
|
|
|
|
{
|
|
|
|
__pm_runtime_set_status(dev, RPM_SUSPENDED);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void pm_runtime_disable(struct device *dev)
|
|
|
|
{
|
|
|
|
__pm_runtime_disable(dev, true);
|
|
|
|
}
|
|
|
|
|
2010-09-26 05:35:21 +08:00
|
|
|
static inline void pm_runtime_use_autosuspend(struct device *dev)
|
|
|
|
{
|
|
|
|
__pm_runtime_use_autosuspend(dev, true);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void pm_runtime_dont_use_autosuspend(struct device *dev)
|
|
|
|
{
|
|
|
|
__pm_runtime_use_autosuspend(dev, false);
|
|
|
|
}
|
|
|
|
|
2009-08-19 05:38:32 +08:00
|
|
|
#endif
|