/* * kernel/power/wakelock.c * * User space wakeup sources support. * * Copyright (C) 2012 Rafael J. Wysocki * * This code is based on the analogous interface allowing user space to * manipulate wakelocks on Android. */ #include #include #include #include #include #include #include #include #include #include "power.h" static DEFINE_MUTEX(wakelocks_lock); struct wakelock { char *name; struct rb_node node; struct wakeup_source ws; #ifdef CONFIG_PM_WAKELOCKS_GC struct list_head lru; #endif }; static struct rb_root wakelocks_tree = RB_ROOT; ssize_t pm_show_wakelocks(char *buf, bool show_active) { struct rb_node *node; struct wakelock *wl; char *str = buf; char *end = buf + PAGE_SIZE; mutex_lock(&wakelocks_lock); for (node = rb_first(&wakelocks_tree); node; node = rb_next(node)) { wl = rb_entry(node, struct wakelock, node); if (wl->ws.active == show_active) str += scnprintf(str, end - str, "%s ", wl->name); } if (str > buf) str--; str += scnprintf(str, end - str, "\n"); mutex_unlock(&wakelocks_lock); return (str - buf); } #if CONFIG_PM_WAKELOCKS_LIMIT > 0 static unsigned int number_of_wakelocks; static inline bool wakelocks_limit_exceeded(void) { return number_of_wakelocks > CONFIG_PM_WAKELOCKS_LIMIT; } static inline void increment_wakelocks_number(void) { number_of_wakelocks++; } static inline void decrement_wakelocks_number(void) { number_of_wakelocks--; } #else /* CONFIG_PM_WAKELOCKS_LIMIT = 0 */ static inline bool wakelocks_limit_exceeded(void) { return false; } static inline void increment_wakelocks_number(void) {} static inline void decrement_wakelocks_number(void) {} #endif /* CONFIG_PM_WAKELOCKS_LIMIT */ #ifdef CONFIG_PM_WAKELOCKS_GC #define WL_GC_COUNT_MAX 100 #define WL_GC_TIME_SEC 300 static void __wakelocks_gc(struct work_struct *work); static LIST_HEAD(wakelocks_lru_list); static DECLARE_WORK(wakelock_work, __wakelocks_gc); static unsigned int wakelocks_gc_count; static inline void wakelocks_lru_add(struct wakelock *wl) { list_add(&wl->lru, &wakelocks_lru_list); } static inline void wakelocks_lru_most_recent(struct wakelock *wl) { list_move(&wl->lru, &wakelocks_lru_list); } static void __wakelocks_gc(struct work_struct *work) { struct wakelock *wl, *aux; ktime_t now; mutex_lock(&wakelocks_lock); now = ktime_get(); list_for_each_entry_safe_reverse(wl, aux, &wakelocks_lru_list, lru) { u64 idle_time_ns; bool active; spin_lock_irq(&wl->ws.lock); idle_time_ns = ktime_to_ns(ktime_sub(now, wl->ws.last_time)); active = wl->ws.active; spin_unlock_irq(&wl->ws.lock); if (idle_time_ns < ((u64)WL_GC_TIME_SEC * NSEC_PER_SEC)) break; if (!active) { wakeup_source_remove(&wl->ws); rb_erase(&wl->node, &wakelocks_tree); list_del(&wl->lru); kfree(wl->name); kfree(wl); decrement_wakelocks_number(); } } wakelocks_gc_count = 0; mutex_unlock(&wakelocks_lock); } static void wakelocks_gc(void) { if (++wakelocks_gc_count <= WL_GC_COUNT_MAX) return; schedule_work(&wakelock_work); } #else /* !CONFIG_PM_WAKELOCKS_GC */ static inline void wakelocks_lru_add(struct wakelock *wl) {} static inline void wakelocks_lru_most_recent(struct wakelock *wl) {} static inline void wakelocks_gc(void) {} #endif /* !CONFIG_PM_WAKELOCKS_GC */ static struct wakelock *wakelock_lookup_add(const char *name, size_t len, bool add_if_not_found) { struct rb_node **node = &wakelocks_tree.rb_node; struct rb_node *parent = *node; struct wakelock *wl; while (*node) { int diff; parent = *node; wl = rb_entry(*node, struct wakelock, node); diff = strncmp(name, wl->name, len); if (diff == 0) { if (wl->name[len]) diff = -1; else return wl; } if (diff < 0) node = &(*node)->rb_left; else node = &(*node)->rb_right; } if (!add_if_not_found) return ERR_PTR(-EINVAL); if (wakelocks_limit_exceeded()) return ERR_PTR(-ENOSPC); /* Not found, we have to add a new one. */ wl = kzalloc(sizeof(*wl), GFP_KERNEL); if (!wl) return ERR_PTR(-ENOMEM); wl->name = kstrndup(name, len, GFP_KERNEL); if (!wl->name) { kfree(wl); return ERR_PTR(-ENOMEM); } wl->ws.name = wl->name; wakeup_source_add(&wl->ws); rb_link_node(&wl->node, parent, node); rb_insert_color(&wl->node, &wakelocks_tree); wakelocks_lru_add(wl); increment_wakelocks_number(); return wl; } int pm_wake_lock(const char *buf) { const char *str = buf; struct wakelock *wl; u64 timeout_ns = 0; size_t len; int ret = 0; if (!capable(CAP_BLOCK_SUSPEND)) return -EPERM; while (*str && !isspace(*str)) str++; len = str - buf; if (!len) return -EINVAL; if (*str && *str != '\n') { /* Find out if there's a valid timeout string appended. */ ret = kstrtou64(skip_spaces(str), 10, &timeout_ns); if (ret) return -EINVAL; } mutex_lock(&wakelocks_lock); wl = wakelock_lookup_add(buf, len, true); if (IS_ERR(wl)) { ret = PTR_ERR(wl); goto out; } if (timeout_ns) { u64 timeout_ms = timeout_ns + NSEC_PER_MSEC - 1; do_div(timeout_ms, NSEC_PER_MSEC); __pm_wakeup_event(&wl->ws, timeout_ms); } else { __pm_stay_awake(&wl->ws); } wakelocks_lru_most_recent(wl); out: mutex_unlock(&wakelocks_lock); return ret; } int pm_wake_unlock(const char *buf) { struct wakelock *wl; size_t len; int ret = 0; if (!capable(CAP_BLOCK_SUSPEND)) return -EPERM; len = strlen(buf); if (!len) return -EINVAL; if (buf[len-1] == '\n') len--; if (!len) return -EINVAL; mutex_lock(&wakelocks_lock); wl = wakelock_lookup_add(buf, len, false); if (IS_ERR(wl)) { ret = PTR_ERR(wl); goto out; } __pm_relax(&wl->ws); wakelocks_lru_most_recent(wl); wakelocks_gc(); out: mutex_unlock(&wakelocks_lock); return ret; } next.git/tree/?h=nds-private-remove&id=3ddc76dfc786cc6f87852693227fb0b1f124f807'>8192b4721e05cf6823087f9696db8c0c8f144b02 /include/soc/tegra/bpmp-abi.h parentb272f732f888d4cf43c943a40c9aaa836f9b7431 (diff)parent1f3a8e49d8f28f498b8694464623ac20aebfe62a (diff)
Merge branch 'timers-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip
Pull timer type cleanups from Thomas Gleixner: "This series does a tree wide cleanup of types related to timers/timekeeping. - Get rid of cycles_t and use a plain u64. The type is not really helpful and caused more confusion than clarity - Get rid of the ktime union. The union has become useless as we use the scalar nanoseconds storage unconditionally now. The 32bit timespec alike storage got removed due to the Y2038 limitations some time ago. That leaves the odd union access around for no reason. Clean it up. Both changes have been done with coccinelle and a small amount of manual mopping up" * 'timers-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip: ktime: Get rid of ktime_equal() ktime: Cleanup ktime_set() usage ktime: Get rid of the union clocksource: Use a plain u64 instead of cycle_t
Diffstat (limited to 'include/soc/tegra/bpmp-abi.h')