/* * This module exposes the interface to kernel space for specifying * QoS dependencies. It provides infrastructure for registration of: * * Dependents on a QoS value : register requirements * Watchers of QoS value : get notified when target QoS value changes * * This QoS design is best effort based. Dependents register their QoS needs. * Watchers register to keep track of the current QoS needs of the system. * * There are 3 basic classes of QoS parameter: latency, timeout, throughput * each have defined units: * latency: usec * timeout: usec <-- currently not used. * throughput: kbs (kilo byte / sec) * * There are lists of pm_qos_objects each one wrapping requirements, notifiers * * User mode requirements on a QOS parameter register themselves to the * subsystem by opening the device node /dev/... and writing there request to * the node. As long as the process holds a file handle open to the node the * client continues to be accounted for. Upon file release the usermode * requirement is removed and a new qos target is computed. This way when the * requirement that the application has is cleaned up when closes the file * pointer or exits the pm_qos_object will get an opportunity to clean up. * * Mark Gross * * Copyright (c) 2009, Code Aurora Forum. All rights reserved. */ #include #include #include #include #include #include #include #include #include #include #include #include #include /* * locking rule: all changes to requirements or notifiers lists * or pm_qos_object list and pm_qos_objects need to happen with pm_qos_lock * held, taken with _irqsave. One lock to rule them all */ static s32 max_compare(s32 v1, s32 v2); static s32 min_compare(s32 v1, s32 v2); struct pm_qos_power_user { int pm_qos_class; char name[sizeof("user_01234567")]; }; static BLOCKING_NOTIFIER_HEAD(cpu_dma_lat_notifier); static struct pm_qos_object cpu_dma_pm_qos = { .requirements = {LIST_HEAD_INIT(cpu_dma_pm_qos.requirements.list)}, .notifiers = &cpu_dma_lat_notifier, .name = "cpu_dma_latency", .default_value = 2000 * USEC_PER_SEC, .target_value = ATOMIC_INIT(2000 * USEC_PER_SEC), .comparitor = min_compare }; static BLOCKING_NOTIFIER_HEAD(network_lat_notifier); static struct pm_qos_object network_lat_pm_qos = { .requirements = {LIST_HEAD_INIT(network_lat_pm_qos.requirements.list)}, .notifiers = &network_lat_notifier, .name = "network_latency", .default_value = 2000 * USEC_PER_SEC, .target_value = ATOMIC_INIT(2000 * USEC_PER_SEC), .comparitor = min_compare }; static BLOCKING_NOTIFIER_HEAD(network_throughput_notifier); static struct pm_qos_object network_throughput_pm_qos = { .requirements = {LIST_HEAD_INIT(network_throughput_pm_qos.requirements.list)}, .notifiers = &network_throughput_notifier, .name = "network_throughput", .default_value = 0, .target_value = ATOMIC_INIT(0), .comparitor = max_compare }; static BLOCKING_NOTIFIER_HEAD(system_bus_freq_notifier); static struct pm_qos_object system_bus_freq_pm_qos = { .requirements = {LIST_HEAD_INIT(system_bus_freq_pm_qos.requirements.list)}, .notifiers = &system_bus_freq_notifier, .name = "system_bus_freq", .default_value = 0, .target_value = ATOMIC_INIT(0), .comparitor = max_compare }; static struct pm_qos_object *pm_qos_array[PM_QOS_NUM_CLASSES] = { [PM_QOS_RESERVED] = NULL, [PM_QOS_CPU_DMA_LATENCY] = &cpu_dma_pm_qos, [PM_QOS_NETWORK_LATENCY] = &network_lat_pm_qos, [PM_QOS_NETWORK_THROUGHPUT] = &network_throughput_pm_qos, [PM_QOS_SYSTEM_BUS_FREQ] = &system_bus_freq_pm_qos, }; static DEFINE_SPINLOCK(pm_qos_lock); static atomic_t pm_qos_user_id = ATOMIC_INIT(0); static ssize_t pm_qos_power_write(struct file *filp, const char __user *buf, size_t count, loff_t *f_pos); static int pm_qos_power_open(struct inode *inode, struct file *filp); static int pm_qos_power_release(struct inode *inode, struct file *filp); static const struct file_operations pm_qos_power_fops = { .write = pm_qos_power_write, .open = pm_qos_power_open, .release = pm_qos_power_release, }; /* static helper functions */ static s32 max_compare(s32 v1, s32 v2) { return max(v1, v2); } static s32 min_compare(s32 v1, s32 v2) { return min(v1, v2); } static void update_target(int pm_qos_class) { s32 extreme_value; struct requirement_list *node; unsigned long flags; int call_notifier = 0; spin_lock_irqsave(&pm_qos_lock, flags); extreme_value = pm_qos_array[pm_qos_class]->default_value; list_for_each_entry(node, &pm_qos_array[pm_qos_class]->requirements.list, list) { extreme_value = pm_qos_array[pm_qos_class]->comparitor( extreme_value, node->value); } if (atomic_read(&pm_qos_array[pm_qos_class]->target_value) != extreme_value) { call_notifier = 1; atomic_set(&pm_qos_array[pm_qos_class]->target_value, extreme_value); pr_debug(KERN_ERR "new target for qos %d is %d\n", pm_qos_class, atomic_read( &pm_qos_array[pm_qos_class]->target_value)); } spin_unlock_irqrestore(&pm_qos_lock, flags); if (call_notifier) blocking_notifier_call_chain( pm_qos_array[pm_qos_class]->notifiers, (unsigned long) extreme_value, NULL); } static int register_pm_qos_misc(struct pm_qos_object *qos) { qos->pm_qos_power_miscdev.minor = MISC_DYNAMIC_MINOR; qos->pm_qos_power_miscdev.name = qos->name; qos->pm_qos_power_miscdev.fops = &pm_qos_power_fops; return misc_register(&qos->pm_qos_power_miscdev); } static int find_pm_qos_object_by_minor(int minor) { int pm_qos_class; for (pm_qos_class = 0; pm_qos_class < PM_QOS_NUM_CLASSES; pm_qos_class++) { if (!pm_qos_array[pm_qos_class]) continue; if (minor == pm_qos_array[pm_qos_class]->pm_qos_power_miscdev.minor) return pm_qos_class; } return -1; } /** * pm_qos_requirement - returns current system wide qos expectation * @pm_qos_class: identification of which qos value is requested * * This function returns the current target value in an atomic manner. */ int pm_qos_requirement(int pm_qos_class) { return atomic_read(&pm_qos_array[pm_qos_class]->target_value); } EXPORT_SYMBOL_GPL(pm_qos_requirement); /** * pm_qos_add_requirement - inserts new qos request into the list * @pm_qos_class: identifies which list of qos request to us * @name: identifies the request * @value: defines the qos request * * This function inserts a new entry in the pm_qos_class list of requested qos * performance characteristics. It recomputes the aggregate QoS expectations * for the pm_qos_class of parameters. */ int pm_qos_add_requirement(int pm_qos_class, char *name, s32 value) { struct requirement_list *dep; unsigned long flags; dep = kzalloc(sizeof(struct requirement_list), GFP_KERNEL); if (dep) { if (value == PM_QOS_DEFAULT_VALUE) dep->value = pm_qos_array[pm_qos_class]->default_value; else dep->value = value; dep->name = kstrdup(name, GFP_KERNEL); if (!dep->name) goto cleanup; spin_lock_irqsave(&pm_qos_lock, flags); list_add(&dep->list, &pm_qos_array[pm_qos_class]->requirements.list); spin_unlock_irqrestore(&pm_qos_lock, flags); update_target(pm_qos_class); return 0; } cleanup: kfree(dep); return -ENOMEM; } EXPORT_SYMBOL_GPL(pm_qos_add_requirement); /** * pm_qos_update_requirement - modifies an existing qos request * @pm_qos_class: identifies which list of qos request to us * @name: identifies the request * @value: defines the qos request * * Updates an existing qos requirement for the pm_qos_class of parameters along * with updating the target pm_qos_class value. * * If the named request isn't in the list then no change is made. */ int pm_qos_update_requirement(int pm_qos_class, char *name, s32 new_value) { unsigned long flags; struct requirement_list *node; int pending_update = 0; spin_lock_irqsave(&pm_qos_lock, flags); list_for_each_entry(node, &pm_qos_array[pm_qos_class]->requirements.list, list) { if (strcmp(node->name, name) == 0) { if (new_value == PM_QOS_DEFAULT_VALUE) node->value = pm_qos_array[pm_qos_class]->default_value; else node->value = new_value; pending_update = 1; break; } } spin_unlock_irqrestore(&pm_qos_lock, flags); if (pending_update) update_target(pm_qos_class); return 0; } EXPORT_SYMBOL_GPL(pm_qos_update_requirement); /** * pm_qos_remove_requirement - modifies an existing qos request * @pm_qos_class: identifies which list of qos request to us * @name: identifies the request * * Will remove named qos request from pm_qos_class list of parameters and * recompute the current target value for the pm_qos_class. */ void pm_qos_remove_requirement(int pm_qos_class, char *name) { unsigned long flags; struct requirement_list *node; int pending_update = 0; spin_lock_irqsave(&pm_qos_lock, flags); list_for_each_entry(node, &pm_qos_array[pm_qos_class]->requirements.list, list) { if (strcmp(node->name, name) == 0) { kfree(node->name); list_del(&node->list); kfree(node); pending_update = 1; break; } } spin_unlock_irqrestore(&pm_qos_lock, flags); if (pending_update) update_target(pm_qos_class); } EXPORT_SYMBOL_GPL(pm_qos_remove_requirement); /** * pm_qos_add_notifier - sets notification entry for changes to target value * @pm_qos_class: identifies which qos target changes should be notified. * @notifier: notifier block managed by caller. * * will register the notifier into a notification chain that gets called * upon changes to the pm_qos_class target value. */ int pm_qos_add_notifier(int pm_qos_class, struct notifier_block *notifier) { int retval; retval = blocking_notifier_chain_register( pm_qos_array[pm_qos_class]->notifiers, notifier); return retval; } EXPORT_SYMBOL_GPL(pm_qos_add_notifier); /** * pm_qos_remove_notifier - deletes notification entry from chain. * @pm_qos_class: identifies which qos target changes are notified. * @notifier: notifier block to be removed. * * will remove the notifier from the notification chain that gets called * upon changes to the pm_qos_class target value. */ int pm_qos_remove_notifier(int pm_qos_class, struct notifier_block *notifier) { int retval; retval = blocking_notifier_chain_unregister( pm_qos_array[pm_qos_class]->notifiers, notifier); return retval; } EXPORT_SYMBOL_GPL(pm_qos_remove_notifier); static int pm_qos_power_open(struct inode *inode, struct file *filp) { int ret; int pm_qos_class; struct pm_qos_power_user *usr; usr = kzalloc(sizeof(struct pm_qos_power_user), GFP_KERNEL); if (!usr) return -ENOMEM; lock_kernel(); pm_qos_class = find_pm_qos_object_by_minor(iminor(inode)); if (pm_qos_class < 0) { unlock_kernel(); kfree(usr); return -EPERM; } usr->pm_qos_class = pm_qos_class; snprintf(usr->name, sizeof(usr->name), "user_%08x", (unsigned)atomic_inc_return(&pm_qos_user_id)); ret = pm_qos_add_requirement(usr->pm_qos_class, usr->name, PM_QOS_DEFAULT_VALUE); unlock_kernel(); if (ret < 0) { kfree(usr); return ret; } filp->private_data = usr; return 0; } static int pm_qos_power_release(struct inode *inode, struct file *filp) { struct pm_qos_power_user *usr; usr = (struct pm_qos_power_user *)filp->private_data; pm_qos_remove_requirement(usr->pm_qos_class, usr->name); filp->private_data = NULL; kfree(usr); return 0; } static ssize_t pm_qos_power_write(struct file *filp, const char __user *buf, size_t count, loff_t *f_pos) { struct pm_qos_power_user *usr; s32 value; usr = (struct pm_qos_power_user *)filp->private_data; if (count != sizeof(s32)) return -EINVAL; if (get_user(value, (s32 *)buf)) return -EFAULT; pm_qos_update_requirement(usr->pm_qos_class, usr->name, value); return sizeof(s32); } static int __init pm_qos_power_init(void) { int ret = 0; ret = register_pm_qos_misc(&cpu_dma_pm_qos); if (ret < 0) { printk(KERN_ERR "pm_qos_param: cpu_dma_latency setup failed\n"); return ret; } ret = register_pm_qos_misc(&network_lat_pm_qos); if (ret < 0) { printk(KERN_ERR "pm_qos_param: network_latency setup failed\n"); return ret; } ret = register_pm_qos_misc(&network_throughput_pm_qos); if (ret < 0) { printk(KERN_ERR "pm_qos_param: network_throughput setup failed\n"); return ret; } ret = register_pm_qos_misc(&system_bus_freq_pm_qos); if (ret < 0) printk(KERN_ERR "pm_qos_param: system_bus_freq setup failed\n"); return ret; } late_initcall(pm_qos_power_init);