mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-01-18 02:46:06 +00:00
Merge branches 'pm-cpuidle' and 'pm-cpufreq'
* pm-cpuidle: cpuidle: menu: Remove get_loadavg() from the performance multiplier sched: Factor out nr_iowait and nr_iowait_cpu * pm-cpufreq: cpufreq: remove unused arm_big_little_dt driver cpufreq: drop ARM_BIG_LITTLE_CPUFREQ support for ARM64 cpufreq: intel_pstate: Fix compilation for !CONFIG_ACPI
This commit is contained in:
commit
c4ac688993
@ -3838,7 +3838,6 @@ W: http://www.arm.com/products/processors/technologies/biglittleprocessing.php
|
|||||||
S: Maintained
|
S: Maintained
|
||||||
F: drivers/cpufreq/arm_big_little.h
|
F: drivers/cpufreq/arm_big_little.h
|
||||||
F: drivers/cpufreq/arm_big_little.c
|
F: drivers/cpufreq/arm_big_little.c
|
||||||
F: drivers/cpufreq/arm_big_little_dt.c
|
|
||||||
|
|
||||||
CPU POWER MONITORING SUBSYSTEM
|
CPU POWER MONITORING SUBSYSTEM
|
||||||
M: Thomas Renninger <trenn@suse.com>
|
M: Thomas Renninger <trenn@suse.com>
|
||||||
|
@ -28,20 +28,13 @@ config ARM_ARMADA_37XX_CPUFREQ
|
|||||||
# big LITTLE core layer and glue drivers
|
# big LITTLE core layer and glue drivers
|
||||||
config ARM_BIG_LITTLE_CPUFREQ
|
config ARM_BIG_LITTLE_CPUFREQ
|
||||||
tristate "Generic ARM big LITTLE CPUfreq driver"
|
tristate "Generic ARM big LITTLE CPUfreq driver"
|
||||||
depends on (ARM_CPU_TOPOLOGY || ARM64) && HAVE_CLK
|
depends on ARM_CPU_TOPOLOGY && HAVE_CLK
|
||||||
# if CPU_THERMAL is on and THERMAL=m, ARM_BIT_LITTLE_CPUFREQ cannot be =y
|
# if CPU_THERMAL is on and THERMAL=m, ARM_BIT_LITTLE_CPUFREQ cannot be =y
|
||||||
depends on !CPU_THERMAL || THERMAL
|
depends on !CPU_THERMAL || THERMAL
|
||||||
select PM_OPP
|
select PM_OPP
|
||||||
help
|
help
|
||||||
This enables the Generic CPUfreq driver for ARM big.LITTLE platforms.
|
This enables the Generic CPUfreq driver for ARM big.LITTLE platforms.
|
||||||
|
|
||||||
config ARM_DT_BL_CPUFREQ
|
|
||||||
tristate "Generic probing via DT for ARM big LITTLE CPUfreq driver"
|
|
||||||
depends on ARM_BIG_LITTLE_CPUFREQ && OF
|
|
||||||
help
|
|
||||||
This enables probing via DT for Generic CPUfreq driver for ARM
|
|
||||||
big.LITTLE platform. This gets frequency tables from DT.
|
|
||||||
|
|
||||||
config ARM_SCPI_CPUFREQ
|
config ARM_SCPI_CPUFREQ
|
||||||
tristate "SCPI based CPUfreq driver"
|
tristate "SCPI based CPUfreq driver"
|
||||||
depends on ARM_SCPI_PROTOCOL && COMMON_CLK_SCPI
|
depends on ARM_SCPI_PROTOCOL && COMMON_CLK_SCPI
|
||||||
|
@ -48,9 +48,6 @@ obj-$(CONFIG_X86_SFI_CPUFREQ) += sfi-cpufreq.o
|
|||||||
##################################################################################
|
##################################################################################
|
||||||
# ARM SoC drivers
|
# ARM SoC drivers
|
||||||
obj-$(CONFIG_ARM_BIG_LITTLE_CPUFREQ) += arm_big_little.o
|
obj-$(CONFIG_ARM_BIG_LITTLE_CPUFREQ) += arm_big_little.o
|
||||||
# big LITTLE per platform glues. Keep DT_BL_CPUFREQ as the last entry in all big
|
|
||||||
# LITTLE drivers, so that it is probed last.
|
|
||||||
obj-$(CONFIG_ARM_DT_BL_CPUFREQ) += arm_big_little_dt.o
|
|
||||||
|
|
||||||
obj-$(CONFIG_ARM_ARMADA_37XX_CPUFREQ) += armada-37xx-cpufreq.o
|
obj-$(CONFIG_ARM_ARMADA_37XX_CPUFREQ) += armada-37xx-cpufreq.o
|
||||||
obj-$(CONFIG_ARM_BRCMSTB_AVS_CPUFREQ) += brcmstb-avs-cpufreq.o
|
obj-$(CONFIG_ARM_BRCMSTB_AVS_CPUFREQ) += brcmstb-avs-cpufreq.o
|
||||||
|
@ -1,100 +0,0 @@
|
|||||||
/*
|
|
||||||
* Generic big.LITTLE CPUFreq Interface driver
|
|
||||||
*
|
|
||||||
* It provides necessary ops to arm_big_little cpufreq driver and gets
|
|
||||||
* Frequency information from Device Tree. Freq table in DT must be in KHz.
|
|
||||||
*
|
|
||||||
* Copyright (C) 2013 Linaro.
|
|
||||||
* Viresh Kumar <viresh.kumar@linaro.org>
|
|
||||||
*
|
|
||||||
* This program is free software; you can redistribute it and/or modify
|
|
||||||
* it under the terms of the GNU General Public License version 2 as
|
|
||||||
* published by the Free Software Foundation.
|
|
||||||
*
|
|
||||||
* This program is distributed "as is" WITHOUT ANY WARRANTY of any
|
|
||||||
* kind, whether express or implied; without even the implied warranty
|
|
||||||
* of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
||||||
* GNU General Public License for more details.
|
|
||||||
*/
|
|
||||||
|
|
||||||
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
|
|
||||||
|
|
||||||
#include <linux/cpufreq.h>
|
|
||||||
#include <linux/device.h>
|
|
||||||
#include <linux/export.h>
|
|
||||||
#include <linux/module.h>
|
|
||||||
#include <linux/of_device.h>
|
|
||||||
#include <linux/pm_opp.h>
|
|
||||||
#include <linux/platform_device.h>
|
|
||||||
#include <linux/slab.h>
|
|
||||||
#include <linux/types.h>
|
|
||||||
#include "arm_big_little.h"
|
|
||||||
|
|
||||||
/* get cpu node with valid operating-points */
|
|
||||||
static struct device_node *get_cpu_node_with_valid_op(int cpu)
|
|
||||||
{
|
|
||||||
struct device_node *np = of_cpu_device_node_get(cpu);
|
|
||||||
|
|
||||||
if (!of_get_property(np, "operating-points", NULL)) {
|
|
||||||
of_node_put(np);
|
|
||||||
np = NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
return np;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int dt_get_transition_latency(struct device *cpu_dev)
|
|
||||||
{
|
|
||||||
struct device_node *np;
|
|
||||||
u32 transition_latency = CPUFREQ_ETERNAL;
|
|
||||||
|
|
||||||
np = of_node_get(cpu_dev->of_node);
|
|
||||||
if (!np) {
|
|
||||||
pr_info("Failed to find cpu node. Use CPUFREQ_ETERNAL transition latency\n");
|
|
||||||
return CPUFREQ_ETERNAL;
|
|
||||||
}
|
|
||||||
|
|
||||||
of_property_read_u32(np, "clock-latency", &transition_latency);
|
|
||||||
of_node_put(np);
|
|
||||||
|
|
||||||
pr_debug("%s: clock-latency: %d\n", __func__, transition_latency);
|
|
||||||
return transition_latency;
|
|
||||||
}
|
|
||||||
|
|
||||||
static const struct cpufreq_arm_bL_ops dt_bL_ops = {
|
|
||||||
.name = "dt-bl",
|
|
||||||
.get_transition_latency = dt_get_transition_latency,
|
|
||||||
.init_opp_table = dev_pm_opp_of_cpumask_add_table,
|
|
||||||
.free_opp_table = dev_pm_opp_of_cpumask_remove_table,
|
|
||||||
};
|
|
||||||
|
|
||||||
static int generic_bL_probe(struct platform_device *pdev)
|
|
||||||
{
|
|
||||||
struct device_node *np;
|
|
||||||
|
|
||||||
np = get_cpu_node_with_valid_op(0);
|
|
||||||
if (!np)
|
|
||||||
return -ENODEV;
|
|
||||||
|
|
||||||
of_node_put(np);
|
|
||||||
return bL_cpufreq_register(&dt_bL_ops);
|
|
||||||
}
|
|
||||||
|
|
||||||
static int generic_bL_remove(struct platform_device *pdev)
|
|
||||||
{
|
|
||||||
bL_cpufreq_unregister(&dt_bL_ops);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static struct platform_driver generic_bL_platdrv = {
|
|
||||||
.driver = {
|
|
||||||
.name = "arm-bL-cpufreq-dt",
|
|
||||||
},
|
|
||||||
.probe = generic_bL_probe,
|
|
||||||
.remove = generic_bL_remove,
|
|
||||||
};
|
|
||||||
module_platform_driver(generic_bL_platdrv);
|
|
||||||
|
|
||||||
MODULE_AUTHOR("Viresh Kumar <viresh.kumar@linaro.org>");
|
|
||||||
MODULE_DESCRIPTION("Generic ARM big LITTLE cpufreq driver via DT");
|
|
||||||
MODULE_LICENSE("GPL v2");
|
|
@ -386,16 +386,11 @@ static int intel_pstate_get_cppc_guranteed(int cpu)
|
|||||||
return cppc_perf.guaranteed_perf;
|
return cppc_perf.guaranteed_perf;
|
||||||
}
|
}
|
||||||
|
|
||||||
#else
|
#else /* CONFIG_ACPI_CPPC_LIB */
|
||||||
static void intel_pstate_set_itmt_prio(int cpu)
|
static void intel_pstate_set_itmt_prio(int cpu)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
#endif /* CONFIG_ACPI_CPPC_LIB */
|
||||||
static int intel_pstate_get_cppc_guranteed(int cpu)
|
|
||||||
{
|
|
||||||
return -ENOTSUPP;
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
|
|
||||||
static void intel_pstate_init_acpi_perf_limits(struct cpufreq_policy *policy)
|
static void intel_pstate_init_acpi_perf_limits(struct cpufreq_policy *policy)
|
||||||
{
|
{
|
||||||
@ -477,7 +472,7 @@ static void intel_pstate_exit_perf_limits(struct cpufreq_policy *policy)
|
|||||||
|
|
||||||
acpi_processor_unregister_performance(policy->cpu);
|
acpi_processor_unregister_performance(policy->cpu);
|
||||||
}
|
}
|
||||||
#else
|
#else /* CONFIG_ACPI */
|
||||||
static inline void intel_pstate_init_acpi_perf_limits(struct cpufreq_policy *policy)
|
static inline void intel_pstate_init_acpi_perf_limits(struct cpufreq_policy *policy)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
@ -490,7 +485,14 @@ static inline bool intel_pstate_acpi_pm_profile_server(void)
|
|||||||
{
|
{
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
#endif
|
#endif /* CONFIG_ACPI */
|
||||||
|
|
||||||
|
#ifndef CONFIG_ACPI_CPPC_LIB
|
||||||
|
static int intel_pstate_get_cppc_guranteed(int cpu)
|
||||||
|
{
|
||||||
|
return -ENOTSUPP;
|
||||||
|
}
|
||||||
|
#endif /* CONFIG_ACPI_CPPC_LIB */
|
||||||
|
|
||||||
static inline void update_turbo_state(void)
|
static inline void update_turbo_state(void)
|
||||||
{
|
{
|
||||||
|
@ -134,11 +134,6 @@ struct menu_device {
|
|||||||
#define LOAD_INT(x) ((x) >> FSHIFT)
|
#define LOAD_INT(x) ((x) >> FSHIFT)
|
||||||
#define LOAD_FRAC(x) LOAD_INT(((x) & (FIXED_1-1)) * 100)
|
#define LOAD_FRAC(x) LOAD_INT(((x) & (FIXED_1-1)) * 100)
|
||||||
|
|
||||||
static inline int get_loadavg(unsigned long load)
|
|
||||||
{
|
|
||||||
return LOAD_INT(load) * 10 + LOAD_FRAC(load) / 10;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline int which_bucket(unsigned int duration, unsigned long nr_iowaiters)
|
static inline int which_bucket(unsigned int duration, unsigned long nr_iowaiters)
|
||||||
{
|
{
|
||||||
int bucket = 0;
|
int bucket = 0;
|
||||||
@ -172,18 +167,10 @@ static inline int which_bucket(unsigned int duration, unsigned long nr_iowaiters
|
|||||||
* to be, the higher this multiplier, and thus the higher
|
* to be, the higher this multiplier, and thus the higher
|
||||||
* the barrier to go to an expensive C state.
|
* the barrier to go to an expensive C state.
|
||||||
*/
|
*/
|
||||||
static inline int performance_multiplier(unsigned long nr_iowaiters, unsigned long load)
|
static inline int performance_multiplier(unsigned long nr_iowaiters)
|
||||||
{
|
{
|
||||||
int mult = 1;
|
/* for IO wait tasks (per cpu!) we add 10x each */
|
||||||
|
return 1 + 10 * nr_iowaiters;
|
||||||
/* for higher loadavg, we are more reluctant */
|
|
||||||
|
|
||||||
mult += 2 * get_loadavg(load);
|
|
||||||
|
|
||||||
/* for IO wait tasks (per cpu!) we add 5x each */
|
|
||||||
mult += 10 * nr_iowaiters;
|
|
||||||
|
|
||||||
return mult;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static DEFINE_PER_CPU(struct menu_device, menu_devices);
|
static DEFINE_PER_CPU(struct menu_device, menu_devices);
|
||||||
@ -301,7 +288,7 @@ static int menu_select(struct cpuidle_driver *drv, struct cpuidle_device *dev,
|
|||||||
int idx;
|
int idx;
|
||||||
unsigned int interactivity_req;
|
unsigned int interactivity_req;
|
||||||
unsigned int predicted_us;
|
unsigned int predicted_us;
|
||||||
unsigned long nr_iowaiters, cpu_load;
|
unsigned long nr_iowaiters;
|
||||||
ktime_t delta_next;
|
ktime_t delta_next;
|
||||||
|
|
||||||
if (data->needs_update) {
|
if (data->needs_update) {
|
||||||
@ -312,7 +299,7 @@ static int menu_select(struct cpuidle_driver *drv, struct cpuidle_device *dev,
|
|||||||
/* determine the expected residency time, round up */
|
/* determine the expected residency time, round up */
|
||||||
data->next_timer_us = ktime_to_us(tick_nohz_get_sleep_length(&delta_next));
|
data->next_timer_us = ktime_to_us(tick_nohz_get_sleep_length(&delta_next));
|
||||||
|
|
||||||
get_iowait_load(&nr_iowaiters, &cpu_load);
|
nr_iowaiters = nr_iowait_cpu(dev->cpu);
|
||||||
data->bucket = which_bucket(data->next_timer_us, nr_iowaiters);
|
data->bucket = which_bucket(data->next_timer_us, nr_iowaiters);
|
||||||
|
|
||||||
if (unlikely(drv->state_count <= 1 || latency_req == 0) ||
|
if (unlikely(drv->state_count <= 1 || latency_req == 0) ||
|
||||||
@ -356,7 +343,7 @@ static int menu_select(struct cpuidle_driver *drv, struct cpuidle_device *dev,
|
|||||||
* Use the performance multiplier and the user-configurable
|
* Use the performance multiplier and the user-configurable
|
||||||
* latency_req to determine the maximum exit latency.
|
* latency_req to determine the maximum exit latency.
|
||||||
*/
|
*/
|
||||||
interactivity_req = predicted_us / performance_multiplier(nr_iowaiters, cpu_load);
|
interactivity_req = predicted_us / performance_multiplier(nr_iowaiters);
|
||||||
if (latency_req > interactivity_req)
|
if (latency_req > interactivity_req)
|
||||||
latency_req = interactivity_req;
|
latency_req = interactivity_req;
|
||||||
}
|
}
|
||||||
|
@ -20,7 +20,6 @@ extern unsigned long nr_running(void);
|
|||||||
extern bool single_task_running(void);
|
extern bool single_task_running(void);
|
||||||
extern unsigned long nr_iowait(void);
|
extern unsigned long nr_iowait(void);
|
||||||
extern unsigned long nr_iowait_cpu(int cpu);
|
extern unsigned long nr_iowait_cpu(int cpu);
|
||||||
extern void get_iowait_load(unsigned long *nr_waiters, unsigned long *load);
|
|
||||||
|
|
||||||
static inline int sched_info_on(void)
|
static inline int sched_info_on(void)
|
||||||
{
|
{
|
||||||
|
@ -2875,6 +2875,18 @@ unsigned long long nr_context_switches(void)
|
|||||||
return sum;
|
return sum;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Consumers of these two interfaces, like for example the cpuidle menu
|
||||||
|
* governor, are using nonsensical data. Preferring shallow idle state selection
|
||||||
|
* for a CPU that has IO-wait which might not even end up running the task when
|
||||||
|
* it does become runnable.
|
||||||
|
*/
|
||||||
|
|
||||||
|
unsigned long nr_iowait_cpu(int cpu)
|
||||||
|
{
|
||||||
|
return atomic_read(&cpu_rq(cpu)->nr_iowait);
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* IO-wait accounting, and how its mostly bollocks (on SMP).
|
* IO-wait accounting, and how its mostly bollocks (on SMP).
|
||||||
*
|
*
|
||||||
@ -2910,31 +2922,11 @@ unsigned long nr_iowait(void)
|
|||||||
unsigned long i, sum = 0;
|
unsigned long i, sum = 0;
|
||||||
|
|
||||||
for_each_possible_cpu(i)
|
for_each_possible_cpu(i)
|
||||||
sum += atomic_read(&cpu_rq(i)->nr_iowait);
|
sum += nr_iowait_cpu(i);
|
||||||
|
|
||||||
return sum;
|
return sum;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
|
||||||
* Consumers of these two interfaces, like for example the cpuidle menu
|
|
||||||
* governor, are using nonsensical data. Preferring shallow idle state selection
|
|
||||||
* for a CPU that has IO-wait which might not even end up running the task when
|
|
||||||
* it does become runnable.
|
|
||||||
*/
|
|
||||||
|
|
||||||
unsigned long nr_iowait_cpu(int cpu)
|
|
||||||
{
|
|
||||||
struct rq *this = cpu_rq(cpu);
|
|
||||||
return atomic_read(&this->nr_iowait);
|
|
||||||
}
|
|
||||||
|
|
||||||
void get_iowait_load(unsigned long *nr_waiters, unsigned long *load)
|
|
||||||
{
|
|
||||||
struct rq *rq = this_rq();
|
|
||||||
*nr_waiters = atomic_read(&rq->nr_iowait);
|
|
||||||
*load = rq->load.weight;
|
|
||||||
}
|
|
||||||
|
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
Loading…
x
Reference in New Issue
Block a user