mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-12-29 01:05:29 +00:00
ea167a7fc2
Commit3c0897c180
("cpufreq: Use scnprintf() for avoiding potential buffer overflow") switched from snprintf to the more secure scnprintf but never updated the exit condition for PAGE_SIZE. As the commit say and as scnprintf document, what scnprintf returns what is actually written not counting the '\0' end char. This results in the case of len exceeding the size, len set to PAGE_SIZE - 1, as it can be written at max PAGE_SIZE - 1 (as '\0' is not counted) Because of len is never set to PAGE_SIZE, the function never break early, never prints the warning and never return -EFBIG. Fix this by changing the condition to PAGE_SIZE - 1 to correctly trigger the error. Cc: 5.10+ <stable@vger.kernel.org> # 5.10+ Fixes:3c0897c180
("cpufreq: Use scnprintf() for avoiding potential buffer overflow") Signed-off-by: Christian Marangi <ansuelsmth@gmail.com> [ rjw: Subject and changelog edits ] Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
290 lines
7.1 KiB
C
290 lines
7.1 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* drivers/cpufreq/cpufreq_stats.c
|
|
*
|
|
* Copyright (C) 2003-2004 Venkatesh Pallipadi <venkatesh.pallipadi@intel.com>.
|
|
* (C) 2004 Zou Nan hai <nanhai.zou@intel.com>.
|
|
*/
|
|
|
|
#include <linux/cpu.h>
|
|
#include <linux/cpufreq.h>
|
|
#include <linux/module.h>
|
|
#include <linux/sched/clock.h>
|
|
#include <linux/slab.h>
|
|
|
|
struct cpufreq_stats {
|
|
unsigned int total_trans;
|
|
unsigned long long last_time;
|
|
unsigned int max_state;
|
|
unsigned int state_num;
|
|
unsigned int last_index;
|
|
u64 *time_in_state;
|
|
unsigned int *freq_table;
|
|
unsigned int *trans_table;
|
|
|
|
/* Deferred reset */
|
|
unsigned int reset_pending;
|
|
unsigned long long reset_time;
|
|
};
|
|
|
|
static void cpufreq_stats_update(struct cpufreq_stats *stats,
|
|
unsigned long long time)
|
|
{
|
|
unsigned long long cur_time = local_clock();
|
|
|
|
stats->time_in_state[stats->last_index] += cur_time - time;
|
|
stats->last_time = cur_time;
|
|
}
|
|
|
|
static void cpufreq_stats_reset_table(struct cpufreq_stats *stats)
|
|
{
|
|
unsigned int count = stats->max_state;
|
|
|
|
memset(stats->time_in_state, 0, count * sizeof(u64));
|
|
memset(stats->trans_table, 0, count * count * sizeof(int));
|
|
stats->last_time = local_clock();
|
|
stats->total_trans = 0;
|
|
|
|
/* Adjust for the time elapsed since reset was requested */
|
|
WRITE_ONCE(stats->reset_pending, 0);
|
|
/*
|
|
* Prevent the reset_time read from being reordered before the
|
|
* reset_pending accesses in cpufreq_stats_record_transition().
|
|
*/
|
|
smp_rmb();
|
|
cpufreq_stats_update(stats, READ_ONCE(stats->reset_time));
|
|
}
|
|
|
|
static ssize_t show_total_trans(struct cpufreq_policy *policy, char *buf)
|
|
{
|
|
struct cpufreq_stats *stats = policy->stats;
|
|
|
|
if (READ_ONCE(stats->reset_pending))
|
|
return sprintf(buf, "%d\n", 0);
|
|
else
|
|
return sprintf(buf, "%u\n", stats->total_trans);
|
|
}
|
|
cpufreq_freq_attr_ro(total_trans);
|
|
|
|
static ssize_t show_time_in_state(struct cpufreq_policy *policy, char *buf)
|
|
{
|
|
struct cpufreq_stats *stats = policy->stats;
|
|
bool pending = READ_ONCE(stats->reset_pending);
|
|
unsigned long long time;
|
|
ssize_t len = 0;
|
|
int i;
|
|
|
|
for (i = 0; i < stats->state_num; i++) {
|
|
if (pending) {
|
|
if (i == stats->last_index) {
|
|
/*
|
|
* Prevent the reset_time read from occurring
|
|
* before the reset_pending read above.
|
|
*/
|
|
smp_rmb();
|
|
time = local_clock() - READ_ONCE(stats->reset_time);
|
|
} else {
|
|
time = 0;
|
|
}
|
|
} else {
|
|
time = stats->time_in_state[i];
|
|
if (i == stats->last_index)
|
|
time += local_clock() - stats->last_time;
|
|
}
|
|
|
|
len += sprintf(buf + len, "%u %llu\n", stats->freq_table[i],
|
|
nsec_to_clock_t(time));
|
|
}
|
|
return len;
|
|
}
|
|
cpufreq_freq_attr_ro(time_in_state);
|
|
|
|
/* We don't care what is written to the attribute */
|
|
static ssize_t store_reset(struct cpufreq_policy *policy, const char *buf,
|
|
size_t count)
|
|
{
|
|
struct cpufreq_stats *stats = policy->stats;
|
|
|
|
/*
|
|
* Defer resetting of stats to cpufreq_stats_record_transition() to
|
|
* avoid races.
|
|
*/
|
|
WRITE_ONCE(stats->reset_time, local_clock());
|
|
/*
|
|
* The memory barrier below is to prevent the readers of reset_time from
|
|
* seeing a stale or partially updated value.
|
|
*/
|
|
smp_wmb();
|
|
WRITE_ONCE(stats->reset_pending, 1);
|
|
|
|
return count;
|
|
}
|
|
cpufreq_freq_attr_wo(reset);
|
|
|
|
static ssize_t show_trans_table(struct cpufreq_policy *policy, char *buf)
|
|
{
|
|
struct cpufreq_stats *stats = policy->stats;
|
|
bool pending = READ_ONCE(stats->reset_pending);
|
|
ssize_t len = 0;
|
|
int i, j, count;
|
|
|
|
len += sysfs_emit_at(buf, len, " From : To\n");
|
|
len += sysfs_emit_at(buf, len, " : ");
|
|
for (i = 0; i < stats->state_num; i++) {
|
|
if (len >= PAGE_SIZE - 1)
|
|
break;
|
|
len += sysfs_emit_at(buf, len, "%9u ", stats->freq_table[i]);
|
|
}
|
|
if (len >= PAGE_SIZE - 1)
|
|
return PAGE_SIZE - 1;
|
|
|
|
len += sysfs_emit_at(buf, len, "\n");
|
|
|
|
for (i = 0; i < stats->state_num; i++) {
|
|
if (len >= PAGE_SIZE - 1)
|
|
break;
|
|
|
|
len += sysfs_emit_at(buf, len, "%9u: ", stats->freq_table[i]);
|
|
|
|
for (j = 0; j < stats->state_num; j++) {
|
|
if (len >= PAGE_SIZE - 1)
|
|
break;
|
|
|
|
if (pending)
|
|
count = 0;
|
|
else
|
|
count = stats->trans_table[i * stats->max_state + j];
|
|
|
|
len += sysfs_emit_at(buf, len, "%9u ", count);
|
|
}
|
|
if (len >= PAGE_SIZE - 1)
|
|
break;
|
|
len += sysfs_emit_at(buf, len, "\n");
|
|
}
|
|
|
|
if (len >= PAGE_SIZE - 1) {
|
|
pr_warn_once("cpufreq transition table exceeds PAGE_SIZE. Disabling\n");
|
|
return -EFBIG;
|
|
}
|
|
return len;
|
|
}
|
|
cpufreq_freq_attr_ro(trans_table);
|
|
|
|
static struct attribute *default_attrs[] = {
|
|
&total_trans.attr,
|
|
&time_in_state.attr,
|
|
&reset.attr,
|
|
&trans_table.attr,
|
|
NULL
|
|
};
|
|
static const struct attribute_group stats_attr_group = {
|
|
.attrs = default_attrs,
|
|
.name = "stats"
|
|
};
|
|
|
|
static int freq_table_get_index(struct cpufreq_stats *stats, unsigned int freq)
|
|
{
|
|
int index;
|
|
for (index = 0; index < stats->max_state; index++)
|
|
if (stats->freq_table[index] == freq)
|
|
return index;
|
|
return -1;
|
|
}
|
|
|
|
void cpufreq_stats_free_table(struct cpufreq_policy *policy)
|
|
{
|
|
struct cpufreq_stats *stats = policy->stats;
|
|
|
|
/* Already freed */
|
|
if (!stats)
|
|
return;
|
|
|
|
pr_debug("%s: Free stats table\n", __func__);
|
|
|
|
sysfs_remove_group(&policy->kobj, &stats_attr_group);
|
|
kfree(stats->time_in_state);
|
|
kfree(stats);
|
|
policy->stats = NULL;
|
|
}
|
|
|
|
void cpufreq_stats_create_table(struct cpufreq_policy *policy)
|
|
{
|
|
unsigned int i = 0, count;
|
|
struct cpufreq_stats *stats;
|
|
unsigned int alloc_size;
|
|
struct cpufreq_frequency_table *pos;
|
|
|
|
count = cpufreq_table_count_valid_entries(policy);
|
|
if (!count)
|
|
return;
|
|
|
|
/* stats already initialized */
|
|
if (policy->stats)
|
|
return;
|
|
|
|
stats = kzalloc(sizeof(*stats), GFP_KERNEL);
|
|
if (!stats)
|
|
return;
|
|
|
|
alloc_size = count * sizeof(int) + count * sizeof(u64);
|
|
|
|
alloc_size += count * count * sizeof(int);
|
|
|
|
/* Allocate memory for time_in_state/freq_table/trans_table in one go */
|
|
stats->time_in_state = kzalloc(alloc_size, GFP_KERNEL);
|
|
if (!stats->time_in_state)
|
|
goto free_stat;
|
|
|
|
stats->freq_table = (unsigned int *)(stats->time_in_state + count);
|
|
|
|
stats->trans_table = stats->freq_table + count;
|
|
|
|
stats->max_state = count;
|
|
|
|
/* Find valid-unique entries */
|
|
cpufreq_for_each_valid_entry(pos, policy->freq_table)
|
|
if (policy->freq_table_sorted != CPUFREQ_TABLE_UNSORTED ||
|
|
freq_table_get_index(stats, pos->frequency) == -1)
|
|
stats->freq_table[i++] = pos->frequency;
|
|
|
|
stats->state_num = i;
|
|
stats->last_time = local_clock();
|
|
stats->last_index = freq_table_get_index(stats, policy->cur);
|
|
|
|
policy->stats = stats;
|
|
if (!sysfs_create_group(&policy->kobj, &stats_attr_group))
|
|
return;
|
|
|
|
/* We failed, release resources */
|
|
policy->stats = NULL;
|
|
kfree(stats->time_in_state);
|
|
free_stat:
|
|
kfree(stats);
|
|
}
|
|
|
|
void cpufreq_stats_record_transition(struct cpufreq_policy *policy,
|
|
unsigned int new_freq)
|
|
{
|
|
struct cpufreq_stats *stats = policy->stats;
|
|
int old_index, new_index;
|
|
|
|
if (unlikely(!stats))
|
|
return;
|
|
|
|
if (unlikely(READ_ONCE(stats->reset_pending)))
|
|
cpufreq_stats_reset_table(stats);
|
|
|
|
old_index = stats->last_index;
|
|
new_index = freq_table_get_index(stats, new_freq);
|
|
|
|
/* We can't do stats->time_in_state[-1]= .. */
|
|
if (unlikely(old_index == -1 || new_index == -1 || old_index == new_index))
|
|
return;
|
|
|
|
cpufreq_stats_update(stats, stats->last_time);
|
|
|
|
stats->last_index = new_index;
|
|
stats->trans_table[old_index * stats->max_state + new_index]++;
|
|
stats->total_trans++;
|
|
}
|