Merge branch 'fixes' of git://git.kernel.org/pub/scm/linux/kernel/git/davej/cpufreq
* 'fixes' of git://git.kernel.org/pub/scm/linux/kernel/git/davej/cpufreq: [CPUFREQ] powernow-k8: Get transition latency from ACPI _PSS table [CPUFREQ] Make ignore_nice_load setting of ondemand work as expected.
This commit is contained in:
commit
6707fbb56c
|
@ -939,10 +939,25 @@ static void powernow_k8_cpu_exit_acpi(struct powernow_k8_data *data)
|
||||||
free_cpumask_var(data->acpi_data.shared_cpu_map);
|
free_cpumask_var(data->acpi_data.shared_cpu_map);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static int get_transition_latency(struct powernow_k8_data *data)
|
||||||
|
{
|
||||||
|
int max_latency = 0;
|
||||||
|
int i;
|
||||||
|
for (i = 0; i < data->acpi_data.state_count; i++) {
|
||||||
|
int cur_latency = data->acpi_data.states[i].transition_latency
|
||||||
|
+ data->acpi_data.states[i].bus_master_latency;
|
||||||
|
if (cur_latency > max_latency)
|
||||||
|
max_latency = cur_latency;
|
||||||
|
}
|
||||||
|
/* value in usecs, needs to be in nanoseconds */
|
||||||
|
return 1000 * max_latency;
|
||||||
|
}
|
||||||
|
|
||||||
#else
|
#else
|
||||||
static int powernow_k8_cpu_init_acpi(struct powernow_k8_data *data) { return -ENODEV; }
|
static int powernow_k8_cpu_init_acpi(struct powernow_k8_data *data) { return -ENODEV; }
|
||||||
static void powernow_k8_cpu_exit_acpi(struct powernow_k8_data *data) { return; }
|
static void powernow_k8_cpu_exit_acpi(struct powernow_k8_data *data) { return; }
|
||||||
static void powernow_k8_acpi_pst_values(struct powernow_k8_data *data, unsigned int index) { return; }
|
static void powernow_k8_acpi_pst_values(struct powernow_k8_data *data, unsigned int index) { return; }
|
||||||
|
static int get_transition_latency(struct powernow_k8_data *data) { return 0; }
|
||||||
#endif /* CONFIG_X86_POWERNOW_K8_ACPI */
|
#endif /* CONFIG_X86_POWERNOW_K8_ACPI */
|
||||||
|
|
||||||
/* Take a frequency, and issue the fid/vid transition command */
|
/* Take a frequency, and issue the fid/vid transition command */
|
||||||
|
@ -1173,7 +1188,13 @@ static int __cpuinit powernowk8_cpu_init(struct cpufreq_policy *pol)
|
||||||
if (rc) {
|
if (rc) {
|
||||||
goto err_out;
|
goto err_out;
|
||||||
}
|
}
|
||||||
}
|
/* Take a crude guess here.
|
||||||
|
* That guess was in microseconds, so multiply with 1000 */
|
||||||
|
pol->cpuinfo.transition_latency = (
|
||||||
|
((data->rvo + 8) * data->vstable * VST_UNITS_20US) +
|
||||||
|
((1 << data->irt) * 30)) * 1000;
|
||||||
|
} else /* ACPI _PSS objects available */
|
||||||
|
pol->cpuinfo.transition_latency = get_transition_latency(data);
|
||||||
|
|
||||||
/* only run on specific CPU from here on */
|
/* only run on specific CPU from here on */
|
||||||
oldmask = current->cpus_allowed;
|
oldmask = current->cpus_allowed;
|
||||||
|
@ -1204,11 +1225,6 @@ static int __cpuinit powernowk8_cpu_init(struct cpufreq_policy *pol)
|
||||||
cpumask_copy(pol->cpus, &per_cpu(cpu_core_map, pol->cpu));
|
cpumask_copy(pol->cpus, &per_cpu(cpu_core_map, pol->cpu));
|
||||||
data->available_cores = pol->cpus;
|
data->available_cores = pol->cpus;
|
||||||
|
|
||||||
/* Take a crude guess here.
|
|
||||||
* That guess was in microseconds, so multiply with 1000 */
|
|
||||||
pol->cpuinfo.transition_latency = (((data->rvo + 8) * data->vstable * VST_UNITS_20US)
|
|
||||||
+ (3 * (1 << data->irt) * 10)) * 1000;
|
|
||||||
|
|
||||||
if (cpu_family == CPU_HW_PSTATE)
|
if (cpu_family == CPU_HW_PSTATE)
|
||||||
pol->cur = find_khz_freq_from_pstate(data->powernow_table, data->currpstate);
|
pol->cur = find_khz_freq_from_pstate(data->powernow_table, data->currpstate);
|
||||||
else
|
else
|
||||||
|
|
|
@ -117,11 +117,7 @@ static inline cputime64_t get_cpu_idle_time_jiffy(unsigned int cpu,
|
||||||
busy_time = cputime64_add(busy_time, kstat_cpu(cpu).cpustat.irq);
|
busy_time = cputime64_add(busy_time, kstat_cpu(cpu).cpustat.irq);
|
||||||
busy_time = cputime64_add(busy_time, kstat_cpu(cpu).cpustat.softirq);
|
busy_time = cputime64_add(busy_time, kstat_cpu(cpu).cpustat.softirq);
|
||||||
busy_time = cputime64_add(busy_time, kstat_cpu(cpu).cpustat.steal);
|
busy_time = cputime64_add(busy_time, kstat_cpu(cpu).cpustat.steal);
|
||||||
|
busy_time = cputime64_add(busy_time, kstat_cpu(cpu).cpustat.nice);
|
||||||
if (!dbs_tuners_ins.ignore_nice) {
|
|
||||||
busy_time = cputime64_add(busy_time,
|
|
||||||
kstat_cpu(cpu).cpustat.nice);
|
|
||||||
}
|
|
||||||
|
|
||||||
idle_time = cputime64_sub(cur_wall_time, busy_time);
|
idle_time = cputime64_sub(cur_wall_time, busy_time);
|
||||||
if (wall)
|
if (wall)
|
||||||
|
@ -137,23 +133,6 @@ static inline cputime64_t get_cpu_idle_time(unsigned int cpu, cputime64_t *wall)
|
||||||
if (idle_time == -1ULL)
|
if (idle_time == -1ULL)
|
||||||
return get_cpu_idle_time_jiffy(cpu, wall);
|
return get_cpu_idle_time_jiffy(cpu, wall);
|
||||||
|
|
||||||
if (dbs_tuners_ins.ignore_nice) {
|
|
||||||
cputime64_t cur_nice;
|
|
||||||
unsigned long cur_nice_jiffies;
|
|
||||||
struct cpu_dbs_info_s *dbs_info;
|
|
||||||
|
|
||||||
dbs_info = &per_cpu(cpu_dbs_info, cpu);
|
|
||||||
cur_nice = cputime64_sub(kstat_cpu(cpu).cpustat.nice,
|
|
||||||
dbs_info->prev_cpu_nice);
|
|
||||||
/*
|
|
||||||
* Assumption: nice time between sampling periods will be
|
|
||||||
* less than 2^32 jiffies for 32 bit sys
|
|
||||||
*/
|
|
||||||
cur_nice_jiffies = (unsigned long)
|
|
||||||
cputime64_to_jiffies64(cur_nice);
|
|
||||||
dbs_info->prev_cpu_nice = kstat_cpu(cpu).cpustat.nice;
|
|
||||||
return idle_time + jiffies_to_usecs(cur_nice_jiffies);
|
|
||||||
}
|
|
||||||
return idle_time;
|
return idle_time;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -319,6 +298,9 @@ static ssize_t store_ignore_nice_load(struct cpufreq_policy *policy,
|
||||||
dbs_info = &per_cpu(cpu_dbs_info, j);
|
dbs_info = &per_cpu(cpu_dbs_info, j);
|
||||||
dbs_info->prev_cpu_idle = get_cpu_idle_time(j,
|
dbs_info->prev_cpu_idle = get_cpu_idle_time(j,
|
||||||
&dbs_info->prev_cpu_wall);
|
&dbs_info->prev_cpu_wall);
|
||||||
|
if (dbs_tuners_ins.ignore_nice)
|
||||||
|
dbs_info->prev_cpu_nice = kstat_cpu(j).cpustat.nice;
|
||||||
|
|
||||||
}
|
}
|
||||||
mutex_unlock(&dbs_mutex);
|
mutex_unlock(&dbs_mutex);
|
||||||
|
|
||||||
|
@ -419,6 +401,23 @@ static void dbs_check_cpu(struct cpu_dbs_info_s *this_dbs_info)
|
||||||
j_dbs_info->prev_cpu_idle);
|
j_dbs_info->prev_cpu_idle);
|
||||||
j_dbs_info->prev_cpu_idle = cur_idle_time;
|
j_dbs_info->prev_cpu_idle = cur_idle_time;
|
||||||
|
|
||||||
|
if (dbs_tuners_ins.ignore_nice) {
|
||||||
|
cputime64_t cur_nice;
|
||||||
|
unsigned long cur_nice_jiffies;
|
||||||
|
|
||||||
|
cur_nice = cputime64_sub(kstat_cpu(j).cpustat.nice,
|
||||||
|
j_dbs_info->prev_cpu_nice);
|
||||||
|
/*
|
||||||
|
* Assumption: nice time between sampling periods will
|
||||||
|
* be less than 2^32 jiffies for 32 bit sys
|
||||||
|
*/
|
||||||
|
cur_nice_jiffies = (unsigned long)
|
||||||
|
cputime64_to_jiffies64(cur_nice);
|
||||||
|
|
||||||
|
j_dbs_info->prev_cpu_nice = kstat_cpu(j).cpustat.nice;
|
||||||
|
idle_time += jiffies_to_usecs(cur_nice_jiffies);
|
||||||
|
}
|
||||||
|
|
||||||
if (unlikely(!wall_time || wall_time < idle_time))
|
if (unlikely(!wall_time || wall_time < idle_time))
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
|
@ -575,6 +574,10 @@ static int cpufreq_governor_dbs(struct cpufreq_policy *policy,
|
||||||
|
|
||||||
j_dbs_info->prev_cpu_idle = get_cpu_idle_time(j,
|
j_dbs_info->prev_cpu_idle = get_cpu_idle_time(j,
|
||||||
&j_dbs_info->prev_cpu_wall);
|
&j_dbs_info->prev_cpu_wall);
|
||||||
|
if (dbs_tuners_ins.ignore_nice) {
|
||||||
|
j_dbs_info->prev_cpu_nice =
|
||||||
|
kstat_cpu(j).cpustat.nice;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
this_dbs_info->cpu = cpu;
|
this_dbs_info->cpu = cpu;
|
||||||
/*
|
/*
|
||||||
|
|
Loading…
Reference in New Issue