/*
* Copyright (c) 2018, NVIDIA Corporation. All rights reserved.
*
* This program is free software; you can redistribute it and/or modify it
* under the terms and conditions of the GNU General Public License,
* version 2, as published by the Free Software Foundation.
*
* This program is distributed in the hope it will be useful, but WITHOUT
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
* more details.
*
* You should have received a copy of the GNU General Public License
* along with this program. If not, see .
*/
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include "denver-knobs.h"
#define BG_CLR_SHIFT 16
#define BG_STATUS_SHIFT 32
#define NVG_CHANNEL_PMIC 0
int smp_call_function_denver(smp_call_func_t func, void *info, int wait)
{
int cpu;
struct cpuinfo_arm64 *cpuinfo;
for_each_online_cpu(cpu) {
/* Skip non-Denver CPUs */
cpuinfo = &per_cpu(cpu_data, cpu);
if (MIDR_IMPLEMENTOR(cpuinfo->reg_midr) != ARM_CPU_IMP_NVIDIA)
continue;
return smp_call_function_single(cpu, func, info, wait);
}
/* return ENXIO when there isn't an online Denver CPU to behave the
same as smp_call_function_single() when the requested CPU is offline
*/
return -ENXIO;
}
EXPORT_SYMBOL_GPL(smp_call_function_denver);
static void _denver_get_bg_allowed(void *ret)
{
unsigned long regval;
int cpu;
if (tegra_get_chip_id() == TEGRA186)
cpu = MPIDR_AFFINITY_LEVEL(cpu_logical_map(
smp_processor_id()), 0);
else
cpu = smp_processor_id();
asm volatile("mrs %0, s3_0_c15_c0_2" : "=r" (regval) : );
regval = (regval >> BG_STATUS_SHIFT) & 0xffff;
*((bool *) ret) = !!(regval & (1 << cpu));
}
bool denver_get_bg_allowed(int cpu)
{
bool enabled;
smp_call_function_single(cpu, _denver_get_bg_allowed,
(void *) &enabled, 1);
return enabled;
}
static void _denver_set_bg_allowed(void *_enable)
{
unsigned long regval;
bool enabled = 1;
bool enable = (bool) _enable;
int cpu;
if (tegra_get_chip_id() == TEGRA186)
cpu = MPIDR_AFFINITY_LEVEL(cpu_logical_map(
smp_processor_id()), 0);
else
cpu = smp_processor_id();
regval = 1 << cpu;
if (!enable)
regval <<= BG_CLR_SHIFT;
asm volatile("msr s3_0_c15_c0_2, %0" : : "r" (regval));
/* Flush all background work for disable */
if (!enable)
while (enabled)
_denver_get_bg_allowed((void *) &enabled);
}
void denver_set_bg_allowed(int cpu, bool enable)
{
smp_call_function_single(cpu, _denver_set_bg_allowed,
(void *) enable, 1);
}
static const char * const pmic_names[] = {
[UNDEFINED] = "none",
[AMS_372x] = "AMS 3722/3720",
[TI_TPS_65913_22] = "TI TPS65913 2.2",
[OPEN_VR] = "Open VR",
[TI_TPS_65913_23] = "TI TPS65913 2.3",
};
static DEFINE_SPINLOCK(nvg_lock);
#ifdef CONFIG_DEBUG_FS
static struct dentry *denver_debugfs_root;
#endif
static int bgallowed_get(void *data, u64 *val)
{
int cpu = (int)((s64)data);
*val = denver_get_bg_allowed(cpu);
return 0;
}
static int bgallowed_set(void *data, u64 val)
{
int cpu = (int)((s64)data);
denver_set_bg_allowed(cpu, (bool)val);
pr_debug("CPU%d: BGALLOWED is %s.\n",
cpu, val ? "enabled" : "disabled");
return 0;
}
DEFINE_SIMPLE_ATTRIBUTE(bgallowed_fops, bgallowed_get,
bgallowed_set, "%llu\n");
#ifdef CONFIG_DEBUG_FS
static int __init create_denver_bgallowed(void)
{
int cpu;
char name[30];
for_each_present_cpu(cpu) {
struct cpuinfo_arm64 *cpuinfo = &per_cpu(cpu_data, cpu);
/* Skip non-denver CPUs */
if (MIDR_IMPLEMENTOR(cpuinfo->reg_midr) != ARM_CPU_IMP_NVIDIA)
continue;
snprintf(name, 30, "bgallowed_cpu%d", cpu);
if (!debugfs_create_file(
name, S_IRUGO, denver_debugfs_root,
(void *)((s64)cpu), &bgallowed_fops)) {
pr_info("ERROR: failed to create bgallow_fs");
return -ENOMEM;
}
}
return 0;
}
#endif
struct nvmstat {
u64 stat0;
u64 tot;
u64 stat1;
u64 bg;
u64 fg;
};
static struct nvmstat nvmstat_agg;
static void _get_stats(void *_stat)
{
u64 stat0;
struct nvmstat *stat = (struct nvmstat *) _stat;
/* read nvmstat0 */
asm volatile("mrs %0, s3_0_c15_c0_0" : "=r" (stat->stat0) : );
stat->tot = (u32)(stat->stat0 >> 32);
nvmstat_agg.tot += stat->tot;
/* read nvmstat1 */
asm volatile("mrs %0, s3_0_c15_c0_1" : "=r" (stat->stat1) : );
stat->bg = (u32)(stat->stat1);
nvmstat_agg.bg += stat->bg;
stat->fg = (u32)(stat->stat1 >> 32);
nvmstat_agg.fg += stat->fg;
/* reset nvmstat0 */
stat0 = 0;
asm volatile("msr s3_0_c15_c0_0, %0" : : "r" (stat0));
}
static int get_stats(struct nvmstat *stats)
{
int cpu;
for_each_online_cpu(cpu) {
struct cpuinfo_arm64 *cpuinfo = &per_cpu(cpu_data, cpu);
/* Skip non-denver CPUs */
if (MIDR_IMPLEMENTOR(cpuinfo->reg_midr) == ARM_CPU_IMP_NVIDIA) {
smp_call_function_single(cpu, _get_stats,
(void *) stats, 1);
return 1;
}
}
return 0;
}
static void print_stats(struct seq_file *s, struct nvmstat *stat)
{
seq_printf(s, "nvmstat0 = %llu\n", stat->stat0);
seq_printf(s, "nvmstat0_tot = %llu\n", stat->tot);
seq_printf(s, "nvmstat1 = %llu\n", stat->stat1);
seq_printf(s, "nvmstat1_bg = %llu\n", stat->bg);
seq_printf(s, "nvmstat1_fg = %llu\n", stat->fg);
}
static int inst_stats_show(struct seq_file *s, void *data)
{
struct nvmstat stat;
if (!get_stats(&stat)) {
seq_puts(s, "No Denver cores online\n");
return 0;
}
print_stats(s, &stat);
return 0;
}
static int inst_stats_open(struct inode *inode, struct file *file)
{
return single_open(file, inst_stats_show, inode->i_private);
}
static const struct file_operations inst_stats_fops = {
.open = inst_stats_open,
.read = seq_read,
.llseek = seq_lseek,
.release = single_release,
};
static int agg_stats_show(struct seq_file *s, void *data)
{
struct nvmstat stat;
if (!get_stats(&stat)) {
seq_puts(s, "No Denver cores online\n");
return 0;
}
print_stats(s, &nvmstat_agg);
return 0;
}
static int agg_stats_open(struct inode *inode, struct file *file)
{
return single_open(file, agg_stats_show, inode->i_private);
}
static const struct file_operations agg_stats_fops = {
.open = agg_stats_open,
.read = seq_read,
.llseek = seq_lseek,
.release = single_release,
};
#ifdef CONFIG_DEBUG_FS
static int __init create_denver_nvmstats(void)
{
struct dentry *nvmstats_dir;
nvmstats_dir = debugfs_create_dir("nvmstats", denver_debugfs_root);
if (!nvmstats_dir) {
pr_err("%s: Couldn't create the \"nvmstats\" debugfs node.\n",
__func__);
return -1;
}
if (!debugfs_create_file("instantaneous_stats",
S_IRUGO,
nvmstats_dir,
NULL,
&inst_stats_fops)) {
pr_err("%s: Couldn't create the \"instantaneous_stats\" debugfs node.\n",
__func__);
return -1;
}
if (!debugfs_create_file("aggregated_stats",
S_IRUGO,
nvmstats_dir,
NULL,
&agg_stats_fops)) {
pr_err("%s: Couldn't create the \"aggregated_stats\" debugfs node.\n",
__func__);
return -1;
}
return 0;
}
#endif
static void denver_set_mts_nvgindex(u32 index)
{
asm volatile("msr s3_0_c15_c1_2, %0" : : "r" (index));
}
static void denver_set_mts_nvgdata(u64 data)
{
asm volatile("msr s3_0_c15_c1_3, %0" : : "r" (data));
}
static void denver_get_mts_nvgdata(u64 *data)
{
asm volatile("mrs %0, s3_0_c15_c1_3" : "=r" (*data));
}
int denver_set_pmic_config(enum denver_pmic_type type,
u16 ret_vol, bool lock)
{
u64 reg;
static bool locked;
if (type >= NR_PMIC_TYPES)
return -EINVAL;
if (locked) {
pr_warn("Denver: PMIC config is locked.\n");
return -EINVAL;
}
spin_lock(&nvg_lock);
denver_set_mts_nvgindex(NVG_CHANNEL_PMIC);
/* retention voltage is sanitized by MTS */
reg = type | (ret_vol << 16) | ((u64)lock << 63);
denver_set_mts_nvgdata(reg);
spin_unlock(&nvg_lock);
locked = lock;
return 0;
}
int denver_get_pmic_config(enum denver_pmic_type *type,
u16 *ret_vol, bool *lock)
{
u64 reg = 0;
spin_lock(&nvg_lock);
denver_set_mts_nvgindex(NVG_CHANNEL_PMIC);
denver_get_mts_nvgdata(®);
spin_unlock(&nvg_lock);
*type = reg & 0xffff;
*ret_vol = (reg >> 16) & 0xffff;
*lock = (reg >> 63);
return 0;
}
static int __init denver_pmic_init(void)
{
u32 voltage;
u32 type;
u32 lock;
int err;
struct device_node *np;
np = of_find_node_by_path("/denver_cpuidle_pmic");
if (!np) {
pr_debug("Denver: using default PMIC setting.\n");
return 0;
}
err = of_property_read_u32(np, "type", &type);
if (err) {
pr_err("%s: failed to read PMIC type\n", __func__);
goto done;
}
if (type >= NR_PMIC_TYPES) {
pr_err("%s: invalid PMIC type: %d\n", __func__, type);
goto done;
}
err = of_property_read_u32(np, "retention-voltage", &voltage);
if (err) {
pr_err("%s: failed to read voltage\n", __func__);
goto done;
}
err = of_property_read_u32(np, "lock", &lock);
if (err) {
pr_err("%s: failed to read lock\n", __func__);
goto done;
}
if (lock != 0 && lock != 1) {
pr_err("%s: invalid lock setting [0|1]: read %d\n", __func__, lock);
goto done;
}
err = denver_set_pmic_config(type, (u16)voltage, lock);
pr_info("Denver: PMIC: type = %s, voltage = %d, locked = %d\n",
pmic_names[type], voltage, lock);
done:
return err;
}
arch_initcall(denver_pmic_init);
static u32 mts_version;
#if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 14, 0)
static int denver_cpu_online(unsigned int cpu)
{
/* Record MTS version if the current CPU is Denver */
if (!mts_version && ((read_cpuid_id() >> 24) == 'N'))
asm volatile ("mrs %0, AIDR_EL1" : "=r" (mts_version));
return 0;
}
#else
static int mts_version_cpu_notify(struct notifier_block *nb,
unsigned long action, void *pcpu)
{
/* Record MTS version if the current CPU is Denver */
if (!mts_version && ((read_cpuid_id() >> 24) == 'N'))
asm volatile ("mrs %0, AIDR_EL1" : "=r" (mts_version));
return NOTIFY_OK;
}
static struct notifier_block mts_version_cpu_nb = {
.notifier_call = mts_version_cpu_notify,
};
#endif
static int __init denver_knobs_init_early(void)
{
#if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 14, 0)
return cpuhp_setup_state_nocalls(CPUHP_AP_ONLINE_DYN,
"denver:cpu:online", denver_cpu_online, NULL);
#else
return register_cpu_notifier(&mts_version_cpu_nb);
#endif
}
early_initcall(denver_knobs_init_early);
static int mts_version_show(struct seq_file *m, void *v)
{
seq_printf(m, "%u\n", mts_version);
return 0;
}
static int mts_version_open(struct inode *inode, struct file *file)
{
return single_open(file, mts_version_show, NULL);
}
static const struct file_operations mts_version_fops = {
.open = mts_version_open,
.read = seq_read,
.llseek = seq_lseek,
.release = single_release,
};
static int __init denver_knobs_init(void)
{
#ifdef CONFIG_DEBUG_FS
int error;
#endif
if (tegra_cpu_is_asim())
return 0;
#ifdef CONFIG_DEBUG_FS
denver_debugfs_root = debugfs_create_dir("tegra_denver", NULL);
#endif
#ifdef CONFIG_DEBUG_FS
error = create_denver_bgallowed();
if (error)
return error;
error = create_denver_nvmstats();
if (error)
return error;
#endif
/* Cancel the notifier as mts_version should be set now. */
#if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 14, 0)
cpuhp_remove_state_nocalls(CPUHP_AP_ONLINE_DYN);
#else
unregister_cpu_notifier(&mts_version_cpu_nb);
#endif
/* mts_version will be non-zero, only if its an NVIDIA CPU */
if (mts_version)
pr_info("%s:MTS_VERSION:%d\n", __func__, mts_version);
if (mts_version && !proc_create("mts_version", 0, NULL, &mts_version_fops)) {
pr_err("Failed to create /proc/mts_version!\n");
return -1;
}
return 0;
}
device_initcall(denver_knobs_init);