1496 lines
38 KiB
C
1496 lines
38 KiB
C
|
/*
|
||
|
* drivers/video/tegra/host/dev.c
|
||
|
*
|
||
|
* Tegra Graphics Host Driver Entrypoint
|
||
|
*
|
||
|
* Copyright (c) 2010-2021, NVIDIA Corporation. All rights reserved.
|
||
|
*
|
||
|
* This program is free software; you can redistribute it and/or modify it
|
||
|
* under the terms and conditions of the GNU General Public License,
|
||
|
* version 2, as published by the Free Software Foundation.
|
||
|
*
|
||
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
||
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||
|
* more details.
|
||
|
*
|
||
|
* You should have received a copy of the GNU General Public License
|
||
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||
|
*/
|
||
|
|
||
|
#include <linux/slab.h>
|
||
|
#include <linux/delay.h>
|
||
|
#include <linux/fs.h>
|
||
|
#include <linux/cdev.h>
|
||
|
#include <linux/uaccess.h>
|
||
|
#include <linux/file.h>
|
||
|
#include <linux/clk.h>
|
||
|
#include <linux/hrtimer.h>
|
||
|
#include <linux/module.h>
|
||
|
#include <linux/pm_runtime.h>
|
||
|
#include <linux/of.h>
|
||
|
#include <linux/of_device.h>
|
||
|
#include <linux/of_platform.h>
|
||
|
#include <soc/tegra/chip-id.h>
|
||
|
#include <linux/tegra_pm_domains.h>
|
||
|
#include <linux/vmalloc.h>
|
||
|
#include <linux/version.h>
|
||
|
#include <linux/poll.h>
|
||
|
#include <linux/anon_inodes.h>
|
||
|
#include <linux/kref.h>
|
||
|
#include <linux/nospec.h>
|
||
|
#include <linux/dma-mapping.h>
|
||
|
|
||
|
#include "dev.h"
|
||
|
#include <trace/events/nvhost.h>
|
||
|
|
||
|
#include <linux/nvhost.h>
|
||
|
#include <uapi/linux/nvhost_ioctl.h>
|
||
|
|
||
|
#include "debug.h"
|
||
|
#include "bus_client.h"
|
||
|
#include "nvhost_vm.h"
|
||
|
#include "nvhost_acm.h"
|
||
|
#include "nvhost_channel.h"
|
||
|
#include "nvhost_job.h"
|
||
|
#include "vhost/vhost.h"
|
||
|
|
||
|
#ifdef CONFIG_TEGRA_GRHOST_SYNC
|
||
|
#include "nvhost_sync.h"
|
||
|
#endif
|
||
|
|
||
|
#include "nvhost_scale.h"
|
||
|
#include "chip_support.h"
|
||
|
#include "t124/t124.h"
|
||
|
#include "t210/t210.h"
|
||
|
|
||
|
#ifdef CONFIG_ARCH_TEGRA_18x_SOC
|
||
|
#include "t186/t186.h"
|
||
|
#endif
|
||
|
#ifdef CONFIG_TEGRA_T19X_GRHOST
|
||
|
#include "t194/t194.h"
|
||
|
#endif
|
||
|
|
||
|
#define DRIVER_NAME "host1x"
|
||
|
|
||
|
static const char *num_syncpts_name = "num_pts";
|
||
|
static const char *num_mutexes_name = "num_mlocks";
|
||
|
static const char *gather_filter_enabled_name = "gather_filter_enabled";
|
||
|
static const char *alloc_syncpts_per_apps_name = "alloc_syncpts_per_apps";
|
||
|
static const char *alloc_chs_per_submits_name = "alloc_chs_per_submits";
|
||
|
static const char *syncpts_pts_base_name = "pts_base";
|
||
|
static const char *syncpts_pts_limit_name = "pts_limit";
|
||
|
|
||
|
struct nvhost_ctrl_userctx {
|
||
|
struct nvhost_master *dev;
|
||
|
u32 *mod_locks;
|
||
|
};
|
||
|
|
||
|
struct nvhost_capability_node {
|
||
|
struct kobj_attribute attr;
|
||
|
struct nvhost_master *host;
|
||
|
int (*read_func)(struct nvhost_syncpt *sp);
|
||
|
int (*store_func)(struct nvhost_syncpt *sp, const char *buf);
|
||
|
};
|
||
|
|
||
|
static LIST_HEAD(nvhost_masters);
|
||
|
static DEFINE_MUTEX(nvhost_masters_lock);
|
||
|
static bool vhost_init_done;
|
||
|
|
||
|
struct nvhost_master *nvhost_get_prim_host(void)
|
||
|
{
|
||
|
struct nvhost_master *host;
|
||
|
|
||
|
mutex_lock(&nvhost_masters_lock);
|
||
|
host = list_first_entry(&nvhost_masters, struct nvhost_master, list);
|
||
|
mutex_unlock(&nvhost_masters_lock);
|
||
|
return host;
|
||
|
}
|
||
|
|
||
|
static int nvhost_ctrlrelease(struct inode *inode, struct file *filp)
|
||
|
{
|
||
|
struct nvhost_ctrl_userctx *priv = filp->private_data;
|
||
|
int i;
|
||
|
|
||
|
trace_nvhost_ctrlrelease(priv->dev->dev->name);
|
||
|
|
||
|
filp->private_data = NULL;
|
||
|
if (priv->mod_locks[0])
|
||
|
nvhost_module_idle(priv->dev->dev);
|
||
|
for (i = 1; i < nvhost_syncpt_nb_mlocks(&priv->dev->syncpt); i++)
|
||
|
if (priv->mod_locks[i])
|
||
|
nvhost_mutex_unlock(&priv->dev->syncpt, i);
|
||
|
kfree(priv->mod_locks);
|
||
|
kfree(priv);
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static int nvhost_ctrlopen(struct inode *inode, struct file *filp)
|
||
|
{
|
||
|
struct nvhost_master *host =
|
||
|
container_of(inode->i_cdev, struct nvhost_master, cdev);
|
||
|
struct nvhost_ctrl_userctx *priv;
|
||
|
u32 *mod_locks;
|
||
|
|
||
|
trace_nvhost_ctrlopen(host->dev->name);
|
||
|
|
||
|
priv = kzalloc(sizeof(*priv), GFP_KERNEL);
|
||
|
mod_locks = kzalloc(sizeof(u32)
|
||
|
* nvhost_syncpt_nb_mlocks(&host->syncpt),
|
||
|
GFP_KERNEL);
|
||
|
|
||
|
if (!(priv && mod_locks)) {
|
||
|
nvhost_err(&host->dev->dev, "failed to allocate mod_locks");
|
||
|
kfree(priv);
|
||
|
kfree(mod_locks);
|
||
|
return -ENOMEM;
|
||
|
}
|
||
|
|
||
|
priv->dev = host;
|
||
|
priv->mod_locks = mod_locks;
|
||
|
filp->private_data = priv;
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static int nvhost_ioctl_ctrl_syncpt_read(struct nvhost_ctrl_userctx *ctx,
|
||
|
struct nvhost_ctrl_syncpt_read_args *args)
|
||
|
{
|
||
|
if (!nvhost_syncpt_is_valid_hw_pt(&ctx->dev->syncpt, args->id))
|
||
|
return -EINVAL;
|
||
|
|
||
|
args->id = array_index_nospec(args->id,
|
||
|
nvhost_syncpt_nb_hw_pts(&ctx->dev->syncpt));
|
||
|
|
||
|
args->value = nvhost_syncpt_read(&ctx->dev->syncpt, args->id);
|
||
|
trace_nvhost_ioctl_ctrl_syncpt_read(args->id, args->value);
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static int nvhost_ioctl_ctrl_syncpt_incr(struct nvhost_ctrl_userctx *ctx,
|
||
|
struct nvhost_ctrl_syncpt_incr_args *args)
|
||
|
{
|
||
|
struct nvhost_master *host = nvhost_get_syncpt_owner(args->id);
|
||
|
|
||
|
if (!host)
|
||
|
return -EINVAL;
|
||
|
trace_nvhost_ioctl_ctrl_syncpt_incr(args->id);
|
||
|
nvhost_syncpt_incr(&host->syncpt, args->id);
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static int nvhost_ioctl_ctrl_syncpt_waitex(struct nvhost_ctrl_userctx *ctx,
|
||
|
struct nvhost_ctrl_syncpt_waitex_args *args)
|
||
|
{
|
||
|
u32 timeout;
|
||
|
int err;
|
||
|
if (!nvhost_syncpt_is_valid_hw_pt_nospec(&ctx->dev->syncpt,
|
||
|
&args->id)) {
|
||
|
nvhost_err(&ctx->dev->dev->dev,
|
||
|
"invalid syncpt %u", args->id);
|
||
|
return -EINVAL;
|
||
|
}
|
||
|
|
||
|
if (args->timeout == NVHOST_NO_TIMEOUT)
|
||
|
/* FIXME: MAX_SCHEDULE_TIMEOUT is ulong which can be bigger
|
||
|
than u32 so we should fix nvhost_syncpt_wait_timeout to
|
||
|
take ulong not u32. */
|
||
|
timeout = (u32)MAX_SCHEDULE_TIMEOUT;
|
||
|
else
|
||
|
timeout = (u32)msecs_to_jiffies(args->timeout);
|
||
|
|
||
|
err = nvhost_syncpt_wait_timeout(&ctx->dev->syncpt, args->id,
|
||
|
args->thresh, timeout, &args->value,
|
||
|
NULL, true);
|
||
|
trace_nvhost_ioctl_ctrl_syncpt_wait(args->id, args->thresh,
|
||
|
args->timeout, args->value, err);
|
||
|
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
static int nvhost_ioctl_ctrl_syncpt_waitmex(struct nvhost_ctrl_userctx *ctx,
|
||
|
struct nvhost_ctrl_syncpt_waitmex_args *args)
|
||
|
{
|
||
|
u32 timeout;
|
||
|
int err;
|
||
|
struct nvhost_timespec nvts;
|
||
|
if (!nvhost_syncpt_is_valid_hw_pt_nospec(&ctx->dev->syncpt, &args->id))
|
||
|
return -EINVAL;
|
||
|
|
||
|
if (args->timeout == NVHOST_NO_TIMEOUT)
|
||
|
timeout = (u32)MAX_SCHEDULE_TIMEOUT;
|
||
|
else
|
||
|
timeout = (u32)msecs_to_jiffies(args->timeout);
|
||
|
|
||
|
err = nvhost_syncpt_wait_timeout(&ctx->dev->syncpt, args->id,
|
||
|
args->thresh, timeout, &args->value,
|
||
|
&nvts, true);
|
||
|
args->tv_sec = nvts.ts.tv_sec;
|
||
|
args->tv_nsec = nvts.ts.tv_nsec;
|
||
|
args->clock_id = nvts.clock;
|
||
|
trace_nvhost_ioctl_ctrl_syncpt_wait(args->id, args->thresh,
|
||
|
args->timeout, args->value, err);
|
||
|
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
static int nvhost_ioctl_ctrl_sync_fence_create(struct nvhost_ctrl_userctx *ctx,
|
||
|
struct nvhost_ctrl_sync_fence_create_args *args)
|
||
|
{
|
||
|
#ifdef CONFIG_TEGRA_GRHOST_SYNC
|
||
|
int err;
|
||
|
int i;
|
||
|
struct nvhost_ctrl_sync_fence_info *pts;
|
||
|
char name[32];
|
||
|
const char __user *args_name =
|
||
|
(const char __user *)(uintptr_t)args->name;
|
||
|
const void __user *args_pts =
|
||
|
(const void __user *)(uintptr_t)args->pts;
|
||
|
|
||
|
if (args_name) {
|
||
|
if (strncpy_from_user(name, args_name, sizeof(name)) < 0) {
|
||
|
nvhost_err(&ctx->dev->dev->dev,
|
||
|
"failed to copy from user: args_name=%px",
|
||
|
args_name);
|
||
|
return -EFAULT;
|
||
|
}
|
||
|
name[sizeof(name) - 1] = '\0';
|
||
|
} else {
|
||
|
name[0] = '\0';
|
||
|
}
|
||
|
|
||
|
pts = kmalloc_array(args->num_pts, sizeof(*pts), GFP_KERNEL);
|
||
|
if (!pts) {
|
||
|
nvhost_err(&ctx->dev->dev->dev, "failed to allocate pts");
|
||
|
return -ENOMEM;
|
||
|
}
|
||
|
|
||
|
/* Multiplication overflow would have errored in kmalloc_array */
|
||
|
if (copy_from_user(pts, args_pts, sizeof(*pts) * args->num_pts)) {
|
||
|
nvhost_err(&ctx->dev->dev->dev,
|
||
|
"failed to copy from user args_pts=%px", args_pts);
|
||
|
err = -EFAULT;
|
||
|
goto out;
|
||
|
}
|
||
|
|
||
|
for (i = 0; i < args->num_pts; i++) {
|
||
|
if (!nvhost_syncpt_is_valid_hw_pt(&ctx->dev->syncpt,
|
||
|
pts[i].id)) {
|
||
|
nvhost_err(&ctx->dev->dev->dev,
|
||
|
"invalid syncpoint id %u", pts[i].id);
|
||
|
err = -EINVAL;
|
||
|
goto out;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
err = nvhost_sync_create_fence_fd(ctx->dev->dev, pts, args->num_pts,
|
||
|
name, &args->fence_fd);
|
||
|
out:
|
||
|
kfree(pts);
|
||
|
return err;
|
||
|
#else
|
||
|
nvhost_err(&ctx->dev->dev->dev, "operation not supported");
|
||
|
return -EINVAL;
|
||
|
#endif
|
||
|
}
|
||
|
|
||
|
static int nvhost_ioctl_ctrl_sync_fence_set_name(
|
||
|
struct nvhost_ctrl_userctx *ctx,
|
||
|
struct nvhost_ctrl_sync_fence_name_args *args)
|
||
|
{
|
||
|
#ifdef CONFIG_TEGRA_GRHOST_SYNC
|
||
|
int err;
|
||
|
char name[32];
|
||
|
const char __user *args_name =
|
||
|
(const char __user *)(uintptr_t)args->name;
|
||
|
|
||
|
if (args_name) {
|
||
|
if (strncpy_from_user(name, args_name, sizeof(name)) < 0) {
|
||
|
nvhost_err(&ctx->dev->dev->dev,
|
||
|
"failed to copy from user: args_name=%px",
|
||
|
args_name);
|
||
|
return -EFAULT;
|
||
|
}
|
||
|
name[sizeof(name) - 1] = '\0';
|
||
|
} else {
|
||
|
name[0] = '\0';
|
||
|
}
|
||
|
|
||
|
err = nvhost_sync_fence_set_name(args->fence_fd, name);
|
||
|
return err;
|
||
|
#else
|
||
|
nvhost_err(&ctx->dev->dev->dev, "operation not supported");
|
||
|
return -EINVAL;
|
||
|
#endif
|
||
|
}
|
||
|
|
||
|
static int nvhost_ioctl_ctrl_module_mutex(struct nvhost_ctrl_userctx *ctx,
|
||
|
struct nvhost_ctrl_module_mutex_args *args)
|
||
|
{
|
||
|
int err = 0;
|
||
|
if (args->id >= nvhost_syncpt_nb_mlocks(&ctx->dev->syncpt) ||
|
||
|
args->lock > 1) {
|
||
|
nvhost_err(&ctx->dev->dev->dev, "invalid modmutex %u",
|
||
|
args->id);
|
||
|
return -EINVAL;
|
||
|
}
|
||
|
|
||
|
args->id = array_index_nospec(args->id,
|
||
|
nvhost_syncpt_nb_mlocks(&ctx->dev->syncpt));
|
||
|
|
||
|
trace_nvhost_ioctl_ctrl_module_mutex(args->lock, args->id);
|
||
|
if (args->lock && !ctx->mod_locks[args->id]) {
|
||
|
if (args->id == 0)
|
||
|
err = nvhost_module_busy(ctx->dev->dev);
|
||
|
else
|
||
|
err = nvhost_mutex_try_lock(&ctx->dev->syncpt,
|
||
|
args->id);
|
||
|
if (!err)
|
||
|
ctx->mod_locks[args->id] = 1;
|
||
|
} else if (!args->lock && ctx->mod_locks[args->id]) {
|
||
|
if (args->id == 0)
|
||
|
nvhost_module_idle(ctx->dev->dev);
|
||
|
else
|
||
|
nvhost_mutex_unlock(&ctx->dev->syncpt, args->id);
|
||
|
ctx->mod_locks[args->id] = 0;
|
||
|
}
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
static int nvhost_ioctl_ctrl_get_version(struct nvhost_ctrl_userctx *ctx,
|
||
|
struct nvhost_get_param_args *args)
|
||
|
{
|
||
|
args->value = NVHOST_SUBMIT_VERSION_MAX_SUPPORTED;
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static int nvhost_ioctl_ctrl_syncpt_read_max(struct nvhost_ctrl_userctx *ctx,
|
||
|
struct nvhost_ctrl_syncpt_read_args *args)
|
||
|
{
|
||
|
if (!nvhost_syncpt_is_valid_hw_pt(&ctx->dev->syncpt, args->id)) {
|
||
|
nvhost_err(&ctx->dev->dev->dev,
|
||
|
"invalid syncpoint id %u", args->id);
|
||
|
return -EINVAL;
|
||
|
}
|
||
|
|
||
|
args->id = array_index_nospec(args->id,
|
||
|
nvhost_syncpt_nb_hw_pts(&ctx->dev->syncpt));
|
||
|
|
||
|
args->value = nvhost_syncpt_read_max(&ctx->dev->syncpt, args->id);
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static int nvhost_ioctl_ctrl_get_characteristics(struct nvhost_ctrl_userctx *ctx,
|
||
|
struct nvhost_ctrl_get_characteristics *args)
|
||
|
{
|
||
|
struct nvhost_characteristics *nvhost_char = &ctx->dev->nvhost_char;
|
||
|
int err = 0;
|
||
|
|
||
|
if (args->nvhost_characteristics_buf_size > 0) {
|
||
|
size_t write_size = sizeof(*nvhost_char);
|
||
|
|
||
|
if (write_size > args->nvhost_characteristics_buf_size)
|
||
|
write_size = args->nvhost_characteristics_buf_size;
|
||
|
|
||
|
err = copy_to_user((void __user *)(uintptr_t)
|
||
|
args->nvhost_characteristics_buf_addr,
|
||
|
nvhost_char, write_size);
|
||
|
if (err) {
|
||
|
nvhost_err(&ctx->dev->dev->dev,
|
||
|
"failed to copy to user");
|
||
|
err = -EFAULT;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
if (err == 0)
|
||
|
args->nvhost_characteristics_buf_size = sizeof(*nvhost_char);
|
||
|
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
static int nvhost_ioctl_ctrl_check_module_support(
|
||
|
struct nvhost_ctrl_userctx *ctx,
|
||
|
struct nvhost_ctrl_check_module_support_args *args)
|
||
|
{
|
||
|
struct platform_device *pdev =
|
||
|
nvhost_device_list_match_by_id(args->module_id);
|
||
|
|
||
|
args->value = pdev != NULL;
|
||
|
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
struct nvhost_event_poll_fd_rec {
|
||
|
int fd;
|
||
|
bool event_posted;
|
||
|
struct mutex lock;
|
||
|
wait_queue_head_t wq;
|
||
|
struct kref ref;
|
||
|
};
|
||
|
|
||
|
static unsigned int nvhost_event_poll(struct file *filp, poll_table *wait)
|
||
|
{
|
||
|
struct nvhost_event_poll_fd_rec *private_data = filp->private_data;
|
||
|
unsigned int mask = 0;
|
||
|
|
||
|
poll_wait(filp, &private_data->wq, wait);
|
||
|
|
||
|
mutex_lock(&private_data->lock);
|
||
|
if (private_data->event_posted) {
|
||
|
mask = (POLLPRI | POLLIN);
|
||
|
private_data->event_posted = false;
|
||
|
}
|
||
|
mutex_unlock(&private_data->lock);
|
||
|
|
||
|
return mask;
|
||
|
}
|
||
|
|
||
|
static void nvhost_event_poll_free_priv_data(struct kref *ref)
|
||
|
{
|
||
|
struct nvhost_event_poll_fd_rec *private_data =
|
||
|
container_of(ref, struct nvhost_event_poll_fd_rec, ref);
|
||
|
|
||
|
kfree(private_data);
|
||
|
}
|
||
|
|
||
|
static int nvhost_event_poll_fd_release(struct inode *inode, struct file *filp)
|
||
|
{
|
||
|
struct nvhost_event_poll_fd_rec *private_data = filp->private_data;
|
||
|
kref_put(&private_data->ref, nvhost_event_poll_free_priv_data);
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static const struct file_operations nvhost_event_poll_fd_ops = {
|
||
|
.owner = THIS_MODULE,
|
||
|
.poll = nvhost_event_poll,
|
||
|
.release = nvhost_event_poll_fd_release,
|
||
|
};
|
||
|
|
||
|
static int nvhost_ioctl_ctrl_poll_fd_create (
|
||
|
struct nvhost_ctrl_userctx *ctx,
|
||
|
struct nvhost_ctrl_poll_fd_create_args *args)
|
||
|
{
|
||
|
struct nvhost_event_poll_fd_rec *private_data;
|
||
|
struct file *file;
|
||
|
char name[64];
|
||
|
int err;
|
||
|
int fd;
|
||
|
|
||
|
err = get_unused_fd_flags(O_RDWR);
|
||
|
if (err < 0) {
|
||
|
nvhost_err(&ctx->dev->dev->dev,
|
||
|
"failed to get unused fd");
|
||
|
return err;
|
||
|
}
|
||
|
fd = err;
|
||
|
|
||
|
snprintf(name, sizeof(name), "nvhost-event-poll-fd");
|
||
|
|
||
|
private_data = kzalloc(sizeof(*private_data), GFP_KERNEL);
|
||
|
if (!private_data) {
|
||
|
nvhost_err(&ctx->dev->dev->dev,
|
||
|
"failed to allocate private_data");
|
||
|
goto fail_alloc;
|
||
|
}
|
||
|
|
||
|
private_data->fd = fd;
|
||
|
init_waitqueue_head(&private_data->wq);
|
||
|
mutex_init(&private_data->lock);
|
||
|
private_data->event_posted = false;
|
||
|
kref_init(&private_data->ref);
|
||
|
|
||
|
file = anon_inode_getfile(name, &nvhost_event_poll_fd_ops,
|
||
|
private_data, O_RDWR);
|
||
|
if (IS_ERR(file)) {
|
||
|
nvhost_err(&ctx->dev->dev->dev, "failed to get file");
|
||
|
err = PTR_ERR(file);
|
||
|
goto fail_file;
|
||
|
}
|
||
|
fd_install(fd, file);
|
||
|
|
||
|
args->fd = fd;
|
||
|
return 0;
|
||
|
|
||
|
fail_file:
|
||
|
kfree(private_data);
|
||
|
fail_alloc:
|
||
|
put_unused_fd(fd);
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
static void nvhost_event_poll_update(void *priv, int nr_completed)
|
||
|
{
|
||
|
struct nvhost_event_poll_fd_rec *private_data =
|
||
|
(struct nvhost_event_poll_fd_rec *)priv;
|
||
|
|
||
|
mutex_lock(&private_data->lock);
|
||
|
private_data->event_posted = true;
|
||
|
mutex_unlock(&private_data->lock);
|
||
|
|
||
|
wake_up_interruptible(&private_data->wq);
|
||
|
|
||
|
kref_put(&private_data->ref, nvhost_event_poll_free_priv_data);
|
||
|
}
|
||
|
|
||
|
static int nvhost_ioctl_ctrl_poll_fd_trigger_event(
|
||
|
struct nvhost_ctrl_userctx *ctx,
|
||
|
struct nvhost_ctrl_poll_fd_trigger_event_args *args)
|
||
|
{
|
||
|
struct nvhost_event_poll_fd_rec *private_data;
|
||
|
struct file *f = fget(args->fd);
|
||
|
int err;
|
||
|
|
||
|
if (!f) {
|
||
|
nvhost_err(&ctx->dev->dev->dev, "failed to fget");
|
||
|
return -EINVAL;
|
||
|
}
|
||
|
|
||
|
if (f->f_op != &nvhost_event_poll_fd_ops) {
|
||
|
nvhost_err(&ctx->dev->dev->dev, "incorrect fops backend");
|
||
|
return -EINVAL;
|
||
|
}
|
||
|
|
||
|
if (!nvhost_syncpt_is_valid_hw_pt(&ctx->dev->syncpt, args->id)) {
|
||
|
nvhost_err(&ctx->dev->dev->dev,
|
||
|
"invalid syncpt id %u", args->id);
|
||
|
return -EINVAL;
|
||
|
}
|
||
|
|
||
|
private_data = (struct nvhost_event_poll_fd_rec *)f->private_data;
|
||
|
kref_get(&private_data->ref);
|
||
|
|
||
|
err = nvhost_intr_register_fast_notifier(ctx->dev->dev, args->id,
|
||
|
args->thresh, nvhost_event_poll_update, private_data);
|
||
|
if (err)
|
||
|
kref_put(&private_data->ref, nvhost_event_poll_free_priv_data);
|
||
|
|
||
|
fput(f);
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
static long nvhost_ctrlctl(struct file *filp,
|
||
|
unsigned int cmd, unsigned long arg)
|
||
|
{
|
||
|
struct nvhost_ctrl_userctx *priv = filp->private_data;
|
||
|
u8 buf[NVHOST_IOCTL_CTRL_MAX_ARG_SIZE] __aligned(sizeof(u64));
|
||
|
int err = 0;
|
||
|
|
||
|
if ((_IOC_TYPE(cmd) != NVHOST_IOCTL_MAGIC) ||
|
||
|
(_IOC_NR(cmd) == 0) ||
|
||
|
(_IOC_NR(cmd) > NVHOST_IOCTL_CTRL_LAST) ||
|
||
|
(_IOC_SIZE(cmd) > NVHOST_IOCTL_CTRL_MAX_ARG_SIZE)) {
|
||
|
nvhost_err_ratelimited(NULL, "invalid cmd 0x%x", cmd);
|
||
|
return -ENOIOCTLCMD;
|
||
|
}
|
||
|
|
||
|
if (_IOC_DIR(cmd) & _IOC_WRITE) {
|
||
|
if (copy_from_user(buf, (void __user *)arg, _IOC_SIZE(cmd))) {
|
||
|
nvhost_err(NULL, "failed to copy from user arg=%px",
|
||
|
(void __user *)arg);
|
||
|
return -EFAULT;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
switch (cmd) {
|
||
|
case NVHOST_IOCTL_CTRL_SYNCPT_READ:
|
||
|
err = nvhost_ioctl_ctrl_syncpt_read(priv, (void *)buf);
|
||
|
break;
|
||
|
case NVHOST_IOCTL_CTRL_SYNCPT_INCR:
|
||
|
err = nvhost_ioctl_ctrl_syncpt_incr(priv, (void *)buf);
|
||
|
break;
|
||
|
case NVHOST_IOCTL_CTRL_SYNCPT_WAIT:
|
||
|
err = nvhost_ioctl_ctrl_syncpt_waitex(priv, (void *)buf);
|
||
|
break;
|
||
|
case NVHOST32_IOCTL_CTRL_SYNC_FENCE_CREATE:
|
||
|
{
|
||
|
struct nvhost32_ctrl_sync_fence_create_args *args32 =
|
||
|
(struct nvhost32_ctrl_sync_fence_create_args *)buf;
|
||
|
struct nvhost_ctrl_sync_fence_create_args args;
|
||
|
args.name = args32->name;
|
||
|
args.pts = args32->pts;
|
||
|
args.num_pts = args32->num_pts;
|
||
|
err = nvhost_ioctl_ctrl_sync_fence_create(priv, &args);
|
||
|
args32->fence_fd = args.fence_fd;
|
||
|
break;
|
||
|
}
|
||
|
case NVHOST_IOCTL_CTRL_SYNC_FENCE_CREATE:
|
||
|
err = nvhost_ioctl_ctrl_sync_fence_create(priv, (void *)buf);
|
||
|
break;
|
||
|
case NVHOST_IOCTL_CTRL_SYNC_FENCE_SET_NAME:
|
||
|
err = nvhost_ioctl_ctrl_sync_fence_set_name(priv, (void *)buf);
|
||
|
break;
|
||
|
case NVHOST_IOCTL_CTRL_MODULE_MUTEX:
|
||
|
err = nvhost_ioctl_ctrl_module_mutex(priv, (void *)buf);
|
||
|
break;
|
||
|
case NVHOST_IOCTL_CTRL_SYNCPT_WAITEX:
|
||
|
err = nvhost_ioctl_ctrl_syncpt_waitex(priv, (void *)buf);
|
||
|
break;
|
||
|
case NVHOST_IOCTL_CTRL_GET_VERSION:
|
||
|
err = nvhost_ioctl_ctrl_get_version(priv, (void *)buf);
|
||
|
break;
|
||
|
case NVHOST_IOCTL_CTRL_SYNCPT_READ_MAX:
|
||
|
err = nvhost_ioctl_ctrl_syncpt_read_max(priv, (void *)buf);
|
||
|
break;
|
||
|
case NVHOST_IOCTL_CTRL_SYNCPT_WAITMEX:
|
||
|
err = nvhost_ioctl_ctrl_syncpt_waitmex(priv, (void *)buf);
|
||
|
break;
|
||
|
case NVHOST_IOCTL_CTRL_GET_CHARACTERISTICS:
|
||
|
err = nvhost_ioctl_ctrl_get_characteristics(priv, (void *)buf);
|
||
|
break;
|
||
|
case NVHOST_IOCTL_CTRL_CHECK_MODULE_SUPPORT:
|
||
|
err = nvhost_ioctl_ctrl_check_module_support(priv, (void *)buf);
|
||
|
break;
|
||
|
case NVHOST_IOCTL_CTRL_POLL_FD_CREATE:
|
||
|
err = nvhost_ioctl_ctrl_poll_fd_create(priv, (void *)buf);
|
||
|
break;
|
||
|
case NVHOST_IOCTL_CTRL_POLL_FD_TRIGGER_EVENT:
|
||
|
err = nvhost_ioctl_ctrl_poll_fd_trigger_event(priv, (void *)buf);
|
||
|
break;
|
||
|
default:
|
||
|
nvhost_err(&priv->dev->dev->dev, "invalid cmd 0x%x", cmd);
|
||
|
err = -ENOIOCTLCMD;
|
||
|
break;
|
||
|
}
|
||
|
|
||
|
if ((err == 0) && (_IOC_DIR(cmd) & _IOC_READ)) {
|
||
|
err = copy_to_user((void __user *)arg, buf, _IOC_SIZE(cmd));
|
||
|
if (err) {
|
||
|
nvhost_err(&priv->dev->dev->dev,
|
||
|
"failed to copy to user");
|
||
|
err = -EFAULT;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
static const struct file_operations nvhost_ctrlops = {
|
||
|
.owner = THIS_MODULE,
|
||
|
.release = nvhost_ctrlrelease,
|
||
|
.open = nvhost_ctrlopen,
|
||
|
.unlocked_ioctl = nvhost_ctrlctl,
|
||
|
#ifdef CONFIG_COMPAT
|
||
|
.compat_ioctl = nvhost_ctrlctl,
|
||
|
#endif
|
||
|
};
|
||
|
|
||
|
static int power_on_host(struct platform_device *dev)
|
||
|
{
|
||
|
struct nvhost_master *host = nvhost_get_private_data(dev);
|
||
|
struct nvhost_device_data *pdata = platform_get_drvdata(dev);
|
||
|
|
||
|
if (host_device_op().load_gating_regs)
|
||
|
host_device_op().load_gating_regs(dev, pdata->engine_can_cg);
|
||
|
|
||
|
nvhost_syncpt_reset(&host->syncpt);
|
||
|
nvhost_syncpt_initialize_unused(&host->syncpt);
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static int power_off_host(struct platform_device *dev)
|
||
|
{
|
||
|
struct nvhost_master *host = nvhost_get_private_data(dev);
|
||
|
|
||
|
nvhost_channel_suspend(host);
|
||
|
nvhost_syncpt_save(&host->syncpt);
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
#ifdef CONFIG_PM
|
||
|
static int enable_irq_host(struct platform_device *dev)
|
||
|
{
|
||
|
struct nvhost_device_data *pdata = platform_get_drvdata(dev);
|
||
|
struct nvhost_master *host = nvhost_get_private_data(dev);
|
||
|
return nvhost_intr_start(&host->intr, clk_get_rate(pdata->clk[0]));
|
||
|
}
|
||
|
|
||
|
static int disable_irq_host(struct platform_device *dev)
|
||
|
{
|
||
|
struct nvhost_master *host = nvhost_get_private_data(dev);
|
||
|
return nvhost_intr_stop(&host->intr);
|
||
|
}
|
||
|
#endif
|
||
|
|
||
|
int nvhost_gather_filter_enabled(struct nvhost_syncpt *sp)
|
||
|
{
|
||
|
/*
|
||
|
* Keep gather filter always enabled
|
||
|
* We still need this API to inform this to user space
|
||
|
*/
|
||
|
return 1;
|
||
|
}
|
||
|
|
||
|
static int alloc_syncpts_per_apps(struct nvhost_syncpt *sp)
|
||
|
{
|
||
|
struct nvhost_master *host = syncpt_to_dev(sp);
|
||
|
return host->info.syncpt_policy == SYNCPT_PER_CHANNEL_INSTANCE;
|
||
|
}
|
||
|
|
||
|
static int alloc_chs_per_submits(struct nvhost_syncpt *sp)
|
||
|
{
|
||
|
struct nvhost_master *host = syncpt_to_dev(sp);
|
||
|
return host->info.channel_policy == MAP_CHANNEL_ON_SUBMIT;
|
||
|
}
|
||
|
|
||
|
static ssize_t nvhost_syncpt_capability_show(struct kobject *kobj,
|
||
|
struct kobj_attribute *attr, char *buf)
|
||
|
{
|
||
|
struct nvhost_capability_node *node =
|
||
|
container_of(attr, struct nvhost_capability_node, attr);
|
||
|
|
||
|
return snprintf(buf, PAGE_SIZE, "%u\n",
|
||
|
node->read_func(&node->host->syncpt));
|
||
|
}
|
||
|
|
||
|
static ssize_t nvhost_syncpt_capability_store(struct kobject *kobj,
|
||
|
struct kobj_attribute *attr, const char *buf, size_t count)
|
||
|
{
|
||
|
struct nvhost_capability_node *node =
|
||
|
container_of(attr, struct nvhost_capability_node, attr);
|
||
|
|
||
|
node->store_func(&node->host->syncpt, buf);
|
||
|
return count;
|
||
|
}
|
||
|
|
||
|
static inline int nvhost_set_sysfs_capability_node(
|
||
|
struct nvhost_master *host, const char *name,
|
||
|
struct nvhost_capability_node *node,
|
||
|
int (*read_func)(struct nvhost_syncpt *sp),
|
||
|
int (*store_func)(struct nvhost_syncpt *sp,
|
||
|
const char *buf))
|
||
|
{
|
||
|
node->attr.attr.name = name;
|
||
|
node->attr.attr.mode = S_IRUGO;
|
||
|
node->attr.show = nvhost_syncpt_capability_show;
|
||
|
node->read_func = read_func;
|
||
|
if (store_func) {
|
||
|
node->attr.store = nvhost_syncpt_capability_store;
|
||
|
node->store_func = store_func;
|
||
|
node->attr.attr.mode = (S_IRWXU|S_IRGRP|S_IROTH);
|
||
|
}
|
||
|
node->host = host;
|
||
|
sysfs_attr_init(&node->attr.attr);
|
||
|
|
||
|
return sysfs_create_file(host->caps_kobj, &node->attr.attr);
|
||
|
}
|
||
|
|
||
|
static int nvhost_user_init(struct nvhost_master *host)
|
||
|
{
|
||
|
struct nvhost_device_data *pdata = platform_get_drvdata(host->dev);
|
||
|
dev_t devno;
|
||
|
int err;
|
||
|
|
||
|
pdata->debug_dump_device = NULL;
|
||
|
|
||
|
host->nvhost_class = class_create(THIS_MODULE,
|
||
|
dev_name(&host->dev->dev));
|
||
|
if (IS_ERR(host->nvhost_class)) {
|
||
|
err = PTR_ERR(host->nvhost_class);
|
||
|
dev_err(&host->dev->dev, "failed to create class\n");
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
err = alloc_chrdev_region(&devno, 0, 1, IFACE_NAME);
|
||
|
if (err < 0) {
|
||
|
dev_err(&host->dev->dev, "failed to reserve chrdev region\n");
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
host->major = MAJOR(devno);
|
||
|
host->next_minor += 1;
|
||
|
mutex_init(&host->chrdev_mutex);
|
||
|
|
||
|
cdev_init(&host->cdev, &nvhost_ctrlops);
|
||
|
host->cdev.owner = THIS_MODULE;
|
||
|
err = cdev_add(&host->cdev, devno, 1);
|
||
|
if (err < 0) {
|
||
|
nvhost_err(&host->dev->dev, "failed to add cdev");
|
||
|
goto fail;
|
||
|
}
|
||
|
host->ctrl = device_create(host->nvhost_class, &host->dev->dev, devno,
|
||
|
NULL, IFACE_NAME "-ctrl");
|
||
|
if (IS_ERR(host->ctrl)) {
|
||
|
err = PTR_ERR(host->ctrl);
|
||
|
dev_err(&host->dev->dev, "failed to create ctrl device\n");
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
host->caps_nodes = devm_kzalloc(&host->dev->dev,
|
||
|
sizeof(struct nvhost_capability_node) * 7, GFP_KERNEL);
|
||
|
if (!host->caps_nodes) {
|
||
|
nvhost_err(&host->dev->dev, "failed to allocate caps_nodes");
|
||
|
err = -ENOMEM;
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
host->caps_kobj = kobject_create_and_add("capabilities",
|
||
|
&host->dev->dev.kobj);
|
||
|
if (!host->caps_kobj) {
|
||
|
nvhost_err(&host->dev->dev,
|
||
|
"failed to create capabilities dir");
|
||
|
err = -EIO;
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
if (nvhost_set_sysfs_capability_node(host, num_syncpts_name,
|
||
|
host->caps_nodes, &nvhost_syncpt_nb_pts,
|
||
|
&nvhost_nb_syncpts_store)) {
|
||
|
err = -EIO;
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
if (nvhost_set_sysfs_capability_node(host, num_mutexes_name,
|
||
|
host->caps_nodes + 1, &nvhost_syncpt_nb_mlocks, NULL)) {
|
||
|
err = -EIO;
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
if (nvhost_set_sysfs_capability_node(host,
|
||
|
gather_filter_enabled_name, host->caps_nodes + 2,
|
||
|
nvhost_gather_filter_enabled, NULL)) {
|
||
|
err = -EIO;
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
if (nvhost_set_sysfs_capability_node(host,
|
||
|
alloc_chs_per_submits_name, host->caps_nodes + 3,
|
||
|
alloc_chs_per_submits, NULL)) {
|
||
|
err = -EIO;
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
if (nvhost_set_sysfs_capability_node(host,
|
||
|
alloc_syncpts_per_apps_name, host->caps_nodes + 4,
|
||
|
alloc_syncpts_per_apps, NULL)) {
|
||
|
err = -EIO;
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
if (nvhost_set_sysfs_capability_node(host, syncpts_pts_limit_name,
|
||
|
host->caps_nodes + 5, &nvhost_syncpt_pts_limit, NULL)) {
|
||
|
err = -EIO;
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
if (nvhost_set_sysfs_capability_node(host, syncpts_pts_base_name,
|
||
|
host->caps_nodes + 6, &nvhost_syncpt_pts_base, NULL)) {
|
||
|
err = -EIO;
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
/* create a symlink for host1x if it is not under platform bus or
|
||
|
* it has been created with different name */
|
||
|
|
||
|
if ((host->dev->dev.parent != &platform_bus) ||
|
||
|
strcmp(dev_name(&host->dev->dev), "host1x")) {
|
||
|
err = sysfs_create_link_nowarn(&platform_bus.kobj,
|
||
|
&host->dev->dev.kobj,
|
||
|
"host1x");
|
||
|
if (err) {
|
||
|
err = sysfs_create_link(&platform_bus.kobj,
|
||
|
&host->dev->dev.kobj,
|
||
|
dev_name(&host->dev->dev));
|
||
|
if (err)
|
||
|
dev_warn(&host->dev->dev, "could not create sysfs links\n");
|
||
|
}
|
||
|
}
|
||
|
|
||
|
return 0;
|
||
|
fail:
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
void nvhost_set_chanops(struct nvhost_channel *ch)
|
||
|
{
|
||
|
host_device_op().set_nvhost_chanops(ch);
|
||
|
}
|
||
|
static void nvhost_free_resources(struct nvhost_master *host)
|
||
|
{
|
||
|
kfree(host->intr.syncpt);
|
||
|
host->intr.syncpt = NULL;
|
||
|
}
|
||
|
|
||
|
static int nvhost_alloc_resources(struct nvhost_master *host)
|
||
|
{
|
||
|
int err;
|
||
|
|
||
|
err = nvhost_init_chip_support(host);
|
||
|
if (err)
|
||
|
return err;
|
||
|
|
||
|
host->intr.syncpt = kzalloc(sizeof(struct nvhost_intr_syncpt) *
|
||
|
nvhost_syncpt_nb_hw_pts(&host->syncpt),
|
||
|
GFP_KERNEL);
|
||
|
|
||
|
if (!host->intr.syncpt) {
|
||
|
nvhost_err(&host->dev->dev, "failed to allocate intr.syncpt");
|
||
|
/* frees happen in the support removal phase */
|
||
|
return -ENOMEM;
|
||
|
}
|
||
|
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static struct of_device_id tegra_host1x_of_match[] = {
|
||
|
#ifdef TEGRA_12X_OR_HIGHER_CONFIG
|
||
|
{ .compatible = "nvidia,tegra124-host1x",
|
||
|
.data = (struct nvhost_device_data *)&t124_host1x_info },
|
||
|
#endif
|
||
|
#ifdef TEGRA_21X_OR_HIGHER_CONFIG
|
||
|
{ .compatible = "nvidia,tegra210-host1x",
|
||
|
.data = (struct nvhost_device_data *)&t21_host1x_info },
|
||
|
#endif
|
||
|
#ifdef CONFIG_ARCH_TEGRA_18x_SOC
|
||
|
{ .name = "host1x",
|
||
|
.compatible = "nvidia,tegra186-host1x",
|
||
|
.data = (struct nvhost_device_data *)&t18_host1x_info },
|
||
|
{ .name = "host1xb",
|
||
|
.compatible = "nvidia,tegra186-host1x",
|
||
|
.data = (struct nvhost_device_data *)&t18_host1xb_info },
|
||
|
{ .compatible = "nvidia,tegra186-host1x-cl34000094",
|
||
|
.data = (struct nvhost_device_data *)&t18_host1x_info },
|
||
|
{ .name = "host1x",
|
||
|
.compatible = "nvidia,tegra186-host1x-hv",
|
||
|
.data = (struct nvhost_device_data *)&t18_host1x_hv_info },
|
||
|
#endif
|
||
|
#ifdef CONFIG_TEGRA_T19X_GRHOST
|
||
|
{ .name = "host1x",
|
||
|
.compatible = "nvidia,tegra194-host1x",
|
||
|
.data = (struct nvhost_device_data *)&t19_host1x_info },
|
||
|
{ .name = "host1x",
|
||
|
.compatible = "nvidia,tegra194-host1x-hv",
|
||
|
.data = (struct nvhost_device_data *)&t19_host1x_hv_info },
|
||
|
#endif
|
||
|
{ },
|
||
|
};
|
||
|
|
||
|
int nvhost_host1x_finalize_poweron(struct platform_device *dev)
|
||
|
{
|
||
|
return power_on_host(dev);
|
||
|
}
|
||
|
|
||
|
int nvhost_host1x_prepare_poweroff(struct platform_device *dev)
|
||
|
{
|
||
|
return power_off_host(dev);
|
||
|
}
|
||
|
|
||
|
static int of_nvhost_parse_platform_data(struct platform_device *dev,
|
||
|
struct nvhost_device_data *pdata)
|
||
|
{
|
||
|
struct device_node *np = dev->dev.of_node;
|
||
|
struct nvhost_master *host = nvhost_get_host(dev);
|
||
|
u32 value;
|
||
|
|
||
|
if (!of_property_read_u32(np, "virtual-dev", &value)) {
|
||
|
if (value)
|
||
|
pdata->virtual_dev = true;
|
||
|
}
|
||
|
|
||
|
if (!of_property_read_u32(np, "vmserver-owns-engines", &value)) {
|
||
|
if (value)
|
||
|
host->info.vmserver_owns_engines = true;
|
||
|
}
|
||
|
|
||
|
if (!of_property_read_u32(np, "nvidia,ch-base", &value))
|
||
|
host->info.ch_base = value;
|
||
|
|
||
|
if (!of_property_read_u32(np, "nvidia,nb-channels", &value))
|
||
|
host->info.nb_channels = value;
|
||
|
|
||
|
host->info.ch_limit = host->info.ch_base + host->info.nb_channels;
|
||
|
|
||
|
if (!of_property_read_u32(np, "nvidia,nb-hw-pts", &value))
|
||
|
host->info.nb_hw_pts = value;
|
||
|
|
||
|
if (!of_property_read_u32(np, "nvidia,pts-base", &value))
|
||
|
host->info.pts_base = value;
|
||
|
|
||
|
if (!of_property_read_u32(np, "nvidia,nb-pts", &value))
|
||
|
host->info.nb_pts = value;
|
||
|
|
||
|
if (host->info.nb_pts > host->info.nb_hw_pts) {
|
||
|
nvhost_err(&dev->dev, "invalid nb_pts property");
|
||
|
return -EINVAL;
|
||
|
}
|
||
|
|
||
|
host->info.pts_limit = host->info.pts_base + host->info.nb_pts;
|
||
|
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
long linsim_cl = 0;
|
||
|
|
||
|
int nvhost_update_characteristics(struct platform_device *dev)
|
||
|
{
|
||
|
struct nvhost_master *host = nvhost_get_host(dev);
|
||
|
|
||
|
if (nvhost_gather_filter_enabled(&host->syncpt))
|
||
|
host->nvhost_char.flags |= NVHOST_CHARACTERISTICS_GFILTER;
|
||
|
|
||
|
if (host->info.channel_policy == MAP_CHANNEL_ON_SUBMIT)
|
||
|
host->nvhost_char.flags |=
|
||
|
NVHOST_CHARACTERISTICS_RESOURCE_PER_CHANNEL_INSTANCE;
|
||
|
|
||
|
host->nvhost_char.flags |= NVHOST_CHARACTERISTICS_SUPPORT_PREFENCES;
|
||
|
|
||
|
host->nvhost_char.num_mlocks = host->info.nb_mlocks;
|
||
|
host->nvhost_char.num_syncpts = host->info.nb_pts;
|
||
|
host->nvhost_char.syncpts_base = host->info.pts_base;
|
||
|
host->nvhost_char.syncpts_limit = host->info.pts_limit;
|
||
|
host->nvhost_char.num_hw_pts = host->info.nb_hw_pts;
|
||
|
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
struct nvhost_master *nvhost_get_syncpt_owner(u32 id)
|
||
|
{
|
||
|
struct nvhost_master *host;
|
||
|
|
||
|
mutex_lock(&nvhost_masters_lock);
|
||
|
list_for_each_entry(host, &nvhost_masters, list)
|
||
|
if (nvhost_syncpt_is_valid_pt(&host->syncpt, id)) {
|
||
|
mutex_unlock(&nvhost_masters_lock);
|
||
|
return host;
|
||
|
}
|
||
|
mutex_unlock(&nvhost_masters_lock);
|
||
|
return NULL;
|
||
|
}
|
||
|
|
||
|
struct nvhost_syncpt *nvhost_get_syncpt_owner_struct(u32 id,
|
||
|
struct nvhost_syncpt *default_syncpt)
|
||
|
{
|
||
|
struct nvhost_master *host = nvhost_get_syncpt_owner(id);
|
||
|
|
||
|
if (!host)
|
||
|
return default_syncpt;
|
||
|
|
||
|
return &host->syncpt;
|
||
|
}
|
||
|
|
||
|
static int nvhost_device_get_host1x_resources_byname(struct nvhost_master *host)
|
||
|
{
|
||
|
int i;
|
||
|
void __iomem *regs = NULL;
|
||
|
struct platform_device *dev = host->dev;
|
||
|
struct nvhost_device_data *pdata = platform_get_drvdata(dev);
|
||
|
int ret;
|
||
|
|
||
|
for (i = 0; i < host->info.nb_resources; ++i) {
|
||
|
struct resource *r = NULL;
|
||
|
|
||
|
r = platform_get_resource_byname(dev, IORESOURCE_MEM,
|
||
|
host->info.resources[i]);
|
||
|
/* Get next resource without failing if current resource is not
|
||
|
* present. Resources vary with configurations.
|
||
|
*/
|
||
|
if (!r)
|
||
|
continue;
|
||
|
|
||
|
regs = devm_ioremap_resource(&dev->dev, r);
|
||
|
if (IS_ERR(regs)) {
|
||
|
nvhost_err(&host->dev->dev,
|
||
|
"failed to ioremap resources %s",
|
||
|
host->info.resources[i]);
|
||
|
ret = PTR_ERR(regs);
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
pdata->aperture[i] = regs;
|
||
|
}
|
||
|
|
||
|
return 0;
|
||
|
fail:
|
||
|
dev_err(&dev->dev, "failed to get register memory\n");
|
||
|
return -ENXIO;
|
||
|
}
|
||
|
|
||
|
/* Verify that we can access syncpts at either end of range. If we fail,
|
||
|
* most likely PCT and DT have conflicting information.
|
||
|
*/
|
||
|
static int check_syncpt_range(struct platform_device *dev,
|
||
|
struct nvhost_syncpt *sp)
|
||
|
{
|
||
|
struct nvhost_master *host = syncpt_to_dev(sp);
|
||
|
int err = 0;
|
||
|
u32 val;
|
||
|
|
||
|
nvhost_syncpt_set_minval(dev, host->info.pts_base, 0xdeadbeef);
|
||
|
val = nvhost_syncpt_update_min(sp, host->info.pts_base);
|
||
|
if (val != 0xdeadbeef) {
|
||
|
nvhost_err(&dev->dev, "syncpt %u: read 0x%x", host->info.pts_base, val);
|
||
|
err = -EINVAL;
|
||
|
goto fail;
|
||
|
}
|
||
|
nvhost_syncpt_set_minval(dev, host->info.pts_base, 0);
|
||
|
|
||
|
nvhost_syncpt_set_minval(dev, host->info.pts_limit - 1, 0xbeefdead);
|
||
|
val = nvhost_syncpt_update_min(sp, host->info.pts_limit - 1);
|
||
|
if (val != 0xbeefdead) {
|
||
|
nvhost_err(&dev->dev, "syncpt %u: read 0x%x", host->info.pts_limit - 1, val);
|
||
|
err = -EINVAL;
|
||
|
goto fail;
|
||
|
}
|
||
|
nvhost_syncpt_set_minval(dev, host->info.pts_limit - 1, 0);
|
||
|
fail:
|
||
|
WARN_ON(err != 0);
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
static int nvhost_probe(struct platform_device *dev)
|
||
|
{
|
||
|
struct nvhost_master *host;
|
||
|
int syncpt_irq, generic_irq;
|
||
|
int err;
|
||
|
struct nvhost_device_data *pdata = NULL;
|
||
|
|
||
|
if (dev->dev.of_node) {
|
||
|
const struct of_device_id *match;
|
||
|
|
||
|
match = of_match_device(tegra_host1x_of_match, &dev->dev);
|
||
|
if (match) {
|
||
|
char *substr = strstr(match->compatible, "cl");
|
||
|
|
||
|
pdata = (struct nvhost_device_data *)match->data;
|
||
|
|
||
|
if (substr) {
|
||
|
substr += 2;
|
||
|
if (kstrtol(substr, 0, &linsim_cl))
|
||
|
WARN(1, "Unable to decode linsim cl\n");
|
||
|
}
|
||
|
}
|
||
|
} else
|
||
|
pdata = (struct nvhost_device_data *)dev->dev.platform_data;
|
||
|
|
||
|
WARN_ON(!pdata);
|
||
|
if (!pdata) {
|
||
|
dev_info(&dev->dev, "no platform data\n");
|
||
|
return -ENODATA;
|
||
|
}
|
||
|
|
||
|
syncpt_irq = platform_get_irq(dev, 0);
|
||
|
if (syncpt_irq < 0) {
|
||
|
dev_err(&dev->dev, "missing syncpt irq\n");
|
||
|
return -ENXIO;
|
||
|
}
|
||
|
|
||
|
generic_irq = platform_get_irq(dev, 1);
|
||
|
if (generic_irq < 0) {
|
||
|
dev_err(&dev->dev, "missing generic irq\n");
|
||
|
generic_irq = 0;
|
||
|
}
|
||
|
|
||
|
host = devm_kzalloc(&dev->dev, sizeof(*host), GFP_KERNEL);
|
||
|
if (!host) {
|
||
|
nvhost_err(NULL, "failed to alloc host structure");
|
||
|
return -ENOMEM;
|
||
|
}
|
||
|
|
||
|
host->dev = dev;
|
||
|
INIT_LIST_HEAD(&host->static_mappings_list);
|
||
|
INIT_LIST_HEAD(&host->vm_list);
|
||
|
host->syncpt_backing_head = RB_ROOT;
|
||
|
mutex_init(&host->vm_mutex);
|
||
|
mutex_init(&host->vm_alloc_mutex);
|
||
|
mutex_init(&pdata->lock);
|
||
|
init_rwsem(&pdata->busy_lock);
|
||
|
|
||
|
/* Copy host1x parameters. The private_data gets replaced
|
||
|
* by nvhost_master later */
|
||
|
memcpy(&host->info, pdata->private_data,
|
||
|
sizeof(struct host1x_device_info));
|
||
|
|
||
|
pdata->pdev = dev;
|
||
|
|
||
|
dma_set_mask_and_coherent(&dev->dev, host->info.dma_mask);
|
||
|
|
||
|
/* set common host1x device data */
|
||
|
platform_set_drvdata(dev, pdata);
|
||
|
|
||
|
/* set private host1x device data */
|
||
|
nvhost_set_private_data(dev, host);
|
||
|
|
||
|
err = of_nvhost_parse_platform_data(dev, pdata);
|
||
|
if (err) {
|
||
|
dev_err(&dev->dev, "error in parsing DT properties, err:%d",
|
||
|
err);
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
if ((pdata->virtual_dev || host->info.vmserver_owns_engines) &&
|
||
|
!vhost_init_done) {
|
||
|
err = nvhost_virt_init(dev, NVHOST_MODULE_NONE);
|
||
|
if (err) {
|
||
|
dev_err(&dev->dev, "failed to init virt support\n");
|
||
|
goto fail;
|
||
|
}
|
||
|
vhost_init_done = true;
|
||
|
}
|
||
|
|
||
|
if (!pdata->virtual_dev) {
|
||
|
if (host->info.nb_resources)
|
||
|
err = nvhost_device_get_host1x_resources_byname(host);
|
||
|
else
|
||
|
err = nvhost_device_get_resources(dev);
|
||
|
if (err) {
|
||
|
dev_err(&dev->dev, "failed to get resources\n");
|
||
|
goto fail;
|
||
|
}
|
||
|
host->aperture = pdata->aperture[0];
|
||
|
}
|
||
|
|
||
|
err = nvhost_alloc_resources(host);
|
||
|
if (err) {
|
||
|
dev_err(&dev->dev, "failed to init chip support\n");
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
err = nvhost_syncpt_init(dev, &host->syncpt);
|
||
|
if (err)
|
||
|
goto fail;
|
||
|
|
||
|
err = nvhost_user_init(host);
|
||
|
if (err)
|
||
|
goto fail;
|
||
|
|
||
|
err = nvhost_module_init(dev);
|
||
|
if (err)
|
||
|
goto fail;
|
||
|
|
||
|
device_set_wakeup_capable(&dev->dev, true);
|
||
|
|
||
|
mutex_init(&host->timeout_mutex);
|
||
|
|
||
|
err = nvhost_module_busy(dev);
|
||
|
if (err)
|
||
|
goto fail;
|
||
|
|
||
|
err = power_on_host(dev);
|
||
|
if (err) {
|
||
|
nvhost_module_idle(dev);
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
err = check_syncpt_range(dev, &host->syncpt);
|
||
|
if (err)
|
||
|
goto fail;
|
||
|
|
||
|
err = nvhost_intr_init(&host->intr, generic_irq, syncpt_irq);
|
||
|
if (err)
|
||
|
goto fail;
|
||
|
|
||
|
err = nvhost_alloc_channels(host);
|
||
|
if (err) {
|
||
|
nvhost_module_idle(dev);
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
if (tegra_cpu_is_asim() || pdata->virtual_dev)
|
||
|
/* for simulation & virtualization, use a fake clock rate */
|
||
|
err = nvhost_intr_start(&host->intr, 12000000);
|
||
|
else
|
||
|
err = nvhost_intr_start(&host->intr,
|
||
|
clk_get_rate(pdata->clk[0]));
|
||
|
|
||
|
if (err) {
|
||
|
nvhost_module_idle(dev);
|
||
|
dev_err(&dev->dev, "Failed to start intr=%d\n", err);
|
||
|
goto fail;
|
||
|
}
|
||
|
|
||
|
nvhost_device_list_init();
|
||
|
|
||
|
/* Disable timeouts in simulation */
|
||
|
if (tegra_platform_is_vdk())
|
||
|
pdata->nvhost_timeout_default = 0;
|
||
|
else
|
||
|
pdata->nvhost_timeout_default =
|
||
|
CONFIG_TEGRA_GRHOST_DEFAULT_TIMEOUT;
|
||
|
|
||
|
nvhost_debug_init(host);
|
||
|
|
||
|
nvhost_module_idle(dev);
|
||
|
|
||
|
nvhost_update_characteristics(dev);
|
||
|
|
||
|
nvhost_vm_init(dev);
|
||
|
|
||
|
INIT_LIST_HEAD(&host->list);
|
||
|
mutex_lock(&nvhost_masters_lock);
|
||
|
list_add_tail(&host->list, &nvhost_masters);
|
||
|
mutex_unlock(&nvhost_masters_lock);
|
||
|
|
||
|
dev_info(&dev->dev, "initialized\n");
|
||
|
return 0;
|
||
|
|
||
|
fail:
|
||
|
nvhost_virt_deinit(dev);
|
||
|
nvhost_free_resources(host);
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
static int __exit nvhost_remove(struct platform_device *dev)
|
||
|
{
|
||
|
struct nvhost_master *host = nvhost_get_private_data(dev);
|
||
|
nvhost_intr_deinit(&host->intr);
|
||
|
nvhost_syncpt_deinit(&host->syncpt);
|
||
|
nvhost_virt_deinit(dev);
|
||
|
nvhost_free_resources(host);
|
||
|
#ifdef CONFIG_PM
|
||
|
pm_runtime_put(&dev->dev);
|
||
|
pm_runtime_disable(&dev->dev);
|
||
|
#else
|
||
|
nvhost_module_disable_clk(&dev->dev);
|
||
|
#endif
|
||
|
nvhost_channel_list_free(host);
|
||
|
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
#ifdef CONFIG_PM
|
||
|
|
||
|
/*
|
||
|
* FIXME: Genpd disables the runtime pm while preparing for system
|
||
|
* suspend. As host1x clients may need host1x in suspend sequence
|
||
|
* for register reads/writes or syncpoint increments, we need to
|
||
|
* re-enable pm_runtime. As we *must* balance pm_runtime counter,
|
||
|
* we drop the reference in suspend complete callback, right before
|
||
|
* the genpd re-enables runtime pm.
|
||
|
*
|
||
|
* We should revisit the power code as this is hacky and
|
||
|
* caused by the way we use power domains.
|
||
|
*/
|
||
|
|
||
|
static int nvhost_suspend_prepare(struct device *dev)
|
||
|
{
|
||
|
if (LINUX_VERSION_CODE < KERNEL_VERSION(4, 9, 0) && dev->pm_domain)
|
||
|
pm_runtime_enable(dev);
|
||
|
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static void nvhost_suspend_complete(struct device *dev)
|
||
|
{
|
||
|
if (LINUX_VERSION_CODE < KERNEL_VERSION(4, 9, 0) && dev->pm_domain)
|
||
|
__pm_runtime_disable(dev, false);
|
||
|
}
|
||
|
|
||
|
static int nvhost_suspend(struct device *dev)
|
||
|
{
|
||
|
struct platform_device *pdev = to_platform_device(dev);
|
||
|
int err = 0;
|
||
|
|
||
|
nvhost_module_enable_clk(dev);
|
||
|
disable_irq_host(pdev);
|
||
|
power_off_host(pdev);
|
||
|
nvhost_module_disable_clk(dev);
|
||
|
|
||
|
err = vhost_suspend(pdev);
|
||
|
if (err)
|
||
|
dev_err(dev, "Failed to do vhost suspend %d\n", err);
|
||
|
else
|
||
|
dev_info(dev, "suspended\n");
|
||
|
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
static int nvhost_resume(struct device *dev)
|
||
|
{
|
||
|
struct platform_device *pdev = to_platform_device(dev);
|
||
|
struct nvhost_master *host = nvhost_get_host(pdev);
|
||
|
int index, err = 0;
|
||
|
|
||
|
err = vhost_resume(pdev);
|
||
|
if (err) {
|
||
|
dev_err(dev, "Failed to do vhost resume %d", err);
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
nvhost_module_enable_clk(dev);
|
||
|
power_on_host(pdev);
|
||
|
err = enable_irq_host(pdev);
|
||
|
if (err) {
|
||
|
dev_err(dev, "Failed to enable host irq=%d\n", err);
|
||
|
return err;
|
||
|
}
|
||
|
|
||
|
for (index = 0; index < nvhost_channel_nb_channels(host); index++) {
|
||
|
/* reinitialise gather filter for each channel */
|
||
|
nvhost_channel_init_gather_filter(host->dev,
|
||
|
host->chlist[index]);
|
||
|
}
|
||
|
|
||
|
nvhost_module_disable_clk(dev);
|
||
|
|
||
|
dev_info(dev, "resuming\n");
|
||
|
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
static const struct dev_pm_ops host1x_pm_ops = {
|
||
|
.prepare = nvhost_suspend_prepare,
|
||
|
.complete = nvhost_suspend_complete,
|
||
|
.suspend_late = nvhost_suspend,
|
||
|
.resume_early = nvhost_resume,
|
||
|
SET_RUNTIME_PM_OPS(nvhost_module_disable_clk,
|
||
|
nvhost_module_enable_clk,
|
||
|
NULL)
|
||
|
};
|
||
|
#endif /* CONFIG_PM */
|
||
|
|
||
|
static struct platform_driver platform_driver = {
|
||
|
.probe = nvhost_probe,
|
||
|
.remove = __exit_p(nvhost_remove),
|
||
|
.driver = {
|
||
|
.owner = THIS_MODULE,
|
||
|
.name = DRIVER_NAME,
|
||
|
#ifdef CONFIG_PM
|
||
|
.pm = &host1x_pm_ops,
|
||
|
#endif
|
||
|
#ifdef CONFIG_OF
|
||
|
.of_match_table = tegra_host1x_of_match,
|
||
|
#endif
|
||
|
.suppress_bind_attrs = true,
|
||
|
},
|
||
|
};
|
||
|
|
||
|
static struct of_device_id tegra_host1x_domain_match[] = {
|
||
|
{.compatible = "nvidia,tegra124-host1x-pd",
|
||
|
.data = (struct nvhost_device_data *)&t124_host1x_info},
|
||
|
{.compatible = "nvidia,tegra132-host1x-pd",
|
||
|
.data = (struct nvhost_device_data *)&t124_host1x_info},
|
||
|
#ifdef TEGRA_21X_OR_HIGHER_CONFIG
|
||
|
{.compatible = "nvidia,tegra210-host1x-pd",
|
||
|
.data = (struct nvhost_device_data *)&t21_host1x_info},
|
||
|
#endif
|
||
|
#ifdef CONFIG_ARCH_TEGRA_18x_SOC
|
||
|
{.compatible = "nvidia,tegra186-host1x-pd",
|
||
|
.data = (struct nvhost_device_data *)&t18_host1x_info},
|
||
|
{.compatible = "nvidia,tegra186-host1x-pd-hv",
|
||
|
.data = (struct nvhost_device_data *)&t18_host1x_hv_info},
|
||
|
#endif
|
||
|
{},
|
||
|
};
|
||
|
|
||
|
static int __init nvhost_mod_init(void)
|
||
|
{
|
||
|
int ret;
|
||
|
|
||
|
ret = nvhost_domain_init(tegra_host1x_domain_match);
|
||
|
if (ret)
|
||
|
return ret;
|
||
|
|
||
|
return platform_driver_register(&platform_driver);
|
||
|
}
|
||
|
|
||
|
static void __exit nvhost_mod_exit(void)
|
||
|
{
|
||
|
platform_driver_unregister(&platform_driver);
|
||
|
}
|
||
|
|
||
|
/* host1x master device needs nvmap to be instantiated first.
|
||
|
* nvmap is instantiated via fs_initcall.
|
||
|
* Hence instantiate host1x master device using rootfs_initcall
|
||
|
* which is one level after fs_initcall. */
|
||
|
rootfs_initcall(nvhost_mod_init);
|
||
|
module_exit(nvhost_mod_exit);
|
||
|
|