2292 lines
64 KiB
C
2292 lines
64 KiB
C
/*
|
|
* SPI driver for NVIDIA's Tegra124 SPI Controller.
|
|
*
|
|
* Copyright (c) 2013-2018, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms and conditions of the GNU General Public License,
|
|
* version 2, as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include <linux/clk.h>
|
|
#include <linux/reset.h>
|
|
#include <linux/completion.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/dmaengine.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/dmapool.h>
|
|
#include <linux/err.h>
|
|
#include <linux/of_gpio.h>
|
|
#include <linux/init.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/io.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/kthread.h>
|
|
#include <linux/module.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/pm_runtime.h>
|
|
#include <linux/of.h>
|
|
#include <linux/of_device.h>
|
|
#include <linux/spi/spi.h>
|
|
#include <linux/spi/spi-tegra.h>
|
|
#include <linux/clk/tegra.h>
|
|
|
|
#define SPI_COMMAND1 0x000
|
|
#define SPI_BIT_LENGTH(x) (((x) & 0x1f) << 0)
|
|
#define SPI_PACKED (1 << 5)
|
|
#define SPI_TX_EN (1 << 11)
|
|
#define SPI_RX_EN (1 << 12)
|
|
#define SPI_BOTH_EN_BYTE (1 << 13)
|
|
#define SPI_BOTH_EN_BIT (1 << 14)
|
|
#define SPI_LSBYTE_FE (1 << 15)
|
|
#define SPI_LSBIT_FE (1 << 16)
|
|
#define SPI_BIDIROE (1 << 17)
|
|
#define SPI_IDLE_SDA_DRIVE_LOW (0 << 18)
|
|
#define SPI_IDLE_SDA_DRIVE_HIGH (1 << 18)
|
|
#define SPI_IDLE_SDA_PULL_LOW (2 << 18)
|
|
#define SPI_IDLE_SDA_PULL_HIGH (3 << 18)
|
|
#define SPI_IDLE_SDA_MASK (3 << 18)
|
|
#define SPI_CS_SS_VAL (1 << 20)
|
|
#define SPI_CS_SW_HW (1 << 21)
|
|
/* SPI_CS_POL_INACTIVE bits are default high */
|
|
#define SPI_CS_POL_INACTIVE 22
|
|
#define SPI_CS_POL_INACTIVE_0 (1 << 22)
|
|
#define SPI_CS_POL_INACTIVE_1 (1 << 23)
|
|
#define SPI_CS_POL_INACTIVE_2 (1 << 24)
|
|
#define SPI_CS_POL_INACTIVE_3 (1 << 25)
|
|
#define SPI_CS_POL_INACTIVE_MASK (0xF << 22)
|
|
|
|
#define SPI_CS_SEL_0 (0 << 26)
|
|
#define SPI_CS_SEL_1 (1 << 26)
|
|
#define SPI_CS_SEL_2 (2 << 26)
|
|
#define SPI_CS_SEL_3 (3 << 26)
|
|
#define SPI_CS_SEL_MASK (3 << 26)
|
|
#define SPI_CS_SEL(x) (((x) & 0x3) << 26)
|
|
#define SPI_CS(x) (((x) >> 26) & 0x3)
|
|
#define SPI_CONTROL_MODE_0 (0 << 28)
|
|
#define SPI_CONTROL_MODE_1 (1 << 28)
|
|
#define SPI_CONTROL_MODE_2 (2 << 28)
|
|
#define SPI_CONTROL_MODE_3 (3 << 28)
|
|
#define SPI_CONTROL_MODE_MASK (3 << 28)
|
|
#define SPI_MODE_SEL(x) (((x) & 0x3) << 28)
|
|
#define SPI_MODE(x) (((x) >> 28) & 0x3)
|
|
#define SPI_M_S (1 << 30)
|
|
#define SPI_PIO (1 << 31)
|
|
|
|
#define SPI_CS_TIMING2 0x00C
|
|
|
|
#define SPI_TRANS_STATUS 0x010
|
|
#define SPI_BLK_CNT(val) (((val) >> 0) & 0xFFFF)
|
|
#define SPI_SLV_IDLE_COUNT(val) (((val) >> 16) & 0xFF)
|
|
#define SPI_RDY (1 << 30)
|
|
|
|
#define SPI_FIFO_STATUS 0x014
|
|
#define SPI_RX_FIFO_EMPTY (1 << 0)
|
|
#define SPI_RX_FIFO_FULL (1 << 1)
|
|
#define SPI_TX_FIFO_EMPTY (1 << 2)
|
|
#define SPI_TX_FIFO_FULL (1 << 3)
|
|
#define SPI_RX_FIFO_UNF (1 << 4)
|
|
#define SPI_RX_FIFO_OVF (1 << 5)
|
|
#define SPI_TX_FIFO_UNF (1 << 6)
|
|
#define SPI_TX_FIFO_OVF (1 << 7)
|
|
#define SPI_ERR (1 << 8)
|
|
#define SPI_TX_FIFO_FLUSH (1 << 14)
|
|
#define SPI_RX_FIFO_FLUSH (1 << 15)
|
|
#define SPI_TX_FIFO_EMPTY_COUNT(val) (((val) >> 16) & 0x7F)
|
|
#define SPI_RX_FIFO_FULL_COUNT(val) (((val) >> 23) & 0x7F)
|
|
#define SPI_FRAME_END (1 << 30)
|
|
#define SPI_CS_INACTIVE (1 << 31)
|
|
#define IS_SPI_CS_INACTIVE(val) (val >> 31)
|
|
#define SPI_FIFO_ERROR (SPI_RX_FIFO_UNF | \
|
|
SPI_RX_FIFO_OVF | SPI_TX_FIFO_UNF | SPI_TX_FIFO_OVF)
|
|
#define SPI_FIFO_EMPTY (SPI_RX_FIFO_EMPTY | SPI_TX_FIFO_EMPTY)
|
|
|
|
#define SPI_TX_DATA 0x018
|
|
#define SPI_RX_DATA 0x01C
|
|
|
|
#define SPI_DMA_CTL 0x020
|
|
#define SPI_TX_TRIG_1 (0 << 15)
|
|
#define SPI_TX_TRIG_4 (1 << 15)
|
|
#define SPI_TX_TRIG_8 (2 << 15)
|
|
#define SPI_TX_TRIG_16 (3 << 15)
|
|
#define SPI_TX_TRIG_MASK (3 << 15)
|
|
#define SPI_TX_TRIG(val) (((val) >> 15) & 0x3)
|
|
#define SPI_RX_TRIG_1 (0 << 19)
|
|
#define SPI_RX_TRIG_4 (1 << 19)
|
|
#define SPI_RX_TRIG_8 (2 << 19)
|
|
#define SPI_RX_TRIG_16 (3 << 19)
|
|
#define SPI_RX_TRIG_MASK (3 << 19)
|
|
#define SPI_RX_TRIG(val) (((val) >> 19) & 0x3)
|
|
#define SPI_IE_TX (1 << 28)
|
|
#define SPI_IE_RX (1 << 29)
|
|
#define SPI_CONT (1 << 30)
|
|
#define SPI_DMA (1 << 31)
|
|
#define SPI_DMA_EN SPI_DMA
|
|
|
|
#define SPI_DMA_BLK 0x024
|
|
#define SPI_DMA_BLK_SET(x) (((x) & 0xFFFF) << 0)
|
|
|
|
#define SPI_TX_FIFO 0x108
|
|
#define SPI_RX_FIFO 0x188
|
|
|
|
#define SPI_INTR_MASK 0x18c
|
|
#define SPI_INTR_RX_FIFO_UNF_MASK (1 << 25)
|
|
#define SPI_INTR_RX_FIFO_OVF_MASK (1 << 26)
|
|
#define SPI_INTR_TX_FIFO_UNF_MASK (1 << 27)
|
|
#define SPI_INTR_TX_FIFO_OVF_MASK (1 << 28)
|
|
#define SPI_INTR_RDY_MASK (1 << 29)
|
|
#define SPI_INTR_FRAME_END_MASK (1 << 30)
|
|
#define SPI_INTR_CS_MASK (1 << 31)
|
|
|
|
#define SPI_MISC_REG 0x194
|
|
#define SPI_MISC_EXT_CLK_EN (1 << 30)
|
|
#define SPI_MISC_CLKEN_OVERRIDE (1 << 31)
|
|
|
|
#define MAX_CHIP_SELECT 4
|
|
#define SPI_FIFO_DEPTH 64
|
|
#define DATA_DIR_TX (1 << 0)
|
|
#define DATA_DIR_RX (1 << 1)
|
|
|
|
#define SPI_DMA_TIMEOUT (msecs_to_jiffies(1000))
|
|
#define DEFAULT_SPI_DMA_BUF_LEN (256*1024)
|
|
#define TX_FIFO_EMPTY_COUNT_MAX SPI_TX_FIFO_EMPTY_COUNT(0x40)
|
|
#define RX_FIFO_FULL_COUNT_ZERO SPI_RX_FIFO_FULL_COUNT(0)
|
|
#define MAX_HOLD_CYCLES 16
|
|
#define SPI_DEFAULT_SPEED 25000000
|
|
|
|
#define SPI_FIFO_DEPTH 64
|
|
#define SPI_FIFO_FLUSH_MAX_DELAY 2000
|
|
#define MAX_PACKETS (1 << 16)
|
|
/* multiplication facotr for slave timeout, See code below */
|
|
#define DELAY_MUL_FACTOR 1000
|
|
|
|
#define PROFILE_SPI_SLAVE /* profile spi sw overhead */
|
|
#define VERBOSE_DUMP_REGS /* register dumps */
|
|
#define TEGRA_SPI_SLAVE_DEBUG /* to enable debug interfaces, sysfs ... */
|
|
|
|
#define dump_xfer(_t, _x) \
|
|
dev_dbg(((_t))->dev, "%s%s len:%d bpw:%d %dus %dHz", \
|
|
(_x)->rx_buf ? "Rx" : "", (_x)->tx_buf ? "Tx" : "", \
|
|
(_x)->len, (_x)->bits_per_word & 0xff, \
|
|
(_x)->delay_usecs & 0xffff, (_x)->speed_hz);
|
|
|
|
#define dump_sw_state(_t, _f, ...) \
|
|
do { \
|
|
dev_dbg((_t)->dev, _f, __VA_ARGS__); \
|
|
dev_dbg((_t)->dev, "[SW] cntrlr: %dHz bypw:%d dma?:%d dir?:%d " \
|
|
"dmasz:%d rx-trig:%d vlen?:%d pack?:%d force_unp?:%d\n",\
|
|
(_t)->cur_speed, (_t)->bytes_per_word, \
|
|
(_t)->is_curr_dma_xfer, (_t)->cur_direction, \
|
|
(_t)->curr_dma_words, (_t)->rx_trig_words, \
|
|
(_t)->variable_length_transfer, (_t)->is_packed, \
|
|
(_t)->force_unpacked_mode); \
|
|
} while (0)
|
|
|
|
|
|
#ifdef VERBOSE_DUMP_REGS
|
|
#define dump_regs(_log, _t, _f, ...) \
|
|
do { \
|
|
uint32_t fifo, cmd, dma_ctl, dma_blk, trans; \
|
|
fifo = tegra_spi_readl((_t), SPI_FIFO_STATUS); \
|
|
cmd = tegra_spi_readl((_t), SPI_COMMAND1); \
|
|
dma_ctl = tegra_spi_readl((_t), SPI_DMA_CTL); \
|
|
dma_blk = tegra_spi_readl((_t), SPI_DMA_BLK); \
|
|
trans = tegra_spi_readl((_t), SPI_TRANS_STATUS); \
|
|
dev_##_log(&((_t)->master->dev), _f, ##__VA_ARGS__); \
|
|
dev_##_log(&((_t)->master->dev), \
|
|
" CMD[%08x]: %s %s M%d CS%d [%c%c%c%c] %s " \
|
|
"%s %s %s %s %db " \
|
|
"TRANS[%08x]:%s I:%d B:%d\n" \
|
|
" FIFO[%08x]:RxF:%d TxE:%d Err[%s%s%s] " \
|
|
"RxSTA[%s%s%s%s] TxSTA[%s%s%s%s]" \
|
|
"DMA[%08x]:%s %s %s %s " \
|
|
"RxTr:%d TxTr:%d B:%d\n", \
|
|
/* command */ \
|
|
cmd, \
|
|
cmd & SPI_PIO ? "Go" : "", \
|
|
cmd & SPI_M_S ? "Ma" : "Sl", \
|
|
SPI_MODE(cmd), SPI_CS(cmd), \
|
|
(cmd & SPI_CS_POL_INACTIVE_0) ? 'H' : 'L', \
|
|
(cmd & SPI_CS_POL_INACTIVE_1) ? 'H' : 'L', \
|
|
(cmd & SPI_CS_POL_INACTIVE_2) ? 'H' : 'L', \
|
|
(cmd & SPI_CS_POL_INACTIVE_3) ? 'H' : 'L', \
|
|
(cmd & SPI_LSBYTE_FE) ? "LSB" : "MSB", \
|
|
(cmd & SPI_LSBIT_FE) ? "LSb" : "MSb", \
|
|
(cmd & SPI_RX_EN) ? "Rx" : "", \
|
|
(cmd & SPI_TX_EN) ? "Tx" : "", \
|
|
(cmd & SPI_PACKED) ? "Pa" : "Un", \
|
|
(SPI_BIT_LENGTH(cmd) + 1), \
|
|
/* transfer status */ \
|
|
trans, \
|
|
(trans & SPI_RDY) ? "RDY" : "BSY", \
|
|
SPI_SLV_IDLE_COUNT(trans), \
|
|
SPI_BLK_CNT(trans), \
|
|
/* fifo */ \
|
|
fifo, \
|
|
SPI_RX_FIFO_FULL_COUNT(fifo), \
|
|
SPI_TX_FIFO_EMPTY_COUNT(fifo), \
|
|
(fifo & SPI_CS_INACTIVE) ? "Cs" : "", \
|
|
(fifo & SPI_FRAME_END) ? "Fe" : "" , \
|
|
(fifo & SPI_FIFO_ERROR) ? "Er" : "" , \
|
|
(fifo & SPI_RX_FIFO_OVF) ? "O" : "" , \
|
|
(fifo & SPI_RX_FIFO_UNF) ? "U" : "" , \
|
|
(fifo & SPI_RX_FIFO_FULL) ? "F" : "" , \
|
|
(fifo & SPI_RX_FIFO_EMPTY) ? "E" : "" , \
|
|
(fifo & SPI_TX_FIFO_OVF) ? "O" : "" , \
|
|
(fifo & SPI_TX_FIFO_UNF) ? "U" : "" , \
|
|
(fifo & SPI_TX_FIFO_FULL) ? "F" : "" , \
|
|
(fifo & SPI_TX_FIFO_EMPTY) ? "E" : "" , \
|
|
/* dma ctl */ \
|
|
dma_ctl, \
|
|
(dma_ctl & SPI_DMA_EN) ? "De" : "", \
|
|
(dma_ctl & SPI_CONT) ? "Co" : "", \
|
|
(dma_ctl & SPI_IE_RX) ? "RxIE" : "", \
|
|
(dma_ctl & SPI_IE_TX) ? "TxIE" : "", \
|
|
SPI_RX_TRIG(dma_ctl), \
|
|
SPI_TX_TRIG(dma_ctl), \
|
|
/* dma blk */ \
|
|
SPI_DMA_BLK_SET(dma_blk)); \
|
|
} while (0)
|
|
#else
|
|
#define dump_regs(_log, _t, _f, ...) \
|
|
dev_##_log(&((_t)->master->dev), _f, ##__VA_ARGS__);
|
|
#endif
|
|
|
|
#define print_fifo_word(_m, _t, _c, _x) \
|
|
dev_dbg((_t)->dev, \
|
|
"%s[%d] %02lx %02lx %02lx %02lx\n", \
|
|
(_m), (_c), \
|
|
(_x) & 0xff, ((_x) >> 8) & 0xff, \
|
|
((_x) >> 16) & 0xff, ((_x) >> 24) & 0xff);
|
|
|
|
/* convert size in fifo bytes to number of packets */
|
|
/* for unpacked, each packet occupies 4 bytes, i.e., a fifo word
|
|
* for packed the each fifo word can be occupied by multiple packets
|
|
* as number of bytes in each packet.
|
|
*/
|
|
#define fifo_bytes_to_packets(_t, _sz) \
|
|
((_t)->is_packed ? (_sz)/(_t)->bytes_per_word : (_sz)/4);
|
|
|
|
struct tegra_spi_chip_data {
|
|
bool intr_mask_reg;
|
|
bool mask_cs_inactive_intr;
|
|
bool new_features;
|
|
};
|
|
|
|
struct tegra_spi_data {
|
|
struct device *dev;
|
|
struct spi_master *master;
|
|
spinlock_t lock;
|
|
|
|
struct clk *clk;
|
|
struct reset_control *rstc;
|
|
void __iomem *base;
|
|
phys_addr_t phys;
|
|
unsigned irq;
|
|
bool clock_always_on;
|
|
u32 spi_max_frequency;
|
|
u32 cur_speed;
|
|
unsigned min_div;
|
|
|
|
struct spi_device *cur_spi;
|
|
unsigned words_per_32bit;
|
|
/* bypw, rounding to nearest byte, irrespective of packed/unpacked */
|
|
unsigned bytes_per_word;
|
|
unsigned curr_dma_words;
|
|
unsigned cur_direction;
|
|
|
|
unsigned dma_buf_size;
|
|
unsigned max_buf_size;
|
|
bool is_curr_dma_xfer;
|
|
bool variable_length_transfer;
|
|
|
|
/* Slave Ready Polarity (true: Active High, false: Active Low) */
|
|
int gpio_slave_ready;
|
|
bool slave_ready_active_high;
|
|
|
|
struct completion rx_dma_complete;
|
|
struct completion tx_dma_complete;
|
|
|
|
u32 tx_status;
|
|
u32 rx_status;
|
|
bool reset_ctrl_status;
|
|
u32 status_reg;
|
|
bool is_packed;
|
|
|
|
u32 command1_reg;
|
|
u32 dma_control_reg;
|
|
u32 def_command1_reg;
|
|
|
|
struct completion xfer_completion;
|
|
struct spi_transfer *curr_xfer;
|
|
struct dma_chan *rx_dma_chan;
|
|
u32 *rx_dma_buf;
|
|
dma_addr_t rx_dma_phys;
|
|
struct dma_async_tx_descriptor *rx_dma_desc;
|
|
|
|
struct dma_chan *tx_dma_chan;
|
|
u32 *tx_dma_buf;
|
|
dma_addr_t tx_dma_phys;
|
|
struct dma_async_tx_descriptor *tx_dma_desc;
|
|
const struct tegra_spi_chip_data *chip_data;
|
|
struct tegra_spi_device_controller_data cdata[MAX_CHIP_SELECT];
|
|
|
|
spi_callback spi_slave_ready_callback;
|
|
spi_callback spi_slave_isr_callback;
|
|
void *client_data;
|
|
atomic_t isr_expected;
|
|
u32 words_to_transfer;
|
|
int curr_rx_pos;
|
|
/* common variable for tx and rx to update the actual length */
|
|
int curr_pos;
|
|
int rx_dma_len;
|
|
int rem_len;
|
|
int rx_trig_words;
|
|
int force_unpacked_mode;
|
|
bool lsbyte_first;
|
|
#ifdef PROFILE_SPI_SLAVE
|
|
ktime_t start_time;
|
|
ktime_t end_time;
|
|
int profile_size;
|
|
#endif
|
|
};
|
|
|
|
|
|
|
|
static int tegra_spi_runtime_suspend(struct device *dev);
|
|
static int tegra_spi_runtime_resume(struct device *dev);
|
|
static int tegra_spi_validate_request(struct spi_device *spi,
|
|
struct tegra_spi_data *tspi, struct spi_transfer *t);
|
|
static int tegra_spi_ext_clk_enable(bool enable, struct tegra_spi_data *tspi);
|
|
|
|
#ifdef TEGRA_SPI_SLAVE_DEBUG
|
|
static ssize_t force_unpacked_mode_set(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
struct tegra_spi_data *tspi;
|
|
struct spi_master *master = dev_get_drvdata(dev);
|
|
if (master) {
|
|
tspi = spi_master_get_devdata(master);
|
|
if (tspi && count) {
|
|
tspi->force_unpacked_mode = ((buf[0] - '0') > 0);
|
|
return count;
|
|
}
|
|
}
|
|
return -ENODEV;
|
|
}
|
|
|
|
static ssize_t force_unpacked_mode_show(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct tegra_spi_data *tspi;
|
|
struct spi_master *master = dev_get_drvdata(dev);
|
|
if (master) {
|
|
tspi = spi_master_get_devdata(master);
|
|
return sprintf(buf, "%d", tspi->force_unpacked_mode);
|
|
}
|
|
return -ENODEV;
|
|
}
|
|
|
|
static DEVICE_ATTR(force_unpacked_mode, 0644, force_unpacked_mode_show,
|
|
force_unpacked_mode_set);
|
|
#endif
|
|
|
|
static inline unsigned long tegra_spi_readl(struct tegra_spi_data *tspi,
|
|
unsigned long reg)
|
|
{
|
|
return readl(tspi->base + reg);
|
|
}
|
|
|
|
static inline void tegra_spi_writel(struct tegra_spi_data *tspi,
|
|
unsigned long val, unsigned long reg)
|
|
{
|
|
writel(val, tspi->base + reg);
|
|
|
|
/* TODO: remove the forced fence read below */
|
|
/* Read back register to make sure that register writes completed */
|
|
if (reg != SPI_TX_FIFO)
|
|
readl(tspi->base + SPI_COMMAND1);
|
|
}
|
|
|
|
static inline void tegra_spi_fence(struct tegra_spi_data *tspi)
|
|
{
|
|
/* Read back register to make sure that register writes completed */
|
|
readl(tspi->base + SPI_COMMAND1);
|
|
}
|
|
|
|
/* FIXME: the name of the function */
|
|
int tegra124_spi_slave_register_callback(struct spi_device *spi,
|
|
spi_callback func_ready, spi_callback func_isr, void *client_data)
|
|
{
|
|
struct tegra_spi_data *tspi = spi_master_get_devdata(spi->master);
|
|
|
|
/* Expect atleast one callback function. */
|
|
if ((!tspi) || (!func_ready && !func_isr))
|
|
return -EINVAL;
|
|
|
|
tspi->spi_slave_ready_callback = func_ready;
|
|
tspi->spi_slave_isr_callback = func_isr;
|
|
|
|
tspi->client_data = client_data;
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(tegra124_spi_slave_register_callback);
|
|
|
|
static void tegra_spi_clear_status(struct tegra_spi_data *tspi)
|
|
{
|
|
unsigned long val;
|
|
|
|
/* Write 1 to clear status register */
|
|
val = tegra_spi_readl(tspi, SPI_TRANS_STATUS);
|
|
tegra_spi_writel(tspi, val, SPI_TRANS_STATUS);
|
|
|
|
if (tspi->chip_data->intr_mask_reg) {
|
|
val = tegra_spi_readl(tspi, SPI_INTR_MASK);
|
|
if (!(val & SPI_INTR_RDY_MASK)) {
|
|
val |= (SPI_INTR_CS_MASK |
|
|
SPI_INTR_FRAME_END_MASK |
|
|
SPI_INTR_RDY_MASK |
|
|
SPI_INTR_RX_FIFO_UNF_MASK |
|
|
SPI_INTR_TX_FIFO_UNF_MASK |
|
|
SPI_INTR_RX_FIFO_OVF_MASK |
|
|
SPI_INTR_TX_FIFO_OVF_MASK);
|
|
tegra_spi_writel(tspi, val, SPI_INTR_MASK);
|
|
}
|
|
}
|
|
|
|
/* Clear fifo status error if any */
|
|
val = tegra_spi_readl(tspi, SPI_FIFO_STATUS);
|
|
tegra_spi_writel(tspi, val &
|
|
(SPI_FRAME_END | SPI_CS_INACTIVE | SPI_ERR |
|
|
SPI_TX_FIFO_UNF | SPI_TX_FIFO_OVF |
|
|
SPI_RX_FIFO_OVF | SPI_RX_FIFO_UNF),
|
|
SPI_FIFO_STATUS);
|
|
}
|
|
|
|
static void reset_controller(struct tegra_spi_data *tspi)
|
|
{
|
|
if (!tspi->reset_ctrl_status)
|
|
return;
|
|
|
|
/* TODO: Debug the reset sequence */
|
|
if (tspi->is_curr_dma_xfer &&
|
|
(tspi->cur_direction & DATA_DIR_TX))
|
|
dmaengine_terminate_all(tspi->tx_dma_chan);
|
|
if (tspi->is_curr_dma_xfer &&
|
|
(tspi->cur_direction & DATA_DIR_RX))
|
|
dmaengine_terminate_all(tspi->rx_dma_chan);
|
|
|
|
wmb(); /* barrier for dma terminate to happen */
|
|
reset_control_reset(tspi->rstc);
|
|
/* restore default value */
|
|
tegra_spi_writel(tspi, tspi->def_command1_reg, SPI_COMMAND1);
|
|
/* set CS inactive b/w packets to mask CS_INACTIVE interrupts */
|
|
if (tspi->chip_data->mask_cs_inactive_intr)
|
|
tegra_spi_writel(tspi, 0, SPI_CS_TIMING2);
|
|
tegra_spi_clear_status(tspi);
|
|
tegra_spi_writel(tspi, tspi->dma_control_reg, SPI_DMA_CTL);
|
|
dump_regs(dbg, tspi, "after controller reset");
|
|
tspi->reset_ctrl_status = false;
|
|
}
|
|
|
|
static unsigned tegra_spi_calculate_curr_xfer_param(
|
|
struct spi_device *spi, struct tegra_spi_data *tspi,
|
|
struct spi_transfer *t)
|
|
{
|
|
unsigned remain_len = t->len;
|
|
unsigned max_word;
|
|
unsigned bits_per_word;
|
|
unsigned max_len;
|
|
unsigned total_fifo_words;
|
|
|
|
bits_per_word = t->bits_per_word ? t->bits_per_word :
|
|
spi->bits_per_word;
|
|
tspi->bytes_per_word = (bits_per_word - 1) / 8 + 1;
|
|
|
|
if (!tspi->force_unpacked_mode &&
|
|
(bits_per_word == 8 || bits_per_word == 16)) {
|
|
tspi->is_packed = 1;
|
|
tspi->words_per_32bit = 32/bits_per_word;
|
|
} else {
|
|
tspi->is_packed = 0;
|
|
tspi->words_per_32bit = 1;
|
|
}
|
|
|
|
if (tspi->is_packed) {
|
|
max_len = min(remain_len, tspi->max_buf_size);
|
|
tspi->curr_dma_words = max_len/tspi->bytes_per_word;
|
|
total_fifo_words = (max_len + 3)/4;
|
|
} else {
|
|
max_word = (remain_len - 1) / tspi->bytes_per_word + 1;
|
|
max_word = min(max_word, tspi->max_buf_size/4);
|
|
tspi->curr_dma_words = max_word;
|
|
total_fifo_words = max_word;
|
|
}
|
|
return total_fifo_words;
|
|
}
|
|
|
|
static unsigned tegra_spi_fill_tx_fifo_from_client_txbuf(
|
|
struct tegra_spi_data *tspi, struct spi_transfer *t)
|
|
{
|
|
unsigned nbytes;
|
|
unsigned tx_empty_count;
|
|
unsigned max_n_32bit;
|
|
unsigned i, count;
|
|
unsigned long x;
|
|
unsigned int written_words;
|
|
unsigned fifo_words_left;
|
|
u8 *tx_buf = (u8 *)t->tx_buf;
|
|
|
|
tx_empty_count = SPI_TX_FIFO_EMPTY_COUNT(
|
|
tegra_spi_readl(tspi, SPI_FIFO_STATUS));
|
|
|
|
if (tspi->is_packed) {
|
|
fifo_words_left = tx_empty_count * tspi->words_per_32bit;
|
|
written_words = min(fifo_words_left, tspi->curr_dma_words);
|
|
nbytes = written_words * tspi->bytes_per_word;
|
|
max_n_32bit = DIV_ROUND_UP(nbytes, 4);
|
|
for (count = 0; count < max_n_32bit; count++) {
|
|
x = 0;
|
|
for (i = 0; (i < 4) && nbytes; i++, nbytes--)
|
|
x |= (*tx_buf++) << (i*8);
|
|
|
|
print_fifo_word("TxFIFO", tspi, count, x);
|
|
tegra_spi_writel(tspi, x, SPI_TX_FIFO);
|
|
}
|
|
} else {
|
|
max_n_32bit = min(tspi->curr_dma_words, tx_empty_count);
|
|
written_words = max_n_32bit;
|
|
nbytes = written_words * tspi->bytes_per_word;
|
|
for (count = 0; count < max_n_32bit; count++) {
|
|
x = 0;
|
|
for (i = 0; nbytes && (i < tspi->bytes_per_word);
|
|
i++, nbytes--)
|
|
x |= ((*tx_buf++) << i*8);
|
|
print_fifo_word("TxFIFO", tspi, count, x);
|
|
tegra_spi_writel(tspi, x, SPI_TX_FIFO);
|
|
}
|
|
}
|
|
return written_words;
|
|
}
|
|
|
|
/* FIFO mode rx data copy from SPI-FIFO to client buffer */
|
|
static unsigned int tegra_spi_read_rx_fifo_to_client_rxbuf(
|
|
struct tegra_spi_data *tspi, struct spi_transfer *t)
|
|
{
|
|
unsigned rx_full_count;
|
|
unsigned i, count;
|
|
unsigned long x;
|
|
unsigned recv_len, npackets;
|
|
u8 *rx_buf = (u8 *)t->rx_buf + tspi->curr_rx_pos;
|
|
|
|
rx_full_count = SPI_RX_FIFO_FULL_COUNT(
|
|
tegra_spi_readl(tspi, SPI_FIFO_STATUS));
|
|
|
|
/* actual bytes to be copied */
|
|
npackets = fifo_bytes_to_packets(tspi, rx_full_count*4);
|
|
npackets = min(npackets, tspi->words_to_transfer);
|
|
recv_len = npackets * tspi->bytes_per_word;
|
|
|
|
if (tspi->is_packed) {
|
|
int j = recv_len;
|
|
if (rx_full_count != DIV_ROUND_UP(tspi->words_to_transfer *
|
|
tspi->bytes_per_word, 4))
|
|
dump_regs(dbg, tspi,
|
|
"fifo-cnt[%d] != trans-cnt[%d]",
|
|
rx_full_count, tspi->words_to_transfer);
|
|
|
|
for (count = 0; j && count < rx_full_count; count++) {
|
|
x = tegra_spi_readl(tspi, SPI_RX_FIFO);
|
|
print_fifo_word("RxFIFO", tspi, count, x);
|
|
for (i = 0; j && (i < 4); i++, j--)
|
|
*rx_buf++ = (x >> i*8) & 0xFF;
|
|
}
|
|
} else {
|
|
unsigned int bits_per_word;
|
|
|
|
bits_per_word = t->bits_per_word ? t->bits_per_word :
|
|
tspi->cur_spi->bits_per_word;
|
|
if (rx_full_count != tspi->words_to_transfer)
|
|
dump_regs(dbg, tspi,
|
|
"fifo-cnt[%d] != trans-cnt[%d]",
|
|
rx_full_count, tspi->words_to_transfer);
|
|
|
|
rx_full_count = min(rx_full_count, tspi->words_to_transfer);
|
|
for (count = 0; count < rx_full_count; count++) {
|
|
x = tegra_spi_readl(tspi, SPI_RX_FIFO);
|
|
print_fifo_word("RxFIFO", tspi, count, x);
|
|
for (i = 0; (i < tspi->bytes_per_word); i++)
|
|
*rx_buf++ = (x >> (i*8)) & 0xFF;
|
|
}
|
|
}
|
|
tspi->words_to_transfer -= npackets;
|
|
tspi->rem_len -= recv_len;
|
|
tspi->curr_rx_pos += recv_len;
|
|
return recv_len;
|
|
}
|
|
|
|
static void tegra_spi_copy_client_txbuf_to_spi_txbuf(
|
|
struct tegra_spi_data *tspi, struct spi_transfer *t)
|
|
{
|
|
unsigned len;
|
|
|
|
/* Make the dma buffer to read by cpu */
|
|
dma_sync_single_for_cpu(tspi->dev, tspi->tx_dma_phys,
|
|
tspi->dma_buf_size, DMA_TO_DEVICE);
|
|
|
|
if (tspi->is_packed) {
|
|
len = tspi->curr_dma_words * tspi->bytes_per_word;
|
|
memcpy(tspi->tx_dma_buf, t->tx_buf, len);
|
|
} else {
|
|
unsigned int i;
|
|
unsigned int count;
|
|
u8 *tx_buf = (u8 *)t->tx_buf;
|
|
unsigned consume = tspi->curr_dma_words * tspi->bytes_per_word;
|
|
unsigned int x;
|
|
|
|
for (count = 0; count < tspi->curr_dma_words; count++) {
|
|
x = 0;
|
|
for (i = 0; consume && (i < tspi->bytes_per_word);
|
|
i++, consume--)
|
|
x |= ((*tx_buf++) << i * 8);
|
|
tspi->tx_dma_buf[count] = x;
|
|
}
|
|
}
|
|
|
|
/* Make the dma buffer to read by dma */
|
|
dma_sync_single_for_device(tspi->dev, tspi->tx_dma_phys,
|
|
tspi->dma_buf_size, DMA_TO_DEVICE);
|
|
}
|
|
|
|
static unsigned int tegra_spi_copy_spi_rxbuf_to_client_rxbuf(
|
|
struct tegra_spi_data *tspi, struct spi_transfer *t)
|
|
{
|
|
unsigned dma_bytes, npackets;
|
|
|
|
/* Make the dma buffer to read by cpu */
|
|
dma_sync_single_for_cpu(tspi->dev, tspi->rx_dma_phys,
|
|
tspi->dma_buf_size, DMA_FROM_DEVICE);
|
|
|
|
dma_bytes = tspi->words_to_transfer * tspi->bytes_per_word;
|
|
npackets = tspi->words_to_transfer;
|
|
if (tspi->is_packed) {
|
|
memcpy(t->rx_buf, tspi->rx_dma_buf, dma_bytes);
|
|
} else {
|
|
unsigned int i;
|
|
unsigned int count;
|
|
unsigned char *rx_buf = t->rx_buf;
|
|
unsigned int x;
|
|
unsigned int bits_per_word;
|
|
|
|
bits_per_word = t->bits_per_word ? t->bits_per_word :
|
|
tspi->cur_spi->bits_per_word;
|
|
/* copy received dma words to rx client buffer */
|
|
for (count = 0; count < tspi->words_to_transfer; count++) {
|
|
x = tspi->rx_dma_buf[count];
|
|
for (i = 0; (i < tspi->bytes_per_word); i++)
|
|
*rx_buf++ = (x >> (i*8)) & 0xFF;
|
|
}
|
|
}
|
|
|
|
tspi->words_to_transfer -= npackets;
|
|
tspi->rem_len -= npackets*tspi->bytes_per_word;
|
|
tspi->curr_rx_pos += npackets*tspi->bytes_per_word;
|
|
|
|
if (tspi->words_to_transfer)
|
|
dma_bytes += tegra_spi_read_rx_fifo_to_client_rxbuf(tspi, t);
|
|
|
|
/* Make the dma buffer to read by dma */
|
|
dma_sync_single_for_device(tspi->dev, tspi->rx_dma_phys,
|
|
tspi->dma_buf_size, DMA_FROM_DEVICE);
|
|
return dma_bytes;
|
|
}
|
|
|
|
static void tegra_spi_rx_dma_complete(void *args)
|
|
{
|
|
struct tegra_spi_data *tspi = args;
|
|
dev_dbg(tspi->dev, "rx-dma-complete\n");
|
|
complete(&tspi->rx_dma_complete);
|
|
/* TODO: how to get transfer size from dma */
|
|
}
|
|
|
|
static void tegra_spi_tx_dma_complete(void *args)
|
|
{
|
|
struct tegra_spi_data *tspi = args;
|
|
dev_dbg(tspi->dev, "tx-dma-complete\n");
|
|
complete(&tspi->tx_dma_complete);
|
|
}
|
|
|
|
static int tegra_spi_start_tx_dma(struct tegra_spi_data *tspi, int len)
|
|
{
|
|
reinit_completion(&tspi->tx_dma_complete);
|
|
dev_dbg(tspi->dev, "Starting tx dma for len:%d\n", len);
|
|
tspi->tx_dma_desc = dmaengine_prep_slave_single(tspi->tx_dma_chan,
|
|
tspi->tx_dma_phys, len, DMA_MEM_TO_DEV,
|
|
DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
|
|
if (!tspi->tx_dma_desc) {
|
|
dev_err(tspi->dev, "Not able to get desc for Tx\n");
|
|
return -EIO;
|
|
}
|
|
|
|
tspi->tx_dma_desc->callback = tegra_spi_tx_dma_complete;
|
|
tspi->tx_dma_desc->callback_param = tspi;
|
|
|
|
dmaengine_submit(tspi->tx_dma_desc);
|
|
dma_async_issue_pending(tspi->tx_dma_chan);
|
|
return 0;
|
|
}
|
|
|
|
static int tegra_spi_start_rx_dma(struct tegra_spi_data *tspi, int len)
|
|
{
|
|
reinit_completion(&tspi->rx_dma_complete);
|
|
tspi->rx_dma_len = len;
|
|
|
|
dev_dbg(tspi->dev, "Starting rx dma for len:%d\n", len);
|
|
tspi->rx_dma_desc = dmaengine_prep_slave_single(tspi->rx_dma_chan,
|
|
tspi->rx_dma_phys, len, DMA_DEV_TO_MEM,
|
|
DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
|
|
if (!tspi->rx_dma_desc) {
|
|
dev_err(tspi->dev, "Not able to get desc for Rx\n");
|
|
return -EIO;
|
|
}
|
|
|
|
tspi->rx_dma_desc->callback = tegra_spi_rx_dma_complete;
|
|
tspi->rx_dma_desc->callback_param = tspi;
|
|
|
|
dmaengine_submit(tspi->rx_dma_desc);
|
|
dma_async_issue_pending(tspi->rx_dma_chan);
|
|
return 0;
|
|
}
|
|
|
|
static int check_and_clear_fifo(struct tegra_spi_data *tspi)
|
|
{
|
|
unsigned long status;
|
|
int cnt = SPI_FIFO_FLUSH_MAX_DELAY;
|
|
|
|
/* Make sure that Rx and Tx fifo are empty */
|
|
status = tegra_spi_readl(tspi, SPI_FIFO_STATUS);
|
|
if ((status & SPI_FIFO_EMPTY) != SPI_FIFO_EMPTY) {
|
|
/* flush the fifo */
|
|
status |= (SPI_RX_FIFO_FLUSH | SPI_TX_FIFO_FLUSH);
|
|
tegra_spi_writel(tspi, status, SPI_FIFO_STATUS);
|
|
do {
|
|
status = tegra_spi_readl(tspi, SPI_FIFO_STATUS);
|
|
if ((status & SPI_FIFO_EMPTY) == SPI_FIFO_EMPTY)
|
|
return 0;
|
|
udelay(1);
|
|
} while (cnt--);
|
|
dev_err(tspi->dev,
|
|
"Rx/Tx fifo are not empty status 0x%08lx\n", status);
|
|
return -EIO;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static inline void tegra_spi_slave_busy(struct tegra_spi_data *tspi)
|
|
{
|
|
int deassert_val;
|
|
|
|
if (tspi->slave_ready_active_high)
|
|
deassert_val = 0;
|
|
else
|
|
deassert_val = 1;
|
|
|
|
/* Deassert ready line to indicate Busy */
|
|
if (gpio_is_valid(tspi->gpio_slave_ready))
|
|
gpio_set_value(tspi->gpio_slave_ready, deassert_val);
|
|
}
|
|
|
|
static inline void tegra_spi_slave_ready(struct tegra_spi_data *tspi)
|
|
{
|
|
int assert_val;
|
|
|
|
if (tspi->slave_ready_active_high)
|
|
assert_val = 1;
|
|
else
|
|
assert_val = 0;
|
|
|
|
/* Assert ready line to indicate Ready */
|
|
if (gpio_is_valid(tspi->gpio_slave_ready))
|
|
gpio_set_value(tspi->gpio_slave_ready, assert_val);
|
|
}
|
|
|
|
static inline int tegra_spi_ext_clk_enable(bool enable,
|
|
struct tegra_spi_data *tspi)
|
|
{
|
|
unsigned long misc_reg = 0;
|
|
int ret = 0;
|
|
|
|
if (tspi->chip_data->new_features) {
|
|
/* Enable external clock bit in SPI_MISC_REG */
|
|
if (enable)
|
|
misc_reg |= SPI_MISC_EXT_CLK_EN;
|
|
else
|
|
misc_reg &= (~SPI_MISC_EXT_CLK_EN);
|
|
|
|
tegra_spi_writel(tspi, misc_reg, SPI_MISC_REG);
|
|
}
|
|
|
|
return ret;
|
|
|
|
}
|
|
static int tegra_spi_start_dma_based_transfer(
|
|
struct tegra_spi_data *tspi, struct spi_transfer *t)
|
|
{
|
|
unsigned long val, flags;
|
|
unsigned long intr_mask;
|
|
unsigned int len;
|
|
int ret = 0;
|
|
int maxburst = 0;
|
|
struct dma_slave_config dma_sconfig;
|
|
|
|
/* Make sure that Rx and Tx fifo are empty */
|
|
ret = check_and_clear_fifo(tspi);
|
|
if (ret != 0)
|
|
return ret;
|
|
|
|
val = SPI_DMA_BLK_SET(tspi->curr_dma_words - 1);
|
|
tegra_spi_writel(tspi, val, SPI_DMA_BLK);
|
|
|
|
val = 0; /* Reset the variable for reuse */
|
|
|
|
if (tspi->is_packed)
|
|
len = DIV_ROUND_UP(tspi->curr_dma_words * tspi->bytes_per_word,
|
|
4) * 4;
|
|
else
|
|
len = tspi->curr_dma_words * 4;
|
|
|
|
if (!tspi->rx_trig_words) {
|
|
/* Set attention level based on length of transfer */
|
|
if (len & 0xF) {
|
|
val |= SPI_TX_TRIG_1 | SPI_RX_TRIG_1;
|
|
maxburst = 1;
|
|
} else if (((len) >> 4) & 0x1) {
|
|
val |= SPI_TX_TRIG_4 | SPI_RX_TRIG_4;
|
|
maxburst = 4;
|
|
} else {
|
|
val |= SPI_TX_TRIG_8 | SPI_RX_TRIG_8;
|
|
maxburst = 8;
|
|
}
|
|
} else { /* dt can override the trigger */
|
|
if (tspi->rx_trig_words == 4) {
|
|
val |= SPI_TX_TRIG_4 | SPI_RX_TRIG_4;
|
|
maxburst = 4;
|
|
} else if (tspi->rx_trig_words == 8) {
|
|
val |= SPI_TX_TRIG_8 | SPI_RX_TRIG_8;
|
|
maxburst = 8;
|
|
}
|
|
}
|
|
|
|
if (tspi->variable_length_transfer &&
|
|
tspi->chip_data->new_features &&
|
|
(tspi->cur_direction & DATA_DIR_RX)) {
|
|
val &= ~SPI_RX_TRIG_MASK;
|
|
val |= SPI_RX_TRIG_1;
|
|
}
|
|
|
|
if (tspi->chip_data->intr_mask_reg) {
|
|
if ((tspi->cur_direction & DATA_DIR_TX) ||
|
|
(tspi->cur_direction & DATA_DIR_RX)) {
|
|
intr_mask = tegra_spi_readl(tspi, SPI_INTR_MASK);
|
|
intr_mask &= ~(SPI_INTR_CS_MASK |
|
|
SPI_INTR_RDY_MASK |
|
|
SPI_INTR_RX_FIFO_UNF_MASK |
|
|
SPI_INTR_TX_FIFO_UNF_MASK |
|
|
SPI_INTR_RX_FIFO_OVF_MASK |
|
|
SPI_INTR_TX_FIFO_OVF_MASK);
|
|
tegra_spi_writel(tspi, intr_mask, SPI_INTR_MASK);
|
|
}
|
|
} else if (!tspi->chip_data->new_features) {
|
|
if (tspi->cur_direction & DATA_DIR_TX)
|
|
val |= SPI_IE_TX;
|
|
if (tspi->cur_direction & DATA_DIR_RX)
|
|
val |= SPI_IE_RX;
|
|
}
|
|
|
|
tegra_spi_writel(tspi, val, SPI_DMA_CTL);
|
|
tspi->dma_control_reg = val;
|
|
|
|
if (tspi->cur_direction & DATA_DIR_TX) {
|
|
dma_sconfig.dst_addr = tspi->phys + SPI_TX_FIFO;
|
|
dma_sconfig.dst_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
|
|
dma_sconfig.dst_maxburst = maxburst;
|
|
dmaengine_slave_config(tspi->tx_dma_chan, &dma_sconfig);
|
|
|
|
tegra_spi_copy_client_txbuf_to_spi_txbuf(tspi, t);
|
|
ret = tegra_spi_start_tx_dma(tspi, len);
|
|
if (ret < 0) {
|
|
dev_err(tspi->dev,
|
|
"Starting tx dma failed, err %d\n", ret);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
if (tspi->cur_direction & DATA_DIR_RX) {
|
|
/* Make the dma buffer to read by dma */
|
|
dma_sync_single_for_device(tspi->dev, tspi->rx_dma_phys,
|
|
tspi->dma_buf_size, DMA_FROM_DEVICE);
|
|
|
|
dma_sconfig.src_addr = tspi->phys + SPI_RX_FIFO;
|
|
dma_sconfig.src_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
|
|
if (tspi->variable_length_transfer &&
|
|
tspi->chip_data->new_features)
|
|
dma_sconfig.src_maxburst = 1;
|
|
else
|
|
dma_sconfig.src_maxburst = maxburst;
|
|
dmaengine_slave_config(tspi->rx_dma_chan, &dma_sconfig);
|
|
|
|
/* align Rx Dma to receive size */
|
|
ret = tegra_spi_start_rx_dma(tspi,
|
|
(len >> tspi->rx_trig_words) << tspi->rx_trig_words);
|
|
if (ret < 0) {
|
|
dev_err(tspi->dev,
|
|
"Starting rx dma failed, err %d\n", ret);
|
|
if (tspi->cur_direction & DATA_DIR_TX)
|
|
dmaengine_terminate_all(tspi->tx_dma_chan);
|
|
return ret;
|
|
}
|
|
}
|
|
tspi->is_curr_dma_xfer = true;
|
|
tspi->dma_control_reg = val;
|
|
|
|
val |= SPI_DMA_EN;
|
|
dump_regs(dbg, tspi, "Before DMA EN");
|
|
|
|
/* atomically set both sw/hw isr enable */
|
|
spin_lock_irqsave(&tspi->lock, flags);
|
|
atomic_set(&tspi->isr_expected, 1);
|
|
wmb(); /* to complete the register writes and the atomic var update */
|
|
tegra_spi_writel(tspi, val, SPI_DMA_CTL);
|
|
tegra_spi_fence(tspi);
|
|
spin_unlock_irqrestore(&tspi->lock, flags);
|
|
ret = tegra_spi_ext_clk_enable(true, tspi);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int tegra_spi_start_cpu_based_transfer(
|
|
struct tegra_spi_data *tspi, struct spi_transfer *t)
|
|
{
|
|
unsigned long val, flags;
|
|
unsigned long intr_mask;
|
|
unsigned cur_words;
|
|
int ret;
|
|
|
|
ret = check_and_clear_fifo(tspi);
|
|
if (ret != 0)
|
|
return ret;
|
|
|
|
if (tspi->cur_direction & DATA_DIR_TX)
|
|
cur_words = tegra_spi_fill_tx_fifo_from_client_txbuf(tspi, t);
|
|
else
|
|
cur_words = tspi->curr_dma_words;
|
|
|
|
val = SPI_DMA_BLK_SET(cur_words - 1);
|
|
tegra_spi_writel(tspi, val, SPI_DMA_BLK);
|
|
|
|
val = 0;
|
|
|
|
if (tspi->chip_data->intr_mask_reg) {
|
|
if ((tspi->cur_direction & DATA_DIR_TX) ||
|
|
(tspi->cur_direction & DATA_DIR_RX)) {
|
|
intr_mask = tegra_spi_readl(tspi, SPI_INTR_MASK);
|
|
intr_mask &= ~(SPI_INTR_CS_MASK |
|
|
SPI_INTR_RDY_MASK |
|
|
SPI_INTR_RX_FIFO_UNF_MASK |
|
|
SPI_INTR_TX_FIFO_UNF_MASK |
|
|
SPI_INTR_RX_FIFO_OVF_MASK |
|
|
SPI_INTR_TX_FIFO_OVF_MASK);
|
|
tegra_spi_writel(tspi, intr_mask, SPI_INTR_MASK);
|
|
}
|
|
} else {
|
|
if (tspi->cur_direction & DATA_DIR_TX)
|
|
val |= SPI_IE_TX;
|
|
if (tspi->cur_direction & DATA_DIR_RX)
|
|
val |= SPI_IE_RX;
|
|
}
|
|
|
|
tegra_spi_writel(tspi, val, SPI_DMA_CTL);
|
|
tspi->dma_control_reg = val;
|
|
|
|
tspi->is_curr_dma_xfer = false;
|
|
val = tspi->command1_reg;
|
|
val |= SPI_PIO;
|
|
dump_regs(dbg, tspi, "Before PIO EN");
|
|
|
|
/* atomically set both sw/hw isr enable */
|
|
spin_lock_irqsave(&tspi->lock, flags);
|
|
atomic_set(&tspi->isr_expected, 1);
|
|
wmb(); /* to complete the register writes and the atomic var update */
|
|
tegra_spi_writel(tspi, val, SPI_COMMAND1);
|
|
tegra_spi_fence(tspi);
|
|
spin_unlock_irqrestore(&tspi->lock, flags);
|
|
ret = tegra_spi_ext_clk_enable(true, tspi);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int tegra_spi_init_dma_param(struct tegra_spi_data *tspi,
|
|
bool dma_to_memory)
|
|
{
|
|
struct dma_chan *dma_chan;
|
|
u32 *dma_buf;
|
|
dma_addr_t dma_phys;
|
|
int ret;
|
|
struct dma_slave_config dma_sconfig;
|
|
|
|
dma_chan = dma_request_slave_channel_reason(tspi->dev,
|
|
dma_to_memory ? "rx" : "tx");
|
|
if (IS_ERR(dma_chan)) {
|
|
ret = PTR_ERR(dma_chan);
|
|
if (ret != -EPROBE_DEFER)
|
|
dev_err(tspi->dev,
|
|
"Dma channel is not available: %d\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
dma_buf = dma_alloc_coherent(tspi->dev, tspi->dma_buf_size,
|
|
&dma_phys, GFP_KERNEL);
|
|
if (!dma_buf) {
|
|
dev_err(tspi->dev, "Not able to allocate the dma buffer\n");
|
|
dma_release_channel(dma_chan);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
if (dma_to_memory) {
|
|
dma_sconfig.src_addr = tspi->phys + SPI_RX_FIFO;
|
|
dma_sconfig.src_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
|
|
dma_sconfig.src_maxburst = tspi->rx_trig_words;
|
|
} else {
|
|
dma_sconfig.dst_addr = tspi->phys + SPI_TX_FIFO;
|
|
dma_sconfig.dst_addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES;
|
|
dma_sconfig.dst_maxburst = tspi->rx_trig_words;
|
|
}
|
|
|
|
ret = dmaengine_slave_config(dma_chan, &dma_sconfig);
|
|
if (ret)
|
|
goto scrub;
|
|
if (dma_to_memory) {
|
|
tspi->rx_dma_chan = dma_chan;
|
|
tspi->rx_dma_buf = dma_buf;
|
|
tspi->rx_dma_phys = dma_phys;
|
|
} else {
|
|
tspi->tx_dma_chan = dma_chan;
|
|
tspi->tx_dma_buf = dma_buf;
|
|
tspi->tx_dma_phys = dma_phys;
|
|
}
|
|
return 0;
|
|
|
|
scrub:
|
|
dma_free_coherent(tspi->dev, tspi->dma_buf_size, dma_buf, dma_phys);
|
|
dma_release_channel(dma_chan);
|
|
return ret;
|
|
}
|
|
|
|
static void tegra_spi_deinit_dma_param(struct tegra_spi_data *tspi,
|
|
bool dma_to_memory)
|
|
{
|
|
u32 *dma_buf;
|
|
dma_addr_t dma_phys;
|
|
struct dma_chan *dma_chan;
|
|
|
|
if (dma_to_memory) {
|
|
dma_buf = tspi->rx_dma_buf;
|
|
dma_chan = tspi->rx_dma_chan;
|
|
dma_phys = tspi->rx_dma_phys;
|
|
tspi->rx_dma_chan = NULL;
|
|
tspi->rx_dma_buf = NULL;
|
|
} else {
|
|
dma_buf = tspi->tx_dma_buf;
|
|
dma_chan = tspi->tx_dma_chan;
|
|
dma_phys = tspi->tx_dma_phys;
|
|
tspi->tx_dma_buf = NULL;
|
|
tspi->tx_dma_chan = NULL;
|
|
}
|
|
if (!dma_chan)
|
|
return;
|
|
|
|
dma_free_coherent(tspi->dev, tspi->dma_buf_size, dma_buf, dma_phys);
|
|
dma_release_channel(dma_chan);
|
|
}
|
|
|
|
static int tegra_spi_validate_request(struct spi_device *spi,
|
|
struct tegra_spi_data *tspi, struct spi_transfer *t)
|
|
{
|
|
int req_mode;
|
|
|
|
req_mode = spi->mode & 0x3;
|
|
if ((req_mode == SPI_MODE_0) || (req_mode == SPI_MODE_2)) {
|
|
if (t->tx_buf) {
|
|
/* Tx is not supported in mode 0 and mode 2 */
|
|
dev_err(tspi->dev, "Tx is not supported in mode %d\n",
|
|
req_mode);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
/* Check that the all words are available */
|
|
if (t->len % tspi->bytes_per_word != 0) {
|
|
dev_dbg(tspi->dev, "length is not a multiple of bypw\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* the transfer needs to be completed in one go */
|
|
if (t->len > tspi->dma_buf_size) {
|
|
dev_dbg(tspi->dev, "message size is greater than dma size\n");
|
|
return -EMSGSIZE;
|
|
}
|
|
|
|
/* In unpacked mode, a fifo-word (4 bytes) is used to store a word
|
|
* irrespective of bytes-per-word setting. This means that the
|
|
* DMA transfers a FIFO word to dma buffer occuping 4 bytes and
|
|
* hence the total number of packets possible is 1/4th of dma buff size
|
|
*/
|
|
if (!tspi->is_packed &&
|
|
((t->len / tspi->bytes_per_word) > tspi->dma_buf_size / 4)) {
|
|
dev_dbg(tspi->dev, "Num words is greater than dma size\n");
|
|
return -EMSGSIZE;
|
|
}
|
|
|
|
/* the total number of packets should be less than max dma packets */
|
|
if ((t->len / tspi->bytes_per_word) > MAX_PACKETS) {
|
|
dev_dbg(tspi->dev, "Num packets is greater than 64K\n");
|
|
return -EMSGSIZE;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void set_best_clk_source(struct spi_device *spi,
|
|
unsigned long rate)
|
|
{
|
|
long new_rate;
|
|
unsigned long err_rate, crate, prate;
|
|
unsigned int cdiv, fin_err = rate;
|
|
int ret;
|
|
struct clk *pclk, *fpclk = NULL;
|
|
const char *pclk_name, *fpclk_name = NULL;
|
|
struct device_node *node;
|
|
struct property *prop;
|
|
struct tegra_spi_data *tspi = spi_master_get_devdata(spi->master);
|
|
|
|
node = spi->master->dev.of_node;
|
|
if (!of_property_count_strings(node, "nvidia,clk-parents"))
|
|
return;
|
|
|
|
/* when parent of a clk changes divider is not changed
|
|
* set a min div with which clk will not cross max rate
|
|
*/
|
|
if (!tspi->min_div) {
|
|
of_property_for_each_string(node, "nvidia,clk-parents",
|
|
prop, pclk_name) {
|
|
pclk = clk_get(tspi->dev, pclk_name);
|
|
if (IS_ERR(pclk))
|
|
continue;
|
|
prate = clk_get_rate(pclk);
|
|
crate = tspi->spi_max_frequency;
|
|
cdiv = DIV_ROUND_UP(prate, crate);
|
|
if (cdiv > tspi->min_div)
|
|
tspi->min_div = cdiv;
|
|
}
|
|
}
|
|
|
|
pclk = clk_get_parent(tspi->clk);
|
|
crate = clk_get_rate(tspi->clk);
|
|
prate = clk_get_rate(pclk);
|
|
if (!crate)
|
|
return;
|
|
cdiv = DIV_ROUND_UP(prate, crate);
|
|
if (cdiv < tspi->min_div) {
|
|
crate = DIV_ROUND_UP(prate, tspi->min_div);
|
|
clk_set_rate(tspi->clk, crate);
|
|
}
|
|
|
|
of_property_for_each_string(node, "nvidia,clk-parents",
|
|
prop, pclk_name) {
|
|
pclk = clk_get(tspi->dev, pclk_name);
|
|
if (IS_ERR(pclk))
|
|
continue;
|
|
|
|
ret = clk_set_parent(tspi->clk, pclk);
|
|
if (ret < 0) {
|
|
dev_warn(tspi->dev,
|
|
"Error in setting parent clk src %s: %d\n",
|
|
pclk_name, ret);
|
|
continue;
|
|
}
|
|
|
|
new_rate = clk_round_rate(tspi->clk, rate);
|
|
if (new_rate < 0)
|
|
continue;
|
|
|
|
err_rate = abs(new_rate - rate);
|
|
if (err_rate < fin_err) {
|
|
fpclk = pclk;
|
|
fin_err = err_rate;
|
|
fpclk_name = pclk_name;
|
|
}
|
|
}
|
|
|
|
if (fpclk) {
|
|
dev_dbg(tspi->dev, "Setting clk_src %s\n",
|
|
fpclk_name);
|
|
clk_set_parent(tspi->clk, fpclk);
|
|
}
|
|
}
|
|
|
|
static int tegra_spi_start_transfer_one(struct spi_device *spi,
|
|
struct spi_transfer *t, bool is_first_of_msg,
|
|
bool is_single_xfer)
|
|
{
|
|
struct tegra_spi_data *tspi = spi_master_get_devdata(spi->master);
|
|
struct tegra_spi_device_controller_data *cdata = spi->controller_data;
|
|
u32 speed;
|
|
u32 core_speed;
|
|
u8 bits_per_word;
|
|
unsigned total_fifo_words;
|
|
int ret;
|
|
unsigned long command1;
|
|
int req_mode;
|
|
|
|
bits_per_word = t->bits_per_word;
|
|
speed = t->speed_hz ? t->speed_hz : spi->max_speed_hz;
|
|
/* Set slave controller clk 1.5 times the bus frequency */
|
|
if (!speed)
|
|
speed = tspi->spi_max_frequency;
|
|
|
|
if (tspi->chip_data->new_features) {
|
|
/* In case of new feature, all DMA interfaces are async.
|
|
* so only 1.5x freq ration required
|
|
Set speed to 2x to avoid border cases where actual
|
|
clock is not same requested rate*/
|
|
core_speed = (speed * 2);
|
|
} else {
|
|
/* To maintain min 1.5x and max 4x ratio between
|
|
* slave core clk and interface clk */
|
|
core_speed = speed * 3;
|
|
}
|
|
if (core_speed > tspi->spi_max_frequency)
|
|
core_speed = tspi->spi_max_frequency;
|
|
|
|
if (core_speed < ((speed * 3) >> 1)) {
|
|
dev_err(tspi->dev, "Cannot set requested clk freq %d\n", speed);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (core_speed != tspi->cur_speed) {
|
|
set_best_clk_source(spi, core_speed);
|
|
ret = clk_set_rate(tspi->clk, core_speed);
|
|
if (ret) {
|
|
dev_err(tspi->dev, "Failed to set clk freq %d\n", ret);
|
|
return -EINVAL;
|
|
}
|
|
tspi->cur_speed = core_speed;
|
|
}
|
|
|
|
tspi->cur_spi = spi;
|
|
tspi->curr_xfer = t;
|
|
tspi->curr_rx_pos = 0;
|
|
tspi->curr_pos = 0;
|
|
tspi->tx_status = 0;
|
|
tspi->rx_status = 0;
|
|
total_fifo_words = tegra_spi_calculate_curr_xfer_param(spi, tspi, t);
|
|
|
|
ret = tegra_spi_validate_request(spi, tspi, t);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
if (is_first_of_msg) {
|
|
tegra_spi_clear_status(tspi);
|
|
|
|
command1 = tspi->def_command1_reg;
|
|
command1 |= SPI_BIT_LENGTH(bits_per_word - 1);
|
|
|
|
command1 &= ~SPI_CONTROL_MODE_MASK;
|
|
req_mode = spi->mode & 0x3;
|
|
if (req_mode == SPI_MODE_0)
|
|
command1 |= SPI_CONTROL_MODE_0;
|
|
else if (req_mode == SPI_MODE_1)
|
|
command1 |= SPI_CONTROL_MODE_1;
|
|
else if (req_mode == SPI_MODE_2)
|
|
command1 |= SPI_CONTROL_MODE_2;
|
|
else if (req_mode == SPI_MODE_3)
|
|
command1 |= SPI_CONTROL_MODE_3;
|
|
|
|
tegra_spi_writel(tspi, command1, SPI_COMMAND1);
|
|
tspi->variable_length_transfer = false;
|
|
if (cdata && cdata->variable_length_transfer)
|
|
tspi->variable_length_transfer = true;
|
|
} else {
|
|
command1 = tspi->command1_reg;
|
|
command1 &= ~SPI_BIT_LENGTH(~0);
|
|
command1 |= SPI_BIT_LENGTH(bits_per_word - 1);
|
|
}
|
|
|
|
if (spi->mode & SPI_LSB_FIRST)
|
|
command1 |= SPI_LSBIT_FE;
|
|
else
|
|
command1 &= ~SPI_LSBIT_FE;
|
|
|
|
if (tspi->is_packed)
|
|
command1 |= SPI_PACKED;
|
|
|
|
command1 &= ~(SPI_CS_SEL_MASK | SPI_TX_EN | SPI_RX_EN);
|
|
tspi->cur_direction = 0;
|
|
if (t->rx_buf) {
|
|
command1 |= SPI_RX_EN;
|
|
tspi->cur_direction |= DATA_DIR_RX;
|
|
}
|
|
if (t->tx_buf) {
|
|
command1 |= SPI_TX_EN;
|
|
tspi->cur_direction |= DATA_DIR_TX;
|
|
}
|
|
command1 |= SPI_CS_SEL(spi->chip_select);
|
|
tegra_spi_writel(tspi, command1, SPI_COMMAND1);
|
|
tspi->command1_reg = command1;
|
|
|
|
dev_dbg(tspi->dev, "The def 0x%x and written 0x%lx\n",
|
|
tspi->def_command1_reg, command1);
|
|
|
|
if (total_fifo_words > SPI_FIFO_DEPTH)
|
|
ret = tegra_spi_start_dma_based_transfer(tspi, t);
|
|
else
|
|
ret = tegra_spi_start_cpu_based_transfer(tspi, t);
|
|
|
|
tegra_spi_slave_ready(tspi);
|
|
|
|
/* Inform client that we are ready now. */
|
|
if (tspi->spi_slave_ready_callback)
|
|
tspi->spi_slave_ready_callback(tspi->client_data);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static struct tegra_spi_device_controller_data
|
|
*tegra_spi_get_cdata_dt(struct spi_device *spi,
|
|
struct tegra_spi_data *tspi)
|
|
{
|
|
struct tegra_spi_device_controller_data *cdata;
|
|
struct device_node *slave_np, *data_np;
|
|
int ret;
|
|
|
|
slave_np = spi->dev.of_node;
|
|
if (!slave_np) {
|
|
dev_dbg(&spi->dev, "device node not found\n");
|
|
return NULL;
|
|
}
|
|
|
|
data_np = of_get_child_by_name(slave_np, "controller-data");
|
|
if (!data_np) {
|
|
dev_dbg(&spi->dev, "child node 'controller-data' not found\n");
|
|
return NULL;
|
|
}
|
|
|
|
cdata = &tspi->cdata[spi->chip_select];
|
|
memset(cdata, 0, sizeof(*cdata));
|
|
|
|
ret = of_property_read_bool(data_np, "nvidia,variable-length-transfer");
|
|
if (ret)
|
|
cdata->variable_length_transfer = 1;
|
|
|
|
tspi->lsbyte_first =
|
|
of_property_read_bool(data_np, "nvidia,lsbyte-first");
|
|
|
|
of_node_put(data_np);
|
|
return cdata;
|
|
}
|
|
|
|
static int tegra_spi_setup(struct spi_device *spi)
|
|
{
|
|
struct tegra_spi_data *tspi = spi_master_get_devdata(spi->master);
|
|
struct tegra_spi_device_controller_data *cdata = spi->controller_data;
|
|
unsigned long val;
|
|
unsigned long flags;
|
|
int ret;
|
|
unsigned int cs_pol_bit[MAX_CHIP_SELECT] = {
|
|
SPI_CS_POL_INACTIVE_0,
|
|
SPI_CS_POL_INACTIVE_1,
|
|
SPI_CS_POL_INACTIVE_2,
|
|
SPI_CS_POL_INACTIVE_3,
|
|
};
|
|
|
|
dev_dbg(&spi->dev, "setup %d bpw, %scpol, %scpha, %dHz\n",
|
|
spi->bits_per_word,
|
|
spi->mode & SPI_CPOL ? "" : "~",
|
|
spi->mode & SPI_CPHA ? "" : "~",
|
|
spi->max_speed_hz);
|
|
|
|
if (!cdata) {
|
|
cdata = tegra_spi_get_cdata_dt(spi, tspi);
|
|
spi->controller_data = cdata;
|
|
}
|
|
|
|
/* Set speed to the spi max fequency if spi device has not set */
|
|
spi->max_speed_hz = spi->max_speed_hz ? : tspi->spi_max_frequency;
|
|
|
|
ret = pm_runtime_get_sync(tspi->dev);
|
|
if (ret < 0) {
|
|
dev_err(tspi->dev, "pm runtime failed, e = %d\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
spin_lock_irqsave(&tspi->lock, flags);
|
|
val = tspi->def_command1_reg;
|
|
if (spi->mode & SPI_CS_HIGH)
|
|
val &= ~cs_pol_bit[spi->chip_select];
|
|
else
|
|
val |= cs_pol_bit[spi->chip_select];
|
|
|
|
if (tspi->lsbyte_first)
|
|
val |= SPI_LSBYTE_FE;
|
|
|
|
tspi->def_command1_reg = val;
|
|
tegra_spi_writel(tspi, tspi->def_command1_reg, SPI_COMMAND1);
|
|
spin_unlock_irqrestore(&tspi->lock, flags);
|
|
|
|
pm_runtime_put(tspi->dev);
|
|
return 0;
|
|
}
|
|
|
|
static void handle_cpu_based_err_xfer(struct tegra_spi_data *tspi)
|
|
{
|
|
if (tspi->tx_status || tspi->rx_status) {
|
|
dump_regs(err_ratelimited, tspi,
|
|
"cpu-xfer-err [status:%08x]", tspi->status_reg);
|
|
tspi->reset_ctrl_status = true;
|
|
}
|
|
|
|
}
|
|
|
|
static void handle_dma_based_err_xfer(struct tegra_spi_data *tspi)
|
|
{
|
|
int err = 0;
|
|
|
|
/* Abort dmas if any error */
|
|
if (tspi->cur_direction & DATA_DIR_TX) {
|
|
if (tspi->tx_status) {
|
|
dmaengine_terminate_all(tspi->tx_dma_chan);
|
|
dump_regs(err_ratelimited, tspi,
|
|
"tx-dma-err [status:%08x]", tspi->status_reg);
|
|
err += 1;
|
|
}
|
|
}
|
|
|
|
if (tspi->cur_direction & DATA_DIR_RX) {
|
|
if (tspi->rx_status) {
|
|
dmaengine_terminate_all(tspi->rx_dma_chan);
|
|
dump_regs(err_ratelimited, tspi,
|
|
"rx-dma-err [status:%08x]", tspi->status_reg);
|
|
err += 2;
|
|
}
|
|
}
|
|
|
|
if (err)
|
|
tspi->reset_ctrl_status = true;
|
|
}
|
|
|
|
|
|
static int tegra_spi_wait_on_message_xfer(struct tegra_spi_data *tspi)
|
|
{
|
|
long ret;
|
|
unsigned long timeout;
|
|
|
|
/* delay_usecs is used as timeout for slave, value of 0 is inf wait.
|
|
* since delay_usecs will be too small for practical use (65ms max)
|
|
* a multiplier of 1000 is used to make as msecs.
|
|
*/
|
|
timeout = tspi->curr_xfer->delay_usecs ?
|
|
usecs_to_jiffies(tspi->curr_xfer->delay_usecs *
|
|
DELAY_MUL_FACTOR) : MAX_SCHEDULE_TIMEOUT;
|
|
/* wait for spi isr */
|
|
ret = wait_for_completion_interruptible_timeout(
|
|
&tspi->xfer_completion, timeout);
|
|
if (ret <= 0) {
|
|
/* interrupted-wait/late interrupt is considered spurious */
|
|
atomic_set(&tspi->isr_expected, 0);
|
|
if (ret == -ERESTARTSYS) {
|
|
flush_signals(current);
|
|
dev_warn(tspi->dev, "waiting for master was interrupted\n");
|
|
}
|
|
if (ret == 0)
|
|
dev_info(tspi->dev, "Timeout waiting for master\n");
|
|
dump_regs(dbg, tspi, "spi-int-timeout [ret:%ld]", ret);
|
|
if (tspi->is_curr_dma_xfer &&
|
|
(tspi->cur_direction & DATA_DIR_TX))
|
|
dmaengine_terminate_all(tspi->tx_dma_chan);
|
|
if (tspi->is_curr_dma_xfer &&
|
|
(tspi->cur_direction & DATA_DIR_RX))
|
|
dmaengine_terminate_all(tspi->rx_dma_chan);
|
|
|
|
/* resetting controller to unarm spi */
|
|
tspi->reset_ctrl_status = true;
|
|
ret = -EIO;
|
|
return ret;
|
|
}
|
|
/* interrupt came and status in tspi->status_reg */
|
|
if (tspi->cur_direction & DATA_DIR_TX)
|
|
tspi->tx_status = tspi->status_reg &
|
|
(SPI_TX_FIFO_UNF | SPI_TX_FIFO_OVF);
|
|
|
|
if (tspi->cur_direction & DATA_DIR_RX)
|
|
tspi->rx_status = tspi->status_reg &
|
|
(SPI_RX_FIFO_OVF | SPI_RX_FIFO_UNF);
|
|
|
|
/* workaround when number of packets is 64K. The transfer status
|
|
* field is 16bit and zero based count (0...2^16-1) and hence for 64K
|
|
* the counter resets to 0. Hence for a transfer, if count is
|
|
* reported as zero, expected is 64K and there is no error, then
|
|
* we correct it to 64K. Packets greater than 64K is not allowed.
|
|
*/
|
|
tspi->words_to_transfer =
|
|
(tspi->words_to_transfer == 0 &&
|
|
tspi->curr_xfer->len == MAX_PACKETS) ?
|
|
MAX_PACKETS : tspi->words_to_transfer;
|
|
|
|
/* check if required bytes where transferred */
|
|
if (!tspi->variable_length_transfer &&
|
|
tspi->words_to_transfer !=
|
|
tspi->curr_xfer->len/tspi->bytes_per_word) {
|
|
dump_regs(err_ratelimited, tspi,
|
|
"transferred[%d] != requested[%d]",
|
|
tspi->words_to_transfer,
|
|
tspi->curr_xfer->len/tspi->bytes_per_word);
|
|
tspi->tx_status = SPI_TX_FIFO_UNF;
|
|
tspi->rx_status = SPI_RX_FIFO_OVF;
|
|
}
|
|
|
|
/* cs inactive intr while it is masked, mark as error */
|
|
if (tspi->chip_data->mask_cs_inactive_intr &&
|
|
(tspi->status_reg & SPI_CS_INACTIVE)) {
|
|
dev_dbg(tspi->dev, "cs inactive intr, status_reg = 0x%x\n",
|
|
tspi->status_reg);
|
|
tspi->tx_status = SPI_TX_FIFO_UNF;
|
|
tspi->rx_status = SPI_RX_FIFO_OVF;
|
|
}
|
|
|
|
if (!tspi->is_curr_dma_xfer)
|
|
handle_cpu_based_err_xfer(tspi);
|
|
else
|
|
handle_dma_based_err_xfer(tspi);
|
|
|
|
/* truncate in case we received more than requested */
|
|
tspi->words_to_transfer = min(tspi->words_to_transfer,
|
|
tspi->curr_xfer->len/tspi->bytes_per_word);
|
|
|
|
return (tspi->tx_status || tspi->rx_status) ? -EIO : 0;
|
|
}
|
|
|
|
static int tegra_spi_handle_message(struct tegra_spi_data *tspi,
|
|
struct spi_transfer *xfer)
|
|
{
|
|
int ret = 0;
|
|
long wait_status;
|
|
unsigned pending_rx_word_count;
|
|
unsigned long fifo_status;
|
|
unsigned long word_tfr_status;
|
|
unsigned long actual_words_xferrd;
|
|
|
|
/* we are here, so no io-error and received the expected num bytes
|
|
* expect in variable length case where received <= expected
|
|
*/
|
|
if (!tspi->is_curr_dma_xfer) {
|
|
if (tspi->cur_direction & DATA_DIR_RX)
|
|
tegra_spi_read_rx_fifo_to_client_rxbuf(tspi, xfer);
|
|
|
|
} else {
|
|
if (tspi->cur_direction & DATA_DIR_TX) {
|
|
if ((IS_SPI_CS_INACTIVE(tspi->status_reg)) &&
|
|
tspi->variable_length_transfer) {
|
|
/* Check for TX DMA completion. If the TX DMA
|
|
* is already completed, No need to terminate
|
|
* the DMA xfer.
|
|
*/
|
|
wait_status = try_wait_for_completion(
|
|
&tspi->tx_dma_complete);
|
|
if (!wait_status) {
|
|
dev_dbg(tspi->dev,
|
|
"terminating tx dma\n");
|
|
dmaengine_terminate_all(
|
|
tspi->tx_dma_chan);
|
|
async_tx_ack(tspi->tx_dma_desc);
|
|
} else
|
|
dev_dbg(tspi->dev,
|
|
"received tx dma completion\n");
|
|
|
|
tspi->reset_ctrl_status = true;
|
|
} else {
|
|
wait_status =
|
|
wait_for_completion_interruptible_timeout(
|
|
&tspi->tx_dma_complete,
|
|
SPI_DMA_TIMEOUT);
|
|
if (wait_status <= 0) {
|
|
dmaengine_terminate_all(
|
|
tspi->tx_dma_chan);
|
|
dump_regs(dbg, tspi,
|
|
"tx-dma-timeout [wait:%ld]",
|
|
wait_status);
|
|
tspi->reset_ctrl_status = true;
|
|
ret = -EIO;
|
|
return ret;
|
|
}
|
|
}
|
|
}
|
|
if (tspi->cur_direction & DATA_DIR_RX) {
|
|
if (tspi->variable_length_transfer) {
|
|
dmaengine_terminate_all(tspi->rx_dma_chan);
|
|
async_tx_ack(tspi->rx_dma_desc);
|
|
|
|
fifo_status =
|
|
tegra_spi_readl(tspi, SPI_FIFO_STATUS);
|
|
|
|
pending_rx_word_count =
|
|
SPI_RX_FIFO_FULL_COUNT(fifo_status);
|
|
|
|
word_tfr_status =
|
|
DIV_ROUND_UP(tspi->words_to_transfer ,
|
|
tspi->words_per_32bit);
|
|
|
|
if (word_tfr_status < pending_rx_word_count) {
|
|
dmaengine_terminate_all(
|
|
tspi->rx_dma_chan);
|
|
dump_regs(err_ratelimited, tspi,
|
|
"available bytes less than expected");
|
|
tspi->reset_ctrl_status = true;
|
|
ret = -EIO;
|
|
return ret;
|
|
}
|
|
|
|
actual_words_xferrd =
|
|
word_tfr_status - pending_rx_word_count;
|
|
|
|
while (pending_rx_word_count > 0) {
|
|
|
|
tspi->rx_dma_buf[actual_words_xferrd] =
|
|
tegra_spi_readl(tspi,
|
|
SPI_RX_FIFO);
|
|
|
|
actual_words_xferrd++;
|
|
pending_rx_word_count--;
|
|
}
|
|
} else {
|
|
wait_status =
|
|
wait_for_completion_interruptible_timeout(
|
|
&tspi->rx_dma_complete,
|
|
SPI_DMA_TIMEOUT);
|
|
if (wait_status <= 0) {
|
|
dmaengine_terminate_all(
|
|
tspi->rx_dma_chan);
|
|
dump_regs(dbg, tspi,
|
|
"rx-dma-timeout [wait:%ld]",
|
|
wait_status);
|
|
tspi->reset_ctrl_status = true;
|
|
ret = -EIO;
|
|
return ret;
|
|
}
|
|
}
|
|
tegra_spi_copy_spi_rxbuf_to_client_rxbuf(tspi, xfer);
|
|
}
|
|
}
|
|
|
|
/* For TX direction, It might be possible that CS deassert happens
|
|
* before the tranfer completion(variable length case). In this scenario
|
|
* actual transferred data is less than the requested data. So to update
|
|
* the curr_pos(and finally xfer->len) read the data transfer count from
|
|
* SPI_TRANS_STATUS and update the xfer->len accordingly.
|
|
*
|
|
* In case of RX direction, tspi->curr_rx_pos contains actual received
|
|
* data. We can directly use it to update the xfer->len.
|
|
*/
|
|
if (tspi->cur_direction & DATA_DIR_TX) {
|
|
tspi->curr_pos += tspi->words_to_transfer * tspi->bytes_per_word;
|
|
tspi->rem_len -= tspi->curr_pos;
|
|
}
|
|
if (tspi->cur_direction & DATA_DIR_RX)
|
|
tspi->curr_pos = tspi->curr_rx_pos;
|
|
|
|
/* If external master de-asserts the CS randomly
|
|
and after that is SW wants to do a fresh transfer,
|
|
SW needs to apply controller reset to flush the fifos
|
|
and internal pipeline as controller might have
|
|
fetched more data from memory by then. */
|
|
if (tspi->curr_pos == xfer->len ||
|
|
(IS_SPI_CS_INACTIVE(tspi->status_reg) &&
|
|
tspi->variable_length_transfer)) {
|
|
xfer->len = tspi->curr_pos;
|
|
if ((IS_SPI_CS_INACTIVE(tspi->status_reg))
|
|
&& tspi->variable_length_transfer
|
|
&& tspi->chip_data->new_features
|
|
&& (tspi->cur_direction & DATA_DIR_TX)) {
|
|
tspi->reset_ctrl_status = true;
|
|
}
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int tegra_spi_transfer_one_message(struct spi_master *master,
|
|
struct spi_message *msg)
|
|
{
|
|
bool is_first_msg = true;
|
|
int single_xfer;
|
|
struct tegra_spi_data *tspi = spi_master_get_devdata(master);
|
|
struct spi_transfer *xfer;
|
|
struct spi_device *spi = msg->spi;
|
|
int ret;
|
|
|
|
msg->status = 0;
|
|
msg->actual_length = 0;
|
|
|
|
ret = pm_runtime_get_sync(tspi->dev);
|
|
if (ret < 0) {
|
|
dev_err(tspi->dev, "runtime PM get failed: %d\n", ret);
|
|
msg->status = ret;
|
|
spi_finalize_current_message(master);
|
|
return ret;
|
|
}
|
|
/* allow SIGINT for interrupting the sleep */
|
|
allow_signal(SIGINT);
|
|
|
|
/* Fix me - Implement proper sequnce for CDC */
|
|
tegra_spi_ext_clk_enable(false, tspi);
|
|
reset_controller(tspi);
|
|
tegra_spi_writel(tspi, tspi->def_command1_reg, SPI_COMMAND1);
|
|
|
|
single_xfer = list_is_singular(&msg->transfers);
|
|
list_for_each_entry(xfer, &msg->transfers, transfer_list) {
|
|
reinit_completion(&tspi->xfer_completion);
|
|
tspi->rem_len = xfer->len;
|
|
tspi->words_to_transfer = 0;
|
|
dump_xfer(tspi, xfer);
|
|
/* program hardware */
|
|
ret = tegra_spi_start_transfer_one(spi, xfer,
|
|
is_first_msg, single_xfer);
|
|
if (ret < 0) {
|
|
dev_err(tspi->dev,
|
|
"spi can not start transfer, err %d\n", ret);
|
|
goto exit;
|
|
}
|
|
is_first_msg = false;
|
|
/* wait for spi-interrupt and handle transfer error */
|
|
ret = tegra_spi_wait_on_message_xfer(tspi);
|
|
if (ret) {
|
|
tegra_spi_slave_busy(tspi);
|
|
goto exit;
|
|
}
|
|
|
|
/* unpack and copy data to client */
|
|
ret = tegra_spi_handle_message(tspi, xfer);
|
|
if (ret < 0)
|
|
goto exit;
|
|
|
|
if (tspi->variable_length_transfer)
|
|
msg->actual_length += xfer->len;
|
|
else
|
|
msg->actual_length += (xfer->len - tspi->rem_len);
|
|
|
|
#ifdef PROFILE_SPI_SLAVE
|
|
{
|
|
ktime_t time;
|
|
tspi->end_time = ktime_get();
|
|
tspi->profile_size = xfer->len - tspi->rem_len;
|
|
time = ktime_sub(tspi->end_time, tspi->start_time);
|
|
dev_dbg(tspi->dev,
|
|
"Profile: size=%dB time-from-spi-int=%lldns\n",
|
|
tspi->profile_size, time.tv64);
|
|
}
|
|
#endif
|
|
if (tspi->rem_len)
|
|
dump_sw_state(tspi,
|
|
"Error transfer was truncated [rem:%d]",
|
|
tspi->rem_len);
|
|
/* checks for any error after s/w handling the transfer */
|
|
ret = tegra_spi_readl(tspi, SPI_FIFO_STATUS);
|
|
if (ret &
|
|
(SPI_FRAME_END | SPI_CS_INACTIVE | SPI_ERR |
|
|
SPI_TX_FIFO_UNF | SPI_TX_FIFO_OVF |
|
|
SPI_RX_FIFO_OVF | SPI_RX_FIFO_UNF))
|
|
dump_regs(dbg, tspi, "sw-handle-error");
|
|
|
|
if (!(ret & SPI_FIFO_EMPTY))
|
|
dump_regs(dbg, tspi, "rx/tx fifo are not empty");
|
|
}
|
|
ret = 0;
|
|
exit:
|
|
tegra_spi_writel(tspi, tspi->def_command1_reg, SPI_COMMAND1);
|
|
pm_runtime_put(tspi->dev);
|
|
msg->status = ret;
|
|
spi_finalize_current_message(master);
|
|
return ret;
|
|
}
|
|
|
|
static irqreturn_t tegra_spi_isr(int irq, void *context_data)
|
|
{
|
|
struct tegra_spi_data *tspi = context_data;
|
|
unsigned long status_reg;
|
|
|
|
status_reg = tegra_spi_readl(tspi, SPI_FIFO_STATUS);
|
|
dump_regs(dbg, tspi, "@isr");
|
|
|
|
tegra_spi_clear_status(tspi);
|
|
|
|
/* if isr was not expected */
|
|
if (!atomic_xchg(&tspi->isr_expected, 0)) {
|
|
dev_err_ratelimited(tspi->dev,
|
|
"spurious interrupt, status_reg = 0x%x\n",
|
|
tspi->status_reg);
|
|
return IRQ_NONE;
|
|
}
|
|
|
|
#ifdef PROFILE_SPI_SLAVE
|
|
tspi->start_time = ktime_get();
|
|
#endif
|
|
tspi->status_reg = status_reg;
|
|
tspi->words_to_transfer = SPI_BLK_CNT(
|
|
tegra_spi_readl(tspi, SPI_TRANS_STATUS));
|
|
|
|
tegra_spi_slave_busy(tspi);
|
|
|
|
/* Inform client about controller interrupt. */
|
|
if (tspi->spi_slave_isr_callback)
|
|
tspi->spi_slave_isr_callback(tspi->client_data);
|
|
|
|
complete(&tspi->xfer_completion);
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static struct tegra_spi_platform_data *tegra_spi_parse_dt(
|
|
struct platform_device *pdev)
|
|
{
|
|
struct tegra_spi_platform_data *pdata;
|
|
const unsigned int *prop;
|
|
struct device_node *np = pdev->dev.of_node;
|
|
enum of_gpio_flags gpio_flags;
|
|
|
|
pdata = devm_kzalloc(&pdev->dev, sizeof(*pdata), GFP_KERNEL);
|
|
if (!pdata) {
|
|
dev_err(&pdev->dev, "Memory alloc for pdata failed\n");
|
|
return NULL;
|
|
}
|
|
|
|
prop = of_get_property(np, "spi-max-frequency", NULL);
|
|
if (prop)
|
|
pdata->spi_max_frequency = be32_to_cpup(prop);
|
|
|
|
prop = of_get_property(np, "nvidia,rx-trigger-words", NULL);
|
|
if (prop)
|
|
pdata->rx_trig_words = be32_to_cpup(prop);
|
|
|
|
prop = of_get_property(np, "nvidia,least-significant-bit", NULL);
|
|
if (prop)
|
|
pdata->ls_bit = (be32_to_cpup(prop)) & 0x1;
|
|
|
|
if (of_find_property(np, "nvidia,clock-always-on", NULL))
|
|
pdata->is_clkon_always = true;
|
|
|
|
pdata->gpio_slave_ready =
|
|
of_get_named_gpio_flags(np, "nvidia,slave-ready-gpio", 0,
|
|
&gpio_flags);
|
|
|
|
if (gpio_flags & OF_GPIO_ACTIVE_LOW)
|
|
pdata->slave_ready_active_high = false;
|
|
else
|
|
pdata->slave_ready_active_high = true;
|
|
|
|
return pdata;
|
|
}
|
|
|
|
static struct tegra_spi_chip_data tegra124_spi_chip_data = {
|
|
.intr_mask_reg = false,
|
|
.mask_cs_inactive_intr = true,
|
|
.new_features = false,
|
|
};
|
|
|
|
static struct tegra_spi_chip_data tegra210_spi_chip_data = {
|
|
.intr_mask_reg = true,
|
|
.mask_cs_inactive_intr = false,
|
|
.new_features = false,
|
|
};
|
|
|
|
static struct tegra_spi_chip_data tegra186_spi_chip_data = {
|
|
.intr_mask_reg = true,
|
|
.mask_cs_inactive_intr = false,
|
|
.new_features = true,
|
|
};
|
|
|
|
static struct of_device_id tegra_spi_of_match[] = {
|
|
{
|
|
.compatible = "nvidia,tegra124-spi-slave",
|
|
.data = &tegra124_spi_chip_data,
|
|
}, {
|
|
.compatible = "nvidia,tegra210-spi-slave",
|
|
.data = &tegra210_spi_chip_data,
|
|
}, {
|
|
.compatible = "nvidia,tegra186-spi-slave",
|
|
.data = &tegra186_spi_chip_data,
|
|
},
|
|
{}
|
|
};
|
|
MODULE_DEVICE_TABLE(of, tegra_spi_of_match);
|
|
|
|
static int tegra_spi_probe(struct platform_device *pdev)
|
|
{
|
|
struct spi_master *master;
|
|
struct tegra_spi_data *tspi;
|
|
struct resource *r;
|
|
struct tegra_spi_platform_data *pdata = pdev->dev.platform_data;
|
|
const struct of_device_id *match;
|
|
const struct tegra_spi_chip_data *chip_data = &tegra124_spi_chip_data;
|
|
int ret, spi_irq;
|
|
int bus_num;
|
|
int deassert_val;
|
|
|
|
if (pdev->dev.of_node) {
|
|
bus_num = of_alias_get_id(pdev->dev.of_node, "spi");
|
|
if (bus_num < 0) {
|
|
dev_warn(&pdev->dev,
|
|
"Dynamic bus number will be registerd\n");
|
|
bus_num = -1;
|
|
}
|
|
} else {
|
|
bus_num = pdev->id;
|
|
}
|
|
|
|
if (!pdata && pdev->dev.of_node)
|
|
pdata = tegra_spi_parse_dt(pdev);
|
|
|
|
if (!pdata) {
|
|
dev_err(&pdev->dev, "No platform data, exiting\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
if (!pdata->spi_max_frequency)
|
|
pdata->spi_max_frequency = 25000000; /* 25MHz */
|
|
|
|
if (pdata->rx_trig_words != 0 &&
|
|
pdata->rx_trig_words != 4 && pdata->rx_trig_words != 8)
|
|
pdata->rx_trig_words = 0;
|
|
|
|
master = spi_alloc_master(&pdev->dev, sizeof(*tspi));
|
|
if (!master) {
|
|
dev_err(&pdev->dev, "master allocation failed\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* the spi->mode bits understood by this driver: */
|
|
master->mode_bits = SPI_CPOL | SPI_CPHA | SPI_CS_HIGH | SPI_LSB_FIRST;
|
|
/* supported bpw 8-32 */
|
|
master->bits_per_word_mask = (u32) ~(BIT(0)|BIT(1)|
|
|
BIT(2)|BIT(3)|BIT(4)|BIT(5)|BIT(6));
|
|
master->setup = tegra_spi_setup;
|
|
master->transfer_one_message = tegra_spi_transfer_one_message;
|
|
master->num_chipselect = MAX_CHIP_SELECT;
|
|
master->bus_num = bus_num;
|
|
|
|
dev_set_drvdata(&pdev->dev, master);
|
|
tspi = spi_master_get_devdata(master);
|
|
tspi->master = master;
|
|
tspi->clock_always_on = pdata->is_clkon_always;
|
|
tspi->rx_trig_words = pdata->rx_trig_words;
|
|
tspi->dev = &pdev->dev;
|
|
dev_dbg(&pdev->dev, "rx-trigger:%d clock-always-on:%d\n",
|
|
tspi->rx_trig_words, tspi->clock_always_on);
|
|
|
|
if (pdev->dev.of_node) {
|
|
match = of_match_device(tegra_spi_of_match,
|
|
&pdev->dev);
|
|
if (match)
|
|
chip_data = match->data;
|
|
}
|
|
tspi->chip_data = chip_data;
|
|
tspi->gpio_slave_ready = pdata->gpio_slave_ready;
|
|
|
|
if (gpio_is_valid(tspi->gpio_slave_ready))
|
|
if (gpio_cansleep(tspi->gpio_slave_ready)) {
|
|
dev_err(&pdev->dev, "Slave Ready GPIO %d is unusable as it can sleep\n",
|
|
tspi->gpio_slave_ready);
|
|
ret = -EINVAL;
|
|
goto exit_free_master;
|
|
}
|
|
|
|
tspi->slave_ready_active_high = pdata->slave_ready_active_high;
|
|
|
|
if (gpio_is_valid(tspi->gpio_slave_ready)) {
|
|
ret = devm_gpio_request(&pdev->dev,
|
|
tspi->gpio_slave_ready, "gpio-spi-slave-ready");
|
|
if (!ret) {
|
|
if (tspi->slave_ready_active_high)
|
|
deassert_val = 0;
|
|
else
|
|
deassert_val = 1;
|
|
|
|
gpio_direction_output(tspi->gpio_slave_ready,
|
|
deassert_val);
|
|
} else {
|
|
dev_err(&pdev->dev, "Slave Ready GPIO %d is busy\n",
|
|
tspi->gpio_slave_ready);
|
|
goto exit_free_master;
|
|
}
|
|
}
|
|
|
|
spin_lock_init(&tspi->lock);
|
|
|
|
r = platform_get_resource(pdev, IORESOURCE_MEM, 0);
|
|
if (!r) {
|
|
dev_err(&pdev->dev, "No IO memory resource\n");
|
|
ret = -ENODEV;
|
|
goto exit_free_master;
|
|
}
|
|
tspi->phys = r->start;
|
|
tspi->base = devm_ioremap_resource(&pdev->dev, r);
|
|
if (IS_ERR(tspi->base)) {
|
|
dev_err(&pdev->dev,
|
|
"Cannot request memregion/iomap dma address\n");
|
|
ret = PTR_ERR(tspi->base);
|
|
goto exit_free_master;
|
|
}
|
|
|
|
tspi->rstc = devm_reset_control_get(&pdev->dev, "spi");
|
|
if (IS_ERR(tspi->rstc)) {
|
|
dev_err(&pdev->dev, "Reset control is not found\n");
|
|
return PTR_ERR(tspi->rstc);
|
|
}
|
|
|
|
atomic_set(&tspi->isr_expected, 0);
|
|
spi_irq = platform_get_irq(pdev, 0);
|
|
tspi->irq = spi_irq;
|
|
ret = request_irq(tspi->irq, tegra_spi_isr, 0,
|
|
dev_name(&pdev->dev), tspi);
|
|
if (ret < 0) {
|
|
dev_err(&pdev->dev, "Failed to register ISR for IRQ %d\n",
|
|
tspi->irq);
|
|
goto exit_free_master;
|
|
}
|
|
|
|
tspi->clk = devm_clk_get(&pdev->dev, "spi");
|
|
if (IS_ERR(tspi->clk)) {
|
|
dev_err(&pdev->dev, "can not get clock\n");
|
|
ret = PTR_ERR(tspi->clk);
|
|
goto exit_free_irq;
|
|
}
|
|
|
|
tspi->max_buf_size = SPI_FIFO_DEPTH << 2;
|
|
tspi->dma_buf_size = DEFAULT_SPI_DMA_BUF_LEN;
|
|
tspi->spi_max_frequency = pdata->spi_max_frequency;
|
|
|
|
ret = tegra_spi_init_dma_param(tspi, true);
|
|
if (ret < 0) {
|
|
if (ret == -EPROBE_DEFER)
|
|
/* GPCDMA is not available at this point.
|
|
* probe will be deferred */
|
|
dev_info(&pdev->dev,
|
|
"Probe Deferred, ret %d\n", ret);
|
|
else
|
|
dev_err(&pdev->dev,
|
|
"RxDma Init failed, err %d\n", ret);
|
|
|
|
goto exit_free_irq;
|
|
}
|
|
|
|
ret = tegra_spi_init_dma_param(tspi, false);
|
|
if (ret < 0) {
|
|
dev_err(&pdev->dev, "TxDma Init failed, err %d\n", ret);
|
|
goto exit_rx_dma_free;
|
|
}
|
|
tspi->max_buf_size = tspi->dma_buf_size;
|
|
init_completion(&tspi->tx_dma_complete);
|
|
init_completion(&tspi->rx_dma_complete);
|
|
init_completion(&tspi->xfer_completion);
|
|
|
|
if (tspi->clock_always_on) {
|
|
ret = clk_prepare_enable(tspi->clk);
|
|
if (ret < 0) {
|
|
dev_err(tspi->dev, "clk_prepare failed: %d\n", ret);
|
|
goto exit_deinit_dma;
|
|
}
|
|
}
|
|
tspi->reset_ctrl_status = false;
|
|
pm_runtime_enable(&pdev->dev);
|
|
if (!pm_runtime_enabled(&pdev->dev)) {
|
|
ret = tegra_spi_runtime_resume(&pdev->dev);
|
|
if (ret)
|
|
goto exit_pm_disable;
|
|
}
|
|
|
|
ret = pm_runtime_get_sync(&pdev->dev);
|
|
if (ret < 0) {
|
|
dev_err(&pdev->dev, "pm runtime get failed, e = %d\n", ret);
|
|
goto exit_pm_disable;
|
|
}
|
|
reset_control_reset(tspi->rstc);
|
|
tegra_spi_ext_clk_enable(false, tspi);
|
|
tspi->def_command1_reg = SPI_CS_SW_HW | SPI_CS_POL_INACTIVE_0 |
|
|
SPI_CS_POL_INACTIVE_1 | SPI_CS_POL_INACTIVE_2 |
|
|
SPI_CS_POL_INACTIVE_3 | SPI_CS_SS_VAL | SPI_CONTROL_MODE_1;
|
|
|
|
tegra_spi_writel(tspi, tspi->def_command1_reg, SPI_COMMAND1);
|
|
/* set CS inactive b/w packets to mask CS_INACTIVE interrupts */
|
|
if (tspi->chip_data->mask_cs_inactive_intr)
|
|
tegra_spi_writel(tspi, 0, SPI_CS_TIMING2);
|
|
pm_runtime_put(&pdev->dev);
|
|
|
|
master->dev.of_node = pdev->dev.of_node;
|
|
ret = spi_register_master(master);
|
|
if (ret < 0) {
|
|
dev_err(&pdev->dev, "can not register to master err %d\n", ret);
|
|
goto exit_pm_disable;
|
|
}
|
|
|
|
#ifdef TEGRA_SPI_SLAVE_DEBUG
|
|
ret = device_create_file(&pdev->dev, &dev_attr_force_unpacked_mode);
|
|
if (ret != 0)
|
|
goto exit_unregister_master;
|
|
|
|
return ret;
|
|
|
|
exit_unregister_master:
|
|
spi_unregister_master(master);
|
|
|
|
#endif
|
|
return ret;
|
|
|
|
exit_pm_disable:
|
|
pm_runtime_disable(&pdev->dev);
|
|
if (!pm_runtime_status_suspended(&pdev->dev))
|
|
tegra_spi_runtime_suspend(&pdev->dev);
|
|
if (tspi->clock_always_on)
|
|
clk_disable_unprepare(tspi->clk);
|
|
exit_deinit_dma:
|
|
tegra_spi_deinit_dma_param(tspi, false);
|
|
exit_rx_dma_free:
|
|
tegra_spi_deinit_dma_param(tspi, true);
|
|
exit_free_irq:
|
|
free_irq(spi_irq, tspi);
|
|
exit_free_master:
|
|
spi_master_put(master);
|
|
return ret;
|
|
}
|
|
|
|
static int tegra_spi_remove(struct platform_device *pdev)
|
|
{
|
|
struct spi_master *master = dev_get_drvdata(&pdev->dev);
|
|
struct tegra_spi_data *tspi = spi_master_get_devdata(master);
|
|
|
|
#ifdef TEGRA_SPI_SLAVE_DEBUG
|
|
device_remove_file(&pdev->dev, &dev_attr_force_unpacked_mode);
|
|
#endif
|
|
free_irq(tspi->irq, tspi);
|
|
spi_unregister_master(master);
|
|
|
|
if (tspi->tx_dma_chan)
|
|
tegra_spi_deinit_dma_param(tspi, false);
|
|
|
|
if (tspi->rx_dma_chan)
|
|
tegra_spi_deinit_dma_param(tspi, true);
|
|
|
|
pm_runtime_disable(&pdev->dev);
|
|
if (!pm_runtime_status_suspended(&pdev->dev))
|
|
tegra_spi_runtime_suspend(&pdev->dev);
|
|
|
|
if (tspi->clock_always_on)
|
|
clk_disable_unprepare(tspi->clk);
|
|
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_PM_SLEEP
|
|
static int tegra_spi_suspend(struct device *dev)
|
|
{
|
|
struct spi_master *master = dev_get_drvdata(dev);
|
|
struct tegra_spi_data *tspi = spi_master_get_devdata(master);
|
|
int ret;
|
|
|
|
ret = spi_master_suspend(master);
|
|
|
|
if (tspi->clock_always_on)
|
|
clk_disable_unprepare(tspi->clk);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int tegra_spi_resume(struct device *dev)
|
|
{
|
|
struct spi_master *master = dev_get_drvdata(dev);
|
|
struct tegra_spi_data *tspi = spi_master_get_devdata(master);
|
|
int ret;
|
|
|
|
if (tspi->clock_always_on) {
|
|
ret = clk_prepare_enable(tspi->clk);
|
|
if (ret < 0) {
|
|
dev_err(tspi->dev, "clk_prepare failed: %d\n", ret);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
ret = pm_runtime_get_sync(dev);
|
|
if (ret < 0) {
|
|
dev_err(dev, "pm runtime failed, e = %d\n", ret);
|
|
return ret;
|
|
}
|
|
tegra_spi_writel(tspi, tspi->command1_reg, SPI_COMMAND1);
|
|
pm_runtime_put(dev);
|
|
|
|
return spi_master_resume(master);
|
|
}
|
|
#endif
|
|
|
|
static int tegra_spi_runtime_suspend(struct device *dev)
|
|
{
|
|
struct spi_master *master = dev_get_drvdata(dev);
|
|
struct tegra_spi_data *tspi = spi_master_get_devdata(master);
|
|
|
|
/* Flush all write which are in PPSB queue by reading back */
|
|
tegra_spi_readl(tspi, SPI_COMMAND1);
|
|
|
|
clk_disable_unprepare(tspi->clk);
|
|
return 0;
|
|
}
|
|
|
|
static int tegra_spi_runtime_resume(struct device *dev)
|
|
{
|
|
struct spi_master *master = dev_get_drvdata(dev);
|
|
struct tegra_spi_data *tspi = spi_master_get_devdata(master);
|
|
int ret;
|
|
|
|
ret = clk_prepare_enable(tspi->clk);
|
|
if (ret < 0) {
|
|
dev_err(tspi->dev, "clk_prepare failed: %d\n", ret);
|
|
return ret;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static const struct dev_pm_ops tegra_spi_pm_ops = {
|
|
SET_RUNTIME_PM_OPS(tegra_spi_runtime_suspend,
|
|
tegra_spi_runtime_resume, NULL)
|
|
SET_SYSTEM_SLEEP_PM_OPS(tegra_spi_suspend, tegra_spi_resume)
|
|
};
|
|
static struct platform_driver tegra_spi_driver = {
|
|
.driver = {
|
|
.name = "spi-tegra124-slave",
|
|
.owner = THIS_MODULE,
|
|
.pm = &tegra_spi_pm_ops,
|
|
.of_match_table = of_match_ptr(tegra_spi_of_match),
|
|
},
|
|
.probe = tegra_spi_probe,
|
|
.remove = tegra_spi_remove,
|
|
};
|
|
module_platform_driver(tegra_spi_driver);
|
|
|
|
MODULE_ALIAS("platform:spi-tegra124");
|
|
MODULE_DESCRIPTION("NVIDIA Tegra124 SPI Controller Driver");
|
|
MODULE_AUTHOR("Laxman Dewangan <ldewangan@nvidia.com> Yousuf A <yousufa@nvidia.com>");
|
|
MODULE_LICENSE("GPL v2");
|