5189 lines
140 KiB
C
5189 lines
140 KiB
C
/*
|
|
* PCIe host controller driver for TEGRA SOCs
|
|
*
|
|
* Copyright (c) 2010, CompuLab, Ltd.
|
|
* Author: Mike Rapoport <mike@compulab.co.il>
|
|
*
|
|
* Based on NVIDIA PCIe driver
|
|
* Copyright (c) 2008-2018, NVIDIA Corporation. All rights reserved.
|
|
*
|
|
* Bits taken from arch/arm/mach-dove/pcie.c
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along
|
|
* with this program; if not, write to the Free Software Foundation, Inc.,
|
|
* 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/debugfs.h>
|
|
#include <linux/uaccess.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/pci-aspm.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/irqdomain.h>
|
|
#include <linux/clk.h>
|
|
#include <linux/reset.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/export.h>
|
|
#include <linux/msi.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/regulator/consumer.h>
|
|
#include <linux/workqueue.h>
|
|
#include <linux/gpio.h>
|
|
#include <linux/clk.h>
|
|
#include <linux/async.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <linux/pm_runtime.h>
|
|
#include <soc/tegra/chip-id.h>
|
|
#include <linux/of_device.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/of_gpio.h>
|
|
#include <linux/of_pci.h>
|
|
#include <linux/tegra_prod.h>
|
|
#include <linux/pinctrl/pinctrl.h>
|
|
#include <linux/pinctrl/consumer.h>
|
|
#include <linux/platform/tegra/emc_bwmgr.h>
|
|
#include <linux/pm_clock.h>
|
|
#ifdef CONFIG_THERMAL
|
|
#include <linux/thermal.h>
|
|
#endif
|
|
|
|
#include <asm/sizes.h>
|
|
#include <asm/mach/pci.h>
|
|
#include <asm/io.h>
|
|
|
|
#include <linux/phy/phy.h>
|
|
#include <linux/pci-tegra.h>
|
|
|
|
#define PCI_CFG_SPACE_SIZE 256
|
|
#define PCI_EXT_CFG_SPACE_SIZE 4096
|
|
|
|
#define AFI_AXI_BAR0_SZ 0x00
|
|
#define AFI_AXI_BAR1_SZ 0x04
|
|
#define AFI_AXI_BAR2_SZ 0x08
|
|
#define AFI_AXI_BAR3_SZ 0x0c
|
|
#define AFI_AXI_BAR4_SZ 0x10
|
|
#define AFI_AXI_BAR5_SZ 0x14
|
|
|
|
#define AFI_AXI_BAR0_START 0x18
|
|
#define AFI_AXI_BAR1_START 0x1c
|
|
#define AFI_AXI_BAR2_START 0x20
|
|
#define AFI_AXI_BAR3_START 0x24
|
|
#define AFI_AXI_BAR4_START 0x28
|
|
#define AFI_AXI_BAR5_START 0x2c
|
|
|
|
#define AFI_FPCI_BAR0 0x30
|
|
#define AFI_FPCI_BAR1 0x34
|
|
#define AFI_FPCI_BAR2 0x38
|
|
#define AFI_FPCI_BAR3 0x3c
|
|
#define AFI_FPCI_BAR4 0x40
|
|
#define AFI_FPCI_BAR5 0x44
|
|
|
|
#define AFI_MSI_BAR_SZ 0x60
|
|
#define AFI_MSI_FPCI_BAR_ST 0x64
|
|
#define AFI_MSI_AXI_BAR_ST 0x68
|
|
|
|
#define AFI_MSI_VEC0_0 0x6c
|
|
#define AFI_MSI_VEC1_0 0x70
|
|
#define AFI_MSI_VEC2_0 0x74
|
|
#define AFI_MSI_VEC3_0 0x78
|
|
#define AFI_MSI_VEC4_0 0x7c
|
|
#define AFI_MSI_VEC5_0 0x80
|
|
#define AFI_MSI_VEC6_0 0x84
|
|
#define AFI_MSI_VEC7_0 0x88
|
|
|
|
#define AFI_MSI_EN_VEC0_0 0x8c
|
|
#define AFI_MSI_EN_VEC1_0 0x90
|
|
#define AFI_MSI_EN_VEC2_0 0x94
|
|
#define AFI_MSI_EN_VEC3_0 0x98
|
|
#define AFI_MSI_EN_VEC4_0 0x9c
|
|
#define AFI_MSI_EN_VEC5_0 0xa0
|
|
#define AFI_MSI_EN_VEC6_0 0xa4
|
|
#define AFI_MSI_EN_VEC7_0 0xa8
|
|
|
|
#define AFI_CONFIGURATION 0xac
|
|
#define AFI_CONFIGURATION_EN_FPCI (1 << 0)
|
|
#define AFI_CONFIGURATION_CLKEN_OVERRIDE (1 << 31)
|
|
|
|
#define AFI_FPCI_ERROR_MASKS 0xb0
|
|
|
|
#define AFI_INTR_MASK 0xb4
|
|
#define AFI_INTR_MASK_INT_MASK (1 << 0)
|
|
#define AFI_INTR_MASK_MSI_MASK (1 << 8)
|
|
|
|
#define AFI_INTR_CODE 0xb8
|
|
#define AFI_INTR_CODE_MASK 0x1f
|
|
#define AFI_INTR_MASTER_ABORT 4
|
|
#define AFI_INTR_LEGACY 6
|
|
#define AFI_INTR_PRSNT_SENSE 10
|
|
|
|
#define AFI_INTR_SIGNATURE 0xbc
|
|
#define AFI_SM_INTR_ENABLE 0xc4
|
|
|
|
#define AFI_AFI_INTR_ENABLE 0xc8
|
|
#define AFI_INTR_EN_INI_SLVERR (1 << 0)
|
|
#define AFI_INTR_EN_INI_DECERR (1 << 1)
|
|
#define AFI_INTR_EN_TGT_SLVERR (1 << 2)
|
|
#define AFI_INTR_EN_TGT_DECERR (1 << 3)
|
|
#define AFI_INTR_EN_TGT_WRERR (1 << 4)
|
|
#define AFI_INTR_EN_DFPCI_DECERR (1 << 5)
|
|
#define AFI_INTR_EN_AXI_DECERR (1 << 6)
|
|
#define AFI_INTR_EN_FPCI_TIMEOUT (1 << 7)
|
|
|
|
#define AFI_PCIE_PME 0x0f0
|
|
#define AFI_PCIE_PME_TURN_OFF 0x101
|
|
#define AFI_PCIE_PME_ACK 0x420
|
|
|
|
#define AFI_PCIE_CONFIG 0x0f8
|
|
#define AFI_PCIE_CONFIG_PCIEC0_DISABLE_DEVICE (1 << 1)
|
|
#define AFI_PCIE_CONFIG_PCIEC1_DISABLE_DEVICE (1 << 2)
|
|
#define AFI_PCIE_CONFIG_PCIEC2_DISABLE_DEVICE (1 << 3)
|
|
#define AFI_PCIE_CONFIG_XBAR_CONFIG_MASK (0xf << 20)
|
|
#define AFI_PCIE_CONFIG_XBAR_CONFIG_X2_X1 (0x0 << 20)
|
|
#define AFI_PCIE_CONFIG_XBAR_CONFIG_X4_X1 (0x1 << 20)
|
|
#define AFI_PCIE_CONFIG_XBAR_CONFIG_X4_X0_X1 (0x0 << 20)
|
|
#define AFI_PCIE_CONFIG_XBAR_CONFIG_X2_X1_X1 (0x1 << 20)
|
|
#define AFI_PCIE_CONFIG_XBAR_CONFIG_X1_X1_X1 (0x2 << 20)
|
|
#define AFI_PCIE_CONFIG_PCIEC0_CLKREQ_AS_GPIO BIT(29)
|
|
#define AFI_PCIE_CONFIG_PCIEC1_CLKREQ_AS_GPIO BIT(30)
|
|
#define AFI_PCIE_CONFIG_PCIEC2_CLKREQ_AS_GPIO BIT(31)
|
|
|
|
#define AFI_FUSE 0x104
|
|
#define AFI_FUSE_PCIE_T0_GEN2_DIS (1 << 2)
|
|
|
|
#define AFI_PEX0_CTRL 0x110
|
|
#define AFI_PEX1_CTRL 0x118
|
|
#define AFI_PEX2_CTRL 0x19C
|
|
#define AFI_PEX_CTRL_RST (1 << 0)
|
|
#define AFI_PEX_CTRL_CLKREQ_EN (1 << 1)
|
|
#define AFI_PEX_CTRL_REFCLK_EN (1 << 3)
|
|
#define AFI_PEX_CTRL_OVERRIDE_EN (1 << 4)
|
|
|
|
#define AFI_PLLE_CONTROL 0x160
|
|
#define AFI_PLLE_CONTROL_BYPASS_PADS2PLLE_CONTROL (1 << 9)
|
|
#define AFI_PLLE_CONTROL_BYPASS_PCIE2PLLE_CONTROL (1 << 8)
|
|
#define AFI_PLLE_CONTROL_PADS2PLLE_CONTROL_EN (1 << 1)
|
|
#define AFI_PLLE_CONTROL_PCIE2PLLE_CONTROL_EN (1 << 0)
|
|
|
|
#define AFI_PEXBIAS_CTRL_0 0x168
|
|
#define AFI_WR_SCRATCH_0 0x120
|
|
#define AFI_WR_SCRATCH_0_RESET_VAL 0x00202020
|
|
#define AFI_WR_SCRATCH_0_DEFAULT_VAL 0x00000000
|
|
|
|
#define AFI_MSG_0 0x190
|
|
#define AFI_MSG_PM_PME_MASK 0x00100010
|
|
#define AFI_MSG_INTX_MASK 0x1f001f00
|
|
#define AFI_MSG_PM_PME0 (1 << 4)
|
|
#define AFI_MSG_PM_PME1 (1 << 20)
|
|
#define AFI_MSG_RP_INT_MASK 0x10001000
|
|
|
|
#define AFI_MSG_1_0 0x194
|
|
#define AFI_MSG_1_PM_PME_MASK 0x00000010
|
|
#define AFI_MSG_1_INTX_MASK 0x00001f00
|
|
#define AFI_MSG_1_PM_PME (1 << 4)
|
|
|
|
#define RP_VEND_XP 0x00000F00
|
|
#define RP_VEND_XP_OPPORTUNISTIC_ACK (1 << 27)
|
|
#define RP_VEND_XP_OPPORTUNISTIC_UPDATEFC (1 << 28)
|
|
#define RP_VEND_XP_DL_UP (1 << 30)
|
|
#define RP_VEND_XP_UPDATE_FC_THRESHOLD (0xFF << 18)
|
|
#define RP_VEND_XP_PRBS_STAT (0xFFFF << 2)
|
|
#define RP_VEND_XP_PRBS_EN (1 << 1)
|
|
|
|
#define RP_LINK_CONTROL_STATUS 0x00000090
|
|
#define RP_LINK_CONTROL_STATUS_BW_MGMT_STATUS 0x40000000
|
|
#define RP_LINK_CONTROL_STATUS_DL_LINK_ACTIVE 0x20000000
|
|
#define RP_LINK_CONTROL_STATUS_LINKSTAT_MASK 0x3fff0000
|
|
#define RP_LINK_CONTROL_STATUS_NEG_LINK_WIDTH (0x3F << 20)
|
|
#define RP_LINK_CONTROL_STATUS_LINK_SPEED (0xF << 16)
|
|
#define RP_LINK_CONTROL_STATUS_L0s_ENABLED 0x00000001
|
|
#define RP_LINK_CONTROL_STATUS_L1_ENABLED 0x00000002
|
|
|
|
#define RP_LINK_CONTROL_STATUS_2 0x000000B0
|
|
#define RP_LINK_CONTROL_STATUS_2_TRGT_LNK_SPD_MASK 0x0000000F
|
|
#define RP_LINK_CONTROL_STATUS_2_TRGT_LNK_SPD_GEN1 0x00000001
|
|
#define RP_LINK_CONTROL_STATUS_2_TRGT_LNK_SPD_GEN2 0x00000002
|
|
|
|
#define NV_PCIE2_RP_RSR 0x000000A0
|
|
#define NV_PCIE2_RP_RSR_PMESTAT (1 << 16)
|
|
|
|
#define NV_PCIE2_RP_INTR_BCR 0x0000003C
|
|
#define NV_PCIE2_RP_INTR_BCR_INTR_LINE (0xFF << 0)
|
|
#define NV_PCIE2_RP_INTR_BCR_SB_RESET (0x1 << 22)
|
|
|
|
#define NV_PCIE2_RP_PRIV_XP_DL 0x00000494
|
|
#define PCIE2_RP_PRIV_XP_DL_GEN2_UPD_FC_TSHOLD (0x1FF << 1)
|
|
|
|
#define NV_PCIE2_RP_RX_HDR_LIMIT 0x00000E00
|
|
#define PCIE2_RP_RX_HDR_LIMIT_PW_MASK (0xFF00)
|
|
#define PCIE2_RP_RX_HDR_LIMIT_PW (0x0E << 8)
|
|
|
|
#define NV_PCIE2_RP_TX_HDR_LIMIT 0x00000E08
|
|
#define PCIE2_RP_TX_HDR_LIMIT_NPT_0 32
|
|
#define PCIE2_RP_TX_HDR_LIMIT_NPT_1 4
|
|
|
|
#define NV_PCIE2_RP_TIMEOUT0 0x00000E24
|
|
#define PCIE2_RP_TIMEOUT0_PAD_PWRUP_MASK (0xFF)
|
|
#define PCIE2_RP_TIMEOUT0_PAD_PWRUP (0xA)
|
|
#define PCIE2_RP_TIMEOUT0_PAD_PWRUP_CM_MASK (0xFFFF00)
|
|
#define PCIE2_RP_TIMEOUT0_PAD_PWRUP_CM (0x180 << 8)
|
|
#define PCIE2_RP_TIMEOUT0_PAD_SPDCHNG_GEN2_MASK (0xFF << 24)
|
|
#define PCIE2_RP_TIMEOUT0_PAD_SPDCHNG_GEN2 (0xA << 24)
|
|
|
|
#define NV_PCIE2_RP_TIMEOUT1 0x00000E28
|
|
#define PCIE2_RP_TIMEOUT1_RCVRY_SPD_SUCCESS_EIDLE_MASK (0xFF << 16)
|
|
#define PCIE2_RP_TIMEOUT1_RCVRY_SPD_SUCCESS_EIDLE (0x10 << 16)
|
|
#define PCIE2_RP_TIMEOUT1_RCVRY_SPD_UNSUCCESS_EIDLE_MASK (0xFF << 24)
|
|
#define PCIE2_RP_TIMEOUT1_RCVRY_SPD_UNSUCCESS_EIDLE (0x74 << 24)
|
|
|
|
#define NV_PCIE2_RP_PRBS 0x00000E34
|
|
#define PCIE2_RP_PRBS_LOCKED (1 << 16)
|
|
|
|
#define NV_PCIE2_RP_LANE_PRBS_ERR_COUNT 0x00000E38
|
|
#define PCIE2_RP_LANE_PRBS_ERR_COUNT (1 << 16)
|
|
#define PCIE2_RP_LANE_PRBS_ERR_SELECT (1 << 0)
|
|
|
|
#define NV_PCIE2_RP_LTSSM_DBGREG 0x00000E44
|
|
#define PCIE2_RP_LTSSM_DBGREG_LINKFSM15 (1 << 15)
|
|
#define PCIE2_RP_LTSSM_DBGREG_LINKFSM16 (1 << 16)
|
|
#define PCIE2_RP_LTSSM_DBGREG_LINKFSM17 (1 << 17)
|
|
|
|
#define NV_PCIE2_RP_LTSSM_TRACE_CONTROL 0x00000E50
|
|
#define LTSSM_TRACE_CONTROL_CLEAR_STORE_EN (1 << 0)
|
|
#define LTSSM_TRACE_CONTROL_CLEAR_RAM (1 << 2)
|
|
#define LTSSM_TRACE_CONTROL_TRIG_ON_EVENT (1 << 3)
|
|
#define LTSSM_TRACE_CONTROL_TRIG_LTSSM_MAJOR_OFFSET 4
|
|
#define LTSSM_TRACE_CONTROL_TRIG_PTX_LTSSM_MINOR_OFFSET 8
|
|
#define LTSSM_TRACE_CONTROL_TRIG_PRX_LTSSM_MAJOR_OFFSET 11
|
|
|
|
#define NV_PCIE2_RP_LTSSM_TRACE_STATUS 0x00000E54
|
|
#define LTSSM_TRACE_STATUS_PRX_MINOR(reg) (((reg) >> 19) & 0x7)
|
|
#define LTSSM_TRACE_STATUS_PTX_MINOR(reg) (((reg) >> 16) & 0x7)
|
|
#define LTSSM_TRACE_STATUS_MAJOR(reg) (((reg) >> 12) & 0xf)
|
|
#define LTSSM_TRACE_STATUS_READ_DATA_VALID(reg) (((reg) >> 11) & 0x1)
|
|
#define LTSSM_TRACE_STATUS_READ_ADDR(reg) ((reg) << 6)
|
|
#define LTSSM_TRACE_STATUS_WRITE_POINTER(reg) (((reg) >> 1) & 0x1f)
|
|
#define LTSSM_TRACE_STATUS_RAM_FULL(reg) (reg & 0x1)
|
|
|
|
|
|
#define NV_PCIE2_RP_XP_REF 0x00000F30
|
|
#define PCIE2_RP_XP_REF_MICROSECOND_LIMIT_MASK (0xFF)
|
|
#define PCIE2_RP_XP_REF_MICROSECOND_LIMIT (0x14)
|
|
#define PCIE2_RP_XP_REF_MICROSECOND_ENABLE (1 << 8)
|
|
#define PCIE2_RP_XP_REF_CPL_TO_OVERRIDE (1 << 13)
|
|
#define PCIE2_RP_XP_REF_CPL_TO_CUSTOM_VALUE_MASK (0x1FFFF << 14)
|
|
#define PCIE2_RP_XP_REF_CPL_TO_CUSTOM_VALUE (0x1770 << 14)
|
|
|
|
#define NV_PCIE2_RP_PRIV_MISC 0x00000FE0
|
|
#define PCIE2_RP_PRIV_MISC_PRSNT_MAP_EP_PRSNT (0xE << 0)
|
|
#define PCIE2_RP_PRIV_MISC_PRSNT_MAP_EP_ABSNT (0xF << 0)
|
|
#define PCIE2_RP_PRIV_MISC_CTLR_CLK_CLAMP_THRESHOLD (0xF << 16)
|
|
#define PCIE2_RP_PRIV_MISC_CTLR_CLK_CLAMP_ENABLE (1 << 23)
|
|
#define PCIE2_RP_PRIV_MISC_TMS_CLK_CLAMP_THRESHOLD (0xF << 24)
|
|
#define PCIE2_RP_PRIV_MISC_TMS_CLK_CLAMP_ENABLE (1 << 31)
|
|
|
|
#define NV_PCIE2_RP_VEND_XP1 0x00000F04
|
|
#define NV_PCIE2_RP_VEND_XP2 0x00000F08
|
|
#define NV_PCIE2_RP_VEND_XP_LINK_PVT_CTL_L1_ASPM_SUPPORT (1 << 21)
|
|
#define NV_PCIE2_RP_VEND_XP1_RNCTRL_MAXWIDTH_MASK (0x3F << 0)
|
|
#define NV_PCIE2_RP_VEND_XP1_RNCTRL_EN (1 << 7)
|
|
|
|
#define NV_PCIE2_RP_VEND_CTL0 0x00000F44
|
|
#define PCIE2_RP_VEND_CTL0_DSK_RST_PULSE_WIDTH_MASK (0xF << 12)
|
|
#define PCIE2_RP_VEND_CTL0_DSK_RST_PULSE_WIDTH (0x9 << 12)
|
|
|
|
#define NV_PCIE2_RP_VEND_CTL1 0x00000F48
|
|
#define PCIE2_RP_VEND_CTL1_ERPT (1 << 13)
|
|
|
|
#define NV_PCIE2_RP_VEND_XP_BIST 0x00000F4C
|
|
#define PCIE2_RP_VEND_XP_BIST_GOTO_L1_L2_AFTER_DLLP_DONE (1 << 28)
|
|
|
|
#define NV_PCIE2_RP_VEND_XP_PAD_PWRDN 0x00000F50
|
|
#define NV_PCIE2_RP_VEND_XP_PAD_PWRDN_L1_EN (1 << 0)
|
|
#define NV_PCIE2_RP_VEND_XP_PAD_PWRDN_DYNAMIC_EN (1 << 1)
|
|
#define NV_PCIE2_RP_VEND_XP_PAD_PWRDN_DISABLED_EN (1 << 2)
|
|
#define NV_PCIE2_RP_VEND_XP_PAD_PWRDN_L1_CLKREQ_EN (1 << 15)
|
|
#define NV_PCIE2_RP_VEND_XP_PAD_PWRDN_SLEEP_MODE_DYNAMIC_L1PP (3 << 5)
|
|
#define NV_PCIE2_RP_VEND_XP_PAD_PWRDN_SLEEP_MODE_L1_L1P (2 << 3)
|
|
#define NV_PCIE2_RP_VEND_XP_PAD_PWRDN_SLEEP_MODE_L1_L1PP (3 << 3)
|
|
#define NV_PCIE2_RP_VEND_XP_PAD_PWRDN_SLEEP_MODE_L1_CLKREQ_L1P (2 << 16)
|
|
#define NV_PCIE2_RP_VEND_XP_PAD_PWRDN_SLEEP_MODE_L1_CLKREQ_L1PP (3 << 16)
|
|
|
|
#define NV_PCIE2_RP_PRIV_XP_RX_L0S_ENTRY_COUNT 0x00000F8C
|
|
#define NV_PCIE2_RP_PRIV_XP_TX_L0S_ENTRY_COUNT 0x00000F90
|
|
#define NV_PCIE2_RP_PRIV_XP_TX_L1_ENTRY_COUNT 0x00000F94
|
|
#define NV_PCIE2_RP_LTR_REP_VAL 0x00000C10
|
|
#define NV_PCIE2_RP_L1_1_ENTRY_COUNT 0x00000C14
|
|
#define PCIE2_RP_L1_1_ENTRY_COUNT_RESET (1 << 31)
|
|
#define NV_PCIE2_RP_L1_2_ENTRY_COUNT 0x00000C18
|
|
#define PCIE2_RP_L1_2_ENTRY_COUNT_RESET (1 << 31)
|
|
|
|
#define NV_PCIE2_RP_VEND_CTL2 0x00000FA8
|
|
#define PCIE2_RP_VEND_CTL2_PCA_ENABLE (1 << 7)
|
|
|
|
#define NV_PCIE2_RP_PRIV_XP_CONFIG 0x00000FAC
|
|
#define NV_PCIE2_RP_PRIV_XP_CONFIG_LOW_PWR_DURATION_MASK 0x3
|
|
#define NV_PCIE2_RP_PRIV_XP_DURATION_IN_LOW_PWR_100NS 0x00000FB0
|
|
|
|
#define NV_PCIE2_RP_XP_CTL_1 0x00000FEC
|
|
#define PCIE2_RP_XP_CTL_1_OLD_IOBIST_EN_BIT25 (1 << 25)
|
|
#define PCIE2_RP_XP_CTL_1_SPARE_BIT29 (1 << 29)
|
|
|
|
#define NV_PCIE2_RP_L1_PM_SUBSTATES_CYA 0x00000C00
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_CYA_CM_RTIME_MASK (0xFF << 8)
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_CYA_CM_RTIME_SHIFT (8)
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_CYA_T_PWRN_SCL_MASK (0x3 << 16)
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_CYA_T_PWRN_SCL_SHIFT (16)
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_CYA_T_PWRN_VAL_MASK (0x1F << 19)
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_CYA_T_PWRN_VAL_SHIFT (19)
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_CYA_HIDE_CAP (0x1 << 24)
|
|
|
|
#define NV_PCIE2_RP_L1_PM_SUBSTATES_1_CYA 0x00000C04
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_1_CYA_PWR_OFF_DLY_MASK (0x1FFF)
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_1_CYA_PWR_OFF_DLY (0x26)
|
|
#define PCIE2_RP_L1SS_1_CYA_CLKREQ_ASSERTED_DLY_MASK (0x1FF << 13)
|
|
#define PCIE2_RP_L1SS_1_CYA_CLKREQ_ASSERTED_DLY (0x27 << 13)
|
|
|
|
#define NV_PCIE2_RP_L1_PM_SUBSTATES_2_CYA 0x00000C08
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_2_CYA_T_L1_2_DLY_MASK (0x1FFF)
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_2_CYA_T_L1_2_DLY (0x4D)
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_2_CYA_MICROSECOND_MASK (0xFF << 13)
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_2_CYA_MICROSECOND (0x13 << 13)
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_2_CYA_MICROSECOND_COMP_MASK (0xF << 21)
|
|
#define PCIE2_RP_L1_PM_SUBSTATES_2_CYA_MICROSECOND_COMP (0x2 << 21)
|
|
|
|
#define PCIE2_RP_L1SS_SPARE 0xC24
|
|
#define PCIE2_RP_L1SS_SPARE_LTR_MSG_INT_EN (1 << 16)
|
|
#define PCIE2_RP_L1SS_SPARE_LTR_MSG_RCV_STS (1 << 17)
|
|
|
|
#define PCIE2_RP_L1_PM_SS_CONTROL 0x00000148
|
|
#define PCIE2_RP_L1_PM_SS_CONTROL_ASPM_L11_ENABLE 0x00000008
|
|
#define PCIE2_RP_L1_PM_SS_CONTROL_ASPM_L12_ENABLE 0x00000004
|
|
|
|
#define INT_PCI_MSI_NR (32 * 8)
|
|
|
|
#define LINK_RETRAIN_TIMEOUT 100000
|
|
|
|
#define PCIE_LANES_X4_X1 0x14
|
|
|
|
#define DEBUG 0
|
|
#if DEBUG || defined(CONFIG_PCI_DEBUG)
|
|
#define PR_FUNC_LINE pr_info("PCIE: %s(%d)\n", __func__, __LINE__)
|
|
#else
|
|
#define PR_FUNC_LINE do {} while (0)
|
|
#endif
|
|
|
|
struct pcie_dvfs {
|
|
u32 afi_clk;
|
|
u32 emc_clk;
|
|
};
|
|
|
|
struct tegra_pcie_soc_data {
|
|
unsigned int num_ports;
|
|
char **pcie_regulator_names;
|
|
int num_pcie_regulators;
|
|
bool config_pex_io_dpd;
|
|
bool program_uphy;
|
|
bool program_clkreq_as_bi_dir;
|
|
bool enable_wrap;
|
|
bool mbist_war;
|
|
bool RAW_violation_war;
|
|
bool perf_war;
|
|
bool updateFC_timer_expire_war;
|
|
bool l1ss_rp_wakeup_war;
|
|
bool link_speed_war;
|
|
bool dvfs_mselect;
|
|
bool dvfs_afi;
|
|
bool update_clamp_threshold;
|
|
struct pcie_dvfs dvfs_tbl[10][2];
|
|
};
|
|
|
|
struct tegra_msi {
|
|
struct msi_controller chip;
|
|
DECLARE_BITMAP(used, INT_PCI_MSI_NR);
|
|
struct irq_domain *domain;
|
|
struct mutex lock;
|
|
void *virt;
|
|
u64 phys;
|
|
int irq;
|
|
};
|
|
|
|
static inline struct tegra_msi *to_tegra_msi(struct msi_controller *chip)
|
|
{
|
|
return container_of(chip, struct tegra_msi, chip);
|
|
}
|
|
|
|
struct tegra_pcie {
|
|
struct device *dev;
|
|
struct pci_host_bridge *host;
|
|
|
|
void __iomem *pads;
|
|
void __iomem *afi;
|
|
int irq;
|
|
|
|
struct list_head buses;
|
|
struct list_head sys;
|
|
struct resource *cs;
|
|
struct resource *afi_res;
|
|
struct resource *pads_res;
|
|
|
|
struct resource io;
|
|
struct resource pio;
|
|
struct resource mem;
|
|
struct resource prefetch;
|
|
struct resource busn;
|
|
|
|
struct {
|
|
resource_size_t mem;
|
|
resource_size_t io;
|
|
} offset;
|
|
|
|
void __iomem *cfg_va_base;
|
|
|
|
struct tegra_msi msi;
|
|
|
|
struct reset_control *afi_rst;
|
|
struct reset_control *pcie_rst;
|
|
struct reset_control *pciex_rst;
|
|
struct tegra_bwmgr_client *emc_bwmgr;
|
|
|
|
#ifdef CONFIG_THERMAL
|
|
struct thermal_cooling_device *cdev;
|
|
atomic_t therm_state;
|
|
struct completion completion;
|
|
bool is_cooling_dev;
|
|
#endif
|
|
|
|
struct list_head ports;
|
|
u32 xbar_config;
|
|
int num_ports;
|
|
|
|
int power_rails_enabled;
|
|
int pcie_power_enabled;
|
|
struct work_struct hotplug_detect;
|
|
|
|
struct regulator **pcie_regulators;
|
|
|
|
struct pinctrl *pex_pin;
|
|
struct pinctrl_state *pex_io_dpd_en_state;
|
|
struct pinctrl_state *pex_io_dpd_dis_state;
|
|
|
|
struct tegra_pci_platform_data *plat_data;
|
|
struct tegra_pcie_soc_data *soc_data;
|
|
struct dentry *debugfs;
|
|
struct delayed_work detect_delay;
|
|
struct tegra_prod *prod_list;
|
|
};
|
|
|
|
static int tegra_pcie_enable_msi(struct tegra_pcie *pcie, bool no_init);
|
|
static void tegra_pcie_check_ports(struct tegra_pcie *pcie);
|
|
static void tegra_pcie_link_speed(struct tegra_pcie *pcie);
|
|
static int tegra_pcie_power_off(struct tegra_pcie *pcie);
|
|
|
|
#define MAX_PWR_GPIOS 5
|
|
|
|
struct tegra_pcie_port {
|
|
struct tegra_pcie *pcie;
|
|
struct list_head list;
|
|
struct resource regs;
|
|
void __iomem *base;
|
|
unsigned int index;
|
|
unsigned int lanes;
|
|
unsigned int num_lanes;
|
|
unsigned int loopback_stat;
|
|
int gpio_presence_detection;
|
|
bool disable_clock_request;
|
|
bool ep_status;
|
|
int status;
|
|
int rst_gpio;
|
|
bool has_mxm_port;
|
|
int pwr_gd_gpio;
|
|
struct dentry *port_debugfs;
|
|
struct phy **phy;
|
|
struct device_node *np;
|
|
int n_gpios;
|
|
int *gpios;
|
|
};
|
|
|
|
struct tegra_pcie_bus {
|
|
struct list_head list;
|
|
unsigned int nr;
|
|
};
|
|
|
|
/* used to avoid successive hotplug disconnect or connect */
|
|
static bool hotplug_event;
|
|
/* pcie mselect, xclk and emc rate */
|
|
static u16 bdf;
|
|
static u16 config_offset;
|
|
static u32 config_val;
|
|
static u16 config_aspm_state;
|
|
|
|
static inline void afi_writel(struct tegra_pcie *pcie, u32 value,
|
|
unsigned long offset)
|
|
{
|
|
writel(value, offset + pcie->afi);
|
|
}
|
|
|
|
static inline u32 afi_readl(struct tegra_pcie *pcie, unsigned long offset)
|
|
{
|
|
return readl(offset + pcie->afi);
|
|
}
|
|
|
|
static inline void __maybe_unused pads_writel(struct tegra_pcie *pcie, u32 value,
|
|
unsigned long offset)
|
|
{
|
|
writel(value, offset + pcie->pads);
|
|
}
|
|
|
|
static inline u32 __maybe_unused pads_readl(struct tegra_pcie *pcie, unsigned long offset)
|
|
{
|
|
return readl(offset + pcie->pads);
|
|
}
|
|
|
|
static inline void rp_writel(struct tegra_pcie_port *port, u32 value,
|
|
unsigned long offset)
|
|
{
|
|
writel(value, offset + port->base);
|
|
}
|
|
|
|
static inline unsigned int rp_readl(struct tegra_pcie_port *port,
|
|
unsigned long offset)
|
|
{
|
|
return readl(offset + port->base);
|
|
}
|
|
|
|
/*
|
|
* The configuration space mapping on Tegra is somewhat similar to the ECAM
|
|
* defined by PCIe. However it deviates a bit in how the 4 bits for extended
|
|
* register accesses are mapped:
|
|
*
|
|
* [27:24] extended register number
|
|
* [23:16] bus number
|
|
* [15:11] device number
|
|
* [10: 8] function number
|
|
* [ 7: 0] register number
|
|
*
|
|
* Mapping the whole extended configuration space would required 256 MiB of
|
|
* virtual address space, only a small part of which will actually be used.
|
|
* To work around this, a 4K of region is used to generate required
|
|
* configuration transaction with relevant B:D:F values. This is achieved by
|
|
* dynamically programming base address and size of AFI_AXI_BAR used for
|
|
* end point config space mapping to make sure that the address (access to
|
|
* which generates correct config transaction) falls in this 4K region
|
|
*/
|
|
static struct tegra_pcie_bus *tegra_pcie_bus_alloc(struct tegra_pcie *pcie,
|
|
unsigned int busnr)
|
|
{
|
|
struct tegra_pcie_bus *bus;
|
|
|
|
PR_FUNC_LINE;
|
|
bus = devm_kzalloc(pcie->dev, sizeof(*bus), GFP_KERNEL);
|
|
if (!bus)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
INIT_LIST_HEAD(&bus->list);
|
|
bus->nr = busnr;
|
|
|
|
if (!pcie->cfg_va_base) {
|
|
pcie->cfg_va_base = ioremap(pcie->cs->start, SZ_4K);
|
|
if (!pcie->cfg_va_base) {
|
|
dev_err(pcie->dev, "failed to ioremap config space\n");
|
|
kfree(bus);
|
|
bus = (struct tegra_pcie_bus *)-ENOMEM;
|
|
}
|
|
}
|
|
|
|
return bus;
|
|
}
|
|
|
|
static void __iomem *tegra_pcie_conf_address(struct pci_bus *bus,
|
|
unsigned int devfn,
|
|
int where)
|
|
{
|
|
struct pci_host_bridge *host = pci_find_host_bridge(bus);
|
|
struct tegra_pcie *pcie = pci_host_bridge_priv(host);
|
|
void __iomem *addr = NULL;
|
|
u32 val = 0;
|
|
|
|
if (bus->number == 0) {
|
|
unsigned int slot = PCI_SLOT(devfn);
|
|
struct tegra_pcie_port *port;
|
|
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
if ((port->index + 1 == slot) && port->status) {
|
|
addr = port->base + (where & ~3);
|
|
break;
|
|
}
|
|
}
|
|
} else {
|
|
addr = pcie->cfg_va_base;
|
|
val = ((((u32)where & 0xf00) >> 8) << 24) |
|
|
(bus->number << 16) | (PCI_SLOT(devfn) << 11) |
|
|
(PCI_FUNC(devfn) << 8) | (where & 0xff);
|
|
addr = (val & (SZ_4K - 1)) + addr;
|
|
val = val & ~(SZ_4K - 1);
|
|
afi_writel(pcie, pcie->cs->start - val, AFI_AXI_BAR0_START);
|
|
afi_writel(pcie, (val + SZ_4K) >> 12, AFI_AXI_BAR0_SZ);
|
|
}
|
|
|
|
return addr;
|
|
}
|
|
|
|
static int tegra_pcie_read_conf(struct pci_bus *bus, unsigned int devfn,
|
|
int where, int size, u32 *value)
|
|
{
|
|
struct pci_host_bridge *host = pci_find_host_bridge(bus);
|
|
struct tegra_pcie *pcie = pci_host_bridge_priv(host);
|
|
struct tegra_pcie_port *port = NULL;
|
|
u32 rp = 0;
|
|
struct pci_dev *dn_dev;
|
|
|
|
if (bus->number) {
|
|
if (bus->number == 1) {
|
|
dn_dev = bus->self;
|
|
rp = PCI_SLOT(dn_dev->devfn);
|
|
list_for_each_entry(port, &pcie->ports, list)
|
|
if (rp == port->index + 1)
|
|
break;
|
|
if (!port->ep_status)
|
|
return PCIBIOS_DEVICE_NOT_FOUND;
|
|
}
|
|
return pci_generic_config_read(bus, devfn, where, size, value);
|
|
}
|
|
return pci_generic_config_read32(bus, devfn, where, size, value);
|
|
}
|
|
|
|
static int tegra_pcie_write_conf(struct pci_bus *bus, unsigned int devfn,
|
|
int where, int size, u32 value)
|
|
{
|
|
struct pci_host_bridge *host = pci_find_host_bridge(bus);
|
|
struct tegra_pcie *pcie = pci_host_bridge_priv(host);
|
|
struct tegra_pcie_port *port = NULL;
|
|
u32 rp = 0;
|
|
struct pci_dev *dn_dev;
|
|
|
|
if (bus->number) {
|
|
if (bus->number == 1) {
|
|
dn_dev = bus->self;
|
|
rp = PCI_SLOT(dn_dev->devfn);
|
|
list_for_each_entry(port, &pcie->ports, list)
|
|
if (rp == port->index + 1)
|
|
break;
|
|
if (!port->ep_status)
|
|
return PCIBIOS_DEVICE_NOT_FOUND;
|
|
}
|
|
return pci_generic_config_write(bus, devfn, where, size, value);
|
|
}
|
|
return pci_generic_config_write32(bus, devfn, where, size, value);
|
|
}
|
|
|
|
static void tegra_pcie_fixup_bridge(struct pci_dev *dev)
|
|
{
|
|
u16 reg;
|
|
|
|
if ((dev->class >> 16) == PCI_BASE_CLASS_BRIDGE) {
|
|
pci_read_config_word(dev, PCI_COMMAND, ®);
|
|
reg |= (PCI_COMMAND_IO | PCI_COMMAND_MEMORY |
|
|
PCI_COMMAND_MASTER | PCI_COMMAND_SERR);
|
|
pci_write_config_word(dev, PCI_COMMAND, reg);
|
|
}
|
|
}
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_ANY_ID, PCI_ANY_ID, tegra_pcie_fixup_bridge);
|
|
|
|
/* Tegra PCIE root complex wrongly reports device class */
|
|
static void tegra_pcie_fixup_class(struct pci_dev *dev)
|
|
{
|
|
dev->class = PCI_CLASS_BRIDGE_PCI << 8;
|
|
}
|
|
|
|
DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_NVIDIA, 0x0e1c, tegra_pcie_fixup_class);
|
|
DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_NVIDIA, 0x0e1d, tegra_pcie_fixup_class);
|
|
|
|
/* Tegra PCIE requires relaxed ordering */
|
|
static void tegra_pcie_relax_enable(struct pci_dev *dev)
|
|
{
|
|
pcie_capability_set_word(dev, PCI_EXP_DEVCTL, PCI_EXP_DEVCTL_RELAX_EN);
|
|
}
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_ANY_ID, PCI_ANY_ID, tegra_pcie_relax_enable);
|
|
|
|
static int tegra_pcie_request_resources(struct tegra_pcie *pcie)
|
|
{
|
|
struct pci_host_bridge *host = pci_host_bridge_from_priv(pcie);
|
|
struct list_head *windows = &host->windows;
|
|
struct device *dev = pcie->dev;
|
|
int err;
|
|
|
|
pci_add_resource_offset(windows, &pcie->pio, pcie->offset.io);
|
|
pci_add_resource_offset(windows, &pcie->mem, pcie->offset.mem);
|
|
pci_add_resource_offset(windows, &pcie->prefetch, pcie->offset.mem);
|
|
pci_add_resource(windows, &pcie->busn);
|
|
|
|
err = devm_request_pci_bus_resources(dev, windows);
|
|
if (err < 0) {
|
|
pci_free_resource_list(windows);
|
|
return err;
|
|
}
|
|
|
|
pci_remap_iospace(&pcie->pio, pcie->io.start);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void tegra_pcie_free_resources(struct tegra_pcie *pcie)
|
|
{
|
|
struct pci_host_bridge *host = pci_host_bridge_from_priv(pcie);
|
|
struct list_head *windows = &host->windows;
|
|
struct resource_entry *win, *tmp;
|
|
|
|
pci_unmap_iospace(&pcie->pio);
|
|
|
|
resource_list_for_each_entry_safe(win, tmp, &host->windows) {
|
|
switch (resource_type(win->res)) {
|
|
case IORESOURCE_IO:
|
|
case IORESOURCE_MEM:
|
|
devm_release_resource(pcie->dev, win->res);
|
|
break;
|
|
default:
|
|
continue;
|
|
}
|
|
}
|
|
pci_free_resource_list(windows);
|
|
}
|
|
|
|
static int tegra_pcie_map_irq(const struct pci_dev *dev, u8 slot, u8 pin)
|
|
{
|
|
struct pci_host_bridge *host = pci_find_host_bridge(dev->bus);
|
|
struct tegra_pcie *pcie = pci_host_bridge_priv(host);
|
|
|
|
return pcie->irq;
|
|
}
|
|
|
|
static int tegra_pcie_add_bus(struct pci_bus *bus)
|
|
{
|
|
struct tegra_pcie_bus *tbus;
|
|
struct pci_host_bridge *host = pci_find_host_bridge(bus);
|
|
struct tegra_pcie *pcie = pci_host_bridge_priv(host);
|
|
|
|
PR_FUNC_LINE;
|
|
/* bus 0 is root complex whose config space is already mapped */
|
|
if (!bus->number)
|
|
return 0;
|
|
if (IS_ENABLED(CONFIG_PCI_MSI))
|
|
bus->msi = &pcie->msi.chip;
|
|
|
|
/* Allocate memory for new bus */
|
|
tbus = tegra_pcie_bus_alloc(pcie, bus->number);
|
|
if (IS_ERR(tbus))
|
|
return 0;
|
|
list_add_tail(&tbus->list, &pcie->buses);
|
|
return 0;
|
|
}
|
|
|
|
static struct pci_ops tegra_pcie_ops = {
|
|
.add_bus = tegra_pcie_add_bus,
|
|
.map_bus = tegra_pcie_conf_address,
|
|
.read = tegra_pcie_read_conf,
|
|
.write = tegra_pcie_write_conf,
|
|
};
|
|
|
|
static unsigned long tegra_pcie_port_get_pex_ctrl(struct tegra_pcie_port *port)
|
|
{
|
|
unsigned long ret = 0;
|
|
|
|
switch (port->index) {
|
|
case 0:
|
|
ret = AFI_PEX0_CTRL;
|
|
break;
|
|
case 1:
|
|
ret = AFI_PEX1_CTRL;
|
|
break;
|
|
case 2:
|
|
ret = AFI_PEX2_CTRL;
|
|
break;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static void tegra_pcie_config_clkreq(struct tegra_pcie *pcie, u32 index,
|
|
int bi_dir)
|
|
{
|
|
struct pinctrl *clkreq_pin = NULL;
|
|
struct pinctrl_state *clkreq_dir = NULL;
|
|
int ret = 0;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
clkreq_pin = devm_pinctrl_get(pcie->dev);
|
|
if (IS_ERR(clkreq_pin)) {
|
|
dev_err(pcie->dev, "config clkreq dir failed: %ld\n",
|
|
PTR_ERR(clkreq_pin));
|
|
return;
|
|
}
|
|
|
|
switch (index) {
|
|
case 0:
|
|
if (bi_dir)
|
|
clkreq_dir =
|
|
pinctrl_lookup_state(clkreq_pin,
|
|
"clkreq-0-bi-dir-enable");
|
|
else
|
|
clkreq_dir =
|
|
pinctrl_lookup_state(clkreq_pin,
|
|
"clkreq-0-in-dir-enable");
|
|
break;
|
|
case 1:
|
|
if (bi_dir)
|
|
clkreq_dir =
|
|
pinctrl_lookup_state(clkreq_pin,
|
|
"clkreq-1-bi-dir-enable");
|
|
else
|
|
clkreq_dir =
|
|
pinctrl_lookup_state(clkreq_pin,
|
|
"clkreq-1-in-dir-enable");
|
|
break;
|
|
default:
|
|
return;
|
|
}
|
|
|
|
if (IS_ERR(clkreq_dir)) {
|
|
dev_err(pcie->dev, "missing clkreq_dir_enable state: %ld\n",
|
|
PTR_ERR(clkreq_dir));
|
|
return;
|
|
}
|
|
ret = pinctrl_select_state(clkreq_pin, clkreq_dir);
|
|
if (ret < 0)
|
|
dev_err(pcie->dev,
|
|
"setting clkreq pin dir state failed: %d\n", ret);
|
|
}
|
|
|
|
#if defined(CONFIG_PCIEASPM)
|
|
static void tegra_pcie_enable_ltr_support(void)
|
|
{
|
|
struct pci_dev *pdev = NULL;
|
|
u16 val = 0;
|
|
u32 data = 0, pos = 0;
|
|
|
|
PR_FUNC_LINE;
|
|
/* enable LTR mechanism for L1.2 support in end points */
|
|
for_each_pci_dev(pdev) {
|
|
pos = pci_find_ext_capability(pdev, PCI_EXT_CAP_ID_L1SS);
|
|
if (!pos)
|
|
continue;
|
|
pci_read_config_dword(pdev, pos + PCI_L1SS_CAP, &data);
|
|
if (!((data & PCI_L1SS_CAP_ASPM_L12S) ||
|
|
(data & PCI_L1SS_CAP_PM_L12S)))
|
|
continue;
|
|
pcie_capability_read_dword(pdev, PCI_EXP_DEVCAP2, &data);
|
|
if (data & PCI_EXP_DEVCAP2_LTR) {
|
|
pcie_capability_read_word(pdev, PCI_EXP_DEVCTL2, &val);
|
|
val |= PCI_EXP_DEVCTL2_LTR_EN;
|
|
pcie_capability_write_word(pdev, PCI_EXP_DEVCTL2, val);
|
|
}
|
|
}
|
|
}
|
|
|
|
struct dev_ids {
|
|
unsigned short vid;
|
|
unsigned short did;
|
|
};
|
|
|
|
static const struct pci_device_id aspm_l0s_blacklist[] = {
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4355), 0, 0, 0 },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x43ef), 0, 0, 0 },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4354), 0, 0, 0 },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_NEC, 0x0194), 0, 0, 0 },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_TOSHIBA, 0x010f), 0, 0, 0 },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x0953), 0, 0, 0 },
|
|
{ 0 }
|
|
};
|
|
|
|
/* Enable ASPM support of all devices based on it's capability */
|
|
static void tegra_pcie_configure_aspm(void)
|
|
{
|
|
struct pci_dev *pdev = NULL;
|
|
struct tegra_pcie *pcie = NULL;
|
|
struct pci_host_bridge *host = NULL;
|
|
|
|
PR_FUNC_LINE;
|
|
if (!pcie_aspm_support_enabled()) {
|
|
pr_info("PCIE: ASPM not enabled\n");
|
|
return;
|
|
}
|
|
|
|
pdev = pci_get_device(PCI_ANY_ID, PCI_ANY_ID, NULL);
|
|
|
|
host = pci_find_host_bridge(pdev->bus);
|
|
pcie = pci_host_bridge_priv(host);
|
|
pdev = NULL;
|
|
|
|
/* disable ASPM-L0s for all links unless the endpoint
|
|
* is a known device with proper ASPM-L0s functionality
|
|
*/
|
|
for_each_pci_dev(pdev) {
|
|
struct pci_dev *parent = NULL;
|
|
struct tegra_pcie_port *port = NULL;
|
|
unsigned long ctrl = 0;
|
|
u32 rp = 0, val = 0, i = 0;
|
|
|
|
if ((pci_pcie_type(pdev) == PCI_EXP_TYPE_ROOT_PORT) ||
|
|
(pci_pcie_type(pdev) == PCI_EXP_TYPE_DOWNSTREAM))
|
|
continue;
|
|
parent = pdev->bus->self;
|
|
|
|
/* following needs to be done only for devices which are
|
|
* directly connected to Tegra root ports */
|
|
if (parent->bus->self)
|
|
continue;
|
|
rp = PCI_SLOT(parent->devfn);
|
|
list_for_each_entry(port, &pcie->ports, list)
|
|
if (rp == port->index + 1)
|
|
break;
|
|
ctrl = tegra_pcie_port_get_pex_ctrl(port);
|
|
/* AFI_PEX_STATUS is AFI_PEX_CTRL + 4 */
|
|
val = afi_readl(port->pcie, ctrl + 4);
|
|
if ((val & 0x1) || (port->disable_clock_request)) {
|
|
i |= PCIE_LINK_STATE_CLKPM;
|
|
/* disable PADS2PLLE control */
|
|
val = afi_readl(port->pcie, AFI_PLLE_CONTROL);
|
|
val &= ~AFI_PLLE_CONTROL_PADS2PLLE_CONTROL_EN;
|
|
afi_writel(port->pcie, val, AFI_PLLE_CONTROL);
|
|
}
|
|
|
|
/* Disable ASPM-l0s for blacklisted devices */
|
|
if (pci_match_id(aspm_l0s_blacklist, pdev))
|
|
i |= PCIE_LINK_STATE_L0S;
|
|
|
|
pci_disable_link_state_locked(pdev, i);
|
|
|
|
if (port->disable_clock_request)
|
|
continue;
|
|
|
|
if (pcie->soc_data->program_clkreq_as_bi_dir) {
|
|
/* check if L1SS capability is supported */
|
|
i = pci_find_ext_capability(pdev, PCI_EXT_CAP_ID_L1SS);
|
|
if (!i)
|
|
continue;
|
|
/* avoid L1SS config if no support of L1SS feature */
|
|
pci_read_config_dword(pdev, i + PCI_L1SS_CAP, &val);
|
|
if ((val & PCI_L1SS_CAP_L1PMS) ||
|
|
(val & PCI_L1SS_CAP_L1PM_MASK))
|
|
tegra_pcie_config_clkreq(pcie, port->index, 1);
|
|
}
|
|
}
|
|
/* L1.2 specific common configuration */
|
|
tegra_pcie_enable_ltr_support();
|
|
}
|
|
#endif
|
|
|
|
static void tegra_pcie_postinit(void)
|
|
{
|
|
#if defined(CONFIG_PCIEASPM)
|
|
tegra_pcie_configure_aspm();
|
|
#endif
|
|
}
|
|
|
|
#ifdef HOTPLUG_ON_SYSTEM_BOOT
|
|
/* It enumerates the devices when dock is connected after system boot */
|
|
/* this is similar to pcibios_init_hw in bios32.c */
|
|
static void __init tegra_pcie_hotplug_init(void)
|
|
{
|
|
struct pci_sys_data *sys = NULL;
|
|
int ret, nr;
|
|
|
|
if (is_dock_conn_at_boot)
|
|
return;
|
|
|
|
PR_FUNC_LINE;
|
|
tegra_pcie_preinit();
|
|
for (nr = 0; nr < tegra_pcie_hw.nr_controllers; nr++) {
|
|
sys = kzalloc(sizeof(struct pci_sys_data), GFP_KERNEL);
|
|
if (!sys)
|
|
panic("PCI: unable to allocate sys data!");
|
|
|
|
#ifdef CONFIG_PCI_DOMAINS
|
|
sys->domain = tegra_pcie_hw.domain;
|
|
#endif
|
|
sys->busnr = nr;
|
|
sys->swizzle = tegra_pcie_hw.swizzle;
|
|
sys->map_irq = tegra_pcie_hw.map_irq;
|
|
INIT_LIST_HEAD(&sys->resources);
|
|
|
|
ret = tegra_pcie_setup(nr, sys);
|
|
if (ret > 0) {
|
|
if (list_empty(&sys->resources)) {
|
|
pci_add_resource_offset(&sys->resources,
|
|
&ioport_resource, sys->io_offset);
|
|
pci_add_resource_offset(&sys->resources,
|
|
&iomem_resource, sys->mem_offset);
|
|
}
|
|
pci_create_root_bus(NULL, nr, &tegra_pcie_ops,
|
|
sys, &sys->resources);
|
|
}
|
|
}
|
|
is_dock_conn_at_boot = true;
|
|
}
|
|
#endif
|
|
|
|
static void tegra_pcie_enable_aer(struct tegra_pcie_port *port, bool enable)
|
|
{
|
|
unsigned int data;
|
|
|
|
PR_FUNC_LINE;
|
|
data = rp_readl(port, NV_PCIE2_RP_VEND_CTL1);
|
|
if (enable)
|
|
data |= PCIE2_RP_VEND_CTL1_ERPT;
|
|
else
|
|
data &= ~PCIE2_RP_VEND_CTL1_ERPT;
|
|
rp_writel(port, data, NV_PCIE2_RP_VEND_CTL1);
|
|
}
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_NVIDIA, 0x0bf0, tegra_pcie_relax_enable);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_NVIDIA, 0x0bf1, tegra_pcie_relax_enable);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_NVIDIA, 0x0e1c, tegra_pcie_relax_enable);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_NVIDIA, 0x0e1d, tegra_pcie_relax_enable);
|
|
|
|
static int tegra_pcie_attach(struct tegra_pcie *pcie)
|
|
{
|
|
struct pci_bus *bus = NULL;
|
|
struct tegra_pcie_port *port;
|
|
|
|
PR_FUNC_LINE;
|
|
if (!hotplug_event)
|
|
return 0;
|
|
|
|
/* rescan and recreate all pcie data structures */
|
|
while ((bus = pci_find_next_bus(bus)) != NULL)
|
|
pci_rescan_bus(bus);
|
|
/* unhide AER capability */
|
|
list_for_each_entry(port, &pcie->ports, list)
|
|
if (port->status)
|
|
tegra_pcie_enable_aer(port, true);
|
|
|
|
hotplug_event = false;
|
|
return 0;
|
|
}
|
|
|
|
static int tegra_pcie_detach(struct tegra_pcie *pcie)
|
|
{
|
|
struct pci_dev *pdev = NULL;
|
|
struct tegra_pcie_port *port;
|
|
|
|
PR_FUNC_LINE;
|
|
if (hotplug_event)
|
|
return 0;
|
|
hotplug_event = true;
|
|
|
|
/* hide AER capability to avoid log spew */
|
|
list_for_each_entry(port, &pcie->ports, list)
|
|
if (port->status)
|
|
tegra_pcie_enable_aer(port, false);
|
|
|
|
/* remove all pcie data structures */
|
|
for_each_pci_dev(pdev) {
|
|
pci_stop_and_remove_bus_device(pdev);
|
|
break;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void tegra_pcie_prsnt_map_override(struct tegra_pcie_port *port,
|
|
bool prsnt)
|
|
{
|
|
unsigned int data;
|
|
|
|
PR_FUNC_LINE;
|
|
/* currently only hotplug on root port 0 supported */
|
|
data = rp_readl(port, NV_PCIE2_RP_PRIV_MISC);
|
|
data &= ~PCIE2_RP_PRIV_MISC_PRSNT_MAP_EP_ABSNT;
|
|
if (prsnt)
|
|
data |= PCIE2_RP_PRIV_MISC_PRSNT_MAP_EP_PRSNT;
|
|
else
|
|
data |= PCIE2_RP_PRIV_MISC_PRSNT_MAP_EP_ABSNT;
|
|
rp_writel(port, data, NV_PCIE2_RP_PRIV_MISC);
|
|
}
|
|
|
|
static void work_hotplug_handler(struct work_struct *work)
|
|
{
|
|
struct tegra_pcie *pcie_driver =
|
|
container_of(work, struct tegra_pcie, hotplug_detect);
|
|
int val;
|
|
|
|
PR_FUNC_LINE;
|
|
if (pcie_driver->plat_data->gpio_hot_plug == -1)
|
|
return;
|
|
val = gpio_get_value(pcie_driver->plat_data->gpio_hot_plug);
|
|
if (val == 0) {
|
|
dev_info(pcie_driver->dev, "PCIE Hotplug: Connected\n");
|
|
tegra_pcie_attach(pcie_driver);
|
|
} else {
|
|
dev_info(pcie_driver->dev, "PCIE Hotplug: DisConnected\n");
|
|
tegra_pcie_detach(pcie_driver);
|
|
}
|
|
}
|
|
|
|
static irqreturn_t gpio_pcie_detect_isr(int irq, void *arg)
|
|
{
|
|
struct tegra_pcie *pcie = arg;
|
|
PR_FUNC_LINE;
|
|
schedule_work(&pcie->hotplug_detect);
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static void handle_sb_intr(struct tegra_pcie *pcie)
|
|
{
|
|
u32 mesg;
|
|
|
|
PR_FUNC_LINE;
|
|
/* Port 0 and 1 */
|
|
mesg = afi_readl(pcie, AFI_MSG_0);
|
|
if (mesg & AFI_MSG_INTX_MASK)
|
|
/* notify device isr for INTx messages from pcie devices */
|
|
dev_dbg(pcie->dev,
|
|
"Legacy INTx interrupt occurred %x on port 0/1\n", mesg);
|
|
else if (mesg & AFI_MSG_PM_PME_MASK) {
|
|
struct tegra_pcie_port *port, *tmp;
|
|
/* handle PME messages */
|
|
list_for_each_entry_safe(port, tmp, &pcie->ports, list) {
|
|
if ((port->index == 0) && (mesg & AFI_MSG_PM_PME0))
|
|
break;
|
|
if ((port->index == 1) && (mesg & AFI_MSG_PM_PME1))
|
|
break;
|
|
}
|
|
mesg = rp_readl(port, NV_PCIE2_RP_RSR);
|
|
mesg |= NV_PCIE2_RP_RSR_PMESTAT;
|
|
rp_writel(port, mesg, NV_PCIE2_RP_RSR);
|
|
} else
|
|
afi_writel(pcie, mesg, AFI_MSG_0);
|
|
|
|
if (pcie->soc_data->num_ports <= 2)
|
|
return;
|
|
|
|
/* Port 2 */
|
|
mesg = afi_readl(pcie, AFI_MSG_1_0);
|
|
if (mesg & AFI_MSG_1_INTX_MASK)
|
|
/* notify device isr for INTx messages from pcie devices */
|
|
dev_dbg(pcie->dev,
|
|
"Legacy INTx interrupt occurred %x on port 2\n", mesg);
|
|
else if (mesg & AFI_MSG_1_PM_PME_MASK) {
|
|
struct tegra_pcie_port *port, *tmp;
|
|
/* handle PME messages */
|
|
list_for_each_entry_safe(port, tmp, &pcie->ports, list) {
|
|
if ((port->index == 2) && (mesg & AFI_MSG_1_PM_PME))
|
|
break;
|
|
}
|
|
mesg = rp_readl(port, NV_PCIE2_RP_RSR);
|
|
mesg |= NV_PCIE2_RP_RSR_PMESTAT;
|
|
rp_writel(port, mesg, NV_PCIE2_RP_RSR);
|
|
} else
|
|
afi_writel(pcie, mesg, AFI_MSG_1_0);
|
|
}
|
|
|
|
static irqreturn_t tegra_pcie_isr(int irq, void *arg)
|
|
{
|
|
const char *err_msg[] = {
|
|
"Unknown",
|
|
"AXI slave error",
|
|
"AXI decode error",
|
|
"Target abort",
|
|
"Master abort",
|
|
"Invalid write",
|
|
"",
|
|
"Response decoding error",
|
|
"AXI response decoding error",
|
|
"Transcation timeout",
|
|
"",
|
|
"Slot Clock request change",
|
|
"TMS Clock clamp change",
|
|
"TMS power down",
|
|
"Peer to Peer error",
|
|
};
|
|
struct tegra_pcie *pcie = arg;
|
|
u32 code, signature;
|
|
|
|
PR_FUNC_LINE;
|
|
code = afi_readl(pcie, AFI_INTR_CODE) & AFI_INTR_CODE_MASK;
|
|
|
|
signature = afi_readl(pcie, AFI_INTR_SIGNATURE);
|
|
|
|
if (code == AFI_INTR_LEGACY)
|
|
handle_sb_intr(pcie);
|
|
afi_writel(pcie, 0, AFI_INTR_CODE);
|
|
afi_readl(pcie, AFI_INTR_CODE); /* read pushes write */
|
|
|
|
if (code >= ARRAY_SIZE(err_msg))
|
|
code = 0;
|
|
|
|
/*
|
|
* do not pollute kernel log with master abort reports since they
|
|
* happen a lot during enumeration
|
|
*/
|
|
if (code == AFI_INTR_MASTER_ABORT)
|
|
dev_dbg(pcie->dev, "PCIE: %s, signature: %08x\n",
|
|
err_msg[code], signature);
|
|
else if ((code != AFI_INTR_LEGACY) && (code != AFI_INTR_PRSNT_SENSE))
|
|
dev_err(pcie->dev, "PCIE: %s, signature: %08x\n",
|
|
err_msg[code], signature);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
/*
|
|
* FPCI map is as follows:
|
|
* - 0xfdfc000000: I/O space
|
|
* - 0xfdfe000000: type 0 configuration space
|
|
* - 0xfdff000000: type 1 configuration space
|
|
* - 0xfe00000000: type 0 extended configuration space
|
|
* - 0xfe10000000: type 1 extended configuration space
|
|
*/
|
|
static void tegra_pcie_setup_translations(struct tegra_pcie *pcie)
|
|
{
|
|
u32 fpci_bar, size, axi_address;
|
|
|
|
/* Bar 0: type 1 extended configuration space */
|
|
fpci_bar = 0xfe100000;
|
|
size = resource_size(pcie->cs);
|
|
axi_address = pcie->cs->start;
|
|
afi_writel(pcie, axi_address, AFI_AXI_BAR0_START);
|
|
afi_writel(pcie, size >> 12, AFI_AXI_BAR0_SZ);
|
|
afi_writel(pcie, fpci_bar, AFI_FPCI_BAR0);
|
|
|
|
/* Bar 1: downstream IO bar */
|
|
fpci_bar = 0xfdfc0000;
|
|
size = resource_size(&pcie->io);
|
|
axi_address = pcie->io.start;
|
|
afi_writel(pcie, axi_address, AFI_AXI_BAR1_START);
|
|
afi_writel(pcie, size >> 12, AFI_AXI_BAR1_SZ);
|
|
afi_writel(pcie, fpci_bar, AFI_FPCI_BAR1);
|
|
|
|
/* Bar 2: prefetchable memory BAR */
|
|
fpci_bar = (((pcie->prefetch.start >> 12) & 0x0fffffff) << 4) | 0x1;
|
|
size = resource_size(&pcie->prefetch);
|
|
axi_address = pcie->prefetch.start;
|
|
afi_writel(pcie, axi_address, AFI_AXI_BAR2_START);
|
|
afi_writel(pcie, size >> 12, AFI_AXI_BAR2_SZ);
|
|
afi_writel(pcie, fpci_bar, AFI_FPCI_BAR2);
|
|
|
|
/* Bar 3: non prefetchable memory BAR */
|
|
fpci_bar = (((pcie->mem.start >> 12) & 0x0fffffff) << 4) | 0x1;
|
|
size = resource_size(&pcie->mem);
|
|
axi_address = pcie->mem.start;
|
|
afi_writel(pcie, axi_address, AFI_AXI_BAR3_START);
|
|
afi_writel(pcie, size >> 12, AFI_AXI_BAR3_SZ);
|
|
afi_writel(pcie, fpci_bar, AFI_FPCI_BAR3);
|
|
|
|
/* NULL out the remaining BARs as they are not used */
|
|
afi_writel(pcie, 0, AFI_AXI_BAR4_START);
|
|
afi_writel(pcie, 0, AFI_AXI_BAR4_SZ);
|
|
afi_writel(pcie, 0, AFI_FPCI_BAR4);
|
|
|
|
afi_writel(pcie, 0, AFI_AXI_BAR5_START);
|
|
afi_writel(pcie, 0, AFI_AXI_BAR5_SZ);
|
|
afi_writel(pcie, 0, AFI_FPCI_BAR5);
|
|
|
|
/* MSI translations are setup only when needed */
|
|
afi_writel(pcie, 0, AFI_MSI_FPCI_BAR_ST);
|
|
afi_writel(pcie, 0, AFI_MSI_BAR_SZ);
|
|
afi_writel(pcie, 0, AFI_MSI_AXI_BAR_ST);
|
|
afi_writel(pcie, 0, AFI_MSI_BAR_SZ);
|
|
}
|
|
|
|
static int tegra_pcie_get_clocks(struct tegra_pcie *pcie)
|
|
{
|
|
int error;
|
|
struct device *dev = pcie->dev;
|
|
|
|
PR_FUNC_LINE;
|
|
error = pm_clk_create(dev);
|
|
if (error) {
|
|
dev_err(dev, "pm_clk_create failed %d\n", error);
|
|
return error;
|
|
}
|
|
|
|
error = pm_clk_add(dev, "afi");
|
|
if (error) {
|
|
dev_err(dev, "missing afi clock");
|
|
return error;
|
|
}
|
|
error = pm_clk_add(dev, "pex");
|
|
if (error) {
|
|
dev_err(dev, "missing pcie clock");
|
|
return error;
|
|
}
|
|
pcie->emc_bwmgr = tegra_bwmgr_register(TEGRA_BWMGR_CLIENT_PCIE);
|
|
if (!pcie->emc_bwmgr) {
|
|
dev_err(pcie->dev, "couldn't register with EMC BwMgr\n");
|
|
return -EINVAL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void tegra_pcie_clocks_put(struct tegra_pcie *pcie)
|
|
{
|
|
pm_clk_destroy(pcie->dev);
|
|
tegra_bwmgr_set_emc(pcie->emc_bwmgr, 0, TEGRA_BWMGR_SET_EMC_FLOOR);
|
|
tegra_bwmgr_unregister(pcie->emc_bwmgr);
|
|
}
|
|
|
|
static int tegra_pcie_port_get_phy(struct tegra_pcie_port *port)
|
|
{
|
|
struct device *dev = port->pcie->dev;
|
|
struct phy *phy;
|
|
int err, i;
|
|
char *name;
|
|
|
|
port->phy = devm_kcalloc(dev, sizeof(phy), port->num_lanes, GFP_KERNEL);
|
|
if (!port->phy)
|
|
return -ENOMEM;
|
|
|
|
for (i = 0; i < port->num_lanes; i++) {
|
|
name = kasprintf(GFP_KERNEL, "pcie-%u", i);
|
|
if (!name)
|
|
return -ENOMEM;
|
|
|
|
phy = devm_of_phy_get(dev, port->np, name);
|
|
kfree(name);
|
|
if (IS_ERR(phy)) {
|
|
if (PTR_ERR(phy) == -EPROBE_DEFER)
|
|
dev_info(dev, "PHY get deferred: %ld\n",
|
|
PTR_ERR(phy));
|
|
else
|
|
dev_err(dev, "failed to get PHY: %ld\n",
|
|
PTR_ERR(phy));
|
|
return PTR_ERR(phy);
|
|
}
|
|
|
|
err = phy_init(phy);
|
|
if (err < 0) {
|
|
dev_err(dev, "failed to initialize PHY: %d\n", err);
|
|
return err;
|
|
}
|
|
|
|
port->phy[i] = phy;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int tegra_pcie_phys_get(struct tegra_pcie *pcie)
|
|
{
|
|
struct tegra_pcie_port *port;
|
|
int err;
|
|
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
err = tegra_pcie_port_get_phy(port);
|
|
if (err < 0)
|
|
return err;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int tegra_pcie_phy_power_on(struct tegra_pcie *pcie)
|
|
{
|
|
struct tegra_pcie_port *port;
|
|
int err, i;
|
|
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
for (i = 0; i < port->num_lanes; i++) {
|
|
err = phy_power_on(port->phy[i]);
|
|
if (err < 0)
|
|
return err;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int tegra_pcie_phy_power_off(struct tegra_pcie *pcie)
|
|
{
|
|
struct tegra_pcie_port *port;
|
|
int err, i;
|
|
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
for (i = 0; i < port->num_lanes; i++) {
|
|
err = phy_power_off(port->phy[i]);
|
|
if (err < 0)
|
|
return err;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int tegra_pcie_phy_exit(struct tegra_pcie *pcie)
|
|
{
|
|
struct tegra_pcie_port *port;
|
|
int err = 0, i;
|
|
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
for (i = 0; i < port->num_lanes; i++) {
|
|
if (port->phy && !IS_ERR_OR_NULL(port->phy[i]))
|
|
err = phy_exit(port->phy[i]);
|
|
if (err < 0)
|
|
return err;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int tegra_pcie_enable_pads(struct tegra_pcie *pcie, bool enable)
|
|
{
|
|
int err = 0;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
if (!tegra_platform_is_silicon())
|
|
return err;
|
|
|
|
if (pcie->soc_data->program_uphy) {
|
|
if (enable)
|
|
err = tegra_pcie_phy_power_on(pcie);
|
|
else
|
|
err = tegra_pcie_phy_power_off(pcie);
|
|
if (err)
|
|
dev_err(pcie->dev, "UPHY operation failed\n");
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static void tegra_pcie_enable_wrap(void)
|
|
{
|
|
u32 val;
|
|
void __iomem *msel_base;
|
|
|
|
PR_FUNC_LINE;
|
|
#define MSELECT_CONFIG_BASE 0x50060000
|
|
#define MSELECT_CONFIG_WRAP_TO_INCR_SLAVE1 BIT(28)
|
|
#define MSELECT_CONFIG_ERR_RESP_EN_SLAVE1 BIT(24)
|
|
|
|
/* Config MSELECT to support wrap trans for normal NC & GRE mapping */
|
|
msel_base = ioremap(MSELECT_CONFIG_BASE, 4);
|
|
val = readl(msel_base);
|
|
/* Enable WRAP_TO_INCR_SLAVE1 */
|
|
val |= MSELECT_CONFIG_WRAP_TO_INCR_SLAVE1;
|
|
/* Disable ERR_RESP_EN_SLAVE1 */
|
|
val &= ~MSELECT_CONFIG_ERR_RESP_EN_SLAVE1;
|
|
writel(val, msel_base);
|
|
iounmap(msel_base);
|
|
}
|
|
|
|
static int tegra_pcie_enable_controller(struct tegra_pcie *pcie)
|
|
{
|
|
u32 val;
|
|
struct tegra_pcie_port *port, *tmp;
|
|
|
|
PR_FUNC_LINE;
|
|
if (pcie->soc_data->enable_wrap)
|
|
tegra_pcie_enable_wrap();
|
|
/* Enable PLL power down */
|
|
val = afi_readl(pcie, AFI_PLLE_CONTROL);
|
|
val &= ~AFI_PLLE_CONTROL_BYPASS_PCIE2PLLE_CONTROL;
|
|
val &= ~AFI_PLLE_CONTROL_BYPASS_PADS2PLLE_CONTROL;
|
|
val |= AFI_PLLE_CONTROL_PADS2PLLE_CONTROL_EN;
|
|
val |= AFI_PLLE_CONTROL_PCIE2PLLE_CONTROL_EN;
|
|
list_for_each_entry_safe(port, tmp, &pcie->ports, list) {
|
|
if (port->disable_clock_request) {
|
|
val &= ~AFI_PLLE_CONTROL_PADS2PLLE_CONTROL_EN;
|
|
break;
|
|
}
|
|
}
|
|
afi_writel(pcie, val, AFI_PLLE_CONTROL);
|
|
|
|
afi_writel(pcie, 0, AFI_PEXBIAS_CTRL_0);
|
|
|
|
/* Enable all PCIE controller and */
|
|
/* system management configuration of PCIE crossbar */
|
|
val = afi_readl(pcie, AFI_PCIE_CONFIG);
|
|
val |= (AFI_PCIE_CONFIG_PCIEC0_DISABLE_DEVICE |
|
|
AFI_PCIE_CONFIG_PCIEC1_DISABLE_DEVICE |
|
|
AFI_PCIE_CONFIG_PCIEC2_DISABLE_DEVICE |
|
|
AFI_PCIE_CONFIG_PCIEC0_CLKREQ_AS_GPIO |
|
|
AFI_PCIE_CONFIG_PCIEC1_CLKREQ_AS_GPIO |
|
|
AFI_PCIE_CONFIG_PCIEC2_CLKREQ_AS_GPIO);
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
val &= ~(1 << (port->index + 1));
|
|
val &= ~(1 << (port->index + 29));
|
|
}
|
|
|
|
/* Configure pcie mode */
|
|
val &= ~AFI_PCIE_CONFIG_XBAR_CONFIG_MASK;
|
|
val |= pcie->xbar_config;
|
|
|
|
afi_writel(pcie, val, AFI_PCIE_CONFIG);
|
|
|
|
/* Enable Gen 2 capability of PCIE */
|
|
val = afi_readl(pcie, AFI_FUSE) & ~AFI_FUSE_PCIE_T0_GEN2_DIS;
|
|
afi_writel(pcie, val, AFI_FUSE);
|
|
|
|
/* Finally enable PCIe */
|
|
val = afi_readl(pcie, AFI_CONFIGURATION);
|
|
val |= (AFI_CONFIGURATION_EN_FPCI |
|
|
AFI_CONFIGURATION_CLKEN_OVERRIDE);
|
|
afi_writel(pcie, val, AFI_CONFIGURATION);
|
|
|
|
val = (AFI_INTR_EN_INI_SLVERR | AFI_INTR_EN_INI_DECERR |
|
|
AFI_INTR_EN_TGT_SLVERR | AFI_INTR_EN_TGT_DECERR |
|
|
AFI_INTR_EN_TGT_WRERR | AFI_INTR_EN_DFPCI_DECERR |
|
|
AFI_INTR_EN_AXI_DECERR );
|
|
afi_writel(pcie, val, AFI_AFI_INTR_ENABLE);
|
|
afi_writel(pcie, 0xffffffff, AFI_SM_INTR_ENABLE);
|
|
|
|
/* FIXME: No MSI for now, only INT */
|
|
afi_writel(pcie, AFI_INTR_MASK_INT_MASK, AFI_INTR_MASK);
|
|
|
|
/* Disable all execptions */
|
|
afi_writel(pcie, 0, AFI_FPCI_ERROR_MASKS);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int tegra_pcie_enable_regulators(struct tegra_pcie *pcie)
|
|
{
|
|
int i;
|
|
PR_FUNC_LINE;
|
|
if (pcie->power_rails_enabled)
|
|
return 0;
|
|
|
|
pcie->power_rails_enabled = 1;
|
|
dev_info(pcie->dev, "PCIE: Enable power rails\n");
|
|
|
|
for (i = 0; i < pcie->soc_data->num_pcie_regulators; i++) {
|
|
if (pcie->pcie_regulators[i])
|
|
if (regulator_enable(pcie->pcie_regulators[i]))
|
|
dev_err(pcie->dev, "%s: can't enable regulator %s\n",
|
|
__func__,
|
|
pcie->soc_data->pcie_regulator_names[i]);
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
static int tegra_pcie_disable_regulators(struct tegra_pcie *pcie)
|
|
{
|
|
int i;
|
|
PR_FUNC_LINE;
|
|
if (pcie->power_rails_enabled == 0)
|
|
return 0;
|
|
|
|
for (i = 0; i < pcie->soc_data->num_pcie_regulators; i++) {
|
|
if (pcie->pcie_regulators[i] != NULL)
|
|
if (regulator_disable(pcie->pcie_regulators[i]))
|
|
dev_err(pcie->dev, "%s: can't disable regulator %s\n",
|
|
__func__,
|
|
pcie->soc_data->pcie_regulator_names[i]);
|
|
}
|
|
|
|
pcie->power_rails_enabled = 0;
|
|
dev_info(pcie->dev, "PCIE: Disable power rails\n");
|
|
return 0;
|
|
|
|
}
|
|
|
|
static int tegra_pcie_map_resources(struct tegra_pcie *pcie)
|
|
{
|
|
struct platform_device *pdev = to_platform_device(pcie->dev);
|
|
struct resource *pads, *afi, *res;
|
|
|
|
PR_FUNC_LINE;
|
|
pads = platform_get_resource_byname(pdev, IORESOURCE_MEM, "pads");
|
|
|
|
pcie->pads_res = __devm_request_region(&pdev->dev, &iomem_resource,
|
|
pads->start, resource_size(pads),
|
|
"pcie-pads");
|
|
|
|
if (!pcie->pads_res) {
|
|
dev_err(&pdev->dev,
|
|
"PCIE: Failed to request region for pad registers\n");
|
|
return -EBUSY;
|
|
}
|
|
|
|
pcie->pads = devm_ioremap_nocache(&pdev->dev, pads->start,
|
|
resource_size(pads));
|
|
if (!(pcie->pads)) {
|
|
dev_err(pcie->dev, "PCIE: Failed to map PAD registers\n");
|
|
return -EADDRNOTAVAIL;
|
|
}
|
|
|
|
afi = platform_get_resource_byname(pdev, IORESOURCE_MEM, "afi");
|
|
|
|
pcie->afi_res = __devm_request_region(&pdev->dev, &iomem_resource,
|
|
afi->start, resource_size(afi),
|
|
"pcie-afi");
|
|
|
|
if (!pcie->afi_res) {
|
|
dev_err(&pdev->dev,
|
|
"PCIE: Failed to request region for afi registers\n");
|
|
return -EBUSY;
|
|
}
|
|
|
|
pcie->afi = devm_ioremap_nocache(&pdev->dev, afi->start,
|
|
resource_size(afi));
|
|
if (!(pcie->afi)) {
|
|
dev_err(pcie->dev, "PCIE: Failed to map AFI registers\n");
|
|
return -EADDRNOTAVAIL;
|
|
}
|
|
|
|
/* request configuration space, but remap later, on demand */
|
|
res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "cs");
|
|
|
|
pcie->cs = __devm_request_region(&pdev->dev, &iomem_resource,
|
|
res->start, resource_size(res), "pcie-config-space");
|
|
if (!pcie->cs) {
|
|
dev_err(&pdev->dev, "PCIE: Failed to request region for CS registers\n");
|
|
return -EBUSY;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void tegra_pcie_unmap_resources(struct tegra_pcie *pcie)
|
|
{
|
|
struct platform_device *pdev = to_platform_device(pcie->dev);
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
if (pcie->cs)
|
|
__devm_release_region(&pdev->dev, &iomem_resource,
|
|
pcie->cs->start,
|
|
resource_size(pcie->cs));
|
|
if (pcie->afi_res)
|
|
__devm_release_region(&pdev->dev, &iomem_resource,
|
|
pcie->afi_res->start,
|
|
resource_size(pcie->afi_res));
|
|
if (pcie->pads_res)
|
|
__devm_release_region(&pdev->dev, &iomem_resource,
|
|
pcie->pads_res->start,
|
|
resource_size(pcie->pads_res));
|
|
|
|
if (pcie->pads) {
|
|
devm_iounmap(&pdev->dev, pcie->pads);
|
|
pcie->pads = NULL;
|
|
}
|
|
if (pcie->afi) {
|
|
devm_iounmap(&pdev->dev, pcie->afi);
|
|
pcie->afi = NULL;
|
|
}
|
|
}
|
|
|
|
static int tegra_pcie_fpga_phy_init(struct tegra_pcie *pcie)
|
|
{
|
|
PR_FUNC_LINE;
|
|
/* Do reset for FPGA pcie phy */
|
|
reset_control_deassert(pcie->afi_rst);
|
|
afi_writel(pcie, AFI_WR_SCRATCH_0_RESET_VAL, AFI_WR_SCRATCH_0);
|
|
udelay(10);
|
|
afi_writel(pcie, AFI_WR_SCRATCH_0_DEFAULT_VAL, AFI_WR_SCRATCH_0);
|
|
udelay(10);
|
|
afi_writel(pcie, AFI_WR_SCRATCH_0_RESET_VAL, AFI_WR_SCRATCH_0);
|
|
return 0;
|
|
}
|
|
|
|
static void tegra_pcie_pme_turnoff(struct tegra_pcie *pcie)
|
|
{
|
|
unsigned int data;
|
|
|
|
PR_FUNC_LINE;
|
|
if (tegra_platform_is_fpga())
|
|
return;
|
|
data = afi_readl(pcie, AFI_PCIE_PME);
|
|
data |= AFI_PCIE_PME_TURN_OFF;
|
|
afi_writel(pcie, data, AFI_PCIE_PME);
|
|
do {
|
|
data = afi_readl(pcie, AFI_PCIE_PME);
|
|
} while (!(data & AFI_PCIE_PME_ACK));
|
|
|
|
/* Required for PLL power down */
|
|
data = afi_readl(pcie, AFI_PLLE_CONTROL);
|
|
data |= AFI_PLLE_CONTROL_BYPASS_PADS2PLLE_CONTROL;
|
|
afi_writel(pcie, data, AFI_PLLE_CONTROL);
|
|
}
|
|
|
|
static int tegra_pcie_module_power_on(struct tegra_pcie *pcie)
|
|
{
|
|
int ret;
|
|
struct device *dev;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
dev = pcie->dev;
|
|
|
|
if (pcie->pcie_power_enabled) {
|
|
dev_info(dev, "PCIE: Already powered on");
|
|
return 0;
|
|
}
|
|
pcie->pcie_power_enabled = 1;
|
|
|
|
if (gpio_is_valid(pcie->plat_data->gpio_wake) &&
|
|
device_may_wakeup(dev)) {
|
|
ret = disable_irq_wake(gpio_to_irq(
|
|
pcie->plat_data->gpio_wake));
|
|
if (ret < 0) {
|
|
dev_err(dev,
|
|
"ID wake-up event failed with error %d\n", ret);
|
|
return ret;
|
|
}
|
|
} else {
|
|
ret = tegra_pcie_enable_regulators(pcie);
|
|
if (ret) {
|
|
dev_err(dev, "PCIE: Failed to enable regulators\n");
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int tegra_pcie_module_power_off(struct tegra_pcie *pcie);
|
|
|
|
static void tegra_pcie_config_plat(struct tegra_pcie *pcie, bool set)
|
|
{
|
|
struct tegra_pcie_port *port;
|
|
int count;
|
|
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
for (count = 0; count < port->n_gpios; ++count)
|
|
gpiod_set_value(gpio_to_desc(port->gpios[count]), set);
|
|
}
|
|
}
|
|
|
|
static int tegra_pcie_restore_device(struct device *dev)
|
|
{
|
|
int err = 0;
|
|
struct tegra_pcie *pcie = dev_get_drvdata(dev);
|
|
|
|
PR_FUNC_LINE;
|
|
if (!pcie)
|
|
return 0;
|
|
|
|
tegra_pcie_config_plat(pcie, 1);
|
|
pm_clk_resume(dev);
|
|
err = tegra_pcie_module_power_on(pcie);
|
|
if (err < 0)
|
|
dev_err(dev, "power on failed: %d\n", err);
|
|
|
|
if (pcie->soc_data->config_pex_io_dpd) {
|
|
err = pinctrl_select_state(pcie->pex_pin,
|
|
pcie->pex_io_dpd_dis_state);
|
|
if (err < 0) {
|
|
dev_err(dev, "disabling pex-io-dpd failed: %d\n", err);
|
|
goto err_exit;
|
|
}
|
|
}
|
|
|
|
err = tegra_pcie_map_resources(pcie);
|
|
if (err) {
|
|
dev_err(dev, "PCIE: Failed to map resources\n");
|
|
goto err_map_resource;
|
|
}
|
|
if (tegra_platform_is_fpga()) {
|
|
err = tegra_pcie_fpga_phy_init(pcie);
|
|
if (err)
|
|
dev_err(dev, "PCIE: Failed to initialize FPGA Phy\n");
|
|
}
|
|
|
|
tegra_pcie_enable_pads(pcie, true);
|
|
reset_control_deassert(pcie->afi_rst);
|
|
tegra_pcie_enable_controller(pcie);
|
|
tegra_pcie_request_resources(pcie);
|
|
tegra_pcie_setup_translations(pcie);
|
|
/* Set up MSI registers, if MSI have been enabled */
|
|
tegra_pcie_enable_msi(pcie, true);
|
|
reset_control_deassert(pcie->pcie_rst);
|
|
tegra_pcie_check_ports(pcie);
|
|
return 0;
|
|
|
|
err_map_resource:
|
|
if (pcie->soc_data->config_pex_io_dpd) {
|
|
err = pinctrl_select_state(pcie->pex_pin,
|
|
pcie->pex_io_dpd_en_state);
|
|
if (err < 0)
|
|
dev_err(dev, "enabling pex-io-dpd failed: %d\n", err);
|
|
}
|
|
err_exit:
|
|
tegra_pcie_module_power_off(pcie);
|
|
pm_clk_suspend(dev);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int tegra_pcie_power_on(struct tegra_pcie *pcie)
|
|
{
|
|
int err = 0;
|
|
|
|
PR_FUNC_LINE;
|
|
err = pm_runtime_get_sync(pcie->dev);
|
|
if (err) {
|
|
pm_runtime_put(pcie->dev);
|
|
pcie->pcie_power_enabled = 0;
|
|
}
|
|
return err;
|
|
}
|
|
|
|
static int tegra_pcie_save_device(struct device *dev)
|
|
{
|
|
struct tegra_pcie *pcie = dev_get_drvdata(dev);
|
|
struct tegra_pcie_port *port;
|
|
int err = 0;
|
|
|
|
PR_FUNC_LINE;
|
|
if (!pcie)
|
|
return 0;
|
|
|
|
if (pcie->pcie_power_enabled == 0) {
|
|
dev_info(dev, "PCIE: Already powered off");
|
|
return 0;
|
|
}
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
tegra_pcie_prsnt_map_override(port, false);
|
|
}
|
|
tegra_pcie_pme_turnoff(pcie);
|
|
tegra_pcie_free_resources(pcie);
|
|
tegra_pcie_enable_pads(pcie, false);
|
|
tegra_pcie_unmap_resources(pcie);
|
|
if (pcie->soc_data->config_pex_io_dpd) {
|
|
err = pinctrl_select_state(pcie->pex_pin,
|
|
pcie->pex_io_dpd_en_state);
|
|
if (err < 0)
|
|
dev_err(dev, "enabling pex-io-dpd failed: %d\n", err);
|
|
}
|
|
reset_control_assert(pcie->pciex_rst);
|
|
reset_control_assert(pcie->pcie_rst);
|
|
reset_control_assert(pcie->afi_rst);
|
|
|
|
err = tegra_pcie_module_power_off(pcie);
|
|
if (err < 0)
|
|
dev_err(dev, "power off failed: %d\n", err);
|
|
|
|
pm_clk_suspend(dev);
|
|
tegra_pcie_config_plat(pcie, 0);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int tegra_pcie_module_power_off(struct tegra_pcie *pcie)
|
|
{
|
|
int ret;
|
|
struct device *dev;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
dev = pcie->dev;
|
|
/* configure PE_WAKE signal as wake sources */
|
|
if (gpio_is_valid(pcie->plat_data->gpio_wake) &&
|
|
device_may_wakeup(dev)) {
|
|
ret = enable_irq_wake(gpio_to_irq(
|
|
pcie->plat_data->gpio_wake));
|
|
if (ret < 0) {
|
|
dev_err(dev,
|
|
"ID wake-up event failed with error %d\n", ret);
|
|
}
|
|
} else {
|
|
ret = tegra_pcie_disable_regulators(pcie);
|
|
}
|
|
|
|
pcie->pcie_power_enabled = 0;
|
|
return ret;
|
|
}
|
|
|
|
static int tegra_pcie_power_off(struct tegra_pcie *pcie)
|
|
{
|
|
int err = 0;
|
|
|
|
PR_FUNC_LINE;
|
|
err = pm_runtime_put_sync(pcie->dev);
|
|
if (err)
|
|
goto err_exit;
|
|
pcie->pcie_power_enabled = 0;
|
|
err_exit:
|
|
return err;
|
|
}
|
|
|
|
static int tegra_pcie_get_resets(struct tegra_pcie *pcie)
|
|
{
|
|
pcie->afi_rst = devm_reset_control_get(pcie->dev, "afi");
|
|
if (IS_ERR(pcie->afi_rst)) {
|
|
dev_err(pcie->dev, "PCIE : afi reset is missing\n");
|
|
return PTR_ERR(pcie->afi_rst);
|
|
}
|
|
|
|
pcie->pcie_rst = devm_reset_control_get(pcie->dev, "pex");
|
|
if (IS_ERR(pcie->pcie_rst)) {
|
|
dev_err(pcie->dev, "PCIE : pcie reset is missing\n");
|
|
return PTR_ERR(pcie->pcie_rst);
|
|
}
|
|
|
|
pcie->pciex_rst = devm_reset_control_get(pcie->dev, "pcie_x");
|
|
if (IS_ERR(pcie->pciex_rst)) {
|
|
dev_err(pcie->dev, "PCIE : pcie-xclk reset is missing\n");
|
|
return PTR_ERR(pcie->pciex_rst);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int tegra_pcie_get_resources(struct tegra_pcie *pcie)
|
|
{
|
|
struct platform_device *pdev = to_platform_device(pcie->dev);
|
|
int err;
|
|
|
|
PR_FUNC_LINE;
|
|
pcie->power_rails_enabled = 0;
|
|
pcie->pcie_power_enabled = 0;
|
|
|
|
err = tegra_pcie_get_clocks(pcie);
|
|
if (err) {
|
|
dev_err(pcie->dev, "PCIE: failed to get clocks: %d\n", err);
|
|
goto err_clk_get;
|
|
}
|
|
err = tegra_pcie_get_resets(pcie);
|
|
if (err) {
|
|
dev_err(pcie->dev, "PCIE: failed to get resets: %d\n", err);
|
|
goto err_reset_get;
|
|
}
|
|
err = tegra_pcie_enable_regulators(pcie);
|
|
if (err) {
|
|
dev_err(pcie->dev, "PCIE: Failed to enable regulators\n");
|
|
goto err_reset_get;
|
|
}
|
|
err = platform_get_irq_byname(pdev, "intr");
|
|
if (err < 0) {
|
|
dev_err(pcie->dev, "failed to get IRQ: %d\n", err);
|
|
goto err_irq;
|
|
}
|
|
|
|
pcie->irq = err;
|
|
|
|
err = devm_request_irq(&pdev->dev, pcie->irq, tegra_pcie_isr,
|
|
IRQF_SHARED, "PCIE", pcie);
|
|
if (err) {
|
|
dev_err(pcie->dev, "PCIE: Failed to register IRQ: %d\n", err);
|
|
goto err_irq;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_irq:
|
|
tegra_pcie_disable_regulators(pcie);
|
|
err_reset_get:
|
|
tegra_pcie_clocks_put(pcie);
|
|
err_clk_get:
|
|
return err;
|
|
}
|
|
|
|
static void tegra_pcie_release_resources(struct tegra_pcie *pcie)
|
|
{
|
|
devm_free_irq(pcie->dev, pcie->irq, pcie);
|
|
tegra_pcie_clocks_put(pcie);
|
|
}
|
|
|
|
static void tegra_pcie_port_reset(struct tegra_pcie_port *port)
|
|
{
|
|
unsigned long ctrl = tegra_pcie_port_get_pex_ctrl(port);
|
|
unsigned long value;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
/* pulse reset signal */
|
|
/* assert PEX_RST_A */
|
|
if (gpio_is_valid(port->rst_gpio)) {
|
|
gpio_set_value(port->rst_gpio, 0);
|
|
} else {
|
|
value = afi_readl(port->pcie, ctrl);
|
|
value &= ~AFI_PEX_CTRL_RST;
|
|
afi_writel(port->pcie, value, ctrl);
|
|
}
|
|
|
|
usleep_range(1000, 2000);
|
|
|
|
/* deAssert PEX_RST_A */
|
|
if (gpio_is_valid(port->rst_gpio)) {
|
|
gpio_set_value(port->rst_gpio, 1);
|
|
} else {
|
|
value = afi_readl(port->pcie, ctrl);
|
|
value |= AFI_PEX_CTRL_RST;
|
|
afi_writel(port->pcie, value, ctrl);
|
|
}
|
|
}
|
|
|
|
static void tegra_pcie_port_enable(struct tegra_pcie_port *port)
|
|
{
|
|
unsigned long ctrl = tegra_pcie_port_get_pex_ctrl(port);
|
|
unsigned long value;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
/* enable reference clock. Enable SW override so as to allow device
|
|
to get enumerated. SW override will be removed after enumeration
|
|
*/
|
|
value = afi_readl(port->pcie, ctrl);
|
|
value |= (AFI_PEX_CTRL_REFCLK_EN | AFI_PEX_CTRL_OVERRIDE_EN);
|
|
/* t124 doesn't support pll power down due to RTL bug and some */
|
|
/* platforms don't support clkreq, both needs to disable clkreq and */
|
|
/* enable refclk override to have refclk always ON independent of EP */
|
|
if (port->disable_clock_request)
|
|
value |= AFI_PEX_CTRL_CLKREQ_EN;
|
|
else
|
|
value &= ~AFI_PEX_CTRL_CLKREQ_EN;
|
|
|
|
afi_writel(port->pcie, value, ctrl);
|
|
|
|
tegra_pcie_port_reset(port);
|
|
|
|
/* On platforms where MXM is not directly connected to Tegra root port,
|
|
* 200 ms delay (worst case) is required after reset, to ensure linkup
|
|
* between PCIe switch and MXM
|
|
*/
|
|
if (port->has_mxm_port)
|
|
mdelay(200);
|
|
}
|
|
|
|
static void tegra_pcie_port_disable(struct tegra_pcie_port *port)
|
|
{
|
|
u32 data;
|
|
|
|
PR_FUNC_LINE;
|
|
data = afi_readl(port->pcie, AFI_PCIE_CONFIG);
|
|
switch (port->index) {
|
|
case 0:
|
|
data |= (AFI_PCIE_CONFIG_PCIEC0_DISABLE_DEVICE |
|
|
AFI_PCIE_CONFIG_PCIEC0_CLKREQ_AS_GPIO);
|
|
break;
|
|
case 1:
|
|
data |= (AFI_PCIE_CONFIG_PCIEC1_DISABLE_DEVICE |
|
|
AFI_PCIE_CONFIG_PCIEC1_CLKREQ_AS_GPIO);
|
|
break;
|
|
case 2:
|
|
data |= (AFI_PCIE_CONFIG_PCIEC2_DISABLE_DEVICE |
|
|
AFI_PCIE_CONFIG_PCIEC2_CLKREQ_AS_GPIO);
|
|
break;
|
|
}
|
|
afi_writel(port->pcie, data, AFI_PCIE_CONFIG);
|
|
}
|
|
|
|
void tegra_pcie_port_enable_per_pdev(struct pci_dev *pdev)
|
|
{
|
|
struct pci_dev *parent = NULL;
|
|
struct tegra_pcie_port *port = NULL;
|
|
struct pci_host_bridge *host = pci_find_host_bridge(pdev->bus);
|
|
struct tegra_pcie *pcie = pci_host_bridge_priv(host);
|
|
u32 rp = 0;
|
|
u32 data;
|
|
|
|
PR_FUNC_LINE;
|
|
parent = pdev->bus->self;
|
|
rp = PCI_SLOT(parent->devfn);
|
|
list_for_each_entry(port, &pcie->ports, list)
|
|
if (rp == port->index + 1)
|
|
break;
|
|
|
|
data = afi_readl(port->pcie, AFI_PCIE_CONFIG);
|
|
if (!((AFI_PCIE_CONFIG_PCIEC0_DISABLE_DEVICE << port->index) & data))
|
|
return;
|
|
|
|
switch (port->index) {
|
|
case 0:
|
|
data &= ~(AFI_PCIE_CONFIG_PCIEC0_DISABLE_DEVICE |
|
|
AFI_PCIE_CONFIG_PCIEC0_CLKREQ_AS_GPIO);
|
|
break;
|
|
case 1:
|
|
data &= ~(AFI_PCIE_CONFIG_PCIEC1_DISABLE_DEVICE |
|
|
AFI_PCIE_CONFIG_PCIEC1_CLKREQ_AS_GPIO);
|
|
break;
|
|
case 2:
|
|
data &= ~(AFI_PCIE_CONFIG_PCIEC2_DISABLE_DEVICE |
|
|
AFI_PCIE_CONFIG_PCIEC2_CLKREQ_AS_GPIO);
|
|
break;
|
|
}
|
|
afi_writel(port->pcie, data, AFI_PCIE_CONFIG);
|
|
}
|
|
EXPORT_SYMBOL(tegra_pcie_port_enable_per_pdev);
|
|
|
|
void tegra_pcie_port_disable_per_pdev(struct pci_dev *pdev)
|
|
{
|
|
struct pci_dev *parent = NULL;
|
|
struct tegra_pcie_port *port = NULL;
|
|
struct pci_host_bridge *host = pci_find_host_bridge(pdev->bus);
|
|
struct tegra_pcie *pcie = pci_host_bridge_priv(host);
|
|
u32 rp = 0;
|
|
u32 data;
|
|
|
|
PR_FUNC_LINE;
|
|
parent = pdev->bus->self;
|
|
rp = PCI_SLOT(parent->devfn);
|
|
list_for_each_entry(port, &pcie->ports, list)
|
|
if (rp == port->index + 1)
|
|
break;
|
|
|
|
data = afi_readl(port->pcie, AFI_PCIE_CONFIG);
|
|
switch (port->index) {
|
|
case 0:
|
|
data |= (AFI_PCIE_CONFIG_PCIEC0_DISABLE_DEVICE |
|
|
AFI_PCIE_CONFIG_PCIEC0_CLKREQ_AS_GPIO);
|
|
break;
|
|
case 1:
|
|
data |= (AFI_PCIE_CONFIG_PCIEC1_DISABLE_DEVICE |
|
|
AFI_PCIE_CONFIG_PCIEC1_CLKREQ_AS_GPIO);
|
|
break;
|
|
case 2:
|
|
data |= (AFI_PCIE_CONFIG_PCIEC2_DISABLE_DEVICE |
|
|
AFI_PCIE_CONFIG_PCIEC2_CLKREQ_AS_GPIO);
|
|
break;
|
|
}
|
|
afi_writel(port->pcie, data, AFI_PCIE_CONFIG);
|
|
}
|
|
EXPORT_SYMBOL(tegra_pcie_port_disable_per_pdev);
|
|
|
|
/*
|
|
* FIXME: If there are no PCIe cards attached, then calling this function
|
|
* can result in the increase of the bootup time as there are big timeout
|
|
* loops.
|
|
*/
|
|
#define TEGRA_PCIE_LINKUP_TIMEOUT 200 /* up to 1.2 seconds */
|
|
static bool tegra_pcie_port_check_link(struct tegra_pcie_port *port)
|
|
{
|
|
struct device *dev = port->pcie->dev;
|
|
unsigned int retries = 3;
|
|
unsigned long value;
|
|
|
|
/* override presence detection */
|
|
value = readl(port->base + NV_PCIE2_RP_PRIV_MISC);
|
|
value &= ~PCIE2_RP_PRIV_MISC_PRSNT_MAP_EP_ABSNT;
|
|
value |= PCIE2_RP_PRIV_MISC_PRSNT_MAP_EP_PRSNT;
|
|
writel(value, port->base + NV_PCIE2_RP_PRIV_MISC);
|
|
|
|
do {
|
|
unsigned int timeout = TEGRA_PCIE_LINKUP_TIMEOUT;
|
|
|
|
do {
|
|
value = readl(port->base + RP_VEND_XP);
|
|
|
|
if (value & RP_VEND_XP_DL_UP)
|
|
break;
|
|
|
|
usleep_range(1000, 2000);
|
|
} while (--timeout);
|
|
|
|
if (!timeout) {
|
|
dev_info(dev, "link %u down, retrying\n", port->index);
|
|
goto retry;
|
|
}
|
|
|
|
timeout = TEGRA_PCIE_LINKUP_TIMEOUT;
|
|
|
|
do {
|
|
value = readl(port->base + RP_LINK_CONTROL_STATUS);
|
|
|
|
if (value & RP_LINK_CONTROL_STATUS_DL_LINK_ACTIVE)
|
|
return true;
|
|
|
|
usleep_range(1000, 2000);
|
|
} while (--timeout);
|
|
|
|
retry:
|
|
tegra_pcie_port_reset(port);
|
|
} while (--retries);
|
|
|
|
return false;
|
|
}
|
|
|
|
static void tegra_pcie_apply_sw_war(struct tegra_pcie_port *port,
|
|
bool enum_done)
|
|
{
|
|
unsigned int data;
|
|
struct tegra_pcie *pcie = port->pcie;
|
|
struct pci_dev *pdev = NULL;
|
|
|
|
PR_FUNC_LINE;
|
|
if (enum_done) {
|
|
/* disable msi for port driver to avoid panic */
|
|
for_each_pci_dev(pdev)
|
|
if (pci_pcie_type(pdev) == PCI_EXP_TYPE_ROOT_PORT)
|
|
pdev->msi_enabled = 0;
|
|
} else {
|
|
|
|
/* Some of the old PCIe end points don't get enumerated
|
|
* if RP advertises both Gen-1 and Gen-2 speeds. Hence, the
|
|
* strategy followed here is to initially advertise only
|
|
* Gen-1 and after link is up, check end point's capability
|
|
* for Gen-2 and retrain link to Gen-2 speed
|
|
*/
|
|
data = rp_readl(port, RP_LINK_CONTROL_STATUS_2);
|
|
data &= ~RP_LINK_CONTROL_STATUS_2_TRGT_LNK_SPD_MASK;
|
|
data |= RP_LINK_CONTROL_STATUS_2_TRGT_LNK_SPD_GEN1;
|
|
rp_writel(port, data, RP_LINK_CONTROL_STATUS_2);
|
|
|
|
/* disable interrupts for LTR messages */
|
|
data = rp_readl(port, PCIE2_RP_L1SS_SPARE);
|
|
data &= ~PCIE2_RP_L1SS_SPARE_LTR_MSG_INT_EN;
|
|
data &= ~PCIE2_RP_L1SS_SPARE_LTR_MSG_RCV_STS;
|
|
rp_writel(port, data, PCIE2_RP_L1SS_SPARE);
|
|
|
|
/* Avoid warning during enumeration for invalid IRQ of RP */
|
|
data = rp_readl(port, NV_PCIE2_RP_INTR_BCR);
|
|
data |= NV_PCIE2_RP_INTR_BCR_INTR_LINE;
|
|
rp_writel(port, data, NV_PCIE2_RP_INTR_BCR);
|
|
|
|
/* Power saving configuration for sleep / idle */
|
|
data = rp_readl(port, NV_PCIE2_RP_VEND_XP_PAD_PWRDN);
|
|
data |= NV_PCIE2_RP_VEND_XP_PAD_PWRDN_DISABLED_EN;
|
|
data |= NV_PCIE2_RP_VEND_XP_PAD_PWRDN_DYNAMIC_EN;
|
|
data |= NV_PCIE2_RP_VEND_XP_PAD_PWRDN_L1_EN;
|
|
data |= NV_PCIE2_RP_VEND_XP_PAD_PWRDN_L1_CLKREQ_EN;
|
|
data |= NV_PCIE2_RP_VEND_XP_PAD_PWRDN_SLEEP_MODE_DYNAMIC_L1PP;
|
|
data |= NV_PCIE2_RP_VEND_XP_PAD_PWRDN_SLEEP_MODE_L1_L1PP;
|
|
data |= NV_PCIE2_RP_VEND_XP_PAD_PWRDN_SLEEP_MODE_L1_CLKREQ_L1PP;
|
|
rp_writel(port, data, NV_PCIE2_RP_VEND_XP_PAD_PWRDN);
|
|
|
|
/* resize buffers for better perf, bug#1447522 */
|
|
/* T210 WAR for perf bugs required when LPDDR4 */
|
|
/* memory is used with both ctlrs in X4_X1 config */
|
|
if (pcie->soc_data->perf_war &&
|
|
(pcie->xbar_config == AFI_PCIE_CONFIG_XBAR_CONFIG_X4_X1) &&
|
|
(pcie->num_ports == pcie->soc_data->num_ports)) {
|
|
struct tegra_pcie_port *temp_port;
|
|
list_for_each_entry(temp_port, &pcie->ports, list) {
|
|
data = rp_readl(temp_port,
|
|
NV_PCIE2_RP_XP_CTL_1);
|
|
data |= PCIE2_RP_XP_CTL_1_SPARE_BIT29;
|
|
rp_writel(temp_port, data,
|
|
NV_PCIE2_RP_XP_CTL_1);
|
|
|
|
data = rp_readl(temp_port,
|
|
NV_PCIE2_RP_TX_HDR_LIMIT);
|
|
if (temp_port->index)
|
|
data |= PCIE2_RP_TX_HDR_LIMIT_NPT_1;
|
|
else
|
|
data |= PCIE2_RP_TX_HDR_LIMIT_NPT_0;
|
|
rp_writel(temp_port, data,
|
|
NV_PCIE2_RP_TX_HDR_LIMIT);
|
|
}
|
|
}
|
|
|
|
if (pcie->soc_data->l1ss_rp_wakeup_war) {
|
|
/* Bug#1461732 WAR, set clkreq asserted delay greater than */
|
|
/* power off time (2us) to avoid RP wakeup in L1.2_ENTRY */
|
|
data = rp_readl(port,
|
|
NV_PCIE2_RP_L1_PM_SUBSTATES_1_CYA);
|
|
data &= ~PCIE2_RP_L1SS_1_CYA_CLKREQ_ASSERTED_DLY_MASK;
|
|
data |= PCIE2_RP_L1SS_1_CYA_CLKREQ_ASSERTED_DLY;
|
|
rp_writel(port, data,
|
|
NV_PCIE2_RP_L1_PM_SUBSTATES_1_CYA);
|
|
}
|
|
|
|
if (pcie->soc_data->link_speed_war) {
|
|
/* take care of link speed change error in corner cases */
|
|
data = rp_readl(port, NV_PCIE2_RP_VEND_CTL0);
|
|
data &= ~PCIE2_RP_VEND_CTL0_DSK_RST_PULSE_WIDTH_MASK;
|
|
data |= PCIE2_RP_VEND_CTL0_DSK_RST_PULSE_WIDTH;
|
|
rp_writel(port, data, NV_PCIE2_RP_VEND_CTL0);
|
|
}
|
|
|
|
if (pcie->soc_data->updateFC_timer_expire_war) {
|
|
data = rp_readl(port, RP_VEND_XP);
|
|
data &= ~RP_VEND_XP_UPDATE_FC_THRESHOLD;
|
|
data |= (0x60 << 18);
|
|
rp_writel(port, data, RP_VEND_XP);
|
|
}
|
|
|
|
/* Do timer settings only if clk25m freq equal to 19.2 MHz */
|
|
if (clk_get_rate(devm_clk_get(pcie->dev, "clk_m")) != 19200000)
|
|
return;
|
|
|
|
data = rp_readl(port, NV_PCIE2_RP_TIMEOUT0);
|
|
data &= ~PCIE2_RP_TIMEOUT0_PAD_PWRUP_MASK;
|
|
data |= PCIE2_RP_TIMEOUT0_PAD_PWRUP;
|
|
data &= ~PCIE2_RP_TIMEOUT0_PAD_PWRUP_CM_MASK;
|
|
data |= PCIE2_RP_TIMEOUT0_PAD_PWRUP_CM;
|
|
data &= ~PCIE2_RP_TIMEOUT0_PAD_SPDCHNG_GEN2_MASK;
|
|
data |= PCIE2_RP_TIMEOUT0_PAD_SPDCHNG_GEN2;
|
|
rp_writel(port, data, NV_PCIE2_RP_TIMEOUT0);
|
|
|
|
data = rp_readl(port, NV_PCIE2_RP_TIMEOUT1);
|
|
data &= ~PCIE2_RP_TIMEOUT1_RCVRY_SPD_SUCCESS_EIDLE_MASK;
|
|
data |= PCIE2_RP_TIMEOUT1_RCVRY_SPD_SUCCESS_EIDLE;
|
|
data &= ~PCIE2_RP_TIMEOUT1_RCVRY_SPD_UNSUCCESS_EIDLE_MASK;
|
|
data |= PCIE2_RP_TIMEOUT1_RCVRY_SPD_UNSUCCESS_EIDLE;
|
|
rp_writel(port, data, NV_PCIE2_RP_TIMEOUT1);
|
|
|
|
data = rp_readl(port, NV_PCIE2_RP_XP_REF);
|
|
data &= ~PCIE2_RP_XP_REF_MICROSECOND_LIMIT_MASK;
|
|
data |= PCIE2_RP_XP_REF_MICROSECOND_LIMIT;
|
|
data |= PCIE2_RP_XP_REF_MICROSECOND_ENABLE;
|
|
data |= PCIE2_RP_XP_REF_CPL_TO_OVERRIDE;
|
|
data &= ~PCIE2_RP_XP_REF_CPL_TO_CUSTOM_VALUE_MASK;
|
|
data |= PCIE2_RP_XP_REF_CPL_TO_CUSTOM_VALUE;
|
|
rp_writel(port, data, NV_PCIE2_RP_XP_REF);
|
|
|
|
data = rp_readl(port, NV_PCIE2_RP_L1_PM_SUBSTATES_1_CYA);
|
|
data &= ~PCIE2_RP_L1_PM_SUBSTATES_1_CYA_PWR_OFF_DLY_MASK;
|
|
data |= PCIE2_RP_L1_PM_SUBSTATES_1_CYA_PWR_OFF_DLY;
|
|
rp_writel(port, data, NV_PCIE2_RP_L1_PM_SUBSTATES_1_CYA);
|
|
|
|
data = rp_readl(port, NV_PCIE2_RP_L1_PM_SUBSTATES_2_CYA);
|
|
data &= ~PCIE2_RP_L1_PM_SUBSTATES_2_CYA_T_L1_2_DLY_MASK;
|
|
data |= PCIE2_RP_L1_PM_SUBSTATES_2_CYA_T_L1_2_DLY;
|
|
data &= ~PCIE2_RP_L1_PM_SUBSTATES_2_CYA_MICROSECOND_MASK;
|
|
data |= PCIE2_RP_L1_PM_SUBSTATES_2_CYA_MICROSECOND;
|
|
data &= ~PCIE2_RP_L1_PM_SUBSTATES_2_CYA_MICROSECOND_COMP_MASK;
|
|
data |= PCIE2_RP_L1_PM_SUBSTATES_2_CYA_MICROSECOND_COMP;
|
|
rp_writel(port, data, NV_PCIE2_RP_L1_PM_SUBSTATES_2_CYA);
|
|
|
|
if (pcie->soc_data->RAW_violation_war) {
|
|
/* WAR for RAW violation on T124/T132 platforms */
|
|
data = rp_readl(port, NV_PCIE2_RP_RX_HDR_LIMIT);
|
|
data &= ~PCIE2_RP_RX_HDR_LIMIT_PW_MASK;
|
|
data |= PCIE2_RP_RX_HDR_LIMIT_PW;
|
|
rp_writel(port, data, NV_PCIE2_RP_RX_HDR_LIMIT);
|
|
|
|
data = rp_readl(port, NV_PCIE2_RP_PRIV_XP_DL);
|
|
data |= PCIE2_RP_PRIV_XP_DL_GEN2_UPD_FC_TSHOLD;
|
|
rp_writel(port, data, NV_PCIE2_RP_PRIV_XP_DL);
|
|
|
|
data = rp_readl(port, RP_VEND_XP);
|
|
data |= RP_VEND_XP_UPDATE_FC_THRESHOLD;
|
|
rp_writel(port, data, RP_VEND_XP);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Enable various features of root port */
|
|
static void tegra_pcie_enable_rp_features(struct tegra_pcie_port *port)
|
|
{
|
|
unsigned int data;
|
|
|
|
PR_FUNC_LINE;
|
|
if (port->pcie->prod_list) {
|
|
if (tegra_prod_set_by_name(
|
|
&(port->pcie->pads),
|
|
"prod_c_pad",
|
|
port->pcie->prod_list)) {
|
|
dev_dbg(port->pcie->dev,
|
|
"pad prod settings are not found in DT\n");
|
|
}
|
|
|
|
if (tegra_prod_set_by_name(
|
|
&(port->base),
|
|
"prod_c_rp",
|
|
port->pcie->prod_list)) {
|
|
dev_dbg(port->pcie->dev,
|
|
"RP prod settings are not found in DT\n");
|
|
}
|
|
}
|
|
|
|
/* Optimal settings to enhance bandwidth */
|
|
data = rp_readl(port, RP_VEND_XP);
|
|
data |= RP_VEND_XP_OPPORTUNISTIC_ACK;
|
|
data |= RP_VEND_XP_OPPORTUNISTIC_UPDATEFC;
|
|
rp_writel(port, data, RP_VEND_XP);
|
|
|
|
/* Power mangagement settings */
|
|
/* Enable clock clamping by default and enable card detect */
|
|
data = rp_readl(port, NV_PCIE2_RP_PRIV_MISC);
|
|
data |= PCIE2_RP_PRIV_MISC_CTLR_CLK_CLAMP_ENABLE |
|
|
PCIE2_RP_PRIV_MISC_TMS_CLK_CLAMP_ENABLE;
|
|
if (port->pcie->soc_data->update_clamp_threshold) {
|
|
data |= PCIE2_RP_PRIV_MISC_CTLR_CLK_CLAMP_THRESHOLD |
|
|
PCIE2_RP_PRIV_MISC_TMS_CLK_CLAMP_THRESHOLD;
|
|
}
|
|
rp_writel(port, data, NV_PCIE2_RP_PRIV_MISC);
|
|
|
|
/* Enable ASPM - L1 state support by default */
|
|
data = rp_readl(port, NV_PCIE2_RP_VEND_XP1);
|
|
data |= NV_PCIE2_RP_VEND_XP_LINK_PVT_CTL_L1_ASPM_SUPPORT;
|
|
rp_writel(port, data, NV_PCIE2_RP_VEND_XP1);
|
|
|
|
/* LTSSM wait for DLLP to finish before entering L1 or L2/L3 */
|
|
/* to avoid truncating of PM mesgs resulting in reciever errors */
|
|
data = rp_readl(port, NV_PCIE2_RP_VEND_XP_BIST);
|
|
data |= PCIE2_RP_VEND_XP_BIST_GOTO_L1_L2_AFTER_DLLP_DONE;
|
|
rp_writel(port, data, NV_PCIE2_RP_VEND_XP_BIST);
|
|
|
|
/* unhide AER capability */
|
|
tegra_pcie_enable_aer(port, true);
|
|
|
|
/* Disable L1SS capability advertisement if CLKREQ is not present */
|
|
if (port->disable_clock_request) {
|
|
data = rp_readl(port, NV_PCIE2_RP_L1_PM_SUBSTATES_CYA);
|
|
data |= PCIE2_RP_L1_PM_SUBSTATES_CYA_HIDE_CAP;
|
|
rp_writel(port, data, NV_PCIE2_RP_L1_PM_SUBSTATES_CYA);
|
|
}
|
|
|
|
/* program timers for L1 substate support */
|
|
/* set cm_rtime = 30us and t_pwr_on = 70us as per HW team */
|
|
data = rp_readl(port, NV_PCIE2_RP_L1_PM_SUBSTATES_CYA);
|
|
data &= ~PCIE2_RP_L1_PM_SUBSTATES_CYA_CM_RTIME_MASK;
|
|
data |= (0x1E << PCIE2_RP_L1_PM_SUBSTATES_CYA_CM_RTIME_SHIFT);
|
|
rp_writel(port, data, NV_PCIE2_RP_L1_PM_SUBSTATES_CYA);
|
|
|
|
data = rp_readl(port, NV_PCIE2_RP_L1_PM_SUBSTATES_CYA);
|
|
data &= ~(PCIE2_RP_L1_PM_SUBSTATES_CYA_T_PWRN_SCL_MASK |
|
|
PCIE2_RP_L1_PM_SUBSTATES_CYA_T_PWRN_VAL_MASK);
|
|
data |= (1 << PCIE2_RP_L1_PM_SUBSTATES_CYA_T_PWRN_SCL_SHIFT) |
|
|
(7 << PCIE2_RP_L1_PM_SUBSTATES_CYA_T_PWRN_VAL_SHIFT);
|
|
rp_writel(port, data, NV_PCIE2_RP_L1_PM_SUBSTATES_CYA);
|
|
|
|
tegra_pcie_apply_sw_war(port, false);
|
|
}
|
|
|
|
static void tegra_pcie_update_lane_width(struct tegra_pcie_port *port)
|
|
{
|
|
port->lanes = rp_readl(port, RP_LINK_CONTROL_STATUS);
|
|
port->lanes = (port->lanes &
|
|
RP_LINK_CONTROL_STATUS_NEG_LINK_WIDTH) >> 20;
|
|
}
|
|
|
|
static void tegra_pcie_update_pads2plle(struct tegra_pcie_port *port)
|
|
{
|
|
unsigned long ctrl = 0;
|
|
u32 val = 0;
|
|
|
|
ctrl = tegra_pcie_port_get_pex_ctrl(port);
|
|
/* AFI_PEX_STATUS is AFI_PEX_CTRL + 4 */
|
|
val = afi_readl(port->pcie, ctrl + 4);
|
|
if (val & 0x1) {
|
|
val = afi_readl(port->pcie, AFI_PLLE_CONTROL);
|
|
val &= ~AFI_PLLE_CONTROL_PADS2PLLE_CONTROL_EN;
|
|
afi_writel(port->pcie, val, AFI_PLLE_CONTROL);
|
|
}
|
|
}
|
|
|
|
static void mbist_war(struct tegra_pcie *pcie, bool apply)
|
|
{
|
|
struct tegra_pcie_port *port, *tmp;
|
|
u32 data;
|
|
|
|
list_for_each_entry_safe(port, tmp, &pcie->ports, list) {
|
|
/* nature of MBIST bug is such that it needs to be applied
|
|
* only for RootPort-0 even if there are no devices
|
|
* connected to it */
|
|
if (port->index == 0) {
|
|
data = rp_readl(port, NV_PCIE2_RP_VEND_CTL2);
|
|
if (apply)
|
|
data |= PCIE2_RP_VEND_CTL2_PCA_ENABLE;
|
|
else
|
|
data &= ~PCIE2_RP_VEND_CTL2_PCA_ENABLE;
|
|
rp_writel(port, data, NV_PCIE2_RP_VEND_CTL2);
|
|
}
|
|
}
|
|
}
|
|
|
|
static int tegra_pcie_mxm_pwr_init(struct tegra_pcie_port *port)
|
|
{
|
|
|
|
mdelay(100);
|
|
if (!(gpio_get_value(port->pwr_gd_gpio)))
|
|
return 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
static void tegra_pcie_check_ports(struct tegra_pcie *pcie)
|
|
{
|
|
struct tegra_pcie_port *port, *tmp;
|
|
|
|
PR_FUNC_LINE;
|
|
pcie->num_ports = 0;
|
|
|
|
if (pcie->soc_data->mbist_war)
|
|
mbist_war(pcie, true);
|
|
|
|
list_for_each_entry_safe(port, tmp, &pcie->ports, list) {
|
|
dev_info(pcie->dev, "probing port %u, using %u lanes\n",
|
|
port->index, port->lanes);
|
|
|
|
tegra_pcie_port_enable(port);
|
|
tegra_pcie_enable_rp_features(port);
|
|
/* override presence detection */
|
|
if (gpio_is_valid(port->gpio_presence_detection))
|
|
tegra_pcie_prsnt_map_override(port,
|
|
!(gpio_get_value_cansleep(
|
|
port->gpio_presence_detection)));
|
|
else
|
|
tegra_pcie_prsnt_map_override(port, true);
|
|
}
|
|
|
|
/* Wait for clock to latch (min of 100us) */
|
|
udelay(100);
|
|
reset_control_deassert(pcie->pciex_rst);
|
|
/* at this point in time, there is no end point which would
|
|
* take more than 20 msec for root port to detect receiver and
|
|
* set AUX_TX_RDET_STATUS bit. This would bring link up checking
|
|
* time from its current value (around 200ms) to flat 20ms
|
|
*/
|
|
usleep_range(19000, 21000);
|
|
list_for_each_entry_safe(port, tmp, &pcie->ports, list) {
|
|
if (tegra_pcie_port_check_link(port)) {
|
|
port->status = 1;
|
|
port->ep_status = 1;
|
|
pcie->num_ports++;
|
|
tegra_pcie_update_lane_width(port);
|
|
tegra_pcie_update_pads2plle(port);
|
|
continue;
|
|
}
|
|
port->ep_status = 0;
|
|
dev_info(pcie->dev, "link %u down, ignoring\n", port->index);
|
|
tegra_pcie_port_disable(port);
|
|
}
|
|
/* configure all links to gen2 speed by default */
|
|
tegra_pcie_link_speed(pcie);
|
|
if (pcie->soc_data->mbist_war)
|
|
mbist_war(pcie, false);
|
|
}
|
|
|
|
static int tegra_pcie_conf_gpios(struct tegra_pcie *pcie)
|
|
{
|
|
int irq, err = 0;
|
|
struct tegra_pcie_port *port, *tmp;
|
|
|
|
PR_FUNC_LINE;
|
|
if (gpio_is_valid(pcie->plat_data->gpio_hot_plug)) {
|
|
/* configure gpio for hotplug detection */
|
|
dev_info(pcie->dev, "acquiring hotplug_detect = %d\n",
|
|
pcie->plat_data->gpio_hot_plug);
|
|
err = devm_gpio_request(pcie->dev,
|
|
pcie->plat_data->gpio_hot_plug,
|
|
"pcie_hotplug_detect");
|
|
if (err < 0) {
|
|
dev_err(pcie->dev, "%s: gpio_request failed %d\n",
|
|
__func__, err);
|
|
return err;
|
|
}
|
|
err = gpio_direction_input(
|
|
pcie->plat_data->gpio_hot_plug);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev,
|
|
"%s: gpio_direction_input failed %d\n",
|
|
__func__, err);
|
|
return err;
|
|
}
|
|
irq = gpio_to_irq(pcie->plat_data->gpio_hot_plug);
|
|
if (irq < 0) {
|
|
dev_err(pcie->dev,
|
|
"Unable to get irq for hotplug_detect\n");
|
|
return err;
|
|
}
|
|
err = devm_request_irq(pcie->dev, (unsigned int)irq,
|
|
gpio_pcie_detect_isr,
|
|
IRQF_TRIGGER_RISING | IRQF_TRIGGER_FALLING,
|
|
"pcie_hotplug_detect",
|
|
(void *)pcie);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev,
|
|
"Unable to claim irq for hotplug_detect\n");
|
|
return err;
|
|
}
|
|
}
|
|
if (gpio_is_valid(pcie->plat_data->gpio_x1_slot)) {
|
|
err = devm_gpio_request(pcie->dev,
|
|
pcie->plat_data->gpio_x1_slot, "pcie_x1_slot");
|
|
if (err < 0) {
|
|
dev_err(pcie->dev,
|
|
"%s: pcie_x1_slot gpio_request failed %d\n",
|
|
__func__, err);
|
|
return err;
|
|
}
|
|
err = gpio_direction_output(
|
|
pcie->plat_data->gpio_x1_slot, 1);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev,
|
|
"%s: pcie_x1_slot gpio_direction_output failed %d\n",
|
|
__func__, err);
|
|
return err;
|
|
}
|
|
gpio_set_value_cansleep(
|
|
pcie->plat_data->gpio_x1_slot, 1);
|
|
}
|
|
if (gpio_is_valid(pcie->plat_data->gpio_wake)) {
|
|
err = devm_gpio_request(pcie->dev,
|
|
pcie->plat_data->gpio_wake, "pcie_wake");
|
|
if (err < 0) {
|
|
dev_err(pcie->dev,
|
|
"%s: pcie_wake gpio_request failed %d\n",
|
|
__func__, err);
|
|
return err;
|
|
}
|
|
err = gpio_direction_input(
|
|
pcie->plat_data->gpio_wake);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev,
|
|
"%s: pcie_wake gpio_direction_input failed %d\n",
|
|
__func__, err);
|
|
return err;
|
|
}
|
|
}
|
|
|
|
list_for_each_entry_safe(port, tmp, &pcie->ports, list) {
|
|
if (gpio_is_valid(port->gpio_presence_detection)) {
|
|
err = devm_gpio_request_one(pcie->dev,
|
|
port->gpio_presence_detection,
|
|
GPIOF_DIR_IN,
|
|
"pcie_presence_detection");
|
|
if (err < 0) {
|
|
dev_err(pcie->dev,
|
|
"%s: pcie_prsnt gpio_request failed %d\n",
|
|
__func__, err);
|
|
return err;
|
|
}
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int tegra_pcie_scale_voltage(struct tegra_pcie *pcie)
|
|
{
|
|
struct tegra_pcie_port *port, *tmp;
|
|
struct tegra_pcie_soc_data *sd = pcie->soc_data;
|
|
int err = 0;
|
|
u32 data = 0;
|
|
u32 active_lanes = 0;
|
|
bool is_gen2 = false;
|
|
|
|
PR_FUNC_LINE;
|
|
list_for_each_entry_safe(port, tmp, &pcie->ports, list) {
|
|
if (!port->status)
|
|
continue;
|
|
data = rp_readl(port, RP_LINK_CONTROL_STATUS);
|
|
active_lanes += ((data &
|
|
RP_LINK_CONTROL_STATUS_NEG_LINK_WIDTH) >> 20);
|
|
if (((data & RP_LINK_CONTROL_STATUS_LINK_SPEED) >> 16) == 2)
|
|
is_gen2 = true;
|
|
}
|
|
|
|
if (sd->dvfs_mselect) {
|
|
struct clk *mselect_clk;
|
|
|
|
active_lanes = 0;
|
|
dev_dbg(pcie->dev, "mselect_clk is set @ %u\n",
|
|
sd->dvfs_tbl[active_lanes][is_gen2].afi_clk);
|
|
mselect_clk = devm_clk_get(pcie->dev, "mselect");
|
|
if (IS_ERR(mselect_clk)) {
|
|
dev_err(pcie->dev, "mselect clk_get failed: %ld\n",
|
|
PTR_ERR(mselect_clk));
|
|
return PTR_ERR(mselect_clk);
|
|
}
|
|
err = clk_set_rate(mselect_clk,
|
|
sd->dvfs_tbl[active_lanes][is_gen2].afi_clk);
|
|
if (err) {
|
|
dev_err(pcie->dev,
|
|
"setting mselect clk to %u failed : %d\n",
|
|
sd->dvfs_tbl[active_lanes][is_gen2].afi_clk,
|
|
err);
|
|
return err;
|
|
}
|
|
}
|
|
|
|
if (sd->dvfs_afi) {
|
|
dev_dbg(pcie->dev, "afi_clk is set @ %u\n",
|
|
sd->dvfs_tbl[active_lanes][is_gen2].afi_clk);
|
|
err = clk_set_rate(devm_clk_get(pcie->dev, "afi"),
|
|
sd->dvfs_tbl[active_lanes][is_gen2].afi_clk);
|
|
if (err) {
|
|
dev_err(pcie->dev,
|
|
"setting afi clk to %u failed : %d\n",
|
|
sd->dvfs_tbl[active_lanes][is_gen2].afi_clk,
|
|
err);
|
|
return err;
|
|
}
|
|
}
|
|
|
|
dev_dbg(pcie->dev, "emc_clk is set @ %u\n",
|
|
sd->dvfs_tbl[active_lanes][is_gen2].emc_clk);
|
|
err = tegra_bwmgr_set_emc(pcie->emc_bwmgr,
|
|
sd->dvfs_tbl[active_lanes][is_gen2].emc_clk,
|
|
TEGRA_BWMGR_SET_EMC_FLOOR);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev, "setting emc clk to %u failed : %d\n",
|
|
sd->dvfs_tbl[active_lanes][is_gen2].emc_clk, err);
|
|
return err;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static void tegra_pcie_change_link_speed(struct tegra_pcie *pcie)
|
|
{
|
|
struct device *dev = pcie->dev;
|
|
struct tegra_pcie_port *port, *tmp;
|
|
ktime_t deadline;
|
|
u32 value;
|
|
|
|
list_for_each_entry_safe(port, tmp, &pcie->ports, list) {
|
|
/*
|
|
* "Supported Link Speeds Vector" in "Link Capabilities 2"
|
|
* is not supported by Tegra. tegra_pcie_change_link_speed()
|
|
* is called only for Tegra chips which support Gen2.
|
|
* So there no harm if supported link speed is not verified.
|
|
*/
|
|
value = readl(port->base + RP_LINK_CONTROL_STATUS_2);
|
|
value &= ~PCI_EXP_LNKSTA_CLS;
|
|
value |= PCI_EXP_LNKSTA_CLS_5_0GB;
|
|
writel(value, port->base + RP_LINK_CONTROL_STATUS_2);
|
|
|
|
/*
|
|
* Poll until link comes back from recovery to avoid race
|
|
* condition.
|
|
*/
|
|
deadline = ktime_add_us(ktime_get(), LINK_RETRAIN_TIMEOUT);
|
|
|
|
while (ktime_before(ktime_get(), deadline)) {
|
|
value = readl(port->base + RP_LINK_CONTROL_STATUS);
|
|
if ((value & PCI_EXP_LNKSTA_LT) == 0)
|
|
break;
|
|
|
|
usleep_range(2000, 3000);
|
|
}
|
|
if (value & PCI_EXP_LNKSTA_LT)
|
|
dev_warn(dev, "PCIe port %u link is in recovery\n",
|
|
port->index);
|
|
|
|
/* Clear BW Management Status */
|
|
value = readl(port->base + RP_LINK_CONTROL_STATUS);
|
|
value |= RP_LINK_CONTROL_STATUS_BW_MGMT_STATUS;
|
|
writel(value, port->base + RP_LINK_CONTROL_STATUS);
|
|
|
|
/* Retrain the link */
|
|
value = readl(port->base + RP_LINK_CONTROL_STATUS);
|
|
value |= PCI_EXP_LNKCTL_RL;
|
|
writel(value, port->base + RP_LINK_CONTROL_STATUS);
|
|
|
|
deadline = ktime_add_us(ktime_get(), LINK_RETRAIN_TIMEOUT);
|
|
|
|
while (ktime_before(ktime_get(), deadline)) {
|
|
value = readl(port->base + RP_LINK_CONTROL_STATUS);
|
|
if (value & RP_LINK_CONTROL_STATUS_BW_MGMT_STATUS)
|
|
break;
|
|
|
|
usleep_range(2000, 3000);
|
|
}
|
|
if (value & PCI_EXP_LNKSTA_LT)
|
|
dev_err(dev, "failed to retrain link of port %u\n",
|
|
port->index);
|
|
}
|
|
}
|
|
|
|
static void tegra_pcie_link_speed(struct tegra_pcie *pcie)
|
|
{
|
|
PR_FUNC_LINE;
|
|
|
|
tegra_pcie_change_link_speed(pcie);
|
|
tegra_pcie_scale_voltage(pcie);
|
|
|
|
return;
|
|
}
|
|
|
|
static void tegra_pcie_enable_features(struct tegra_pcie *pcie)
|
|
{
|
|
struct tegra_pcie_port *port;
|
|
|
|
PR_FUNC_LINE;
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
if (port->status)
|
|
tegra_pcie_apply_sw_war(port, true);
|
|
}
|
|
}
|
|
static int tegra_pcie_enable_msi(struct tegra_pcie *, bool);
|
|
static int tegra_pcie_disable_msi(struct tegra_pcie *pcie);
|
|
|
|
static int tegra_pcie_init(struct tegra_pcie *pcie)
|
|
{
|
|
int err = 0;
|
|
struct pci_bus *child;
|
|
struct pci_host_bridge *host = pcie->host;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
INIT_WORK(&pcie->hotplug_detect, work_hotplug_handler);
|
|
err = tegra_pcie_get_resources(pcie);
|
|
if (err) {
|
|
dev_err(pcie->dev, "PCIE: get resources failed\n");
|
|
return err;
|
|
}
|
|
err = tegra_pcie_power_on(pcie);
|
|
if (err) {
|
|
dev_err(pcie->dev, "PCIE: Failed to power on: %d\n", err);
|
|
goto fail_release_resource;
|
|
}
|
|
err = tegra_pcie_conf_gpios(pcie);
|
|
if (err) {
|
|
dev_err(pcie->dev, "PCIE: configuring gpios failed\n");
|
|
goto fail_power_off;
|
|
}
|
|
|
|
if (!pcie->num_ports) {
|
|
dev_info(pcie->dev, "PCIE: no end points detected\n");
|
|
err = -ENODEV;
|
|
goto fail_power_off;
|
|
}
|
|
if (IS_ENABLED(CONFIG_PCI_MSI)) {
|
|
err = tegra_pcie_enable_msi(pcie, false);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev,
|
|
"failed to enable MSI support: %d\n",
|
|
err);
|
|
goto fail_release_resource;
|
|
}
|
|
}
|
|
|
|
pci_add_flags(PCI_REASSIGN_ALL_RSRC | PCI_REASSIGN_ALL_BUS);
|
|
host->busnr = pcie->busn.start;
|
|
host->dev.parent = pcie->dev;
|
|
host->ops = &tegra_pcie_ops;
|
|
|
|
err = pci_register_host_bridge(host);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev, "failed to register host: %d\n", err);
|
|
goto fail_power_off;
|
|
}
|
|
|
|
pci_scan_child_bus(host->bus);
|
|
|
|
pci_fixup_irqs(pci_common_swizzle, tegra_pcie_map_irq);
|
|
pci_bus_size_bridges(host->bus);
|
|
pci_bus_assign_resources(host->bus);
|
|
|
|
list_for_each_entry(child, &host->bus->children, node)
|
|
pcie_bus_configure_settings(child);
|
|
|
|
tegra_pcie_postinit();
|
|
tegra_pcie_enable_features(pcie);
|
|
|
|
pci_bus_add_devices(host->bus);
|
|
|
|
/* register pcie device as wakeup source */
|
|
device_init_wakeup(pcie->dev, true);
|
|
|
|
return 0;
|
|
|
|
fail_power_off:
|
|
tegra_pcie_power_off(pcie);
|
|
fail_release_resource:
|
|
tegra_pcie_release_resources(pcie);
|
|
return err;
|
|
}
|
|
|
|
/* 1:1 matching of these to the MSI vectors, 1 per bit */
|
|
/* and each mapping matches one of the available interrupts */
|
|
struct msi_map_entry {
|
|
bool used;
|
|
u8 index;
|
|
int irq;
|
|
};
|
|
|
|
/* hardware supports 256 max*/
|
|
#if (INT_PCI_MSI_NR > 256)
|
|
#error "INT_PCI_MSI_NR too big"
|
|
#endif
|
|
|
|
static int tegra_msi_alloc(struct tegra_msi *chip)
|
|
{
|
|
int msi;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
mutex_lock(&chip->lock);
|
|
|
|
msi = find_first_zero_bit(chip->used, INT_PCI_MSI_NR);
|
|
if (msi < INT_PCI_MSI_NR)
|
|
set_bit(msi, chip->used);
|
|
else
|
|
msi = -ENOSPC;
|
|
|
|
mutex_unlock(&chip->lock);
|
|
|
|
return msi;
|
|
}
|
|
|
|
static void tegra_msi_free(struct tegra_msi *chip, unsigned long irq)
|
|
{
|
|
struct device *dev = chip->chip.dev;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
mutex_lock(&chip->lock);
|
|
|
|
if (!test_bit(irq, chip->used))
|
|
dev_err(dev, "trying to free unused MSI#%lu\n", irq);
|
|
else
|
|
clear_bit(irq, chip->used);
|
|
|
|
mutex_unlock(&chip->lock);
|
|
}
|
|
|
|
|
|
static irqreturn_t tegra_pcie_msi_irq(int irq, void *data)
|
|
{
|
|
struct tegra_pcie *pcie = data;
|
|
struct tegra_msi *msi = &pcie->msi;
|
|
unsigned int i, processed = 0;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
for (i = 0; i < 8; i++) {
|
|
unsigned long reg = afi_readl(pcie, AFI_MSI_VEC0_0 + i * 4);
|
|
|
|
while (reg) {
|
|
unsigned int offset = find_first_bit(®, 32);
|
|
unsigned int index = i * 32 + offset;
|
|
unsigned int irq_num;
|
|
|
|
/* check if there are any interrupts in this reg */
|
|
if (offset == 32)
|
|
break;
|
|
|
|
/* clear the interrupt */
|
|
afi_writel(pcie, 1 << offset, AFI_MSI_VEC0_0 + i * 4);
|
|
|
|
irq_num = irq_find_mapping(msi->domain, index);
|
|
if (irq_num) {
|
|
if (test_bit(index, msi->used))
|
|
generic_handle_irq(irq_num);
|
|
else
|
|
dev_info(pcie->dev, "unhandled MSI\n");
|
|
} else {
|
|
/*
|
|
* that's weird who triggered this?
|
|
* just clear it
|
|
*/
|
|
dev_info(pcie->dev, "unexpected MSI\n");
|
|
}
|
|
|
|
/* see if there's any more pending in this vector */
|
|
reg = afi_readl(pcie, AFI_MSI_VEC0_0 + i * 4);
|
|
|
|
processed++;
|
|
}
|
|
}
|
|
|
|
return processed > 0 ? IRQ_HANDLED : IRQ_NONE;
|
|
}
|
|
|
|
static int tegra_msi_setup_irq(struct msi_controller *chip, struct pci_dev *pdev,
|
|
struct msi_desc *desc)
|
|
{
|
|
struct tegra_msi *msi = to_tegra_msi(chip);
|
|
struct msi_msg msg;
|
|
unsigned int irq;
|
|
int hwirq;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
hwirq = tegra_msi_alloc(msi);
|
|
if (hwirq < 0)
|
|
return hwirq;
|
|
|
|
irq = irq_create_mapping(msi->domain, hwirq);
|
|
if (!irq)
|
|
return -EINVAL;
|
|
|
|
irq_set_msi_desc(irq, desc);
|
|
|
|
msg.address_lo = lower_32_bits(msi->phys);
|
|
#ifdef CONFIG_ARM64
|
|
msg.address_hi = upper_32_bits(msi->phys);
|
|
#else
|
|
msg.address_hi = 0;
|
|
#endif
|
|
msg.data = hwirq;
|
|
|
|
write_msi_msg(irq, &msg);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void tegra_msi_teardown_irq(struct msi_controller *chip, unsigned int irq)
|
|
{
|
|
struct tegra_msi *msi = to_tegra_msi(chip);
|
|
struct irq_data *d = irq_get_irq_data(irq);
|
|
|
|
PR_FUNC_LINE;
|
|
tegra_msi_free(msi, d->hwirq);
|
|
}
|
|
|
|
static struct irq_chip tegra_msi_irq_chip = {
|
|
.name = "Tegra PCIe MSI",
|
|
.irq_enable = unmask_msi_irq,
|
|
.irq_disable = mask_msi_irq,
|
|
.irq_mask = mask_msi_irq,
|
|
.irq_unmask = unmask_msi_irq,
|
|
};
|
|
|
|
static int tegra_msi_map(struct irq_domain *domain, unsigned int irq,
|
|
irq_hw_number_t hwirq)
|
|
{
|
|
PR_FUNC_LINE;
|
|
irq_set_chip_and_handler(irq, &tegra_msi_irq_chip, handle_simple_irq);
|
|
irq_set_chip_data(irq, domain->host_data);
|
|
return 0;
|
|
}
|
|
|
|
static const struct irq_domain_ops msi_domain_ops = {
|
|
.map = tegra_msi_map,
|
|
};
|
|
|
|
static int tegra_pcie_enable_msi(struct tegra_pcie *pcie, bool no_init)
|
|
{
|
|
struct platform_device *pdev = to_platform_device(pcie->dev);
|
|
struct tegra_msi *msi = &pcie->msi;
|
|
int err;
|
|
u32 reg;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
if (!msi->virt) {
|
|
if (no_init)
|
|
return true;
|
|
|
|
mutex_init(&msi->lock);
|
|
|
|
msi->chip.dev = pcie->dev;
|
|
msi->chip.setup_irq = tegra_msi_setup_irq;
|
|
msi->chip.teardown_irq = tegra_msi_teardown_irq;
|
|
|
|
msi->domain = irq_domain_add_linear(pcie->dev->of_node,
|
|
INT_PCI_MSI_NR, &msi_domain_ops, &msi->chip);
|
|
if (!msi->domain) {
|
|
dev_err(&pdev->dev, "failed to create IRQ domain\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
err = platform_get_irq_byname(pdev, "msi");
|
|
if (err < 0) {
|
|
dev_err(&pdev->dev, "failed to get IRQ: %d\n", err);
|
|
goto free_irq_domain;
|
|
}
|
|
|
|
msi->irq = err;
|
|
err = request_irq(msi->irq, tegra_pcie_msi_irq, IRQF_NO_THREAD,
|
|
tegra_msi_irq_chip.name, pcie);
|
|
if (err < 0) {
|
|
dev_err(&pdev->dev, "failed to request IRQ: %d\n", err);
|
|
goto free_irq_domain;
|
|
}
|
|
|
|
/* setup AFI/FPCI range */
|
|
err = dma_set_coherent_mask(pcie->dev, DMA_BIT_MASK(32));
|
|
if (err < 0) {
|
|
dev_err(&pdev->dev, "dma_set_coherent_mask() failed: %d\n",
|
|
err);
|
|
goto free_irq;
|
|
}
|
|
|
|
msi->virt = dma_alloc_coherent(pcie->dev, PAGE_SIZE,
|
|
&msi->phys, GFP_KERNEL);
|
|
if (!msi->virt) {
|
|
dev_err(&pdev->dev, "%s: failed to alloc dma mem\n", __func__);
|
|
err = -ENOMEM;
|
|
goto free_irq;
|
|
}
|
|
}
|
|
|
|
afi_writel(pcie, msi->phys >> 8, AFI_MSI_FPCI_BAR_ST);
|
|
afi_writel(pcie, msi->phys, AFI_MSI_AXI_BAR_ST);
|
|
/* this register is in 4K increments */
|
|
afi_writel(pcie, 1, AFI_MSI_BAR_SZ);
|
|
|
|
/* enable all MSI vectors */
|
|
afi_writel(pcie, 0xffffffff, AFI_MSI_EN_VEC0_0);
|
|
afi_writel(pcie, 0xffffffff, AFI_MSI_EN_VEC1_0);
|
|
afi_writel(pcie, 0xffffffff, AFI_MSI_EN_VEC2_0);
|
|
afi_writel(pcie, 0xffffffff, AFI_MSI_EN_VEC3_0);
|
|
afi_writel(pcie, 0xffffffff, AFI_MSI_EN_VEC4_0);
|
|
afi_writel(pcie, 0xffffffff, AFI_MSI_EN_VEC5_0);
|
|
afi_writel(pcie, 0xffffffff, AFI_MSI_EN_VEC6_0);
|
|
afi_writel(pcie, 0xffffffff, AFI_MSI_EN_VEC7_0);
|
|
|
|
/* and unmask the MSI interrupt */
|
|
reg = afi_readl(pcie, AFI_INTR_MASK);
|
|
reg |= AFI_INTR_MASK_MSI_MASK;
|
|
afi_writel(pcie, reg, AFI_INTR_MASK);
|
|
|
|
return 0;
|
|
|
|
free_irq:
|
|
free_irq(msi->irq, pcie);
|
|
free_irq_domain:
|
|
irq_domain_remove(msi->domain);
|
|
return err;
|
|
}
|
|
|
|
static int tegra_pcie_disable_msi(struct tegra_pcie *pcie)
|
|
{
|
|
struct tegra_msi *msi = &pcie->msi;
|
|
unsigned int i, irq;
|
|
u32 value;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
if (pcie->pcie_power_enabled == 0)
|
|
return 0;
|
|
|
|
/* mask the MSI interrupt */
|
|
value = afi_readl(pcie, AFI_INTR_MASK);
|
|
value &= ~AFI_INTR_MASK_MSI_MASK;
|
|
afi_writel(pcie, value, AFI_INTR_MASK);
|
|
|
|
/* disable all MSI vectors */
|
|
afi_writel(pcie, 0, AFI_MSI_EN_VEC0_0);
|
|
afi_writel(pcie, 0, AFI_MSI_EN_VEC1_0);
|
|
afi_writel(pcie, 0, AFI_MSI_EN_VEC2_0);
|
|
afi_writel(pcie, 0, AFI_MSI_EN_VEC3_0);
|
|
afi_writel(pcie, 0, AFI_MSI_EN_VEC4_0);
|
|
afi_writel(pcie, 0, AFI_MSI_EN_VEC5_0);
|
|
afi_writel(pcie, 0, AFI_MSI_EN_VEC6_0);
|
|
afi_writel(pcie, 0, AFI_MSI_EN_VEC7_0);
|
|
|
|
dma_free_coherent(pcie->dev, PAGE_SIZE, msi->virt, msi->phys);
|
|
|
|
if (msi->irq > 0)
|
|
free_irq(msi->irq, pcie);
|
|
|
|
for (i = 0; i < INT_PCI_MSI_NR; i++) {
|
|
irq = irq_find_mapping(msi->domain, i);
|
|
if (irq > 0)
|
|
irq_dispose_mapping(irq);
|
|
}
|
|
|
|
irq_domain_remove(msi->domain);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void update_rp_lanes(struct tegra_pcie *pcie, u32 lanes)
|
|
{
|
|
struct tegra_pcie_port *port = NULL;
|
|
|
|
list_for_each_entry(port, &pcie->ports, list)
|
|
port->lanes = (lanes >> (port->index << 3)) & 0xFF;
|
|
|
|
}
|
|
|
|
static int tegra_pcie_get_xbar_config(struct tegra_pcie *pcie, u32 lanes,
|
|
u32 *xbar)
|
|
{
|
|
struct device_node *np = pcie->dev->of_node;
|
|
|
|
if (of_device_is_compatible(np, "nvidia,tegra210b01-pcie")) {
|
|
switch (lanes) {
|
|
case 0x0104:
|
|
dev_info(pcie->dev, "4x1, 1x1 configuration\n");
|
|
*xbar = AFI_PCIE_CONFIG_XBAR_CONFIG_X4_X1;
|
|
return 0;
|
|
default:
|
|
dev_info(pcie->dev, "wrong configuration updated in DT, "
|
|
"switching to default 4x1, 1x1 configuration\n");
|
|
*xbar = AFI_PCIE_CONFIG_XBAR_CONFIG_X4_X1;
|
|
update_rp_lanes(pcie, 0x0104);
|
|
return 0;
|
|
}
|
|
} else if (of_device_is_compatible(np, "nvidia,tegra124-pcie") ||
|
|
of_device_is_compatible(np, "nvidia,tegra210-pcie")) {
|
|
switch (lanes) {
|
|
case 0x0104:
|
|
dev_info(pcie->dev, "4x1, 1x1 configuration\n");
|
|
*xbar = AFI_PCIE_CONFIG_XBAR_CONFIG_X4_X1;
|
|
return 0;
|
|
|
|
case 0x0102:
|
|
dev_info(pcie->dev, "2x1, 1x1 configuration\n");
|
|
*xbar = AFI_PCIE_CONFIG_XBAR_CONFIG_X2_X1;
|
|
return 0;
|
|
default:
|
|
dev_info(pcie->dev, "wrong configuration updated in DT, "
|
|
"switching to default 4x1, 1x1 configuration\n");
|
|
*xbar = AFI_PCIE_CONFIG_XBAR_CONFIG_X4_X1;
|
|
update_rp_lanes(pcie, 0x0104);
|
|
return 0;
|
|
}
|
|
} else if (of_device_is_compatible(np, "nvidia,tegra186-pcie")) {
|
|
switch (lanes) {
|
|
case 0x010004:
|
|
dev_info(pcie->dev, "4x1, 1x1 configuration\n");
|
|
*xbar = AFI_PCIE_CONFIG_XBAR_CONFIG_X4_X0_X1;
|
|
return 0;
|
|
|
|
case 0x010102:
|
|
dev_info(pcie->dev, "2x1, 1x1, 1x1 configuration\n");
|
|
*xbar = AFI_PCIE_CONFIG_XBAR_CONFIG_X2_X1_X1;
|
|
return 0;
|
|
|
|
case 0x010101:
|
|
dev_info(pcie->dev, "1x1, 1x1, 1x1 configuration\n");
|
|
*xbar = AFI_PCIE_CONFIG_XBAR_CONFIG_X1_X1_X1;
|
|
return 0;
|
|
default:
|
|
dev_info(pcie->dev, "wrong configuration updated in DT,"
|
|
" switching to default 2x1, 1x1, 1x1 "
|
|
"configuration\n");
|
|
*xbar = AFI_PCIE_CONFIG_XBAR_CONFIG_X2_X1_X1;
|
|
update_rp_lanes(pcie, 0x010102);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return -EINVAL;
|
|
}
|
|
|
|
|
|
static void tegra_pcie_read_plat_data(struct tegra_pcie *pcie)
|
|
{
|
|
struct device_node *node = pcie->dev->of_node;
|
|
|
|
PR_FUNC_LINE;
|
|
of_property_read_u32(node, "nvidia,boot-detect-delay",
|
|
&pcie->plat_data->boot_detect_delay);
|
|
pcie->plat_data->gpio_hot_plug =
|
|
of_get_named_gpio(node, "nvidia,hot-plug-gpio", 0);
|
|
pcie->plat_data->gpio_wake =
|
|
of_get_named_gpio(node, "nvidia,wake-gpio", 0);
|
|
pcie->plat_data->gpio_x1_slot =
|
|
of_get_named_gpio(node, "nvidia,x1-slot-gpio", 0);
|
|
pcie->plat_data->has_memtype_lpddr4 =
|
|
of_property_read_bool(node, "nvidia,has_memtype_lpddr4");
|
|
}
|
|
|
|
static char *t124_rail_names[] = {"hvdd-pex", "hvdd-pex-pll-e", "dvddio-pex",
|
|
"avddio-pex", "avdd-pex-pll", "vddio-pex-ctl"};
|
|
|
|
static char *t210_rail_names[] = { "avdd-pll-uerefe", "hvddio-pex",
|
|
"dvddio-pex", "dvdd-pex-pll",
|
|
"hvdd-pex-pll-e", "vddio-pex-ctl" };
|
|
|
|
static char *t186_rail_names[] = {"vddio-pexctl-aud"};
|
|
|
|
static const struct tegra_pcie_soc_data tegra186_pcie_data = {
|
|
.num_ports = 3,
|
|
.pcie_regulator_names = t186_rail_names,
|
|
.num_pcie_regulators =
|
|
sizeof(t186_rail_names) / sizeof(t186_rail_names[0]),
|
|
.dvfs_afi = true,
|
|
.update_clamp_threshold = true,
|
|
.dvfs_tbl = {
|
|
{{0, 0}, {0, 0} },
|
|
{{102000000, 480000000}, {102000000, 480000000} },
|
|
{{102000000, 480000000}, {204000000, 480000000} },
|
|
{{102000000, 480000000}, {204000000, 480000000} },
|
|
{{204000000, 480000000}, {408000000, 480000000} },
|
|
{{204000000, 480000000}, {408000000, 640000000} } },
|
|
};
|
|
|
|
static const struct tegra_pcie_soc_data tegra210b01_pcie_data = {
|
|
.num_ports = 2,
|
|
.pcie_regulator_names = t210_rail_names,
|
|
.num_pcie_regulators =
|
|
sizeof(t210_rail_names) / sizeof(t210_rail_names[0]),
|
|
.program_uphy = true,
|
|
.program_clkreq_as_bi_dir = true,
|
|
.enable_wrap = true,
|
|
.perf_war = true,
|
|
.updateFC_timer_expire_war = true,
|
|
.l1ss_rp_wakeup_war = true,
|
|
.link_speed_war = true,
|
|
.dvfs_mselect = true,
|
|
.dvfs_tbl = {
|
|
{{204000000, 102000000}, {408000000, 528000000} } },
|
|
};
|
|
|
|
static const struct tegra_pcie_soc_data tegra210_pcie_data = {
|
|
.num_ports = 2,
|
|
.pcie_regulator_names = t210_rail_names,
|
|
.num_pcie_regulators =
|
|
sizeof(t210_rail_names) / sizeof(t210_rail_names[0]),
|
|
.config_pex_io_dpd = true,
|
|
.program_uphy = true,
|
|
.program_clkreq_as_bi_dir = true,
|
|
.enable_wrap = true,
|
|
.mbist_war = true,
|
|
.perf_war = true,
|
|
.updateFC_timer_expire_war = true,
|
|
.l1ss_rp_wakeup_war = true,
|
|
.link_speed_war = true,
|
|
.dvfs_mselect = true,
|
|
.update_clamp_threshold = true,
|
|
.dvfs_tbl = {
|
|
{{204000000, 102000000}, {408000000, 528000000} } },
|
|
};
|
|
|
|
static const struct tegra_pcie_soc_data tegra124_pcie_data = {
|
|
.num_ports = 2,
|
|
.pcie_regulator_names = t124_rail_names,
|
|
.num_pcie_regulators =
|
|
sizeof(t124_rail_names) / sizeof(t124_rail_names[0]),
|
|
.RAW_violation_war = true,
|
|
};
|
|
|
|
static struct of_device_id tegra_pcie_of_match[] = {
|
|
{ .compatible = "nvidia,tegra186-pcie", .data = &tegra186_pcie_data },
|
|
{ .compatible = "nvidia,tegra210b01-pcie", .data = &tegra210b01_pcie_data },
|
|
{ .compatible = "nvidia,tegra210-pcie", .data = &tegra210_pcie_data },
|
|
{ .compatible = "nvidia,tegra124-pcie", .data = &tegra124_pcie_data },
|
|
{ }
|
|
};
|
|
MODULE_DEVICE_TABLE(of, tegra_pcie_of_match);
|
|
|
|
static int tegra_pcie_parse_dt(struct tegra_pcie *pcie)
|
|
{
|
|
struct tegra_pcie_soc_data *soc = pcie->soc_data;
|
|
struct device_node *np = pcie->dev->of_node, *port;
|
|
struct of_pci_range_parser parser;
|
|
struct of_pci_range range;
|
|
u32 lanes = 0, mask = 0;
|
|
unsigned int lane = 0;
|
|
struct resource res = {0};
|
|
int err;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
if (of_pci_range_parser_init(&parser, np)) {
|
|
dev_err(pcie->dev, "missing \"ranges\" property\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
for_each_of_pci_range(&parser, &range) {
|
|
err = of_pci_range_to_resource(&range, np, &res);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
switch (res.flags & IORESOURCE_TYPE_BITS) {
|
|
case IORESOURCE_IO:
|
|
/* Track the bus -> CPU I/O mapping offset. */
|
|
pcie->offset.io = res.start - range.pci_addr;
|
|
|
|
memcpy(&pcie->pio, &res, sizeof(res));
|
|
pcie->pio.name = np->full_name;
|
|
|
|
/*
|
|
* The Tegra PCIe host bridge uses this to program the
|
|
* mapping of the I/O space to the physical address,
|
|
* so we override the .start and .end fields here that
|
|
* of_pci_range_to_resource() converted to I/O space.
|
|
* We also set the IORESOURCE_MEM type to clarify that
|
|
* the resource is in the physical memory space.
|
|
*/
|
|
pcie->io.start = range.cpu_addr;
|
|
pcie->io.end = range.cpu_addr + range.size - 1;
|
|
pcie->io.flags = IORESOURCE_MEM;
|
|
pcie->io.name = "I/O";
|
|
|
|
memcpy(&res, &pcie->io, sizeof(res));
|
|
break;
|
|
|
|
case IORESOURCE_MEM:
|
|
/*
|
|
* Track the bus -> CPU memory mapping offset. This
|
|
* assumes that the prefetchable and non-prefetchable
|
|
* regions will be the last of type IORESOURCE_MEM in
|
|
* the ranges property.
|
|
* */
|
|
pcie->offset.mem = res.start - range.pci_addr;
|
|
|
|
if (res.flags & IORESOURCE_PREFETCH) {
|
|
memcpy(&pcie->prefetch, &res, sizeof(res));
|
|
pcie->prefetch.name = "prefetchable";
|
|
} else {
|
|
memcpy(&pcie->mem, &res, sizeof(res));
|
|
pcie->mem.name = "non-prefetchable";
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
|
|
err = of_pci_parse_bus_range(np, &pcie->busn);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev, "failed to parse ranges property: %d\n",
|
|
err);
|
|
pcie->busn.name = np->name;
|
|
pcie->busn.start = 0;
|
|
pcie->busn.end = 0xff;
|
|
pcie->busn.flags = IORESOURCE_BUS;
|
|
}
|
|
|
|
/* parse root ports */
|
|
for_each_child_of_node(np, port) {
|
|
struct tegra_pcie_port *rp;
|
|
unsigned int index;
|
|
u32 value;
|
|
|
|
#ifdef CONFIG_THERMAL
|
|
if (!strncmp(port->name, "pcie-cool-dev",
|
|
sizeof("pcie-cool-dev")))
|
|
pcie->is_cooling_dev = true;
|
|
#endif
|
|
|
|
if (strncmp(port->type, "pci", sizeof("pci")))
|
|
continue;
|
|
|
|
err = of_pci_get_devfn(port);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev, "failed to parse address: %d\n",
|
|
err);
|
|
goto err_node_put;
|
|
}
|
|
|
|
index = PCI_SLOT(err);
|
|
if (index < 1 || index > soc->num_ports) {
|
|
dev_err(pcie->dev, "invalid port number: %d\n", index);
|
|
err = -EINVAL;
|
|
goto err_node_put;
|
|
}
|
|
|
|
index--;
|
|
|
|
err = of_property_read_u32(port, "nvidia,num-lanes", &value);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev, "failed to parse # of lanes: %d\n",
|
|
err);
|
|
goto err_node_put;
|
|
}
|
|
|
|
if (value > 16) {
|
|
dev_err(pcie->dev, "invalid # of lanes: %u\n", value);
|
|
err = -EINVAL;
|
|
goto err_node_put;
|
|
}
|
|
lanes |= value << (index << 3);
|
|
|
|
if (!of_device_is_available(port)) {
|
|
continue;
|
|
}
|
|
|
|
mask |= ((1 << value) - 1) << lane;
|
|
lane += value;
|
|
|
|
rp = devm_kzalloc(pcie->dev, sizeof(*rp), GFP_KERNEL);
|
|
if (!rp) {
|
|
err = -ENOMEM;
|
|
goto err_node_put;
|
|
}
|
|
|
|
err = of_address_to_resource(port, 0, &rp->regs);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev, "failed to parse address: %d\n", err);
|
|
goto err_node_put;
|
|
}
|
|
|
|
rp->gpio_presence_detection =
|
|
of_get_named_gpio(port,
|
|
"nvidia,presence-detection-gpio", 0);
|
|
|
|
INIT_LIST_HEAD(&rp->list);
|
|
rp->index = index;
|
|
rp->lanes = rp->num_lanes = value;
|
|
rp->pcie = pcie;
|
|
rp->np = port;
|
|
rp->base = devm_ioremap_resource(pcie->dev, &rp->regs);
|
|
if (!(rp->base))
|
|
return -EADDRNOTAVAIL;
|
|
rp->disable_clock_request = of_property_read_bool(port,
|
|
"nvidia,disable-clock-request");
|
|
|
|
rp->rst_gpio = of_get_named_gpio(port, "nvidia,rst-gpio", 0);
|
|
if (gpio_is_valid(rp->rst_gpio)) {
|
|
err = devm_gpio_request(pcie->dev,
|
|
rp->rst_gpio, "pex_rst_gpio");
|
|
if (err < 0) {
|
|
dev_err(pcie->dev,
|
|
"%s: pex_rst_gpio request failed %d\n",
|
|
__func__, err);
|
|
return err;
|
|
}
|
|
err = gpio_direction_output(rp->rst_gpio, 0);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev,
|
|
"%s: pex_rst_gpio direction_output failed %d\n",
|
|
__func__, err);
|
|
return err;
|
|
}
|
|
}
|
|
|
|
rp->has_mxm_port = of_property_read_bool(port,
|
|
"nvidia,has-mxm-port");
|
|
if (rp->has_mxm_port) {
|
|
rp->pwr_gd_gpio = of_get_named_gpio(port,
|
|
"nvidia,pwr-gd-gpio", 0);
|
|
if (gpio_is_valid(rp->pwr_gd_gpio)) {
|
|
err = devm_gpio_request(pcie->dev,
|
|
rp->pwr_gd_gpio,
|
|
"pwr_gd_gpio");
|
|
if (err < 0) {
|
|
dev_err(pcie->dev,
|
|
"%s: pwr_gd_gpio request failed %d\n",
|
|
__func__, err);
|
|
return err;
|
|
}
|
|
err = gpio_direction_input(rp->pwr_gd_gpio);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev,
|
|
"%s: pwr_gd_gpio direction_input failed %d\n",
|
|
__func__, err);
|
|
}
|
|
}
|
|
}
|
|
|
|
rp->n_gpios = of_gpio_named_count(port, "nvidia,plat-gpios");
|
|
if (rp->n_gpios > 0) {
|
|
int count, gpio;
|
|
enum of_gpio_flags flags;
|
|
unsigned long f;
|
|
|
|
rp->gpios = devm_kzalloc(pcie->dev,
|
|
rp->n_gpios * sizeof(int),
|
|
GFP_KERNEL);
|
|
if (!rp->gpios)
|
|
return -ENOMEM;
|
|
|
|
for (count = 0; count < rp->n_gpios; ++count) {
|
|
gpio = of_get_named_gpio_flags(port,
|
|
"nvidia,plat-gpios",
|
|
count, &flags);
|
|
if (!gpio_is_valid(gpio))
|
|
return gpio;
|
|
|
|
f = (flags & OF_GPIO_ACTIVE_LOW) ?
|
|
(GPIOF_OUT_INIT_LOW | GPIOF_ACTIVE_LOW) :
|
|
GPIOF_OUT_INIT_HIGH;
|
|
|
|
err = devm_gpio_request_one(pcie->dev, gpio, f,
|
|
NULL);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev, "gpio %d request failed\n",
|
|
gpio);
|
|
return err;
|
|
}
|
|
rp->gpios[count] = gpio;
|
|
}
|
|
}
|
|
|
|
list_add_tail(&rp->list, &pcie->ports);
|
|
}
|
|
err = tegra_pcie_get_xbar_config(pcie, lanes, &pcie->xbar_config);
|
|
if (err < 0) {
|
|
dev_err(pcie->dev, "invalid lane configuration\n");
|
|
return err;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_node_put:
|
|
of_node_put(port);
|
|
return err;
|
|
}
|
|
|
|
static int list_devices(struct seq_file *s, void *data)
|
|
{
|
|
struct pci_dev *pdev = NULL;
|
|
u16 vendor, device, devclass, speed;
|
|
bool pass = false;
|
|
int ret = 0;
|
|
|
|
for_each_pci_dev(pdev) {
|
|
pass = true;
|
|
ret = pci_read_config_word(pdev, PCI_VENDOR_ID, &vendor);
|
|
if (ret) {
|
|
pass = false;
|
|
break;
|
|
}
|
|
ret = pci_read_config_word(pdev, PCI_DEVICE_ID, &device);
|
|
if (ret) {
|
|
pass = false;
|
|
break;
|
|
}
|
|
ret = pci_read_config_word(pdev, PCI_CLASS_DEVICE, &devclass);
|
|
if (ret) {
|
|
pass = false;
|
|
break;
|
|
}
|
|
pcie_capability_read_word(pdev, PCI_EXP_LNKSTA, &speed);
|
|
|
|
seq_printf(s, "%s Vendor:%04x Device id:%04x ",
|
|
kobject_name(&pdev->dev.kobj), vendor,
|
|
device);
|
|
seq_printf(s, "Class:%04x Speed:%s Driver:%s(%s)\n", devclass,
|
|
((speed & PCI_EXP_LNKSTA_CLS_5_0GB) ==
|
|
PCI_EXP_LNKSTA_CLS_5_0GB) ?
|
|
"Gen2" : "Gen1",
|
|
(pdev->driver) ? "enabled" : "disabled",
|
|
(pdev->driver) ? pdev->driver->name : NULL);
|
|
}
|
|
if (!pass)
|
|
seq_printf(s, "Couldn't read devices\n");
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int apply_link_speed(struct seq_file *s, void *data)
|
|
{
|
|
struct tegra_pcie *pcie = (struct tegra_pcie *)(s->private);
|
|
|
|
seq_puts(s, "Changing link speed to Gen2\n");
|
|
tegra_pcie_link_speed(pcie);
|
|
seq_printf(s, "Done\n");
|
|
return 0;
|
|
|
|
}
|
|
|
|
static int check_d3hot(struct seq_file *s, void *data)
|
|
{
|
|
u16 val;
|
|
struct pci_dev *pdev = NULL;
|
|
|
|
/* Force all the devices (including RPs) in d3 hot state */
|
|
for_each_pci_dev(pdev) {
|
|
if (pci_pcie_type(pdev) == PCI_EXP_TYPE_ROOT_PORT ||
|
|
pci_pcie_type(pdev) == PCI_EXP_TYPE_DOWNSTREAM)
|
|
continue;
|
|
/* First, keep Downstream component in D3_Hot */
|
|
pci_read_config_word(pdev, pdev->pm_cap + PCI_PM_CTRL,
|
|
&val);
|
|
if ((val & PCI_PM_CTRL_STATE_MASK) == PCI_D3hot)
|
|
seq_printf(s, "device[%x:%x] is already in D3_hot]\n",
|
|
pdev->vendor, pdev->device);
|
|
val &= ~PCI_PM_CTRL_STATE_MASK;
|
|
val |= PCI_D3hot;
|
|
pci_write_config_word(pdev, pdev->pm_cap + PCI_PM_CTRL,
|
|
val);
|
|
/* Keep corresponding upstream component in D3_Hot */
|
|
pci_read_config_word(pdev->bus->self,
|
|
pdev->bus->self->pm_cap + PCI_PM_CTRL, &val);
|
|
val &= ~PCI_PM_CTRL_STATE_MASK;
|
|
val |= PCI_D3hot;
|
|
pci_write_config_word(pdev->bus->self,
|
|
pdev->bus->self->pm_cap + PCI_PM_CTRL, val);
|
|
mdelay(100);
|
|
/* check if they have changed their state */
|
|
pci_read_config_word(pdev, pdev->pm_cap + PCI_PM_CTRL,
|
|
&val);
|
|
if ((val & PCI_PM_CTRL_STATE_MASK) == PCI_D3hot)
|
|
seq_printf(s, "device[%x:%x] transitioned to D3_hot]\n",
|
|
pdev->vendor, pdev->device);
|
|
else
|
|
seq_printf(s, "device[%x:%x] couldn't transition to D3_hot]\n",
|
|
pdev->vendor, pdev->device);
|
|
pci_read_config_word(pdev->bus->self,
|
|
pdev->bus->self->pm_cap + PCI_PM_CTRL, &val);
|
|
if ((val & PCI_PM_CTRL_STATE_MASK) == PCI_D3hot)
|
|
seq_printf(s, "device[%x:%x] transitioned to D3_hot]\n",
|
|
pdev->bus->self->vendor,
|
|
pdev->bus->self->device);
|
|
else
|
|
seq_printf(s, "device[%x:%x] couldn't transition to D3_hot]\n",
|
|
pdev->bus->self->vendor,
|
|
pdev->bus->self->device);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int dump_config_space(struct seq_file *s, void *data)
|
|
{
|
|
u8 val;
|
|
int row, col;
|
|
struct pci_dev *pdev = NULL;
|
|
|
|
for_each_pci_dev(pdev) {
|
|
int row_cnt = pci_is_pcie(pdev) ?
|
|
PCI_EXT_CFG_SPACE_SIZE : PCI_CFG_SPACE_SIZE;
|
|
seq_printf(s, "%s\n", kobject_name(&pdev->dev.kobj));
|
|
seq_printf(s, "%s\n", "------------");
|
|
|
|
for (row = 0; row < (row_cnt / 16); row++) {
|
|
seq_printf(s, "%02x: ", (row * 16));
|
|
for (col = 0; col < 16; col++) {
|
|
pci_read_config_byte(pdev, ((row * 16) + col),
|
|
&val);
|
|
seq_printf(s, "%02x ", val);
|
|
}
|
|
seq_printf(s, "\n");
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int dump_afi_space(struct seq_file *s, void *data)
|
|
{
|
|
u32 val, offset;
|
|
struct tegra_pcie_port *port = NULL;
|
|
struct tegra_pcie *pcie = (struct tegra_pcie *)(s->private);
|
|
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
seq_puts(s, "Offset: Values\n");
|
|
for (offset = 0; offset < 0x200; offset += 0x10) {
|
|
val = afi_readl(port->pcie, offset);
|
|
seq_printf(s, "%6x: %8x %8x %8x %8x\n", offset,
|
|
afi_readl(port->pcie, offset),
|
|
afi_readl(port->pcie, offset + 4),
|
|
afi_readl(port->pcie, offset + 8),
|
|
afi_readl(port->pcie, offset + 12));
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int config_read(struct seq_file *s, void *data)
|
|
{
|
|
struct pci_dev *pdev = NULL;
|
|
|
|
pdev = pci_get_bus_and_slot((bdf >> 8), (bdf & 0xFF));
|
|
if (!pdev) {
|
|
seq_printf(s, "%02d:%02d.%02d : Doesn't exist\n",
|
|
(bdf >> 8), PCI_SLOT(bdf), PCI_FUNC(bdf));
|
|
seq_printf(s,
|
|
"Enter (bus<<8 | dev<<3 | func) value to bdf file\n");
|
|
goto end;
|
|
}
|
|
if (config_offset >= PCI_EXT_CFG_SPACE_SIZE) {
|
|
seq_printf(s, "Config offset exceeds max (i.e %d) value\n",
|
|
PCI_EXT_CFG_SPACE_SIZE);
|
|
}
|
|
if (!(config_offset & 0x3)) {
|
|
u32 val;
|
|
/* read 32 */
|
|
pci_read_config_dword(pdev, config_offset, &val);
|
|
seq_printf(s, "%08x\n", val);
|
|
config_val = val;
|
|
} else if (!(config_offset & 0x1)) {
|
|
u16 val;
|
|
/* read 16 */
|
|
pci_read_config_word(pdev, config_offset, &val);
|
|
seq_printf(s, "%04x\n", val);
|
|
config_val = val;
|
|
} else {
|
|
u8 val;
|
|
/* read 8 */
|
|
pci_read_config_byte(pdev, config_offset, &val);
|
|
seq_printf(s, "%02x\n", val);
|
|
config_val = val;
|
|
}
|
|
|
|
end:
|
|
return 0;
|
|
}
|
|
|
|
static int config_write(struct seq_file *s, void *data)
|
|
{
|
|
struct pci_dev *pdev = NULL;
|
|
|
|
pdev = pci_get_bus_and_slot((bdf >> 8), (bdf & 0xFF));
|
|
if (!pdev) {
|
|
seq_printf(s, "%02d:%02d.%02d : Doesn't exist\n",
|
|
(bdf >> 8), PCI_SLOT(bdf), PCI_FUNC(bdf));
|
|
seq_printf(s,
|
|
"Enter (bus<<8 | dev<<3 | func) value to bdf file\n");
|
|
goto end;
|
|
}
|
|
if (config_offset >= PCI_EXT_CFG_SPACE_SIZE) {
|
|
seq_printf(s, "Config offset exceeds max (i.e %d) value\n",
|
|
PCI_EXT_CFG_SPACE_SIZE);
|
|
}
|
|
if (!(config_offset & 0x3)) {
|
|
/* write 32 */
|
|
pci_write_config_dword(pdev, config_offset, config_val);
|
|
} else if (!(config_offset & 0x1)) {
|
|
/* write 16 */
|
|
pci_write_config_word(pdev, config_offset,
|
|
(u16)(config_val & 0xFFFF));
|
|
} else {
|
|
/* write 8 */
|
|
pci_write_config_byte(pdev, config_offset,
|
|
(u8)(config_val & 0xFF));
|
|
}
|
|
|
|
end:
|
|
return 0;
|
|
}
|
|
|
|
static int power_down(struct seq_file *s, void *data)
|
|
{
|
|
struct tegra_pcie_port *port = NULL;
|
|
struct tegra_pcie *pcie = (struct tegra_pcie *)(s->private);
|
|
u32 val;
|
|
bool pass = false;
|
|
|
|
val = afi_readl(pcie, AFI_PCIE_PME);
|
|
val |= AFI_PCIE_PME_TURN_OFF;
|
|
afi_writel(pcie, val, AFI_PCIE_PME);
|
|
do {
|
|
val = afi_readl(pcie, AFI_PCIE_PME);
|
|
} while(!(val & AFI_PCIE_PME_ACK));
|
|
|
|
mdelay(1000);
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
val = rp_readl(port, NV_PCIE2_RP_LTSSM_DBGREG);
|
|
if (val & PCIE2_RP_LTSSM_DBGREG_LINKFSM16) {
|
|
pass = true;
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
out:
|
|
if (pass)
|
|
seq_printf(s, "[pass: pcie_power_down]\n");
|
|
else
|
|
seq_printf(s, "[fail: pcie_power_down]\n");
|
|
pr_info("PCIE power_down test END..\n");
|
|
return 0;
|
|
}
|
|
|
|
static int loopback(struct seq_file *s, void *data)
|
|
{
|
|
struct tegra_pcie_port *port = (struct tegra_pcie_port *)(s->private);
|
|
unsigned int new, i, val;
|
|
|
|
new = rp_readl(port, RP_LINK_CONTROL_STATUS);
|
|
|
|
if (!(new & RP_LINK_CONTROL_STATUS_DL_LINK_ACTIVE)) {
|
|
pr_info("PCIE port %d not active\n", port->index);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* trigger trace ram on loopback states */
|
|
val = LTSSM_TRACE_CONTROL_CLEAR_STORE_EN |
|
|
LTSSM_TRACE_CONTROL_TRIG_ON_EVENT |
|
|
(0x08 << LTSSM_TRACE_CONTROL_TRIG_LTSSM_MAJOR_OFFSET) |
|
|
(0x00 << LTSSM_TRACE_CONTROL_TRIG_PTX_LTSSM_MINOR_OFFSET) |
|
|
(0x00 << LTSSM_TRACE_CONTROL_TRIG_PRX_LTSSM_MAJOR_OFFSET);
|
|
rp_writel(port, val, NV_PCIE2_RP_LTSSM_TRACE_CONTROL);
|
|
|
|
/* clear trace ram */
|
|
val = rp_readl(port, NV_PCIE2_RP_LTSSM_TRACE_CONTROL);
|
|
val |= LTSSM_TRACE_CONTROL_CLEAR_RAM;
|
|
rp_writel(port, val, NV_PCIE2_RP_LTSSM_TRACE_CONTROL);
|
|
val &= ~LTSSM_TRACE_CONTROL_CLEAR_RAM;
|
|
rp_writel(port, val, NV_PCIE2_RP_LTSSM_TRACE_CONTROL);
|
|
|
|
/* reset and clear status */
|
|
port->loopback_stat = 0;
|
|
|
|
new = rp_readl(port, RP_VEND_XP);
|
|
new &= ~RP_VEND_XP_PRBS_EN;
|
|
rp_writel(port, new, RP_VEND_XP);
|
|
|
|
new = rp_readl(port, NV_PCIE2_RP_XP_CTL_1);
|
|
new &= ~PCIE2_RP_XP_CTL_1_OLD_IOBIST_EN_BIT25;
|
|
rp_writel(port, new, NV_PCIE2_RP_XP_CTL_1);
|
|
|
|
rp_writel(port, 0x10000001, NV_PCIE2_RP_VEND_XP_BIST);
|
|
rp_writel(port, 0, NV_PCIE2_RP_PRBS);
|
|
|
|
mdelay(1);
|
|
|
|
rp_writel(port, 0x90820001, NV_PCIE2_RP_VEND_XP_BIST);
|
|
new = rp_readl(port, NV_PCIE2_RP_VEND_XP_BIST);
|
|
|
|
new = rp_readl(port, NV_PCIE2_RP_XP_CTL_1);
|
|
new |= PCIE2_RP_XP_CTL_1_OLD_IOBIST_EN_BIT25;
|
|
rp_writel(port, new, NV_PCIE2_RP_XP_CTL_1);
|
|
|
|
new = rp_readl(port, RP_VEND_XP);
|
|
new |= RP_VEND_XP_PRBS_EN;
|
|
rp_writel(port, new, RP_VEND_XP);
|
|
|
|
mdelay(1000);
|
|
|
|
new = rp_readl(port, RP_VEND_XP);
|
|
port->loopback_stat = (new & RP_VEND_XP_PRBS_STAT) >> 2;
|
|
pr_info("--- loopback status ---\n");
|
|
for (i = 0; i < port->lanes; ++i)
|
|
pr_info("@lane %d: %s\n", i,
|
|
(port->loopback_stat & 0x01 << i) ? "pass" : "fail");
|
|
|
|
new = rp_readl(port, NV_PCIE2_RP_PRBS);
|
|
pr_info("--- PRBS pattern locked ---\n");
|
|
for (i = 0; i < port->lanes; ++i)
|
|
pr_info("@lane %d: %s\n", i,
|
|
(new >> 16 & 0x01 << i) ? "Y" : "N");
|
|
pr_info("--- err overflow bits ---\n");
|
|
for (i = 0; i < port->lanes; ++i)
|
|
pr_info("@lane %d: %s\n", i,
|
|
((new & 0xffff) & 0x01 << i) ? "Y" : "N");
|
|
|
|
new = rp_readl(port, NV_PCIE2_RP_XP_CTL_1);
|
|
new &= ~PCIE2_RP_XP_CTL_1_OLD_IOBIST_EN_BIT25;
|
|
rp_writel(port, new, NV_PCIE2_RP_XP_CTL_1);
|
|
|
|
pr_info("--- err counts ---\n");
|
|
for (i = 0; i < port->lanes; ++i) {
|
|
rp_writel(port, i, NV_PCIE2_RP_LANE_PRBS_ERR_COUNT);
|
|
new = rp_readl(port, NV_PCIE2_RP_LANE_PRBS_ERR_COUNT);
|
|
pr_info("@lane %d: %u\n", i, new >> 16);
|
|
}
|
|
|
|
rp_writel(port, 0x90000001, NV_PCIE2_RP_VEND_XP_BIST);
|
|
|
|
new = rp_readl(port, RP_VEND_XP);
|
|
new &= ~RP_VEND_XP_PRBS_EN;
|
|
rp_writel(port, new, RP_VEND_XP);
|
|
|
|
mdelay(1);
|
|
|
|
rp_writel(port, 0x92000001, NV_PCIE2_RP_VEND_XP_BIST);
|
|
rp_writel(port, 0x90000001, NV_PCIE2_RP_VEND_XP_BIST);
|
|
pr_info("pcie loopback test is done\n");
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int apply_lane_width(struct seq_file *s, void *data)
|
|
{
|
|
unsigned int new;
|
|
struct tegra_pcie_port *port = (struct tegra_pcie_port *)(s->private);
|
|
|
|
if (port->lanes > 0x10) {
|
|
seq_printf(s, "link width cannot be grater than 16\n");
|
|
new = rp_readl(port, RP_LINK_CONTROL_STATUS);
|
|
port->lanes = (new &
|
|
RP_LINK_CONTROL_STATUS_NEG_LINK_WIDTH) >> 20;
|
|
return 0;
|
|
}
|
|
new = rp_readl(port, NV_PCIE2_RP_VEND_XP1);
|
|
new &= ~NV_PCIE2_RP_VEND_XP1_RNCTRL_MAXWIDTH_MASK;
|
|
new |= port->lanes | NV_PCIE2_RP_VEND_XP1_RNCTRL_EN;
|
|
rp_writel(port, new, NV_PCIE2_RP_VEND_XP1);
|
|
mdelay(1);
|
|
|
|
new = rp_readl(port, RP_LINK_CONTROL_STATUS);
|
|
new = (new & RP_LINK_CONTROL_STATUS_NEG_LINK_WIDTH) >> 20;
|
|
if (new != port->lanes)
|
|
seq_printf(s, "can't set link width %u, falling back to %u\n",
|
|
port->lanes, new);
|
|
else
|
|
seq_printf(s, "lane width %d applied\n", new);
|
|
port->lanes = new;
|
|
return 0;
|
|
}
|
|
|
|
static int aspm_state_cnt(struct seq_file *s, void *data)
|
|
{
|
|
u32 val, cs;
|
|
struct tegra_pcie_port *port = (struct tegra_pcie_port *)(s->private);
|
|
|
|
cs = rp_readl(port, RP_LINK_CONTROL_STATUS);
|
|
/* check if L0s is enabled on this port */
|
|
if (cs & RP_LINK_CONTROL_STATUS_L0s_ENABLED) {
|
|
val = rp_readl(port, NV_PCIE2_RP_PRIV_XP_TX_L0S_ENTRY_COUNT);
|
|
seq_printf(s, "Tx L0s entry count : %u\n", val);
|
|
} else
|
|
seq_printf(s, "Tx L0s entry count : %s\n", "disabled");
|
|
|
|
val = rp_readl(port, NV_PCIE2_RP_PRIV_XP_RX_L0S_ENTRY_COUNT);
|
|
seq_printf(s, "Rx L0s entry count : %u\n", val);
|
|
|
|
/* check if L1 is enabled on this port */
|
|
if (cs & RP_LINK_CONTROL_STATUS_L1_ENABLED) {
|
|
val = rp_readl(port, NV_PCIE2_RP_PRIV_XP_TX_L1_ENTRY_COUNT);
|
|
seq_printf(s, "Link L1 entry count : %u\n", val);
|
|
} else
|
|
seq_printf(s, "Link L1 entry count : %s\n", "disabled");
|
|
|
|
cs = rp_readl(port, PCIE2_RP_L1_PM_SS_CONTROL);
|
|
/* RESETting the count value is not possible by any means
|
|
because of HW Bug : 200034278 */
|
|
/* check if L1.1 is enabled */
|
|
if (cs & PCIE2_RP_L1_PM_SS_CONTROL_ASPM_L11_ENABLE) {
|
|
val = rp_readl(port, NV_PCIE2_RP_L1_1_ENTRY_COUNT);
|
|
val |= PCIE2_RP_L1_1_ENTRY_COUNT_RESET;
|
|
rp_writel(port, val, NV_PCIE2_RP_L1_1_ENTRY_COUNT);
|
|
seq_printf(s, "Link L1.1 entry count : %u\n", (val & 0xFFFF));
|
|
} else
|
|
seq_printf(s, "Link L1.1 entry count : %s\n", "disabled");
|
|
/* check if L1.2 is enabled */
|
|
if (cs & PCIE2_RP_L1_PM_SS_CONTROL_ASPM_L12_ENABLE) {
|
|
val = rp_readl(port, NV_PCIE2_RP_L1_2_ENTRY_COUNT);
|
|
val |= PCIE2_RP_L1_2_ENTRY_COUNT_RESET;
|
|
rp_writel(port, val, NV_PCIE2_RP_L1_2_ENTRY_COUNT);
|
|
seq_printf(s, "Link L1.2 entry count : %u\n", (val & 0xFFFF));
|
|
} else
|
|
seq_printf(s, "Link L1.2 entry count : %s\n", "disabled");
|
|
|
|
return 0;
|
|
}
|
|
|
|
static char *aspm_states[] = {
|
|
"Tx-L0s",
|
|
"Rx-L0s",
|
|
"L1",
|
|
"IDLE ((Tx-L0s && Rx-L0s) + L1)"
|
|
};
|
|
|
|
static int list_aspm_states(struct seq_file *s, void *data)
|
|
{
|
|
u32 i = 0;
|
|
seq_printf(s, "----------------------------------------------------\n");
|
|
seq_printf(s, "Note: Duration of link's residency is calcualated\n");
|
|
seq_printf(s, " only for one of the ASPM states at a time\n");
|
|
seq_printf(s, "----------------------------------------------------\n");
|
|
seq_printf(s, "write(echo) number from below table corresponding to\n");
|
|
seq_printf(s, "one of the ASPM states for which link duration needs\n");
|
|
seq_printf(s, "to be calculated to 'config_aspm_state'\n");
|
|
seq_printf(s, "-----------------\n");
|
|
for (i = 0; i < ARRAY_SIZE(aspm_states); i++)
|
|
seq_printf(s, "%d : %s\n", i, aspm_states[i]);
|
|
seq_printf(s, "-----------------\n");
|
|
return 0;
|
|
}
|
|
|
|
static int apply_aspm_state(struct seq_file *s, void *data)
|
|
{
|
|
u32 val;
|
|
struct tegra_pcie_port *port = (struct tegra_pcie_port *)(s->private);
|
|
|
|
if (config_aspm_state >= ARRAY_SIZE(aspm_states)) {
|
|
seq_printf(s, "Invalid ASPM state : %u\n", config_aspm_state);
|
|
list_aspm_states(s, data);
|
|
} else {
|
|
val = rp_readl(port, NV_PCIE2_RP_PRIV_XP_CONFIG);
|
|
val &= ~NV_PCIE2_RP_PRIV_XP_CONFIG_LOW_PWR_DURATION_MASK;
|
|
val |= config_aspm_state;
|
|
rp_writel(port, val, NV_PCIE2_RP_PRIV_XP_CONFIG);
|
|
seq_printf(s, "Configured for ASPM-%s state...\n",
|
|
aspm_states[config_aspm_state]);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int get_aspm_duration(struct seq_file *s, void *data)
|
|
{
|
|
u32 val;
|
|
struct tegra_pcie_port *port = (struct tegra_pcie_port *)(s->private);
|
|
|
|
val = rp_readl(port, NV_PCIE2_RP_PRIV_XP_DURATION_IN_LOW_PWR_100NS);
|
|
/* 52.08 = 1000 / 19.2MHz is rounded to 52 */
|
|
seq_printf(s, "ASPM-%s duration = %d ns\n",
|
|
aspm_states[config_aspm_state], (u32)((val * 100)/52));
|
|
return 0;
|
|
}
|
|
|
|
static int secondary_bus_reset(struct seq_file *s, void *data)
|
|
{
|
|
u32 val;
|
|
struct tegra_pcie_port *port = (struct tegra_pcie_port *)(s->private);
|
|
|
|
val = rp_readl(port, NV_PCIE2_RP_INTR_BCR);
|
|
val |= NV_PCIE2_RP_INTR_BCR_SB_RESET;
|
|
rp_writel(port, val, NV_PCIE2_RP_INTR_BCR);
|
|
udelay(10);
|
|
val = rp_readl(port, NV_PCIE2_RP_INTR_BCR);
|
|
val &= ~NV_PCIE2_RP_INTR_BCR_SB_RESET;
|
|
rp_writel(port, val, NV_PCIE2_RP_INTR_BCR);
|
|
|
|
seq_printf(s, "Secondary Bus Reset applied successfully...\n");
|
|
return 0;
|
|
}
|
|
|
|
static void reset_l1ss_counter(struct tegra_pcie_port *port, u32 val,
|
|
unsigned long offset)
|
|
{
|
|
int c = 0;
|
|
|
|
if ((val & 0xFFFF) == 0xFFFF) {
|
|
pr_info(" Trying reset L1ss entry count to 0\n");
|
|
while (val) {
|
|
if (c++ > 50) {
|
|
pr_info("Timeout: reset did not happen!\n");
|
|
break;
|
|
}
|
|
val |= PCIE2_RP_L1_1_ENTRY_COUNT_RESET;
|
|
rp_writel(port, val, offset);
|
|
mdelay(1);
|
|
val = rp_readl(port, offset);
|
|
}
|
|
if (!val)
|
|
pr_info("L1ss entry count reset to 0\n");
|
|
}
|
|
}
|
|
static int aspm_l11(struct seq_file *s, void *data)
|
|
{
|
|
struct pci_dev *pdev = NULL;
|
|
u32 val = 0, pos = 0;
|
|
struct tegra_pcie_port *port = NULL;
|
|
struct tegra_pcie *pcie = (struct tegra_pcie *)(s->private);
|
|
|
|
pr_info("\nPCIE aspm l1.1 test START..\n");
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
/* reset RP L1.1 counter */
|
|
val = rp_readl(port, NV_PCIE2_RP_L1_1_ENTRY_COUNT);
|
|
val |= PCIE2_RP_L1_1_ENTRY_COUNT_RESET;
|
|
rp_writel(port, val, NV_PCIE2_RP_L1_1_ENTRY_COUNT);
|
|
|
|
val = rp_readl(port, NV_PCIE2_RP_L1_1_ENTRY_COUNT);
|
|
pr_info("L1.1 Entry count before %x\n", val);
|
|
reset_l1ss_counter(port, val, NV_PCIE2_RP_L1_1_ENTRY_COUNT);
|
|
}
|
|
/* disable automatic l1ss exit by gpu */
|
|
for_each_pci_dev(pdev)
|
|
if (pci_pcie_type(pdev) != PCI_EXP_TYPE_ROOT_PORT) {
|
|
pci_write_config_dword(pdev, 0x658, 0);
|
|
pci_write_config_dword(pdev, 0x150, 0xE0000015);
|
|
}
|
|
for_each_pci_dev(pdev) {
|
|
u16 aspm;
|
|
pcie_capability_read_word(pdev, PCI_EXP_LNKCTL, &aspm);
|
|
aspm |= PCI_EXP_LNKCTL_ASPM_L1;
|
|
pcie_capability_write_word(pdev, PCI_EXP_LNKCTL, aspm);
|
|
pos = pci_find_ext_capability(pdev, PCI_EXT_CAP_ID_L1SS);
|
|
pci_read_config_dword(pdev, pos + PCI_L1SS_CTRL1, &val);
|
|
val &= ~PCI_L1SS_CAP_L1PM_MASK;
|
|
val |= PCI_L1SS_CTRL1_ASPM_L11S;
|
|
pci_write_config_dword(pdev, pos + PCI_L1SS_CTRL1, val);
|
|
if (pci_pcie_type(pdev) != PCI_EXP_TYPE_ROOT_PORT)
|
|
break;
|
|
}
|
|
mdelay(2000);
|
|
for_each_pci_dev(pdev) {
|
|
pos = pci_find_ext_capability(pdev, PCI_EXT_CAP_ID_L1SS);
|
|
pci_read_config_dword(pdev, pos + PCI_L1SS_CTRL1, &val);
|
|
if (pci_pcie_type(pdev) != PCI_EXP_TYPE_ROOT_PORT)
|
|
break;
|
|
}
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
val = rp_readl(port, NV_PCIE2_RP_L1_1_ENTRY_COUNT);
|
|
pr_info("L1.1 Entry count after %x\n", val);
|
|
}
|
|
|
|
pr_info("PCIE aspm l1.1 test END..\n");
|
|
return 0;
|
|
}
|
|
|
|
static int aspm_l1ss(struct seq_file *s, void *data)
|
|
{
|
|
struct pci_dev *pdev = NULL;
|
|
u32 val = 0, pos = 0;
|
|
struct tegra_pcie_port *port = NULL;
|
|
struct tegra_pcie *pcie = (struct tegra_pcie *)(s->private);
|
|
|
|
pr_info("\nPCIE aspm l1ss test START..\n");
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
/* reset RP L1.1 L1.2 counters */
|
|
val = rp_readl(port, NV_PCIE2_RP_L1_1_ENTRY_COUNT);
|
|
val |= PCIE2_RP_L1_1_ENTRY_COUNT_RESET;
|
|
rp_writel(port, val, NV_PCIE2_RP_L1_1_ENTRY_COUNT);
|
|
val = rp_readl(port, NV_PCIE2_RP_L1_1_ENTRY_COUNT);
|
|
pr_info("L1.1 Entry count before %x\n", val);
|
|
reset_l1ss_counter(port, val, NV_PCIE2_RP_L1_1_ENTRY_COUNT);
|
|
|
|
val = rp_readl(port, NV_PCIE2_RP_L1_2_ENTRY_COUNT);
|
|
val |= PCIE2_RP_L1_2_ENTRY_COUNT_RESET;
|
|
rp_writel(port, val, NV_PCIE2_RP_L1_2_ENTRY_COUNT);
|
|
val = rp_readl(port, NV_PCIE2_RP_L1_2_ENTRY_COUNT);
|
|
pr_info("L1.2 Entry count before %x\n", val);
|
|
reset_l1ss_counter(port, val, NV_PCIE2_RP_L1_2_ENTRY_COUNT);
|
|
}
|
|
/* disable automatic l1ss exit by gpu */
|
|
for_each_pci_dev(pdev)
|
|
if (pci_pcie_type(pdev) != PCI_EXP_TYPE_ROOT_PORT) {
|
|
pci_write_config_dword(pdev, 0x658, 0);
|
|
pci_write_config_dword(pdev, 0x150, 0xE0000015);
|
|
}
|
|
|
|
for_each_pci_dev(pdev) {
|
|
u16 aspm;
|
|
pcie_capability_read_word(pdev, PCI_EXP_LNKCTL, &aspm);
|
|
aspm |= PCI_EXP_LNKCTL_ASPM_L1;
|
|
pcie_capability_write_word(pdev, PCI_EXP_LNKCTL, aspm);
|
|
pos = pci_find_ext_capability(pdev, PCI_EXT_CAP_ID_L1SS);
|
|
pci_read_config_dword(pdev, pos + PCI_L1SS_CTRL1, &val);
|
|
val &= ~PCI_L1SS_CAP_L1PM_MASK;
|
|
val |= (PCI_L1SS_CTRL1_ASPM_L11S | PCI_L1SS_CTRL1_ASPM_L12S);
|
|
pci_write_config_dword(pdev, pos + PCI_L1SS_CTRL1, val);
|
|
if (pci_pcie_type(pdev) != PCI_EXP_TYPE_ROOT_PORT)
|
|
break;
|
|
}
|
|
mdelay(2000);
|
|
for_each_pci_dev(pdev) {
|
|
pos = pci_find_ext_capability(pdev, PCI_EXT_CAP_ID_L1SS);
|
|
pci_read_config_dword(pdev, pos + PCI_L1SS_CTRL1, &val);
|
|
if (pci_pcie_type(pdev) != PCI_EXP_TYPE_ROOT_PORT)
|
|
break;
|
|
}
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
u32 ltr_val;
|
|
val = rp_readl(port, NV_PCIE2_RP_L1_1_ENTRY_COUNT);
|
|
pr_info("L1.1 Entry count after %x\n", val);
|
|
val = rp_readl(port, NV_PCIE2_RP_L1_2_ENTRY_COUNT);
|
|
pr_info("L1.2 Entry count after %x\n", val);
|
|
|
|
val = rp_readl(port, NV_PCIE2_RP_LTR_REP_VAL);
|
|
pr_info("LTR reproted by EP %x\n", val);
|
|
ltr_val = (val & 0x1FF) * (1 << (5 * ((val & 0x1C00) >> 10)));
|
|
if (ltr_val > (106 * 1000)) {
|
|
pr_info("EP's LTR = %u ns is > RP's threshold = %u ns\n",
|
|
ltr_val, 106 * 1000);
|
|
pr_info("Hence only L1.2 entry allowed\n");
|
|
} else {
|
|
pr_info("EP's LTR = %u ns is < RP's threshold = %u ns\n",
|
|
ltr_val, 106 * 1000);
|
|
pr_info("Hence only L1.1 entry allowed\n");
|
|
}
|
|
}
|
|
|
|
pr_info("PCIE aspm l1ss test END..\n");
|
|
return 0;
|
|
}
|
|
|
|
struct ltssm_major_state {
|
|
const char *name;
|
|
const char *minor[8];
|
|
};
|
|
|
|
struct ltssm_state {
|
|
struct ltssm_major_state major[12];
|
|
};
|
|
|
|
static struct ltssm_state ltssm_state = {
|
|
.major[0] = {"detect", {"quiet", "active", "retry", "wait", "entry"}},
|
|
.major[1] = {"polling", {"active", "config", "idle", NULL, "compliance", "cspeed"}},
|
|
.major[2] = {"config", {"link start", "link accept", "lane accept", "lane wait", "idle", "pwrup", "complete"}},
|
|
.major[3] = {NULL, {NULL}},
|
|
.major[4] = {"l0", {"normal", "l0s entry", "l0s idle", "l0s wait", "l0s fts", "pwrup"}},
|
|
.major[5] = {"l1", {"entry", "waitrx", "idle", "wait", "pwrup", "beacon entry", "beacon exit"}},
|
|
.major[6] = {"l2", {"entry", "waitrx", "transmitwake", "idle"}},
|
|
.major[7] = {"recovery", {"rcvrlock", "rcvrcfg", "speed", "idle", NULL, NULL, NULL, "finish pkt"}},
|
|
.major[8] = {"loopback", {"entry", "active", "idle", "exit", "speed", "pre speed"}},
|
|
.major[9] = {"hotreset", {NULL}},
|
|
.major[10] = {"disabled", {NULL}},
|
|
.major[11] = {"txchar", {NULL}},
|
|
};
|
|
|
|
static const char *ltssm_get_major(unsigned int major)
|
|
{
|
|
const char *state;
|
|
|
|
state = ltssm_state.major[major].name;
|
|
if (!state)
|
|
return "unknown";
|
|
return state;
|
|
}
|
|
|
|
static const char *ltssm_get_minor(unsigned int major, unsigned int minor)
|
|
{
|
|
const char *state;
|
|
|
|
state = ltssm_state.major[major].minor[minor];
|
|
if (!state)
|
|
return "unknown";
|
|
return state;
|
|
}
|
|
|
|
static int dump_ltssm_trace(struct seq_file *s, void *data)
|
|
{
|
|
struct tegra_pcie_port *port = (struct tegra_pcie_port *)(s->private);
|
|
unsigned int val, ridx, widx, entries;
|
|
|
|
seq_printf(s, "LTSSM trace dump:\n");
|
|
val = rp_readl(port, NV_PCIE2_RP_LTSSM_TRACE_STATUS);
|
|
widx = LTSSM_TRACE_STATUS_WRITE_POINTER(val);
|
|
entries = LTSSM_TRACE_STATUS_RAM_FULL(val) ? 32 : widx;
|
|
seq_printf(s, "LTSSM trace dump - %d entries:\n", entries);
|
|
for (ridx = 0; ridx < entries; ridx++) {
|
|
val = LTSSM_TRACE_STATUS_READ_ADDR(ridx);
|
|
rp_writel(port, val, NV_PCIE2_RP_LTSSM_TRACE_STATUS);
|
|
val = rp_readl(port, NV_PCIE2_RP_LTSSM_TRACE_STATUS);
|
|
|
|
seq_printf(s, " [0x%08x] major: %-10s minor_tx: %-15s minor_rx: %s\n", val,
|
|
ltssm_get_major(LTSSM_TRACE_STATUS_MAJOR(val)),
|
|
ltssm_get_minor(LTSSM_TRACE_STATUS_MAJOR(val), LTSSM_TRACE_STATUS_PTX_MINOR(val)),
|
|
ltssm_get_minor(LTSSM_TRACE_STATUS_MAJOR(val), LTSSM_TRACE_STATUS_PRX_MINOR(val)));
|
|
}
|
|
/* clear trace ram */
|
|
val = rp_readl(port, NV_PCIE2_RP_LTSSM_TRACE_CONTROL);
|
|
val |= LTSSM_TRACE_CONTROL_CLEAR_RAM;
|
|
rp_writel(port, val, NV_PCIE2_RP_LTSSM_TRACE_CONTROL);
|
|
val &= ~LTSSM_TRACE_CONTROL_CLEAR_RAM;
|
|
rp_writel(port, val, NV_PCIE2_RP_LTSSM_TRACE_CONTROL);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct dentry *create_tegra_pcie_debufs_file(char *name,
|
|
const struct file_operations *ops,
|
|
struct dentry *parent,
|
|
void *data)
|
|
{
|
|
struct dentry *d;
|
|
|
|
d = debugfs_create_file(name, S_IRUGO, parent, data, ops);
|
|
if (!d)
|
|
debugfs_remove_recursive(parent);
|
|
|
|
return d;
|
|
}
|
|
|
|
#define DEFINE_ENTRY(__name) \
|
|
static int __name ## _open(struct inode *inode, struct file *file) \
|
|
{ \
|
|
return single_open(file, __name, inode->i_private); \
|
|
} \
|
|
static const struct file_operations __name ## _fops = { \
|
|
.open = __name ## _open, \
|
|
.read = seq_read, \
|
|
.llseek = seq_lseek, \
|
|
.release = single_release, \
|
|
};
|
|
|
|
/* common */
|
|
DEFINE_ENTRY(list_devices)
|
|
DEFINE_ENTRY(apply_link_speed)
|
|
DEFINE_ENTRY(check_d3hot)
|
|
DEFINE_ENTRY(dump_config_space)
|
|
DEFINE_ENTRY(dump_afi_space)
|
|
DEFINE_ENTRY(config_read)
|
|
DEFINE_ENTRY(config_write)
|
|
DEFINE_ENTRY(aspm_l11)
|
|
DEFINE_ENTRY(aspm_l1ss)
|
|
DEFINE_ENTRY(power_down)
|
|
|
|
/* Port specific */
|
|
DEFINE_ENTRY(loopback)
|
|
DEFINE_ENTRY(apply_lane_width)
|
|
DEFINE_ENTRY(aspm_state_cnt)
|
|
DEFINE_ENTRY(list_aspm_states)
|
|
DEFINE_ENTRY(apply_aspm_state)
|
|
DEFINE_ENTRY(get_aspm_duration)
|
|
DEFINE_ENTRY(secondary_bus_reset)
|
|
DEFINE_ENTRY(dump_ltssm_trace)
|
|
|
|
static int tegra_pcie_port_debugfs_init(struct tegra_pcie_port *port)
|
|
{
|
|
struct dentry *d;
|
|
char port_name[2] = {0};
|
|
|
|
snprintf(port_name, sizeof(port_name), "%d", port->index);
|
|
port->port_debugfs = debugfs_create_dir(port_name,
|
|
port->pcie->debugfs);
|
|
if (!port->port_debugfs)
|
|
return -ENOMEM;
|
|
|
|
d = debugfs_create_u32("lane_width", S_IWUGO | S_IRUGO,
|
|
port->port_debugfs,
|
|
&(port->lanes));
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = debugfs_create_x32("loopback_status", S_IWUGO | S_IRUGO,
|
|
port->port_debugfs,
|
|
&(port->loopback_stat));
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = debugfs_create_file("loopback", S_IRUGO,
|
|
port->port_debugfs, (void *)port,
|
|
&loopback_fops);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = debugfs_create_file("apply_lane_width", S_IRUGO,
|
|
port->port_debugfs, (void *)port,
|
|
&apply_lane_width_fops);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = debugfs_create_file("aspm_state_cnt", S_IRUGO,
|
|
port->port_debugfs, (void *)port,
|
|
&aspm_state_cnt_fops);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = debugfs_create_u16("config_aspm_state", S_IWUGO | S_IRUGO,
|
|
port->port_debugfs,
|
|
&config_aspm_state);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = debugfs_create_file("apply_aspm_state", S_IRUGO,
|
|
port->port_debugfs, (void *)port,
|
|
&apply_aspm_state_fops);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = debugfs_create_file("list_aspm_states", S_IRUGO,
|
|
port->port_debugfs, (void *)port,
|
|
&list_aspm_states_fops);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = debugfs_create_file("dump_ltssm_trace", S_IRUGO,
|
|
port->port_debugfs, (void *)port,
|
|
&dump_ltssm_trace_fops);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = debugfs_create_file("get_aspm_duration", S_IRUGO,
|
|
port->port_debugfs, (void *)port,
|
|
&get_aspm_duration_fops);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = debugfs_create_file("secondary_bus_reset", S_IRUGO,
|
|
port->port_debugfs, (void *)port,
|
|
&secondary_bus_reset_fops);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
return 0;
|
|
|
|
remove:
|
|
debugfs_remove_recursive(port->port_debugfs);
|
|
port->port_debugfs = NULL;
|
|
return -ENOMEM;
|
|
}
|
|
|
|
static void *tegra_pcie_ports_seq_start(struct seq_file *s, loff_t *pos)
|
|
{
|
|
struct tegra_pcie *pcie = s->private;
|
|
|
|
if (list_empty(&pcie->ports))
|
|
return NULL;
|
|
|
|
seq_printf(s, "Index Status\n");
|
|
|
|
return seq_list_start(&pcie->ports, *pos);
|
|
}
|
|
|
|
static void *tegra_pcie_ports_seq_next(struct seq_file *s, void *v, loff_t *pos)
|
|
{
|
|
struct tegra_pcie *pcie = s->private;
|
|
|
|
return seq_list_next(v, &pcie->ports, pos);
|
|
}
|
|
|
|
static void tegra_pcie_ports_seq_stop(struct seq_file *s, void *v)
|
|
{
|
|
}
|
|
|
|
static int tegra_pcie_ports_seq_show(struct seq_file *s, void *v)
|
|
{
|
|
bool up = false, active = false;
|
|
struct tegra_pcie_port *port;
|
|
unsigned int value;
|
|
|
|
port = list_entry(v, struct tegra_pcie_port, list);
|
|
|
|
if (!port->status)
|
|
return 0;
|
|
|
|
value = readl(port->base + RP_VEND_XP);
|
|
|
|
if (value & RP_VEND_XP_DL_UP)
|
|
up = true;
|
|
|
|
value = readl(port->base + RP_LINK_CONTROL_STATUS);
|
|
|
|
if (value & RP_LINK_CONTROL_STATUS_DL_LINK_ACTIVE)
|
|
active = true;
|
|
|
|
seq_printf(s, "%2u ", port->index);
|
|
|
|
if (up)
|
|
seq_printf(s, "up");
|
|
|
|
if (active) {
|
|
if (up)
|
|
seq_printf(s, ", ");
|
|
|
|
seq_printf(s, "active");
|
|
}
|
|
|
|
seq_printf(s, "\n");
|
|
return 0;
|
|
}
|
|
|
|
static const struct seq_operations tegra_pcie_ports_seq_ops = {
|
|
.start = tegra_pcie_ports_seq_start,
|
|
.next = tegra_pcie_ports_seq_next,
|
|
.stop = tegra_pcie_ports_seq_stop,
|
|
.show = tegra_pcie_ports_seq_show,
|
|
};
|
|
|
|
static int tegra_pcie_ports_open(struct inode *inode, struct file *file)
|
|
{
|
|
struct tegra_pcie *pcie = inode->i_private;
|
|
struct seq_file *s;
|
|
int err;
|
|
|
|
err = seq_open(file, &tegra_pcie_ports_seq_ops);
|
|
if (err)
|
|
return err;
|
|
|
|
s = file->private_data;
|
|
s->private = pcie;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct file_operations tegra_pcie_ports_ops = {
|
|
.owner = THIS_MODULE,
|
|
.open = tegra_pcie_ports_open,
|
|
.read = seq_read,
|
|
.llseek = seq_lseek,
|
|
.release = seq_release,
|
|
};
|
|
|
|
static void tegra_pcie_debugfs_exit(struct tegra_pcie *pcie)
|
|
{
|
|
if (pcie->debugfs)
|
|
debugfs_remove_recursive(pcie->debugfs);
|
|
}
|
|
|
|
static int tegra_pcie_debugfs_init(struct tegra_pcie *pcie)
|
|
{
|
|
struct dentry *file, *d;
|
|
struct tegra_pcie_port *port;
|
|
|
|
pcie->debugfs = debugfs_create_dir("pcie", NULL);
|
|
if (!pcie->debugfs)
|
|
return -ENOMEM;
|
|
|
|
file = debugfs_create_file("ports", S_IFREG | S_IRUGO, pcie->debugfs,
|
|
pcie, &tegra_pcie_ports_ops);
|
|
if (!file)
|
|
goto remove;
|
|
|
|
d = create_tegra_pcie_debufs_file("list_devices",
|
|
&list_devices_fops, pcie->debugfs,
|
|
(void *)pcie);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = create_tegra_pcie_debufs_file("apply_link_speed",
|
|
&apply_link_speed_fops, pcie->debugfs,
|
|
(void *)pcie);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = create_tegra_pcie_debufs_file("check_d3hot",
|
|
&check_d3hot_fops, pcie->debugfs,
|
|
(void *)pcie);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = create_tegra_pcie_debufs_file("power_down",
|
|
&power_down_fops, pcie->debugfs,
|
|
(void *)pcie);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = create_tegra_pcie_debufs_file("dump_config_space",
|
|
&dump_config_space_fops, pcie->debugfs,
|
|
(void *)pcie);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = create_tegra_pcie_debufs_file("dump_afi_space",
|
|
&dump_afi_space_fops, pcie->debugfs,
|
|
(void *)pcie);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = debugfs_create_u16("bus_dev_func", S_IWUGO | S_IRUGO,
|
|
pcie->debugfs,
|
|
&bdf);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = debugfs_create_u16("config_offset", S_IWUGO | S_IRUGO,
|
|
pcie->debugfs,
|
|
&config_offset);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = debugfs_create_u32("config_val", S_IWUGO | S_IRUGO,
|
|
pcie->debugfs,
|
|
&config_val);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = create_tegra_pcie_debufs_file("config_read",
|
|
&config_read_fops, pcie->debugfs,
|
|
(void *)pcie);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
d = create_tegra_pcie_debufs_file("config_write",
|
|
&config_write_fops, pcie->debugfs,
|
|
(void *)pcie);
|
|
if (!d)
|
|
goto remove;
|
|
d = create_tegra_pcie_debufs_file("aspm_l11",
|
|
&aspm_l11_fops, pcie->debugfs,
|
|
(void *)pcie);
|
|
if (!d)
|
|
goto remove;
|
|
d = create_tegra_pcie_debufs_file("aspm_l1ss",
|
|
&aspm_l1ss_fops, pcie->debugfs,
|
|
(void *)pcie);
|
|
if (!d)
|
|
goto remove;
|
|
|
|
list_for_each_entry(port, &pcie->ports, list) {
|
|
if (port->status)
|
|
if (tegra_pcie_port_debugfs_init(port))
|
|
goto remove;
|
|
}
|
|
|
|
return 0;
|
|
|
|
remove:
|
|
tegra_pcie_debugfs_exit(pcie);
|
|
pcie->debugfs = NULL;
|
|
return -ENOMEM;
|
|
}
|
|
|
|
static int tegra_pcie_probe_complete(struct tegra_pcie *pcie)
|
|
{
|
|
int ret = 0;
|
|
struct platform_device *pdev = to_platform_device(pcie->dev);
|
|
|
|
PR_FUNC_LINE;
|
|
ret = tegra_pcie_init(pcie);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (IS_ENABLED(CONFIG_DEBUG_FS))
|
|
if (pcie->num_ports) {
|
|
ret = tegra_pcie_debugfs_init(pcie);
|
|
if (ret < 0)
|
|
dev_err(&pdev->dev, "failed to setup debugfs: %d\n",
|
|
ret);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void pcie_delayed_detect(struct work_struct *work)
|
|
{
|
|
struct tegra_pcie *pcie;
|
|
struct platform_device *pdev;
|
|
int ret = 0;
|
|
|
|
pcie = container_of(work, struct tegra_pcie, detect_delay.work);
|
|
pdev = to_platform_device(pcie->dev);
|
|
#ifdef CONFIG_THERMAL
|
|
if (pcie->is_cooling_dev) {
|
|
dev_info(pcie->dev,
|
|
"Going to wait till end point temp is above 0-C\n");
|
|
wait_for_completion_interruptible(&pcie->completion);
|
|
dev_info(pcie->dev,
|
|
"proceeding with PCIe hierarchy enumeraton\n");
|
|
}
|
|
#endif
|
|
ret = tegra_pcie_probe_complete(pcie);
|
|
if (ret || !pcie->num_ports) {
|
|
pm_runtime_put_sync(pcie->dev);
|
|
goto release_regulators;
|
|
}
|
|
return;
|
|
|
|
release_regulators:
|
|
devm_kfree(pcie->dev, pcie->pcie_regulators);
|
|
devm_kfree(pcie->dev, pcie->plat_data);
|
|
pci_free_host_bridge(pcie->host);
|
|
platform_set_drvdata(pdev, NULL);
|
|
return;
|
|
}
|
|
|
|
#ifdef CONFIG_THERMAL
|
|
#define TEGRA_PCIE_THERM_MAX_STATE 2
|
|
static int tegra_pcie_max_state(struct thermal_cooling_device *tcd,
|
|
unsigned long *state)
|
|
{
|
|
struct tegra_pcie *pcie = tcd->devdata;
|
|
|
|
dev_info(pcie->dev, "%s state=%d\n", __func__,
|
|
pcie->therm_state.counter);
|
|
*state = TEGRA_PCIE_THERM_MAX_STATE;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int tegra_pcie_cur_state(struct thermal_cooling_device *tcd,
|
|
unsigned long *state)
|
|
{
|
|
struct tegra_pcie *pcie = tcd->devdata;
|
|
|
|
dev_info(pcie->dev, "%s state=%d\n", __func__,
|
|
pcie->therm_state.counter);
|
|
*state = (unsigned long)atomic_read(&pcie->therm_state);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int tegra_pcie_set_state(struct thermal_cooling_device *tcd,
|
|
unsigned long state)
|
|
{
|
|
struct tegra_pcie *pcie = tcd->devdata;
|
|
|
|
if (state != (TEGRA_PCIE_THERM_MAX_STATE - 1)) {
|
|
if ((unsigned long)atomic_read(&pcie->therm_state))
|
|
dev_err(pcie->dev, "dGPU temp is below zero...!\n");
|
|
else
|
|
return 0;
|
|
} else {
|
|
atomic_set(&pcie->therm_state, state);
|
|
complete(&pcie->completion);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/* Cooling device support */
|
|
static struct thermal_cooling_device_ops pcie_cdev_ops = {
|
|
.get_max_state = tegra_pcie_max_state,
|
|
.get_cur_state = tegra_pcie_cur_state,
|
|
.set_cur_state = tegra_pcie_set_state,
|
|
};
|
|
|
|
static int pcie_therm_init(struct tegra_pcie *pcie)
|
|
{
|
|
pcie->cdev = thermal_cooling_device_register("tegra-pcie", pcie,
|
|
&pcie_cdev_ops);
|
|
if (IS_ERR(pcie->cdev))
|
|
return PTR_ERR(pcie->cdev);
|
|
if (pcie->cdev == NULL)
|
|
return -ENODEV;
|
|
dev_info(pcie->dev, "PCIE cooling dev registered\n");
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
static int tegra_pcie_probe(struct platform_device *pdev)
|
|
{
|
|
int ret = 0;
|
|
int i;
|
|
const struct of_device_id *match;
|
|
struct tegra_pcie *pcie;
|
|
struct tegra_pcie_port *port, *tmp;
|
|
struct pci_host_bridge *host;
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
host = pci_alloc_host_bridge(sizeof(*pcie));
|
|
if (!host)
|
|
return -ENOMEM;
|
|
|
|
pcie = pci_host_bridge_priv(host);
|
|
|
|
platform_set_drvdata(pdev, pcie);
|
|
pcie->dev = &pdev->dev;
|
|
pcie->host = host;
|
|
|
|
/* use DT way to init platform data */
|
|
pcie->plat_data = devm_kzalloc(pcie->dev,
|
|
sizeof(*(pcie->plat_data)), GFP_KERNEL);
|
|
if (!(pcie->plat_data)) {
|
|
dev_err(pcie->dev, "memory alloc failed\n");
|
|
ret = -ENOMEM;
|
|
goto release_drvdata;
|
|
}
|
|
tegra_pcie_read_plat_data(pcie);
|
|
|
|
match = of_match_device(tegra_pcie_of_match, &pdev->dev);
|
|
if (!match) {
|
|
ret = -ENODEV;
|
|
goto release_platdata;
|
|
}
|
|
pcie->soc_data = (struct tegra_pcie_soc_data *)match->data;
|
|
|
|
if (pcie->soc_data->config_pex_io_dpd) {
|
|
pcie->pex_pin = devm_pinctrl_get(pcie->dev);
|
|
if (IS_ERR(pcie->pex_pin)) {
|
|
ret = PTR_ERR(pcie->pex_pin);
|
|
dev_err(pcie->dev, "pex io-dpd config failed: %d\n",
|
|
ret);
|
|
return ret;
|
|
}
|
|
|
|
pcie->pex_io_dpd_en_state = pinctrl_lookup_state(pcie->pex_pin,
|
|
"pex-io-dpd-en");
|
|
if (IS_ERR(pcie->pex_io_dpd_en_state)) {
|
|
ret = PTR_ERR(pcie->pex_io_dpd_en_state);
|
|
dev_err(pcie->dev, "missing pex-io-dpd en state: %d\n",
|
|
ret);
|
|
return ret;
|
|
}
|
|
pcie->pex_io_dpd_dis_state = pinctrl_lookup_state(pcie->pex_pin,
|
|
"pex-io-dpd-dis");
|
|
if (IS_ERR(pcie->pex_io_dpd_dis_state)) {
|
|
ret = PTR_ERR(pcie->pex_io_dpd_dis_state);
|
|
dev_err(pcie->dev, "missing pex-io-dpd dis state:%ld\n",
|
|
PTR_ERR(pcie->pex_io_dpd_dis_state));
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
pcie->pcie_regulators = devm_kzalloc(pcie->dev,
|
|
pcie->soc_data->num_pcie_regulators
|
|
* sizeof(struct regulator *), GFP_KERNEL);
|
|
|
|
for (i = 0; i < pcie->soc_data->num_pcie_regulators; i++) {
|
|
pcie->pcie_regulators[i] =
|
|
devm_regulator_get(pcie->dev,
|
|
pcie->soc_data->pcie_regulator_names[i]);
|
|
if (IS_ERR(pcie->pcie_regulators[i])) {
|
|
dev_err(pcie->dev, "%s: unable to get regulator %s\n",
|
|
__func__,
|
|
pcie->soc_data->pcie_regulator_names[i]);
|
|
pcie->pcie_regulators[i] = NULL;
|
|
ret = IS_ERR(pcie->pcie_regulators[i]);
|
|
goto release_regulators;
|
|
}
|
|
}
|
|
|
|
INIT_LIST_HEAD(&pcie->buses);
|
|
INIT_LIST_HEAD(&pcie->ports);
|
|
INIT_LIST_HEAD(&pcie->sys);
|
|
INIT_DELAYED_WORK(&pcie->detect_delay, pcie_delayed_detect);
|
|
|
|
ret = tegra_pcie_parse_dt(pcie);
|
|
if (ret < 0)
|
|
goto release_regulators;
|
|
|
|
if (pcie->soc_data->program_uphy) {
|
|
ret = tegra_pcie_phys_get(pcie);
|
|
if (ret < 0) {
|
|
if (ret == -EPROBE_DEFER)
|
|
dev_info(pcie->dev, "failed to get PHYs: %d\n", ret);
|
|
else
|
|
dev_err(pcie->dev, "failed to get PHYs: %d\n", ret);
|
|
goto release_regulators;
|
|
}
|
|
}
|
|
|
|
pcie->prod_list = devm_tegra_prod_get(pcie->dev);
|
|
if (IS_ERR(pcie->prod_list)) {
|
|
dev_info(pcie->dev, "No prod values found\n");
|
|
pcie->prod_list = NULL;
|
|
}
|
|
|
|
list_for_each_entry_safe(port, tmp, &pcie->ports, list) {
|
|
if (port->has_mxm_port) {
|
|
if (tegra_pcie_mxm_pwr_init(port))
|
|
dev_info(pcie->dev,
|
|
"pwr_good is down for port %d, ignoring\n",
|
|
port->index);
|
|
}
|
|
if (pcie->soc_data->program_clkreq_as_bi_dir) {
|
|
/* set clkreq as input to avoid root port control it */
|
|
tegra_pcie_config_clkreq(pcie, port->index, 0);
|
|
}
|
|
}
|
|
|
|
/* Enable Runtime PM for PCIe, TODO: Need to add PCIe host device */
|
|
pm_runtime_enable(pcie->dev);
|
|
|
|
#ifdef CONFIG_THERMAL
|
|
if (pcie->is_cooling_dev) {
|
|
init_completion(&pcie->completion);
|
|
/* register cooling device */
|
|
ret = pcie_therm_init(pcie);
|
|
if (ret != 0) {
|
|
dev_err(pcie->dev,
|
|
"unable to register cooling device err: %d\n",
|
|
ret);
|
|
goto release_regulators;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
schedule_delayed_work(&pcie->detect_delay,
|
|
msecs_to_jiffies(
|
|
pcie->plat_data->boot_detect_delay));
|
|
return ret;
|
|
|
|
release_regulators:
|
|
if (pcie->soc_data->program_uphy)
|
|
tegra_pcie_phy_exit(pcie);
|
|
devm_kfree(&pdev->dev, pcie->pcie_regulators);
|
|
release_platdata:
|
|
devm_kfree(&pdev->dev, pcie->plat_data);
|
|
release_drvdata:
|
|
pci_free_host_bridge(host);
|
|
platform_set_drvdata(pdev, NULL);
|
|
return ret;
|
|
}
|
|
|
|
static int tegra_pcie_remove(struct platform_device *pdev)
|
|
{
|
|
struct tegra_pcie *pcie = platform_get_drvdata(pdev);
|
|
|
|
PR_FUNC_LINE;
|
|
|
|
if (!pcie)
|
|
return 0;
|
|
|
|
if (cancel_delayed_work_sync(&pcie->detect_delay))
|
|
return 0;
|
|
if (IS_ENABLED(CONFIG_DEBUG_FS))
|
|
tegra_pcie_debugfs_exit(pcie);
|
|
pci_stop_root_bus(pcie->host->bus);
|
|
pci_bus_remove_resources(pcie->host->bus);
|
|
pci_remove_root_bus(pcie->host->bus);
|
|
iounmap(pcie->cfg_va_base);
|
|
if (IS_ENABLED(CONFIG_PCI_MSI))
|
|
tegra_pcie_disable_msi(pcie);
|
|
tegra_pcie_detach(pcie);
|
|
tegra_pcie_power_off(pcie);
|
|
if (pcie->soc_data->program_uphy)
|
|
tegra_pcie_phy_exit(pcie);
|
|
tegra_pcie_release_resources(pcie);
|
|
pm_runtime_disable(pcie->dev);
|
|
tegra_pcie_free_resources(pcie);
|
|
pci_free_host_bridge(pcie->host);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline u32 get_pme_port_offset(struct tegra_pcie_port *port)
|
|
{
|
|
u32 ret = 0;
|
|
|
|
switch (port->index) {
|
|
case 0:
|
|
ret = 0;
|
|
break;
|
|
case 1:
|
|
ret = 8;
|
|
break;
|
|
case 2:
|
|
ret = 12;
|
|
break;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static inline u32 get_pme_ack_offset(struct tegra_pcie_port *port)
|
|
{
|
|
u32 ret = 0;
|
|
|
|
switch (port->index) {
|
|
case 0:
|
|
ret = 5;
|
|
break;
|
|
case 1:
|
|
ret = 10;
|
|
break;
|
|
case 2:
|
|
ret = 14;
|
|
break;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
int tegra_pcie_pm_control(enum tegra_pcie_pm_opt pm_opt, void *user)
|
|
{
|
|
struct pci_dev *epdev = (struct pci_dev *)user;
|
|
struct pci_dev *rpdev = epdev->bus->self;
|
|
struct pci_host_bridge *host = pci_find_host_bridge(epdev->bus);
|
|
struct tegra_pcie *pcie = pci_host_bridge_priv(host);
|
|
|
|
struct tegra_pcie_port *port = NULL;
|
|
unsigned long ctrl = 0;
|
|
u32 rp;
|
|
u32 val;
|
|
u32 timeout = 100; /* 10 ms */
|
|
u16 val_16 = 0;
|
|
|
|
rp = PCI_SLOT(rpdev->devfn);
|
|
list_for_each_entry(port, &pcie->ports, list)
|
|
if (rp == port->index + 1)
|
|
break;
|
|
|
|
switch (pm_opt) {
|
|
case TEGRA_PCIE_SUSPEND:
|
|
pr_debug("---> in suspend\n");
|
|
|
|
port->ep_status = 0;
|
|
/* now setting it to '1' */
|
|
val = afi_readl(pcie, AFI_PCIE_PME);
|
|
val |= (0x1 << get_pme_port_offset(port));
|
|
afi_writel(pcie, val, AFI_PCIE_PME);
|
|
/* wait till ack is received */
|
|
do {
|
|
udelay(1);
|
|
val = afi_readl(pcie, AFI_PCIE_PME);
|
|
val = val & (0x1 << get_pme_ack_offset(port));
|
|
} while (!(val));
|
|
usleep_range(10000, 11000); /* 10ms delay */
|
|
/* clear PME_TO */
|
|
val = afi_readl(pcie, AFI_PCIE_PME);
|
|
val &= ~(0x1 << get_pme_port_offset(port));
|
|
afi_writel(pcie, val, AFI_PCIE_PME);
|
|
|
|
/* by this time, link would have gone into L2/L3 ready */
|
|
/* assert reset to EP */
|
|
pr_debug("---> asserting EP reset through AFI\n");
|
|
ctrl = tegra_pcie_port_get_pex_ctrl(port);
|
|
val = afi_readl(port->pcie, ctrl);
|
|
val &= ~AFI_PEX_CTRL_RST;
|
|
afi_writel(port->pcie, val, ctrl);
|
|
|
|
val = afi_readl(port->pcie, ctrl);
|
|
val &= ~AFI_PEX_CTRL_REFCLK_EN;
|
|
afi_writel(port->pcie, val, ctrl);
|
|
|
|
msleep(20);
|
|
break;
|
|
|
|
case TEGRA_PCIE_RESUME_PRE:
|
|
pr_debug("---> in resume (pre)\n");
|
|
|
|
port->ep_status = 1;
|
|
/* assert SBR on RP */
|
|
pr_debug("---> perform assert,de-assert of SBR\n");
|
|
pci_read_config_word(rpdev, PCI_BRIDGE_CONTROL, &val_16);
|
|
val_16 |= PCI_BRIDGE_CTL_BUS_RESET;
|
|
pci_write_config_word(rpdev, PCI_BRIDGE_CONTROL, val_16);
|
|
msleep(20);
|
|
val_16 &= ~PCI_BRIDGE_CTL_BUS_RESET;
|
|
pci_write_config_word(rpdev, PCI_BRIDGE_CONTROL, val_16);
|
|
msleep(100);
|
|
break;
|
|
|
|
case TEGRA_PCIE_RESUME_POST:
|
|
pr_debug("---> in resume (post)\n");
|
|
|
|
/* de-assert reset to EP */
|
|
pr_debug("---> de-asserting EP reset through AFI\n");
|
|
ctrl = tegra_pcie_port_get_pex_ctrl(port);
|
|
val = afi_readl(port->pcie, ctrl);
|
|
val |= AFI_PEX_CTRL_RST;
|
|
afi_writel(port->pcie, val, ctrl);
|
|
|
|
val = afi_readl(port->pcie, ctrl);
|
|
val |= AFI_PEX_CTRL_REFCLK_EN;
|
|
afi_writel(port->pcie, val, ctrl);
|
|
|
|
msleep(100);
|
|
|
|
/* make sure that link is up before doing anything */
|
|
do {
|
|
val = readl(port->base + RP_VEND_XP);
|
|
pr_debug("---> checking for link up\n");
|
|
if (val & RP_VEND_XP_DL_UP)
|
|
break;
|
|
usleep_range(100, 200);
|
|
} while (--timeout);
|
|
if (!timeout) {
|
|
dev_err(port->pcie->dev, "link %u is down\n",
|
|
port->index);
|
|
return -1;
|
|
}
|
|
|
|
/* try to read one config space register as this would
|
|
* result in completion timeout, hence we wouldn't be losing
|
|
* anything later on from second access onwards */
|
|
/* EP device state */
|
|
pr_debug("---> First config read START\n");
|
|
pci_read_config_word(epdev, PCI_DEVICE_ID, &val_16);
|
|
pr_debug("EP device ID = 0x%04X\n", val_16);
|
|
pr_debug("---> First config read END\n");
|
|
msleep(100);
|
|
break;
|
|
}
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(tegra_pcie_pm_control);
|
|
|
|
#ifdef CONFIG_PM
|
|
|
|
static int tegra_pcie_enable_msi(struct tegra_pcie *, bool);
|
|
|
|
static int tegra_pcie_resume(struct device *dev)
|
|
{
|
|
struct tegra_pcie *pcie = dev_get_drvdata(dev);
|
|
PR_FUNC_LINE;
|
|
if (!pcie)
|
|
return 0;
|
|
tegra_pcie_enable_features(pcie);
|
|
return 0;
|
|
}
|
|
|
|
/* Since BCM4359 WiFi driver is not informing the system about its absence
|
|
* when Wifi is turned off, PCIe subsystem tries to do save/restore as part
|
|
* of its routine during SC7 cycle will lead to error interrupt generation
|
|
* which prevents system entering into SC7 state. Hence, it is better to
|
|
* disable interrupts in suspend_late as there are no interrupts after this
|
|
* stage anyway and re-enable in resume_early
|
|
*/
|
|
static int tegra_pcie_suspend_late(struct device *dev)
|
|
{
|
|
struct tegra_pcie *pcie = dev_get_drvdata(dev);
|
|
u32 val = 0;
|
|
|
|
PR_FUNC_LINE;
|
|
if (!pcie)
|
|
return 0;
|
|
val = afi_readl(pcie, AFI_INTR_MASK);
|
|
val &= ~AFI_INTR_MASK_INT_MASK;
|
|
val &= ~AFI_INTR_MASK_MSI_MASK;
|
|
afi_writel(pcie, val, AFI_INTR_MASK);
|
|
return 0;
|
|
}
|
|
|
|
static int tegra_pcie_resume_early(struct device *dev)
|
|
{
|
|
struct tegra_pcie *pcie = dev_get_drvdata(dev);
|
|
u32 val = 0;
|
|
|
|
PR_FUNC_LINE;
|
|
if (!pcie)
|
|
return 0;
|
|
val = afi_readl(pcie, AFI_INTR_MASK);
|
|
val |= AFI_INTR_MASK_INT_MASK;
|
|
val |= AFI_INTR_MASK_MSI_MASK;
|
|
afi_writel(pcie, val, AFI_INTR_MASK);
|
|
return 0;
|
|
}
|
|
|
|
static const struct dev_pm_ops tegra_pcie_pm_ops = {
|
|
.resume = tegra_pcie_resume,
|
|
.suspend_late = tegra_pcie_suspend_late,
|
|
.resume_early = tegra_pcie_resume_early,
|
|
.runtime_suspend = tegra_pcie_save_device,
|
|
.runtime_resume = tegra_pcie_restore_device,
|
|
.suspend_noirq = tegra_pcie_save_device,
|
|
.resume_noirq = tegra_pcie_restore_device,
|
|
};
|
|
#endif /* CONFIG_PM */
|
|
|
|
/* driver data is accessed after init, so use __refdata instead of __initdata */
|
|
static struct platform_driver __refdata tegra_pcie_driver = {
|
|
.probe = tegra_pcie_probe,
|
|
.remove = tegra_pcie_remove,
|
|
.driver = {
|
|
.name = "tegra-pcie",
|
|
.owner = THIS_MODULE,
|
|
#ifdef CONFIG_PM
|
|
.pm = &tegra_pcie_pm_ops,
|
|
#endif
|
|
.of_match_table = tegra_pcie_of_match,
|
|
},
|
|
};
|
|
|
|
static int __init tegra_pcie_init_driver(void)
|
|
{
|
|
return platform_driver_register(&tegra_pcie_driver);
|
|
}
|
|
|
|
static void __exit tegra_pcie_exit_driver(void)
|
|
{
|
|
platform_driver_unregister(&tegra_pcie_driver);
|
|
}
|
|
|
|
module_init(tegra_pcie_init_driver);
|
|
module_exit(tegra_pcie_exit_driver);
|
|
MODULE_LICENSE("GPL v2");
|