mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-09-04 20:19:47 +08:00

The look up table (LUT) setting would be lost during the PCIe suspend on
i.MX95 SoC. So to ensure proper functionality after resume, save it during
suspend and restore it while resuming.
Fixes: 9d6b1bd6b3
("PCI: imx6: Add i.MX8MQ, i.MX8Q and i.MX95 PM support")
Signed-off-by: Richard Zhu <hongxing.zhu@nxp.com>
[mani: subject and description rewording]
Signed-off-by: Manivannan Sadhasivam <manivannan.sadhasivam@linaro.org>
Reviewed-by: Frank Li <Frank.Li@nxp.com>
Reviewed-by: Manivannan Sadhasivam <manivannan.sadhasivam@linaro.org>
Link: https://patch.msgid.link/20250416081314.3929794-8-hongxing.zhu@nxp.com
2049 lines
56 KiB
C
2049 lines
56 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* PCIe host controller driver for Freescale i.MX6 SoCs
|
|
*
|
|
* Copyright (C) 2013 Kosagi
|
|
* https://www.kosagi.com
|
|
*
|
|
* Author: Sean Cross <xobs@kosagi.com>
|
|
*/
|
|
|
|
#include <linux/bitfield.h>
|
|
#include <linux/clk.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/gpio/consumer.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/mfd/syscon.h>
|
|
#include <linux/mfd/syscon/imx6q-iomuxc-gpr.h>
|
|
#include <linux/mfd/syscon/imx7-iomuxc-gpr.h>
|
|
#include <linux/module.h>
|
|
#include <linux/of.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/regmap.h>
|
|
#include <linux/regulator/consumer.h>
|
|
#include <linux/resource.h>
|
|
#include <linux/signal.h>
|
|
#include <linux/types.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/reset.h>
|
|
#include <linux/phy/pcie.h>
|
|
#include <linux/phy/phy.h>
|
|
#include <linux/pm_domain.h>
|
|
#include <linux/pm_runtime.h>
|
|
|
|
#include "../../pci.h"
|
|
#include "pcie-designware.h"
|
|
|
|
#define IMX8MQ_GPR_PCIE_REF_USE_PAD BIT(9)
|
|
#define IMX8MQ_GPR_PCIE_CLK_REQ_OVERRIDE_EN BIT(10)
|
|
#define IMX8MQ_GPR_PCIE_CLK_REQ_OVERRIDE BIT(11)
|
|
#define IMX8MQ_GPR_PCIE_VREG_BYPASS BIT(12)
|
|
#define IMX8MQ_GPR12_PCIE2_CTRL_DEVICE_TYPE GENMASK(11, 8)
|
|
|
|
#define IMX95_PCIE_PHY_GEN_CTRL 0x0
|
|
#define IMX95_PCIE_REF_USE_PAD BIT(17)
|
|
|
|
#define IMX95_PCIE_PHY_MPLLA_CTRL 0x10
|
|
#define IMX95_PCIE_PHY_MPLL_STATE BIT(30)
|
|
|
|
#define IMX95_PCIE_SS_RW_REG_0 0xf0
|
|
#define IMX95_PCIE_REF_CLKEN BIT(23)
|
|
#define IMX95_PCIE_PHY_CR_PARA_SEL BIT(9)
|
|
#define IMX95_PCIE_SS_RW_REG_1 0xf4
|
|
#define IMX95_PCIE_SYS_AUX_PWR_DET BIT(31)
|
|
|
|
#define IMX95_PE0_GEN_CTRL_1 0x1050
|
|
#define IMX95_PCIE_DEVICE_TYPE GENMASK(3, 0)
|
|
|
|
#define IMX95_PE0_GEN_CTRL_3 0x1058
|
|
#define IMX95_PCIE_LTSSM_EN BIT(0)
|
|
|
|
#define IMX95_PE0_LUT_ACSCTRL 0x1008
|
|
#define IMX95_PEO_LUT_RWA BIT(16)
|
|
#define IMX95_PE0_LUT_ENLOC GENMASK(4, 0)
|
|
|
|
#define IMX95_PE0_LUT_DATA1 0x100c
|
|
#define IMX95_PE0_LUT_VLD BIT(31)
|
|
#define IMX95_PE0_LUT_DAC_ID GENMASK(10, 8)
|
|
#define IMX95_PE0_LUT_STREAM_ID GENMASK(5, 0)
|
|
|
|
#define IMX95_PE0_LUT_DATA2 0x1010
|
|
#define IMX95_PE0_LUT_REQID GENMASK(31, 16)
|
|
#define IMX95_PE0_LUT_MASK GENMASK(15, 0)
|
|
|
|
#define IMX95_SID_MASK GENMASK(5, 0)
|
|
#define IMX95_MAX_LUT 32
|
|
|
|
#define IMX95_PCIE_RST_CTRL 0x3010
|
|
#define IMX95_PCIE_COLD_RST BIT(0)
|
|
|
|
#define to_imx_pcie(x) dev_get_drvdata((x)->dev)
|
|
|
|
enum imx_pcie_variants {
|
|
IMX6Q,
|
|
IMX6SX,
|
|
IMX6QP,
|
|
IMX7D,
|
|
IMX8MQ,
|
|
IMX8MM,
|
|
IMX8MP,
|
|
IMX8Q,
|
|
IMX95,
|
|
IMX8MQ_EP,
|
|
IMX8MM_EP,
|
|
IMX8MP_EP,
|
|
IMX8Q_EP,
|
|
IMX95_EP,
|
|
};
|
|
|
|
#define IMX_PCIE_FLAG_IMX_PHY BIT(0)
|
|
#define IMX_PCIE_FLAG_SPEED_CHANGE_WORKAROUND BIT(1)
|
|
#define IMX_PCIE_FLAG_SUPPORTS_SUSPEND BIT(2)
|
|
#define IMX_PCIE_FLAG_HAS_PHYDRV BIT(3)
|
|
#define IMX_PCIE_FLAG_HAS_APP_RESET BIT(4)
|
|
#define IMX_PCIE_FLAG_HAS_PHY_RESET BIT(5)
|
|
#define IMX_PCIE_FLAG_HAS_SERDES BIT(6)
|
|
#define IMX_PCIE_FLAG_SUPPORT_64BIT BIT(7)
|
|
#define IMX_PCIE_FLAG_CPU_ADDR_FIXUP BIT(8)
|
|
/*
|
|
* Because of ERR005723 (PCIe does not support L2 power down) we need to
|
|
* workaround suspend resume on some devices which are affected by this errata.
|
|
*/
|
|
#define IMX_PCIE_FLAG_BROKEN_SUSPEND BIT(9)
|
|
#define IMX_PCIE_FLAG_HAS_LUT BIT(10)
|
|
#define IMX_PCIE_FLAG_8GT_ECN_ERR051586 BIT(11)
|
|
|
|
#define imx_check_flag(pci, val) (pci->drvdata->flags & val)
|
|
|
|
#define IMX_PCIE_MAX_INSTANCES 2
|
|
|
|
struct imx_pcie;
|
|
|
|
struct imx_pcie_drvdata {
|
|
enum imx_pcie_variants variant;
|
|
enum dw_pcie_device_mode mode;
|
|
u32 flags;
|
|
int dbi_length;
|
|
const char *gpr;
|
|
const u32 ltssm_off;
|
|
const u32 ltssm_mask;
|
|
const u32 mode_off[IMX_PCIE_MAX_INSTANCES];
|
|
const u32 mode_mask[IMX_PCIE_MAX_INSTANCES];
|
|
const struct pci_epc_features *epc_features;
|
|
int (*init_phy)(struct imx_pcie *pcie);
|
|
int (*enable_ref_clk)(struct imx_pcie *pcie, bool enable);
|
|
int (*core_reset)(struct imx_pcie *pcie, bool assert);
|
|
int (*wait_pll_lock)(struct imx_pcie *pcie);
|
|
const struct dw_pcie_host_ops *ops;
|
|
};
|
|
|
|
struct imx_lut_data {
|
|
u32 data1;
|
|
u32 data2;
|
|
};
|
|
|
|
struct imx_pcie {
|
|
struct dw_pcie *pci;
|
|
struct gpio_desc *reset_gpiod;
|
|
struct clk_bulk_data *clks;
|
|
int num_clks;
|
|
struct regmap *iomuxc_gpr;
|
|
u16 msi_ctrl;
|
|
u32 controller_id;
|
|
struct reset_control *pciephy_reset;
|
|
struct reset_control *apps_reset;
|
|
u32 tx_deemph_gen1;
|
|
u32 tx_deemph_gen2_3p5db;
|
|
u32 tx_deemph_gen2_6db;
|
|
u32 tx_swing_full;
|
|
u32 tx_swing_low;
|
|
struct regulator *vpcie;
|
|
struct regulator *vph;
|
|
void __iomem *phy_base;
|
|
|
|
/* LUT data for pcie */
|
|
struct imx_lut_data luts[IMX95_MAX_LUT];
|
|
/* power domain for pcie */
|
|
struct device *pd_pcie;
|
|
/* power domain for pcie phy */
|
|
struct device *pd_pcie_phy;
|
|
struct phy *phy;
|
|
const struct imx_pcie_drvdata *drvdata;
|
|
|
|
/* Ensure that only one device's LUT is configured at any given time */
|
|
struct mutex lock;
|
|
};
|
|
|
|
/* Parameters for the waiting for PCIe PHY PLL to lock on i.MX7 */
|
|
#define PHY_PLL_LOCK_WAIT_USLEEP_MAX 200
|
|
#define PHY_PLL_LOCK_WAIT_TIMEOUT (2000 * PHY_PLL_LOCK_WAIT_USLEEP_MAX)
|
|
|
|
/* PCIe Port Logic registers (memory-mapped) */
|
|
#define PL_OFFSET 0x700
|
|
|
|
#define PCIE_PHY_CTRL (PL_OFFSET + 0x114)
|
|
#define PCIE_PHY_CTRL_DATA(x) FIELD_PREP(GENMASK(15, 0), (x))
|
|
#define PCIE_PHY_CTRL_CAP_ADR BIT(16)
|
|
#define PCIE_PHY_CTRL_CAP_DAT BIT(17)
|
|
#define PCIE_PHY_CTRL_WR BIT(18)
|
|
#define PCIE_PHY_CTRL_RD BIT(19)
|
|
|
|
#define PCIE_PHY_STAT (PL_OFFSET + 0x110)
|
|
#define PCIE_PHY_STAT_ACK BIT(16)
|
|
|
|
/* PHY registers (not memory-mapped) */
|
|
#define PCIE_PHY_ATEOVRD 0x10
|
|
#define PCIE_PHY_ATEOVRD_EN BIT(2)
|
|
#define PCIE_PHY_ATEOVRD_REF_CLKDIV_SHIFT 0
|
|
#define PCIE_PHY_ATEOVRD_REF_CLKDIV_MASK 0x1
|
|
|
|
#define PCIE_PHY_MPLL_OVRD_IN_LO 0x11
|
|
#define PCIE_PHY_MPLL_MULTIPLIER_SHIFT 2
|
|
#define PCIE_PHY_MPLL_MULTIPLIER_MASK 0x7f
|
|
#define PCIE_PHY_MPLL_MULTIPLIER_OVRD BIT(9)
|
|
|
|
#define PCIE_PHY_RX_ASIC_OUT 0x100D
|
|
#define PCIE_PHY_RX_ASIC_OUT_VALID (1 << 0)
|
|
|
|
/* iMX7 PCIe PHY registers */
|
|
#define PCIE_PHY_CMN_REG4 0x14
|
|
/* These are probably the bits that *aren't* DCC_FB_EN */
|
|
#define PCIE_PHY_CMN_REG4_DCC_FB_EN 0x29
|
|
|
|
#define PCIE_PHY_CMN_REG15 0x54
|
|
#define PCIE_PHY_CMN_REG15_DLY_4 BIT(2)
|
|
#define PCIE_PHY_CMN_REG15_PLL_PD BIT(5)
|
|
#define PCIE_PHY_CMN_REG15_OVRD_PLL_PD BIT(7)
|
|
|
|
#define PCIE_PHY_CMN_REG24 0x90
|
|
#define PCIE_PHY_CMN_REG24_RX_EQ BIT(6)
|
|
#define PCIE_PHY_CMN_REG24_RX_EQ_SEL BIT(3)
|
|
|
|
#define PCIE_PHY_CMN_REG26 0x98
|
|
#define PCIE_PHY_CMN_REG26_ATT_MODE 0xBC
|
|
|
|
#define PHY_RX_OVRD_IN_LO 0x1005
|
|
#define PHY_RX_OVRD_IN_LO_RX_DATA_EN BIT(5)
|
|
#define PHY_RX_OVRD_IN_LO_RX_PLL_EN BIT(3)
|
|
|
|
static unsigned int imx_pcie_grp_offset(const struct imx_pcie *imx_pcie)
|
|
{
|
|
WARN_ON(imx_pcie->drvdata->variant != IMX8MQ &&
|
|
imx_pcie->drvdata->variant != IMX8MQ_EP &&
|
|
imx_pcie->drvdata->variant != IMX8MM &&
|
|
imx_pcie->drvdata->variant != IMX8MM_EP &&
|
|
imx_pcie->drvdata->variant != IMX8MP &&
|
|
imx_pcie->drvdata->variant != IMX8MP_EP);
|
|
return imx_pcie->controller_id == 1 ? IOMUXC_GPR16 : IOMUXC_GPR14;
|
|
}
|
|
|
|
static int imx95_pcie_init_phy(struct imx_pcie *imx_pcie)
|
|
{
|
|
/*
|
|
* ERR051624: The Controller Without Vaux Cannot Exit L23 Ready
|
|
* Through Beacon or PERST# De-assertion
|
|
*
|
|
* When the auxiliary power is not available, the controller
|
|
* cannot exit from L23 Ready with beacon or PERST# de-assertion
|
|
* when main power is not removed.
|
|
*
|
|
* Workaround: Set SS_RW_REG_1[SYS_AUX_PWR_DET] to 1.
|
|
*/
|
|
regmap_set_bits(imx_pcie->iomuxc_gpr, IMX95_PCIE_SS_RW_REG_1,
|
|
IMX95_PCIE_SYS_AUX_PWR_DET);
|
|
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr,
|
|
IMX95_PCIE_SS_RW_REG_0,
|
|
IMX95_PCIE_PHY_CR_PARA_SEL,
|
|
IMX95_PCIE_PHY_CR_PARA_SEL);
|
|
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr,
|
|
IMX95_PCIE_PHY_GEN_CTRL,
|
|
IMX95_PCIE_REF_USE_PAD, 0);
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr,
|
|
IMX95_PCIE_SS_RW_REG_0,
|
|
IMX95_PCIE_REF_CLKEN,
|
|
IMX95_PCIE_REF_CLKEN);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void imx_pcie_configure_type(struct imx_pcie *imx_pcie)
|
|
{
|
|
const struct imx_pcie_drvdata *drvdata = imx_pcie->drvdata;
|
|
unsigned int mask, val, mode, id;
|
|
|
|
if (drvdata->mode == DW_PCIE_EP_TYPE)
|
|
mode = PCI_EXP_TYPE_ENDPOINT;
|
|
else
|
|
mode = PCI_EXP_TYPE_ROOT_PORT;
|
|
|
|
id = imx_pcie->controller_id;
|
|
|
|
/* If mode_mask is 0, generic PHY driver is used to set the mode */
|
|
if (!drvdata->mode_mask[0])
|
|
return;
|
|
|
|
/* If mode_mask[id] is 0, each controller has its individual GPR */
|
|
if (!drvdata->mode_mask[id])
|
|
id = 0;
|
|
|
|
mask = drvdata->mode_mask[id];
|
|
val = mode << (ffs(mask) - 1);
|
|
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, drvdata->mode_off[id], mask, val);
|
|
}
|
|
|
|
static int pcie_phy_poll_ack(struct imx_pcie *imx_pcie, bool exp_val)
|
|
{
|
|
struct dw_pcie *pci = imx_pcie->pci;
|
|
bool val;
|
|
u32 max_iterations = 10;
|
|
u32 wait_counter = 0;
|
|
|
|
do {
|
|
val = dw_pcie_readl_dbi(pci, PCIE_PHY_STAT) &
|
|
PCIE_PHY_STAT_ACK;
|
|
wait_counter++;
|
|
|
|
if (val == exp_val)
|
|
return 0;
|
|
|
|
udelay(1);
|
|
} while (wait_counter < max_iterations);
|
|
|
|
return -ETIMEDOUT;
|
|
}
|
|
|
|
static int pcie_phy_wait_ack(struct imx_pcie *imx_pcie, int addr)
|
|
{
|
|
struct dw_pcie *pci = imx_pcie->pci;
|
|
u32 val;
|
|
int ret;
|
|
|
|
val = PCIE_PHY_CTRL_DATA(addr);
|
|
dw_pcie_writel_dbi(pci, PCIE_PHY_CTRL, val);
|
|
|
|
val |= PCIE_PHY_CTRL_CAP_ADR;
|
|
dw_pcie_writel_dbi(pci, PCIE_PHY_CTRL, val);
|
|
|
|
ret = pcie_phy_poll_ack(imx_pcie, true);
|
|
if (ret)
|
|
return ret;
|
|
|
|
val = PCIE_PHY_CTRL_DATA(addr);
|
|
dw_pcie_writel_dbi(pci, PCIE_PHY_CTRL, val);
|
|
|
|
return pcie_phy_poll_ack(imx_pcie, false);
|
|
}
|
|
|
|
/* Read from the 16-bit PCIe PHY control registers (not memory-mapped) */
|
|
static int pcie_phy_read(struct imx_pcie *imx_pcie, int addr, u16 *data)
|
|
{
|
|
struct dw_pcie *pci = imx_pcie->pci;
|
|
u32 phy_ctl;
|
|
int ret;
|
|
|
|
ret = pcie_phy_wait_ack(imx_pcie, addr);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* assert Read signal */
|
|
phy_ctl = PCIE_PHY_CTRL_RD;
|
|
dw_pcie_writel_dbi(pci, PCIE_PHY_CTRL, phy_ctl);
|
|
|
|
ret = pcie_phy_poll_ack(imx_pcie, true);
|
|
if (ret)
|
|
return ret;
|
|
|
|
*data = dw_pcie_readl_dbi(pci, PCIE_PHY_STAT);
|
|
|
|
/* deassert Read signal */
|
|
dw_pcie_writel_dbi(pci, PCIE_PHY_CTRL, 0x00);
|
|
|
|
return pcie_phy_poll_ack(imx_pcie, false);
|
|
}
|
|
|
|
static int pcie_phy_write(struct imx_pcie *imx_pcie, int addr, u16 data)
|
|
{
|
|
struct dw_pcie *pci = imx_pcie->pci;
|
|
u32 var;
|
|
int ret;
|
|
|
|
/* write addr */
|
|
/* cap addr */
|
|
ret = pcie_phy_wait_ack(imx_pcie, addr);
|
|
if (ret)
|
|
return ret;
|
|
|
|
var = PCIE_PHY_CTRL_DATA(data);
|
|
dw_pcie_writel_dbi(pci, PCIE_PHY_CTRL, var);
|
|
|
|
/* capture data */
|
|
var |= PCIE_PHY_CTRL_CAP_DAT;
|
|
dw_pcie_writel_dbi(pci, PCIE_PHY_CTRL, var);
|
|
|
|
ret = pcie_phy_poll_ack(imx_pcie, true);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* deassert cap data */
|
|
var = PCIE_PHY_CTRL_DATA(data);
|
|
dw_pcie_writel_dbi(pci, PCIE_PHY_CTRL, var);
|
|
|
|
/* wait for ack de-assertion */
|
|
ret = pcie_phy_poll_ack(imx_pcie, false);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* assert wr signal */
|
|
var = PCIE_PHY_CTRL_WR;
|
|
dw_pcie_writel_dbi(pci, PCIE_PHY_CTRL, var);
|
|
|
|
/* wait for ack */
|
|
ret = pcie_phy_poll_ack(imx_pcie, true);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* deassert wr signal */
|
|
var = PCIE_PHY_CTRL_DATA(data);
|
|
dw_pcie_writel_dbi(pci, PCIE_PHY_CTRL, var);
|
|
|
|
/* wait for ack de-assertion */
|
|
ret = pcie_phy_poll_ack(imx_pcie, false);
|
|
if (ret)
|
|
return ret;
|
|
|
|
dw_pcie_writel_dbi(pci, PCIE_PHY_CTRL, 0x0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx8mq_pcie_init_phy(struct imx_pcie *imx_pcie)
|
|
{
|
|
/* TODO: This code assumes external oscillator is being used */
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr,
|
|
imx_pcie_grp_offset(imx_pcie),
|
|
IMX8MQ_GPR_PCIE_REF_USE_PAD,
|
|
IMX8MQ_GPR_PCIE_REF_USE_PAD);
|
|
/*
|
|
* Per the datasheet, the PCIE_VPH is suggested to be 1.8V. If the
|
|
* PCIE_VPH is supplied by 3.3V, the VREG_BYPASS should be cleared
|
|
* to zero.
|
|
*/
|
|
if (imx_pcie->vph && regulator_get_voltage(imx_pcie->vph) > 3000000)
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr,
|
|
imx_pcie_grp_offset(imx_pcie),
|
|
IMX8MQ_GPR_PCIE_VREG_BYPASS,
|
|
0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx_pcie_init_phy(struct imx_pcie *imx_pcie)
|
|
{
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR12,
|
|
IMX6Q_GPR12_PCIE_CTL_2, 0 << 10);
|
|
|
|
/* configure constant input signal to the pcie ctrl and phy */
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR12,
|
|
IMX6Q_GPR12_LOS_LEVEL, 9 << 4);
|
|
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR8,
|
|
IMX6Q_GPR8_TX_DEEMPH_GEN1,
|
|
imx_pcie->tx_deemph_gen1 << 0);
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR8,
|
|
IMX6Q_GPR8_TX_DEEMPH_GEN2_3P5DB,
|
|
imx_pcie->tx_deemph_gen2_3p5db << 6);
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR8,
|
|
IMX6Q_GPR8_TX_DEEMPH_GEN2_6DB,
|
|
imx_pcie->tx_deemph_gen2_6db << 12);
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR8,
|
|
IMX6Q_GPR8_TX_SWING_FULL,
|
|
imx_pcie->tx_swing_full << 18);
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR8,
|
|
IMX6Q_GPR8_TX_SWING_LOW,
|
|
imx_pcie->tx_swing_low << 25);
|
|
return 0;
|
|
}
|
|
|
|
static int imx6sx_pcie_init_phy(struct imx_pcie *imx_pcie)
|
|
{
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR12,
|
|
IMX6SX_GPR12_PCIE_RX_EQ_MASK, IMX6SX_GPR12_PCIE_RX_EQ_2);
|
|
|
|
return imx_pcie_init_phy(imx_pcie);
|
|
}
|
|
|
|
static void imx7d_pcie_wait_for_phy_pll_lock(struct imx_pcie *imx_pcie)
|
|
{
|
|
u32 val;
|
|
struct device *dev = imx_pcie->pci->dev;
|
|
|
|
if (regmap_read_poll_timeout(imx_pcie->iomuxc_gpr,
|
|
IOMUXC_GPR22, val,
|
|
val & IMX7D_GPR22_PCIE_PHY_PLL_LOCKED,
|
|
PHY_PLL_LOCK_WAIT_USLEEP_MAX,
|
|
PHY_PLL_LOCK_WAIT_TIMEOUT))
|
|
dev_err(dev, "PCIe PLL lock timeout\n");
|
|
}
|
|
|
|
static int imx95_pcie_wait_for_phy_pll_lock(struct imx_pcie *imx_pcie)
|
|
{
|
|
u32 val;
|
|
struct device *dev = imx_pcie->pci->dev;
|
|
|
|
if (regmap_read_poll_timeout(imx_pcie->iomuxc_gpr,
|
|
IMX95_PCIE_PHY_MPLLA_CTRL, val,
|
|
val & IMX95_PCIE_PHY_MPLL_STATE,
|
|
PHY_PLL_LOCK_WAIT_USLEEP_MAX,
|
|
PHY_PLL_LOCK_WAIT_TIMEOUT)) {
|
|
dev_err(dev, "PCIe PLL lock timeout\n");
|
|
return -ETIMEDOUT;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx_setup_phy_mpll(struct imx_pcie *imx_pcie)
|
|
{
|
|
unsigned long phy_rate = 0;
|
|
int mult, div;
|
|
u16 val;
|
|
int i;
|
|
struct clk_bulk_data *clks = imx_pcie->clks;
|
|
|
|
if (!(imx_pcie->drvdata->flags & IMX_PCIE_FLAG_IMX_PHY))
|
|
return 0;
|
|
|
|
for (i = 0; i < imx_pcie->num_clks; i++)
|
|
if (strncmp(clks[i].id, "pcie_phy", 8) == 0)
|
|
phy_rate = clk_get_rate(clks[i].clk);
|
|
|
|
switch (phy_rate) {
|
|
case 125000000:
|
|
/*
|
|
* The default settings of the MPLL are for a 125MHz input
|
|
* clock, so no need to reconfigure anything in that case.
|
|
*/
|
|
return 0;
|
|
case 100000000:
|
|
mult = 25;
|
|
div = 0;
|
|
break;
|
|
case 200000000:
|
|
mult = 25;
|
|
div = 1;
|
|
break;
|
|
default:
|
|
dev_err(imx_pcie->pci->dev,
|
|
"Unsupported PHY reference clock rate %lu\n", phy_rate);
|
|
return -EINVAL;
|
|
}
|
|
|
|
pcie_phy_read(imx_pcie, PCIE_PHY_MPLL_OVRD_IN_LO, &val);
|
|
val &= ~(PCIE_PHY_MPLL_MULTIPLIER_MASK <<
|
|
PCIE_PHY_MPLL_MULTIPLIER_SHIFT);
|
|
val |= mult << PCIE_PHY_MPLL_MULTIPLIER_SHIFT;
|
|
val |= PCIE_PHY_MPLL_MULTIPLIER_OVRD;
|
|
pcie_phy_write(imx_pcie, PCIE_PHY_MPLL_OVRD_IN_LO, val);
|
|
|
|
pcie_phy_read(imx_pcie, PCIE_PHY_ATEOVRD, &val);
|
|
val &= ~(PCIE_PHY_ATEOVRD_REF_CLKDIV_MASK <<
|
|
PCIE_PHY_ATEOVRD_REF_CLKDIV_SHIFT);
|
|
val |= div << PCIE_PHY_ATEOVRD_REF_CLKDIV_SHIFT;
|
|
val |= PCIE_PHY_ATEOVRD_EN;
|
|
pcie_phy_write(imx_pcie, PCIE_PHY_ATEOVRD, val);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void imx_pcie_reset_phy(struct imx_pcie *imx_pcie)
|
|
{
|
|
u16 tmp;
|
|
|
|
if (!(imx_pcie->drvdata->flags & IMX_PCIE_FLAG_IMX_PHY))
|
|
return;
|
|
|
|
pcie_phy_read(imx_pcie, PHY_RX_OVRD_IN_LO, &tmp);
|
|
tmp |= (PHY_RX_OVRD_IN_LO_RX_DATA_EN |
|
|
PHY_RX_OVRD_IN_LO_RX_PLL_EN);
|
|
pcie_phy_write(imx_pcie, PHY_RX_OVRD_IN_LO, tmp);
|
|
|
|
usleep_range(2000, 3000);
|
|
|
|
pcie_phy_read(imx_pcie, PHY_RX_OVRD_IN_LO, &tmp);
|
|
tmp &= ~(PHY_RX_OVRD_IN_LO_RX_DATA_EN |
|
|
PHY_RX_OVRD_IN_LO_RX_PLL_EN);
|
|
pcie_phy_write(imx_pcie, PHY_RX_OVRD_IN_LO, tmp);
|
|
}
|
|
|
|
#ifdef CONFIG_ARM
|
|
/* Added for PCI abort handling */
|
|
static int imx6q_pcie_abort_handler(unsigned long addr,
|
|
unsigned int fsr, struct pt_regs *regs)
|
|
{
|
|
unsigned long pc = instruction_pointer(regs);
|
|
unsigned long instr = *(unsigned long *)pc;
|
|
int reg = (instr >> 12) & 15;
|
|
|
|
/*
|
|
* If the instruction being executed was a read,
|
|
* make it look like it read all-ones.
|
|
*/
|
|
if ((instr & 0x0c100000) == 0x04100000) {
|
|
unsigned long val;
|
|
|
|
if (instr & 0x00400000)
|
|
val = 255;
|
|
else
|
|
val = -1;
|
|
|
|
regs->uregs[reg] = val;
|
|
regs->ARM_pc += 4;
|
|
return 0;
|
|
}
|
|
|
|
if ((instr & 0x0e100090) == 0x00100090) {
|
|
regs->uregs[reg] = -1;
|
|
regs->ARM_pc += 4;
|
|
return 0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
#endif
|
|
|
|
static int imx_pcie_attach_pd(struct device *dev)
|
|
{
|
|
struct imx_pcie *imx_pcie = dev_get_drvdata(dev);
|
|
struct device_link *link;
|
|
|
|
/* Do nothing when in a single power domain */
|
|
if (dev->pm_domain)
|
|
return 0;
|
|
|
|
imx_pcie->pd_pcie = dev_pm_domain_attach_by_name(dev, "pcie");
|
|
if (IS_ERR(imx_pcie->pd_pcie))
|
|
return PTR_ERR(imx_pcie->pd_pcie);
|
|
/* Do nothing when power domain missing */
|
|
if (!imx_pcie->pd_pcie)
|
|
return 0;
|
|
link = device_link_add(dev, imx_pcie->pd_pcie,
|
|
DL_FLAG_STATELESS |
|
|
DL_FLAG_PM_RUNTIME |
|
|
DL_FLAG_RPM_ACTIVE);
|
|
if (!link) {
|
|
dev_err(dev, "Failed to add device_link to pcie pd\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
imx_pcie->pd_pcie_phy = dev_pm_domain_attach_by_name(dev, "pcie_phy");
|
|
if (IS_ERR(imx_pcie->pd_pcie_phy))
|
|
return PTR_ERR(imx_pcie->pd_pcie_phy);
|
|
|
|
link = device_link_add(dev, imx_pcie->pd_pcie_phy,
|
|
DL_FLAG_STATELESS |
|
|
DL_FLAG_PM_RUNTIME |
|
|
DL_FLAG_RPM_ACTIVE);
|
|
if (!link) {
|
|
dev_err(dev, "Failed to add device_link to pcie_phy pd\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx6sx_pcie_enable_ref_clk(struct imx_pcie *imx_pcie, bool enable)
|
|
{
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR12,
|
|
IMX6SX_GPR12_PCIE_TEST_POWERDOWN,
|
|
enable ? 0 : IMX6SX_GPR12_PCIE_TEST_POWERDOWN);
|
|
return 0;
|
|
}
|
|
|
|
static int imx6q_pcie_enable_ref_clk(struct imx_pcie *imx_pcie, bool enable)
|
|
{
|
|
if (enable) {
|
|
/* power up core phy and enable ref clock */
|
|
regmap_clear_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR1, IMX6Q_GPR1_PCIE_TEST_PD);
|
|
/*
|
|
* The async reset input need ref clock to sync internally,
|
|
* when the ref clock comes after reset, internal synced
|
|
* reset time is too short, cannot meet the requirement.
|
|
* Add a ~10us delay here.
|
|
*/
|
|
usleep_range(10, 100);
|
|
regmap_set_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR1, IMX6Q_GPR1_PCIE_REF_CLK_EN);
|
|
} else {
|
|
regmap_clear_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR1, IMX6Q_GPR1_PCIE_REF_CLK_EN);
|
|
regmap_set_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR1, IMX6Q_GPR1_PCIE_TEST_PD);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx8mm_pcie_enable_ref_clk(struct imx_pcie *imx_pcie, bool enable)
|
|
{
|
|
int offset = imx_pcie_grp_offset(imx_pcie);
|
|
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, offset,
|
|
IMX8MQ_GPR_PCIE_CLK_REQ_OVERRIDE,
|
|
enable ? 0 : IMX8MQ_GPR_PCIE_CLK_REQ_OVERRIDE);
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, offset,
|
|
IMX8MQ_GPR_PCIE_CLK_REQ_OVERRIDE_EN,
|
|
enable ? IMX8MQ_GPR_PCIE_CLK_REQ_OVERRIDE_EN : 0);
|
|
return 0;
|
|
}
|
|
|
|
static int imx7d_pcie_enable_ref_clk(struct imx_pcie *imx_pcie, bool enable)
|
|
{
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR12,
|
|
IMX7D_GPR12_PCIE_PHY_REFCLK_SEL,
|
|
enable ? 0 : IMX7D_GPR12_PCIE_PHY_REFCLK_SEL);
|
|
return 0;
|
|
}
|
|
|
|
static int imx_pcie_clk_enable(struct imx_pcie *imx_pcie)
|
|
{
|
|
struct dw_pcie *pci = imx_pcie->pci;
|
|
struct device *dev = pci->dev;
|
|
int ret;
|
|
|
|
ret = clk_bulk_prepare_enable(imx_pcie->num_clks, imx_pcie->clks);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (imx_pcie->drvdata->enable_ref_clk) {
|
|
ret = imx_pcie->drvdata->enable_ref_clk(imx_pcie, true);
|
|
if (ret) {
|
|
dev_err(dev, "Failed to enable PCIe REFCLK\n");
|
|
goto err_ref_clk;
|
|
}
|
|
}
|
|
|
|
/* allow the clocks to stabilize */
|
|
usleep_range(200, 500);
|
|
return 0;
|
|
|
|
err_ref_clk:
|
|
clk_bulk_disable_unprepare(imx_pcie->num_clks, imx_pcie->clks);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void imx_pcie_clk_disable(struct imx_pcie *imx_pcie)
|
|
{
|
|
if (imx_pcie->drvdata->enable_ref_clk)
|
|
imx_pcie->drvdata->enable_ref_clk(imx_pcie, false);
|
|
clk_bulk_disable_unprepare(imx_pcie->num_clks, imx_pcie->clks);
|
|
}
|
|
|
|
static int imx6sx_pcie_core_reset(struct imx_pcie *imx_pcie, bool assert)
|
|
{
|
|
if (assert)
|
|
regmap_set_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR12,
|
|
IMX6SX_GPR12_PCIE_TEST_POWERDOWN);
|
|
|
|
/* Force PCIe PHY reset */
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR5, IMX6SX_GPR5_PCIE_BTNRST_RESET,
|
|
assert ? IMX6SX_GPR5_PCIE_BTNRST_RESET : 0);
|
|
return 0;
|
|
}
|
|
|
|
static int imx6qp_pcie_core_reset(struct imx_pcie *imx_pcie, bool assert)
|
|
{
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR1, IMX6Q_GPR1_PCIE_SW_RST,
|
|
assert ? IMX6Q_GPR1_PCIE_SW_RST : 0);
|
|
if (!assert)
|
|
usleep_range(200, 500);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx6q_pcie_core_reset(struct imx_pcie *imx_pcie, bool assert)
|
|
{
|
|
if (!assert)
|
|
return 0;
|
|
|
|
regmap_set_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR1, IMX6Q_GPR1_PCIE_TEST_PD);
|
|
regmap_set_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR1, IMX6Q_GPR1_PCIE_REF_CLK_EN);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx7d_pcie_core_reset(struct imx_pcie *imx_pcie, bool assert)
|
|
{
|
|
struct dw_pcie *pci = imx_pcie->pci;
|
|
struct device *dev = pci->dev;
|
|
|
|
if (assert)
|
|
return 0;
|
|
|
|
/*
|
|
* Workaround for ERR010728 (IMX7DS_2N09P, Rev. 1.1, 4/2023):
|
|
*
|
|
* PCIe: PLL may fail to lock under corner conditions.
|
|
*
|
|
* Initial VCO oscillation may fail under corner conditions such as
|
|
* cold temperature which will cause the PCIe PLL fail to lock in the
|
|
* initialization phase.
|
|
*
|
|
* The Duty-cycle Corrector calibration must be disabled.
|
|
*
|
|
* 1. De-assert the G_RST signal by clearing
|
|
* SRC_PCIEPHY_RCR[PCIEPHY_G_RST].
|
|
* 2. De-assert DCC_FB_EN by writing data “0x29” to the register
|
|
* address 0x306d0014 (PCIE_PHY_CMN_REG4).
|
|
* 3. Assert RX_EQS, RX_EQ_SEL by writing data “0x48” to the register
|
|
* address 0x306d0090 (PCIE_PHY_CMN_REG24).
|
|
* 4. Assert ATT_MODE by writing data “0xbc” to the register
|
|
* address 0x306d0098 (PCIE_PHY_CMN_REG26).
|
|
* 5. De-assert the CMN_RST signal by clearing register bit
|
|
* SRC_PCIEPHY_RCR[PCIEPHY_BTN]
|
|
*/
|
|
|
|
if (likely(imx_pcie->phy_base)) {
|
|
/* De-assert DCC_FB_EN */
|
|
writel(PCIE_PHY_CMN_REG4_DCC_FB_EN, imx_pcie->phy_base + PCIE_PHY_CMN_REG4);
|
|
/* Assert RX_EQS and RX_EQS_SEL */
|
|
writel(PCIE_PHY_CMN_REG24_RX_EQ_SEL | PCIE_PHY_CMN_REG24_RX_EQ,
|
|
imx_pcie->phy_base + PCIE_PHY_CMN_REG24);
|
|
/* Assert ATT_MODE */
|
|
writel(PCIE_PHY_CMN_REG26_ATT_MODE, imx_pcie->phy_base + PCIE_PHY_CMN_REG26);
|
|
} else {
|
|
dev_warn(dev, "Unable to apply ERR010728 workaround. DT missing fsl,imx7d-pcie-phy phandle ?\n");
|
|
}
|
|
imx7d_pcie_wait_for_phy_pll_lock(imx_pcie);
|
|
return 0;
|
|
}
|
|
|
|
static int imx95_pcie_core_reset(struct imx_pcie *imx_pcie, bool assert)
|
|
{
|
|
u32 val;
|
|
|
|
if (assert) {
|
|
/*
|
|
* From i.MX95 PCIe PHY perspective, the COLD reset toggle
|
|
* should be complete after power-up by the following sequence.
|
|
* > 10us(at power-up)
|
|
* > 10ns(warm reset)
|
|
* |<------------>|
|
|
* ______________
|
|
* phy_reset ____/ \________________
|
|
* ____________
|
|
* ref_clk_en_______________________/
|
|
* Toggle COLD reset aligned with this sequence for i.MX95 PCIe.
|
|
*/
|
|
regmap_set_bits(imx_pcie->iomuxc_gpr, IMX95_PCIE_RST_CTRL,
|
|
IMX95_PCIE_COLD_RST);
|
|
/*
|
|
* Make sure the write to IMX95_PCIE_RST_CTRL is flushed to the
|
|
* hardware by doing a read. Otherwise, there is no guarantee
|
|
* that the write has reached the hardware before udelay().
|
|
*/
|
|
regmap_read_bypassed(imx_pcie->iomuxc_gpr, IMX95_PCIE_RST_CTRL,
|
|
&val);
|
|
udelay(15);
|
|
regmap_clear_bits(imx_pcie->iomuxc_gpr, IMX95_PCIE_RST_CTRL,
|
|
IMX95_PCIE_COLD_RST);
|
|
regmap_read_bypassed(imx_pcie->iomuxc_gpr, IMX95_PCIE_RST_CTRL,
|
|
&val);
|
|
udelay(10);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void imx_pcie_assert_core_reset(struct imx_pcie *imx_pcie)
|
|
{
|
|
reset_control_assert(imx_pcie->pciephy_reset);
|
|
reset_control_assert(imx_pcie->apps_reset);
|
|
|
|
if (imx_pcie->drvdata->core_reset)
|
|
imx_pcie->drvdata->core_reset(imx_pcie, true);
|
|
|
|
/* Some boards don't have PCIe reset GPIO. */
|
|
gpiod_set_value_cansleep(imx_pcie->reset_gpiod, 1);
|
|
}
|
|
|
|
static int imx_pcie_deassert_core_reset(struct imx_pcie *imx_pcie)
|
|
{
|
|
reset_control_deassert(imx_pcie->pciephy_reset);
|
|
reset_control_deassert(imx_pcie->apps_reset);
|
|
|
|
if (imx_pcie->drvdata->core_reset)
|
|
imx_pcie->drvdata->core_reset(imx_pcie, false);
|
|
|
|
/* Some boards don't have PCIe reset GPIO. */
|
|
if (imx_pcie->reset_gpiod) {
|
|
msleep(100);
|
|
gpiod_set_value_cansleep(imx_pcie->reset_gpiod, 0);
|
|
/* Wait for 100ms after PERST# deassertion (PCIe r5.0, 6.6.1) */
|
|
msleep(100);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx_pcie_wait_for_speed_change(struct imx_pcie *imx_pcie)
|
|
{
|
|
struct dw_pcie *pci = imx_pcie->pci;
|
|
struct device *dev = pci->dev;
|
|
u32 tmp;
|
|
unsigned int retries;
|
|
|
|
for (retries = 0; retries < 200; retries++) {
|
|
tmp = dw_pcie_readl_dbi(pci, PCIE_LINK_WIDTH_SPEED_CONTROL);
|
|
/* Test if the speed change finished. */
|
|
if (!(tmp & PORT_LOGIC_SPEED_CHANGE))
|
|
return 0;
|
|
usleep_range(100, 1000);
|
|
}
|
|
|
|
dev_err(dev, "Speed change timeout\n");
|
|
return -ETIMEDOUT;
|
|
}
|
|
|
|
static void imx_pcie_ltssm_enable(struct device *dev)
|
|
{
|
|
struct imx_pcie *imx_pcie = dev_get_drvdata(dev);
|
|
const struct imx_pcie_drvdata *drvdata = imx_pcie->drvdata;
|
|
u8 offset = dw_pcie_find_capability(imx_pcie->pci, PCI_CAP_ID_EXP);
|
|
u32 tmp;
|
|
|
|
tmp = dw_pcie_readl_dbi(imx_pcie->pci, offset + PCI_EXP_LNKCAP);
|
|
phy_set_speed(imx_pcie->phy, FIELD_GET(PCI_EXP_LNKCAP_SLS, tmp));
|
|
if (drvdata->ltssm_mask)
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, drvdata->ltssm_off, drvdata->ltssm_mask,
|
|
drvdata->ltssm_mask);
|
|
|
|
reset_control_deassert(imx_pcie->apps_reset);
|
|
}
|
|
|
|
static void imx_pcie_ltssm_disable(struct device *dev)
|
|
{
|
|
struct imx_pcie *imx_pcie = dev_get_drvdata(dev);
|
|
const struct imx_pcie_drvdata *drvdata = imx_pcie->drvdata;
|
|
|
|
phy_set_speed(imx_pcie->phy, 0);
|
|
if (drvdata->ltssm_mask)
|
|
regmap_update_bits(imx_pcie->iomuxc_gpr, drvdata->ltssm_off,
|
|
drvdata->ltssm_mask, 0);
|
|
|
|
reset_control_assert(imx_pcie->apps_reset);
|
|
}
|
|
|
|
static int imx_pcie_start_link(struct dw_pcie *pci)
|
|
{
|
|
struct imx_pcie *imx_pcie = to_imx_pcie(pci);
|
|
struct device *dev = pci->dev;
|
|
u8 offset = dw_pcie_find_capability(pci, PCI_CAP_ID_EXP);
|
|
u32 tmp;
|
|
int ret;
|
|
|
|
if (!(imx_pcie->drvdata->flags &
|
|
IMX_PCIE_FLAG_SPEED_CHANGE_WORKAROUND)) {
|
|
imx_pcie_ltssm_enable(dev);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Force Gen1 operation when starting the link. In case the link is
|
|
* started in Gen2 mode, there is a possibility the devices on the
|
|
* bus will not be detected at all. This happens with PCIe switches.
|
|
*/
|
|
dw_pcie_dbi_ro_wr_en(pci);
|
|
tmp = dw_pcie_readl_dbi(pci, offset + PCI_EXP_LNKCAP);
|
|
tmp &= ~PCI_EXP_LNKCAP_SLS;
|
|
tmp |= PCI_EXP_LNKCAP_SLS_2_5GB;
|
|
dw_pcie_writel_dbi(pci, offset + PCI_EXP_LNKCAP, tmp);
|
|
dw_pcie_dbi_ro_wr_dis(pci);
|
|
|
|
/* Start LTSSM. */
|
|
imx_pcie_ltssm_enable(dev);
|
|
|
|
if (pci->max_link_speed > 1) {
|
|
ret = dw_pcie_wait_for_link(pci);
|
|
if (ret)
|
|
goto err_reset_phy;
|
|
|
|
/* Allow faster modes after the link is up */
|
|
dw_pcie_dbi_ro_wr_en(pci);
|
|
tmp = dw_pcie_readl_dbi(pci, offset + PCI_EXP_LNKCAP);
|
|
tmp &= ~PCI_EXP_LNKCAP_SLS;
|
|
tmp |= pci->max_link_speed;
|
|
dw_pcie_writel_dbi(pci, offset + PCI_EXP_LNKCAP, tmp);
|
|
|
|
/*
|
|
* Start Directed Speed Change so the best possible
|
|
* speed both link partners support can be negotiated.
|
|
*/
|
|
tmp = dw_pcie_readl_dbi(pci, PCIE_LINK_WIDTH_SPEED_CONTROL);
|
|
tmp |= PORT_LOGIC_SPEED_CHANGE;
|
|
dw_pcie_writel_dbi(pci, PCIE_LINK_WIDTH_SPEED_CONTROL, tmp);
|
|
dw_pcie_dbi_ro_wr_dis(pci);
|
|
|
|
ret = imx_pcie_wait_for_speed_change(imx_pcie);
|
|
if (ret) {
|
|
dev_err(dev, "Failed to bring link up!\n");
|
|
goto err_reset_phy;
|
|
}
|
|
} else {
|
|
dev_info(dev, "Link: Only Gen1 is enabled\n");
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_reset_phy:
|
|
dev_dbg(dev, "PHY DEBUG_R0=0x%08x DEBUG_R1=0x%08x\n",
|
|
dw_pcie_readl_dbi(pci, PCIE_PORT_DEBUG0),
|
|
dw_pcie_readl_dbi(pci, PCIE_PORT_DEBUG1));
|
|
imx_pcie_reset_phy(imx_pcie);
|
|
return 0;
|
|
}
|
|
|
|
static void imx_pcie_stop_link(struct dw_pcie *pci)
|
|
{
|
|
struct device *dev = pci->dev;
|
|
|
|
/* Turn off PCIe LTSSM */
|
|
imx_pcie_ltssm_disable(dev);
|
|
}
|
|
|
|
static int imx_pcie_add_lut(struct imx_pcie *imx_pcie, u16 rid, u8 sid)
|
|
{
|
|
struct dw_pcie *pci = imx_pcie->pci;
|
|
struct device *dev = pci->dev;
|
|
u32 data1, data2;
|
|
int free = -1;
|
|
int i;
|
|
|
|
if (sid >= 64) {
|
|
dev_err(dev, "Invalid SID for index %d\n", sid);
|
|
return -EINVAL;
|
|
}
|
|
|
|
guard(mutex)(&imx_pcie->lock);
|
|
|
|
/*
|
|
* Iterate through all LUT entries to check for duplicate RID and
|
|
* identify the first available entry. Configure this available entry
|
|
* immediately after verification to avoid rescanning it.
|
|
*/
|
|
for (i = 0; i < IMX95_MAX_LUT; i++) {
|
|
regmap_write(imx_pcie->iomuxc_gpr,
|
|
IMX95_PE0_LUT_ACSCTRL, IMX95_PEO_LUT_RWA | i);
|
|
regmap_read(imx_pcie->iomuxc_gpr, IMX95_PE0_LUT_DATA1, &data1);
|
|
|
|
if (!(data1 & IMX95_PE0_LUT_VLD)) {
|
|
if (free < 0)
|
|
free = i;
|
|
continue;
|
|
}
|
|
|
|
regmap_read(imx_pcie->iomuxc_gpr, IMX95_PE0_LUT_DATA2, &data2);
|
|
|
|
/* Do not add duplicate RID */
|
|
if (rid == FIELD_GET(IMX95_PE0_LUT_REQID, data2)) {
|
|
dev_warn(dev, "Existing LUT entry available for RID (%d)", rid);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
if (free < 0) {
|
|
dev_err(dev, "LUT entry is not available\n");
|
|
return -ENOSPC;
|
|
}
|
|
|
|
data1 = FIELD_PREP(IMX95_PE0_LUT_DAC_ID, 0);
|
|
data1 |= FIELD_PREP(IMX95_PE0_LUT_STREAM_ID, sid);
|
|
data1 |= IMX95_PE0_LUT_VLD;
|
|
regmap_write(imx_pcie->iomuxc_gpr, IMX95_PE0_LUT_DATA1, data1);
|
|
|
|
data2 = IMX95_PE0_LUT_MASK; /* Match all bits of RID */
|
|
data2 |= FIELD_PREP(IMX95_PE0_LUT_REQID, rid);
|
|
regmap_write(imx_pcie->iomuxc_gpr, IMX95_PE0_LUT_DATA2, data2);
|
|
|
|
regmap_write(imx_pcie->iomuxc_gpr, IMX95_PE0_LUT_ACSCTRL, free);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void imx_pcie_remove_lut(struct imx_pcie *imx_pcie, u16 rid)
|
|
{
|
|
u32 data2;
|
|
int i;
|
|
|
|
guard(mutex)(&imx_pcie->lock);
|
|
|
|
for (i = 0; i < IMX95_MAX_LUT; i++) {
|
|
regmap_write(imx_pcie->iomuxc_gpr,
|
|
IMX95_PE0_LUT_ACSCTRL, IMX95_PEO_LUT_RWA | i);
|
|
regmap_read(imx_pcie->iomuxc_gpr, IMX95_PE0_LUT_DATA2, &data2);
|
|
if (FIELD_GET(IMX95_PE0_LUT_REQID, data2) == rid) {
|
|
regmap_write(imx_pcie->iomuxc_gpr,
|
|
IMX95_PE0_LUT_DATA1, 0);
|
|
regmap_write(imx_pcie->iomuxc_gpr,
|
|
IMX95_PE0_LUT_DATA2, 0);
|
|
regmap_write(imx_pcie->iomuxc_gpr,
|
|
IMX95_PE0_LUT_ACSCTRL, i);
|
|
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
static int imx_pcie_enable_device(struct pci_host_bridge *bridge,
|
|
struct pci_dev *pdev)
|
|
{
|
|
struct imx_pcie *imx_pcie = to_imx_pcie(to_dw_pcie_from_pp(bridge->sysdata));
|
|
u32 sid_i, sid_m, rid = pci_dev_id(pdev);
|
|
struct device_node *target;
|
|
struct device *dev;
|
|
int err_i, err_m;
|
|
u32 sid = 0;
|
|
|
|
dev = imx_pcie->pci->dev;
|
|
|
|
target = NULL;
|
|
err_i = of_map_id(dev->of_node, rid, "iommu-map", "iommu-map-mask",
|
|
&target, &sid_i);
|
|
if (target) {
|
|
of_node_put(target);
|
|
} else {
|
|
/*
|
|
* "target == NULL && err_i == 0" means RID out of map range.
|
|
* Use 1:1 map RID to streamID. Hardware can't support this
|
|
* because the streamID is only 6 bits
|
|
*/
|
|
err_i = -EINVAL;
|
|
}
|
|
|
|
target = NULL;
|
|
err_m = of_map_id(dev->of_node, rid, "msi-map", "msi-map-mask",
|
|
&target, &sid_m);
|
|
|
|
/*
|
|
* err_m target
|
|
* 0 NULL RID out of range. Use 1:1 map RID to
|
|
* streamID, Current hardware can't
|
|
* support it, so return -EINVAL.
|
|
* != 0 NULL msi-map does not exist, use built-in MSI
|
|
* 0 != NULL Get correct streamID from RID
|
|
* != 0 != NULL Invalid combination
|
|
*/
|
|
if (!err_m && !target)
|
|
return -EINVAL;
|
|
else if (target)
|
|
of_node_put(target); /* Find streamID map entry for RID in msi-map */
|
|
|
|
/*
|
|
* msi-map iommu-map
|
|
* N N DWC MSI Ctrl
|
|
* Y Y ITS + SMMU, require the same SID
|
|
* Y N ITS
|
|
* N Y DWC MSI Ctrl + SMMU
|
|
*/
|
|
if (err_i && err_m)
|
|
return 0;
|
|
|
|
if (!err_i && !err_m) {
|
|
/*
|
|
* Glue Layer
|
|
* <==========>
|
|
* ┌─────┐ ┌──────────┐
|
|
* │ LUT │ 6-bit streamID │ │
|
|
* │ │─────────────────►│ MSI │
|
|
* └─────┘ 2-bit ctrl ID │ │
|
|
* ┌───────────►│ │
|
|
* (i.MX95) │ │ │
|
|
* 00 PCIe0 │ │ │
|
|
* 01 ENETC │ │ │
|
|
* 10 PCIe1 │ │ │
|
|
* │ └──────────┘
|
|
* The MSI glue layer auto adds 2 bits controller ID ahead of
|
|
* streamID, so mask these 2 bits to get streamID. The
|
|
* IOMMU glue layer doesn't do that.
|
|
*/
|
|
if (sid_i != (sid_m & IMX95_SID_MASK)) {
|
|
dev_err(dev, "iommu-map and msi-map entries mismatch!\n");
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
if (!err_i)
|
|
sid = sid_i;
|
|
else if (!err_m)
|
|
sid = sid_m & IMX95_SID_MASK;
|
|
|
|
return imx_pcie_add_lut(imx_pcie, rid, sid);
|
|
}
|
|
|
|
static void imx_pcie_disable_device(struct pci_host_bridge *bridge,
|
|
struct pci_dev *pdev)
|
|
{
|
|
struct imx_pcie *imx_pcie;
|
|
|
|
imx_pcie = to_imx_pcie(to_dw_pcie_from_pp(bridge->sysdata));
|
|
imx_pcie_remove_lut(imx_pcie, pci_dev_id(pdev));
|
|
}
|
|
|
|
static int imx_pcie_host_init(struct dw_pcie_rp *pp)
|
|
{
|
|
struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
|
|
struct device *dev = pci->dev;
|
|
struct imx_pcie *imx_pcie = to_imx_pcie(pci);
|
|
int ret;
|
|
|
|
if (imx_pcie->vpcie) {
|
|
ret = regulator_enable(imx_pcie->vpcie);
|
|
if (ret) {
|
|
dev_err(dev, "failed to enable vpcie regulator: %d\n",
|
|
ret);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
if (pp->bridge && imx_check_flag(imx_pcie, IMX_PCIE_FLAG_HAS_LUT)) {
|
|
pp->bridge->enable_device = imx_pcie_enable_device;
|
|
pp->bridge->disable_device = imx_pcie_disable_device;
|
|
}
|
|
|
|
imx_pcie_assert_core_reset(imx_pcie);
|
|
|
|
if (imx_pcie->drvdata->init_phy)
|
|
imx_pcie->drvdata->init_phy(imx_pcie);
|
|
|
|
imx_pcie_configure_type(imx_pcie);
|
|
|
|
ret = imx_pcie_clk_enable(imx_pcie);
|
|
if (ret) {
|
|
dev_err(dev, "unable to enable pcie clocks: %d\n", ret);
|
|
goto err_reg_disable;
|
|
}
|
|
|
|
if (imx_pcie->phy) {
|
|
ret = phy_init(imx_pcie->phy);
|
|
if (ret) {
|
|
dev_err(dev, "pcie PHY power up failed\n");
|
|
goto err_clk_disable;
|
|
}
|
|
|
|
ret = phy_set_mode_ext(imx_pcie->phy, PHY_MODE_PCIE,
|
|
imx_pcie->drvdata->mode == DW_PCIE_EP_TYPE ?
|
|
PHY_MODE_PCIE_EP : PHY_MODE_PCIE_RC);
|
|
if (ret) {
|
|
dev_err(dev, "unable to set PCIe PHY mode\n");
|
|
goto err_phy_exit;
|
|
}
|
|
|
|
ret = phy_power_on(imx_pcie->phy);
|
|
if (ret) {
|
|
dev_err(dev, "waiting for PHY ready timeout!\n");
|
|
goto err_phy_exit;
|
|
}
|
|
}
|
|
|
|
ret = imx_pcie_deassert_core_reset(imx_pcie);
|
|
if (ret < 0) {
|
|
dev_err(dev, "pcie deassert core reset failed: %d\n", ret);
|
|
goto err_phy_off;
|
|
}
|
|
|
|
if (imx_pcie->drvdata->wait_pll_lock) {
|
|
ret = imx_pcie->drvdata->wait_pll_lock(imx_pcie);
|
|
if (ret < 0)
|
|
goto err_phy_off;
|
|
}
|
|
|
|
imx_setup_phy_mpll(imx_pcie);
|
|
|
|
return 0;
|
|
|
|
err_phy_off:
|
|
phy_power_off(imx_pcie->phy);
|
|
err_phy_exit:
|
|
phy_exit(imx_pcie->phy);
|
|
err_clk_disable:
|
|
imx_pcie_clk_disable(imx_pcie);
|
|
err_reg_disable:
|
|
if (imx_pcie->vpcie)
|
|
regulator_disable(imx_pcie->vpcie);
|
|
return ret;
|
|
}
|
|
|
|
static void imx_pcie_host_exit(struct dw_pcie_rp *pp)
|
|
{
|
|
struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
|
|
struct imx_pcie *imx_pcie = to_imx_pcie(pci);
|
|
|
|
if (imx_pcie->phy) {
|
|
if (phy_power_off(imx_pcie->phy))
|
|
dev_err(pci->dev, "unable to power off PHY\n");
|
|
phy_exit(imx_pcie->phy);
|
|
}
|
|
imx_pcie_clk_disable(imx_pcie);
|
|
|
|
if (imx_pcie->vpcie)
|
|
regulator_disable(imx_pcie->vpcie);
|
|
}
|
|
|
|
static void imx_pcie_host_post_init(struct dw_pcie_rp *pp)
|
|
{
|
|
struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
|
|
struct imx_pcie *imx_pcie = to_imx_pcie(pci);
|
|
u32 val;
|
|
|
|
if (imx_pcie->drvdata->flags & IMX_PCIE_FLAG_8GT_ECN_ERR051586) {
|
|
/*
|
|
* ERR051586: Compliance with 8GT/s Receiver Impedance ECN
|
|
*
|
|
* The default value of GEN3_RELATED_OFF[GEN3_ZRXDC_NONCOMPL]
|
|
* is 1 which makes receiver non-compliant with the ZRX-DC
|
|
* parameter for 2.5 GT/s when operating at 8 GT/s or higher.
|
|
* It causes unnecessary timeout in L1.
|
|
*
|
|
* Workaround: Program GEN3_RELATED_OFF[GEN3_ZRXDC_NONCOMPL]
|
|
* to 0.
|
|
*/
|
|
dw_pcie_dbi_ro_wr_en(pci);
|
|
val = dw_pcie_readl_dbi(pci, GEN3_RELATED_OFF);
|
|
val &= ~GEN3_RELATED_OFF_GEN3_ZRXDC_NONCOMPL;
|
|
dw_pcie_writel_dbi(pci, GEN3_RELATED_OFF, val);
|
|
dw_pcie_dbi_ro_wr_dis(pci);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* In old DWC implementations, PCIE_ATU_INHIBIT_PAYLOAD in iATU Ctrl2
|
|
* register is reserved, so the generic DWC implementation of sending the
|
|
* PME_Turn_Off message using a dummy MMIO write cannot be used.
|
|
*/
|
|
static void imx_pcie_pme_turn_off(struct dw_pcie_rp *pp)
|
|
{
|
|
struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
|
|
struct imx_pcie *imx_pcie = to_imx_pcie(pci);
|
|
|
|
regmap_set_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR12, IMX6SX_GPR12_PCIE_PM_TURN_OFF);
|
|
regmap_clear_bits(imx_pcie->iomuxc_gpr, IOMUXC_GPR12, IMX6SX_GPR12_PCIE_PM_TURN_OFF);
|
|
|
|
usleep_range(PCIE_PME_TO_L2_TIMEOUT_US/10, PCIE_PME_TO_L2_TIMEOUT_US);
|
|
}
|
|
|
|
static const struct dw_pcie_host_ops imx_pcie_host_ops = {
|
|
.init = imx_pcie_host_init,
|
|
.deinit = imx_pcie_host_exit,
|
|
.pme_turn_off = imx_pcie_pme_turn_off,
|
|
};
|
|
|
|
static const struct dw_pcie_host_ops imx_pcie_host_dw_pme_ops = {
|
|
.init = imx_pcie_host_init,
|
|
.deinit = imx_pcie_host_exit,
|
|
.post_init = imx_pcie_host_post_init,
|
|
};
|
|
|
|
static const struct dw_pcie_ops dw_pcie_ops = {
|
|
.start_link = imx_pcie_start_link,
|
|
.stop_link = imx_pcie_stop_link,
|
|
};
|
|
|
|
static void imx_pcie_ep_init(struct dw_pcie_ep *ep)
|
|
{
|
|
enum pci_barno bar;
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
|
|
for (bar = BAR_0; bar <= BAR_5; bar++)
|
|
dw_pcie_ep_reset_bar(pci, bar);
|
|
}
|
|
|
|
static int imx_pcie_ep_raise_irq(struct dw_pcie_ep *ep, u8 func_no,
|
|
unsigned int type, u16 interrupt_num)
|
|
{
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
|
|
switch (type) {
|
|
case PCI_IRQ_INTX:
|
|
return dw_pcie_ep_raise_intx_irq(ep, func_no);
|
|
case PCI_IRQ_MSI:
|
|
return dw_pcie_ep_raise_msi_irq(ep, func_no, interrupt_num);
|
|
case PCI_IRQ_MSIX:
|
|
return dw_pcie_ep_raise_msix_irq(ep, func_no, interrupt_num);
|
|
default:
|
|
dev_err(pci->dev, "UNKNOWN IRQ type\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct pci_epc_features imx8m_pcie_epc_features = {
|
|
.linkup_notifier = false,
|
|
.msi_capable = true,
|
|
.msix_capable = false,
|
|
.bar[BAR_1] = { .type = BAR_RESERVED, },
|
|
.bar[BAR_3] = { .type = BAR_RESERVED, },
|
|
.align = SZ_64K,
|
|
};
|
|
|
|
static const struct pci_epc_features imx8q_pcie_epc_features = {
|
|
.linkup_notifier = false,
|
|
.msi_capable = true,
|
|
.msix_capable = false,
|
|
.bar[BAR_1] = { .type = BAR_RESERVED, },
|
|
.bar[BAR_3] = { .type = BAR_RESERVED, },
|
|
.bar[BAR_5] = { .type = BAR_RESERVED, },
|
|
.align = SZ_64K,
|
|
};
|
|
|
|
/*
|
|
* | Default | Default | Default | BAR Sizing
|
|
* BAR# | Enable? | Type | Size | Scheme
|
|
* =======================================================
|
|
* BAR0 | Enable | 64-bit | 1 MB | Programmable Size
|
|
* BAR1 | Disable | 32-bit | 64 KB | Fixed Size
|
|
* (BAR1 should be disabled if BAR0 is 64-bit)
|
|
* BAR2 | Enable | 32-bit | 1 MB | Programmable Size
|
|
* BAR3 | Enable | 32-bit | 64 KB | Programmable Size
|
|
* BAR4 | Enable | 32-bit | 1 MB | Programmable Size
|
|
* BAR5 | Enable | 32-bit | 64 KB | Programmable Size
|
|
*/
|
|
static const struct pci_epc_features imx95_pcie_epc_features = {
|
|
.msi_capable = true,
|
|
.bar[BAR_1] = { .type = BAR_FIXED, .fixed_size = SZ_64K, },
|
|
.align = SZ_4K,
|
|
};
|
|
|
|
static const struct pci_epc_features*
|
|
imx_pcie_ep_get_features(struct dw_pcie_ep *ep)
|
|
{
|
|
struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
|
|
struct imx_pcie *imx_pcie = to_imx_pcie(pci);
|
|
|
|
return imx_pcie->drvdata->epc_features;
|
|
}
|
|
|
|
static const struct dw_pcie_ep_ops pcie_ep_ops = {
|
|
.init = imx_pcie_ep_init,
|
|
.raise_irq = imx_pcie_ep_raise_irq,
|
|
.get_features = imx_pcie_ep_get_features,
|
|
};
|
|
|
|
static int imx_add_pcie_ep(struct imx_pcie *imx_pcie,
|
|
struct platform_device *pdev)
|
|
{
|
|
int ret;
|
|
struct dw_pcie_ep *ep;
|
|
struct dw_pcie *pci = imx_pcie->pci;
|
|
struct dw_pcie_rp *pp = &pci->pp;
|
|
struct device *dev = pci->dev;
|
|
|
|
imx_pcie_host_init(pp);
|
|
ep = &pci->ep;
|
|
ep->ops = &pcie_ep_ops;
|
|
|
|
if (imx_check_flag(imx_pcie, IMX_PCIE_FLAG_SUPPORT_64BIT))
|
|
dma_set_mask_and_coherent(dev, DMA_BIT_MASK(64));
|
|
|
|
ep->page_size = imx_pcie->drvdata->epc_features->align;
|
|
|
|
ret = dw_pcie_ep_init(ep);
|
|
if (ret) {
|
|
dev_err(dev, "failed to initialize endpoint\n");
|
|
return ret;
|
|
}
|
|
imx_pcie_host_post_init(pp);
|
|
|
|
ret = dw_pcie_ep_init_registers(ep);
|
|
if (ret) {
|
|
dev_err(dev, "Failed to initialize DWC endpoint registers\n");
|
|
dw_pcie_ep_deinit(ep);
|
|
return ret;
|
|
}
|
|
|
|
pci_epc_init_notify(ep->epc);
|
|
|
|
/* Start LTSSM. */
|
|
imx_pcie_ltssm_enable(dev);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void imx_pcie_msi_save_restore(struct imx_pcie *imx_pcie, bool save)
|
|
{
|
|
u8 offset;
|
|
u16 val;
|
|
struct dw_pcie *pci = imx_pcie->pci;
|
|
|
|
if (pci_msi_enabled()) {
|
|
offset = dw_pcie_find_capability(pci, PCI_CAP_ID_MSI);
|
|
if (save) {
|
|
val = dw_pcie_readw_dbi(pci, offset + PCI_MSI_FLAGS);
|
|
imx_pcie->msi_ctrl = val;
|
|
} else {
|
|
dw_pcie_dbi_ro_wr_en(pci);
|
|
val = imx_pcie->msi_ctrl;
|
|
dw_pcie_writew_dbi(pci, offset + PCI_MSI_FLAGS, val);
|
|
dw_pcie_dbi_ro_wr_dis(pci);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void imx_pcie_lut_save(struct imx_pcie *imx_pcie)
|
|
{
|
|
u32 data1, data2;
|
|
int i;
|
|
|
|
for (i = 0; i < IMX95_MAX_LUT; i++) {
|
|
regmap_write(imx_pcie->iomuxc_gpr, IMX95_PE0_LUT_ACSCTRL,
|
|
IMX95_PEO_LUT_RWA | i);
|
|
regmap_read(imx_pcie->iomuxc_gpr, IMX95_PE0_LUT_DATA1, &data1);
|
|
regmap_read(imx_pcie->iomuxc_gpr, IMX95_PE0_LUT_DATA2, &data2);
|
|
if (data1 & IMX95_PE0_LUT_VLD) {
|
|
imx_pcie->luts[i].data1 = data1;
|
|
imx_pcie->luts[i].data2 = data2;
|
|
} else {
|
|
imx_pcie->luts[i].data1 = 0;
|
|
imx_pcie->luts[i].data2 = 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void imx_pcie_lut_restore(struct imx_pcie *imx_pcie)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < IMX95_MAX_LUT; i++) {
|
|
if ((imx_pcie->luts[i].data1 & IMX95_PE0_LUT_VLD) == 0)
|
|
continue;
|
|
|
|
regmap_write(imx_pcie->iomuxc_gpr, IMX95_PE0_LUT_DATA1,
|
|
imx_pcie->luts[i].data1);
|
|
regmap_write(imx_pcie->iomuxc_gpr, IMX95_PE0_LUT_DATA2,
|
|
imx_pcie->luts[i].data2);
|
|
regmap_write(imx_pcie->iomuxc_gpr, IMX95_PE0_LUT_ACSCTRL, i);
|
|
}
|
|
}
|
|
|
|
static int imx_pcie_suspend_noirq(struct device *dev)
|
|
{
|
|
struct imx_pcie *imx_pcie = dev_get_drvdata(dev);
|
|
|
|
if (!(imx_pcie->drvdata->flags & IMX_PCIE_FLAG_SUPPORTS_SUSPEND))
|
|
return 0;
|
|
|
|
imx_pcie_msi_save_restore(imx_pcie, true);
|
|
if (imx_check_flag(imx_pcie, IMX_PCIE_FLAG_HAS_LUT))
|
|
imx_pcie_lut_save(imx_pcie);
|
|
if (imx_check_flag(imx_pcie, IMX_PCIE_FLAG_BROKEN_SUSPEND)) {
|
|
/*
|
|
* The minimum for a workaround would be to set PERST# and to
|
|
* set the PCIE_TEST_PD flag. However, we can also disable the
|
|
* clock which saves some power.
|
|
*/
|
|
imx_pcie_assert_core_reset(imx_pcie);
|
|
imx_pcie->drvdata->enable_ref_clk(imx_pcie, false);
|
|
} else {
|
|
return dw_pcie_suspend_noirq(imx_pcie->pci);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int imx_pcie_resume_noirq(struct device *dev)
|
|
{
|
|
int ret;
|
|
struct imx_pcie *imx_pcie = dev_get_drvdata(dev);
|
|
|
|
if (!(imx_pcie->drvdata->flags & IMX_PCIE_FLAG_SUPPORTS_SUSPEND))
|
|
return 0;
|
|
|
|
if (imx_check_flag(imx_pcie, IMX_PCIE_FLAG_BROKEN_SUSPEND)) {
|
|
ret = imx_pcie->drvdata->enable_ref_clk(imx_pcie, true);
|
|
if (ret)
|
|
return ret;
|
|
ret = imx_pcie_deassert_core_reset(imx_pcie);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/*
|
|
* Using PCIE_TEST_PD seems to disable MSI and powers down the
|
|
* root complex. This is why we have to setup the rc again and
|
|
* why we have to restore the MSI register.
|
|
*/
|
|
ret = dw_pcie_setup_rc(&imx_pcie->pci->pp);
|
|
if (ret)
|
|
return ret;
|
|
} else {
|
|
ret = dw_pcie_resume_noirq(imx_pcie->pci);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
if (imx_check_flag(imx_pcie, IMX_PCIE_FLAG_HAS_LUT))
|
|
imx_pcie_lut_restore(imx_pcie);
|
|
imx_pcie_msi_save_restore(imx_pcie, false);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct dev_pm_ops imx_pcie_pm_ops = {
|
|
NOIRQ_SYSTEM_SLEEP_PM_OPS(imx_pcie_suspend_noirq,
|
|
imx_pcie_resume_noirq)
|
|
};
|
|
|
|
static int imx_pcie_probe(struct platform_device *pdev)
|
|
{
|
|
struct device *dev = &pdev->dev;
|
|
struct dw_pcie *pci;
|
|
struct imx_pcie *imx_pcie;
|
|
struct device_node *np;
|
|
struct device_node *node = dev->of_node;
|
|
int ret, domain;
|
|
u16 val;
|
|
|
|
imx_pcie = devm_kzalloc(dev, sizeof(*imx_pcie), GFP_KERNEL);
|
|
if (!imx_pcie)
|
|
return -ENOMEM;
|
|
|
|
pci = devm_kzalloc(dev, sizeof(*pci), GFP_KERNEL);
|
|
if (!pci)
|
|
return -ENOMEM;
|
|
|
|
pci->dev = dev;
|
|
pci->ops = &dw_pcie_ops;
|
|
|
|
imx_pcie->pci = pci;
|
|
imx_pcie->drvdata = of_device_get_match_data(dev);
|
|
|
|
mutex_init(&imx_pcie->lock);
|
|
|
|
if (imx_pcie->drvdata->ops)
|
|
pci->pp.ops = imx_pcie->drvdata->ops;
|
|
else
|
|
pci->pp.ops = &imx_pcie_host_dw_pme_ops;
|
|
|
|
/* Find the PHY if one is defined, only imx7d uses it */
|
|
np = of_parse_phandle(node, "fsl,imx7d-pcie-phy", 0);
|
|
if (np) {
|
|
struct resource res;
|
|
|
|
ret = of_address_to_resource(np, 0, &res);
|
|
if (ret) {
|
|
dev_err(dev, "Unable to map PCIe PHY\n");
|
|
return ret;
|
|
}
|
|
imx_pcie->phy_base = devm_ioremap_resource(dev, &res);
|
|
if (IS_ERR(imx_pcie->phy_base))
|
|
return PTR_ERR(imx_pcie->phy_base);
|
|
}
|
|
|
|
/* Fetch GPIOs */
|
|
imx_pcie->reset_gpiod = devm_gpiod_get_optional(dev, "reset", GPIOD_OUT_HIGH);
|
|
if (IS_ERR(imx_pcie->reset_gpiod))
|
|
return dev_err_probe(dev, PTR_ERR(imx_pcie->reset_gpiod),
|
|
"unable to get reset gpio\n");
|
|
gpiod_set_consumer_name(imx_pcie->reset_gpiod, "PCIe reset");
|
|
|
|
/* Fetch clocks */
|
|
imx_pcie->num_clks = devm_clk_bulk_get_all(dev, &imx_pcie->clks);
|
|
if (imx_pcie->num_clks < 0)
|
|
return dev_err_probe(dev, imx_pcie->num_clks,
|
|
"failed to get clocks\n");
|
|
|
|
if (imx_check_flag(imx_pcie, IMX_PCIE_FLAG_HAS_PHYDRV)) {
|
|
imx_pcie->phy = devm_phy_get(dev, "pcie-phy");
|
|
if (IS_ERR(imx_pcie->phy))
|
|
return dev_err_probe(dev, PTR_ERR(imx_pcie->phy),
|
|
"failed to get pcie phy\n");
|
|
}
|
|
|
|
if (imx_check_flag(imx_pcie, IMX_PCIE_FLAG_HAS_APP_RESET)) {
|
|
imx_pcie->apps_reset = devm_reset_control_get_exclusive(dev, "apps");
|
|
if (IS_ERR(imx_pcie->apps_reset))
|
|
return dev_err_probe(dev, PTR_ERR(imx_pcie->apps_reset),
|
|
"failed to get pcie apps reset control\n");
|
|
}
|
|
|
|
if (imx_check_flag(imx_pcie, IMX_PCIE_FLAG_HAS_PHY_RESET)) {
|
|
imx_pcie->pciephy_reset = devm_reset_control_get_exclusive(dev, "pciephy");
|
|
if (IS_ERR(imx_pcie->pciephy_reset))
|
|
return dev_err_probe(dev, PTR_ERR(imx_pcie->pciephy_reset),
|
|
"Failed to get PCIEPHY reset control\n");
|
|
}
|
|
|
|
switch (imx_pcie->drvdata->variant) {
|
|
case IMX8MQ:
|
|
case IMX8MQ_EP:
|
|
domain = of_get_pci_domain_nr(node);
|
|
if (domain < 0 || domain > 1)
|
|
return dev_err_probe(dev, -ENODEV, "no \"linux,pci-domain\" property in devicetree\n");
|
|
|
|
imx_pcie->controller_id = domain;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
if (imx_pcie->drvdata->gpr) {
|
|
/* Grab GPR config register range */
|
|
imx_pcie->iomuxc_gpr =
|
|
syscon_regmap_lookup_by_compatible(imx_pcie->drvdata->gpr);
|
|
if (IS_ERR(imx_pcie->iomuxc_gpr))
|
|
return dev_err_probe(dev, PTR_ERR(imx_pcie->iomuxc_gpr),
|
|
"unable to find iomuxc registers\n");
|
|
}
|
|
|
|
if (imx_check_flag(imx_pcie, IMX_PCIE_FLAG_HAS_SERDES)) {
|
|
void __iomem *off = devm_platform_ioremap_resource_byname(pdev, "app");
|
|
|
|
if (IS_ERR(off))
|
|
return dev_err_probe(dev, PTR_ERR(off),
|
|
"unable to find serdes registers\n");
|
|
|
|
static const struct regmap_config regmap_config = {
|
|
.reg_bits = 32,
|
|
.val_bits = 32,
|
|
.reg_stride = 4,
|
|
};
|
|
|
|
imx_pcie->iomuxc_gpr = devm_regmap_init_mmio(dev, off, ®map_config);
|
|
if (IS_ERR(imx_pcie->iomuxc_gpr))
|
|
return dev_err_probe(dev, PTR_ERR(imx_pcie->iomuxc_gpr),
|
|
"unable to find iomuxc registers\n");
|
|
}
|
|
|
|
/* Grab PCIe PHY Tx Settings */
|
|
if (of_property_read_u32(node, "fsl,tx-deemph-gen1",
|
|
&imx_pcie->tx_deemph_gen1))
|
|
imx_pcie->tx_deemph_gen1 = 0;
|
|
|
|
if (of_property_read_u32(node, "fsl,tx-deemph-gen2-3p5db",
|
|
&imx_pcie->tx_deemph_gen2_3p5db))
|
|
imx_pcie->tx_deemph_gen2_3p5db = 0;
|
|
|
|
if (of_property_read_u32(node, "fsl,tx-deemph-gen2-6db",
|
|
&imx_pcie->tx_deemph_gen2_6db))
|
|
imx_pcie->tx_deemph_gen2_6db = 20;
|
|
|
|
if (of_property_read_u32(node, "fsl,tx-swing-full",
|
|
&imx_pcie->tx_swing_full))
|
|
imx_pcie->tx_swing_full = 127;
|
|
|
|
if (of_property_read_u32(node, "fsl,tx-swing-low",
|
|
&imx_pcie->tx_swing_low))
|
|
imx_pcie->tx_swing_low = 127;
|
|
|
|
/* Limit link speed */
|
|
pci->max_link_speed = 1;
|
|
of_property_read_u32(node, "fsl,max-link-speed", &pci->max_link_speed);
|
|
|
|
imx_pcie->vpcie = devm_regulator_get_optional(&pdev->dev, "vpcie");
|
|
if (IS_ERR(imx_pcie->vpcie)) {
|
|
if (PTR_ERR(imx_pcie->vpcie) != -ENODEV)
|
|
return PTR_ERR(imx_pcie->vpcie);
|
|
imx_pcie->vpcie = NULL;
|
|
}
|
|
|
|
imx_pcie->vph = devm_regulator_get_optional(&pdev->dev, "vph");
|
|
if (IS_ERR(imx_pcie->vph)) {
|
|
if (PTR_ERR(imx_pcie->vph) != -ENODEV)
|
|
return PTR_ERR(imx_pcie->vph);
|
|
imx_pcie->vph = NULL;
|
|
}
|
|
|
|
platform_set_drvdata(pdev, imx_pcie);
|
|
|
|
ret = imx_pcie_attach_pd(dev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
pci->use_parent_dt_ranges = true;
|
|
if (imx_pcie->drvdata->mode == DW_PCIE_EP_TYPE) {
|
|
ret = imx_add_pcie_ep(imx_pcie, pdev);
|
|
if (ret < 0)
|
|
return ret;
|
|
} else {
|
|
pci->pp.use_atu_msg = true;
|
|
ret = dw_pcie_host_init(&pci->pp);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
if (pci_msi_enabled()) {
|
|
u8 offset = dw_pcie_find_capability(pci, PCI_CAP_ID_MSI);
|
|
|
|
val = dw_pcie_readw_dbi(pci, offset + PCI_MSI_FLAGS);
|
|
val |= PCI_MSI_FLAGS_ENABLE;
|
|
dw_pcie_writew_dbi(pci, offset + PCI_MSI_FLAGS, val);
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void imx_pcie_shutdown(struct platform_device *pdev)
|
|
{
|
|
struct imx_pcie *imx_pcie = platform_get_drvdata(pdev);
|
|
|
|
/* bring down link, so bootloader gets clean state in case of reboot */
|
|
imx_pcie_assert_core_reset(imx_pcie);
|
|
}
|
|
|
|
static const struct imx_pcie_drvdata drvdata[] = {
|
|
[IMX6Q] = {
|
|
.variant = IMX6Q,
|
|
.flags = IMX_PCIE_FLAG_IMX_PHY |
|
|
IMX_PCIE_FLAG_SPEED_CHANGE_WORKAROUND |
|
|
IMX_PCIE_FLAG_BROKEN_SUSPEND |
|
|
IMX_PCIE_FLAG_SUPPORTS_SUSPEND,
|
|
.dbi_length = 0x200,
|
|
.gpr = "fsl,imx6q-iomuxc-gpr",
|
|
.ltssm_off = IOMUXC_GPR12,
|
|
.ltssm_mask = IMX6Q_GPR12_PCIE_CTL_2,
|
|
.mode_off[0] = IOMUXC_GPR12,
|
|
.mode_mask[0] = IMX6Q_GPR12_DEVICE_TYPE,
|
|
.init_phy = imx_pcie_init_phy,
|
|
.enable_ref_clk = imx6q_pcie_enable_ref_clk,
|
|
.core_reset = imx6q_pcie_core_reset,
|
|
},
|
|
[IMX6SX] = {
|
|
.variant = IMX6SX,
|
|
.flags = IMX_PCIE_FLAG_IMX_PHY |
|
|
IMX_PCIE_FLAG_SPEED_CHANGE_WORKAROUND |
|
|
IMX_PCIE_FLAG_SUPPORTS_SUSPEND,
|
|
.gpr = "fsl,imx6q-iomuxc-gpr",
|
|
.ltssm_off = IOMUXC_GPR12,
|
|
.ltssm_mask = IMX6Q_GPR12_PCIE_CTL_2,
|
|
.mode_off[0] = IOMUXC_GPR12,
|
|
.mode_mask[0] = IMX6Q_GPR12_DEVICE_TYPE,
|
|
.init_phy = imx6sx_pcie_init_phy,
|
|
.enable_ref_clk = imx6sx_pcie_enable_ref_clk,
|
|
.core_reset = imx6sx_pcie_core_reset,
|
|
.ops = &imx_pcie_host_ops,
|
|
},
|
|
[IMX6QP] = {
|
|
.variant = IMX6QP,
|
|
.flags = IMX_PCIE_FLAG_IMX_PHY |
|
|
IMX_PCIE_FLAG_SPEED_CHANGE_WORKAROUND |
|
|
IMX_PCIE_FLAG_SUPPORTS_SUSPEND,
|
|
.dbi_length = 0x200,
|
|
.gpr = "fsl,imx6q-iomuxc-gpr",
|
|
.ltssm_off = IOMUXC_GPR12,
|
|
.ltssm_mask = IMX6Q_GPR12_PCIE_CTL_2,
|
|
.mode_off[0] = IOMUXC_GPR12,
|
|
.mode_mask[0] = IMX6Q_GPR12_DEVICE_TYPE,
|
|
.init_phy = imx_pcie_init_phy,
|
|
.enable_ref_clk = imx6q_pcie_enable_ref_clk,
|
|
.core_reset = imx6qp_pcie_core_reset,
|
|
.ops = &imx_pcie_host_ops,
|
|
},
|
|
[IMX7D] = {
|
|
.variant = IMX7D,
|
|
.flags = IMX_PCIE_FLAG_SUPPORTS_SUSPEND |
|
|
IMX_PCIE_FLAG_HAS_APP_RESET |
|
|
IMX_PCIE_FLAG_HAS_PHY_RESET,
|
|
.gpr = "fsl,imx7d-iomuxc-gpr",
|
|
.mode_off[0] = IOMUXC_GPR12,
|
|
.mode_mask[0] = IMX6Q_GPR12_DEVICE_TYPE,
|
|
.enable_ref_clk = imx7d_pcie_enable_ref_clk,
|
|
.core_reset = imx7d_pcie_core_reset,
|
|
},
|
|
[IMX8MQ] = {
|
|
.variant = IMX8MQ,
|
|
.flags = IMX_PCIE_FLAG_HAS_APP_RESET |
|
|
IMX_PCIE_FLAG_HAS_PHY_RESET |
|
|
IMX_PCIE_FLAG_SUPPORTS_SUSPEND,
|
|
.gpr = "fsl,imx8mq-iomuxc-gpr",
|
|
.mode_off[0] = IOMUXC_GPR12,
|
|
.mode_mask[0] = IMX6Q_GPR12_DEVICE_TYPE,
|
|
.mode_off[1] = IOMUXC_GPR12,
|
|
.mode_mask[1] = IMX8MQ_GPR12_PCIE2_CTRL_DEVICE_TYPE,
|
|
.init_phy = imx8mq_pcie_init_phy,
|
|
.enable_ref_clk = imx8mm_pcie_enable_ref_clk,
|
|
},
|
|
[IMX8MM] = {
|
|
.variant = IMX8MM,
|
|
.flags = IMX_PCIE_FLAG_SUPPORTS_SUSPEND |
|
|
IMX_PCIE_FLAG_HAS_PHYDRV |
|
|
IMX_PCIE_FLAG_HAS_APP_RESET,
|
|
.gpr = "fsl,imx8mm-iomuxc-gpr",
|
|
.mode_off[0] = IOMUXC_GPR12,
|
|
.mode_mask[0] = IMX6Q_GPR12_DEVICE_TYPE,
|
|
.enable_ref_clk = imx8mm_pcie_enable_ref_clk,
|
|
},
|
|
[IMX8MP] = {
|
|
.variant = IMX8MP,
|
|
.flags = IMX_PCIE_FLAG_SUPPORTS_SUSPEND |
|
|
IMX_PCIE_FLAG_HAS_PHYDRV |
|
|
IMX_PCIE_FLAG_HAS_APP_RESET,
|
|
.gpr = "fsl,imx8mp-iomuxc-gpr",
|
|
.mode_off[0] = IOMUXC_GPR12,
|
|
.mode_mask[0] = IMX6Q_GPR12_DEVICE_TYPE,
|
|
.enable_ref_clk = imx8mm_pcie_enable_ref_clk,
|
|
},
|
|
[IMX8Q] = {
|
|
.variant = IMX8Q,
|
|
.flags = IMX_PCIE_FLAG_HAS_PHYDRV |
|
|
IMX_PCIE_FLAG_CPU_ADDR_FIXUP |
|
|
IMX_PCIE_FLAG_SUPPORTS_SUSPEND,
|
|
},
|
|
[IMX95] = {
|
|
.variant = IMX95,
|
|
.flags = IMX_PCIE_FLAG_HAS_SERDES |
|
|
IMX_PCIE_FLAG_HAS_LUT |
|
|
IMX_PCIE_FLAG_8GT_ECN_ERR051586 |
|
|
IMX_PCIE_FLAG_SUPPORTS_SUSPEND,
|
|
.ltssm_off = IMX95_PE0_GEN_CTRL_3,
|
|
.ltssm_mask = IMX95_PCIE_LTSSM_EN,
|
|
.mode_off[0] = IMX95_PE0_GEN_CTRL_1,
|
|
.mode_mask[0] = IMX95_PCIE_DEVICE_TYPE,
|
|
.core_reset = imx95_pcie_core_reset,
|
|
.init_phy = imx95_pcie_init_phy,
|
|
.wait_pll_lock = imx95_pcie_wait_for_phy_pll_lock,
|
|
},
|
|
[IMX8MQ_EP] = {
|
|
.variant = IMX8MQ_EP,
|
|
.flags = IMX_PCIE_FLAG_HAS_APP_RESET |
|
|
IMX_PCIE_FLAG_HAS_PHY_RESET,
|
|
.mode = DW_PCIE_EP_TYPE,
|
|
.gpr = "fsl,imx8mq-iomuxc-gpr",
|
|
.mode_off[0] = IOMUXC_GPR12,
|
|
.mode_mask[0] = IMX6Q_GPR12_DEVICE_TYPE,
|
|
.mode_off[1] = IOMUXC_GPR12,
|
|
.mode_mask[1] = IMX8MQ_GPR12_PCIE2_CTRL_DEVICE_TYPE,
|
|
.epc_features = &imx8m_pcie_epc_features,
|
|
.init_phy = imx8mq_pcie_init_phy,
|
|
.enable_ref_clk = imx8mm_pcie_enable_ref_clk,
|
|
},
|
|
[IMX8MM_EP] = {
|
|
.variant = IMX8MM_EP,
|
|
.flags = IMX_PCIE_FLAG_HAS_APP_RESET |
|
|
IMX_PCIE_FLAG_HAS_PHYDRV,
|
|
.mode = DW_PCIE_EP_TYPE,
|
|
.gpr = "fsl,imx8mm-iomuxc-gpr",
|
|
.mode_off[0] = IOMUXC_GPR12,
|
|
.mode_mask[0] = IMX6Q_GPR12_DEVICE_TYPE,
|
|
.epc_features = &imx8m_pcie_epc_features,
|
|
.enable_ref_clk = imx8mm_pcie_enable_ref_clk,
|
|
},
|
|
[IMX8MP_EP] = {
|
|
.variant = IMX8MP_EP,
|
|
.flags = IMX_PCIE_FLAG_HAS_APP_RESET |
|
|
IMX_PCIE_FLAG_HAS_PHYDRV,
|
|
.mode = DW_PCIE_EP_TYPE,
|
|
.gpr = "fsl,imx8mp-iomuxc-gpr",
|
|
.mode_off[0] = IOMUXC_GPR12,
|
|
.mode_mask[0] = IMX6Q_GPR12_DEVICE_TYPE,
|
|
.epc_features = &imx8m_pcie_epc_features,
|
|
.enable_ref_clk = imx8mm_pcie_enable_ref_clk,
|
|
},
|
|
[IMX8Q_EP] = {
|
|
.variant = IMX8Q_EP,
|
|
.flags = IMX_PCIE_FLAG_HAS_PHYDRV,
|
|
.mode = DW_PCIE_EP_TYPE,
|
|
.epc_features = &imx8q_pcie_epc_features,
|
|
},
|
|
[IMX95_EP] = {
|
|
.variant = IMX95_EP,
|
|
.flags = IMX_PCIE_FLAG_HAS_SERDES |
|
|
IMX_PCIE_FLAG_8GT_ECN_ERR051586 |
|
|
IMX_PCIE_FLAG_SUPPORT_64BIT,
|
|
.ltssm_off = IMX95_PE0_GEN_CTRL_3,
|
|
.ltssm_mask = IMX95_PCIE_LTSSM_EN,
|
|
.mode_off[0] = IMX95_PE0_GEN_CTRL_1,
|
|
.mode_mask[0] = IMX95_PCIE_DEVICE_TYPE,
|
|
.init_phy = imx95_pcie_init_phy,
|
|
.core_reset = imx95_pcie_core_reset,
|
|
.wait_pll_lock = imx95_pcie_wait_for_phy_pll_lock,
|
|
.epc_features = &imx95_pcie_epc_features,
|
|
.mode = DW_PCIE_EP_TYPE,
|
|
},
|
|
};
|
|
|
|
static const struct of_device_id imx_pcie_of_match[] = {
|
|
{ .compatible = "fsl,imx6q-pcie", .data = &drvdata[IMX6Q], },
|
|
{ .compatible = "fsl,imx6sx-pcie", .data = &drvdata[IMX6SX], },
|
|
{ .compatible = "fsl,imx6qp-pcie", .data = &drvdata[IMX6QP], },
|
|
{ .compatible = "fsl,imx7d-pcie", .data = &drvdata[IMX7D], },
|
|
{ .compatible = "fsl,imx8mq-pcie", .data = &drvdata[IMX8MQ], },
|
|
{ .compatible = "fsl,imx8mm-pcie", .data = &drvdata[IMX8MM], },
|
|
{ .compatible = "fsl,imx8mp-pcie", .data = &drvdata[IMX8MP], },
|
|
{ .compatible = "fsl,imx8q-pcie", .data = &drvdata[IMX8Q], },
|
|
{ .compatible = "fsl,imx95-pcie", .data = &drvdata[IMX95], },
|
|
{ .compatible = "fsl,imx8mq-pcie-ep", .data = &drvdata[IMX8MQ_EP], },
|
|
{ .compatible = "fsl,imx8mm-pcie-ep", .data = &drvdata[IMX8MM_EP], },
|
|
{ .compatible = "fsl,imx8mp-pcie-ep", .data = &drvdata[IMX8MP_EP], },
|
|
{ .compatible = "fsl,imx8q-pcie-ep", .data = &drvdata[IMX8Q_EP], },
|
|
{ .compatible = "fsl,imx95-pcie-ep", .data = &drvdata[IMX95_EP], },
|
|
{},
|
|
};
|
|
|
|
static struct platform_driver imx_pcie_driver = {
|
|
.driver = {
|
|
.name = "imx6q-pcie",
|
|
.of_match_table = imx_pcie_of_match,
|
|
.suppress_bind_attrs = true,
|
|
.pm = &imx_pcie_pm_ops,
|
|
.probe_type = PROBE_PREFER_ASYNCHRONOUS,
|
|
},
|
|
.probe = imx_pcie_probe,
|
|
.shutdown = imx_pcie_shutdown,
|
|
};
|
|
|
|
static void imx_pcie_quirk(struct pci_dev *dev)
|
|
{
|
|
struct pci_bus *bus = dev->bus;
|
|
struct dw_pcie_rp *pp = bus->sysdata;
|
|
|
|
/* Bus parent is the PCI bridge, its parent is this platform driver */
|
|
if (!bus->dev.parent || !bus->dev.parent->parent)
|
|
return;
|
|
|
|
/* Make sure we only quirk devices associated with this driver */
|
|
if (bus->dev.parent->parent->driver != &imx_pcie_driver.driver)
|
|
return;
|
|
|
|
if (pci_is_root_bus(bus)) {
|
|
struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
|
|
struct imx_pcie *imx_pcie = to_imx_pcie(pci);
|
|
|
|
/*
|
|
* Limit config length to avoid the kernel reading beyond
|
|
* the register set and causing an abort on i.MX 6Quad
|
|
*/
|
|
if (imx_pcie->drvdata->dbi_length) {
|
|
dev->cfg_size = imx_pcie->drvdata->dbi_length;
|
|
dev_info(&dev->dev, "Limiting cfg_size to %d\n",
|
|
dev->cfg_size);
|
|
}
|
|
}
|
|
}
|
|
DECLARE_PCI_FIXUP_CLASS_HEADER(PCI_VENDOR_ID_SYNOPSYS, 0xabcd,
|
|
PCI_CLASS_BRIDGE_PCI, 8, imx_pcie_quirk);
|
|
|
|
static int __init imx_pcie_init(void)
|
|
{
|
|
#ifdef CONFIG_ARM
|
|
struct device_node *np;
|
|
|
|
np = of_find_matching_node(NULL, imx_pcie_of_match);
|
|
if (!np)
|
|
return -ENODEV;
|
|
of_node_put(np);
|
|
|
|
/*
|
|
* Since probe() can be deferred we need to make sure that
|
|
* hook_fault_code is not called after __init memory is freed
|
|
* by kernel and since imx6q_pcie_abort_handler() is a no-op,
|
|
* we can install the handler here without risking it
|
|
* accessing some uninitialized driver state.
|
|
*/
|
|
hook_fault_code(8, imx6q_pcie_abort_handler, SIGBUS, 0,
|
|
"external abort on non-linefetch");
|
|
#endif
|
|
|
|
return platform_driver_register(&imx_pcie_driver);
|
|
}
|
|
device_initcall(imx_pcie_init);
|