1713 строки
48 KiB
C
1713 строки
48 KiB
C
/*
|
|
* Rockchip AXI PCIe host controller driver
|
|
*
|
|
* Copyright (c) 2016 Rockchip, Inc.
|
|
*
|
|
* Author: Shawn Lin <shawn.lin@rock-chips.com>
|
|
* Wenrui Li <wenrui.li@rock-chips.com>
|
|
*
|
|
* Bits taken from Synopsys DesignWare Host controller driver and
|
|
* ARM PCI Host generic driver.
|
|
*
|
|
* This program is free software: you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation, either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*/
|
|
|
|
#include <linux/bitrev.h>
|
|
#include <linux/clk.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/gpio/consumer.h>
|
|
#include <linux/init.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/iopoll.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/irqchip/chained_irq.h>
|
|
#include <linux/irqdomain.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/mfd/syscon.h>
|
|
#include <linux/module.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/of_device.h>
|
|
#include <linux/of_pci.h>
|
|
#include <linux/of_platform.h>
|
|
#include <linux/of_irq.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/pci_ids.h>
|
|
#include <linux/phy/phy.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/reset.h>
|
|
#include <linux/regmap.h>
|
|
|
|
/*
|
|
* The upper 16 bits of PCIE_CLIENT_CONFIG are a write mask for the lower 16
|
|
* bits. This allows atomic updates of the register without locking.
|
|
*/
|
|
#define HIWORD_UPDATE(mask, val) (((mask) << 16) | (val))
|
|
#define HIWORD_UPDATE_BIT(val) HIWORD_UPDATE(val, val)
|
|
|
|
#define ENCODE_LANES(x) ((((x) >> 1) & 3) << 4)
|
|
#define MAX_LANE_NUM 4
|
|
|
|
#define PCIE_CLIENT_BASE 0x0
|
|
#define PCIE_CLIENT_CONFIG (PCIE_CLIENT_BASE + 0x00)
|
|
#define PCIE_CLIENT_CONF_ENABLE HIWORD_UPDATE_BIT(0x0001)
|
|
#define PCIE_CLIENT_LINK_TRAIN_ENABLE HIWORD_UPDATE_BIT(0x0002)
|
|
#define PCIE_CLIENT_ARI_ENABLE HIWORD_UPDATE_BIT(0x0008)
|
|
#define PCIE_CLIENT_CONF_LANE_NUM(x) HIWORD_UPDATE(0x0030, ENCODE_LANES(x))
|
|
#define PCIE_CLIENT_MODE_RC HIWORD_UPDATE_BIT(0x0040)
|
|
#define PCIE_CLIENT_GEN_SEL_1 HIWORD_UPDATE(0x0080, 0)
|
|
#define PCIE_CLIENT_GEN_SEL_2 HIWORD_UPDATE_BIT(0x0080)
|
|
#define PCIE_CLIENT_DEBUG_OUT_0 (PCIE_CLIENT_BASE + 0x3c)
|
|
#define PCIE_CLIENT_DEBUG_LTSSM_MASK GENMASK(5, 0)
|
|
#define PCIE_CLIENT_DEBUG_LTSSM_L1 0x18
|
|
#define PCIE_CLIENT_DEBUG_LTSSM_L2 0x19
|
|
#define PCIE_CLIENT_BASIC_STATUS1 (PCIE_CLIENT_BASE + 0x48)
|
|
#define PCIE_CLIENT_LINK_STATUS_UP 0x00300000
|
|
#define PCIE_CLIENT_LINK_STATUS_MASK 0x00300000
|
|
#define PCIE_CLIENT_INT_MASK (PCIE_CLIENT_BASE + 0x4c)
|
|
#define PCIE_CLIENT_INT_STATUS (PCIE_CLIENT_BASE + 0x50)
|
|
#define PCIE_CLIENT_INTR_MASK GENMASK(8, 5)
|
|
#define PCIE_CLIENT_INTR_SHIFT 5
|
|
#define PCIE_CLIENT_INT_LEGACY_DONE BIT(15)
|
|
#define PCIE_CLIENT_INT_MSG BIT(14)
|
|
#define PCIE_CLIENT_INT_HOT_RST BIT(13)
|
|
#define PCIE_CLIENT_INT_DPA BIT(12)
|
|
#define PCIE_CLIENT_INT_FATAL_ERR BIT(11)
|
|
#define PCIE_CLIENT_INT_NFATAL_ERR BIT(10)
|
|
#define PCIE_CLIENT_INT_CORR_ERR BIT(9)
|
|
#define PCIE_CLIENT_INT_INTD BIT(8)
|
|
#define PCIE_CLIENT_INT_INTC BIT(7)
|
|
#define PCIE_CLIENT_INT_INTB BIT(6)
|
|
#define PCIE_CLIENT_INT_INTA BIT(5)
|
|
#define PCIE_CLIENT_INT_LOCAL BIT(4)
|
|
#define PCIE_CLIENT_INT_UDMA BIT(3)
|
|
#define PCIE_CLIENT_INT_PHY BIT(2)
|
|
#define PCIE_CLIENT_INT_HOT_PLUG BIT(1)
|
|
#define PCIE_CLIENT_INT_PWR_STCG BIT(0)
|
|
|
|
#define PCIE_CLIENT_INT_LEGACY \
|
|
(PCIE_CLIENT_INT_INTA | PCIE_CLIENT_INT_INTB | \
|
|
PCIE_CLIENT_INT_INTC | PCIE_CLIENT_INT_INTD)
|
|
|
|
#define PCIE_CLIENT_INT_CLI \
|
|
(PCIE_CLIENT_INT_CORR_ERR | PCIE_CLIENT_INT_NFATAL_ERR | \
|
|
PCIE_CLIENT_INT_FATAL_ERR | PCIE_CLIENT_INT_DPA | \
|
|
PCIE_CLIENT_INT_HOT_RST | PCIE_CLIENT_INT_MSG | \
|
|
PCIE_CLIENT_INT_LEGACY_DONE | PCIE_CLIENT_INT_LEGACY | \
|
|
PCIE_CLIENT_INT_PHY)
|
|
|
|
#define PCIE_CORE_CTRL_MGMT_BASE 0x900000
|
|
#define PCIE_CORE_CTRL (PCIE_CORE_CTRL_MGMT_BASE + 0x000)
|
|
#define PCIE_CORE_PL_CONF_SPEED_5G 0x00000008
|
|
#define PCIE_CORE_PL_CONF_SPEED_MASK 0x00000018
|
|
#define PCIE_CORE_PL_CONF_LANE_MASK 0x00000006
|
|
#define PCIE_CORE_PL_CONF_LANE_SHIFT 1
|
|
#define PCIE_CORE_CTRL_PLC1 (PCIE_CORE_CTRL_MGMT_BASE + 0x004)
|
|
#define PCIE_CORE_CTRL_PLC1_FTS_MASK GENMASK(23, 8)
|
|
#define PCIE_CORE_CTRL_PLC1_FTS_SHIFT 8
|
|
#define PCIE_CORE_CTRL_PLC1_FTS_CNT 0xffff
|
|
#define PCIE_CORE_TXCREDIT_CFG1 (PCIE_CORE_CTRL_MGMT_BASE + 0x020)
|
|
#define PCIE_CORE_TXCREDIT_CFG1_MUI_MASK 0xFFFF0000
|
|
#define PCIE_CORE_TXCREDIT_CFG1_MUI_SHIFT 16
|
|
#define PCIE_CORE_TXCREDIT_CFG1_MUI_ENCODE(x) \
|
|
(((x) >> 3) << PCIE_CORE_TXCREDIT_CFG1_MUI_SHIFT)
|
|
#define PCIE_CORE_LANE_MAP (PCIE_CORE_CTRL_MGMT_BASE + 0x200)
|
|
#define PCIE_CORE_LANE_MAP_MASK 0x0000000f
|
|
#define PCIE_CORE_LANE_MAP_REVERSE BIT(16)
|
|
#define PCIE_CORE_INT_STATUS (PCIE_CORE_CTRL_MGMT_BASE + 0x20c)
|
|
#define PCIE_CORE_INT_PRFPE BIT(0)
|
|
#define PCIE_CORE_INT_CRFPE BIT(1)
|
|
#define PCIE_CORE_INT_RRPE BIT(2)
|
|
#define PCIE_CORE_INT_PRFO BIT(3)
|
|
#define PCIE_CORE_INT_CRFO BIT(4)
|
|
#define PCIE_CORE_INT_RT BIT(5)
|
|
#define PCIE_CORE_INT_RTR BIT(6)
|
|
#define PCIE_CORE_INT_PE BIT(7)
|
|
#define PCIE_CORE_INT_MTR BIT(8)
|
|
#define PCIE_CORE_INT_UCR BIT(9)
|
|
#define PCIE_CORE_INT_FCE BIT(10)
|
|
#define PCIE_CORE_INT_CT BIT(11)
|
|
#define PCIE_CORE_INT_UTC BIT(18)
|
|
#define PCIE_CORE_INT_MMVC BIT(19)
|
|
#define PCIE_CORE_CONFIG_VENDOR (PCIE_CORE_CTRL_MGMT_BASE + 0x44)
|
|
#define PCIE_CORE_INT_MASK (PCIE_CORE_CTRL_MGMT_BASE + 0x210)
|
|
#define PCIE_RC_BAR_CONF (PCIE_CORE_CTRL_MGMT_BASE + 0x300)
|
|
|
|
#define PCIE_CORE_INT \
|
|
(PCIE_CORE_INT_PRFPE | PCIE_CORE_INT_CRFPE | \
|
|
PCIE_CORE_INT_RRPE | PCIE_CORE_INT_CRFO | \
|
|
PCIE_CORE_INT_RT | PCIE_CORE_INT_RTR | \
|
|
PCIE_CORE_INT_PE | PCIE_CORE_INT_MTR | \
|
|
PCIE_CORE_INT_UCR | PCIE_CORE_INT_FCE | \
|
|
PCIE_CORE_INT_CT | PCIE_CORE_INT_UTC | \
|
|
PCIE_CORE_INT_MMVC)
|
|
|
|
#define PCIE_RC_CONFIG_NORMAL_BASE 0x800000
|
|
#define PCIE_RC_CONFIG_BASE 0xa00000
|
|
#define PCIE_RC_CONFIG_RID_CCR (PCIE_RC_CONFIG_BASE + 0x08)
|
|
#define PCIE_RC_CONFIG_SCC_SHIFT 16
|
|
#define PCIE_RC_CONFIG_DCR (PCIE_RC_CONFIG_BASE + 0xc4)
|
|
#define PCIE_RC_CONFIG_DCR_CSPL_SHIFT 18
|
|
#define PCIE_RC_CONFIG_DCR_CSPL_LIMIT 0xff
|
|
#define PCIE_RC_CONFIG_DCR_CPLS_SHIFT 26
|
|
#define PCIE_RC_CONFIG_DCSR (PCIE_RC_CONFIG_BASE + 0xc8)
|
|
#define PCIE_RC_CONFIG_DCSR_MPS_MASK GENMASK(7, 5)
|
|
#define PCIE_RC_CONFIG_DCSR_MPS_256 (0x1 << 5)
|
|
#define PCIE_RC_CONFIG_LINK_CAP (PCIE_RC_CONFIG_BASE + 0xcc)
|
|
#define PCIE_RC_CONFIG_LINK_CAP_L0S BIT(10)
|
|
#define PCIE_RC_CONFIG_LCS (PCIE_RC_CONFIG_BASE + 0xd0)
|
|
#define PCIE_RC_CONFIG_L1_SUBSTATE_CTRL2 (PCIE_RC_CONFIG_BASE + 0x90c)
|
|
#define PCIE_RC_CONFIG_THP_CAP (PCIE_RC_CONFIG_BASE + 0x274)
|
|
#define PCIE_RC_CONFIG_THP_CAP_NEXT_MASK GENMASK(31, 20)
|
|
|
|
#define PCIE_CORE_AXI_CONF_BASE 0xc00000
|
|
#define PCIE_CORE_OB_REGION_ADDR0 (PCIE_CORE_AXI_CONF_BASE + 0x0)
|
|
#define PCIE_CORE_OB_REGION_ADDR0_NUM_BITS 0x3f
|
|
#define PCIE_CORE_OB_REGION_ADDR0_LO_ADDR 0xffffff00
|
|
#define PCIE_CORE_OB_REGION_ADDR1 (PCIE_CORE_AXI_CONF_BASE + 0x4)
|
|
#define PCIE_CORE_OB_REGION_DESC0 (PCIE_CORE_AXI_CONF_BASE + 0x8)
|
|
#define PCIE_CORE_OB_REGION_DESC1 (PCIE_CORE_AXI_CONF_BASE + 0xc)
|
|
|
|
#define PCIE_CORE_AXI_INBOUND_BASE 0xc00800
|
|
#define PCIE_RP_IB_ADDR0 (PCIE_CORE_AXI_INBOUND_BASE + 0x0)
|
|
#define PCIE_CORE_IB_REGION_ADDR0_NUM_BITS 0x3f
|
|
#define PCIE_CORE_IB_REGION_ADDR0_LO_ADDR 0xffffff00
|
|
#define PCIE_RP_IB_ADDR1 (PCIE_CORE_AXI_INBOUND_BASE + 0x4)
|
|
|
|
/* Size of one AXI Region (not Region 0) */
|
|
#define AXI_REGION_SIZE BIT(20)
|
|
/* Size of Region 0, equal to sum of sizes of other regions */
|
|
#define AXI_REGION_0_SIZE (32 * (0x1 << 20))
|
|
#define OB_REG_SIZE_SHIFT 5
|
|
#define IB_ROOT_PORT_REG_SIZE_SHIFT 3
|
|
#define AXI_WRAPPER_IO_WRITE 0x6
|
|
#define AXI_WRAPPER_MEM_WRITE 0x2
|
|
#define AXI_WRAPPER_TYPE0_CFG 0xa
|
|
#define AXI_WRAPPER_TYPE1_CFG 0xb
|
|
#define AXI_WRAPPER_NOR_MSG 0xc
|
|
|
|
#define MAX_AXI_IB_ROOTPORT_REGION_NUM 3
|
|
#define MIN_AXI_ADDR_BITS_PASSED 8
|
|
#define PCIE_RC_SEND_PME_OFF 0x11960
|
|
#define ROCKCHIP_VENDOR_ID 0x1d87
|
|
#define PCIE_ECAM_BUS(x) (((x) & 0xff) << 20)
|
|
#define PCIE_ECAM_DEV(x) (((x) & 0x1f) << 15)
|
|
#define PCIE_ECAM_FUNC(x) (((x) & 0x7) << 12)
|
|
#define PCIE_ECAM_REG(x) (((x) & 0xfff) << 0)
|
|
#define PCIE_ECAM_ADDR(bus, dev, func, reg) \
|
|
(PCIE_ECAM_BUS(bus) | PCIE_ECAM_DEV(dev) | \
|
|
PCIE_ECAM_FUNC(func) | PCIE_ECAM_REG(reg))
|
|
#define PCIE_LINK_IS_L2(x) \
|
|
(((x) & PCIE_CLIENT_DEBUG_LTSSM_MASK) == PCIE_CLIENT_DEBUG_LTSSM_L2)
|
|
#define PCIE_LINK_UP(x) \
|
|
(((x) & PCIE_CLIENT_LINK_STATUS_MASK) == PCIE_CLIENT_LINK_STATUS_UP)
|
|
#define PCIE_LINK_IS_GEN2(x) \
|
|
(((x) & PCIE_CORE_PL_CONF_SPEED_MASK) == PCIE_CORE_PL_CONF_SPEED_5G)
|
|
|
|
#define RC_REGION_0_ADDR_TRANS_H 0x00000000
|
|
#define RC_REGION_0_ADDR_TRANS_L 0x00000000
|
|
#define RC_REGION_0_PASS_BITS (25 - 1)
|
|
#define RC_REGION_0_TYPE_MASK GENMASK(3, 0)
|
|
#define MAX_AXI_WRAPPER_REGION_NUM 33
|
|
|
|
struct rockchip_pcie {
|
|
void __iomem *reg_base; /* DT axi-base */
|
|
void __iomem *apb_base; /* DT apb-base */
|
|
bool legacy_phy;
|
|
struct phy *phys[MAX_LANE_NUM];
|
|
struct reset_control *core_rst;
|
|
struct reset_control *mgmt_rst;
|
|
struct reset_control *mgmt_sticky_rst;
|
|
struct reset_control *pipe_rst;
|
|
struct reset_control *pm_rst;
|
|
struct reset_control *aclk_rst;
|
|
struct reset_control *pclk_rst;
|
|
struct clk *aclk_pcie;
|
|
struct clk *aclk_perf_pcie;
|
|
struct clk *hclk_pcie;
|
|
struct clk *clk_pcie_pm;
|
|
struct regulator *vpcie12v; /* 12V power supply */
|
|
struct regulator *vpcie3v3; /* 3.3V power supply */
|
|
struct regulator *vpcie1v8; /* 1.8V power supply */
|
|
struct regulator *vpcie0v9; /* 0.9V power supply */
|
|
struct gpio_desc *ep_gpio;
|
|
u32 lanes;
|
|
u8 lanes_map;
|
|
u8 root_bus_nr;
|
|
int link_gen;
|
|
struct device *dev;
|
|
struct irq_domain *irq_domain;
|
|
int offset;
|
|
struct pci_bus *root_bus;
|
|
struct resource *io;
|
|
phys_addr_t io_bus_addr;
|
|
u32 io_size;
|
|
void __iomem *msg_region;
|
|
u32 mem_size;
|
|
phys_addr_t msg_bus_addr;
|
|
phys_addr_t mem_bus_addr;
|
|
};
|
|
|
|
static u32 rockchip_pcie_read(struct rockchip_pcie *rockchip, u32 reg)
|
|
{
|
|
return readl(rockchip->apb_base + reg);
|
|
}
|
|
|
|
static void rockchip_pcie_write(struct rockchip_pcie *rockchip, u32 val,
|
|
u32 reg)
|
|
{
|
|
writel(val, rockchip->apb_base + reg);
|
|
}
|
|
|
|
static void rockchip_pcie_enable_bw_int(struct rockchip_pcie *rockchip)
|
|
{
|
|
u32 status;
|
|
|
|
status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_LCS);
|
|
status |= (PCI_EXP_LNKCTL_LBMIE | PCI_EXP_LNKCTL_LABIE);
|
|
rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_LCS);
|
|
}
|
|
|
|
static void rockchip_pcie_clr_bw_int(struct rockchip_pcie *rockchip)
|
|
{
|
|
u32 status;
|
|
|
|
status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_LCS);
|
|
status |= (PCI_EXP_LNKSTA_LBMS | PCI_EXP_LNKSTA_LABS) << 16;
|
|
rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_LCS);
|
|
}
|
|
|
|
static void rockchip_pcie_update_txcredit_mui(struct rockchip_pcie *rockchip)
|
|
{
|
|
u32 val;
|
|
|
|
/* Update Tx credit maximum update interval */
|
|
val = rockchip_pcie_read(rockchip, PCIE_CORE_TXCREDIT_CFG1);
|
|
val &= ~PCIE_CORE_TXCREDIT_CFG1_MUI_MASK;
|
|
val |= PCIE_CORE_TXCREDIT_CFG1_MUI_ENCODE(24000); /* ns */
|
|
rockchip_pcie_write(rockchip, val, PCIE_CORE_TXCREDIT_CFG1);
|
|
}
|
|
|
|
static int rockchip_pcie_valid_device(struct rockchip_pcie *rockchip,
|
|
struct pci_bus *bus, int dev)
|
|
{
|
|
/* access only one slot on each root port */
|
|
if (bus->number == rockchip->root_bus_nr && dev > 0)
|
|
return 0;
|
|
|
|
/*
|
|
* do not read more than one device on the bus directly attached
|
|
* to RC's downstream side.
|
|
*/
|
|
if (bus->primary == rockchip->root_bus_nr && dev > 0)
|
|
return 0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
static u8 rockchip_pcie_lane_map(struct rockchip_pcie *rockchip)
|
|
{
|
|
u32 val;
|
|
u8 map;
|
|
|
|
if (rockchip->legacy_phy)
|
|
return GENMASK(MAX_LANE_NUM - 1, 0);
|
|
|
|
val = rockchip_pcie_read(rockchip, PCIE_CORE_LANE_MAP);
|
|
map = val & PCIE_CORE_LANE_MAP_MASK;
|
|
|
|
/* The link may be using a reverse-indexed mapping. */
|
|
if (val & PCIE_CORE_LANE_MAP_REVERSE)
|
|
map = bitrev8(map) >> 4;
|
|
|
|
return map;
|
|
}
|
|
|
|
static int rockchip_pcie_rd_own_conf(struct rockchip_pcie *rockchip,
|
|
int where, int size, u32 *val)
|
|
{
|
|
void __iomem *addr;
|
|
|
|
addr = rockchip->apb_base + PCIE_RC_CONFIG_NORMAL_BASE + where;
|
|
|
|
if (!IS_ALIGNED((uintptr_t)addr, size)) {
|
|
*val = 0;
|
|
return PCIBIOS_BAD_REGISTER_NUMBER;
|
|
}
|
|
|
|
if (size == 4) {
|
|
*val = readl(addr);
|
|
} else if (size == 2) {
|
|
*val = readw(addr);
|
|
} else if (size == 1) {
|
|
*val = readb(addr);
|
|
} else {
|
|
*val = 0;
|
|
return PCIBIOS_BAD_REGISTER_NUMBER;
|
|
}
|
|
return PCIBIOS_SUCCESSFUL;
|
|
}
|
|
|
|
static int rockchip_pcie_wr_own_conf(struct rockchip_pcie *rockchip,
|
|
int where, int size, u32 val)
|
|
{
|
|
u32 mask, tmp, offset;
|
|
void __iomem *addr;
|
|
|
|
offset = where & ~0x3;
|
|
addr = rockchip->apb_base + PCIE_RC_CONFIG_NORMAL_BASE + offset;
|
|
|
|
if (size == 4) {
|
|
writel(val, addr);
|
|
return PCIBIOS_SUCCESSFUL;
|
|
}
|
|
|
|
mask = ~(((1 << (size * 8)) - 1) << ((where & 0x3) * 8));
|
|
|
|
/*
|
|
* N.B. This read/modify/write isn't safe in general because it can
|
|
* corrupt RW1C bits in adjacent registers. But the hardware
|
|
* doesn't support smaller writes.
|
|
*/
|
|
tmp = readl(addr) & mask;
|
|
tmp |= val << ((where & 0x3) * 8);
|
|
writel(tmp, addr);
|
|
|
|
return PCIBIOS_SUCCESSFUL;
|
|
}
|
|
|
|
static void rockchip_pcie_cfg_configuration_accesses(
|
|
struct rockchip_pcie *rockchip, u32 type)
|
|
{
|
|
u32 ob_desc_0;
|
|
|
|
/* Configuration Accesses for region 0 */
|
|
rockchip_pcie_write(rockchip, 0x0, PCIE_RC_BAR_CONF);
|
|
|
|
rockchip_pcie_write(rockchip,
|
|
(RC_REGION_0_ADDR_TRANS_L + RC_REGION_0_PASS_BITS),
|
|
PCIE_CORE_OB_REGION_ADDR0);
|
|
rockchip_pcie_write(rockchip, RC_REGION_0_ADDR_TRANS_H,
|
|
PCIE_CORE_OB_REGION_ADDR1);
|
|
ob_desc_0 = rockchip_pcie_read(rockchip, PCIE_CORE_OB_REGION_DESC0);
|
|
ob_desc_0 &= ~(RC_REGION_0_TYPE_MASK);
|
|
ob_desc_0 |= (type | (0x1 << 23));
|
|
rockchip_pcie_write(rockchip, ob_desc_0, PCIE_CORE_OB_REGION_DESC0);
|
|
rockchip_pcie_write(rockchip, 0x0, PCIE_CORE_OB_REGION_DESC1);
|
|
}
|
|
|
|
static int rockchip_pcie_rd_other_conf(struct rockchip_pcie *rockchip,
|
|
struct pci_bus *bus, u32 devfn,
|
|
int where, int size, u32 *val)
|
|
{
|
|
u32 busdev;
|
|
|
|
busdev = PCIE_ECAM_ADDR(bus->number, PCI_SLOT(devfn),
|
|
PCI_FUNC(devfn), where);
|
|
|
|
if (!IS_ALIGNED(busdev, size)) {
|
|
*val = 0;
|
|
return PCIBIOS_BAD_REGISTER_NUMBER;
|
|
}
|
|
|
|
if (bus->parent->number == rockchip->root_bus_nr)
|
|
rockchip_pcie_cfg_configuration_accesses(rockchip,
|
|
AXI_WRAPPER_TYPE0_CFG);
|
|
else
|
|
rockchip_pcie_cfg_configuration_accesses(rockchip,
|
|
AXI_WRAPPER_TYPE1_CFG);
|
|
|
|
if (size == 4) {
|
|
*val = readl(rockchip->reg_base + busdev);
|
|
} else if (size == 2) {
|
|
*val = readw(rockchip->reg_base + busdev);
|
|
} else if (size == 1) {
|
|
*val = readb(rockchip->reg_base + busdev);
|
|
} else {
|
|
*val = 0;
|
|
return PCIBIOS_BAD_REGISTER_NUMBER;
|
|
}
|
|
return PCIBIOS_SUCCESSFUL;
|
|
}
|
|
|
|
static int rockchip_pcie_wr_other_conf(struct rockchip_pcie *rockchip,
|
|
struct pci_bus *bus, u32 devfn,
|
|
int where, int size, u32 val)
|
|
{
|
|
u32 busdev;
|
|
|
|
busdev = PCIE_ECAM_ADDR(bus->number, PCI_SLOT(devfn),
|
|
PCI_FUNC(devfn), where);
|
|
if (!IS_ALIGNED(busdev, size))
|
|
return PCIBIOS_BAD_REGISTER_NUMBER;
|
|
|
|
if (bus->parent->number == rockchip->root_bus_nr)
|
|
rockchip_pcie_cfg_configuration_accesses(rockchip,
|
|
AXI_WRAPPER_TYPE0_CFG);
|
|
else
|
|
rockchip_pcie_cfg_configuration_accesses(rockchip,
|
|
AXI_WRAPPER_TYPE1_CFG);
|
|
|
|
if (size == 4)
|
|
writel(val, rockchip->reg_base + busdev);
|
|
else if (size == 2)
|
|
writew(val, rockchip->reg_base + busdev);
|
|
else if (size == 1)
|
|
writeb(val, rockchip->reg_base + busdev);
|
|
else
|
|
return PCIBIOS_BAD_REGISTER_NUMBER;
|
|
|
|
return PCIBIOS_SUCCESSFUL;
|
|
}
|
|
|
|
static int rockchip_pcie_rd_conf(struct pci_bus *bus, u32 devfn, int where,
|
|
int size, u32 *val)
|
|
{
|
|
struct rockchip_pcie *rockchip = bus->sysdata;
|
|
|
|
if (!rockchip_pcie_valid_device(rockchip, bus, PCI_SLOT(devfn))) {
|
|
*val = 0xffffffff;
|
|
return PCIBIOS_DEVICE_NOT_FOUND;
|
|
}
|
|
|
|
if (bus->number == rockchip->root_bus_nr)
|
|
return rockchip_pcie_rd_own_conf(rockchip, where, size, val);
|
|
|
|
return rockchip_pcie_rd_other_conf(rockchip, bus, devfn, where, size, val);
|
|
}
|
|
|
|
static int rockchip_pcie_wr_conf(struct pci_bus *bus, u32 devfn,
|
|
int where, int size, u32 val)
|
|
{
|
|
struct rockchip_pcie *rockchip = bus->sysdata;
|
|
|
|
if (!rockchip_pcie_valid_device(rockchip, bus, PCI_SLOT(devfn)))
|
|
return PCIBIOS_DEVICE_NOT_FOUND;
|
|
|
|
if (bus->number == rockchip->root_bus_nr)
|
|
return rockchip_pcie_wr_own_conf(rockchip, where, size, val);
|
|
|
|
return rockchip_pcie_wr_other_conf(rockchip, bus, devfn, where, size, val);
|
|
}
|
|
|
|
static struct pci_ops rockchip_pcie_ops = {
|
|
.read = rockchip_pcie_rd_conf,
|
|
.write = rockchip_pcie_wr_conf,
|
|
};
|
|
|
|
static void rockchip_pcie_set_power_limit(struct rockchip_pcie *rockchip)
|
|
{
|
|
int curr;
|
|
u32 status, scale, power;
|
|
|
|
if (IS_ERR(rockchip->vpcie3v3))
|
|
return;
|
|
|
|
/*
|
|
* Set RC's captured slot power limit and scale if
|
|
* vpcie3v3 available. The default values are both zero
|
|
* which means the software should set these two according
|
|
* to the actual power supply.
|
|
*/
|
|
curr = regulator_get_current_limit(rockchip->vpcie3v3);
|
|
if (curr <= 0)
|
|
return;
|
|
|
|
scale = 3; /* 0.001x */
|
|
curr = curr / 1000; /* convert to mA */
|
|
power = (curr * 3300) / 1000; /* milliwatt */
|
|
while (power > PCIE_RC_CONFIG_DCR_CSPL_LIMIT) {
|
|
if (!scale) {
|
|
dev_warn(rockchip->dev, "invalid power supply\n");
|
|
return;
|
|
}
|
|
scale--;
|
|
power = power / 10;
|
|
}
|
|
|
|
status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_DCR);
|
|
status |= (power << PCIE_RC_CONFIG_DCR_CSPL_SHIFT) |
|
|
(scale << PCIE_RC_CONFIG_DCR_CPLS_SHIFT);
|
|
rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_DCR);
|
|
}
|
|
|
|
/**
|
|
* rockchip_pcie_init_port - Initialize hardware
|
|
* @rockchip: PCIe port information
|
|
*/
|
|
static int rockchip_pcie_init_port(struct rockchip_pcie *rockchip)
|
|
{
|
|
struct device *dev = rockchip->dev;
|
|
int err, i;
|
|
u32 status;
|
|
|
|
gpiod_set_value_cansleep(rockchip->ep_gpio, 0);
|
|
|
|
err = reset_control_assert(rockchip->aclk_rst);
|
|
if (err) {
|
|
dev_err(dev, "assert aclk_rst err %d\n", err);
|
|
return err;
|
|
}
|
|
|
|
err = reset_control_assert(rockchip->pclk_rst);
|
|
if (err) {
|
|
dev_err(dev, "assert pclk_rst err %d\n", err);
|
|
return err;
|
|
}
|
|
|
|
err = reset_control_assert(rockchip->pm_rst);
|
|
if (err) {
|
|
dev_err(dev, "assert pm_rst err %d\n", err);
|
|
return err;
|
|
}
|
|
|
|
for (i = 0; i < MAX_LANE_NUM; i++) {
|
|
err = phy_init(rockchip->phys[i]);
|
|
if (err) {
|
|
dev_err(dev, "init phy%d err %d\n", i, err);
|
|
goto err_exit_phy;
|
|
}
|
|
}
|
|
|
|
err = reset_control_assert(rockchip->core_rst);
|
|
if (err) {
|
|
dev_err(dev, "assert core_rst err %d\n", err);
|
|
goto err_exit_phy;
|
|
}
|
|
|
|
err = reset_control_assert(rockchip->mgmt_rst);
|
|
if (err) {
|
|
dev_err(dev, "assert mgmt_rst err %d\n", err);
|
|
goto err_exit_phy;
|
|
}
|
|
|
|
err = reset_control_assert(rockchip->mgmt_sticky_rst);
|
|
if (err) {
|
|
dev_err(dev, "assert mgmt_sticky_rst err %d\n", err);
|
|
goto err_exit_phy;
|
|
}
|
|
|
|
err = reset_control_assert(rockchip->pipe_rst);
|
|
if (err) {
|
|
dev_err(dev, "assert pipe_rst err %d\n", err);
|
|
goto err_exit_phy;
|
|
}
|
|
|
|
udelay(10);
|
|
|
|
err = reset_control_deassert(rockchip->pm_rst);
|
|
if (err) {
|
|
dev_err(dev, "deassert pm_rst err %d\n", err);
|
|
goto err_exit_phy;
|
|
}
|
|
|
|
err = reset_control_deassert(rockchip->aclk_rst);
|
|
if (err) {
|
|
dev_err(dev, "deassert aclk_rst err %d\n", err);
|
|
goto err_exit_phy;
|
|
}
|
|
|
|
err = reset_control_deassert(rockchip->pclk_rst);
|
|
if (err) {
|
|
dev_err(dev, "deassert pclk_rst err %d\n", err);
|
|
goto err_exit_phy;
|
|
}
|
|
|
|
if (rockchip->link_gen == 2)
|
|
rockchip_pcie_write(rockchip, PCIE_CLIENT_GEN_SEL_2,
|
|
PCIE_CLIENT_CONFIG);
|
|
else
|
|
rockchip_pcie_write(rockchip, PCIE_CLIENT_GEN_SEL_1,
|
|
PCIE_CLIENT_CONFIG);
|
|
|
|
rockchip_pcie_write(rockchip,
|
|
PCIE_CLIENT_CONF_ENABLE |
|
|
PCIE_CLIENT_LINK_TRAIN_ENABLE |
|
|
PCIE_CLIENT_ARI_ENABLE |
|
|
PCIE_CLIENT_CONF_LANE_NUM(rockchip->lanes) |
|
|
PCIE_CLIENT_MODE_RC,
|
|
PCIE_CLIENT_CONFIG);
|
|
|
|
for (i = 0; i < MAX_LANE_NUM; i++) {
|
|
err = phy_power_on(rockchip->phys[i]);
|
|
if (err) {
|
|
dev_err(dev, "power on phy%d err %d\n", i, err);
|
|
goto err_power_off_phy;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Please don't reorder the deassert sequence of the following
|
|
* four reset pins.
|
|
*/
|
|
err = reset_control_deassert(rockchip->mgmt_sticky_rst);
|
|
if (err) {
|
|
dev_err(dev, "deassert mgmt_sticky_rst err %d\n", err);
|
|
goto err_power_off_phy;
|
|
}
|
|
|
|
err = reset_control_deassert(rockchip->core_rst);
|
|
if (err) {
|
|
dev_err(dev, "deassert core_rst err %d\n", err);
|
|
goto err_power_off_phy;
|
|
}
|
|
|
|
err = reset_control_deassert(rockchip->mgmt_rst);
|
|
if (err) {
|
|
dev_err(dev, "deassert mgmt_rst err %d\n", err);
|
|
goto err_power_off_phy;
|
|
}
|
|
|
|
err = reset_control_deassert(rockchip->pipe_rst);
|
|
if (err) {
|
|
dev_err(dev, "deassert pipe_rst err %d\n", err);
|
|
goto err_power_off_phy;
|
|
}
|
|
|
|
/* Fix the transmitted FTS count desired to exit from L0s. */
|
|
status = rockchip_pcie_read(rockchip, PCIE_CORE_CTRL_PLC1);
|
|
status = (status & ~PCIE_CORE_CTRL_PLC1_FTS_MASK) |
|
|
(PCIE_CORE_CTRL_PLC1_FTS_CNT << PCIE_CORE_CTRL_PLC1_FTS_SHIFT);
|
|
rockchip_pcie_write(rockchip, status, PCIE_CORE_CTRL_PLC1);
|
|
|
|
rockchip_pcie_set_power_limit(rockchip);
|
|
|
|
/* Set RC's clock architecture as common clock */
|
|
status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_LCS);
|
|
status |= PCI_EXP_LNKSTA_SLC << 16;
|
|
rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_LCS);
|
|
|
|
/* Set RC's RCB to 128 */
|
|
status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_LCS);
|
|
status |= PCI_EXP_LNKCTL_RCB;
|
|
rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_LCS);
|
|
|
|
/* Enable Gen1 training */
|
|
rockchip_pcie_write(rockchip, PCIE_CLIENT_LINK_TRAIN_ENABLE,
|
|
PCIE_CLIENT_CONFIG);
|
|
|
|
gpiod_set_value_cansleep(rockchip->ep_gpio, 1);
|
|
|
|
/* 500ms timeout value should be enough for Gen1/2 training */
|
|
err = readl_poll_timeout(rockchip->apb_base + PCIE_CLIENT_BASIC_STATUS1,
|
|
status, PCIE_LINK_UP(status), 20,
|
|
500 * USEC_PER_MSEC);
|
|
if (err) {
|
|
dev_err(dev, "PCIe link training gen1 timeout!\n");
|
|
goto err_power_off_phy;
|
|
}
|
|
|
|
if (rockchip->link_gen == 2) {
|
|
/*
|
|
* Enable retrain for gen2. This should be configured only after
|
|
* gen1 finished.
|
|
*/
|
|
status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_LCS);
|
|
status |= PCI_EXP_LNKCTL_RL;
|
|
rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_LCS);
|
|
|
|
err = readl_poll_timeout(rockchip->apb_base + PCIE_CORE_CTRL,
|
|
status, PCIE_LINK_IS_GEN2(status), 20,
|
|
500 * USEC_PER_MSEC);
|
|
if (err)
|
|
dev_dbg(dev, "PCIe link training gen2 timeout, fall back to gen1!\n");
|
|
}
|
|
|
|
/* Check the final link width from negotiated lane counter from MGMT */
|
|
status = rockchip_pcie_read(rockchip, PCIE_CORE_CTRL);
|
|
status = 0x1 << ((status & PCIE_CORE_PL_CONF_LANE_MASK) >>
|
|
PCIE_CORE_PL_CONF_LANE_SHIFT);
|
|
dev_dbg(dev, "current link width is x%d\n", status);
|
|
|
|
/* Power off unused lane(s) */
|
|
rockchip->lanes_map = rockchip_pcie_lane_map(rockchip);
|
|
for (i = 0; i < MAX_LANE_NUM; i++) {
|
|
if (!(rockchip->lanes_map & BIT(i))) {
|
|
dev_dbg(dev, "idling lane %d\n", i);
|
|
phy_power_off(rockchip->phys[i]);
|
|
}
|
|
}
|
|
|
|
rockchip_pcie_write(rockchip, ROCKCHIP_VENDOR_ID,
|
|
PCIE_CORE_CONFIG_VENDOR);
|
|
rockchip_pcie_write(rockchip,
|
|
PCI_CLASS_BRIDGE_PCI << PCIE_RC_CONFIG_SCC_SHIFT,
|
|
PCIE_RC_CONFIG_RID_CCR);
|
|
|
|
/* Clear THP cap's next cap pointer to remove L1 substate cap */
|
|
status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_THP_CAP);
|
|
status &= ~PCIE_RC_CONFIG_THP_CAP_NEXT_MASK;
|
|
rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_THP_CAP);
|
|
|
|
/* Clear L0s from RC's link cap */
|
|
if (of_property_read_bool(dev->of_node, "aspm-no-l0s")) {
|
|
status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_LINK_CAP);
|
|
status &= ~PCIE_RC_CONFIG_LINK_CAP_L0S;
|
|
rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_LINK_CAP);
|
|
}
|
|
|
|
status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_DCSR);
|
|
status &= ~PCIE_RC_CONFIG_DCSR_MPS_MASK;
|
|
status |= PCIE_RC_CONFIG_DCSR_MPS_256;
|
|
rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_DCSR);
|
|
|
|
return 0;
|
|
err_power_off_phy:
|
|
while (i--)
|
|
phy_power_off(rockchip->phys[i]);
|
|
i = MAX_LANE_NUM;
|
|
err_exit_phy:
|
|
while (i--)
|
|
phy_exit(rockchip->phys[i]);
|
|
return err;
|
|
}
|
|
|
|
static void rockchip_pcie_deinit_phys(struct rockchip_pcie *rockchip)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < MAX_LANE_NUM; i++) {
|
|
/* inactive lanes are already powered off */
|
|
if (rockchip->lanes_map & BIT(i))
|
|
phy_power_off(rockchip->phys[i]);
|
|
phy_exit(rockchip->phys[i]);
|
|
}
|
|
}
|
|
|
|
static irqreturn_t rockchip_pcie_subsys_irq_handler(int irq, void *arg)
|
|
{
|
|
struct rockchip_pcie *rockchip = arg;
|
|
struct device *dev = rockchip->dev;
|
|
u32 reg;
|
|
u32 sub_reg;
|
|
|
|
reg = rockchip_pcie_read(rockchip, PCIE_CLIENT_INT_STATUS);
|
|
if (reg & PCIE_CLIENT_INT_LOCAL) {
|
|
dev_dbg(dev, "local interrupt received\n");
|
|
sub_reg = rockchip_pcie_read(rockchip, PCIE_CORE_INT_STATUS);
|
|
if (sub_reg & PCIE_CORE_INT_PRFPE)
|
|
dev_dbg(dev, "parity error detected while reading from the PNP receive FIFO RAM\n");
|
|
|
|
if (sub_reg & PCIE_CORE_INT_CRFPE)
|
|
dev_dbg(dev, "parity error detected while reading from the Completion Receive FIFO RAM\n");
|
|
|
|
if (sub_reg & PCIE_CORE_INT_RRPE)
|
|
dev_dbg(dev, "parity error detected while reading from replay buffer RAM\n");
|
|
|
|
if (sub_reg & PCIE_CORE_INT_PRFO)
|
|
dev_dbg(dev, "overflow occurred in the PNP receive FIFO\n");
|
|
|
|
if (sub_reg & PCIE_CORE_INT_CRFO)
|
|
dev_dbg(dev, "overflow occurred in the completion receive FIFO\n");
|
|
|
|
if (sub_reg & PCIE_CORE_INT_RT)
|
|
dev_dbg(dev, "replay timer timed out\n");
|
|
|
|
if (sub_reg & PCIE_CORE_INT_RTR)
|
|
dev_dbg(dev, "replay timer rolled over after 4 transmissions of the same TLP\n");
|
|
|
|
if (sub_reg & PCIE_CORE_INT_PE)
|
|
dev_dbg(dev, "phy error detected on receive side\n");
|
|
|
|
if (sub_reg & PCIE_CORE_INT_MTR)
|
|
dev_dbg(dev, "malformed TLP received from the link\n");
|
|
|
|
if (sub_reg & PCIE_CORE_INT_UCR)
|
|
dev_dbg(dev, "malformed TLP received from the link\n");
|
|
|
|
if (sub_reg & PCIE_CORE_INT_FCE)
|
|
dev_dbg(dev, "an error was observed in the flow control advertisements from the other side\n");
|
|
|
|
if (sub_reg & PCIE_CORE_INT_CT)
|
|
dev_dbg(dev, "a request timed out waiting for completion\n");
|
|
|
|
if (sub_reg & PCIE_CORE_INT_UTC)
|
|
dev_dbg(dev, "unmapped TC error\n");
|
|
|
|
if (sub_reg & PCIE_CORE_INT_MMVC)
|
|
dev_dbg(dev, "MSI mask register changes\n");
|
|
|
|
rockchip_pcie_write(rockchip, sub_reg, PCIE_CORE_INT_STATUS);
|
|
} else if (reg & PCIE_CLIENT_INT_PHY) {
|
|
dev_dbg(dev, "phy link changes\n");
|
|
rockchip_pcie_update_txcredit_mui(rockchip);
|
|
rockchip_pcie_clr_bw_int(rockchip);
|
|
}
|
|
|
|
rockchip_pcie_write(rockchip, reg & PCIE_CLIENT_INT_LOCAL,
|
|
PCIE_CLIENT_INT_STATUS);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static irqreturn_t rockchip_pcie_client_irq_handler(int irq, void *arg)
|
|
{
|
|
struct rockchip_pcie *rockchip = arg;
|
|
struct device *dev = rockchip->dev;
|
|
u32 reg;
|
|
|
|
reg = rockchip_pcie_read(rockchip, PCIE_CLIENT_INT_STATUS);
|
|
if (reg & PCIE_CLIENT_INT_LEGACY_DONE)
|
|
dev_dbg(dev, "legacy done interrupt received\n");
|
|
|
|
if (reg & PCIE_CLIENT_INT_MSG)
|
|
dev_dbg(dev, "message done interrupt received\n");
|
|
|
|
if (reg & PCIE_CLIENT_INT_HOT_RST)
|
|
dev_dbg(dev, "hot reset interrupt received\n");
|
|
|
|
if (reg & PCIE_CLIENT_INT_DPA)
|
|
dev_dbg(dev, "dpa interrupt received\n");
|
|
|
|
if (reg & PCIE_CLIENT_INT_FATAL_ERR)
|
|
dev_dbg(dev, "fatal error interrupt received\n");
|
|
|
|
if (reg & PCIE_CLIENT_INT_NFATAL_ERR)
|
|
dev_dbg(dev, "no fatal error interrupt received\n");
|
|
|
|
if (reg & PCIE_CLIENT_INT_CORR_ERR)
|
|
dev_dbg(dev, "correctable error interrupt received\n");
|
|
|
|
if (reg & PCIE_CLIENT_INT_PHY)
|
|
dev_dbg(dev, "phy interrupt received\n");
|
|
|
|
rockchip_pcie_write(rockchip, reg & (PCIE_CLIENT_INT_LEGACY_DONE |
|
|
PCIE_CLIENT_INT_MSG | PCIE_CLIENT_INT_HOT_RST |
|
|
PCIE_CLIENT_INT_DPA | PCIE_CLIENT_INT_FATAL_ERR |
|
|
PCIE_CLIENT_INT_NFATAL_ERR |
|
|
PCIE_CLIENT_INT_CORR_ERR |
|
|
PCIE_CLIENT_INT_PHY),
|
|
PCIE_CLIENT_INT_STATUS);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static void rockchip_pcie_legacy_int_handler(struct irq_desc *desc)
|
|
{
|
|
struct irq_chip *chip = irq_desc_get_chip(desc);
|
|
struct rockchip_pcie *rockchip = irq_desc_get_handler_data(desc);
|
|
struct device *dev = rockchip->dev;
|
|
u32 reg;
|
|
u32 hwirq;
|
|
u32 virq;
|
|
|
|
chained_irq_enter(chip, desc);
|
|
|
|
reg = rockchip_pcie_read(rockchip, PCIE_CLIENT_INT_STATUS);
|
|
reg = (reg & PCIE_CLIENT_INTR_MASK) >> PCIE_CLIENT_INTR_SHIFT;
|
|
|
|
while (reg) {
|
|
hwirq = ffs(reg) - 1;
|
|
reg &= ~BIT(hwirq);
|
|
|
|
virq = irq_find_mapping(rockchip->irq_domain, hwirq);
|
|
if (virq)
|
|
generic_handle_irq(virq);
|
|
else
|
|
dev_err(dev, "unexpected IRQ, INT%d\n", hwirq);
|
|
}
|
|
|
|
chained_irq_exit(chip, desc);
|
|
}
|
|
|
|
static int rockchip_pcie_get_phys(struct rockchip_pcie *rockchip)
|
|
{
|
|
struct device *dev = rockchip->dev;
|
|
struct phy *phy;
|
|
char *name;
|
|
u32 i;
|
|
|
|
phy = devm_phy_get(dev, "pcie-phy");
|
|
if (!IS_ERR(phy)) {
|
|
rockchip->legacy_phy = true;
|
|
rockchip->phys[0] = phy;
|
|
dev_warn(dev, "legacy phy model is deprecated!\n");
|
|
return 0;
|
|
}
|
|
|
|
if (PTR_ERR(phy) == -EPROBE_DEFER)
|
|
return PTR_ERR(phy);
|
|
|
|
dev_dbg(dev, "missing legacy phy; search for per-lane PHY\n");
|
|
|
|
for (i = 0; i < MAX_LANE_NUM; i++) {
|
|
name = kasprintf(GFP_KERNEL, "pcie-phy-%u", i);
|
|
if (!name)
|
|
return -ENOMEM;
|
|
|
|
phy = devm_of_phy_get(dev, dev->of_node, name);
|
|
kfree(name);
|
|
|
|
if (IS_ERR(phy)) {
|
|
if (PTR_ERR(phy) != -EPROBE_DEFER)
|
|
dev_err(dev, "missing phy for lane %d: %ld\n",
|
|
i, PTR_ERR(phy));
|
|
return PTR_ERR(phy);
|
|
}
|
|
|
|
rockchip->phys[i] = phy;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int rockchip_pcie_setup_irq(struct rockchip_pcie *rockchip)
|
|
{
|
|
int irq, err;
|
|
struct device *dev = rockchip->dev;
|
|
struct platform_device *pdev = to_platform_device(dev);
|
|
|
|
irq = platform_get_irq_byname(pdev, "sys");
|
|
if (irq < 0) {
|
|
dev_err(dev, "missing sys IRQ resource\n");
|
|
return irq;
|
|
}
|
|
|
|
err = devm_request_irq(dev, irq, rockchip_pcie_subsys_irq_handler,
|
|
IRQF_SHARED, "pcie-sys", rockchip);
|
|
if (err) {
|
|
dev_err(dev, "failed to request PCIe subsystem IRQ\n");
|
|
return err;
|
|
}
|
|
|
|
irq = platform_get_irq_byname(pdev, "legacy");
|
|
if (irq < 0) {
|
|
dev_err(dev, "missing legacy IRQ resource\n");
|
|
return irq;
|
|
}
|
|
|
|
irq_set_chained_handler_and_data(irq,
|
|
rockchip_pcie_legacy_int_handler,
|
|
rockchip);
|
|
|
|
irq = platform_get_irq_byname(pdev, "client");
|
|
if (irq < 0) {
|
|
dev_err(dev, "missing client IRQ resource\n");
|
|
return irq;
|
|
}
|
|
|
|
err = devm_request_irq(dev, irq, rockchip_pcie_client_irq_handler,
|
|
IRQF_SHARED, "pcie-client", rockchip);
|
|
if (err) {
|
|
dev_err(dev, "failed to request PCIe client IRQ\n");
|
|
return err;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* rockchip_pcie_parse_dt - Parse Device Tree
|
|
* @rockchip: PCIe port information
|
|
*
|
|
* Return: '0' on success and error value on failure
|
|
*/
|
|
static int rockchip_pcie_parse_dt(struct rockchip_pcie *rockchip)
|
|
{
|
|
struct device *dev = rockchip->dev;
|
|
struct platform_device *pdev = to_platform_device(dev);
|
|
struct device_node *node = dev->of_node;
|
|
struct resource *regs;
|
|
int err;
|
|
|
|
regs = platform_get_resource_byname(pdev,
|
|
IORESOURCE_MEM,
|
|
"axi-base");
|
|
rockchip->reg_base = devm_pci_remap_cfg_resource(dev, regs);
|
|
if (IS_ERR(rockchip->reg_base))
|
|
return PTR_ERR(rockchip->reg_base);
|
|
|
|
regs = platform_get_resource_byname(pdev,
|
|
IORESOURCE_MEM,
|
|
"apb-base");
|
|
rockchip->apb_base = devm_ioremap_resource(dev, regs);
|
|
if (IS_ERR(rockchip->apb_base))
|
|
return PTR_ERR(rockchip->apb_base);
|
|
|
|
err = rockchip_pcie_get_phys(rockchip);
|
|
if (err)
|
|
return err;
|
|
|
|
rockchip->lanes = 1;
|
|
err = of_property_read_u32(node, "num-lanes", &rockchip->lanes);
|
|
if (!err && (rockchip->lanes == 0 ||
|
|
rockchip->lanes == 3 ||
|
|
rockchip->lanes > 4)) {
|
|
dev_warn(dev, "invalid num-lanes, default to use one lane\n");
|
|
rockchip->lanes = 1;
|
|
}
|
|
|
|
rockchip->link_gen = of_pci_get_max_link_speed(node);
|
|
if (rockchip->link_gen < 0 || rockchip->link_gen > 2)
|
|
rockchip->link_gen = 2;
|
|
|
|
rockchip->core_rst = devm_reset_control_get_exclusive(dev, "core");
|
|
if (IS_ERR(rockchip->core_rst)) {
|
|
if (PTR_ERR(rockchip->core_rst) != -EPROBE_DEFER)
|
|
dev_err(dev, "missing core reset property in node\n");
|
|
return PTR_ERR(rockchip->core_rst);
|
|
}
|
|
|
|
rockchip->mgmt_rst = devm_reset_control_get_exclusive(dev, "mgmt");
|
|
if (IS_ERR(rockchip->mgmt_rst)) {
|
|
if (PTR_ERR(rockchip->mgmt_rst) != -EPROBE_DEFER)
|
|
dev_err(dev, "missing mgmt reset property in node\n");
|
|
return PTR_ERR(rockchip->mgmt_rst);
|
|
}
|
|
|
|
rockchip->mgmt_sticky_rst = devm_reset_control_get_exclusive(dev,
|
|
"mgmt-sticky");
|
|
if (IS_ERR(rockchip->mgmt_sticky_rst)) {
|
|
if (PTR_ERR(rockchip->mgmt_sticky_rst) != -EPROBE_DEFER)
|
|
dev_err(dev, "missing mgmt-sticky reset property in node\n");
|
|
return PTR_ERR(rockchip->mgmt_sticky_rst);
|
|
}
|
|
|
|
rockchip->pipe_rst = devm_reset_control_get_exclusive(dev, "pipe");
|
|
if (IS_ERR(rockchip->pipe_rst)) {
|
|
if (PTR_ERR(rockchip->pipe_rst) != -EPROBE_DEFER)
|
|
dev_err(dev, "missing pipe reset property in node\n");
|
|
return PTR_ERR(rockchip->pipe_rst);
|
|
}
|
|
|
|
rockchip->pm_rst = devm_reset_control_get_exclusive(dev, "pm");
|
|
if (IS_ERR(rockchip->pm_rst)) {
|
|
if (PTR_ERR(rockchip->pm_rst) != -EPROBE_DEFER)
|
|
dev_err(dev, "missing pm reset property in node\n");
|
|
return PTR_ERR(rockchip->pm_rst);
|
|
}
|
|
|
|
rockchip->pclk_rst = devm_reset_control_get_exclusive(dev, "pclk");
|
|
if (IS_ERR(rockchip->pclk_rst)) {
|
|
if (PTR_ERR(rockchip->pclk_rst) != -EPROBE_DEFER)
|
|
dev_err(dev, "missing pclk reset property in node\n");
|
|
return PTR_ERR(rockchip->pclk_rst);
|
|
}
|
|
|
|
rockchip->aclk_rst = devm_reset_control_get_exclusive(dev, "aclk");
|
|
if (IS_ERR(rockchip->aclk_rst)) {
|
|
if (PTR_ERR(rockchip->aclk_rst) != -EPROBE_DEFER)
|
|
dev_err(dev, "missing aclk reset property in node\n");
|
|
return PTR_ERR(rockchip->aclk_rst);
|
|
}
|
|
|
|
rockchip->ep_gpio = devm_gpiod_get(dev, "ep", GPIOD_OUT_HIGH);
|
|
if (IS_ERR(rockchip->ep_gpio)) {
|
|
dev_err(dev, "missing ep-gpios property in node\n");
|
|
return PTR_ERR(rockchip->ep_gpio);
|
|
}
|
|
|
|
rockchip->aclk_pcie = devm_clk_get(dev, "aclk");
|
|
if (IS_ERR(rockchip->aclk_pcie)) {
|
|
dev_err(dev, "aclk clock not found\n");
|
|
return PTR_ERR(rockchip->aclk_pcie);
|
|
}
|
|
|
|
rockchip->aclk_perf_pcie = devm_clk_get(dev, "aclk-perf");
|
|
if (IS_ERR(rockchip->aclk_perf_pcie)) {
|
|
dev_err(dev, "aclk_perf clock not found\n");
|
|
return PTR_ERR(rockchip->aclk_perf_pcie);
|
|
}
|
|
|
|
rockchip->hclk_pcie = devm_clk_get(dev, "hclk");
|
|
if (IS_ERR(rockchip->hclk_pcie)) {
|
|
dev_err(dev, "hclk clock not found\n");
|
|
return PTR_ERR(rockchip->hclk_pcie);
|
|
}
|
|
|
|
rockchip->clk_pcie_pm = devm_clk_get(dev, "pm");
|
|
if (IS_ERR(rockchip->clk_pcie_pm)) {
|
|
dev_err(dev, "pm clock not found\n");
|
|
return PTR_ERR(rockchip->clk_pcie_pm);
|
|
}
|
|
|
|
err = rockchip_pcie_setup_irq(rockchip);
|
|
if (err)
|
|
return err;
|
|
|
|
rockchip->vpcie12v = devm_regulator_get_optional(dev, "vpcie12v");
|
|
if (IS_ERR(rockchip->vpcie12v)) {
|
|
if (PTR_ERR(rockchip->vpcie12v) == -EPROBE_DEFER)
|
|
return -EPROBE_DEFER;
|
|
dev_info(dev, "no vpcie12v regulator found\n");
|
|
}
|
|
|
|
rockchip->vpcie3v3 = devm_regulator_get_optional(dev, "vpcie3v3");
|
|
if (IS_ERR(rockchip->vpcie3v3)) {
|
|
if (PTR_ERR(rockchip->vpcie3v3) == -EPROBE_DEFER)
|
|
return -EPROBE_DEFER;
|
|
dev_info(dev, "no vpcie3v3 regulator found\n");
|
|
}
|
|
|
|
rockchip->vpcie1v8 = devm_regulator_get_optional(dev, "vpcie1v8");
|
|
if (IS_ERR(rockchip->vpcie1v8)) {
|
|
if (PTR_ERR(rockchip->vpcie1v8) == -EPROBE_DEFER)
|
|
return -EPROBE_DEFER;
|
|
dev_info(dev, "no vpcie1v8 regulator found\n");
|
|
}
|
|
|
|
rockchip->vpcie0v9 = devm_regulator_get_optional(dev, "vpcie0v9");
|
|
if (IS_ERR(rockchip->vpcie0v9)) {
|
|
if (PTR_ERR(rockchip->vpcie0v9) == -EPROBE_DEFER)
|
|
return -EPROBE_DEFER;
|
|
dev_info(dev, "no vpcie0v9 regulator found\n");
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int rockchip_pcie_set_vpcie(struct rockchip_pcie *rockchip)
|
|
{
|
|
struct device *dev = rockchip->dev;
|
|
int err;
|
|
|
|
if (!IS_ERR(rockchip->vpcie12v)) {
|
|
err = regulator_enable(rockchip->vpcie12v);
|
|
if (err) {
|
|
dev_err(dev, "fail to enable vpcie12v regulator\n");
|
|
goto err_out;
|
|
}
|
|
}
|
|
|
|
if (!IS_ERR(rockchip->vpcie3v3)) {
|
|
err = regulator_enable(rockchip->vpcie3v3);
|
|
if (err) {
|
|
dev_err(dev, "fail to enable vpcie3v3 regulator\n");
|
|
goto err_disable_12v;
|
|
}
|
|
}
|
|
|
|
if (!IS_ERR(rockchip->vpcie1v8)) {
|
|
err = regulator_enable(rockchip->vpcie1v8);
|
|
if (err) {
|
|
dev_err(dev, "fail to enable vpcie1v8 regulator\n");
|
|
goto err_disable_3v3;
|
|
}
|
|
}
|
|
|
|
if (!IS_ERR(rockchip->vpcie0v9)) {
|
|
err = regulator_enable(rockchip->vpcie0v9);
|
|
if (err) {
|
|
dev_err(dev, "fail to enable vpcie0v9 regulator\n");
|
|
goto err_disable_1v8;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_disable_1v8:
|
|
if (!IS_ERR(rockchip->vpcie1v8))
|
|
regulator_disable(rockchip->vpcie1v8);
|
|
err_disable_3v3:
|
|
if (!IS_ERR(rockchip->vpcie3v3))
|
|
regulator_disable(rockchip->vpcie3v3);
|
|
err_disable_12v:
|
|
if (!IS_ERR(rockchip->vpcie12v))
|
|
regulator_disable(rockchip->vpcie12v);
|
|
err_out:
|
|
return err;
|
|
}
|
|
|
|
static void rockchip_pcie_enable_interrupts(struct rockchip_pcie *rockchip)
|
|
{
|
|
rockchip_pcie_write(rockchip, (PCIE_CLIENT_INT_CLI << 16) &
|
|
(~PCIE_CLIENT_INT_CLI), PCIE_CLIENT_INT_MASK);
|
|
rockchip_pcie_write(rockchip, (u32)(~PCIE_CORE_INT),
|
|
PCIE_CORE_INT_MASK);
|
|
|
|
rockchip_pcie_enable_bw_int(rockchip);
|
|
}
|
|
|
|
static int rockchip_pcie_intx_map(struct irq_domain *domain, unsigned int irq,
|
|
irq_hw_number_t hwirq)
|
|
{
|
|
irq_set_chip_and_handler(irq, &dummy_irq_chip, handle_simple_irq);
|
|
irq_set_chip_data(irq, domain->host_data);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct irq_domain_ops intx_domain_ops = {
|
|
.map = rockchip_pcie_intx_map,
|
|
};
|
|
|
|
static int rockchip_pcie_init_irq_domain(struct rockchip_pcie *rockchip)
|
|
{
|
|
struct device *dev = rockchip->dev;
|
|
struct device_node *intc = of_get_next_child(dev->of_node, NULL);
|
|
|
|
if (!intc) {
|
|
dev_err(dev, "missing child interrupt-controller node\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
rockchip->irq_domain = irq_domain_add_linear(intc, PCI_NUM_INTX,
|
|
&intx_domain_ops, rockchip);
|
|
if (!rockchip->irq_domain) {
|
|
dev_err(dev, "failed to get a INTx IRQ domain\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int rockchip_pcie_prog_ob_atu(struct rockchip_pcie *rockchip,
|
|
int region_no, int type, u8 num_pass_bits,
|
|
u32 lower_addr, u32 upper_addr)
|
|
{
|
|
u32 ob_addr_0;
|
|
u32 ob_addr_1;
|
|
u32 ob_desc_0;
|
|
u32 aw_offset;
|
|
|
|
if (region_no >= MAX_AXI_WRAPPER_REGION_NUM)
|
|
return -EINVAL;
|
|
if (num_pass_bits + 1 < 8)
|
|
return -EINVAL;
|
|
if (num_pass_bits > 63)
|
|
return -EINVAL;
|
|
if (region_no == 0) {
|
|
if (AXI_REGION_0_SIZE < (2ULL << num_pass_bits))
|
|
return -EINVAL;
|
|
}
|
|
if (region_no != 0) {
|
|
if (AXI_REGION_SIZE < (2ULL << num_pass_bits))
|
|
return -EINVAL;
|
|
}
|
|
|
|
aw_offset = (region_no << OB_REG_SIZE_SHIFT);
|
|
|
|
ob_addr_0 = num_pass_bits & PCIE_CORE_OB_REGION_ADDR0_NUM_BITS;
|
|
ob_addr_0 |= lower_addr & PCIE_CORE_OB_REGION_ADDR0_LO_ADDR;
|
|
ob_addr_1 = upper_addr;
|
|
ob_desc_0 = (1 << 23 | type);
|
|
|
|
rockchip_pcie_write(rockchip, ob_addr_0,
|
|
PCIE_CORE_OB_REGION_ADDR0 + aw_offset);
|
|
rockchip_pcie_write(rockchip, ob_addr_1,
|
|
PCIE_CORE_OB_REGION_ADDR1 + aw_offset);
|
|
rockchip_pcie_write(rockchip, ob_desc_0,
|
|
PCIE_CORE_OB_REGION_DESC0 + aw_offset);
|
|
rockchip_pcie_write(rockchip, 0,
|
|
PCIE_CORE_OB_REGION_DESC1 + aw_offset);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int rockchip_pcie_prog_ib_atu(struct rockchip_pcie *rockchip,
|
|
int region_no, u8 num_pass_bits,
|
|
u32 lower_addr, u32 upper_addr)
|
|
{
|
|
u32 ib_addr_0;
|
|
u32 ib_addr_1;
|
|
u32 aw_offset;
|
|
|
|
if (region_no > MAX_AXI_IB_ROOTPORT_REGION_NUM)
|
|
return -EINVAL;
|
|
if (num_pass_bits + 1 < MIN_AXI_ADDR_BITS_PASSED)
|
|
return -EINVAL;
|
|
if (num_pass_bits > 63)
|
|
return -EINVAL;
|
|
|
|
aw_offset = (region_no << IB_ROOT_PORT_REG_SIZE_SHIFT);
|
|
|
|
ib_addr_0 = num_pass_bits & PCIE_CORE_IB_REGION_ADDR0_NUM_BITS;
|
|
ib_addr_0 |= (lower_addr << 8) & PCIE_CORE_IB_REGION_ADDR0_LO_ADDR;
|
|
ib_addr_1 = upper_addr;
|
|
|
|
rockchip_pcie_write(rockchip, ib_addr_0, PCIE_RP_IB_ADDR0 + aw_offset);
|
|
rockchip_pcie_write(rockchip, ib_addr_1, PCIE_RP_IB_ADDR1 + aw_offset);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int rockchip_pcie_cfg_atu(struct rockchip_pcie *rockchip)
|
|
{
|
|
struct device *dev = rockchip->dev;
|
|
int offset;
|
|
int err;
|
|
int reg_no;
|
|
|
|
rockchip_pcie_cfg_configuration_accesses(rockchip,
|
|
AXI_WRAPPER_TYPE0_CFG);
|
|
|
|
for (reg_no = 0; reg_no < (rockchip->mem_size >> 20); reg_no++) {
|
|
err = rockchip_pcie_prog_ob_atu(rockchip, reg_no + 1,
|
|
AXI_WRAPPER_MEM_WRITE,
|
|
20 - 1,
|
|
rockchip->mem_bus_addr +
|
|
(reg_no << 20),
|
|
0);
|
|
if (err) {
|
|
dev_err(dev, "program RC mem outbound ATU failed\n");
|
|
return err;
|
|
}
|
|
}
|
|
|
|
err = rockchip_pcie_prog_ib_atu(rockchip, 2, 32 - 1, 0x0, 0);
|
|
if (err) {
|
|
dev_err(dev, "program RC mem inbound ATU failed\n");
|
|
return err;
|
|
}
|
|
|
|
offset = rockchip->mem_size >> 20;
|
|
for (reg_no = 0; reg_no < (rockchip->io_size >> 20); reg_no++) {
|
|
err = rockchip_pcie_prog_ob_atu(rockchip,
|
|
reg_no + 1 + offset,
|
|
AXI_WRAPPER_IO_WRITE,
|
|
20 - 1,
|
|
rockchip->io_bus_addr +
|
|
(reg_no << 20),
|
|
0);
|
|
if (err) {
|
|
dev_err(dev, "program RC io outbound ATU failed\n");
|
|
return err;
|
|
}
|
|
}
|
|
|
|
/* assign message regions */
|
|
rockchip_pcie_prog_ob_atu(rockchip, reg_no + 1 + offset,
|
|
AXI_WRAPPER_NOR_MSG,
|
|
20 - 1, 0, 0);
|
|
|
|
rockchip->msg_bus_addr = rockchip->mem_bus_addr +
|
|
((reg_no + offset) << 20);
|
|
return err;
|
|
}
|
|
|
|
static int rockchip_pcie_wait_l2(struct rockchip_pcie *rockchip)
|
|
{
|
|
u32 value;
|
|
int err;
|
|
|
|
/* send PME_TURN_OFF message */
|
|
writel(0x0, rockchip->msg_region + PCIE_RC_SEND_PME_OFF);
|
|
|
|
/* read LTSSM and wait for falling into L2 link state */
|
|
err = readl_poll_timeout(rockchip->apb_base + PCIE_CLIENT_DEBUG_OUT_0,
|
|
value, PCIE_LINK_IS_L2(value), 20,
|
|
jiffies_to_usecs(5 * HZ));
|
|
if (err) {
|
|
dev_err(rockchip->dev, "PCIe link enter L2 timeout!\n");
|
|
return err;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int rockchip_pcie_enable_clocks(struct rockchip_pcie *rockchip)
|
|
{
|
|
struct device *dev = rockchip->dev;
|
|
int err;
|
|
|
|
err = clk_prepare_enable(rockchip->aclk_pcie);
|
|
if (err) {
|
|
dev_err(dev, "unable to enable aclk_pcie clock\n");
|
|
return err;
|
|
}
|
|
|
|
err = clk_prepare_enable(rockchip->aclk_perf_pcie);
|
|
if (err) {
|
|
dev_err(dev, "unable to enable aclk_perf_pcie clock\n");
|
|
goto err_aclk_perf_pcie;
|
|
}
|
|
|
|
err = clk_prepare_enable(rockchip->hclk_pcie);
|
|
if (err) {
|
|
dev_err(dev, "unable to enable hclk_pcie clock\n");
|
|
goto err_hclk_pcie;
|
|
}
|
|
|
|
err = clk_prepare_enable(rockchip->clk_pcie_pm);
|
|
if (err) {
|
|
dev_err(dev, "unable to enable clk_pcie_pm clock\n");
|
|
goto err_clk_pcie_pm;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_clk_pcie_pm:
|
|
clk_disable_unprepare(rockchip->hclk_pcie);
|
|
err_hclk_pcie:
|
|
clk_disable_unprepare(rockchip->aclk_perf_pcie);
|
|
err_aclk_perf_pcie:
|
|
clk_disable_unprepare(rockchip->aclk_pcie);
|
|
return err;
|
|
}
|
|
|
|
static void rockchip_pcie_disable_clocks(void *data)
|
|
{
|
|
struct rockchip_pcie *rockchip = data;
|
|
|
|
clk_disable_unprepare(rockchip->clk_pcie_pm);
|
|
clk_disable_unprepare(rockchip->hclk_pcie);
|
|
clk_disable_unprepare(rockchip->aclk_perf_pcie);
|
|
clk_disable_unprepare(rockchip->aclk_pcie);
|
|
}
|
|
|
|
static int __maybe_unused rockchip_pcie_suspend_noirq(struct device *dev)
|
|
{
|
|
struct rockchip_pcie *rockchip = dev_get_drvdata(dev);
|
|
int ret;
|
|
|
|
/* disable core and cli int since we don't need to ack PME_ACK */
|
|
rockchip_pcie_write(rockchip, (PCIE_CLIENT_INT_CLI << 16) |
|
|
PCIE_CLIENT_INT_CLI, PCIE_CLIENT_INT_MASK);
|
|
rockchip_pcie_write(rockchip, (u32)PCIE_CORE_INT, PCIE_CORE_INT_MASK);
|
|
|
|
ret = rockchip_pcie_wait_l2(rockchip);
|
|
if (ret) {
|
|
rockchip_pcie_enable_interrupts(rockchip);
|
|
return ret;
|
|
}
|
|
|
|
rockchip_pcie_deinit_phys(rockchip);
|
|
|
|
rockchip_pcie_disable_clocks(rockchip);
|
|
|
|
if (!IS_ERR(rockchip->vpcie0v9))
|
|
regulator_disable(rockchip->vpcie0v9);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int __maybe_unused rockchip_pcie_resume_noirq(struct device *dev)
|
|
{
|
|
struct rockchip_pcie *rockchip = dev_get_drvdata(dev);
|
|
int err;
|
|
|
|
if (!IS_ERR(rockchip->vpcie0v9)) {
|
|
err = regulator_enable(rockchip->vpcie0v9);
|
|
if (err) {
|
|
dev_err(dev, "fail to enable vpcie0v9 regulator\n");
|
|
return err;
|
|
}
|
|
}
|
|
|
|
err = rockchip_pcie_enable_clocks(rockchip);
|
|
if (err)
|
|
goto err_disable_0v9;
|
|
|
|
err = rockchip_pcie_init_port(rockchip);
|
|
if (err)
|
|
goto err_pcie_resume;
|
|
|
|
err = rockchip_pcie_cfg_atu(rockchip);
|
|
if (err)
|
|
goto err_err_deinit_port;
|
|
|
|
/* Need this to enter L1 again */
|
|
rockchip_pcie_update_txcredit_mui(rockchip);
|
|
rockchip_pcie_enable_interrupts(rockchip);
|
|
|
|
return 0;
|
|
|
|
err_err_deinit_port:
|
|
rockchip_pcie_deinit_phys(rockchip);
|
|
err_pcie_resume:
|
|
rockchip_pcie_disable_clocks(rockchip);
|
|
err_disable_0v9:
|
|
if (!IS_ERR(rockchip->vpcie0v9))
|
|
regulator_disable(rockchip->vpcie0v9);
|
|
return err;
|
|
}
|
|
|
|
static int rockchip_pcie_probe(struct platform_device *pdev)
|
|
{
|
|
struct rockchip_pcie *rockchip;
|
|
struct device *dev = &pdev->dev;
|
|
struct pci_bus *bus, *child;
|
|
struct pci_host_bridge *bridge;
|
|
struct resource_entry *win;
|
|
resource_size_t io_base;
|
|
struct resource *mem;
|
|
struct resource *io;
|
|
int err;
|
|
|
|
LIST_HEAD(res);
|
|
|
|
if (!dev->of_node)
|
|
return -ENODEV;
|
|
|
|
bridge = devm_pci_alloc_host_bridge(dev, sizeof(*rockchip));
|
|
if (!bridge)
|
|
return -ENOMEM;
|
|
|
|
rockchip = pci_host_bridge_priv(bridge);
|
|
|
|
platform_set_drvdata(pdev, rockchip);
|
|
rockchip->dev = dev;
|
|
|
|
err = rockchip_pcie_parse_dt(rockchip);
|
|
if (err)
|
|
return err;
|
|
|
|
err = rockchip_pcie_enable_clocks(rockchip);
|
|
if (err)
|
|
return err;
|
|
|
|
err = rockchip_pcie_set_vpcie(rockchip);
|
|
if (err) {
|
|
dev_err(dev, "failed to set vpcie regulator\n");
|
|
goto err_set_vpcie;
|
|
}
|
|
|
|
err = rockchip_pcie_init_port(rockchip);
|
|
if (err)
|
|
goto err_vpcie;
|
|
|
|
rockchip_pcie_enable_interrupts(rockchip);
|
|
|
|
err = rockchip_pcie_init_irq_domain(rockchip);
|
|
if (err < 0)
|
|
goto err_deinit_port;
|
|
|
|
err = of_pci_get_host_bridge_resources(dev->of_node, 0, 0xff,
|
|
&res, &io_base);
|
|
if (err)
|
|
goto err_remove_irq_domain;
|
|
|
|
err = devm_request_pci_bus_resources(dev, &res);
|
|
if (err)
|
|
goto err_free_res;
|
|
|
|
/* Get the I/O and memory ranges from DT */
|
|
resource_list_for_each_entry(win, &res) {
|
|
switch (resource_type(win->res)) {
|
|
case IORESOURCE_IO:
|
|
io = win->res;
|
|
io->name = "I/O";
|
|
rockchip->io_size = resource_size(io);
|
|
rockchip->io_bus_addr = io->start - win->offset;
|
|
err = pci_remap_iospace(io, io_base);
|
|
if (err) {
|
|
dev_warn(dev, "error %d: failed to map resource %pR\n",
|
|
err, io);
|
|
continue;
|
|
}
|
|
rockchip->io = io;
|
|
break;
|
|
case IORESOURCE_MEM:
|
|
mem = win->res;
|
|
mem->name = "MEM";
|
|
rockchip->mem_size = resource_size(mem);
|
|
rockchip->mem_bus_addr = mem->start - win->offset;
|
|
break;
|
|
case IORESOURCE_BUS:
|
|
rockchip->root_bus_nr = win->res->start;
|
|
break;
|
|
default:
|
|
continue;
|
|
}
|
|
}
|
|
|
|
err = rockchip_pcie_cfg_atu(rockchip);
|
|
if (err)
|
|
goto err_unmap_iospace;
|
|
|
|
rockchip->msg_region = devm_ioremap(dev, rockchip->msg_bus_addr, SZ_1M);
|
|
if (!rockchip->msg_region) {
|
|
err = -ENOMEM;
|
|
goto err_unmap_iospace;
|
|
}
|
|
|
|
list_splice_init(&res, &bridge->windows);
|
|
bridge->dev.parent = dev;
|
|
bridge->sysdata = rockchip;
|
|
bridge->busnr = 0;
|
|
bridge->ops = &rockchip_pcie_ops;
|
|
bridge->map_irq = of_irq_parse_and_map_pci;
|
|
bridge->swizzle_irq = pci_common_swizzle;
|
|
|
|
err = pci_scan_root_bus_bridge(bridge);
|
|
if (err < 0)
|
|
goto err_unmap_iospace;
|
|
|
|
bus = bridge->bus;
|
|
|
|
rockchip->root_bus = bus;
|
|
|
|
pci_bus_size_bridges(bus);
|
|
pci_bus_assign_resources(bus);
|
|
list_for_each_entry(child, &bus->children, node)
|
|
pcie_bus_configure_settings(child);
|
|
|
|
pci_bus_add_devices(bus);
|
|
return 0;
|
|
|
|
err_unmap_iospace:
|
|
pci_unmap_iospace(rockchip->io);
|
|
err_free_res:
|
|
pci_free_resource_list(&res);
|
|
err_remove_irq_domain:
|
|
irq_domain_remove(rockchip->irq_domain);
|
|
err_deinit_port:
|
|
rockchip_pcie_deinit_phys(rockchip);
|
|
err_vpcie:
|
|
if (!IS_ERR(rockchip->vpcie12v))
|
|
regulator_disable(rockchip->vpcie12v);
|
|
if (!IS_ERR(rockchip->vpcie3v3))
|
|
regulator_disable(rockchip->vpcie3v3);
|
|
if (!IS_ERR(rockchip->vpcie1v8))
|
|
regulator_disable(rockchip->vpcie1v8);
|
|
if (!IS_ERR(rockchip->vpcie0v9))
|
|
regulator_disable(rockchip->vpcie0v9);
|
|
err_set_vpcie:
|
|
rockchip_pcie_disable_clocks(rockchip);
|
|
return err;
|
|
}
|
|
|
|
static int rockchip_pcie_remove(struct platform_device *pdev)
|
|
{
|
|
struct device *dev = &pdev->dev;
|
|
struct rockchip_pcie *rockchip = dev_get_drvdata(dev);
|
|
|
|
pci_stop_root_bus(rockchip->root_bus);
|
|
pci_remove_root_bus(rockchip->root_bus);
|
|
pci_unmap_iospace(rockchip->io);
|
|
irq_domain_remove(rockchip->irq_domain);
|
|
|
|
rockchip_pcie_deinit_phys(rockchip);
|
|
|
|
rockchip_pcie_disable_clocks(rockchip);
|
|
|
|
if (!IS_ERR(rockchip->vpcie12v))
|
|
regulator_disable(rockchip->vpcie12v);
|
|
if (!IS_ERR(rockchip->vpcie3v3))
|
|
regulator_disable(rockchip->vpcie3v3);
|
|
if (!IS_ERR(rockchip->vpcie1v8))
|
|
regulator_disable(rockchip->vpcie1v8);
|
|
if (!IS_ERR(rockchip->vpcie0v9))
|
|
regulator_disable(rockchip->vpcie0v9);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct dev_pm_ops rockchip_pcie_pm_ops = {
|
|
SET_NOIRQ_SYSTEM_SLEEP_PM_OPS(rockchip_pcie_suspend_noirq,
|
|
rockchip_pcie_resume_noirq)
|
|
};
|
|
|
|
static const struct of_device_id rockchip_pcie_of_match[] = {
|
|
{ .compatible = "rockchip,rk3399-pcie", },
|
|
{}
|
|
};
|
|
MODULE_DEVICE_TABLE(of, rockchip_pcie_of_match);
|
|
|
|
static struct platform_driver rockchip_pcie_driver = {
|
|
.driver = {
|
|
.name = "rockchip-pcie",
|
|
.of_match_table = rockchip_pcie_of_match,
|
|
.pm = &rockchip_pcie_pm_ops,
|
|
},
|
|
.probe = rockchip_pcie_probe,
|
|
.remove = rockchip_pcie_remove,
|
|
};
|
|
module_platform_driver(rockchip_pcie_driver);
|
|
|
|
MODULE_AUTHOR("Rockchip Inc");
|
|
MODULE_DESCRIPTION("Rockchip AXI PCIe driver");
|
|
MODULE_LICENSE("GPL v2");
|