| /* |
| * Rockchip AXI PCIe host controller driver |
| * |
| * Copyright (c) 2016 Rockchip, Inc. |
| * |
| * Author: Shawn Lin <shawn.lin@rock-chips.com> |
| * Wenrui Li <wenrui.li@rock-chips.com> |
| * |
| * Bits taken from Synopsys Designware Host controller driver and |
| * ARM PCI Host generic driver. |
| * |
| * This program is free software: you can redistribute it and/or modify |
| * it under the terms of the GNU General Public License as published by |
| * the Free Software Foundation, either version 2 of the License, or |
| * (at your option) any later version. |
| */ |
| |
| #include <linux/bitrev.h> |
| #include <linux/clk.h> |
| #include <linux/delay.h> |
| #include <linux/gpio/consumer.h> |
| #include <linux/init.h> |
| #include <linux/interrupt.h> |
| #include <linux/iopoll.h> |
| #include <linux/irq.h> |
| #include <linux/irqchip/chained_irq.h> |
| #include <linux/irqdomain.h> |
| #include <linux/kernel.h> |
| #include <linux/mfd/syscon.h> |
| #include <linux/module.h> |
| #include <linux/of_address.h> |
| #include <linux/of_device.h> |
| #include <linux/of_pci.h> |
| #include <linux/of_platform.h> |
| #include <linux/of_irq.h> |
| #include <linux/pci.h> |
| #include <linux/pci_ids.h> |
| #include <linux/phy/phy.h> |
| #include <linux/platform_device.h> |
| #include <linux/reset.h> |
| #include <linux/regmap.h> |
| |
| /* |
| * The upper 16 bits of PCIE_CLIENT_CONFIG are a write mask for the lower 16 |
| * bits. This allows atomic updates of the register without locking. |
| */ |
| #define HIWORD_UPDATE(mask, val) (((mask) << 16) | (val)) |
| #define HIWORD_UPDATE_BIT(val) HIWORD_UPDATE(val, val) |
| |
| #define ENCODE_LANES(x) ((((x) >> 1) & 3) << 4) |
| #define MAX_LANE_NUM 4 |
| |
| #define PCIE_CLIENT_BASE 0x0 |
| #define PCIE_CLIENT_CONFIG (PCIE_CLIENT_BASE + 0x00) |
| #define PCIE_CLIENT_CONF_ENABLE HIWORD_UPDATE_BIT(0x0001) |
| #define PCIE_CLIENT_LINK_TRAIN_ENABLE HIWORD_UPDATE_BIT(0x0002) |
| #define PCIE_CLIENT_ARI_ENABLE HIWORD_UPDATE_BIT(0x0008) |
| #define PCIE_CLIENT_CONF_LANE_NUM(x) HIWORD_UPDATE(0x0030, ENCODE_LANES(x)) |
| #define PCIE_CLIENT_MODE_RC HIWORD_UPDATE_BIT(0x0040) |
| #define PCIE_CLIENT_GEN_SEL_1 HIWORD_UPDATE(0x0080, 0) |
| #define PCIE_CLIENT_GEN_SEL_2 HIWORD_UPDATE_BIT(0x0080) |
| #define PCIE_CLIENT_DEBUG_OUT_0 (PCIE_CLIENT_BASE + 0x3c) |
| #define PCIE_CLIENT_DEBUG_LTSSM_MASK GENMASK(5, 0) |
| #define PCIE_CLIENT_DEBUG_LTSSM_L1 0x18 |
| #define PCIE_CLIENT_DEBUG_LTSSM_L2 0x19 |
| #define PCIE_CLIENT_BASIC_STATUS1 (PCIE_CLIENT_BASE + 0x48) |
| #define PCIE_CLIENT_LINK_STATUS_UP 0x00300000 |
| #define PCIE_CLIENT_LINK_STATUS_MASK 0x00300000 |
| #define PCIE_CLIENT_INT_MASK (PCIE_CLIENT_BASE + 0x4c) |
| #define PCIE_CLIENT_INT_STATUS (PCIE_CLIENT_BASE + 0x50) |
| #define PCIE_CLIENT_INTR_MASK GENMASK(8, 5) |
| #define PCIE_CLIENT_INTR_SHIFT 5 |
| #define PCIE_CLIENT_INT_LEGACY_DONE BIT(15) |
| #define PCIE_CLIENT_INT_MSG BIT(14) |
| #define PCIE_CLIENT_INT_HOT_RST BIT(13) |
| #define PCIE_CLIENT_INT_DPA BIT(12) |
| #define PCIE_CLIENT_INT_FATAL_ERR BIT(11) |
| #define PCIE_CLIENT_INT_NFATAL_ERR BIT(10) |
| #define PCIE_CLIENT_INT_CORR_ERR BIT(9) |
| #define PCIE_CLIENT_INT_INTD BIT(8) |
| #define PCIE_CLIENT_INT_INTC BIT(7) |
| #define PCIE_CLIENT_INT_INTB BIT(6) |
| #define PCIE_CLIENT_INT_INTA BIT(5) |
| #define PCIE_CLIENT_INT_LOCAL BIT(4) |
| #define PCIE_CLIENT_INT_UDMA BIT(3) |
| #define PCIE_CLIENT_INT_PHY BIT(2) |
| #define PCIE_CLIENT_INT_HOT_PLUG BIT(1) |
| #define PCIE_CLIENT_INT_PWR_STCG BIT(0) |
| |
| #define PCIE_CLIENT_INT_LEGACY \ |
| (PCIE_CLIENT_INT_INTA | PCIE_CLIENT_INT_INTB | \ |
| PCIE_CLIENT_INT_INTC | PCIE_CLIENT_INT_INTD) |
| |
| #define PCIE_CLIENT_INT_CLI \ |
| (PCIE_CLIENT_INT_CORR_ERR | PCIE_CLIENT_INT_NFATAL_ERR | \ |
| PCIE_CLIENT_INT_FATAL_ERR | PCIE_CLIENT_INT_DPA | \ |
| PCIE_CLIENT_INT_HOT_RST | PCIE_CLIENT_INT_MSG | \ |
| PCIE_CLIENT_INT_LEGACY_DONE | PCIE_CLIENT_INT_LEGACY | \ |
| PCIE_CLIENT_INT_PHY) |
| |
| #define PCIE_CORE_CTRL_MGMT_BASE 0x900000 |
| #define PCIE_CORE_CTRL (PCIE_CORE_CTRL_MGMT_BASE + 0x000) |
| #define PCIE_CORE_PL_CONF_SPEED_5G 0x00000008 |
| #define PCIE_CORE_PL_CONF_SPEED_MASK 0x00000018 |
| #define PCIE_CORE_PL_CONF_LANE_MASK 0x00000006 |
| #define PCIE_CORE_PL_CONF_LANE_SHIFT 1 |
| #define PCIE_CORE_CTRL_PLC1 (PCIE_CORE_CTRL_MGMT_BASE + 0x004) |
| #define PCIE_CORE_CTRL_PLC1_FTS_MASK GENMASK(23, 8) |
| #define PCIE_CORE_CTRL_PLC1_FTS_SHIFT 8 |
| #define PCIE_CORE_CTRL_PLC1_FTS_CNT 0xffff |
| #define PCIE_CORE_TXCREDIT_CFG1 (PCIE_CORE_CTRL_MGMT_BASE + 0x020) |
| #define PCIE_CORE_TXCREDIT_CFG1_MUI_MASK 0xFFFF0000 |
| #define PCIE_CORE_TXCREDIT_CFG1_MUI_SHIFT 16 |
| #define PCIE_CORE_TXCREDIT_CFG1_MUI_ENCODE(x) \ |
| (((x) >> 3) << PCIE_CORE_TXCREDIT_CFG1_MUI_SHIFT) |
| #define PCIE_CORE_LANE_MAP (PCIE_CORE_CTRL_MGMT_BASE + 0x200) |
| #define PCIE_CORE_LANE_MAP_MASK 0x0000000f |
| #define PCIE_CORE_LANE_MAP_REVERSE BIT(16) |
| #define PCIE_CORE_INT_STATUS (PCIE_CORE_CTRL_MGMT_BASE + 0x20c) |
| #define PCIE_CORE_INT_PRFPE BIT(0) |
| #define PCIE_CORE_INT_CRFPE BIT(1) |
| #define PCIE_CORE_INT_RRPE BIT(2) |
| #define PCIE_CORE_INT_PRFO BIT(3) |
| #define PCIE_CORE_INT_CRFO BIT(4) |
| #define PCIE_CORE_INT_RT BIT(5) |
| #define PCIE_CORE_INT_RTR BIT(6) |
| #define PCIE_CORE_INT_PE BIT(7) |
| #define PCIE_CORE_INT_MTR BIT(8) |
| #define PCIE_CORE_INT_UCR BIT(9) |
| #define PCIE_CORE_INT_FCE BIT(10) |
| #define PCIE_CORE_INT_CT BIT(11) |
| #define PCIE_CORE_INT_UTC BIT(18) |
| #define PCIE_CORE_INT_MMVC BIT(19) |
| #define PCIE_CORE_CONFIG_VENDOR (PCIE_CORE_CTRL_MGMT_BASE + 0x44) |
| #define PCIE_CORE_INT_MASK (PCIE_CORE_CTRL_MGMT_BASE + 0x210) |
| #define PCIE_RC_BAR_CONF (PCIE_CORE_CTRL_MGMT_BASE + 0x300) |
| |
| #define PCIE_CORE_INT \ |
| (PCIE_CORE_INT_PRFPE | PCIE_CORE_INT_CRFPE | \ |
| PCIE_CORE_INT_RRPE | PCIE_CORE_INT_CRFO | \ |
| PCIE_CORE_INT_RT | PCIE_CORE_INT_RTR | \ |
| PCIE_CORE_INT_PE | PCIE_CORE_INT_MTR | \ |
| PCIE_CORE_INT_UCR | PCIE_CORE_INT_FCE | \ |
| PCIE_CORE_INT_CT | PCIE_CORE_INT_UTC | \ |
| PCIE_CORE_INT_MMVC) |
| |
| #define PCIE_RC_CONFIG_NORMAL_BASE 0x800000 |
| #define PCIE_RC_CONFIG_BASE 0xa00000 |
| #define PCIE_RC_CONFIG_RID_CCR (PCIE_RC_CONFIG_BASE + 0x08) |
| #define PCIE_RC_CONFIG_SCC_SHIFT 16 |
| #define PCIE_RC_CONFIG_DCR (PCIE_RC_CONFIG_BASE + 0xc4) |
| #define PCIE_RC_CONFIG_DCR_CSPL_SHIFT 18 |
| #define PCIE_RC_CONFIG_DCR_CSPL_LIMIT 0xff |
| #define PCIE_RC_CONFIG_DCR_CPLS_SHIFT 26 |
| #define PCIE_RC_CONFIG_DCSR (PCIE_RC_CONFIG_BASE + 0xc8) |
| #define PCIE_RC_CONFIG_DCSR_MPS_MASK GENMASK(7, 5) |
| #define PCIE_RC_CONFIG_DCSR_MPS_256 (0x1 << 5) |
| #define PCIE_RC_CONFIG_LINK_CAP (PCIE_RC_CONFIG_BASE + 0xcc) |
| #define PCIE_RC_CONFIG_LINK_CAP_L0S BIT(10) |
| #define PCIE_RC_CONFIG_LCS (PCIE_RC_CONFIG_BASE + 0xd0) |
| #define PCIE_RC_CONFIG_L1_SUBSTATE_CTRL2 (PCIE_RC_CONFIG_BASE + 0x90c) |
| #define PCIE_RC_CONFIG_THP_CAP (PCIE_RC_CONFIG_BASE + 0x274) |
| #define PCIE_RC_CONFIG_THP_CAP_NEXT_MASK GENMASK(31, 20) |
| |
| #define PCIE_CORE_AXI_CONF_BASE 0xc00000 |
| #define PCIE_CORE_OB_REGION_ADDR0 (PCIE_CORE_AXI_CONF_BASE + 0x0) |
| #define PCIE_CORE_OB_REGION_ADDR0_NUM_BITS 0x3f |
| #define PCIE_CORE_OB_REGION_ADDR0_LO_ADDR 0xffffff00 |
| #define PCIE_CORE_OB_REGION_ADDR1 (PCIE_CORE_AXI_CONF_BASE + 0x4) |
| #define PCIE_CORE_OB_REGION_DESC0 (PCIE_CORE_AXI_CONF_BASE + 0x8) |
| #define PCIE_CORE_OB_REGION_DESC1 (PCIE_CORE_AXI_CONF_BASE + 0xc) |
| |
| #define PCIE_CORE_AXI_INBOUND_BASE 0xc00800 |
| #define PCIE_RP_IB_ADDR0 (PCIE_CORE_AXI_INBOUND_BASE + 0x0) |
| #define PCIE_CORE_IB_REGION_ADDR0_NUM_BITS 0x3f |
| #define PCIE_CORE_IB_REGION_ADDR0_LO_ADDR 0xffffff00 |
| #define PCIE_RP_IB_ADDR1 (PCIE_CORE_AXI_INBOUND_BASE + 0x4) |
| |
| /* Size of one AXI Region (not Region 0) */ |
| #define AXI_REGION_SIZE BIT(20) |
| /* Size of Region 0, equal to sum of sizes of other regions */ |
| #define AXI_REGION_0_SIZE (32 * (0x1 << 20)) |
| #define OB_REG_SIZE_SHIFT 5 |
| #define IB_ROOT_PORT_REG_SIZE_SHIFT 3 |
| #define AXI_WRAPPER_IO_WRITE 0x6 |
| #define AXI_WRAPPER_MEM_WRITE 0x2 |
| #define AXI_WRAPPER_TYPE0_CFG 0xa |
| #define AXI_WRAPPER_TYPE1_CFG 0xb |
| #define AXI_WRAPPER_NOR_MSG 0xc |
| |
| #define MAX_AXI_IB_ROOTPORT_REGION_NUM 3 |
| #define MIN_AXI_ADDR_BITS_PASSED 8 |
| #define PCIE_RC_SEND_PME_OFF 0x11960 |
| #define ROCKCHIP_VENDOR_ID 0x1d87 |
| #define PCIE_ECAM_BUS(x) (((x) & 0xff) << 20) |
| #define PCIE_ECAM_DEV(x) (((x) & 0x1f) << 15) |
| #define PCIE_ECAM_FUNC(x) (((x) & 0x7) << 12) |
| #define PCIE_ECAM_REG(x) (((x) & 0xfff) << 0) |
| #define PCIE_ECAM_ADDR(bus, dev, func, reg) \ |
| (PCIE_ECAM_BUS(bus) | PCIE_ECAM_DEV(dev) | \ |
| PCIE_ECAM_FUNC(func) | PCIE_ECAM_REG(reg)) |
| #define PCIE_LINK_IS_L2(x) \ |
| (((x) & PCIE_CLIENT_DEBUG_LTSSM_MASK) == PCIE_CLIENT_DEBUG_LTSSM_L2) |
| #define PCIE_LINK_UP(x) \ |
| (((x) & PCIE_CLIENT_LINK_STATUS_MASK) == PCIE_CLIENT_LINK_STATUS_UP) |
| #define PCIE_LINK_IS_GEN2(x) \ |
| (((x) & PCIE_CORE_PL_CONF_SPEED_MASK) == PCIE_CORE_PL_CONF_SPEED_5G) |
| |
| #define RC_REGION_0_ADDR_TRANS_H 0x00000000 |
| #define RC_REGION_0_ADDR_TRANS_L 0x00000000 |
| #define RC_REGION_0_PASS_BITS (25 - 1) |
| #define RC_REGION_0_TYPE_MASK GENMASK(3, 0) |
| #define MAX_AXI_WRAPPER_REGION_NUM 33 |
| |
| struct rockchip_pcie { |
| void __iomem *reg_base; /* DT axi-base */ |
| void __iomem *apb_base; /* DT apb-base */ |
| bool legacy_phy; |
| struct phy *phys[MAX_LANE_NUM]; |
| struct reset_control *core_rst; |
| struct reset_control *mgmt_rst; |
| struct reset_control *mgmt_sticky_rst; |
| struct reset_control *pipe_rst; |
| struct reset_control *pm_rst; |
| struct reset_control *aclk_rst; |
| struct reset_control *pclk_rst; |
| struct clk *aclk_pcie; |
| struct clk *aclk_perf_pcie; |
| struct clk *hclk_pcie; |
| struct clk *clk_pcie_pm; |
| struct regulator *vpcie12v; /* 12V power supply */ |
| struct regulator *vpcie3v3; /* 3.3V power supply */ |
| struct regulator *vpcie1v8; /* 1.8V power supply */ |
| struct regulator *vpcie0v9; /* 0.9V power supply */ |
| struct gpio_desc *ep_gpio; |
| u32 lanes; |
| u8 lanes_map; |
| u8 root_bus_nr; |
| int link_gen; |
| struct device *dev; |
| struct irq_domain *irq_domain; |
| int offset; |
| struct pci_bus *root_bus; |
| struct resource *io; |
| phys_addr_t io_bus_addr; |
| u32 io_size; |
| void __iomem *msg_region; |
| u32 mem_size; |
| phys_addr_t msg_bus_addr; |
| phys_addr_t mem_bus_addr; |
| }; |
| |
| static u32 rockchip_pcie_read(struct rockchip_pcie *rockchip, u32 reg) |
| { |
| return readl(rockchip->apb_base + reg); |
| } |
| |
| static void rockchip_pcie_write(struct rockchip_pcie *rockchip, u32 val, |
| u32 reg) |
| { |
| writel(val, rockchip->apb_base + reg); |
| } |
| |
| static void rockchip_pcie_enable_bw_int(struct rockchip_pcie *rockchip) |
| { |
| u32 status; |
| |
| status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_LCS); |
| status |= (PCI_EXP_LNKCTL_LBMIE | PCI_EXP_LNKCTL_LABIE); |
| rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_LCS); |
| } |
| |
| static void rockchip_pcie_clr_bw_int(struct rockchip_pcie *rockchip) |
| { |
| u32 status; |
| |
| status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_LCS); |
| status |= (PCI_EXP_LNKSTA_LBMS | PCI_EXP_LNKSTA_LABS) << 16; |
| rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_LCS); |
| } |
| |
| static void rockchip_pcie_update_txcredit_mui(struct rockchip_pcie *rockchip) |
| { |
| u32 val; |
| |
| /* Update Tx credit maximum update interval */ |
| val = rockchip_pcie_read(rockchip, PCIE_CORE_TXCREDIT_CFG1); |
| val &= ~PCIE_CORE_TXCREDIT_CFG1_MUI_MASK; |
| val |= PCIE_CORE_TXCREDIT_CFG1_MUI_ENCODE(24000); /* ns */ |
| rockchip_pcie_write(rockchip, val, PCIE_CORE_TXCREDIT_CFG1); |
| } |
| |
| static int rockchip_pcie_valid_device(struct rockchip_pcie *rockchip, |
| struct pci_bus *bus, int dev) |
| { |
| /* access only one slot on each root port */ |
| if (bus->number == rockchip->root_bus_nr && dev > 0) |
| return 0; |
| |
| /* |
| * do not read more than one device on the bus directly attached |
| * to RC's downstream side. |
| */ |
| if (bus->primary == rockchip->root_bus_nr && dev > 0) |
| return 0; |
| |
| return 1; |
| } |
| |
| static u8 rockchip_pcie_lane_map(struct rockchip_pcie *rockchip) |
| { |
| u32 val; |
| u8 map; |
| |
| if (rockchip->legacy_phy) |
| return GENMASK(MAX_LANE_NUM - 1, 0); |
| |
| val = rockchip_pcie_read(rockchip, PCIE_CORE_LANE_MAP); |
| map = val & PCIE_CORE_LANE_MAP_MASK; |
| |
| /* The link may be using a reverse-indexed mapping. */ |
| if (val & PCIE_CORE_LANE_MAP_REVERSE) |
| map = bitrev8(map) >> 4; |
| |
| return map; |
| } |
| |
| static int rockchip_pcie_rd_own_conf(struct rockchip_pcie *rockchip, |
| int where, int size, u32 *val) |
| { |
| void __iomem *addr; |
| |
| addr = rockchip->apb_base + PCIE_RC_CONFIG_NORMAL_BASE + where; |
| |
| if (!IS_ALIGNED((uintptr_t)addr, size)) { |
| *val = 0; |
| return PCIBIOS_BAD_REGISTER_NUMBER; |
| } |
| |
| if (size == 4) { |
| *val = readl(addr); |
| } else if (size == 2) { |
| *val = readw(addr); |
| } else if (size == 1) { |
| *val = readb(addr); |
| } else { |
| *val = 0; |
| return PCIBIOS_BAD_REGISTER_NUMBER; |
| } |
| return PCIBIOS_SUCCESSFUL; |
| } |
| |
| static int rockchip_pcie_wr_own_conf(struct rockchip_pcie *rockchip, |
| int where, int size, u32 val) |
| { |
| u32 mask, tmp, offset; |
| void __iomem *addr; |
| |
| offset = where & ~0x3; |
| addr = rockchip->apb_base + PCIE_RC_CONFIG_NORMAL_BASE + offset; |
| |
| if (size == 4) { |
| writel(val, addr); |
| return PCIBIOS_SUCCESSFUL; |
| } |
| |
| mask = ~(((1 << (size * 8)) - 1) << ((where & 0x3) * 8)); |
| |
| /* |
| * N.B. This read/modify/write isn't safe in general because it can |
| * corrupt RW1C bits in adjacent registers. But the hardware |
| * doesn't support smaller writes. |
| */ |
| tmp = readl(addr) & mask; |
| tmp |= val << ((where & 0x3) * 8); |
| writel(tmp, addr); |
| |
| return PCIBIOS_SUCCESSFUL; |
| } |
| |
| static void rockchip_pcie_cfg_configuration_accesses( |
| struct rockchip_pcie *rockchip, u32 type) |
| { |
| u32 ob_desc_0; |
| |
| /* Configuration Accesses for region 0 */ |
| rockchip_pcie_write(rockchip, 0x0, PCIE_RC_BAR_CONF); |
| |
| rockchip_pcie_write(rockchip, |
| (RC_REGION_0_ADDR_TRANS_L + RC_REGION_0_PASS_BITS), |
| PCIE_CORE_OB_REGION_ADDR0); |
| rockchip_pcie_write(rockchip, RC_REGION_0_ADDR_TRANS_H, |
| PCIE_CORE_OB_REGION_ADDR1); |
| ob_desc_0 = rockchip_pcie_read(rockchip, PCIE_CORE_OB_REGION_DESC0); |
| ob_desc_0 &= ~(RC_REGION_0_TYPE_MASK); |
| ob_desc_0 |= (type | (0x1 << 23)); |
| rockchip_pcie_write(rockchip, ob_desc_0, PCIE_CORE_OB_REGION_DESC0); |
| rockchip_pcie_write(rockchip, 0x0, PCIE_CORE_OB_REGION_DESC1); |
| } |
| |
| static int rockchip_pcie_rd_other_conf(struct rockchip_pcie *rockchip, |
| struct pci_bus *bus, u32 devfn, |
| int where, int size, u32 *val) |
| { |
| u32 busdev; |
| |
| busdev = PCIE_ECAM_ADDR(bus->number, PCI_SLOT(devfn), |
| PCI_FUNC(devfn), where); |
| |
| if (!IS_ALIGNED(busdev, size)) { |
| *val = 0; |
| return PCIBIOS_BAD_REGISTER_NUMBER; |
| } |
| |
| if (bus->parent->number == rockchip->root_bus_nr) |
| rockchip_pcie_cfg_configuration_accesses(rockchip, |
| AXI_WRAPPER_TYPE0_CFG); |
| else |
| rockchip_pcie_cfg_configuration_accesses(rockchip, |
| AXI_WRAPPER_TYPE1_CFG); |
| |
| if (size == 4) { |
| *val = readl(rockchip->reg_base + busdev); |
| } else if (size == 2) { |
| *val = readw(rockchip->reg_base + busdev); |
| } else if (size == 1) { |
| *val = readb(rockchip->reg_base + busdev); |
| } else { |
| *val = 0; |
| return PCIBIOS_BAD_REGISTER_NUMBER; |
| } |
| return PCIBIOS_SUCCESSFUL; |
| } |
| |
| static int rockchip_pcie_wr_other_conf(struct rockchip_pcie *rockchip, |
| struct pci_bus *bus, u32 devfn, |
| int where, int size, u32 val) |
| { |
| u32 busdev; |
| |
| busdev = PCIE_ECAM_ADDR(bus->number, PCI_SLOT(devfn), |
| PCI_FUNC(devfn), where); |
| if (!IS_ALIGNED(busdev, size)) |
| return PCIBIOS_BAD_REGISTER_NUMBER; |
| |
| if (bus->parent->number == rockchip->root_bus_nr) |
| rockchip_pcie_cfg_configuration_accesses(rockchip, |
| AXI_WRAPPER_TYPE0_CFG); |
| else |
| rockchip_pcie_cfg_configuration_accesses(rockchip, |
| AXI_WRAPPER_TYPE1_CFG); |
| |
| if (size == 4) |
| writel(val, rockchip->reg_base + busdev); |
| else if (size == 2) |
| writew(val, rockchip->reg_base + busdev); |
| else if (size == 1) |
| writeb(val, rockchip->reg_base + busdev); |
| else |
| return PCIBIOS_BAD_REGISTER_NUMBER; |
| |
| return PCIBIOS_SUCCESSFUL; |
| } |
| |
| static int rockchip_pcie_rd_conf(struct pci_bus *bus, u32 devfn, int where, |
| int size, u32 *val) |
| { |
| struct rockchip_pcie *rockchip = bus->sysdata; |
| |
| if (!rockchip_pcie_valid_device(rockchip, bus, PCI_SLOT(devfn))) { |
| *val = 0xffffffff; |
| return PCIBIOS_DEVICE_NOT_FOUND; |
| } |
| |
| if (bus->number == rockchip->root_bus_nr) |
| return rockchip_pcie_rd_own_conf(rockchip, where, size, val); |
| |
| return rockchip_pcie_rd_other_conf(rockchip, bus, devfn, where, size, val); |
| } |
| |
| static int rockchip_pcie_wr_conf(struct pci_bus *bus, u32 devfn, |
| int where, int size, u32 val) |
| { |
| struct rockchip_pcie *rockchip = bus->sysdata; |
| |
| if (!rockchip_pcie_valid_device(rockchip, bus, PCI_SLOT(devfn))) |
| return PCIBIOS_DEVICE_NOT_FOUND; |
| |
| if (bus->number == rockchip->root_bus_nr) |
| return rockchip_pcie_wr_own_conf(rockchip, where, size, val); |
| |
| return rockchip_pcie_wr_other_conf(rockchip, bus, devfn, where, size, val); |
| } |
| |
| static struct pci_ops rockchip_pcie_ops = { |
| .read = rockchip_pcie_rd_conf, |
| .write = rockchip_pcie_wr_conf, |
| }; |
| |
| static void rockchip_pcie_set_power_limit(struct rockchip_pcie *rockchip) |
| { |
| int curr; |
| u32 status, scale, power; |
| |
| if (IS_ERR(rockchip->vpcie3v3)) |
| return; |
| |
| /* |
| * Set RC's captured slot power limit and scale if |
| * vpcie3v3 available. The default values are both zero |
| * which means the software should set these two according |
| * to the actual power supply. |
| */ |
| curr = regulator_get_current_limit(rockchip->vpcie3v3); |
| if (curr <= 0) |
| return; |
| |
| scale = 3; /* 0.001x */ |
| curr = curr / 1000; /* convert to mA */ |
| power = (curr * 3300) / 1000; /* milliwatt */ |
| while (power > PCIE_RC_CONFIG_DCR_CSPL_LIMIT) { |
| if (!scale) { |
| dev_warn(rockchip->dev, "invalid power supply\n"); |
| return; |
| } |
| scale--; |
| power = power / 10; |
| } |
| |
| status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_DCR); |
| status |= (power << PCIE_RC_CONFIG_DCR_CSPL_SHIFT) | |
| (scale << PCIE_RC_CONFIG_DCR_CPLS_SHIFT); |
| rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_DCR); |
| } |
| |
| /** |
| * rockchip_pcie_init_port - Initialize hardware |
| * @rockchip: PCIe port information |
| */ |
| static int rockchip_pcie_init_port(struct rockchip_pcie *rockchip) |
| { |
| struct device *dev = rockchip->dev; |
| int err, i; |
| u32 status; |
| |
| gpiod_set_value_cansleep(rockchip->ep_gpio, 0); |
| |
| err = reset_control_assert(rockchip->aclk_rst); |
| if (err) { |
| dev_err(dev, "assert aclk_rst err %d\n", err); |
| return err; |
| } |
| |
| err = reset_control_assert(rockchip->pclk_rst); |
| if (err) { |
| dev_err(dev, "assert pclk_rst err %d\n", err); |
| return err; |
| } |
| |
| err = reset_control_assert(rockchip->pm_rst); |
| if (err) { |
| dev_err(dev, "assert pm_rst err %d\n", err); |
| return err; |
| } |
| |
| for (i = 0; i < MAX_LANE_NUM; i++) { |
| err = phy_init(rockchip->phys[i]); |
| if (err) { |
| dev_err(dev, "init phy%d err %d\n", i, err); |
| goto err_exit_phy; |
| } |
| } |
| |
| err = reset_control_assert(rockchip->core_rst); |
| if (err) { |
| dev_err(dev, "assert core_rst err %d\n", err); |
| goto err_exit_phy; |
| } |
| |
| err = reset_control_assert(rockchip->mgmt_rst); |
| if (err) { |
| dev_err(dev, "assert mgmt_rst err %d\n", err); |
| goto err_exit_phy; |
| } |
| |
| err = reset_control_assert(rockchip->mgmt_sticky_rst); |
| if (err) { |
| dev_err(dev, "assert mgmt_sticky_rst err %d\n", err); |
| goto err_exit_phy; |
| } |
| |
| err = reset_control_assert(rockchip->pipe_rst); |
| if (err) { |
| dev_err(dev, "assert pipe_rst err %d\n", err); |
| goto err_exit_phy; |
| } |
| |
| udelay(10); |
| |
| err = reset_control_deassert(rockchip->pm_rst); |
| if (err) { |
| dev_err(dev, "deassert pm_rst err %d\n", err); |
| goto err_exit_phy; |
| } |
| |
| err = reset_control_deassert(rockchip->aclk_rst); |
| if (err) { |
| dev_err(dev, "deassert aclk_rst err %d\n", err); |
| goto err_exit_phy; |
| } |
| |
| err = reset_control_deassert(rockchip->pclk_rst); |
| if (err) { |
| dev_err(dev, "deassert pclk_rst err %d\n", err); |
| goto err_exit_phy; |
| } |
| |
| if (rockchip->link_gen == 2) |
| rockchip_pcie_write(rockchip, PCIE_CLIENT_GEN_SEL_2, |
| PCIE_CLIENT_CONFIG); |
| else |
| rockchip_pcie_write(rockchip, PCIE_CLIENT_GEN_SEL_1, |
| PCIE_CLIENT_CONFIG); |
| |
| rockchip_pcie_write(rockchip, |
| PCIE_CLIENT_CONF_ENABLE | |
| PCIE_CLIENT_LINK_TRAIN_ENABLE | |
| PCIE_CLIENT_ARI_ENABLE | |
| PCIE_CLIENT_CONF_LANE_NUM(rockchip->lanes) | |
| PCIE_CLIENT_MODE_RC, |
| PCIE_CLIENT_CONFIG); |
| |
| for (i = 0; i < MAX_LANE_NUM; i++) { |
| err = phy_power_on(rockchip->phys[i]); |
| if (err) { |
| dev_err(dev, "power on phy%d err %d\n", i, err); |
| goto err_power_off_phy; |
| } |
| } |
| |
| /* |
| * Please don't reorder the deassert sequence of the following |
| * four reset pins. |
| */ |
| err = reset_control_deassert(rockchip->mgmt_sticky_rst); |
| if (err) { |
| dev_err(dev, "deassert mgmt_sticky_rst err %d\n", err); |
| goto err_power_off_phy; |
| } |
| |
| err = reset_control_deassert(rockchip->core_rst); |
| if (err) { |
| dev_err(dev, "deassert core_rst err %d\n", err); |
| goto err_power_off_phy; |
| } |
| |
| err = reset_control_deassert(rockchip->mgmt_rst); |
| if (err) { |
| dev_err(dev, "deassert mgmt_rst err %d\n", err); |
| goto err_power_off_phy; |
| } |
| |
| err = reset_control_deassert(rockchip->pipe_rst); |
| if (err) { |
| dev_err(dev, "deassert pipe_rst err %d\n", err); |
| goto err_power_off_phy; |
| } |
| |
| /* Fix the transmitted FTS count desired to exit from L0s. */ |
| status = rockchip_pcie_read(rockchip, PCIE_CORE_CTRL_PLC1); |
| status = (status & ~PCIE_CORE_CTRL_PLC1_FTS_MASK) | |
| (PCIE_CORE_CTRL_PLC1_FTS_CNT << PCIE_CORE_CTRL_PLC1_FTS_SHIFT); |
| rockchip_pcie_write(rockchip, status, PCIE_CORE_CTRL_PLC1); |
| |
| rockchip_pcie_set_power_limit(rockchip); |
| |
| /* Set RC's clock architecture as common clock */ |
| status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_LCS); |
| status |= PCI_EXP_LNKSTA_SLC << 16; |
| rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_LCS); |
| |
| /* Set RC's RCB to 128 */ |
| status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_LCS); |
| status |= PCI_EXP_LNKCTL_RCB; |
| rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_LCS); |
| |
| /* Enable Gen1 training */ |
| rockchip_pcie_write(rockchip, PCIE_CLIENT_LINK_TRAIN_ENABLE, |
| PCIE_CLIENT_CONFIG); |
| |
| gpiod_set_value_cansleep(rockchip->ep_gpio, 1); |
| |
| /* 500ms timeout value should be enough for Gen1/2 training */ |
| err = readl_poll_timeout(rockchip->apb_base + PCIE_CLIENT_BASIC_STATUS1, |
| status, PCIE_LINK_UP(status), 20, |
| 500 * USEC_PER_MSEC); |
| if (err) { |
| dev_err(dev, "PCIe link training gen1 timeout!\n"); |
| goto err_power_off_phy; |
| } |
| |
| if (rockchip->link_gen == 2) { |
| /* |
| * Enable retrain for gen2. This should be configured only after |
| * gen1 finished. |
| */ |
| status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_LCS); |
| status |= PCI_EXP_LNKCTL_RL; |
| rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_LCS); |
| |
| err = readl_poll_timeout(rockchip->apb_base + PCIE_CORE_CTRL, |
| status, PCIE_LINK_IS_GEN2(status), 20, |
| 500 * USEC_PER_MSEC); |
| if (err) |
| dev_dbg(dev, "PCIe link training gen2 timeout, fall back to gen1!\n"); |
| } |
| |
| /* Check the final link width from negotiated lane counter from MGMT */ |
| status = rockchip_pcie_read(rockchip, PCIE_CORE_CTRL); |
| status = 0x1 << ((status & PCIE_CORE_PL_CONF_LANE_MASK) >> |
| PCIE_CORE_PL_CONF_LANE_SHIFT); |
| dev_dbg(dev, "current link width is x%d\n", status); |
| |
| /* Power off unused lane(s) */ |
| rockchip->lanes_map = rockchip_pcie_lane_map(rockchip); |
| for (i = 0; i < MAX_LANE_NUM; i++) { |
| if (!(rockchip->lanes_map & BIT(i))) { |
| dev_dbg(dev, "idling lane %d\n", i); |
| phy_power_off(rockchip->phys[i]); |
| } |
| } |
| |
| rockchip_pcie_write(rockchip, ROCKCHIP_VENDOR_ID, |
| PCIE_CORE_CONFIG_VENDOR); |
| rockchip_pcie_write(rockchip, |
| PCI_CLASS_BRIDGE_PCI << PCIE_RC_CONFIG_SCC_SHIFT, |
| PCIE_RC_CONFIG_RID_CCR); |
| |
| /* Clear THP cap's next cap pointer to remove L1 substate cap */ |
| status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_THP_CAP); |
| status &= ~PCIE_RC_CONFIG_THP_CAP_NEXT_MASK; |
| rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_THP_CAP); |
| |
| /* Clear L0s from RC's link cap */ |
| if (of_property_read_bool(dev->of_node, "aspm-no-l0s")) { |
| status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_LINK_CAP); |
| status &= ~PCIE_RC_CONFIG_LINK_CAP_L0S; |
| rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_LINK_CAP); |
| } |
| |
| status = rockchip_pcie_read(rockchip, PCIE_RC_CONFIG_DCSR); |
| status &= ~PCIE_RC_CONFIG_DCSR_MPS_MASK; |
| status |= PCIE_RC_CONFIG_DCSR_MPS_256; |
| rockchip_pcie_write(rockchip, status, PCIE_RC_CONFIG_DCSR); |
| |
| return 0; |
| err_power_off_phy: |
| while (i--) |
| phy_power_off(rockchip->phys[i]); |
| i = MAX_LANE_NUM; |
| err_exit_phy: |
| while (i--) |
| phy_exit(rockchip->phys[i]); |
| return err; |
| } |
| |
| static void rockchip_pcie_deinit_phys(struct rockchip_pcie *rockchip) |
| { |
| int i; |
| |
| for (i = 0; i < MAX_LANE_NUM; i++) { |
| /* inactive lanes are already powered off */ |
| if (rockchip->lanes_map & BIT(i)) |
| phy_power_off(rockchip->phys[i]); |
| phy_exit(rockchip->phys[i]); |
| } |
| } |
| |
| static irqreturn_t rockchip_pcie_subsys_irq_handler(int irq, void *arg) |
| { |
| struct rockchip_pcie *rockchip = arg; |
| struct device *dev = rockchip->dev; |
| u32 reg; |
| u32 sub_reg; |
| |
| reg = rockchip_pcie_read(rockchip, PCIE_CLIENT_INT_STATUS); |
| if (reg & PCIE_CLIENT_INT_LOCAL) { |
| dev_dbg(dev, "local interrupt received\n"); |
| sub_reg = rockchip_pcie_read(rockchip, PCIE_CORE_INT_STATUS); |
| if (sub_reg & PCIE_CORE_INT_PRFPE) |
| dev_dbg(dev, "parity error detected while reading from the PNP receive FIFO RAM\n"); |
| |
| if (sub_reg & PCIE_CORE_INT_CRFPE) |
| dev_dbg(dev, "parity error detected while reading from the Completion Receive FIFO RAM\n"); |
| |
| if (sub_reg & PCIE_CORE_INT_RRPE) |
| dev_dbg(dev, "parity error detected while reading from replay buffer RAM\n"); |
| |
| if (sub_reg & PCIE_CORE_INT_PRFO) |
| dev_dbg(dev, "overflow occurred in the PNP receive FIFO\n"); |
| |
| if (sub_reg & PCIE_CORE_INT_CRFO) |
| dev_dbg(dev, "overflow occurred in the completion receive FIFO\n"); |
| |
| if (sub_reg & PCIE_CORE_INT_RT) |
| dev_dbg(dev, "replay timer timed out\n"); |
| |
| if (sub_reg & PCIE_CORE_INT_RTR) |
| dev_dbg(dev, "replay timer rolled over after 4 transmissions of the same TLP\n"); |
| |
| if (sub_reg & PCIE_CORE_INT_PE) |
| dev_dbg(dev, "phy error detected on receive side\n"); |
| |
| if (sub_reg & PCIE_CORE_INT_MTR) |
| dev_dbg(dev, "malformed TLP received from the link\n"); |
| |
| if (sub_reg & PCIE_CORE_INT_UCR) |
| dev_dbg(dev, "malformed TLP received from the link\n"); |
| |
| if (sub_reg & PCIE_CORE_INT_FCE) |
| dev_dbg(dev, "an error was observed in the flow control advertisements from the other side\n"); |
| |
| if (sub_reg & PCIE_CORE_INT_CT) |
| dev_dbg(dev, "a request timed out waiting for completion\n"); |
| |
| if (sub_reg & PCIE_CORE_INT_UTC) |
| dev_dbg(dev, "unmapped TC error\n"); |
| |
| if (sub_reg & PCIE_CORE_INT_MMVC) |
| dev_dbg(dev, "MSI mask register changes\n"); |
| |
| rockchip_pcie_write(rockchip, sub_reg, PCIE_CORE_INT_STATUS); |
| } else if (reg & PCIE_CLIENT_INT_PHY) { |
| dev_dbg(dev, "phy link changes\n"); |
| rockchip_pcie_update_txcredit_mui(rockchip); |
| rockchip_pcie_clr_bw_int(rockchip); |
| } |
| |
| rockchip_pcie_write(rockchip, reg & PCIE_CLIENT_INT_LOCAL, |
| PCIE_CLIENT_INT_STATUS); |
| |
| return IRQ_HANDLED; |
| } |
| |
| static irqreturn_t rockchip_pcie_client_irq_handler(int irq, void *arg) |
| { |
| struct rockchip_pcie *rockchip = arg; |
| struct device *dev = rockchip->dev; |
| u32 reg; |
| |
| reg = rockchip_pcie_read(rockchip, PCIE_CLIENT_INT_STATUS); |
| if (reg & PCIE_CLIENT_INT_LEGACY_DONE) |
| dev_dbg(dev, "legacy done interrupt received\n"); |
| |
| if (reg & PCIE_CLIENT_INT_MSG) |
| dev_dbg(dev, "message done interrupt received\n"); |
| |
| if (reg & PCIE_CLIENT_INT_HOT_RST) |
| dev_dbg(dev, "hot reset interrupt received\n"); |
| |
| if (reg & PCIE_CLIENT_INT_DPA) |
| dev_dbg(dev, "dpa interrupt received\n"); |
| |
| if (reg & PCIE_CLIENT_INT_FATAL_ERR) |
| dev_dbg(dev, "fatal error interrupt received\n"); |
| |
| if (reg & PCIE_CLIENT_INT_NFATAL_ERR) |
| dev_dbg(dev, "no fatal error interrupt received\n"); |
| |
| if (reg & PCIE_CLIENT_INT_CORR_ERR) |
| dev_dbg(dev, "correctable error interrupt received\n"); |
| |
| if (reg & PCIE_CLIENT_INT_PHY) |
| dev_dbg(dev, "phy interrupt received\n"); |
| |
| rockchip_pcie_write(rockchip, reg & (PCIE_CLIENT_INT_LEGACY_DONE | |
| PCIE_CLIENT_INT_MSG | PCIE_CLIENT_INT_HOT_RST | |
| PCIE_CLIENT_INT_DPA | PCIE_CLIENT_INT_FATAL_ERR | |
| PCIE_CLIENT_INT_NFATAL_ERR | |
| PCIE_CLIENT_INT_CORR_ERR | |
| PCIE_CLIENT_INT_PHY), |
| PCIE_CLIENT_INT_STATUS); |
| |
| return IRQ_HANDLED; |
| } |
| |
| static void rockchip_pcie_legacy_int_handler(struct irq_desc *desc) |
| { |
| struct irq_chip *chip = irq_desc_get_chip(desc); |
| struct rockchip_pcie *rockchip = irq_desc_get_handler_data(desc); |
| struct device *dev = rockchip->dev; |
| u32 reg; |
| u32 hwirq; |
| u32 virq; |
| |
| chained_irq_enter(chip, desc); |
| |
| reg = rockchip_pcie_read(rockchip, PCIE_CLIENT_INT_STATUS); |
| reg = (reg & PCIE_CLIENT_INTR_MASK) >> PCIE_CLIENT_INTR_SHIFT; |
| |
| while (reg) { |
| hwirq = ffs(reg) - 1; |
| reg &= ~BIT(hwirq); |
| |
| virq = irq_find_mapping(rockchip->irq_domain, hwirq); |
| if (virq) |
| generic_handle_irq(virq); |
| else |
| dev_err(dev, "unexpected IRQ, INT%d\n", hwirq); |
| } |
| |
| chained_irq_exit(chip, desc); |
| } |
| |
| static int rockchip_pcie_get_phys(struct rockchip_pcie *rockchip) |
| { |
| struct device *dev = rockchip->dev; |
| struct phy *phy; |
| char *name; |
| u32 i; |
| |
| phy = devm_phy_get(dev, "pcie-phy"); |
| if (!IS_ERR(phy)) { |
| rockchip->legacy_phy = true; |
| rockchip->phys[0] = phy; |
| dev_warn(dev, "legacy phy model is deprecated!\n"); |
| return 0; |
| } |
| |
| if (PTR_ERR(phy) == -EPROBE_DEFER) |
| return PTR_ERR(phy); |
| |
| dev_dbg(dev, "missing legacy phy; search for per-lane PHY\n"); |
| |
| for (i = 0; i < MAX_LANE_NUM; i++) { |
| name = kasprintf(GFP_KERNEL, "pcie-phy-%u", i); |
| if (!name) |
| return -ENOMEM; |
| |
| phy = devm_of_phy_get(dev, dev->of_node, name); |
| kfree(name); |
| |
| if (IS_ERR(phy)) { |
| if (PTR_ERR(phy) != -EPROBE_DEFER) |
| dev_err(dev, "missing phy for lane %d: %ld\n", |
| i, PTR_ERR(phy)); |
| return PTR_ERR(phy); |
| } |
| |
| rockchip->phys[i] = phy; |
| } |
| |
| return 0; |
| } |
| |
| static int rockchip_pcie_setup_irq(struct rockchip_pcie *rockchip) |
| { |
| int irq, err; |
| struct device *dev = rockchip->dev; |
| struct platform_device *pdev = to_platform_device(dev); |
| |
| irq = platform_get_irq_byname(pdev, "sys"); |
| if (irq < 0) { |
| dev_err(dev, "missing sys IRQ resource\n"); |
| return -EINVAL; |
| } |
| |
| err = devm_request_irq(dev, irq, rockchip_pcie_subsys_irq_handler, |
| IRQF_SHARED, "pcie-sys", rockchip); |
| if (err) { |
| dev_err(dev, "failed to request PCIe subsystem IRQ\n"); |
| return err; |
| } |
| |
| irq = platform_get_irq_byname(pdev, "legacy"); |
| if (irq < 0) { |
| dev_err(dev, "missing legacy IRQ resource\n"); |
| return -EINVAL; |
| } |
| |
| irq_set_chained_handler_and_data(irq, |
| rockchip_pcie_legacy_int_handler, |
| rockchip); |
| |
| irq = platform_get_irq_byname(pdev, "client"); |
| if (irq < 0) { |
| dev_err(dev, "missing client IRQ resource\n"); |
| return -EINVAL; |
| } |
| |
| err = devm_request_irq(dev, irq, rockchip_pcie_client_irq_handler, |
| IRQF_SHARED, "pcie-client", rockchip); |
| if (err) { |
| dev_err(dev, "failed to request PCIe client IRQ\n"); |
| return err; |
| } |
| |
| return 0; |
| } |
| |
| /** |
| * rockchip_pcie_parse_dt - Parse Device Tree |
| * @rockchip: PCIe port information |
| * |
| * Return: '0' on success and error value on failure |
| */ |
| static int rockchip_pcie_parse_dt(struct rockchip_pcie *rockchip) |
| { |
| struct device *dev = rockchip->dev; |
| struct platform_device *pdev = to_platform_device(dev); |
| struct device_node *node = dev->of_node; |
| struct resource *regs; |
| int err; |
| |
| regs = platform_get_resource_byname(pdev, |
| IORESOURCE_MEM, |
| "axi-base"); |
| rockchip->reg_base = devm_pci_remap_cfg_resource(dev, regs); |
| if (IS_ERR(rockchip->reg_base)) |
| return PTR_ERR(rockchip->reg_base); |
| |
| regs = platform_get_resource_byname(pdev, |
| IORESOURCE_MEM, |
| "apb-base"); |
| rockchip->apb_base = devm_ioremap_resource(dev, regs); |
| if (IS_ERR(rockchip->apb_base)) |
| return PTR_ERR(rockchip->apb_base); |
| |
| err = rockchip_pcie_get_phys(rockchip); |
| if (err) |
| return err; |
| |
| rockchip->lanes = 1; |
| err = of_property_read_u32(node, "num-lanes", &rockchip->lanes); |
| if (!err && (rockchip->lanes == 0 || |
| rockchip->lanes == 3 || |
| rockchip->lanes > 4)) { |
| dev_warn(dev, "invalid num-lanes, default to use one lane\n"); |
| rockchip->lanes = 1; |
| } |
| |
| rockchip->link_gen = of_pci_get_max_link_speed(node); |
| if (rockchip->link_gen < 0 || rockchip->link_gen > 2) |
| rockchip->link_gen = 2; |
| |
| rockchip->core_rst = devm_reset_control_get_exclusive(dev, "core"); |
| if (IS_ERR(rockchip->core_rst)) { |
| if (PTR_ERR(rockchip->core_rst) != -EPROBE_DEFER) |
| dev_err(dev, "missing core reset property in node\n"); |
| return PTR_ERR(rockchip->core_rst); |
| } |
| |
| rockchip->mgmt_rst = devm_reset_control_get_exclusive(dev, "mgmt"); |
| if (IS_ERR(rockchip->mgmt_rst)) { |
| if (PTR_ERR(rockchip->mgmt_rst) != -EPROBE_DEFER) |
| dev_err(dev, "missing mgmt reset property in node\n"); |
| return PTR_ERR(rockchip->mgmt_rst); |
| } |
| |
| rockchip->mgmt_sticky_rst = devm_reset_control_get_exclusive(dev, |
| "mgmt-sticky"); |
| if (IS_ERR(rockchip->mgmt_sticky_rst)) { |
| if (PTR_ERR(rockchip->mgmt_sticky_rst) != -EPROBE_DEFER) |
| dev_err(dev, "missing mgmt-sticky reset property in node\n"); |
| return PTR_ERR(rockchip->mgmt_sticky_rst); |
| } |
| |
| rockchip->pipe_rst = devm_reset_control_get_exclusive(dev, "pipe"); |
| if (IS_ERR(rockchip->pipe_rst)) { |
| if (PTR_ERR(rockchip->pipe_rst) != -EPROBE_DEFER) |
| dev_err(dev, "missing pipe reset property in node\n"); |
| return PTR_ERR(rockchip->pipe_rst); |
| } |
| |
| rockchip->pm_rst = devm_reset_control_get_exclusive(dev, "pm"); |
| if (IS_ERR(rockchip->pm_rst)) { |
| if (PTR_ERR(rockchip->pm_rst) != -EPROBE_DEFER) |
| dev_err(dev, "missing pm reset property in node\n"); |
| return PTR_ERR(rockchip->pm_rst); |
| } |
| |
| rockchip->pclk_rst = devm_reset_control_get_exclusive(dev, "pclk"); |
| if (IS_ERR(rockchip->pclk_rst)) { |
| if (PTR_ERR(rockchip->pclk_rst) != -EPROBE_DEFER) |
| dev_err(dev, "missing pclk reset property in node\n"); |
| return PTR_ERR(rockchip->pclk_rst); |
| } |
| |
| rockchip->aclk_rst = devm_reset_control_get_exclusive(dev, "aclk"); |
| if (IS_ERR(rockchip->aclk_rst)) { |
| if (PTR_ERR(rockchip->aclk_rst) != -EPROBE_DEFER) |
| dev_err(dev, "missing aclk reset property in node\n"); |
| return PTR_ERR(rockchip->aclk_rst); |
| } |
| |
| rockchip->ep_gpio = devm_gpiod_get(dev, "ep", GPIOD_OUT_HIGH); |
| if (IS_ERR(rockchip->ep_gpio)) { |
| dev_err(dev, "missing ep-gpios property in node\n"); |
| return PTR_ERR(rockchip->ep_gpio); |
| } |
| |
| rockchip->aclk_pcie = devm_clk_get(dev, "aclk"); |
| if (IS_ERR(rockchip->aclk_pcie)) { |
| dev_err(dev, "aclk clock not found\n"); |
| return PTR_ERR(rockchip->aclk_pcie); |
| } |
| |
| rockchip->aclk_perf_pcie = devm_clk_get(dev, "aclk-perf"); |
| if (IS_ERR(rockchip->aclk_perf_pcie)) { |
| dev_err(dev, "aclk_perf clock not found\n"); |
| return PTR_ERR(rockchip->aclk_perf_pcie); |
| } |
| |
| rockchip->hclk_pcie = devm_clk_get(dev, "hclk"); |
| if (IS_ERR(rockchip->hclk_pcie)) { |
| dev_err(dev, "hclk clock not found\n"); |
| return PTR_ERR(rockchip->hclk_pcie); |
| } |
| |
| rockchip->clk_pcie_pm = devm_clk_get(dev, "pm"); |
| if (IS_ERR(rockchip->clk_pcie_pm)) { |
| dev_err(dev, "pm clock not found\n"); |
| return PTR_ERR(rockchip->clk_pcie_pm); |
| } |
| |
| err = rockchip_pcie_setup_irq(rockchip); |
| if (err) |
| return err; |
| |
| rockchip->vpcie12v = devm_regulator_get_optional(dev, "vpcie12v"); |
| if (IS_ERR(rockchip->vpcie12v)) { |
| if (PTR_ERR(rockchip->vpcie12v) == -EPROBE_DEFER) |
| return -EPROBE_DEFER; |
| dev_info(dev, "no vpcie12v regulator found\n"); |
| } |
| |
| rockchip->vpcie3v3 = devm_regulator_get_optional(dev, "vpcie3v3"); |
| if (IS_ERR(rockchip->vpcie3v3)) { |
| if (PTR_ERR(rockchip->vpcie3v3) == -EPROBE_DEFER) |
| return -EPROBE_DEFER; |
| dev_info(dev, "no vpcie3v3 regulator found\n"); |
| } |
| |
| rockchip->vpcie1v8 = devm_regulator_get_optional(dev, "vpcie1v8"); |
| if (IS_ERR(rockchip->vpcie1v8)) { |
| if (PTR_ERR(rockchip->vpcie1v8) == -EPROBE_DEFER) |
| return -EPROBE_DEFER; |
| dev_info(dev, "no vpcie1v8 regulator found\n"); |
| } |
| |
| rockchip->vpcie0v9 = devm_regulator_get_optional(dev, "vpcie0v9"); |
| if (IS_ERR(rockchip->vpcie0v9)) { |
| if (PTR_ERR(rockchip->vpcie0v9) == -EPROBE_DEFER) |
| return -EPROBE_DEFER; |
| dev_info(dev, "no vpcie0v9 regulator found\n"); |
| } |
| |
| return 0; |
| } |
| |
| static int rockchip_pcie_set_vpcie(struct rockchip_pcie *rockchip) |
| { |
| struct device *dev = rockchip->dev; |
| int err; |
| |
| if (!IS_ERR(rockchip->vpcie12v)) { |
| err = regulator_enable(rockchip->vpcie12v); |
| if (err) { |
| dev_err(dev, "fail to enable vpcie12v regulator\n"); |
| goto err_out; |
| } |
| } |
| |
| if (!IS_ERR(rockchip->vpcie3v3)) { |
| err = regulator_enable(rockchip->vpcie3v3); |
| if (err) { |
| dev_err(dev, "fail to enable vpcie3v3 regulator\n"); |
| goto err_disable_12v; |
| } |
| } |
| |
| if (!IS_ERR(rockchip->vpcie1v8)) { |
| err = regulator_enable(rockchip->vpcie1v8); |
| if (err) { |
| dev_err(dev, "fail to enable vpcie1v8 regulator\n"); |
| goto err_disable_3v3; |
| } |
| } |
| |
| if (!IS_ERR(rockchip->vpcie0v9)) { |
| err = regulator_enable(rockchip->vpcie0v9); |
| if (err) { |
| dev_err(dev, "fail to enable vpcie0v9 regulator\n"); |
| goto err_disable_1v8; |
| } |
| } |
| |
| return 0; |
| |
| err_disable_1v8: |
| if (!IS_ERR(rockchip->vpcie1v8)) |
| regulator_disable(rockchip->vpcie1v8); |
| err_disable_3v3: |
| if (!IS_ERR(rockchip->vpcie3v3)) |
| regulator_disable(rockchip->vpcie3v3); |
| err_disable_12v: |
| if (!IS_ERR(rockchip->vpcie12v)) |
| regulator_disable(rockchip->vpcie12v); |
| err_out: |
| return err; |
| } |
| |
| static void rockchip_pcie_enable_interrupts(struct rockchip_pcie *rockchip) |
| { |
| rockchip_pcie_write(rockchip, (PCIE_CLIENT_INT_CLI << 16) & |
| (~PCIE_CLIENT_INT_CLI), PCIE_CLIENT_INT_MASK); |
| rockchip_pcie_write(rockchip, (u32)(~PCIE_CORE_INT), |
| PCIE_CORE_INT_MASK); |
| |
| rockchip_pcie_enable_bw_int(rockchip); |
| } |
| |
| static int rockchip_pcie_intx_map(struct irq_domain *domain, unsigned int irq, |
| irq_hw_number_t hwirq) |
| { |
| irq_set_chip_and_handler(irq, &dummy_irq_chip, handle_simple_irq); |
| irq_set_chip_data(irq, domain->host_data); |
| |
| return 0; |
| } |
| |
| static const struct irq_domain_ops intx_domain_ops = { |
| .map = rockchip_pcie_intx_map, |
| }; |
| |
| static int rockchip_pcie_init_irq_domain(struct rockchip_pcie *rockchip) |
| { |
| struct device *dev = rockchip->dev; |
| struct device_node *intc = of_get_next_child(dev->of_node, NULL); |
| |
| if (!intc) { |
| dev_err(dev, "missing child interrupt-controller node\n"); |
| return -EINVAL; |
| } |
| |
| rockchip->irq_domain = irq_domain_add_linear(intc, PCI_NUM_INTX, |
| &intx_domain_ops, rockchip); |
| if (!rockchip->irq_domain) { |
| dev_err(dev, "failed to get a INTx IRQ domain\n"); |
| return -EINVAL; |
| } |
| |
| return 0; |
| } |
| |
| static int rockchip_pcie_prog_ob_atu(struct rockchip_pcie *rockchip, |
| int region_no, int type, u8 num_pass_bits, |
| u32 lower_addr, u32 upper_addr) |
| { |
| u32 ob_addr_0; |
| u32 ob_addr_1; |
| u32 ob_desc_0; |
| u32 aw_offset; |
| |
| if (region_no >= MAX_AXI_WRAPPER_REGION_NUM) |
| return -EINVAL; |
| if (num_pass_bits + 1 < 8) |
| return -EINVAL; |
| if (num_pass_bits > 63) |
| return -EINVAL; |
| if (region_no == 0) { |
| if (AXI_REGION_0_SIZE < (2ULL << num_pass_bits)) |
| return -EINVAL; |
| } |
| if (region_no != 0) { |
| if (AXI_REGION_SIZE < (2ULL << num_pass_bits)) |
| return -EINVAL; |
| } |
| |
| aw_offset = (region_no << OB_REG_SIZE_SHIFT); |
| |
| ob_addr_0 = num_pass_bits & PCIE_CORE_OB_REGION_ADDR0_NUM_BITS; |
| ob_addr_0 |= lower_addr & PCIE_CORE_OB_REGION_ADDR0_LO_ADDR; |
| ob_addr_1 = upper_addr; |
| ob_desc_0 = (1 << 23 | type); |
| |
| rockchip_pcie_write(rockchip, ob_addr_0, |
| PCIE_CORE_OB_REGION_ADDR0 + aw_offset); |
| rockchip_pcie_write(rockchip, ob_addr_1, |
| PCIE_CORE_OB_REGION_ADDR1 + aw_offset); |
| rockchip_pcie_write(rockchip, ob_desc_0, |
| PCIE_CORE_OB_REGION_DESC0 + aw_offset); |
| rockchip_pcie_write(rockchip, 0, |
| PCIE_CORE_OB_REGION_DESC1 + aw_offset); |
| |
| return 0; |
| } |
| |
| static int rockchip_pcie_prog_ib_atu(struct rockchip_pcie *rockchip, |
| int region_no, u8 num_pass_bits, |
| u32 lower_addr, u32 upper_addr) |
| { |
| u32 ib_addr_0; |
| u32 ib_addr_1; |
| u32 aw_offset; |
| |
| if (region_no > MAX_AXI_IB_ROOTPORT_REGION_NUM) |
| return -EINVAL; |
| if (num_pass_bits + 1 < MIN_AXI_ADDR_BITS_PASSED) |
| return -EINVAL; |
| if (num_pass_bits > 63) |
| return -EINVAL; |
| |
| aw_offset = (region_no << IB_ROOT_PORT_REG_SIZE_SHIFT); |
| |
| ib_addr_0 = num_pass_bits & PCIE_CORE_IB_REGION_ADDR0_NUM_BITS; |
| ib_addr_0 |= (lower_addr << 8) & PCIE_CORE_IB_REGION_ADDR0_LO_ADDR; |
| ib_addr_1 = upper_addr; |
| |
| rockchip_pcie_write(rockchip, ib_addr_0, PCIE_RP_IB_ADDR0 + aw_offset); |
| rockchip_pcie_write(rockchip, ib_addr_1, PCIE_RP_IB_ADDR1 + aw_offset); |
| |
| return 0; |
| } |
| |
| static int rockchip_pcie_cfg_atu(struct rockchip_pcie *rockchip) |
| { |
| struct device *dev = rockchip->dev; |
| int offset; |
| int err; |
| int reg_no; |
| |
| rockchip_pcie_cfg_configuration_accesses(rockchip, |
| AXI_WRAPPER_TYPE0_CFG); |
| |
| for (reg_no = 0; reg_no < (rockchip->mem_size >> 20); reg_no++) { |
| err = rockchip_pcie_prog_ob_atu(rockchip, reg_no + 1, |
| AXI_WRAPPER_MEM_WRITE, |
| 20 - 1, |
| rockchip->mem_bus_addr + |
| (reg_no << 20), |
| 0); |
| if (err) { |
| dev_err(dev, "program RC mem outbound ATU failed\n"); |
| return err; |
| } |
| } |
| |
| err = rockchip_pcie_prog_ib_atu(rockchip, 2, 32 - 1, 0x0, 0); |
| if (err) { |
| dev_err(dev, "program RC mem inbound ATU failed\n"); |
| return err; |
| } |
| |
| offset = rockchip->mem_size >> 20; |
| for (reg_no = 0; reg_no < (rockchip->io_size >> 20); reg_no++) { |
| err = rockchip_pcie_prog_ob_atu(rockchip, |
| reg_no + 1 + offset, |
| AXI_WRAPPER_IO_WRITE, |
| 20 - 1, |
| rockchip->io_bus_addr + |
| (reg_no << 20), |
| 0); |
| if (err) { |
| dev_err(dev, "program RC io outbound ATU failed\n"); |
| return err; |
| } |
| } |
| |
| /* assign message regions */ |
| rockchip_pcie_prog_ob_atu(rockchip, reg_no + 1 + offset, |
| AXI_WRAPPER_NOR_MSG, |
| 20 - 1, 0, 0); |
| |
| rockchip->msg_bus_addr = rockchip->mem_bus_addr + |
| ((reg_no + offset) << 20); |
| return err; |
| } |
| |
| static int rockchip_pcie_wait_l2(struct rockchip_pcie *rockchip) |
| { |
| u32 value; |
| int err; |
| |
| /* send PME_TURN_OFF message */ |
| writel(0x0, rockchip->msg_region + PCIE_RC_SEND_PME_OFF); |
| |
| /* read LTSSM and wait for falling into L2 link state */ |
| err = readl_poll_timeout(rockchip->apb_base + PCIE_CLIENT_DEBUG_OUT_0, |
| value, PCIE_LINK_IS_L2(value), 20, |
| jiffies_to_usecs(5 * HZ)); |
| if (err) { |
| dev_err(rockchip->dev, "PCIe link enter L2 timeout!\n"); |
| return err; |
| } |
| |
| return 0; |
| } |
| |
| static int rockchip_pcie_enable_clocks(struct rockchip_pcie *rockchip) |
| { |
| struct device *dev = rockchip->dev; |
| int err; |
| |
| err = clk_prepare_enable(rockchip->aclk_pcie); |
| if (err) { |
| dev_err(dev, "unable to enable aclk_pcie clock\n"); |
| return err; |
| } |
| |
| err = clk_prepare_enable(rockchip->aclk_perf_pcie); |
| if (err) { |
| dev_err(dev, "unable to enable aclk_perf_pcie clock\n"); |
| goto err_aclk_perf_pcie; |
| } |
| |
| err = clk_prepare_enable(rockchip->hclk_pcie); |
| if (err) { |
| dev_err(dev, "unable to enable hclk_pcie clock\n"); |
| goto err_hclk_pcie; |
| } |
| |
| err = clk_prepare_enable(rockchip->clk_pcie_pm); |
| if (err) { |
| dev_err(dev, "unable to enable clk_pcie_pm clock\n"); |
| goto err_clk_pcie_pm; |
| } |
| |
| return 0; |
| |
| err_clk_pcie_pm: |
| clk_disable_unprepare(rockchip->hclk_pcie); |
| err_hclk_pcie: |
| clk_disable_unprepare(rockchip->aclk_perf_pcie); |
| err_aclk_perf_pcie: |
| clk_disable_unprepare(rockchip->aclk_pcie); |
| return err; |
| } |
| |
| static void rockchip_pcie_disable_clocks(void *data) |
| { |
| struct rockchip_pcie *rockchip = data; |
| |
| clk_disable_unprepare(rockchip->clk_pcie_pm); |
| clk_disable_unprepare(rockchip->hclk_pcie); |
| clk_disable_unprepare(rockchip->aclk_perf_pcie); |
| clk_disable_unprepare(rockchip->aclk_pcie); |
| } |
| |
| static int __maybe_unused rockchip_pcie_suspend_noirq(struct device *dev) |
| { |
| struct rockchip_pcie *rockchip = dev_get_drvdata(dev); |
| int ret; |
| |
| /* disable core and cli int since we don't need to ack PME_ACK */ |
| rockchip_pcie_write(rockchip, (PCIE_CLIENT_INT_CLI << 16) | |
| PCIE_CLIENT_INT_CLI, PCIE_CLIENT_INT_MASK); |
| rockchip_pcie_write(rockchip, (u32)PCIE_CORE_INT, PCIE_CORE_INT_MASK); |
| |
| ret = rockchip_pcie_wait_l2(rockchip); |
| if (ret) { |
| rockchip_pcie_enable_interrupts(rockchip); |
| return ret; |
| } |
| |
| rockchip_pcie_deinit_phys(rockchip); |
| |
| rockchip_pcie_disable_clocks(rockchip); |
| |
| if (!IS_ERR(rockchip->vpcie0v9)) |
| regulator_disable(rockchip->vpcie0v9); |
| |
| return ret; |
| } |
| |
| static int __maybe_unused rockchip_pcie_resume_noirq(struct device *dev) |
| { |
| struct rockchip_pcie *rockchip = dev_get_drvdata(dev); |
| int err; |
| |
| if (!IS_ERR(rockchip->vpcie0v9)) { |
| err = regulator_enable(rockchip->vpcie0v9); |
| if (err) { |
| dev_err(dev, "fail to enable vpcie0v9 regulator\n"); |
| return err; |
| } |
| } |
| |
| err = rockchip_pcie_enable_clocks(rockchip); |
| if (err) |
| goto err_disable_0v9; |
| |
| err = rockchip_pcie_init_port(rockchip); |
| if (err) |
| goto err_pcie_resume; |
| |
| err = rockchip_pcie_cfg_atu(rockchip); |
| if (err) |
| goto err_err_deinit_port; |
| |
| /* Need this to enter L1 again */ |
| rockchip_pcie_update_txcredit_mui(rockchip); |
| rockchip_pcie_enable_interrupts(rockchip); |
| |
| return 0; |
| |
| err_err_deinit_port: |
| rockchip_pcie_deinit_phys(rockchip); |
| err_pcie_resume: |
| rockchip_pcie_disable_clocks(rockchip); |
| err_disable_0v9: |
| if (!IS_ERR(rockchip->vpcie0v9)) |
| regulator_disable(rockchip->vpcie0v9); |
| return err; |
| } |
| |
| static int rockchip_pcie_probe(struct platform_device *pdev) |
| { |
| struct rockchip_pcie *rockchip; |
| struct device *dev = &pdev->dev; |
| struct pci_bus *bus, *child; |
| struct pci_host_bridge *bridge; |
| struct resource_entry *win; |
| resource_size_t io_base; |
| struct resource *mem; |
| struct resource *io; |
| int err; |
| |
| LIST_HEAD(res); |
| |
| if (!dev->of_node) |
| return -ENODEV; |
| |
| bridge = devm_pci_alloc_host_bridge(dev, sizeof(*rockchip)); |
| if (!bridge) |
| return -ENOMEM; |
| |
| rockchip = pci_host_bridge_priv(bridge); |
| |
| platform_set_drvdata(pdev, rockchip); |
| rockchip->dev = dev; |
| |
| err = rockchip_pcie_parse_dt(rockchip); |
| if (err) |
| return err; |
| |
| err = rockchip_pcie_enable_clocks(rockchip); |
| if (err) |
| return err; |
| |
| err = rockchip_pcie_set_vpcie(rockchip); |
| if (err) { |
| dev_err(dev, "failed to set vpcie regulator\n"); |
| goto err_set_vpcie; |
| } |
| |
| err = rockchip_pcie_init_port(rockchip); |
| if (err) |
| goto err_vpcie; |
| |
| rockchip_pcie_enable_interrupts(rockchip); |
| |
| err = rockchip_pcie_init_irq_domain(rockchip); |
| if (err < 0) |
| goto err_deinit_port; |
| |
| err = of_pci_get_host_bridge_resources(dev->of_node, 0, 0xff, |
| &res, &io_base); |
| if (err) |
| goto err_deinit_port; |
| |
| err = devm_request_pci_bus_resources(dev, &res); |
| if (err) |
| goto err_free_res; |
| |
| /* Get the I/O and memory ranges from DT */ |
| resource_list_for_each_entry(win, &res) { |
| switch (resource_type(win->res)) { |
| case IORESOURCE_IO: |
| io = win->res; |
| io->name = "I/O"; |
| rockchip->io_size = resource_size(io); |
| rockchip->io_bus_addr = io->start - win->offset; |
| err = pci_remap_iospace(io, io_base); |
| if (err) { |
| dev_warn(dev, "error %d: failed to map resource %pR\n", |
| err, io); |
| continue; |
| } |
| rockchip->io = io; |
| break; |
| case IORESOURCE_MEM: |
| mem = win->res; |
| mem->name = "MEM"; |
| rockchip->mem_size = resource_size(mem); |
| rockchip->mem_bus_addr = mem->start - win->offset; |
| break; |
| case IORESOURCE_BUS: |
| rockchip->root_bus_nr = win->res->start; |
| break; |
| default: |
| continue; |
| } |
| } |
| |
| err = rockchip_pcie_cfg_atu(rockchip); |
| if (err) |
| goto err_free_res; |
| |
| rockchip->msg_region = devm_ioremap(dev, rockchip->msg_bus_addr, SZ_1M); |
| if (!rockchip->msg_region) { |
| err = -ENOMEM; |
| goto err_free_res; |
| } |
| |
| list_splice_init(&res, &bridge->windows); |
| bridge->dev.parent = dev; |
| bridge->sysdata = rockchip; |
| bridge->busnr = 0; |
| bridge->ops = &rockchip_pcie_ops; |
| bridge->map_irq = of_irq_parse_and_map_pci; |
| bridge->swizzle_irq = pci_common_swizzle; |
| |
| err = pci_scan_root_bus_bridge(bridge); |
| if (err < 0) |
| goto err_free_res; |
| |
| bus = bridge->bus; |
| |
| rockchip->root_bus = bus; |
| |
| pci_bus_size_bridges(bus); |
| pci_bus_assign_resources(bus); |
| list_for_each_entry(child, &bus->children, node) |
| pcie_bus_configure_settings(child); |
| |
| pci_bus_add_devices(bus); |
| return 0; |
| |
| err_free_res: |
| pci_free_resource_list(&res); |
| err_deinit_port: |
| rockchip_pcie_deinit_phys(rockchip); |
| err_vpcie: |
| if (!IS_ERR(rockchip->vpcie12v)) |
| regulator_disable(rockchip->vpcie12v); |
| if (!IS_ERR(rockchip->vpcie3v3)) |
| regulator_disable(rockchip->vpcie3v3); |
| if (!IS_ERR(rockchip->vpcie1v8)) |
| regulator_disable(rockchip->vpcie1v8); |
| if (!IS_ERR(rockchip->vpcie0v9)) |
| regulator_disable(rockchip->vpcie0v9); |
| err_set_vpcie: |
| rockchip_pcie_disable_clocks(rockchip); |
| return err; |
| } |
| |
| static int rockchip_pcie_remove(struct platform_device *pdev) |
| { |
| struct device *dev = &pdev->dev; |
| struct rockchip_pcie *rockchip = dev_get_drvdata(dev); |
| |
| pci_stop_root_bus(rockchip->root_bus); |
| pci_remove_root_bus(rockchip->root_bus); |
| pci_unmap_iospace(rockchip->io); |
| irq_domain_remove(rockchip->irq_domain); |
| |
| rockchip_pcie_deinit_phys(rockchip); |
| |
| rockchip_pcie_disable_clocks(rockchip); |
| |
| if (!IS_ERR(rockchip->vpcie12v)) |
| regulator_disable(rockchip->vpcie12v); |
| if (!IS_ERR(rockchip->vpcie3v3)) |
| regulator_disable(rockchip->vpcie3v3); |
| if (!IS_ERR(rockchip->vpcie1v8)) |
| regulator_disable(rockchip->vpcie1v8); |
| if (!IS_ERR(rockchip->vpcie0v9)) |
| regulator_disable(rockchip->vpcie0v9); |
| |
| return 0; |
| } |
| |
| static const struct dev_pm_ops rockchip_pcie_pm_ops = { |
| SET_NOIRQ_SYSTEM_SLEEP_PM_OPS(rockchip_pcie_suspend_noirq, |
| rockchip_pcie_resume_noirq) |
| }; |
| |
| static const struct of_device_id rockchip_pcie_of_match[] = { |
| { .compatible = "rockchip,rk3399-pcie", }, |
| {} |
| }; |
| MODULE_DEVICE_TABLE(of, rockchip_pcie_of_match); |
| |
| static struct platform_driver rockchip_pcie_driver = { |
| .driver = { |
| .name = "rockchip-pcie", |
| .of_match_table = rockchip_pcie_of_match, |
| .pm = &rockchip_pcie_pm_ops, |
| }, |
| .probe = rockchip_pcie_probe, |
| .remove = rockchip_pcie_remove, |
| }; |
| module_platform_driver(rockchip_pcie_driver); |
| |
| MODULE_AUTHOR("Rockchip Inc"); |
| MODULE_DESCRIPTION("Rockchip AXI PCIe driver"); |
| MODULE_LICENSE("GPL v2"); |