// SPDX-License-Identifier: GPL-2.0 /* * PCIe host controller driver for Rockchip SoCs * * Copyright (C) 2018 Rockchip Electronics Co., Ltd. * http://www.rock-chips.com * * Author: Simon Xue */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "pcie-designware.h" #include "../../pci.h" #include "../rockchip-pcie-dma.h" enum rk_pcie_device_mode { RK_PCIE_EP_TYPE, RK_PCIE_RC_TYPE, }; struct reset_bulk_data { const char *id; struct reset_control *rst; }; #define PCIE_DMA_OFFSET 0x380000 #define PCIE_DMA_WR_ENB 0xc #define PCIE_DMA_WR_CTRL_LO 0x200 #define PCIE_DMA_WR_CTRL_HI 0x204 #define PCIE_DMA_WR_XFERSIZE 0x208 #define PCIE_DMA_WR_SAR_PTR_LO 0x20c #define PCIE_DMA_WR_SAR_PTR_HI 0x210 #define PCIE_DMA_WR_DAR_PTR_LO 0x214 #define PCIE_DMA_WR_DAR_PTR_HI 0x218 #define PCIE_DMA_WR_WEILO 0x18 #define PCIE_DMA_WR_WEIHI 0x1c #define PCIE_DMA_WR_DOORBELL 0x10 #define PCIE_DMA_WR_INT_STATUS 0x4c #define PCIE_DMA_WR_INT_MASK 0x54 #define PCIE_DMA_WR_INT_CLEAR 0x58 #define PCIE_DMA_RD_ENB 0x2c #define PCIE_DMA_RD_CTRL_LO 0x300 #define PCIE_DMA_RD_CTRL_HI 0x304 #define PCIE_DMA_RD_XFERSIZE 0x308 #define PCIE_DMA_RD_SAR_PTR_LO 0x30c #define PCIE_DMA_RD_SAR_PTR_HI 0x310 #define PCIE_DMA_RD_DAR_PTR_LO 0x314 #define PCIE_DMA_RD_DAR_PTR_HI 0x318 #define PCIE_DMA_RD_WEILO 0x38 #define PCIE_DMA_RD_WEIHI 0x3c #define PCIE_DMA_RD_DOORBELL 0x30 #define PCIE_DMA_RD_INT_STATUS 0xa0 #define PCIE_DMA_RD_INT_MASK 0xa8 #define PCIE_DMA_RD_INT_CLEAR 0xac /* Parameters for the waiting for iATU enabled routine */ #define LINK_WAIT_IATU_MIN 9000 #define LINK_WAIT_IATU_MAX 10000 #define PCIE_DIRECT_SPEED_CHANGE (0x1 << 17) #define PCIE_TYPE0_STATUS_COMMAND_REG 0x4 #define PCIE_TYPE0_BAR0_REG 0x10 #define PCIE_CAP_LINK_CONTROL2_LINK_STATUS 0xa0 #define PCIE_CLIENT_INTR_STATUS_LEGACY 0x08 #define PCIE_CLIENT_INTR_STATUS_MISC 0x10 #define PCIE_CLIENT_INTR_MASK_LEGACY 0x1c #define UNMASK_ALL_LEGACY_INT 0xffff0000 #define MASK_LEGACY_INT(x) (0x00110011 << x) #define UNMASK_LEGACY_INT(x) (0x00110000 << x) #define PCIE_CLIENT_INTR_MASK 0x24 #define PCIE_CLIENT_GENERAL_DEBUG 0x104 #define PCIE_CLIENT_HOT_RESET_CTRL 0x180 #define PCIE_LTSSM_ENABLE_ENHANCE BIT(4) #define PCIE_CLIENT_LTSSM_STATUS 0x300 #define SMLH_LINKUP BIT(16) #define RDLH_LINKUP BIT(17) #define PCIE_CLIENT_DBG_FIFO_MODE_CON 0x310 #define PCIE_CLIENT_DBG_FIFO_PTN_HIT_D0 0x320 #define PCIE_CLIENT_DBG_FIFO_PTN_HIT_D1 0x324 #define PCIE_CLIENT_DBG_FIFO_TRN_HIT_D0 0x328 #define PCIE_CLIENT_DBG_FIFO_TRN_HIT_D1 0x32c #define PCIE_CLIENT_DBG_FIFO_STATUS 0x350 #define PCIE_CLIENT_DBG_TRANSITION_DATA 0xffff0000 #define PCIE_CLIENT_DBF_EN 0xffff0003 #define RK_PCIE_DBG 0 #define PCIE_PHY_LINKUP BIT(0) #define PCIE_DATA_LINKUP BIT(1) #define PCIE_SB_BAR0_MASK_REG 0x100010 #define PCIE_PL_ORDER_RULE_CTRL_OFF 0x8B4 struct rk_pcie { struct dw_pcie *pci; enum rk_pcie_device_mode mode; enum phy_mode phy_mode; int phy_sub_mode; unsigned char bar_to_atu[6]; phys_addr_t *outbound_addr; unsigned long *ib_window_map; unsigned long *ob_window_map; unsigned int num_ib_windows; unsigned int num_ob_windows; void __iomem *dbi_base; void __iomem *apb_base; struct phy *phy; struct clk_bulk_data *clks; unsigned int clk_cnt; struct reset_bulk_data *rsts; struct gpio_desc *rst_gpio; phys_addr_t mem_start; size_t mem_size; struct pcie_port pp; struct regmap *usb_pcie_grf; struct regmap *pmu_grf; struct dma_trx_obj *dma_obj; bool in_suspend; bool skip_scan_in_resume; bool is_rk1808; bool is_signal_test; bool bifurcation; struct regulator *vpcie3v3; struct irq_domain *irq_domain; raw_spinlock_t intx_lock; }; struct rk_pcie_of_data { enum rk_pcie_device_mode mode; }; #define to_rk_pcie(x) dev_get_drvdata((x)->dev) static int rk_pcie_read(void __iomem *addr, int size, u32 *val) { if ((uintptr_t)addr & (size - 1)) { *val = 0; return PCIBIOS_BAD_REGISTER_NUMBER; } if (size == 4) { *val = readl(addr); } else if (size == 2) { *val = readw(addr); } else if (size == 1) { *val = readb(addr); } else { *val = 0; return PCIBIOS_BAD_REGISTER_NUMBER; } return PCIBIOS_SUCCESSFUL; } static int rk_pcie_write(void __iomem *addr, int size, u32 val) { if ((uintptr_t)addr & (size - 1)) return PCIBIOS_BAD_REGISTER_NUMBER; if (size == 4) writel(val, addr); else if (size == 2) writew(val, addr); else if (size == 1) writeb(val, addr); else return PCIBIOS_BAD_REGISTER_NUMBER; return PCIBIOS_SUCCESSFUL; } static u32 __rk_pcie_read_apb(struct rk_pcie *rk_pcie, void __iomem *base, u32 reg, size_t size) { int ret; u32 val; ret = rk_pcie_read(base + reg, size, &val); if (ret) dev_err(rk_pcie->pci->dev, "Read APB address failed\n"); return val; } static void __rk_pcie_write_apb(struct rk_pcie *rk_pcie, void __iomem *base, u32 reg, size_t size, u32 val) { int ret; ret = rk_pcie_write(base + reg, size, val); if (ret) dev_err(rk_pcie->pci->dev, "Write APB address failed\n"); } static inline u32 rk_pcie_readl_apb(struct rk_pcie *rk_pcie, u32 reg) { return __rk_pcie_read_apb(rk_pcie, rk_pcie->apb_base, reg, 0x4); } static inline void rk_pcie_writel_apb(struct rk_pcie *rk_pcie, u32 reg, u32 val) { __rk_pcie_write_apb(rk_pcie, rk_pcie->apb_base, reg, 0x4, val); } static u8 rk_pcie_iatu_unroll_enabled(struct dw_pcie *pci) { u32 val; val = dw_pcie_readl_dbi(pci, PCIE_ATU_VIEWPORT); if (val == 0xffffffff) return 1; return 0; } static void rk_pcie_writel_atu(struct dw_pcie *pci, u32 reg, u32 val) { int ret; if (pci->ops->write_dbi) { pci->ops->write_dbi(pci, pci->atu_base, reg, 4, val); return; } ret = dw_pcie_write(pci->atu_base + reg, 4, val); if (ret) dev_err(pci->dev, "Write ATU address failed\n"); } static void rk_pcie_writel_ib_unroll(struct dw_pcie *pci, u32 index, u32 reg, u32 val) { u32 offset = PCIE_GET_ATU_INB_UNR_REG_OFFSET(index); rk_pcie_writel_atu(pci, offset + reg, val); } static u32 rk_pcie_readl_atu(struct dw_pcie *pci, u32 reg) { int ret; u32 val; if (pci->ops->read_dbi) return pci->ops->read_dbi(pci, pci->atu_base, reg, 4); ret = dw_pcie_read(pci->atu_base + reg, 4, &val); if (ret) dev_err(pci->dev, "Read ATU address failed\n"); return val; } static u32 rk_pcie_readl_ib_unroll(struct dw_pcie *pci, u32 index, u32 reg) { u32 offset = PCIE_GET_ATU_INB_UNR_REG_OFFSET(index); return rk_pcie_readl_atu(pci, offset + reg); } static int rk_pcie_prog_inbound_atu_unroll(struct dw_pcie *pci, u8 func_no, int index, int bar, u64 cpu_addr, enum dw_pcie_as_type as_type) { int type; u32 retries, val; rk_pcie_writel_ib_unroll(pci, index, PCIE_ATU_UNR_LOWER_TARGET, lower_32_bits(cpu_addr)); rk_pcie_writel_ib_unroll(pci, index, PCIE_ATU_UNR_UPPER_TARGET, upper_32_bits(cpu_addr)); switch (as_type) { case DW_PCIE_AS_MEM: type = PCIE_ATU_TYPE_MEM; break; case DW_PCIE_AS_IO: type = PCIE_ATU_TYPE_IO; break; default: return -EINVAL; } rk_pcie_writel_ib_unroll(pci, index, PCIE_ATU_UNR_REGION_CTRL1, type | PCIE_ATU_FUNC_NUM(func_no)); rk_pcie_writel_ib_unroll(pci, index, PCIE_ATU_UNR_REGION_CTRL2, PCIE_ATU_FUNC_NUM_MATCH_EN | PCIE_ATU_ENABLE | PCIE_ATU_BAR_MODE_ENABLE | (bar << 8)); /* * Make sure ATU enable takes effect before any subsequent config * and I/O accesses. */ for (retries = 0; retries < LINK_WAIT_MAX_IATU_RETRIES; retries++) { val = rk_pcie_readl_ib_unroll(pci, index, PCIE_ATU_UNR_REGION_CTRL2); if (val & PCIE_ATU_ENABLE) return 0; mdelay(LINK_WAIT_IATU); } dev_err(pci->dev, "Inbound iATU is not being enabled\n"); return -EBUSY; } static int rk_pcie_prog_inbound_atu(struct dw_pcie *pci, u8 func_no, int index, int bar, u64 cpu_addr, enum dw_pcie_as_type as_type) { int type; u32 retries, val; if (pci->iatu_unroll_enabled) return rk_pcie_prog_inbound_atu_unroll(pci, func_no, index, bar, cpu_addr, as_type); dw_pcie_writel_dbi(pci, PCIE_ATU_VIEWPORT, PCIE_ATU_REGION_INBOUND | index); dw_pcie_writel_dbi(pci, PCIE_ATU_LOWER_TARGET, lower_32_bits(cpu_addr)); dw_pcie_writel_dbi(pci, PCIE_ATU_UPPER_TARGET, upper_32_bits(cpu_addr)); switch (as_type) { case DW_PCIE_AS_MEM: type = PCIE_ATU_TYPE_MEM; break; case DW_PCIE_AS_IO: type = PCIE_ATU_TYPE_IO; break; default: return -EINVAL; } dw_pcie_writel_dbi(pci, PCIE_ATU_CR1, type | PCIE_ATU_FUNC_NUM(func_no)); dw_pcie_writel_dbi(pci, PCIE_ATU_CR2, PCIE_ATU_ENABLE | PCIE_ATU_FUNC_NUM_MATCH_EN | PCIE_ATU_BAR_MODE_ENABLE | (bar << 8)); /* * Make sure ATU enable takes effect before any subsequent config * and I/O accesses. */ for (retries = 0; retries < LINK_WAIT_MAX_IATU_RETRIES; retries++) { val = dw_pcie_readl_dbi(pci, PCIE_ATU_CR2); if (val & PCIE_ATU_ENABLE) return 0; mdelay(LINK_WAIT_IATU); } dev_err(pci->dev, "Inbound iATU is not being enabled\n"); return -EBUSY; } static int rk_pcie_ep_inbound_atu(struct rk_pcie *rk_pcie, enum pci_barno bar, dma_addr_t cpu_addr, enum dw_pcie_as_type as_type) { int ret; u32 free_win; u8 func_no = 0x0; if (rk_pcie->in_suspend) { free_win = rk_pcie->bar_to_atu[bar]; } else { free_win = find_first_zero_bit(rk_pcie->ib_window_map, rk_pcie->num_ib_windows); if (free_win >= rk_pcie->num_ib_windows) { dev_err(rk_pcie->pci->dev, "No free inbound window\n"); return -EINVAL; } } ret = rk_pcie_prog_inbound_atu(rk_pcie->pci, func_no, free_win, bar, cpu_addr, as_type); if (ret < 0) { dev_err(rk_pcie->pci->dev, "Failed to program IB window\n"); return ret; } if (rk_pcie->in_suspend) return 0; rk_pcie->bar_to_atu[bar] = free_win; set_bit(free_win, rk_pcie->ib_window_map); return 0; } static void rk_pcie_writel_ob_unroll(struct dw_pcie *pci, u32 index, u32 reg, u32 val) { u32 offset = PCIE_GET_ATU_OUTB_UNR_REG_OFFSET(index); rk_pcie_writel_atu(pci, offset + reg, val); } static u32 rk_pcie_readl_ob_unroll(struct dw_pcie *pci, u32 index, u32 reg) { u32 offset = PCIE_GET_ATU_OUTB_UNR_REG_OFFSET(index); return rk_pcie_readl_atu(pci, offset + reg); } static void rk_pcie_prog_outbound_atu_unroll(struct dw_pcie *pci, u8 func_no, int index, int type, u64 cpu_addr, u64 pci_addr, u32 size) { u32 retries, val; u64 limit_addr = cpu_addr + size - 1; rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_LOWER_BASE, lower_32_bits(cpu_addr)); rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_UPPER_BASE, upper_32_bits(cpu_addr)); rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_LOWER_LIMIT, lower_32_bits(limit_addr)); rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_UPPER_LIMIT, upper_32_bits(limit_addr)); rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_LOWER_TARGET, lower_32_bits(pci_addr)); rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_UPPER_TARGET, upper_32_bits(pci_addr)); rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_REGION_CTRL1, type | PCIE_ATU_FUNC_NUM(func_no)); rk_pcie_writel_ob_unroll(pci, index, PCIE_ATU_UNR_REGION_CTRL2, PCIE_ATU_ENABLE); /* * Make sure ATU enable takes effect before any subsequent config * and I/O accesses. */ for (retries = 0; retries < LINK_WAIT_MAX_IATU_RETRIES; retries++) { val = rk_pcie_readl_ob_unroll(pci, index, PCIE_ATU_UNR_REGION_CTRL2); if (val & PCIE_ATU_ENABLE) return; mdelay(LINK_WAIT_IATU); } dev_err(pci->dev, "Outbound iATU is not being enabled\n"); } static void rk_pcie_prog_outbound_atu(struct dw_pcie *pci, int index, int type, u64 cpu_addr, u64 pci_addr, u32 size) { u32 retries, val; if (pci->ops->cpu_addr_fixup) cpu_addr = pci->ops->cpu_addr_fixup(pci, cpu_addr); if (pci->iatu_unroll_enabled) { rk_pcie_prog_outbound_atu_unroll(pci, 0x0, index, type, cpu_addr, pci_addr, size); return; } dw_pcie_writel_dbi(pci, PCIE_ATU_VIEWPORT, PCIE_ATU_REGION_OUTBOUND | index); dw_pcie_writel_dbi(pci, PCIE_ATU_LOWER_BASE, lower_32_bits(cpu_addr)); dw_pcie_writel_dbi(pci, PCIE_ATU_UPPER_BASE, upper_32_bits(cpu_addr)); dw_pcie_writel_dbi(pci, PCIE_ATU_LIMIT, lower_32_bits(cpu_addr + size - 1)); dw_pcie_writel_dbi(pci, PCIE_ATU_LOWER_TARGET, lower_32_bits(pci_addr)); dw_pcie_writel_dbi(pci, PCIE_ATU_UPPER_TARGET, upper_32_bits(pci_addr)); dw_pcie_writel_dbi(pci, PCIE_ATU_CR1, type | PCIE_ATU_FUNC_NUM(0x0)); dw_pcie_writel_dbi(pci, PCIE_ATU_CR2, PCIE_ATU_ENABLE); /* * Make sure ATU enable takes effect before any subsequent config * and I/O accesses. */ for (retries = 0; retries < LINK_WAIT_MAX_IATU_RETRIES; retries++) { val = dw_pcie_readl_dbi(pci, PCIE_ATU_CR2); if (val & PCIE_ATU_ENABLE) return; mdelay(LINK_WAIT_IATU); } dev_err(pci->dev, "Outbound iATU is not being enabled\n"); } static int rk_pcie_ep_outbound_atu(struct rk_pcie *rk_pcie, phys_addr_t phys_addr, u64 pci_addr, size_t size) { u32 free_win; if (rk_pcie->in_suspend) { free_win = find_first_bit(rk_pcie->ob_window_map, rk_pcie->num_ob_windows); } else { free_win = find_first_zero_bit(rk_pcie->ob_window_map, rk_pcie->num_ob_windows); if (free_win >= rk_pcie->num_ob_windows) { dev_err(rk_pcie->pci->dev, "No free outbound window\n"); return -EINVAL; } } rk_pcie_prog_outbound_atu(rk_pcie->pci, free_win, PCIE_ATU_TYPE_MEM, phys_addr, pci_addr, size); if (rk_pcie->in_suspend) return 0; set_bit(free_win, rk_pcie->ob_window_map); rk_pcie->outbound_addr[free_win] = phys_addr; return 0; } static void __rk_pcie_ep_reset_bar(struct rk_pcie *rk_pcie, enum pci_barno bar, int flags) { u32 reg; reg = PCI_BASE_ADDRESS_0 + (4 * bar); dw_pcie_writel_dbi(rk_pcie->pci, reg, 0x0); if (flags & PCI_BASE_ADDRESS_MEM_TYPE_64) dw_pcie_writel_dbi(rk_pcie->pci, reg + 4, 0x0); } static void rk_pcie_ep_reset_bar(struct rk_pcie *rk_pcie, enum pci_barno bar) { __rk_pcie_ep_reset_bar(rk_pcie, bar, 0); } static int rk_pcie_ep_atu_init(struct rk_pcie *rk_pcie) { int ret; enum pci_barno bar; enum dw_pcie_as_type as_type; dma_addr_t cpu_addr; phys_addr_t phys_addr; u64 pci_addr; size_t size; for (bar = BAR_0; bar <= BAR_5; bar++) rk_pcie_ep_reset_bar(rk_pcie, bar); cpu_addr = rk_pcie->mem_start; as_type = DW_PCIE_AS_MEM; ret = rk_pcie_ep_inbound_atu(rk_pcie, BAR_0, cpu_addr, as_type); if (ret) return ret; phys_addr = 0x0; pci_addr = 0x0; size = SZ_2G; ret = rk_pcie_ep_outbound_atu(rk_pcie, phys_addr, pci_addr, size); if (ret) return ret; return 0; } static inline void rk_pcie_set_mode(struct rk_pcie *rk_pcie) { switch (rk_pcie->mode) { case RK_PCIE_EP_TYPE: rk_pcie_writel_apb(rk_pcie, 0x0, 0xf00000); break; case RK_PCIE_RC_TYPE: rk_pcie_writel_apb(rk_pcie, 0x0, 0xf00040); /* * Disable order rule for CPL can't pass halted P queue. * Need to check producer-consumer model. * Just for RK1808 platform. */ if (rk_pcie->is_rk1808) dw_pcie_writel_dbi(rk_pcie->pci, PCIE_PL_ORDER_RULE_CTRL_OFF, 0xff00); break; } } static inline void rk_pcie_link_status_clear(struct rk_pcie *rk_pcie) { rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_GENERAL_DEBUG, 0x0); } static inline void rk_pcie_disable_ltssm(struct rk_pcie *rk_pcie) { rk_pcie_writel_apb(rk_pcie, 0x0, 0xc0008); } static inline void rk_pcie_enable_ltssm(struct rk_pcie *rk_pcie) { rk_pcie_writel_apb(rk_pcie, 0x0, 0xC000C); } static int rk_pcie_link_up(struct dw_pcie *pci) { struct rk_pcie *rk_pcie = to_rk_pcie(pci); u32 val; if (rk_pcie->is_rk1808) { val = rk_pcie_readl_apb(rk_pcie, PCIE_CLIENT_GENERAL_DEBUG); if ((val & (PCIE_PHY_LINKUP | PCIE_DATA_LINKUP)) == 0x3 && ((val & GENMASK(15, 10)) >> 10) == 0x11) return 1; } else { val = rk_pcie_readl_apb(rk_pcie, PCIE_CLIENT_LTSSM_STATUS); if ((val & (RDLH_LINKUP | SMLH_LINKUP)) == 0x30000) return 1; } return 0; } static void rk_pcie_enable_debug(struct rk_pcie *rk_pcie) { #if RK_PCIE_DBG if (rk_pcie->is_rk1808 == true) return; rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_DBG_FIFO_PTN_HIT_D0, PCIE_CLIENT_DBG_TRANSITION_DATA); rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_DBG_FIFO_PTN_HIT_D1, PCIE_CLIENT_DBG_TRANSITION_DATA); rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_DBG_FIFO_TRN_HIT_D0, PCIE_CLIENT_DBG_TRANSITION_DATA); rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_DBG_FIFO_TRN_HIT_D1, PCIE_CLIENT_DBG_TRANSITION_DATA); rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_DBG_FIFO_MODE_CON, PCIE_CLIENT_DBF_EN); #endif } static void rk_pcie_debug_dump(struct rk_pcie *rk_pcie) { #if RK_PCIE_DBG u32 loop; struct dw_pcie *pci = rk_pcie->pci; dev_info(pci->dev, "ltssm = 0x%x\n", rk_pcie_readl_apb(rk_pcie, PCIE_CLIENT_LTSSM_STATUS)); for (loop = 0; loop < 64; loop++) dev_info(pci->dev, "fifo_status = 0x%x\n", rk_pcie_readl_apb(rk_pcie, PCIE_CLIENT_DBG_FIFO_STATUS)); #endif } static int rk_pcie_establish_link(struct dw_pcie *pci) { int retries, power; struct rk_pcie *rk_pcie = to_rk_pcie(pci); bool std_rc = rk_pcie->mode == RK_PCIE_RC_TYPE && !rk_pcie->dma_obj; /* * For standard RC, even if the link has been setup by firmware, * we still need to reset link as we need to remove all resource info * from devices, for instance BAR, as it wasn't assigned by kernel. */ if (dw_pcie_link_up(pci) && !std_rc) { dev_err(pci->dev, "link is already up\n"); return 0; } /* Rest the device */ gpiod_set_value_cansleep(rk_pcie->rst_gpio, 0); rk_pcie_disable_ltssm(rk_pcie); rk_pcie_link_status_clear(rk_pcie); rk_pcie_enable_debug(rk_pcie); /* Enable client reset or link down interrupt */ rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_INTR_MASK, 0x40000); /* Enable LTSSM */ rk_pcie_enable_ltssm(rk_pcie); /* * In resume routine, function devices' resume function must be late after * controllers'. Some devices, such as Wi-Fi, need special IO setting before * finishing training. So there must be timeout here. These kinds of devices * need rescan devices by its driver when used. So no need to waste time waiting * for training pass. */ if (rk_pcie->in_suspend && rk_pcie->skip_scan_in_resume) { rfkill_get_wifi_power_state(&power); if (!power) { gpiod_set_value_cansleep(rk_pcie->rst_gpio, 1); return 0; } } /* * PCIe requires the refclk to be stable for 100µs prior to releasing * PERST and T_PVPERL (Power stable to PERST# inactive) should be a * minimum of 100ms. See table 2-4 in section 2.6.2 AC, the PCI Express * Card Electromechanical Specification 3.0. So 100ms in total is the min * requuirement here. We add a 200ms for sake of hoping everthings work fine. */ msleep(200); gpiod_set_value_cansleep(rk_pcie->rst_gpio, 1); /* * Add this 1ms delay because we observe link is always up stably after it and * could help us save 20ms for scanning devices. */ usleep_range(1000, 1100); for (retries = 0; retries < 100; retries++) { if (dw_pcie_link_up(pci)) { /* * We may be here in case of L0 in Gen1. But if EP is capable * of Gen2 or Gen3, Gen switch may happen just in this time, but * we keep on accessing devices in unstable link status. Given * that LTSSM max timeout is 24ms per period, we can wait a bit * more for Gen switch. */ msleep(50); dev_info(pci->dev, "PCIe Link up, LTSSM is 0x%x\n", rk_pcie_readl_apb(rk_pcie, PCIE_CLIENT_LTSSM_STATUS)); rk_pcie_debug_dump(rk_pcie); return 0; } dev_info_ratelimited(pci->dev, "PCIe Linking... LTSSM is 0x%x\n", rk_pcie_readl_apb(rk_pcie, PCIE_CLIENT_LTSSM_STATUS)); rk_pcie_debug_dump(rk_pcie); msleep(20); } dev_err(pci->dev, "PCIe Link Fail\n"); return rk_pcie->is_signal_test == true ? 0 : -EINVAL; } static int rk_pcie_host_init_dma_trx(struct rk_pcie *rk_pcie) { rk_pcie->dma_obj = rk_pcie_dma_obj_probe(rk_pcie->pci->dev); if (IS_ERR(rk_pcie->dma_obj)) { dev_err(rk_pcie->pci->dev, "failed to prepare dma object\n"); return -EINVAL; } /* Enable client write and read interrupt */ rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_INTR_MASK, 0xc000000); /* Enable core write interrupt */ dw_pcie_writel_dbi(rk_pcie->pci, PCIE_DMA_OFFSET + PCIE_DMA_WR_INT_MASK, 0x0); /* Enable core read interrupt */ dw_pcie_writel_dbi(rk_pcie->pci, PCIE_DMA_OFFSET + PCIE_DMA_RD_INT_MASK, 0x0); return 0; } static int rk_pci_find_resbar_capability(struct rk_pcie *rk_pcie) { u32 header; int ttl; int start = 0; int pos = PCI_CFG_SPACE_SIZE; int cap = PCI_EXT_CAP_ID_REBAR; /* minimum 8 bytes per capability */ ttl = (PCI_CFG_SPACE_EXP_SIZE - PCI_CFG_SPACE_SIZE) / 8; header = dw_pcie_readl_dbi(rk_pcie->pci, pos); /* * If we have no capabilities, this is indicated by cap ID, * cap version and next pointer all being 0. */ if (header == 0) return 0; while (ttl-- > 0) { if (PCI_EXT_CAP_ID(header) == cap && pos != start) return pos; pos = PCI_EXT_CAP_NEXT(header); if (pos < PCI_CFG_SPACE_SIZE) break; header = dw_pcie_readl_dbi(rk_pcie->pci, pos); if (!header) break; } return 0; } static void rk_pcie_ep_setup(struct rk_pcie *rk_pcie) { int ret; u32 val; u32 lanes; struct device *dev = rk_pcie->pci->dev; struct device_node *np = dev->of_node; int resbar_base; int bar; /* Enable client write and read interrupt */ rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_INTR_MASK, 0xc000000); /* Enable core write interrupt */ dw_pcie_writel_dbi(rk_pcie->pci, PCIE_DMA_OFFSET + PCIE_DMA_WR_INT_MASK, 0x0); /* Enable core read interrupt */ dw_pcie_writel_dbi(rk_pcie->pci, PCIE_DMA_OFFSET + PCIE_DMA_RD_INT_MASK, 0x0); ret = of_property_read_u32(np, "num-lanes", &lanes); if (ret) lanes = 0; /* Set the number of lanes */ val = dw_pcie_readl_dbi(rk_pcie->pci, PCIE_PORT_LINK_CONTROL); val &= ~PORT_LINK_MODE_MASK; switch (lanes) { case 1: val |= PORT_LINK_MODE_1_LANES; break; case 2: val |= PORT_LINK_MODE_2_LANES; break; case 4: val |= PORT_LINK_MODE_4_LANES; break; case 8: val |= PORT_LINK_MODE_8_LANES; break; default: dev_err(dev, "num-lanes %u: invalid value\n", lanes); return; } dw_pcie_writel_dbi(rk_pcie->pci, PCIE_PORT_LINK_CONTROL, val); /* Set link width speed control register */ val = dw_pcie_readl_dbi(rk_pcie->pci, PCIE_LINK_WIDTH_SPEED_CONTROL); val &= ~PORT_LOGIC_LINK_WIDTH_MASK; switch (lanes) { case 1: val |= PORT_LOGIC_LINK_WIDTH_1_LANES; break; case 2: val |= PORT_LOGIC_LINK_WIDTH_2_LANES; break; case 4: val |= PORT_LOGIC_LINK_WIDTH_4_LANES; break; case 8: val |= PORT_LOGIC_LINK_WIDTH_8_LANES; break; } val |= PCIE_DIRECT_SPEED_CHANGE; dw_pcie_writel_dbi(rk_pcie->pci, PCIE_LINK_WIDTH_SPEED_CONTROL, val); /* Enable bus master and memory space */ dw_pcie_writel_dbi(rk_pcie->pci, PCIE_TYPE0_STATUS_COMMAND_REG, 0x6); resbar_base = rk_pci_find_resbar_capability(rk_pcie); if (!resbar_base) { dev_warn(dev, "failed to find resbar_base\n"); } else { /* Resize BAR0 to support 512GB */ dw_pcie_writel_dbi(rk_pcie->pci, resbar_base + 0x4, 0xfffff0); /* Bit13-8 set to 19 means 2^19MB (512GB) */ dw_pcie_writel_dbi(rk_pcie->pci, resbar_base + 0x8, 0x13c0); /* Resize bar1 - bar6 to 64M */ for (bar = 1; bar < 6; bar++) { dw_pcie_writel_dbi(rk_pcie->pci, resbar_base + 0x4 + bar * 0x8, 0xfffff0); dw_pcie_writel_dbi(rk_pcie->pci, resbar_base + 0x8 + bar * 0x8, 0x6c0); } } /* Device id and class id needed for request bar address */ dw_pcie_writew_dbi(rk_pcie->pci, PCI_DEVICE_ID, 0x356a); dw_pcie_writew_dbi(rk_pcie->pci, PCI_CLASS_DEVICE, 0x0580); /* Set shadow BAR0 */ if (rk_pcie->is_rk1808) { val = rk_pcie->mem_size - 1; dw_pcie_writel_dbi(rk_pcie->pci, PCIE_SB_BAR0_MASK_REG, val); } } static int rk_pcie_ep_win_parse(struct rk_pcie *rk_pcie) { int ret; void *addr; struct device *dev = rk_pcie->pci->dev; struct device_node *np = dev->of_node; ret = of_property_read_u32(np, "num-ib-windows", &rk_pcie->num_ib_windows); if (ret < 0) { dev_err(dev, "unable to read *num-ib-windows* property\n"); return ret; } if (rk_pcie->num_ib_windows > MAX_IATU_IN) { dev_err(dev, "Invalid *num-ib-windows*\n"); return -EINVAL; } ret = of_property_read_u32(np, "num-ob-windows", &rk_pcie->num_ob_windows); if (ret < 0) { dev_err(dev, "Unable to read *num-ob-windows* property\n"); return ret; } if (rk_pcie->num_ob_windows > MAX_IATU_OUT) { dev_err(dev, "Invalid *num-ob-windows*\n"); return -EINVAL; } rk_pcie->ib_window_map = devm_kcalloc(dev, BITS_TO_LONGS(rk_pcie->num_ib_windows), sizeof(long), GFP_KERNEL); if (!rk_pcie->ib_window_map) return -ENOMEM; rk_pcie->ob_window_map = devm_kcalloc(dev, BITS_TO_LONGS(rk_pcie->num_ob_windows), sizeof(long), GFP_KERNEL); if (!rk_pcie->ob_window_map) return -ENOMEM; addr = devm_kcalloc(dev, rk_pcie->num_ob_windows, sizeof(phys_addr_t), GFP_KERNEL); if (!addr) return -ENOMEM; rk_pcie->outbound_addr = addr; return 0; } static int rk_pcie_msi_host_init(struct pcie_port *pp) { return 0; } static int rk_pcie_host_init(struct pcie_port *pp) { int ret; struct dw_pcie *pci = to_dw_pcie_from_pp(pp); dw_pcie_setup_rc(pp); ret = rk_pcie_establish_link(pci); return ret; } static const struct dw_pcie_host_ops rk_pcie_host_ops = { .host_init = rk_pcie_host_init, .msi_host_init = rk_pcie_msi_host_init, }; static int rk_add_pcie_port(struct rk_pcie *rk_pcie) { int ret; struct dw_pcie *pci = rk_pcie->pci; struct pcie_port *pp = &pci->pp; struct device *dev = pci->dev; pp->ops = &rk_pcie_host_ops; ret = dw_pcie_host_init(pp); if (ret) { dev_err(dev, "failed to initialize host\n"); return ret; } ret = rk_pcie_host_init_dma_trx(rk_pcie); if (ret) { dev_err(dev, "failed to init host dma trx\n"); return ret; } return 0; } static int rk_pcie_add_ep(struct rk_pcie *rk_pcie) { int ret; struct device *dev = rk_pcie->pci->dev; struct device_node *np = dev->of_node; struct device_node *mem; struct resource reg; mem = of_parse_phandle(np, "memory-region", 0); if (!mem) { dev_err(dev, "missing \"memory-region\" property\n"); return -ENODEV; } ret = of_address_to_resource(mem, 0, ®); if (ret < 0) { dev_err(dev, "missing \"reg\" property\n"); return ret; } rk_pcie->mem_start = reg.start; rk_pcie->mem_size = resource_size(®); ret = rk_pcie_ep_win_parse(rk_pcie); if (ret) { dev_err(dev, "failed to parse ep dts\n"); return ret; } rk_pcie->pci->atu_base = rk_pcie->pci->dbi_base + DEFAULT_DBI_ATU_OFFSET; rk_pcie->pci->iatu_unroll_enabled = rk_pcie_iatu_unroll_enabled(rk_pcie->pci); ret = rk_pcie_ep_atu_init(rk_pcie); if (ret) { dev_err(dev, "failed to init ep device\n"); return ret; } rk_pcie_ep_setup(rk_pcie); ret = rk_pcie_establish_link(rk_pcie->pci); if (ret) { dev_err(dev, "failed to establish pcie link\n"); return ret; } rk_pcie->dma_obj = rk_pcie_dma_obj_probe(dev); if (IS_ERR(rk_pcie->dma_obj)) { dev_err(dev, "failed to prepare dma object\n"); return -EINVAL; } return 0; } static void rk_pcie_clk_deinit(struct rk_pcie *rk_pcie) { clk_bulk_disable(rk_pcie->clk_cnt, rk_pcie->clks); clk_bulk_unprepare(rk_pcie->clk_cnt, rk_pcie->clks); } static int rk_pcie_clk_init(struct rk_pcie *rk_pcie) { struct device *dev = rk_pcie->pci->dev; struct property *prop; const char *name; int i = 0, ret, count; count = of_property_count_strings(dev->of_node, "clock-names"); if (count < 1) return -ENODEV; rk_pcie->clks = devm_kcalloc(dev, count, sizeof(struct clk_bulk_data), GFP_KERNEL); if (!rk_pcie->clks) return -ENOMEM; rk_pcie->clk_cnt = count; of_property_for_each_string(dev->of_node, "clock-names", prop, name) { rk_pcie->clks[i].id = name; if (!rk_pcie->clks[i].id) return -ENOMEM; i++; } ret = devm_clk_bulk_get(dev, count, rk_pcie->clks); if (ret) return ret; ret = clk_bulk_prepare(count, rk_pcie->clks); if (ret) return ret; ret = clk_bulk_enable(count, rk_pcie->clks); if (ret) { clk_bulk_unprepare(count, rk_pcie->clks); return ret; } return 0; } static int rk_pcie_resource_get(struct platform_device *pdev, struct rk_pcie *rk_pcie) { struct resource *dbi_base; struct resource *apb_base; dbi_base = platform_get_resource_byname(pdev, IORESOURCE_MEM, "pcie-dbi"); if (!dbi_base) { dev_err(&pdev->dev, "get pcie-dbi failed\n"); return -ENODEV; } rk_pcie->dbi_base = devm_ioremap_resource(&pdev->dev, dbi_base); if (IS_ERR(rk_pcie->dbi_base)) return PTR_ERR(rk_pcie->dbi_base); rk_pcie->pci->dbi_base = rk_pcie->dbi_base; apb_base = platform_get_resource_byname(pdev, IORESOURCE_MEM, "pcie-apb"); if (!apb_base) { dev_err(&pdev->dev, "get pcie-apb failed\n"); return -ENODEV; } rk_pcie->apb_base = devm_ioremap_resource(&pdev->dev, apb_base); if (IS_ERR(rk_pcie->apb_base)) return PTR_ERR(rk_pcie->apb_base); /* * Rest the device before enabling power because some of the * platforms may use external refclk input with the some power * rail connect to 100MHz OSC chip. So once the power is up for * the slot and the refclk is available, which isn't quite follow * the spec. We should make sure it is in reset state before * everthing's ready. */ rk_pcie->rst_gpio = devm_gpiod_get_optional(&pdev->dev, "reset", GPIOD_OUT_LOW); if (IS_ERR(rk_pcie->rst_gpio)) { dev_err(&pdev->dev, "invalid reset-gpios property in node\n"); return PTR_ERR(rk_pcie->rst_gpio); } return 0; } static int rk_pcie_phy_init(struct rk_pcie *rk_pcie) { int ret; struct device *dev = rk_pcie->pci->dev; rk_pcie->phy = devm_phy_get(dev, "pcie-phy"); if (IS_ERR(rk_pcie->phy)) { if (PTR_ERR(rk_pcie->phy) != -EPROBE_DEFER) dev_info(dev, "missing phy\n"); return PTR_ERR(rk_pcie->phy); } switch (rk_pcie->mode) { case RK_PCIE_RC_TYPE: rk_pcie->phy_mode = PHY_MODE_PCIE; /* make no sense */ rk_pcie->phy_sub_mode = PHY_MODE_PCIE_RC; break; case RK_PCIE_EP_TYPE: rk_pcie->phy_mode = PHY_MODE_PCIE; rk_pcie->phy_sub_mode = PHY_MODE_PCIE_EP; break; } ret = phy_set_mode_ext(rk_pcie->phy, rk_pcie->phy_mode, rk_pcie->phy_sub_mode); if (ret) { dev_err(dev, "fail to set phy to mode %s, err %d\n", (rk_pcie->phy_sub_mode == PHY_MODE_PCIE_RC) ? "RC" : "EP", ret); return ret; } if (rk_pcie->bifurcation) ret = phy_set_mode_ext(rk_pcie->phy, rk_pcie->phy_mode, PHY_MODE_PCIE_BIFURCATION); ret = phy_init(rk_pcie->phy); if (ret < 0) { dev_err(dev, "fail to init phy, err %d\n", ret); return ret; } phy_power_on(rk_pcie->phy); return 0; } static int rk_pcie_reset_control_release(struct rk_pcie *rk_pcie) { struct device *dev = rk_pcie->pci->dev; struct property *prop; const char *name; int ret, count, i = 0; count = of_property_count_strings(dev->of_node, "reset-names"); if (count < 1) return -ENODEV; rk_pcie->rsts = devm_kcalloc(dev, count, sizeof(struct reset_bulk_data), GFP_KERNEL); if (!rk_pcie->rsts) return -ENOMEM; of_property_for_each_string(dev->of_node, "reset-names", prop, name) { rk_pcie->rsts[i].id = name; if (!rk_pcie->rsts[i].id) return -ENOMEM; i++; } for (i = 0; i < count; i++) { rk_pcie->rsts[i].rst = devm_reset_control_get_exclusive(dev, rk_pcie->rsts[i].id); if (IS_ERR_OR_NULL(rk_pcie->rsts[i].rst)) { dev_err(dev, "failed to get %s\n", rk_pcie->clks[i].id); return -PTR_ERR(rk_pcie->rsts[i].rst); } } for (i = 0; i < count; i++) { ret = reset_control_deassert(rk_pcie->rsts[i].rst); if (ret) { dev_err(dev, "failed to release %s\n", rk_pcie->rsts[i].id); return ret; } } return 0; } static int rk_pcie_reset_grant_ctrl(struct rk_pcie *rk_pcie, bool enable) { int ret; u32 val = (0x1 << 18); /* Write mask bit */ if (enable) val |= (0x1 << 2); ret = regmap_write(rk_pcie->usb_pcie_grf, 0x0, val); return ret; } static void rk_pcie_start_dma_rd(struct dma_trx_obj *obj, int ctr_off) { struct rk_pcie *rk_pcie = dev_get_drvdata(obj->dev); struct dma_table *cur = obj->cur; dw_pcie_writel_dbi(rk_pcie->pci, PCIE_DMA_OFFSET + PCIE_DMA_RD_ENB, cur->enb.asdword); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_RD_CTRL_LO, cur->ctx_reg.ctrllo.asdword); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_RD_CTRL_HI, cur->ctx_reg.ctrlhi.asdword); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_RD_XFERSIZE, cur->ctx_reg.xfersize); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_RD_SAR_PTR_LO, cur->ctx_reg.sarptrlo); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_RD_SAR_PTR_HI, cur->ctx_reg.sarptrhi); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_RD_DAR_PTR_LO, cur->ctx_reg.darptrlo); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_RD_DAR_PTR_HI, cur->ctx_reg.darptrhi); dw_pcie_writel_dbi(rk_pcie->pci, PCIE_DMA_OFFSET + PCIE_DMA_RD_DOORBELL, cur->start.asdword); } static void rk_pcie_start_dma_wr(struct dma_trx_obj *obj, int ctr_off) { struct rk_pcie *rk_pcie = dev_get_drvdata(obj->dev); struct dma_table *cur = obj->cur; dw_pcie_writel_dbi(rk_pcie->pci, PCIE_DMA_OFFSET + PCIE_DMA_WR_ENB, cur->enb.asdword); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_WR_CTRL_LO, cur->ctx_reg.ctrllo.asdword); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_WR_CTRL_HI, cur->ctx_reg.ctrlhi.asdword); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_WR_XFERSIZE, cur->ctx_reg.xfersize); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_WR_SAR_PTR_LO, cur->ctx_reg.sarptrlo); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_WR_SAR_PTR_HI, cur->ctx_reg.sarptrhi); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_WR_DAR_PTR_LO, cur->ctx_reg.darptrlo); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_WR_DAR_PTR_HI, cur->ctx_reg.darptrhi); dw_pcie_writel_dbi(rk_pcie->pci, ctr_off + PCIE_DMA_WR_WEILO, cur->weilo.asdword); dw_pcie_writel_dbi(rk_pcie->pci, PCIE_DMA_OFFSET + PCIE_DMA_WR_DOORBELL, cur->start.asdword); } static void rk_pcie_start_dma_dwc(struct dma_trx_obj *obj) { int dir = obj->cur->dir; int chn = obj->cur->chn; int ctr_off = PCIE_DMA_OFFSET + chn * 0x200; if (dir == DMA_FROM_BUS) rk_pcie_start_dma_rd(obj, ctr_off); else if (dir == DMA_TO_BUS) rk_pcie_start_dma_wr(obj, ctr_off); } static void rk_pcie_config_dma_dwc(struct dma_table *table) { table->enb.enb = 0x1; table->ctx_reg.ctrllo.lie = 0x1; table->ctx_reg.ctrllo.rie = 0x0; table->ctx_reg.ctrllo.td = 0x1; table->ctx_reg.ctrlhi.asdword = 0x0; table->ctx_reg.xfersize = table->buf_size; if (table->dir == DMA_FROM_BUS) { table->ctx_reg.sarptrlo = (u32)(table->bus & 0xffffffff); table->ctx_reg.sarptrhi = (u32)(table->bus >> 32); table->ctx_reg.darptrlo = (u32)(table->local & 0xffffffff); table->ctx_reg.darptrhi = (u32)(table->local >> 32); } else if (table->dir == DMA_TO_BUS) { table->ctx_reg.sarptrlo = (u32)(table->local & 0xffffffff); table->ctx_reg.sarptrhi = (u32)(table->local >> 32); table->ctx_reg.darptrlo = (u32)(table->bus & 0xffffffff); table->ctx_reg.darptrhi = (u32)(table->bus >> 32); } table->weilo.weight0 = 0x0; table->start.stop = 0x0; table->start.chnl = table->chn; } static inline void rk_pcie_handle_dma_interrupt(struct rk_pcie *rk_pcie) { struct dma_trx_obj *obj = rk_pcie->dma_obj; struct dma_table *cur; if (!obj) return; cur = obj->cur; if (!cur) { pr_err("no pcie dma table\n"); return; } obj->dma_free = true; obj->irq_num++; if (cur->dir == DMA_TO_BUS) { if (list_empty(&obj->tbl_list)) { if (obj->dma_free && obj->loop_count >= obj->loop_count_threshold) complete(&obj->done); } } } static irqreturn_t rk_pcie_sys_irq_handler(int irq, void *arg) { struct rk_pcie *rk_pcie = arg; u32 chn = 0; union int_status status; union int_clear clears; u32 reg, val; status.asdword = dw_pcie_readl_dbi(rk_pcie->pci, PCIE_DMA_OFFSET + PCIE_DMA_WR_INT_STATUS); if (rk_pcie->dma_obj && rk_pcie->dma_obj->cur) chn = rk_pcie->dma_obj->cur->chn; if (status.donesta & BIT(chn)) { clears.doneclr = 0x1 << chn; dw_pcie_writel_dbi(rk_pcie->pci, PCIE_DMA_OFFSET + PCIE_DMA_WR_INT_CLEAR, clears.asdword); rk_pcie_handle_dma_interrupt(rk_pcie); } if (status.abortsta & BIT(chn)) { dev_err(rk_pcie->pci->dev, "%s, abort\n", __func__); clears.abortclr = 0x1 << chn; dw_pcie_writel_dbi(rk_pcie->pci, PCIE_DMA_OFFSET + PCIE_DMA_WR_INT_CLEAR, clears.asdword); } status.asdword = dw_pcie_readl_dbi(rk_pcie->pci, PCIE_DMA_OFFSET + PCIE_DMA_RD_INT_STATUS); if (status.donesta & BIT(chn)) { clears.doneclr = 0x1 << chn; dw_pcie_writel_dbi(rk_pcie->pci, PCIE_DMA_OFFSET + PCIE_DMA_RD_INT_CLEAR, clears.asdword); rk_pcie_handle_dma_interrupt(rk_pcie); } if (status.abortsta & BIT(chn)) { dev_err(rk_pcie->pci->dev, "%s, abort\n", __func__); clears.abortclr = 0x1 << chn; dw_pcie_writel_dbi(rk_pcie->pci, PCIE_DMA_OFFSET + PCIE_DMA_RD_INT_CLEAR, clears.asdword); } reg = rk_pcie_readl_apb(rk_pcie, PCIE_CLIENT_INTR_STATUS_MISC); if (reg & BIT(2)) { /* Setup command register */ val = dw_pcie_readl_dbi(rk_pcie->pci, PCI_COMMAND); val &= 0xffff0000; val |= PCI_COMMAND_IO | PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER | PCI_COMMAND_SERR; dw_pcie_writel_dbi(rk_pcie->pci, PCI_COMMAND, val); } rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_INTR_STATUS_MISC, reg); return IRQ_HANDLED; } static int rk_pcie_request_sys_irq(struct rk_pcie *rk_pcie, struct platform_device *pdev) { int irq; int ret; irq = platform_get_irq_byname(pdev, "sys"); if (irq < 0) { dev_err(rk_pcie->pci->dev, "missing sys IRQ resource\n"); return -EINVAL; } ret = devm_request_irq(rk_pcie->pci->dev, irq, rk_pcie_sys_irq_handler, IRQF_SHARED, "pcie-sys", rk_pcie); if (ret) { dev_err(rk_pcie->pci->dev, "failed to request PCIe subsystem IRQ\n"); return ret; } return 0; } static const struct rk_pcie_of_data rk_pcie_rc_of_data = { .mode = RK_PCIE_RC_TYPE, }; static const struct rk_pcie_of_data rk_pcie_ep_of_data = { .mode = RK_PCIE_EP_TYPE, }; static const struct of_device_id rk_pcie_of_match[] = { { .compatible = "rockchip,rk1808-pcie", .data = &rk_pcie_rc_of_data, }, { .compatible = "rockchip,rk1808-pcie-ep", .data = &rk_pcie_ep_of_data, }, { .compatible = "rockchip,rk3568-pcie", .data = &rk_pcie_rc_of_data, }, { .compatible = "rockchip,rk3568-pcie-ep", .data = &rk_pcie_ep_of_data, }, { .compatible = "rockchip,rk3588-pcie", .data = &rk_pcie_rc_of_data, }, { .compatible = "rockchip,rk3588-pcie-ep", .data = &rk_pcie_ep_of_data, }, {}, }; MODULE_DEVICE_TABLE(of, rk_pcie_of_match); static const struct dw_pcie_ops dw_pcie_ops = { .start_link = rk_pcie_establish_link, .link_up = rk_pcie_link_up, }; static int rk1808_pcie_fixup(struct rk_pcie *rk_pcie, struct device_node *np) { int ret; struct device *dev = rk_pcie->pci->dev; rk_pcie->usb_pcie_grf = syscon_regmap_lookup_by_phandle(np, "rockchip,usbpciegrf"); if (IS_ERR(rk_pcie->usb_pcie_grf)) { dev_err(dev, "failed to find usb_pcie_grf regmap\n"); return PTR_ERR(rk_pcie->usb_pcie_grf); } rk_pcie->pmu_grf = syscon_regmap_lookup_by_phandle(np, "rockchip,pmugrf"); if (IS_ERR(rk_pcie->pmu_grf)) { dev_err(dev, "failed to find pmugrf regmap\n"); return PTR_ERR(rk_pcie->pmu_grf); } /* Workaround for pcie, switch to PCIe_PRSTNm0 */ ret = regmap_write(rk_pcie->pmu_grf, 0x100, 0x01000100); if (ret) return ret; ret = regmap_write(rk_pcie->pmu_grf, 0x0, 0x0c000000); if (ret) return ret; /* release link reset grant */ ret = rk_pcie_reset_grant_ctrl(rk_pcie, true); return ret; } static void rk_pcie_fast_link_setup(struct rk_pcie *rk_pcie) { u32 val; /* LTSSM EN ctrl mode */ val = rk_pcie_readl_apb(rk_pcie, PCIE_CLIENT_HOT_RESET_CTRL); val |= PCIE_LTSSM_ENABLE_ENHANCE | (PCIE_LTSSM_ENABLE_ENHANCE << 16); rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_HOT_RESET_CTRL, val); } static void rk_pcie_legacy_irq_mask(struct irq_data *d) { struct rk_pcie *rk_pcie = irq_data_get_irq_chip_data(d); unsigned long flags; raw_spin_lock_irqsave(&rk_pcie->intx_lock, flags); rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_INTR_MASK_LEGACY, MASK_LEGACY_INT(d->hwirq)); raw_spin_unlock_irqrestore(&rk_pcie->intx_lock, flags); } static void rk_pcie_legacy_irq_unmask(struct irq_data *d) { struct rk_pcie *rk_pcie = irq_data_get_irq_chip_data(d); unsigned long flags; raw_spin_lock_irqsave(&rk_pcie->intx_lock, flags); rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_INTR_MASK_LEGACY, UNMASK_LEGACY_INT(d->hwirq)); raw_spin_unlock_irqrestore(&rk_pcie->intx_lock, flags); } static struct irq_chip rk_pcie_legacy_irq_chip = { .name = "rk-pcie-legacy-int", .irq_enable = rk_pcie_legacy_irq_unmask, .irq_disable = rk_pcie_legacy_irq_mask, .irq_mask = rk_pcie_legacy_irq_mask, .irq_unmask = rk_pcie_legacy_irq_unmask, .flags = IRQCHIP_SKIP_SET_WAKE | IRQCHIP_MASK_ON_SUSPEND, }; static int rk_pcie_intx_map(struct irq_domain *domain, unsigned int irq, irq_hw_number_t hwirq) { irq_set_chip_and_handler(irq, &rk_pcie_legacy_irq_chip, handle_simple_irq); irq_set_chip_data(irq, domain->host_data); return 0; } static const struct irq_domain_ops intx_domain_ops = { .map = rk_pcie_intx_map, }; static void rk_pcie_legacy_int_handler(struct irq_desc *desc) { struct irq_chip *chip = irq_desc_get_chip(desc); struct rk_pcie *rockchip = irq_desc_get_handler_data(desc); struct device *dev = rockchip->pci->dev; u32 reg; u32 hwirq; u32 virq; chained_irq_enter(chip, desc); reg = rk_pcie_readl_apb(rockchip, PCIE_CLIENT_INTR_STATUS_LEGACY); reg = reg & 0xf; while (reg) { hwirq = ffs(reg) - 1; reg &= ~BIT(hwirq); virq = irq_find_mapping(rockchip->irq_domain, hwirq); if (virq) generic_handle_irq(virq); else dev_err(dev, "unexpected IRQ, INT%d\n", hwirq); } chained_irq_exit(chip, desc); } static int rk_pcie_init_irq_domain(struct rk_pcie *rockchip) { struct device *dev = rockchip->pci->dev; struct device_node *intc = of_get_next_child(dev->of_node, NULL); if (!intc) { dev_err(dev, "missing child interrupt-controller node\n"); return -EINVAL; } raw_spin_lock_init(&rockchip->intx_lock); rockchip->irq_domain = irq_domain_add_linear(intc, PCI_NUM_INTX, &intx_domain_ops, rockchip); if (!rockchip->irq_domain) { dev_err(dev, "failed to get a INTx IRQ domain\n"); return -EINVAL; } return 0; } static int rk_pcie_enable_power(struct rk_pcie *rk_pcie) { int ret = 0; struct device *dev = rk_pcie->pci->dev; if (IS_ERR(rk_pcie->vpcie3v3)) return ret; ret = regulator_enable(rk_pcie->vpcie3v3); if (ret) dev_err(dev, "fail to enable vpcie3v3 regulator\n"); return ret; } static int rk_pcie_disable_power(struct rk_pcie *rk_pcie) { int ret = 0; struct device *dev = rk_pcie->pci->dev; if (IS_ERR(rk_pcie->vpcie3v3)) return ret; ret = regulator_disable(rk_pcie->vpcie3v3); if (ret) dev_err(dev, "fail to disable vpcie3v3 regulator\n"); return ret; } static int rk_pcie_really_probe(void *p) { struct platform_device *pdev = p; struct device *dev = &pdev->dev; struct rk_pcie *rk_pcie; struct dw_pcie *pci; int ret; const struct of_device_id *match; const struct rk_pcie_of_data *data; enum rk_pcie_device_mode mode; struct device_node *np = pdev->dev.of_node; u32 val; int irq; match = of_match_device(rk_pcie_of_match, dev); if (!match) return -EINVAL; data = (struct rk_pcie_of_data *)match->data; mode = (enum rk_pcie_device_mode)data->mode; rk_pcie = devm_kzalloc(dev, sizeof(*rk_pcie), GFP_KERNEL); if (!rk_pcie) return -ENOMEM; pci = devm_kzalloc(dev, sizeof(*pci), GFP_KERNEL); if (!pci) return -ENOMEM; pci->dev = dev; pci->ops = &dw_pcie_ops; rk_pcie->mode = mode; rk_pcie->pci = pci; if (of_device_is_compatible(np, "rockchip,rk1808-pcie") || of_device_is_compatible(np, "rockchip,rk1808-pcie-ep")) rk_pcie->is_rk1808 = true; else rk_pcie->is_rk1808 = false; if (device_property_read_bool(dev, "rockchip,bifurcation")) rk_pcie->bifurcation = true; ret = rk_pcie_resource_get(pdev, rk_pcie); if (ret) { dev_err(dev, "resource init failed\n"); return ret; } /* DON'T MOVE ME: must be enable before phy init */ rk_pcie->vpcie3v3 = devm_regulator_get_optional(dev, "vpcie3v3"); if (IS_ERR(rk_pcie->vpcie3v3)) { if (PTR_ERR(rk_pcie->vpcie3v3) != -ENODEV) return PTR_ERR(rk_pcie->vpcie3v3); dev_info(dev, "no vpcie3v3 regulator found\n"); } ret = rk_pcie_enable_power(rk_pcie); if (ret) return ret; ret = rk_pcie_phy_init(rk_pcie); if (ret) { dev_err(dev, "phy init failed\n"); goto disable_vpcie3v3; } ret = rk_pcie_reset_control_release(rk_pcie); if (ret) { dev_err(dev, "reset control init failed\n"); goto disable_phy; } ret = rk_pcie_request_sys_irq(rk_pcie, pdev); if (ret) { dev_err(dev, "pcie irq init failed\n"); goto disable_phy; } platform_set_drvdata(pdev, rk_pcie); ret = rk_pcie_clk_init(rk_pcie); if (ret) { dev_err(dev, "clock init failed\n"); goto disable_phy; } dw_pcie_dbi_ro_wr_en(pci); if (rk_pcie->is_rk1808) { ret = rk1808_pcie_fixup(rk_pcie, np); if (ret) goto deinit_clk; } else { rk_pcie_fast_link_setup(rk_pcie); } /* Legacy interrupt is optional */ ret = rk_pcie_init_irq_domain(rk_pcie); if (!ret) { irq = platform_get_irq_byname(pdev, "legacy"); if (irq >= 0) { irq_set_chained_handler_and_data(irq, rk_pcie_legacy_int_handler, rk_pcie); /* Unmask all legacy interrupt from INTA~INTD */ rk_pcie_writel_apb(rk_pcie, PCIE_CLIENT_INTR_MASK_LEGACY, UNMASK_ALL_LEGACY_INT); } dev_info(dev, "missing legacy IRQ resource\n"); } /* Set PCIe mode */ rk_pcie_set_mode(rk_pcie); /* Force into loopback master mode */ if (device_property_read_bool(dev, "rockchip,lpbk-master")) { val = dw_pcie_readl_dbi(pci, PCIE_PORT_LINK_CONTROL); val |= PORT_LINK_LPBK_ENABLE; dw_pcie_writel_dbi(pci, PCIE_PORT_LINK_CONTROL, val); rk_pcie->is_signal_test = true; } /* Force into compliance mode */ if (device_property_read_bool(dev, "rockchip,compliance-mode")) { val = dw_pcie_readl_dbi(pci, PCIE_CAP_LINK_CONTROL2_LINK_STATUS); val |= BIT(4); dw_pcie_writel_dbi(pci, PCIE_CAP_LINK_CONTROL2_LINK_STATUS, val); rk_pcie->is_signal_test = true; } /* Skip waiting for training to pass in system PM routine */ if (device_property_read_bool(dev, "rockchip,skip-scan-in-resume")) rk_pcie->skip_scan_in_resume = true; switch (rk_pcie->mode) { case RK_PCIE_RC_TYPE: ret = rk_add_pcie_port(rk_pcie); break; case RK_PCIE_EP_TYPE: ret = rk_pcie_add_ep(rk_pcie); break; } if (rk_pcie->is_signal_test == true) return 0; if (ret) goto remove_irq_domain; if (rk_pcie->dma_obj) { rk_pcie->dma_obj->start_dma_func = rk_pcie_start_dma_dwc; rk_pcie->dma_obj->config_dma_func = rk_pcie_config_dma_dwc; } if (rk_pcie->is_rk1808) { /* hold link reset grant after link-up */ ret = rk_pcie_reset_grant_ctrl(rk_pcie, false); if (ret) goto remove_irq_domain; } dw_pcie_dbi_ro_wr_dis(pci); device_init_wakeup(dev, true); /* Enable async system PM for multiports SoC */ device_enable_async_suspend(dev); return 0; remove_irq_domain: if (rk_pcie->irq_domain) irq_domain_remove(rk_pcie->irq_domain); disable_phy: phy_power_off(rk_pcie->phy); phy_exit(rk_pcie->phy); deinit_clk: rk_pcie_clk_deinit(rk_pcie); disable_vpcie3v3: rk_pcie_disable_power(rk_pcie); device_release_driver(dev); return ret; } static int rk_pcie_probe(struct platform_device *pdev) { struct task_struct *tsk; tsk = kthread_run(rk_pcie_really_probe, pdev, "rk-pcie"); if (IS_ERR(tsk)) { dev_err(&pdev->dev, "start rk-pcie thread failed\n"); return PTR_ERR(tsk); } return 0; } static int __maybe_unused rockchip_dw_pcie_suspend(struct device *dev) { struct rk_pcie *rk_pcie = dev_get_drvdata(dev); int ret; rk_pcie_link_status_clear(rk_pcie); rk_pcie_disable_ltssm(rk_pcie); /* make sure assert phy success */ usleep_range(200, 300); phy_power_off(rk_pcie->phy); phy_exit(rk_pcie->phy); clk_bulk_disable(rk_pcie->clk_cnt, rk_pcie->clks); rk_pcie->in_suspend = true; gpiod_set_value_cansleep(rk_pcie->rst_gpio, 0); ret = rk_pcie_disable_power(rk_pcie); return ret; } static int __maybe_unused rockchip_dw_pcie_resume(struct device *dev) { struct rk_pcie *rk_pcie = dev_get_drvdata(dev); bool std_rc = rk_pcie->mode == RK_PCIE_RC_TYPE && !rk_pcie->dma_obj; int ret; ret = rk_pcie_enable_power(rk_pcie); if (ret) return ret; ret = clk_bulk_enable(rk_pcie->clk_cnt, rk_pcie->clks); if (ret) { clk_bulk_unprepare(rk_pcie->clk_cnt, rk_pcie->clks); return ret; } ret = phy_set_mode_ext(rk_pcie->phy, rk_pcie->phy_mode, rk_pcie->phy_sub_mode); if (ret) { dev_err(dev, "fail to set phy to mode %s, err %d\n", (rk_pcie->phy_sub_mode == PHY_MODE_PCIE_RC) ? "RC" : "EP", ret); return ret; } ret = phy_init(rk_pcie->phy); if (ret < 0) { dev_err(dev, "fail to init phy, err %d\n", ret); return ret; } phy_power_on(rk_pcie->phy); dw_pcie_dbi_ro_wr_en(rk_pcie->pci); if (rk_pcie->is_rk1808) { /* release link reset grant */ ret = rk_pcie_reset_grant_ctrl(rk_pcie, true); if (ret) return ret; } else { rk_pcie_fast_link_setup(rk_pcie); } /* Set PCIe mode */ rk_pcie_set_mode(rk_pcie); if (std_rc) dw_pcie_setup_rc(&rk_pcie->pci->pp); ret = rk_pcie_establish_link(rk_pcie->pci); if (ret) { dev_err(dev, "failed to establish pcie link\n"); goto err; } if (std_rc) goto std_rc_done; ret = rk_pcie_ep_atu_init(rk_pcie); if (ret) { dev_err(dev, "failed to init ep device\n"); goto err; } rk_pcie_ep_setup(rk_pcie); rk_pcie->in_suspend = false; std_rc_done: dw_pcie_dbi_ro_wr_dis(rk_pcie->pci); /* hold link reset grant after link-up */ if (rk_pcie->is_rk1808) { ret = rk_pcie_reset_grant_ctrl(rk_pcie, false); if (ret) goto err; } return 0; err: rk_pcie_disable_power(rk_pcie); return ret; } static const struct dev_pm_ops rockchip_dw_pcie_pm_ops = { SET_NOIRQ_SYSTEM_SLEEP_PM_OPS(rockchip_dw_pcie_suspend, rockchip_dw_pcie_resume) }; static struct platform_driver rk_plat_pcie_driver = { .driver = { .name = "rk-pcie", .of_match_table = rk_pcie_of_match, .suppress_bind_attrs = true, .pm = &rockchip_dw_pcie_pm_ops, }, .probe = rk_pcie_probe, }; module_platform_driver(rk_plat_pcie_driver); MODULE_AUTHOR("Simon Xue "); MODULE_DESCRIPTION("RockChip PCIe Controller driver"); MODULE_LICENSE("GPL v2");