1962 lines
58 KiB
C
1962 lines
58 KiB
C
/*
|
|
* PCIe host controller driver for Samsung EXYNOS SoCs
|
|
*
|
|
* Copyright (C) 2013 Samsung Electronics Co., Ltd.
|
|
* http://www.samsung.com
|
|
*
|
|
* Author: Jingoo Han <jg1.han@samsung.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation.
|
|
*/
|
|
|
|
#include <linux/clk.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/gpio.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/mfd/syscon.h>
|
|
#include <linux/module.h>
|
|
#include <linux/of_gpio.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/regmap.h>
|
|
#include <linux/resource.h>
|
|
#include <linux/signal.h>
|
|
#include <linux/types.h>
|
|
#include <linux/exynos-pci-noti.h>
|
|
#include <linux/pm_qos.h>
|
|
#include <linux/exynos-pci-ctrl.h>
|
|
|
|
#include <soc/samsung/exynos-pm.h>
|
|
#include <soc/samsung/exynos-powermode.h>
|
|
|
|
#include "pcie-designware.h"
|
|
#include "pci-exynos.h"
|
|
|
|
struct exynos_pcie g_pcie[MAX_RC_NUM];
|
|
#ifdef CONFIG_PM_DEVFREQ
|
|
static struct pm_qos_request exynos_pcie_int_qos[MAX_RC_NUM];
|
|
#endif
|
|
|
|
#ifdef CONFIG_CPU_IDLE
|
|
static int exynos_pci_lpa_event(struct notifier_block *nb, unsigned long event, void *data);
|
|
#endif
|
|
static void exynos_pcie_resumed_phydown(struct pcie_port *pp);
|
|
static void exynos_pcie_assert_phy_reset(struct pcie_port *pp);
|
|
static int exynos_pcie_rd_own_conf(struct pcie_port *pp, int where, int size, u32 *val);
|
|
void exynos_pcie_send_pme_turn_off(struct exynos_pcie *exynos_pcie);
|
|
void exynos_pcie_poweroff(int ch_num);
|
|
int exynos_pcie_poweron(int ch_num);
|
|
static int exynos_pcie_wr_own_conf(struct pcie_port *pp, int where, int size,
|
|
u32 val);
|
|
static int exynos_pcie_rd_own_conf(struct pcie_port *pp, int where, int size,
|
|
u32 *val);
|
|
|
|
static inline void exynos_elb_writel(struct exynos_pcie *pcie, u32 val, u32 reg)
|
|
{
|
|
writel(val, pcie->elbi_base + reg);
|
|
}
|
|
|
|
static inline u32 exynos_elb_readl(struct exynos_pcie *pcie, u32 reg)
|
|
{
|
|
return readl(pcie->elbi_base + reg);
|
|
}
|
|
|
|
static inline void exynos_phy_writel(struct exynos_pcie *pcie, u32 val, u32 reg)
|
|
{
|
|
writel(val, pcie->phy_base + reg);
|
|
}
|
|
|
|
static inline u32 exynos_phy_readl(struct exynos_pcie *pcie, u32 reg)
|
|
{
|
|
return readl(pcie->phy_base + reg);
|
|
}
|
|
|
|
static inline void exynos_blk_writel(struct exynos_pcie *pcie, u32 val, u32 reg)
|
|
{
|
|
writel(val, pcie->block_base + reg);
|
|
}
|
|
|
|
static inline u32 exynos_blk_readl(struct exynos_pcie *pcie, u32 reg)
|
|
{
|
|
return readl(pcie->block_base + reg);
|
|
}
|
|
|
|
static int exynos_pcie_set_l1ss(int enable, struct pcie_port *pp, int id)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
u32 val;
|
|
unsigned long flags;
|
|
void __iomem *ep_dbi_base = pp->va_cfg0_base;
|
|
|
|
dev_info(pp->dev, "%s: START (state = 0x%x, id = 0x%x, enable = %d)\n",
|
|
__func__, exynos_pcie->l1ss_ctrl_id_state, id, enable);
|
|
spin_lock_irqsave(&exynos_pcie->conf_lock, flags);
|
|
if (exynos_pcie->state != STATE_LINK_UP) {
|
|
if (enable)
|
|
exynos_pcie->l1ss_ctrl_id_state &= ~(id);
|
|
else
|
|
exynos_pcie->l1ss_ctrl_id_state |= id;
|
|
spin_unlock_irqrestore(&exynos_pcie->conf_lock, flags);
|
|
dev_info(pp->dev, "%s: Link is not up. This will be set later. (state = 0x%x, id = 0x%x)\n",
|
|
__func__, exynos_pcie->l1ss_ctrl_id_state, id);
|
|
return -1;
|
|
}
|
|
if (enable) {
|
|
exynos_pcie->l1ss_ctrl_id_state &= ~(id);
|
|
if (exynos_pcie->l1ss_ctrl_id_state == 0) {
|
|
val = readl(ep_dbi_base + 0xbc);
|
|
val &= ~0x3;
|
|
val |= 0x142;
|
|
writel(val, ep_dbi_base + 0xBC);
|
|
val = readl(ep_dbi_base + 0x248);
|
|
writel(val | 0xa0f, ep_dbi_base + 0x248);
|
|
dev_info(pp->dev, "%s: L1ss enabled. (state = 0x%x, id = 0x%x)\n",
|
|
__func__, exynos_pcie->l1ss_ctrl_id_state, id);
|
|
} else {
|
|
dev_info(pp->dev, "%s: Can't enable L1ss. (state = 0x%x, id = 0x%x)\n",
|
|
__func__, exynos_pcie->l1ss_ctrl_id_state, id);
|
|
}
|
|
} else if (enable == 0) {
|
|
if (exynos_pcie->l1ss_ctrl_id_state) {
|
|
exynos_pcie->l1ss_ctrl_id_state |= id;
|
|
dev_info(pp->dev, "%s: L1ss is already disabled. (state = 0x%x, id = 0x%x)\n",
|
|
__func__, exynos_pcie->l1ss_ctrl_id_state, id);
|
|
} else {
|
|
exynos_pcie->l1ss_ctrl_id_state |= id;
|
|
val = readl(ep_dbi_base + 0xbc);
|
|
writel(val & ~0x3, ep_dbi_base + 0xBC);
|
|
val = readl(ep_dbi_base + 0x248);
|
|
writel(val & ~0xf, ep_dbi_base + 0x248);
|
|
dev_info(pp->dev, "%s: L1ss disabled. (state = 0x%x, id = 0x%x)\n",
|
|
__func__, exynos_pcie->l1ss_ctrl_id_state, id);
|
|
}
|
|
}
|
|
spin_unlock_irqrestore(&exynos_pcie->conf_lock, flags);
|
|
dev_info(pp->dev, "%s: END (state = 0x%x, id = 0x%x, enable = %d)\n",
|
|
__func__, exynos_pcie->l1ss_ctrl_id_state, id, enable);
|
|
return 0;
|
|
}
|
|
int exynos_pcie_l1ss_ctrl(int enable, int id)
|
|
{
|
|
struct pcie_port *pp = &g_pcie[0].pp;
|
|
|
|
return exynos_pcie_set_l1ss(enable, pp, id);
|
|
}
|
|
void exynos_pcie_register_dump(int ch_num)
|
|
{
|
|
struct pcie_port *pp = &g_pcie[ch_num].pp;
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
u32 i, j, val;
|
|
|
|
/* Link Reg : 0x0 ~ 0x2C4 */
|
|
for (i = 0; i < 45; i++) {
|
|
printk("RC Driver Reg 0x%04x: ", i * 0x10);
|
|
for (j = 0; j < 4; j++) {
|
|
if (((i * 0x10) + (j * 4)) < 0x2C4)
|
|
printk("0x%08x ", readl(exynos_pcie->elbi_base + (i * 0x10) + (j * 4)));
|
|
}
|
|
printk("\n");
|
|
}
|
|
|
|
/* PHY Reg : 0x0 ~ 0x180 */
|
|
for (i = 0; i < 24; i++) {
|
|
printk("PHY reg 0x%04x(%02x): ", i * 0x10, i * 4);
|
|
for (j = 0; j < 4; j++) {
|
|
if (((i * 0x10) + (j * 4)) < 0x180)
|
|
printk("0x%08x ", readl(exynos_pcie->phy_base + (i * 0x10) + (j * 4)));
|
|
}
|
|
printk("\n");
|
|
}
|
|
|
|
/* RC Conf : 0x0 ~ 0x40 */
|
|
for (i = 0; i < 4; i++) {
|
|
printk("RC Conf 0x%04x: ", i * 0x10);
|
|
for (j = 0; j < 4; j++) {
|
|
if (((i * 0x10) + (j * 4)) < 0x40) {
|
|
exynos_pcie_rd_own_conf(pp, (i * 0x10) + (j * 4), 4, &val);
|
|
printk("0x%08x ", val);
|
|
}
|
|
}
|
|
printk("\n");
|
|
}
|
|
|
|
exynos_pcie_rd_own_conf(pp, 0x78, 4, &val);
|
|
printk("RC Conf 0x0078(Device Status Register): 0x%08x\n", val);
|
|
exynos_pcie_rd_own_conf(pp, 0x80, 4, &val);
|
|
printk("RC Conf 0x0080(Link Status Register): 0x%08x\n", val);
|
|
exynos_pcie_rd_own_conf(pp, 0x104, 4, &val);
|
|
printk("RC Conf 0x0104(AER Registers): 0x%08x\n", val);
|
|
exynos_pcie_rd_own_conf(pp, 0x110, 4, &val);
|
|
printk("RC Conf 0x0110(AER Registers): 0x%08x\n", val);
|
|
exynos_pcie_rd_own_conf(pp, 0x130, 4, &val);
|
|
printk("RC Conf 0x0130(AER Registers): 0x%08x\n", val);
|
|
}
|
|
EXPORT_SYMBOL(exynos_pcie_register_dump);
|
|
|
|
static ssize_t show_pcie(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
|
|
return snprintf(buf, PAGE_SIZE, "0: send pme turn off message\n");
|
|
}
|
|
|
|
static ssize_t store_pcie(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
int enable;
|
|
u32 val;
|
|
struct exynos_pcie *exynos_pcie = dev_get_drvdata(dev);
|
|
|
|
if (sscanf(buf, "%10d", &enable) != 1)
|
|
return -EINVAL;
|
|
|
|
if (enable == 0) {
|
|
exynos_elb_writel(exynos_pcie, 0x1, PCIE_APP_REQ_EXIT_L1);
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_APP_REQ_EXIT_L1_MODE);
|
|
val &= ~APP_REQ_EXIT_L1_MODE;
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_APP_REQ_EXIT_L1_MODE);
|
|
dev_info(dev, "%s: Before set perst, gpio val = %d\n",
|
|
__func__, gpio_get_value(exynos_pcie->perst_gpio));
|
|
gpio_set_value(exynos_pcie->perst_gpio, 0);
|
|
dev_info(dev, "%s: After set perst, gpio val = %d\n",
|
|
__func__, gpio_get_value(exynos_pcie->perst_gpio));
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_APP_REQ_EXIT_L1_MODE);
|
|
val |= APP_REQ_EXIT_L1_MODE;
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_APP_REQ_EXIT_L1_MODE);
|
|
exynos_elb_writel(exynos_pcie, 0x0, PCIE_APP_REQ_EXIT_L1);
|
|
#ifdef CONFIG_PCI_EXYNOS_TEST
|
|
} else if (enable == 1) {
|
|
gpio_set_value(exynos_pcie->bt_gpio, 1);
|
|
gpio_set_value(exynos_pcie->wlan_gpio, 1);
|
|
mdelay(100);
|
|
exynos_pcie_poweron(exynos_pcie->ch_num);
|
|
} else if (enable == 2) {
|
|
exynos_pcie_poweroff(exynos_pcie->ch_num);
|
|
gpio_set_value(exynos_pcie->bt_gpio, 0);
|
|
gpio_set_value(exynos_pcie->wlan_gpio, 0);
|
|
} else if (enable == 3) {
|
|
exynos_pcie_send_pme_turn_off(exynos_pcie);
|
|
#endif
|
|
} else if (enable == 4) {
|
|
BUG_ON(1);
|
|
}
|
|
|
|
return count;
|
|
}
|
|
|
|
static DEVICE_ATTR(pcie_sysfs, S_IWUSR | S_IWGRP | S_IRUSR | S_IRGRP,
|
|
show_pcie, store_pcie);
|
|
|
|
static inline int create_pcie_sys_file(struct device *dev)
|
|
{
|
|
return device_create_file(dev, &dev_attr_pcie_sysfs);
|
|
}
|
|
|
|
static inline void remove_pcie_sys_file(struct device *dev)
|
|
{
|
|
device_remove_file(dev, &dev_attr_pcie_sysfs);
|
|
}
|
|
|
|
static void __maybe_unused exynos_pcie_notify_callback(struct pcie_port *pp, int event)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
if (exynos_pcie->event_reg && exynos_pcie->event_reg->callback &&
|
|
(exynos_pcie->event_reg->events & event)) {
|
|
struct exynos_pcie_notify *notify = &exynos_pcie->event_reg->notify;
|
|
notify->event = event;
|
|
notify->user = exynos_pcie->event_reg->user;
|
|
dev_info(pp->dev, "Callback for the event : %d\n", event);
|
|
exynos_pcie->event_reg->callback(notify);
|
|
} else
|
|
dev_info(pp->dev,
|
|
"Client driver does not have registration of the event : %d\n", event);
|
|
}
|
|
|
|
int exynos_pcie_dump_link_down_status(int ch_num)
|
|
{
|
|
struct pcie_port *pp = &g_pcie[ch_num].pp;
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
|
|
if (exynos_pcie->state == STATE_LINK_UP) {
|
|
dev_info(pp->dev, "LTSSM: 0x%08x\n",
|
|
readl(exynos_pcie->elbi_base + PCIE_ELBI_RDLH_LINKUP));
|
|
dev_info(pp->dev, "LTSSM_H: 0x%08x\n",
|
|
readl(exynos_pcie->elbi_base + PCIE_CXPL_DEBUG_INFO_H));
|
|
|
|
dev_info(pp->dev, "DMA_MONITOR1: 0x%08x\n",
|
|
readl(exynos_pcie->elbi_base + PCIE_DMA_MONITOR1));
|
|
dev_info(pp->dev, "DMA_MONITOR2: 0x%08x\n",
|
|
readl(exynos_pcie->elbi_base + PCIE_DMA_MONITOR2));
|
|
dev_info(pp->dev, "DMA_MONITOR3: 0x%08x\n",
|
|
readl(exynos_pcie->elbi_base + PCIE_DMA_MONITOR3));
|
|
} else
|
|
dev_info(pp->dev, "PCIE link state is %d\n", exynos_pcie->state);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(exynos_pcie_dump_link_down_status);
|
|
|
|
static int exynos_pcie_dma_mon_event(struct notifier_block *nb, unsigned long event, void *data)
|
|
{
|
|
int ret;
|
|
struct exynos_pcie *exynos_pcie = container_of(nb, struct exynos_pcie, ss_dma_mon_nb);
|
|
|
|
ret = exynos_pcie_dump_link_down_status(exynos_pcie->ch_num);
|
|
if (exynos_pcie->state == STATE_LINK_UP)
|
|
exynos_pcie_register_dump(exynos_pcie->ch_num);
|
|
|
|
return notifier_from_errno(ret);
|
|
}
|
|
|
|
static int exynos_pcie_clock_get(struct pcie_port *pp)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
struct exynos_pcie_clks *clks = &exynos_pcie->clks;
|
|
int i, total_clk_num, phy_count;
|
|
|
|
/*
|
|
* CAUTION - PCIe and phy clock have to define in order.
|
|
* You must define related PCIe clock first in DT.
|
|
*/
|
|
total_clk_num = exynos_pcie->pcie_clk_num + exynos_pcie->phy_clk_num;
|
|
|
|
for (i = 0; i < total_clk_num; i++) {
|
|
if (i < exynos_pcie->pcie_clk_num) {
|
|
clks->pcie_clks[i] = of_clk_get(pp->dev->of_node, i);
|
|
if (IS_ERR(clks->pcie_clks[i])) {
|
|
dev_err(pp->dev, "Failed to get pcie clock\n");
|
|
return -ENODEV;
|
|
}
|
|
} else {
|
|
phy_count = i - exynos_pcie->pcie_clk_num;
|
|
clks->phy_clks[phy_count] = of_clk_get(pp->dev->of_node, i);
|
|
if (IS_ERR(clks->phy_clks[i])) {
|
|
dev_err(pp->dev, "Failed to get pcie clock\n");
|
|
return -ENODEV;
|
|
}
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int exynos_pcie_clock_enable(struct pcie_port *pp, int enable)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
struct exynos_pcie_clks *clks = &exynos_pcie->clks;
|
|
int i;
|
|
int ret = 0;
|
|
|
|
if (enable) {
|
|
for (i = 0; i < exynos_pcie->pcie_clk_num; i++)
|
|
ret = clk_prepare_enable(clks->pcie_clks[i]);
|
|
} else {
|
|
for (i = 0; i < exynos_pcie->pcie_clk_num; i++)
|
|
clk_disable_unprepare(clks->pcie_clks[i]);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int exynos_pcie_phy_clock_enable(struct pcie_port *pp, int enable)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
struct exynos_pcie_clks *clks = &exynos_pcie->clks;
|
|
int i;
|
|
int ret = 0;
|
|
|
|
if (enable) {
|
|
for (i = 0; i < exynos_pcie->phy_clk_num; i++)
|
|
ret = clk_prepare_enable(clks->phy_clks[i]);
|
|
} else {
|
|
for (i = 0; i < exynos_pcie->phy_clk_num; i++)
|
|
clk_disable_unprepare(clks->phy_clks[i]);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
void exynos_pcie_print_link_history(struct pcie_port *pp)
|
|
{
|
|
struct device *dev = pp->dev;
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
u32 history_buffer[32];
|
|
int i;
|
|
|
|
for (i = 31; i >= 0; i--)
|
|
history_buffer[i] = exynos_elb_readl(exynos_pcie,
|
|
PCIE_HISTORY_REG(i));
|
|
for (i = 31; i >= 0; i--)
|
|
dev_info(dev, "LTSSM: 0x%02x, L1sub: 0x%x, D state: 0x%x\n",
|
|
LTSSM_STATE(history_buffer[i]),
|
|
L1SUB_STATE(history_buffer[i]),
|
|
PM_DSTATE(history_buffer[i]));
|
|
}
|
|
|
|
static void exynos_pcie_setup_rc(struct pcie_port *pp)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
u32 exp_cap_off = EXP_CAP_ID_OFFSET;
|
|
u32 pm_cap_off = PM_CAP_ID_OFFSET;
|
|
u32 val;
|
|
|
|
/* enable writing to DBI read-only registers */
|
|
exynos_pcie_wr_own_conf(pp, PCIE_MISC_CONTROL, 4, DBI_RO_WR_EN);
|
|
|
|
/* change vendor ID and device ID for PCIe */
|
|
exynos_pcie_wr_own_conf(pp, PCI_VENDOR_ID, 2, PCI_VENDOR_ID_SAMSUNG);
|
|
exynos_pcie_wr_own_conf(pp, PCI_DEVICE_ID, 2,
|
|
PCI_DEVICE_ID_EXYNOS + exynos_pcie->ch_num);
|
|
|
|
/* set max link width & speed : Gen2, Lane1 */
|
|
exynos_pcie_rd_own_conf(pp, exp_cap_off + PCI_EXP_LNKCAP, 4, &val);
|
|
val &= ~(PCI_EXP_LNKCAP_L1EL|PCI_EXP_LNKCAP_MLW|PCI_EXP_LNKCAP_SLS);
|
|
val |= PCI_EXP_LNKCAP_L1EL_64USEC|PCI_EXP_LNKCAP_MLW_X1|PCI_EXP_LNKCAP_SLS_5_0GB;
|
|
exynos_pcie_wr_own_conf(pp, exp_cap_off + PCI_EXP_LNKCAP, 4, val);
|
|
|
|
/* set auxiliary clock frequency: 26MHz */
|
|
exynos_pcie_wr_own_conf(pp, PCIE_AUX_CLK_FREQ_OFF, 4, PCIE_AUX_CLK_FREQ_26MHZ);
|
|
|
|
/* set duration of L1.2 & L1.2.Entry */
|
|
exynos_pcie_wr_own_conf(pp, PCIE_L1_SUBSTATES_OFF, 4, 0xD2);
|
|
|
|
/* clear power management control and status register */
|
|
exynos_pcie_wr_own_conf(pp, pm_cap_off + PCI_PM_CTRL, 4, 0x0);
|
|
|
|
/* initiate link retraining */
|
|
exynos_pcie_rd_own_conf(pp, exp_cap_off + PCI_EXP_LNKCTL, 4, &val);
|
|
val |= PCI_EXP_LNKCTL_RL;
|
|
exynos_pcie_wr_own_conf(pp, exp_cap_off + PCI_EXP_LNKCTL, 4, val);
|
|
|
|
/* set target speed to GEN1 only */
|
|
exynos_pcie_rd_own_conf(pp, exp_cap_off + PCI_EXP_LNKCTL2, 4, &val);
|
|
val &= ~PCI_EXP_LNKCTL2_TLS;
|
|
val |= PCI_EXP_LNKCTL2_TLS_2_5GB;
|
|
exynos_pcie_wr_own_conf(pp, exp_cap_off + PCI_EXP_LNKCTL2, 4, val);
|
|
|
|
/* enable PHY 2Lane Bifurcation mode */
|
|
val = exynos_blk_readl(exynos_pcie, PCIE_WIFI0_PCIE_PHY_CONTROL);
|
|
val &= ~BIFURCATION_MODE_DISABLE;
|
|
val |= LINK0_ENABLE | PCS_LANE0_EANBLE;
|
|
exynos_blk_writel(exynos_pcie, val, PCIE_WIFI0_PCIE_PHY_CONTROL);
|
|
}
|
|
|
|
static void exynos_pcie_prog_viewport_cfg0(struct pcie_port *pp, u32 busdev)
|
|
{
|
|
/* Program viewport 0 : OUTBOUND : CFG0 */
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_VIEWPORT, 4,
|
|
PCIE_ATU_REGION_OUTBOUND | PCIE_ATU_REGION_INDEX0);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_LOWER_BASE, 4, pp->cfg0_base);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_UPPER_BASE, 4,
|
|
(pp->cfg0_base >> 32));
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_LIMIT, 4,
|
|
pp->cfg0_base + pp->cfg0_size - 1);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_LOWER_TARGET, 4, busdev);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_UPPER_TARGET, 4, 0);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_CR1, 4, PCIE_ATU_TYPE_CFG0);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_CR2, 4, PCIE_ATU_ENABLE);
|
|
}
|
|
|
|
static void exynos_pcie_prog_viewport_cfg1(struct pcie_port *pp, u32 busdev)
|
|
{
|
|
/* Program viewport 1 : OUTBOUND : CFG1 */
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_VIEWPORT, 4,
|
|
PCIE_ATU_REGION_OUTBOUND | PCIE_ATU_REGION_INDEX2);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_CR1, 4, PCIE_ATU_TYPE_CFG1);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_LOWER_BASE, 4, pp->cfg1_base);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_UPPER_BASE, 4,
|
|
(pp->cfg1_base >> 32));
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_LIMIT, 4,
|
|
pp->cfg1_base + pp->cfg1_size - 1);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_LOWER_TARGET, 4, busdev);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_UPPER_TARGET, 4, 0);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_CR2, 4, PCIE_ATU_ENABLE);
|
|
}
|
|
|
|
#ifdef CONFIG_SOC_EXYNOS8895
|
|
void dw_pcie_prog_viewport_mem_outbound_index2(struct pcie_port *pp)
|
|
{
|
|
/* Program viewport 0 : OUTBOUND : MEM */
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_VIEWPORT, 4,
|
|
PCIE_ATU_REGION_OUTBOUND | PCIE_ATU_REGION_INDEX2);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_CR1, 4, PCIE_ATU_TYPE_MEM);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_LOWER_BASE, 4, 0x11B00000);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_UPPER_BASE, 4, 0x0);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_LIMIT, 4, 0x11B0ffff);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_LOWER_TARGET, 4, 0x11C00000);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_UPPER_TARGET, 4, 0x0);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_CR2, 4, PCIE_ATU_ENABLE);
|
|
}
|
|
#endif
|
|
|
|
static void exynos_pcie_prog_viewport_mem_outbound(struct pcie_port *pp)
|
|
{
|
|
/* Program viewport 0 : OUTBOUND : MEM */
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_VIEWPORT, 4,
|
|
PCIE_ATU_REGION_OUTBOUND | PCIE_ATU_REGION_INDEX1);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_CR1, 4, PCIE_ATU_TYPE_MEM);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_LOWER_BASE, 4, pp->mem_base);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_UPPER_BASE, 4,
|
|
(pp->mem_base >> 32));
|
|
#ifdef CONFIG_SOC_EXYNOS8895
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_LIMIT, 4,
|
|
pp->mem_base + pp->mem_size - 1 - 0x100000);
|
|
#else
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_LIMIT, 4,
|
|
pp->mem_base + pp->mem_size - 1);
|
|
#endif
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_LOWER_TARGET, 4, pp->mem_bus_addr);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_UPPER_TARGET, 4,
|
|
upper_32_bits(pp->mem_bus_addr));
|
|
exynos_pcie_wr_own_conf(pp, PCIE_ATU_CR2, 4, PCIE_ATU_ENABLE);
|
|
|
|
#ifdef CONFIG_SOC_EXYNOS8895
|
|
dw_pcie_prog_viewport_mem_outbound_index2(pp);
|
|
#endif
|
|
}
|
|
|
|
static int exynos_pcie_establish_link(struct pcie_port *pp)
|
|
{
|
|
struct device *dev = pp->dev;
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
u32 val;
|
|
int count = 0, try_cnt = 0;
|
|
|
|
retry:
|
|
/* avoid checking rx elecidle when access DBI */
|
|
if (exynos_pcie->phy_ops.phy_check_rx_elecidle != NULL)
|
|
exynos_pcie->phy_ops.phy_check_rx_elecidle(
|
|
exynos_pcie->phy_pcs_base, 1);
|
|
|
|
exynos_elb_writel(exynos_pcie, 0x0, PCIE_SOFT_CORE_RESET);
|
|
udelay(20);
|
|
exynos_elb_writel(exynos_pcie, 0x1, PCIE_SOFT_CORE_RESET);
|
|
|
|
/* set #PERST high */
|
|
dev_info(dev, "%s: Before set perst, gpio val = %d\n",
|
|
__func__, gpio_get_value(exynos_pcie->perst_gpio));
|
|
gpio_set_value(exynos_pcie->perst_gpio, 1);
|
|
dev_info(dev, "%s: After set perst, gpio val = %d\n",
|
|
__func__, gpio_get_value(exynos_pcie->perst_gpio));
|
|
usleep_range(18000, 20000);
|
|
|
|
/* APP_REQ_EXIT_L1_MODE : BIT0 (0x0 : S/W mode, 0x1 : H/W mode) */
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_APP_REQ_EXIT_L1_MODE);
|
|
val |= APP_REQ_EXIT_L1_MODE;
|
|
val |= L1_REQ_NAK_CONTROL_MASTER;
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_APP_REQ_EXIT_L1_MODE);
|
|
exynos_elb_writel(exynos_pcie, PCIE_LINKDOWN_RST_MANUAL,
|
|
PCIE_LINKDOWN_RST_CTRL_SEL);
|
|
|
|
/* Q-Channel support and L1 NAK enable when AXI pending */
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_QCH_SEL);
|
|
if (exynos_pcie->ip_ver >= 0x889500) {
|
|
val &= ~(CLOCK_GATING_PMU_MASK | CLOCK_GATING_APB_MASK | CLOCK_GATING_AXI_MASK);
|
|
} else {
|
|
val &= ~CLOCK_GATING_MASK;
|
|
val |= CLOCK_NOT_GATING;
|
|
}
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_QCH_SEL);
|
|
|
|
exynos_pcie_assert_phy_reset(pp);
|
|
|
|
exynos_elb_writel(exynos_pcie, 0x1, PCIE_L1_BUG_FIX_ENABLE);
|
|
|
|
/* setup root complex */
|
|
dw_pcie_setup_rc(pp);
|
|
exynos_pcie_setup_rc(pp);
|
|
|
|
if (exynos_pcie->phy_ops.phy_check_rx_elecidle != NULL)
|
|
exynos_pcie->phy_ops.phy_check_rx_elecidle(
|
|
exynos_pcie->phy_pcs_base, 0);
|
|
|
|
dev_info(dev, "D state: %x, %x\n",
|
|
exynos_elb_readl(exynos_pcie, PCIE_PM_DSTATE) & 0x7,
|
|
exynos_elb_readl(exynos_pcie, PCIE_ELBI_RDLH_LINKUP));
|
|
|
|
/* assert LTSSM enable */
|
|
exynos_elb_writel(exynos_pcie, PCIE_ELBI_LTSSM_ENABLE, PCIE_APP_LTSSM_ENABLE);
|
|
count = 0;
|
|
while (count < MAX_TIMEOUT) {
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_ELBI_RDLH_LINKUP) & 0x1f;
|
|
if (val >= 0x0d && val <= 0x15)
|
|
break;
|
|
|
|
count++;
|
|
|
|
udelay(10);
|
|
}
|
|
|
|
if (count >= MAX_TIMEOUT) {
|
|
try_cnt++;
|
|
dev_err(dev, "%s: Link is not up, try count: %d, %x\n", __func__,
|
|
try_cnt, exynos_elb_readl(exynos_pcie, PCIE_ELBI_RDLH_LINKUP));
|
|
if (try_cnt < 10) {
|
|
dev_info(dev, "%s: Before set perst, gpio val = %d\n",
|
|
__func__, gpio_get_value(exynos_pcie->perst_gpio));
|
|
gpio_set_value(exynos_pcie->perst_gpio, 0);
|
|
dev_info(dev, "%s: After set perst, gpio val = %d\n",
|
|
__func__, gpio_get_value(exynos_pcie->perst_gpio));
|
|
/* LTSSM disable */
|
|
exynos_elb_writel(exynos_pcie, PCIE_ELBI_LTSSM_DISABLE,
|
|
PCIE_APP_LTSSM_ENABLE);
|
|
exynos_pcie_phy_clock_enable(pp, 0);
|
|
goto retry;
|
|
} else {
|
|
exynos_pcie_print_link_history(pp);
|
|
if ((exynos_pcie->ip_ver >= 0x889000) && (exynos_pcie->ch_num == 0)) {
|
|
return -EPIPE;
|
|
}
|
|
BUG_ON(1);
|
|
return -EPIPE;
|
|
}
|
|
} else {
|
|
dev_info(dev, "%s: Link up:%x\n", __func__,
|
|
exynos_elb_readl(exynos_pcie, PCIE_ELBI_RDLH_LINKUP));
|
|
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_IRQ_PULSE);
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_IRQ_PULSE);
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_IRQ_EN_PULSE);
|
|
val |= IRQ_LINKDOWN_ENABLE;
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_IRQ_EN_PULSE);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
void exynos_pcie_work(struct work_struct *work)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = container_of(work, struct exynos_pcie, work.work);
|
|
struct pcie_port *pp = &exynos_pcie->pp;
|
|
struct device *dev = pp->dev;
|
|
|
|
if (exynos_pcie->state == STATE_LINK_DOWN)
|
|
return;
|
|
|
|
exynos_pcie->linkdown_cnt++;
|
|
dev_info(dev, "link down and recovery cnt: %d\n", exynos_pcie->linkdown_cnt);
|
|
|
|
#ifdef CONFIG_PCI_EXYNOS_TEST
|
|
exynos_pcie_poweroff(exynos_pcie->ch_num);
|
|
exynos_pcie_poweron(exynos_pcie->ch_num);
|
|
#else
|
|
exynos_pcie_notify_callback(pp, EXYNOS_PCIE_EVENT_LINKDOWN);
|
|
#endif
|
|
}
|
|
|
|
void exynos_pcie_work_l1ss(struct work_struct *work)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = container_of(work, struct exynos_pcie, work_l1ss.work);
|
|
struct pcie_port *pp = &exynos_pcie->pp;
|
|
struct device *dev = pp->dev;
|
|
if (exynos_pcie->work_l1ss_cnt == 0) {
|
|
dev_info(dev, "[%s]boot_cnt: %d, work_l1ss_cnt: %d\n", __func__, exynos_pcie->boot_cnt, exynos_pcie->work_l1ss_cnt);
|
|
exynos_pcie_set_l1ss(1, pp, PCIE_L1SS_CTRL_BOOT);
|
|
exynos_pcie->work_l1ss_cnt++;
|
|
} else {
|
|
dev_info(dev, "[%s]boot_cnt: %d, work_l1ss_cnt: %d\n", __func__, exynos_pcie->boot_cnt, exynos_pcie->work_l1ss_cnt);
|
|
return;
|
|
}
|
|
}
|
|
static void exynos_pcie_assert_phy_reset(struct pcie_port *pp)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
u32 val;
|
|
|
|
if (exynos_pcie->phy_ops.phy_config != NULL)
|
|
exynos_pcie->phy_ops.phy_config(exynos_pcie->phy_base,
|
|
exynos_pcie->phy_pcs_base, exynos_pcie->block_base,
|
|
exynos_pcie->elbi_base, exynos_pcie->ch_num);
|
|
dev_err(pp->dev, "phy clk enable, ret value = %d\n",
|
|
exynos_pcie_phy_clock_enable(&exynos_pcie->pp, 1));
|
|
|
|
/* Bus number enable */
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_SW_WAKE);
|
|
val &= ~(0x1<<1);
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_SW_WAKE);
|
|
}
|
|
|
|
static void exynos_pcie_enable_interrupts(struct pcie_port *pp)
|
|
{
|
|
u32 val;
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
|
|
/* enable INTX interrupt */
|
|
val = IRQ_INTA_ASSERT | IRQ_INTB_ASSERT |
|
|
IRQ_INTC_ASSERT | IRQ_INTD_ASSERT;
|
|
|
|
if (exynos_pcie->pcie_irq_toe_enable)
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_IRQ_TOE_EN_PULSE);
|
|
else
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_IRQ_EN_PULSE);
|
|
/* disable TOE PULSE2 interrupt */
|
|
exynos_elb_writel(exynos_pcie, 0x0, PCIE_IRQ_TOE_EN_PULSE2);
|
|
|
|
/* disable TOE LEVEL interrupt */
|
|
exynos_elb_writel(exynos_pcie, 0x0, PCIE_IRQ_TOE_EN_LEVEL);
|
|
|
|
/* disable LEVEL interrupt */
|
|
exynos_elb_writel(exynos_pcie, 0x0, PCIE_IRQ_EN_LEVEL);
|
|
|
|
/* disable SPECIAL interrupt */
|
|
exynos_elb_writel(exynos_pcie, 0x0, PCIE_IRQ_EN_SPECIAL);
|
|
|
|
return;
|
|
}
|
|
|
|
static irqreturn_t exynos_pcie_irq_handler(int irq, void *arg)
|
|
{
|
|
struct pcie_port *pp = arg;
|
|
u32 val;
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
|
|
/* handle PULSE interrupt */
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_IRQ_PULSE);
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_IRQ_PULSE);
|
|
|
|
if (val & IRQ_LINK_DOWN) {
|
|
dev_info(pp->dev, "!!!PCIE LINK DOWN!!!\n");
|
|
exynos_pcie_print_link_history(pp);
|
|
exynos_pcie_dump_link_down_status(exynos_pcie->ch_num);
|
|
exynos_pcie_register_dump(exynos_pcie->ch_num);
|
|
queue_work(exynos_pcie->pcie_wq, &exynos_pcie->work.work);
|
|
}
|
|
#ifdef CONFIG_PCI_MSI
|
|
if (val & IRQ_MSI_RISING_ASSERT && exynos_pcie->use_msi)
|
|
dw_handle_msi_irq(pp);
|
|
#endif
|
|
|
|
/* handle LEVEL interrupt */
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_IRQ_LEVEL);
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_IRQ_LEVEL);
|
|
|
|
/* handle SPECIAL interrupt */
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_IRQ_SPECIAL);
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_IRQ_SPECIAL);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
#ifdef CONFIG_PCI_MSI
|
|
static void exynos_pcie_msi_init(struct pcie_port *pp)
|
|
{
|
|
u32 val;
|
|
u64 msi_target;
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
|
|
if (!exynos_pcie->use_msi)
|
|
return;
|
|
|
|
/*
|
|
* dw_pcie_msi_init() function allocates msi_data.
|
|
* The following code is added to avoid duplicated allocation.
|
|
*/
|
|
msi_target = virt_to_phys((void *)pp->msi_data);
|
|
|
|
/* program the msi_data */
|
|
exynos_pcie_wr_own_conf(pp, PCIE_MSI_ADDR_LO, 4,
|
|
(u32)(msi_target & 0xffffffff));
|
|
exynos_pcie_wr_own_conf(pp, PCIE_MSI_ADDR_HI, 4,
|
|
(u32)(msi_target >> 32 & 0xffffffff));
|
|
|
|
/* enable MSI interrupt */
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_IRQ_EN_PULSE);
|
|
val |= IRQ_MSI_CTRL_EN_RISING_EDG;
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_IRQ_EN_PULSE);
|
|
|
|
if (exynos_pcie->pcie_irq_msi_cp_enable) {
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_IRQ_CP_EN_PULSE);
|
|
val |= IRQ_MSI_CTRL_EN_RISING_EDG;
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_IRQ_CP_EN_PULSE);
|
|
}
|
|
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
static int exynos_pcie_rd_own_conf(struct pcie_port *pp, int where, int size,
|
|
u32 *val)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
int is_linked = 0;
|
|
int ret = 0;
|
|
|
|
if (exynos_pcie->state == STATE_LINK_UP)
|
|
is_linked = 1;
|
|
|
|
if (is_linked == 0) {
|
|
if (exynos_pcie->phy_ops.phy_check_rx_elecidle != NULL)
|
|
exynos_pcie->phy_ops.phy_check_rx_elecidle(
|
|
exynos_pcie->phy_pcs_base, 1);
|
|
}
|
|
|
|
ret = dw_pcie_cfg_read(exynos_pcie->rc_dbi_base + (where), size, val);
|
|
|
|
if (is_linked == 0) {
|
|
if (exynos_pcie->phy_ops.phy_check_rx_elecidle != NULL)
|
|
exynos_pcie->phy_ops.phy_check_rx_elecidle(
|
|
exynos_pcie->phy_pcs_base, 0);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int exynos_pcie_wr_own_conf(struct pcie_port *pp, int where, int size,
|
|
u32 val)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
int is_linked = 0;
|
|
int ret = 0;
|
|
|
|
if (exynos_pcie->state == STATE_LINK_UP)
|
|
is_linked = 1;
|
|
|
|
if (is_linked == 0) {
|
|
if (exynos_pcie->phy_ops.phy_check_rx_elecidle != NULL)
|
|
exynos_pcie->phy_ops.phy_check_rx_elecidle(
|
|
exynos_pcie->phy_pcs_base, 1);
|
|
}
|
|
|
|
ret = dw_pcie_cfg_write(exynos_pcie->rc_dbi_base + (where), size, val);
|
|
|
|
if (is_linked == 0) {
|
|
if (exynos_pcie->phy_ops.phy_check_rx_elecidle != NULL)
|
|
exynos_pcie->phy_ops.phy_check_rx_elecidle(
|
|
exynos_pcie->phy_pcs_base, 0);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int exynos_pcie_rd_other_conf(struct pcie_port *pp, struct pci_bus *bus,
|
|
u32 devfn, int where, int size, u32 *val)
|
|
{
|
|
int ret, type;
|
|
u32 busdev, cfg_size;
|
|
u64 cpu_addr;
|
|
void __iomem *va_cfg_base;
|
|
|
|
busdev = PCIE_ATU_BUS(bus->number) | PCIE_ATU_DEV(PCI_SLOT(devfn)) |
|
|
PCIE_ATU_FUNC(PCI_FUNC(devfn));
|
|
|
|
if (bus->parent->number == pp->root_bus_nr) {
|
|
type = PCIE_ATU_TYPE_CFG0;
|
|
cpu_addr = pp->cfg0_base;
|
|
cfg_size = pp->cfg0_size;
|
|
va_cfg_base = pp->va_cfg0_base;
|
|
/* setup ATU for cfg/mem outbound */
|
|
exynos_pcie_prog_viewport_cfg0(pp, busdev);
|
|
} else {
|
|
type = PCIE_ATU_TYPE_CFG1;
|
|
cpu_addr = pp->cfg1_base;
|
|
cfg_size = pp->cfg1_size;
|
|
va_cfg_base = pp->va_cfg1_base;
|
|
/* setup ATU for cfg/mem outbound */
|
|
exynos_pcie_prog_viewport_cfg1(pp, busdev);
|
|
}
|
|
|
|
ret = dw_pcie_cfg_read(va_cfg_base + where, size, val);
|
|
exynos_pcie_prog_viewport_mem_outbound(pp);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int exynos_pcie_wr_other_conf(struct pcie_port *pp, struct pci_bus *bus,
|
|
u32 devfn, int where, int size, u32 val)
|
|
{
|
|
|
|
int ret, type;
|
|
u32 busdev, cfg_size;
|
|
u64 cpu_addr;
|
|
void __iomem *va_cfg_base;
|
|
|
|
busdev = PCIE_ATU_BUS(bus->number) | PCIE_ATU_DEV(PCI_SLOT(devfn)) |
|
|
PCIE_ATU_FUNC(PCI_FUNC(devfn));
|
|
|
|
if (bus->parent->number == pp->root_bus_nr) {
|
|
type = PCIE_ATU_TYPE_CFG0;
|
|
cpu_addr = pp->cfg0_base;
|
|
cfg_size = pp->cfg0_size;
|
|
va_cfg_base = pp->va_cfg0_base;
|
|
/* setup ATU for cfg/mem outbound */
|
|
exynos_pcie_prog_viewport_cfg0(pp, busdev);
|
|
} else {
|
|
type = PCIE_ATU_TYPE_CFG1;
|
|
cpu_addr = pp->cfg1_base;
|
|
cfg_size = pp->cfg1_size;
|
|
va_cfg_base = pp->va_cfg1_base;
|
|
/* setup ATU for cfg/mem outbound */
|
|
exynos_pcie_prog_viewport_cfg1(pp, busdev);
|
|
}
|
|
|
|
ret = dw_pcie_cfg_write(va_cfg_base + where, size, val);
|
|
exynos_pcie_prog_viewport_mem_outbound(pp);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int exynos_pcie_link_up(struct pcie_port *pp)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
|
|
if (exynos_pcie->ip_ver >= 0x889500) {
|
|
if (exynos_pcie->state == STATE_LINK_UP)
|
|
return 1;
|
|
} else {
|
|
u32 val;
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_ELBI_RDLH_LINKUP) & 0x1f;
|
|
if (val >= 0x0d && val <= 0x15)
|
|
return 1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct pcie_host_ops exynos_pcie_host_ops = {
|
|
.rd_own_conf = exynos_pcie_rd_own_conf,
|
|
.wr_own_conf = exynos_pcie_wr_own_conf,
|
|
.rd_other_conf = exynos_pcie_rd_other_conf,
|
|
.wr_other_conf = exynos_pcie_wr_other_conf,
|
|
.link_up = exynos_pcie_link_up,
|
|
};
|
|
|
|
static int __init add_pcie_port(struct pcie_port *pp,
|
|
struct platform_device *pdev)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
int ret;
|
|
#ifdef MODIFY_MSI_ADDR
|
|
u64 msi_target;
|
|
#endif
|
|
|
|
pp->irq = platform_get_irq(pdev, 0);
|
|
if (!pp->irq) {
|
|
dev_err(&pdev->dev, "failed to get irq\n");
|
|
return -ENODEV;
|
|
}
|
|
ret = devm_request_irq(&pdev->dev, pp->irq, exynos_pcie_irq_handler,
|
|
IRQF_SHARED, "exynos-pcie", pp);
|
|
if (ret) {
|
|
dev_err(&pdev->dev, "failed to request irq\n");
|
|
return ret;
|
|
}
|
|
|
|
pp->root_bus_nr = 0;
|
|
pp->ops = &exynos_pcie_host_ops;
|
|
|
|
exynos_pcie_setup_rc(pp);
|
|
spin_lock_init(&exynos_pcie->conf_lock);
|
|
ret = dw_pcie_host_init(pp);
|
|
#ifdef CONFIG_PCI_MSI
|
|
dw_pcie_msi_init(pp);
|
|
#ifdef MODIFY_MSI_ADDR
|
|
/* Change MSI address to fit within a 32bit address boundary */
|
|
free_page(pp->msi_data);
|
|
|
|
pp->msi_data = (u64)kmalloc(4, GFP_DMA);
|
|
msi_target = virt_to_phys((void *)pp->msi_data);
|
|
|
|
if ((msi_target >> 32) != 0)
|
|
dev_info(&pdev->dev, "MSI memory is allocated over 32bit boundary\n");
|
|
|
|
/* program the msi_data */
|
|
exynos_pcie_wr_own_conf(pp, PCIE_MSI_ADDR_LO, 4,
|
|
(u32)(msi_target & 0xffffffff));
|
|
exynos_pcie_wr_own_conf(pp, PCIE_MSI_ADDR_HI, 4,
|
|
(u32)(msi_target >> 32 & 0xffffffff));
|
|
#endif
|
|
#endif
|
|
if (ret) {
|
|
dev_err(&pdev->dev, "failed to initialize host\n");
|
|
return ret;
|
|
}
|
|
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void enable_cache_cohernecy(struct pcie_port *pp, int enable)
|
|
{
|
|
int set_val;
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
|
|
if (exynos_pcie->ip_ver > 0x889500) {
|
|
/* Set enable value */
|
|
if (enable) {
|
|
dev_info(pp->dev, "Enable cache coherency.\n");
|
|
set_val = PCIE_SYSREG_SHARABLE_ENABLE;
|
|
} else {
|
|
dev_info(pp->dev, "Disable cache coherency.\n");
|
|
set_val = PCIE_SYSREG_SHARABLE_DISABLE;
|
|
}
|
|
|
|
/* Set configurations */
|
|
regmap_update_bits(exynos_pcie->sysreg,
|
|
PCIE_SYSREG_SHARABILITY_CTRL, /* SYSREG address */
|
|
(0x3 << (PCIE_SYSREG_SHARABLE_OFFSET
|
|
+ exynos_pcie->ch_num * 4)), /* Mask */
|
|
(set_val << (PCIE_SYSREG_SHARABLE_OFFSET
|
|
+ exynos_pcie->ch_num * 4))); /* Set value */
|
|
}
|
|
}
|
|
static int exynos_pcie_parse_dt(struct device_node *np, struct pcie_port *pp)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
const char *use_cache_coherency;
|
|
const char *use_msi;
|
|
const char *use_sicd;
|
|
|
|
if (of_property_read_u32(np, "ip-ver", &exynos_pcie->ip_ver)) {
|
|
dev_err(pp->dev, "Failed to parse the number of ip-ver\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (of_property_read_u32(np, "pcie-clk-num", &exynos_pcie->pcie_clk_num)) {
|
|
dev_err(pp->dev, "Failed to parse the number of pcie clock\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (of_property_read_u32(np, "phy-clk-num", &exynos_pcie->phy_clk_num)) {
|
|
dev_err(pp->dev, "Failed to parse the number of phy clock\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (of_property_read_u32(np, "pcie-irq-toe-enable", &exynos_pcie->pcie_irq_toe_enable)) {
|
|
dev_err(pp->dev, "Failed to parse toe irq enable\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (of_property_read_u32(np, "pcie-irq-toe-num", &exynos_pcie->pcie_irq_toe_num)) {
|
|
dev_err(pp->dev, "Failed to parse the number of toe irq\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (of_property_read_u32(np, "pcie-irq-msi-cp-enable", &exynos_pcie->pcie_irq_msi_cp_enable)) {
|
|
dev_err(pp->dev, "Failed to parse msi cp irq enable\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (of_property_read_u32(np, "pcie-irq-msi-cp-base-bit", &exynos_pcie->pcie_irq_msi_cp_base_bit)) {
|
|
dev_err(pp->dev, "Failed to parse msi cp base bit\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (of_property_read_u32(np, "pcie-irq-msi-cp-num", &exynos_pcie->pcie_irq_msi_cp_num)) {
|
|
dev_err(pp->dev, "Failed to parse the number of msi cp\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (of_property_read_string(np, "use-cache-coherency", &use_cache_coherency)) {
|
|
return -ENODEV;
|
|
} else {
|
|
if (!strcmp(use_cache_coherency, "true")) {
|
|
exynos_pcie->use_cache_coherency = true;
|
|
} else if (!strcmp(use_cache_coherency, "false")) {
|
|
exynos_pcie->use_cache_coherency = false;
|
|
} else {
|
|
dev_err(pp->dev, "invalid use_cache_coherency : (%s)\n", use_cache_coherency);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
if (of_property_read_string(np, "use-msi", &use_msi)) {
|
|
return -ENODEV;
|
|
} else {
|
|
if (!strcmp(use_msi, "true")) {
|
|
exynos_pcie->use_msi = true;
|
|
} else if (!strcmp(use_msi, "false")) {
|
|
exynos_pcie->use_msi = false;
|
|
} else {
|
|
dev_err(pp->dev, "invalid use_msi : (%s)\n", use_msi);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
if (of_property_read_string(np, "use-sicd", &use_sicd)) {
|
|
return -ENODEV;
|
|
} else {
|
|
if (!strcmp(use_sicd, "true")) {
|
|
exynos_pcie->use_sicd = true;
|
|
} else if (!strcmp(use_sicd, "false")) {
|
|
exynos_pcie->use_sicd = false;
|
|
} else {
|
|
dev_err(pp->dev, "invalid use_sicd : (%s)\n", use_sicd);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
#ifdef CONFIG_PM_DEVFREQ
|
|
if (of_property_read_u32(np, "pcie-pm-qos-int", &exynos_pcie->int_min_lock))
|
|
exynos_pcie->int_min_lock = 0;
|
|
|
|
if (exynos_pcie->int_min_lock)
|
|
pm_qos_add_request(&exynos_pcie_int_qos[exynos_pcie->ch_num],
|
|
PM_QOS_DEVICE_THROUGHPUT, 0);
|
|
#endif
|
|
|
|
exynos_pcie->pmureg = syscon_regmap_lookup_by_phandle(np,
|
|
"samsung,syscon-phandle");
|
|
if (IS_ERR(exynos_pcie->pmureg)) {
|
|
dev_err(pp->dev, "syscon regmap lookup failed.\n");
|
|
return PTR_ERR(exynos_pcie->pmureg);
|
|
}
|
|
|
|
exynos_pcie->sysreg = syscon_regmap_lookup_by_phandle(np,
|
|
"samsung,sysreg-phandle");
|
|
/* Check definitions to access SYSREG in DT*/
|
|
if (IS_ERR(exynos_pcie->sysreg) && IS_ERR(exynos_pcie->block_base)) {
|
|
dev_err(pp->dev, "SYSREG is not defined.\n");
|
|
return PTR_ERR(exynos_pcie->sysreg);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __init exynos_pcie_probe(struct platform_device *pdev)
|
|
{
|
|
struct exynos_pcie *exynos_pcie;
|
|
struct pcie_port *pp;
|
|
struct device_node *np = pdev->dev.of_node;
|
|
struct resource *temp_resource;
|
|
struct pinctrl *pinctrl_reset;
|
|
int ret = 0;
|
|
int ch_num;
|
|
|
|
if (create_pcie_sys_file(&pdev->dev))
|
|
dev_err(&pdev->dev, "Failed to create pcie sys file\n");
|
|
|
|
if (of_property_read_u32(np, "ch-num", &ch_num)) {
|
|
dev_err(&pdev->dev, "Failed to parse the channel number\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
exynos_pcie = &g_pcie[ch_num];
|
|
pp = &exynos_pcie->pp;
|
|
pp->dev = &pdev->dev;
|
|
|
|
exynos_pcie->ch_num = ch_num;
|
|
exynos_pcie->l1ss_enable = 1;
|
|
exynos_pcie->state = STATE_LINK_DOWN;
|
|
|
|
exynos_pcie->perst_gpio = of_get_gpio(np, 0);
|
|
|
|
exynos_pcie->linkdown_cnt = 0;
|
|
exynos_pcie->boot_cnt = 0;
|
|
exynos_pcie->work_l1ss_cnt = 0;
|
|
exynos_pcie->l1ss_ctrl_id_state = 0;
|
|
|
|
/* parsing pcie dts data for exynos */
|
|
ret = exynos_pcie_parse_dt(pdev->dev.of_node, pp);
|
|
if (ret)
|
|
goto probe_fail;
|
|
|
|
if (exynos_pcie->perst_gpio < 0) {
|
|
dev_err(&pdev->dev, "cannot get perst_gpio\n");
|
|
} else {
|
|
ret = devm_gpio_request_one(pp->dev, exynos_pcie->perst_gpio,
|
|
GPIOF_OUT_INIT_LOW,
|
|
dev_name(pp->dev));
|
|
if (ret)
|
|
goto probe_fail;
|
|
}
|
|
|
|
pinctrl_reset = devm_pinctrl_get_select(&pdev->dev, "clkreq_output");
|
|
if (IS_ERR(pinctrl_reset)) {
|
|
dev_err(&pdev->dev, "failed to set pcie clkerq pin to output high\n");
|
|
goto probe_fail;
|
|
}
|
|
|
|
#ifdef CONFIG_PCI_EXYNOS_TEST
|
|
exynos_pcie->wlan_gpio = of_get_named_gpio(np, "pcie,wlan-gpio", 0);
|
|
exynos_pcie->bt_gpio = of_get_named_gpio(np, "pcie,bt-gpio", 0);
|
|
|
|
if (exynos_pcie->wlan_gpio < 0) {
|
|
dev_err(&pdev->dev, "cannot get wlan_gpio\n");
|
|
} else {
|
|
ret = devm_gpio_request_one(pp->dev, exynos_pcie->wlan_gpio,
|
|
GPIOF_OUT_INIT_LOW,
|
|
dev_name(pp->dev));
|
|
if (ret)
|
|
goto probe_fail;
|
|
}
|
|
|
|
if (exynos_pcie->bt_gpio < 0) {
|
|
dev_err(&pdev->dev, "cannot get bt_gpio\n");
|
|
} else {
|
|
ret = devm_gpio_request_one(pp->dev, exynos_pcie->bt_gpio,
|
|
GPIOF_OUT_INIT_LOW,
|
|
dev_name(pp->dev));
|
|
if (ret)
|
|
goto probe_fail;
|
|
}
|
|
#endif
|
|
|
|
ret = exynos_pcie_clock_get(pp);
|
|
if (ret)
|
|
goto probe_fail;
|
|
|
|
temp_resource = platform_get_resource_byname(pdev, IORESOURCE_MEM, "elbi");
|
|
exynos_pcie->elbi_base = devm_ioremap_resource(&pdev->dev, temp_resource);
|
|
if (IS_ERR(exynos_pcie->elbi_base)) {
|
|
ret = PTR_ERR(exynos_pcie->elbi_base);
|
|
goto probe_fail;
|
|
}
|
|
|
|
temp_resource = platform_get_resource_byname(pdev, IORESOURCE_MEM, "phy");
|
|
exynos_pcie->phy_base = devm_ioremap_resource(&pdev->dev, temp_resource);
|
|
if (IS_ERR(exynos_pcie->phy_base)) {
|
|
ret = PTR_ERR(exynos_pcie->phy_base);
|
|
goto probe_fail;
|
|
}
|
|
|
|
temp_resource = platform_get_resource_byname(pdev, IORESOURCE_MEM, "sysreg");
|
|
exynos_pcie->block_base = devm_ioremap_resource(&pdev->dev, temp_resource);
|
|
if (IS_ERR(exynos_pcie->block_base)) {
|
|
ret = PTR_ERR(exynos_pcie->block_base);
|
|
goto probe_fail;
|
|
}
|
|
|
|
temp_resource = platform_get_resource_byname(pdev, IORESOURCE_MEM, "dbi");
|
|
exynos_pcie->rc_dbi_base = devm_ioremap_resource(&pdev->dev, temp_resource);
|
|
if (IS_ERR(exynos_pcie->rc_dbi_base)) {
|
|
ret = PTR_ERR(exynos_pcie->rc_dbi_base);
|
|
goto probe_fail;
|
|
}
|
|
|
|
pp->dbi_base = exynos_pcie->rc_dbi_base;
|
|
|
|
temp_resource = platform_get_resource_byname(pdev, IORESOURCE_MEM, "pcs");
|
|
exynos_pcie->phy_pcs_base = devm_ioremap_resource(&pdev->dev, temp_resource);
|
|
if (IS_ERR(exynos_pcie->phy_pcs_base)) {
|
|
ret = PTR_ERR(exynos_pcie->phy_pcs_base);
|
|
goto probe_fail;
|
|
}
|
|
|
|
/* Mapping PHY functions */
|
|
exynos_pcie_phy_init(pp);
|
|
|
|
if (exynos_pcie->use_cache_coherency)
|
|
enable_cache_cohernecy(pp, 1);
|
|
|
|
exynos_pcie_resumed_phydown(pp);
|
|
|
|
regmap_update_bits(exynos_pcie->pmureg,
|
|
PCIE_PHY_CONTROL + exynos_pcie->ch_num * 4,
|
|
PCIE_PHY_CONTROL_MASK, 1);
|
|
ret = add_pcie_port(pp, pdev);
|
|
if (ret)
|
|
goto probe_fail;
|
|
|
|
disable_irq(pp->irq);
|
|
|
|
#ifdef CONFIG_CPU_IDLE
|
|
exynos_pcie->idle_ip_index = exynos_get_idle_ip_index(dev_name(&pdev->dev));
|
|
exynos_update_ip_idle_status(exynos_pcie->idle_ip_index, 1);
|
|
exynos_pcie->lpa_nb.notifier_call = exynos_pci_lpa_event;
|
|
exynos_pcie->lpa_nb.next = NULL;
|
|
exynos_pcie->lpa_nb.priority = 0;
|
|
|
|
ret = exynos_pm_register_notifier(&exynos_pcie->lpa_nb);
|
|
if (ret) {
|
|
dev_err(&pdev->dev, "Failed to register lpa notifier\n");
|
|
goto probe_fail;
|
|
}
|
|
|
|
if (exynos_pcie->use_sicd) {
|
|
if (exynos_pcie->ip_ver >= 0x889500) {
|
|
/* Disable checking PCIe idle(L1.2) signal for SICD */
|
|
regmap_update_bits(exynos_pcie->pmureg, IDLE_IP3_MASK,
|
|
(0x1 << (IDLE_IP_RC0_SHIFT + exynos_pcie->ch_num)),
|
|
(0x1 << (IDLE_IP_RC0_SHIFT + exynos_pcie->ch_num)));
|
|
}
|
|
}
|
|
#endif
|
|
|
|
/* Register Panic notifier */
|
|
exynos_pcie->ss_dma_mon_nb.notifier_call = exynos_pcie_dma_mon_event;
|
|
exynos_pcie->ss_dma_mon_nb.next = NULL;
|
|
exynos_pcie->ss_dma_mon_nb.priority = 0;
|
|
atomic_notifier_chain_register(&panic_notifier_list, &exynos_pcie->ss_dma_mon_nb);
|
|
|
|
exynos_pcie->pcie_wq = create_freezable_workqueue("pcie_wq");
|
|
if (IS_ERR(exynos_pcie->pcie_wq)) {
|
|
dev_err(pp->dev, "couldn't create workqueue\n");
|
|
ret = EBUSY;
|
|
goto probe_fail;
|
|
}
|
|
INIT_DELAYED_WORK(&exynos_pcie->work, exynos_pcie_work);
|
|
|
|
INIT_DELAYED_WORK(&exynos_pcie->work_l1ss, exynos_pcie_work_l1ss);
|
|
platform_set_drvdata(pdev, exynos_pcie);
|
|
|
|
probe_fail:
|
|
if (ret)
|
|
dev_err(&pdev->dev, "%s: pcie probe failed(%d)\n", __func__, ret);
|
|
else
|
|
dev_info(&pdev->dev, "%s: pcie probe success\n", __func__);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int __exit exynos_pcie_remove(struct platform_device *pdev)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = platform_get_drvdata(pdev);
|
|
struct pcie_port *pp = &exynos_pcie->pp;
|
|
u32 val;
|
|
|
|
#ifdef CONFIG_CPU_IDLE
|
|
exynos_pm_unregister_notifier(&exynos_pcie->lpa_nb);
|
|
#endif
|
|
if (exynos_pcie->state > STATE_LINK_DOWN) {
|
|
val = exynos_phy_readl(exynos_pcie, 0x15*4);
|
|
val |= 0xf << 3;
|
|
exynos_phy_writel(exynos_pcie, val, 0x15*4);
|
|
exynos_phy_writel(exynos_pcie, 0xff, 0x4E*4);
|
|
exynos_phy_writel(exynos_pcie, 0x3f, 0x4F*4);
|
|
|
|
exynos_pcie_phy_clock_enable(pp, 0);
|
|
|
|
exynos_pcie_clock_enable(pp, 0);
|
|
#ifdef CONFIG_CPU_IDLE
|
|
if (exynos_pcie->use_sicd) {
|
|
/* need to exclude 0x889500 after aging testing */
|
|
if (exynos_pcie->ip_ver <= 0x889500)
|
|
exynos_update_ip_idle_status(exynos_pcie->idle_ip_index, 1);
|
|
}
|
|
#endif
|
|
}
|
|
|
|
remove_pcie_sys_file(&pdev->dev);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void exynos_pcie_shutdown(struct platform_device *pdev)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = platform_get_drvdata(pdev);
|
|
int ret;
|
|
|
|
ret = atomic_notifier_chain_unregister(&panic_notifier_list, &exynos_pcie->ss_dma_mon_nb);
|
|
|
|
if (ret)
|
|
dev_err(&pdev->dev, "Failed to unregister snapshot kernel panic notifier\n");
|
|
}
|
|
|
|
static const struct of_device_id exynos_pcie_of_match[] = {
|
|
{ .compatible = "samsung,exynos8890-pcie", },
|
|
{ .compatible = "samsung,exynos-pcie", },
|
|
{},
|
|
};
|
|
MODULE_DEVICE_TABLE(of, exynos_pcie_of_match);
|
|
|
|
static void exynos_pcie_resumed_phydown(struct pcie_port *pp)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
|
|
/* phy all power down on wifi off during suspend/resume */
|
|
dev_err(pp->dev, "pcie clk enable, ret value = %d\n", exynos_pcie_clock_enable(pp, 1));
|
|
|
|
exynos_pcie_enable_interrupts(pp);
|
|
regmap_update_bits(exynos_pcie->pmureg,
|
|
PCIE_PHY_CONTROL + exynos_pcie->ch_num * 4,
|
|
PCIE_PHY_CONTROL_MASK, 1);
|
|
|
|
exynos_pcie_assert_phy_reset(pp);
|
|
|
|
/* phy all power down */
|
|
if (exynos_pcie->phy_ops.phy_all_pwrdn != NULL)
|
|
exynos_pcie->phy_ops.phy_all_pwrdn(exynos_pcie->phy_base, exynos_pcie->block_base, exynos_pcie->ch_num);
|
|
|
|
exynos_pcie_phy_clock_enable(pp, 0);
|
|
|
|
exynos_pcie_clock_enable(pp, 0);
|
|
}
|
|
|
|
void exynos_pcie_config_l1ss(struct pcie_port *pp)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
u32 val;
|
|
void __iomem *ep_dbi_base = pp->va_cfg0_base;
|
|
u32 exp_cap_off = EXP_CAP_ID_OFFSET;
|
|
|
|
/* setup ATU for cfg/mem outbound */
|
|
exynos_pcie_prog_viewport_cfg0(pp, 0x1000000);
|
|
exynos_pcie_prog_viewport_mem_outbound(pp);
|
|
|
|
if (exynos_pcie->l1ss_ctrl_id_state == 0) {
|
|
/* Enable L1SS on Root Complex */
|
|
val = readl(ep_dbi_base + 0xbc);
|
|
val &= ~0x3;
|
|
val |= 0x142;
|
|
writel(val, ep_dbi_base + 0xBC);
|
|
val = readl(ep_dbi_base + 0x248);
|
|
writel(val | 0xa0f, ep_dbi_base + 0x248);
|
|
dev_err(pp->dev, "l1ss enabled(0x%x)\n", exynos_pcie->l1ss_ctrl_id_state);
|
|
} else {
|
|
/* disable L1SS on Root Complex */
|
|
val = readl(ep_dbi_base + 0xbc);
|
|
writel(val & ~0x3, ep_dbi_base + 0xBC);
|
|
val = readl(ep_dbi_base + 0x248);
|
|
writel(val & ~0xf, ep_dbi_base + 0x248);
|
|
dev_err(pp->dev, "l1ss disabled(0x%x)\n", exynos_pcie->l1ss_ctrl_id_state);
|
|
}
|
|
|
|
writel(PORT_LINK_TPOWERON_130US, ep_dbi_base + 0x24C);
|
|
writel(0x10031003, ep_dbi_base + 0x1B4);
|
|
val = readl(ep_dbi_base + 0xD4);
|
|
writel(val | (1 << 10), ep_dbi_base + 0xD4);
|
|
|
|
exynos_pcie_rd_own_conf(pp, PCIE_LINK_L1SS_CONTROL, 4, &val);
|
|
val |= PORT_LINK_TCOMMON_32US | PORT_LINK_L1SS_ENABLE;
|
|
exynos_pcie_wr_own_conf(pp, PCIE_LINK_L1SS_CONTROL, 4, val);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_LINK_L1SS_CONTROL2, 4,
|
|
PORT_LINK_TPOWERON_130US);
|
|
val = PORT_LINK_L1SS_T_PCLKACK | PORT_LINK_L1SS_T_L1_2 |
|
|
PORT_LINK_L1SS_T_POWER_OFF;
|
|
exynos_pcie_wr_own_conf(pp, PCIE_LINK_L1SS_OFF, 4, val);
|
|
|
|
exynos_pcie_rd_own_conf(pp, exp_cap_off + PCI_EXP_LNKCTL, 4, &val);
|
|
val &= ~PCI_EXP_LNKCTL_ASPMC;
|
|
val |= PCI_EXP_LNKCTL_CCC | PCI_EXP_LNKCTL_ASPM_L1;
|
|
exynos_pcie_wr_own_conf(pp, exp_cap_off + PCI_EXP_LNKCTL, 4, val);
|
|
exynos_pcie_wr_own_conf(pp, exp_cap_off + PCI_EXP_DEVCTL2, 4,
|
|
PCI_EXP_DEVCTL2_LTR_EN);
|
|
}
|
|
|
|
int exynos_pcie_poweron(int ch_num)
|
|
{
|
|
struct pcie_port *pp = &g_pcie[ch_num].pp;
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
struct pinctrl *pinctrl_reset;
|
|
u32 val, vendor_id, device_id;
|
|
|
|
dev_info(pp->dev, "%s, start of poweron, pcie state: %d\n", __func__,
|
|
exynos_pcie->state);
|
|
|
|
if (exynos_pcie->state == STATE_LINK_DOWN) {
|
|
#ifdef CONFIG_CPU_IDLE
|
|
if (exynos_pcie->use_sicd) {
|
|
if (exynos_pcie->ip_ver >= 0x889500) {
|
|
/* Disable checking PCIe idle(L1.2) signal for SICD */
|
|
regmap_update_bits(exynos_pcie->pmureg, IDLE_IP3_MASK,
|
|
(0x1 << (IDLE_IP_RC0_SHIFT + exynos_pcie->ch_num)),
|
|
(0x1 << (IDLE_IP_RC0_SHIFT + exynos_pcie->ch_num)));
|
|
} else {
|
|
exynos_update_ip_idle_status(exynos_pcie->idle_ip_index, 0);
|
|
}
|
|
}
|
|
#endif
|
|
#ifdef CONFIG_PM_DEVFREQ
|
|
if (exynos_pcie->int_min_lock)
|
|
pm_qos_update_request(&exynos_pcie_int_qos[ch_num],
|
|
exynos_pcie->int_min_lock);
|
|
#endif
|
|
pinctrl_reset = devm_pinctrl_get_select_default(pp->dev);
|
|
if (IS_ERR(pinctrl_reset)) {
|
|
dev_err(pp->dev, "faied to set pcie pin state\n");
|
|
goto poweron_fail;
|
|
}
|
|
|
|
dev_err(pp->dev, "pcie clk enable, ret value = %d\n", exynos_pcie_clock_enable(pp, 1));
|
|
regmap_update_bits(exynos_pcie->pmureg,
|
|
PCIE_PHY_CONTROL + exynos_pcie->ch_num * 4,
|
|
PCIE_PHY_CONTROL_MASK, 1);
|
|
|
|
/* phy all power down clear */
|
|
if (exynos_pcie->phy_ops.phy_all_pwrdn_clear != NULL)
|
|
exynos_pcie->phy_ops.phy_all_pwrdn_clear(exynos_pcie->phy_base, exynos_pcie->block_base, exynos_pcie->ch_num);
|
|
|
|
exynos_pcie->state = STATE_LINK_UP_TRY;
|
|
|
|
/* Enable history buffer */
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_STATE_HISTORY_CHECK);
|
|
val |= HISTORY_BUFFER_ENABLE;
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_STATE_HISTORY_CHECK);
|
|
|
|
if (exynos_pcie_establish_link(pp)) {
|
|
dev_err(pp->dev, "pcie link up fail\n");
|
|
goto poweron_fail;
|
|
}
|
|
exynos_pcie->state = STATE_LINK_UP;
|
|
|
|
if (!exynos_pcie->probe_ok) {
|
|
/* dw_pcie_scan function is removed in Kernel4.4.
|
|
if (dw_pcie_scan(pp)) {
|
|
dev_err(pp->dev, "pcie scan fail\n");
|
|
goto poweron_fail;
|
|
}
|
|
*/
|
|
#ifdef ORGCFG_IN_DWSCAN
|
|
/* Maybe it's not needed. */
|
|
exynos_pcie_wr_own_conf(pp, PCI_BASE_ADDRESS_0, 4, 0);
|
|
|
|
/* program correct class for RC */
|
|
exynos_pcie_wr_own_conf(pp, PCI_CLASS_DEVICE, 2, PCI_CLASS_BRIDGE_PCI);
|
|
|
|
exynos_pcie_rd_own_conf(pp, PCIE_LINK_WIDTH_SPEED_CONTROL, 4, &val);
|
|
val |= PORT_LOGIC_SPEED_CHANGE;
|
|
exynos_pcie_wr_own_conf(pp, PCIE_LINK_WIDTH_SPEED_CONTROL, 4, val);
|
|
#endif
|
|
exynos_pcie_rd_own_conf(pp, PCI_VENDOR_ID, 4, &val);
|
|
vendor_id = val & ID_MASK;
|
|
device_id = (val >> 16) & ID_MASK;
|
|
|
|
exynos_pcie->pci_dev = pci_get_device(vendor_id, device_id, NULL);
|
|
if (!exynos_pcie->pci_dev) {
|
|
dev_err(pp->dev, "Failed to get pci device\n");
|
|
goto poweron_fail;
|
|
}
|
|
|
|
pci_scan_child_bus(exynos_pcie->pci_dev->bus);
|
|
pci_assign_unassigned_bus_resources(exynos_pcie->pci_dev->bus);
|
|
pci_bus_add_devices(exynos_pcie->pci_dev->bus);
|
|
|
|
/* L1.2 ASPM enable */
|
|
exynos_pcie_config_l1ss(&exynos_pcie->pp);
|
|
|
|
#ifdef CONFIG_PCI_MSI
|
|
exynos_pcie_msi_init(pp);
|
|
#endif
|
|
|
|
if (pci_save_state(exynos_pcie->pci_dev)) {
|
|
dev_err(pp->dev, "Failed to save pcie state\n");
|
|
goto poweron_fail;
|
|
}
|
|
exynos_pcie->pci_saved_configs =
|
|
pci_store_saved_state(exynos_pcie->pci_dev);
|
|
exynos_pcie->probe_ok = 1;
|
|
} else if (exynos_pcie->probe_ok) {
|
|
if (exynos_pcie->boot_cnt == 0) {
|
|
schedule_delayed_work(&exynos_pcie->work_l1ss, msecs_to_jiffies(40000));
|
|
exynos_pcie->boot_cnt++;
|
|
exynos_pcie->l1ss_ctrl_id_state = PCIE_L1SS_CTRL_BOOT;
|
|
}
|
|
/* L1.2 ASPM enable */
|
|
exynos_pcie_config_l1ss(pp);
|
|
#ifdef CONFIG_PCI_MSI
|
|
exynos_pcie_msi_init(pp);
|
|
#endif
|
|
|
|
if (pci_load_saved_state(exynos_pcie->pci_dev,
|
|
exynos_pcie->pci_saved_configs)) {
|
|
dev_err(pp->dev, "Failed to load pcie state\n");
|
|
goto poweron_fail;
|
|
}
|
|
pci_restore_state(exynos_pcie->pci_dev);
|
|
}
|
|
enable_irq(pp->irq);
|
|
}
|
|
|
|
dev_info(pp->dev, "%s, end of poweron, pcie state: %d\n", __func__,
|
|
exynos_pcie->state);
|
|
|
|
return 0;
|
|
|
|
poweron_fail:
|
|
exynos_pcie->state = STATE_LINK_UP;
|
|
exynos_pcie_poweroff(exynos_pcie->ch_num);
|
|
|
|
return -EPIPE;
|
|
}
|
|
EXPORT_SYMBOL(exynos_pcie_poweron);
|
|
|
|
void exynos_pcie_poweroff(int ch_num)
|
|
{
|
|
struct pcie_port *pp = &g_pcie[ch_num].pp;
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
struct pinctrl *pinctrl_reset;
|
|
unsigned long flags;
|
|
u32 val;
|
|
|
|
dev_info(pp->dev, "%s, start of poweroff, pcie state: %d\n", __func__,
|
|
exynos_pcie->state);
|
|
|
|
exynos_pcie_send_pme_turn_off(exynos_pcie);
|
|
|
|
if (exynos_pcie->state == STATE_LINK_UP ||
|
|
exynos_pcie->state == STATE_LINK_DOWN_TRY) {
|
|
exynos_pcie->state = STATE_LINK_DOWN_TRY;
|
|
|
|
disable_irq(pp->irq);
|
|
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_IRQ_EN_PULSE);
|
|
val &= ~IRQ_LINKDOWN_ENABLE;
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_IRQ_EN_PULSE);
|
|
|
|
spin_lock_irqsave(&exynos_pcie->conf_lock, flags);
|
|
exynos_pcie->state = STATE_LINK_DOWN;
|
|
|
|
/* Disable history buffer */
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_STATE_HISTORY_CHECK);
|
|
val &= ~HISTORY_BUFFER_ENABLE;
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_STATE_HISTORY_CHECK);
|
|
|
|
dev_info(pp->dev, "%s: Before set perst, gpio val = %d\n",
|
|
__func__, gpio_get_value(exynos_pcie->perst_gpio));
|
|
gpio_set_value(exynos_pcie->perst_gpio, 0);
|
|
dev_info(pp->dev, "%s: After set perst, gpio val = %d\n",
|
|
__func__, gpio_get_value(exynos_pcie->perst_gpio));
|
|
|
|
/* LTSSM disable */
|
|
writel(PCIE_ELBI_LTSSM_DISABLE, exynos_pcie->elbi_base + PCIE_APP_LTSSM_ENABLE);
|
|
|
|
/* phy all power down */
|
|
if (exynos_pcie->phy_ops.phy_all_pwrdn != NULL)
|
|
exynos_pcie->phy_ops.phy_all_pwrdn(exynos_pcie->phy_base, exynos_pcie->block_base, exynos_pcie->ch_num);
|
|
|
|
spin_unlock_irqrestore(&exynos_pcie->conf_lock, flags);
|
|
exynos_pcie_phy_clock_enable(pp, 0);
|
|
|
|
exynos_pcie_clock_enable(pp, 0);
|
|
|
|
pinctrl_reset = devm_pinctrl_get_select(pp->dev, "clkreq_output");
|
|
if (IS_ERR(pinctrl_reset)) {
|
|
dev_err(pp->dev, "failed to set pcie clkerq pin to output high\n");
|
|
return;
|
|
}
|
|
#ifdef CONFIG_PM_DEVFREQ
|
|
if (exynos_pcie->int_min_lock)
|
|
pm_qos_update_request(&exynos_pcie_int_qos[ch_num], 0);
|
|
#endif
|
|
#ifdef CONFIG_CPU_IDLE
|
|
if (exynos_pcie->use_sicd) {
|
|
/* need to exclude 0x889500 after aging testing */
|
|
if (exynos_pcie->ip_ver <= 0x889500)
|
|
exynos_update_ip_idle_status(exynos_pcie->idle_ip_index, 1);
|
|
}
|
|
#endif
|
|
}
|
|
|
|
dev_info(pp->dev, "%s, end of poweroff, pcie state: %d\n", __func__,
|
|
exynos_pcie->state);
|
|
}
|
|
EXPORT_SYMBOL(exynos_pcie_poweroff);
|
|
|
|
void exynos_pcie_send_pme_turn_off(struct exynos_pcie *exynos_pcie)
|
|
{
|
|
struct pcie_port *pp = &exynos_pcie->pp;
|
|
struct device *dev = pp->dev;
|
|
int __maybe_unused count = 0;
|
|
u32 __maybe_unused val;
|
|
|
|
val = readl(exynos_pcie->elbi_base + PCIE_ELBI_RDLH_LINKUP) & 0x1f;
|
|
if (!(val >= 0x0d && val <= 0x14)) {
|
|
dev_info(dev, "%s, pcie link is not up\n", __func__);
|
|
return;
|
|
}
|
|
|
|
exynos_elb_writel(exynos_pcie, 0x1, PCIE_APP_REQ_EXIT_L1);
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_APP_REQ_EXIT_L1_MODE);
|
|
val &= ~APP_REQ_EXIT_L1_MODE;
|
|
val |= L1_REQ_NAK_CONTROL_MASTER;
|
|
exynos_elb_writel(exynos_pcie, val, PCIE_APP_REQ_EXIT_L1_MODE);
|
|
exynos_elb_writel(exynos_pcie, 0x0, 0xa8);
|
|
exynos_elb_writel(exynos_pcie, 0x1, 0xac);
|
|
exynos_elb_writel(exynos_pcie, 0x13, 0xb0);
|
|
exynos_elb_writel(exynos_pcie, 0x19, 0xd0);
|
|
exynos_elb_writel(exynos_pcie, 0x1, 0xa8);
|
|
|
|
while (count < MAX_TIMEOUT) {
|
|
if ((exynos_elb_readl(exynos_pcie, PCIE_IRQ_PULSE)
|
|
& IRQ_RADM_PM_TO_ACK)) {
|
|
dev_err(dev, "ack message is ok\n");
|
|
break;
|
|
}
|
|
|
|
udelay(10);
|
|
count++;
|
|
}
|
|
|
|
if (count >= MAX_TIMEOUT)
|
|
dev_err(dev, "cannot receive ack message from wifi\n");
|
|
|
|
exynos_elb_writel(exynos_pcie, 0x0, PCIE_APP_REQ_EXIT_L1);
|
|
|
|
do {
|
|
val = exynos_elb_readl(exynos_pcie, PCIE_ELBI_RDLH_LINKUP);
|
|
val = val & 0x1f;
|
|
if (val == 0x15) {
|
|
dev_err(dev, "received Enter_L23_READY DLLP packet\n");
|
|
break;
|
|
}
|
|
udelay(10);
|
|
count++;
|
|
} while (count < MAX_TIMEOUT);
|
|
|
|
if (count >= MAX_TIMEOUT)
|
|
dev_err(dev, "cannot receive L23_READY DLLP packet\n");
|
|
}
|
|
|
|
void exynos_pcie_pm_suspend(int ch_num)
|
|
{
|
|
struct pcie_port *pp = &g_pcie[ch_num].pp;
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
unsigned long flags;
|
|
|
|
if (exynos_pcie->state == STATE_LINK_DOWN) {
|
|
dev_info(pp->dev, "RC%d already off\n", exynos_pcie->ch_num);
|
|
return;
|
|
}
|
|
|
|
spin_lock_irqsave(&exynos_pcie->conf_lock, flags);
|
|
exynos_pcie->state = STATE_LINK_DOWN_TRY;
|
|
spin_unlock_irqrestore(&exynos_pcie->conf_lock, flags);
|
|
|
|
exynos_pcie_poweroff(ch_num);
|
|
}
|
|
EXPORT_SYMBOL(exynos_pcie_pm_suspend);
|
|
|
|
int exynos_pcie_pm_resume(int ch_num)
|
|
{
|
|
return exynos_pcie_poweron(ch_num);
|
|
}
|
|
EXPORT_SYMBOL(exynos_pcie_pm_resume);
|
|
|
|
#ifdef CONFIG_PM
|
|
static int exynos_pcie_suspend_noirq(struct device *dev)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = dev_get_drvdata(dev);
|
|
|
|
if (exynos_pcie->state == STATE_LINK_DOWN) {
|
|
dev_info(dev, "RC%d already off\n", exynos_pcie->ch_num);
|
|
return 0;
|
|
}
|
|
|
|
exynos_pcie_send_pme_turn_off(exynos_pcie);
|
|
dev_info(dev, "%s: Before set perst, gpio val = %d\n",
|
|
__func__, gpio_get_value(exynos_pcie->perst_gpio));
|
|
gpio_set_value(exynos_pcie->perst_gpio, 0);
|
|
dev_info(dev, "%s: After set perst, gpio val = %d\n",
|
|
__func__, gpio_get_value(exynos_pcie->perst_gpio));
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int exynos_pcie_resume_noirq(struct device *dev)
|
|
{
|
|
struct exynos_pcie *exynos_pcie = dev_get_drvdata(dev);
|
|
|
|
if (exynos_pcie->state == STATE_LINK_DOWN) {
|
|
exynos_pcie_resumed_phydown(&exynos_pcie->pp);
|
|
return 0;
|
|
}
|
|
|
|
exynos_pcie_enable_interrupts(&exynos_pcie->pp);
|
|
|
|
regmap_update_bits(exynos_pcie->pmureg,
|
|
PCIE_PHY_CONTROL + exynos_pcie->ch_num * 4,
|
|
PCIE_PHY_CONTROL_MASK, 1);
|
|
|
|
exynos_pcie_establish_link(&exynos_pcie->pp);
|
|
|
|
/* L1.2 ASPM enable */
|
|
exynos_pcie_config_l1ss(&exynos_pcie->pp);
|
|
|
|
return 0;
|
|
}
|
|
|
|
#else
|
|
#define exynos_pcie_suspend_noirq NULL
|
|
#define exynos_pcie_resume_noirq NULL
|
|
#endif
|
|
|
|
static const struct dev_pm_ops exynos_pcie_dev_pm_ops = {
|
|
.suspend_noirq = exynos_pcie_suspend_noirq,
|
|
.resume_noirq = exynos_pcie_resume_noirq,
|
|
};
|
|
|
|
static struct platform_driver exynos_pcie_driver = {
|
|
.remove = __exit_p(exynos_pcie_remove),
|
|
.shutdown = exynos_pcie_shutdown,
|
|
.driver = {
|
|
.name = "exynos-pcie",
|
|
.owner = THIS_MODULE,
|
|
.of_match_table = exynos_pcie_of_match,
|
|
.pm = &exynos_pcie_dev_pm_ops,
|
|
},
|
|
};
|
|
|
|
#ifdef CONFIG_CPU_IDLE
|
|
static void __maybe_unused exynos_pcie_set_tpoweron(struct pcie_port *pp, int max)
|
|
{
|
|
void __iomem *ep_dbi_base = pp->va_cfg0_base;
|
|
struct exynos_pcie *exynos_pcie = to_exynos_pcie(pp);
|
|
u32 val;
|
|
|
|
if (exynos_pcie->state != STATE_LINK_UP)
|
|
return;
|
|
|
|
/* setup ATU for cfg/mem outbound */
|
|
exynos_pcie_prog_viewport_cfg0(pp, 0x1000000);
|
|
exynos_pcie_prog_viewport_mem_outbound(pp);
|
|
|
|
/* Disable ASPM */
|
|
val = readl(ep_dbi_base + 0xbc);
|
|
val &= ~0x3;
|
|
writel(val, ep_dbi_base + 0xBC);
|
|
|
|
val = readl(ep_dbi_base + 0x248);
|
|
val &= ~0xF;
|
|
writel(val, ep_dbi_base + 0x248);
|
|
|
|
if (max) {
|
|
writel(PORT_LINK_TPOWERON_3100US, ep_dbi_base + 0x24C);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_LINK_L1SS_CONTROL2, 4, PORT_LINK_TPOWERON_3100US);
|
|
} else {
|
|
writel(PORT_LINK_TPOWERON_130US, ep_dbi_base + 0x24C);
|
|
exynos_pcie_wr_own_conf(pp, PCIE_LINK_L1SS_CONTROL2, 4, PORT_LINK_TPOWERON_130US);
|
|
}
|
|
|
|
/* Enable L1ss */
|
|
val = readl(ep_dbi_base + 0xBC);
|
|
val |= 0x2;
|
|
writel(val, ep_dbi_base + 0xBC);
|
|
|
|
val = readl(ep_dbi_base + 0x248);
|
|
val |= PORT_LINK_L1SS_ENABLE;
|
|
writel(val, ep_dbi_base + 0x248);
|
|
|
|
}
|
|
|
|
static int exynos_pci_lpa_event(struct notifier_block *nb, unsigned long event, void *data)
|
|
{
|
|
int ret = NOTIFY_DONE;
|
|
struct exynos_pcie *exynos_pcie = container_of(nb,
|
|
struct exynos_pcie, lpa_nb);
|
|
u32 val;
|
|
struct pcie_port *pp = &exynos_pcie->pp;
|
|
|
|
switch (event) {
|
|
case LPA_EXIT:
|
|
if (exynos_pcie->state == STATE_LINK_DOWN)
|
|
exynos_pcie_resumed_phydown(&exynos_pcie->pp);
|
|
break;
|
|
case SICD_ENTER:
|
|
if (exynos_pcie->use_sicd) {
|
|
if (exynos_pcie->ip_ver >= 0x889500) {
|
|
if (exynos_pcie->state != STATE_LINK_DOWN) {
|
|
val = readl(exynos_pcie->elbi_base + PCIE_ELBI_RDLH_LINKUP) & 0x1f;
|
|
if (val == 0x14 || val == 0x15) {
|
|
ret = NOTIFY_DONE;
|
|
/* Change tpower on time to MAX value */
|
|
exynos_pcie_set_tpoweron(pp, 1);
|
|
} else {
|
|
ret = NOTIFY_BAD;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
case SICD_EXIT:
|
|
if (exynos_pcie->use_sicd) {
|
|
if (exynos_pcie->ip_ver >= 0x889500) {
|
|
if (exynos_pcie->state != STATE_LINK_DOWN) {
|
|
/* Change tpower on time to NORMAL value */
|
|
exynos_pcie_set_tpoweron(pp, 0);
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
default:
|
|
ret = NOTIFY_DONE;
|
|
}
|
|
|
|
return notifier_from_errno(ret);
|
|
}
|
|
#endif
|
|
|
|
/* Exynos PCIe driver does not allow module unload */
|
|
|
|
static int __init pcie_init(void)
|
|
{
|
|
return platform_driver_probe(&exynos_pcie_driver, exynos_pcie_probe);
|
|
}
|
|
device_initcall(pcie_init);
|
|
|
|
int exynos_pcie_register_event(struct exynos_pcie_register_event *reg)
|
|
{
|
|
int ret = 0;
|
|
struct pcie_port *pp;
|
|
struct exynos_pcie *exynos_pcie;
|
|
if (!reg) {
|
|
pr_err("PCIe: Event registration is NULL\n");
|
|
return -ENODEV;
|
|
}
|
|
if (!reg->user) {
|
|
pr_err("PCIe: User of event registration is NULL\n");
|
|
return -ENODEV;
|
|
}
|
|
pp = PCIE_BUS_PRIV_DATA(((struct pci_dev *)reg->user));
|
|
exynos_pcie = to_exynos_pcie(pp);
|
|
if (pp) {
|
|
exynos_pcie->event_reg = reg;
|
|
dev_info(pp->dev,
|
|
"Event 0x%x is registered for RC %d\n",
|
|
reg->events, exynos_pcie->ch_num);
|
|
} else {
|
|
pr_err("PCIe: did not find RC for pci endpoint device\n");
|
|
ret = -ENODEV;
|
|
}
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(exynos_pcie_register_event);
|
|
|
|
int exynos_pcie_deregister_event(struct exynos_pcie_register_event *reg)
|
|
{
|
|
int ret = 0;
|
|
struct pcie_port *pp;
|
|
struct exynos_pcie *exynos_pcie;
|
|
if (!reg) {
|
|
pr_err("PCIe: Event deregistration is NULL\n");
|
|
return -ENODEV;
|
|
}
|
|
if (!reg->user) {
|
|
pr_err("PCIe: User of event deregistration is NULL\n");
|
|
return -ENODEV;
|
|
}
|
|
pp = PCIE_BUS_PRIV_DATA(((struct pci_dev *)reg->user));
|
|
exynos_pcie = to_exynos_pcie(pp);
|
|
if (pp) {
|
|
exynos_pcie->event_reg = NULL;
|
|
dev_info(pp->dev, "Event is deregistered for RC %d\n",
|
|
exynos_pcie->ch_num);
|
|
} else {
|
|
pr_err("PCIe: did not find RC for pci endpoint device\n");
|
|
ret = -ENODEV;
|
|
}
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(exynos_pcie_deregister_event);
|
|
|
|
int exynos_pcie_gpio_debug(int ch_num, int option)
|
|
{
|
|
void __iomem *sysreg;
|
|
u32 addr, val;
|
|
|
|
addr = (GPIO_DEBUG_SFR);
|
|
|
|
sysreg = ioremap_nocache(addr, 4);
|
|
if (!sysreg) {
|
|
pr_err("Cannot get the sysreg address\n");
|
|
return -EPIPE;
|
|
}
|
|
val = readl(sysreg);
|
|
val &= ~FSYS1_MON_SEL_MASK;
|
|
val |= ch_num + 2;
|
|
val &= ~(PCIE_MON_SEL_MASK << ((ch_num + 2) * 8));
|
|
val |= option << ((ch_num + 2) * 8);
|
|
writel(val, sysreg);
|
|
iounmap(sysreg);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(exynos_pcie_gpio_debug);
|
|
|
|
MODULE_AUTHOR("Jingoo Han <jg1.han@samsung.com>");
|
|
MODULE_DESCRIPTION("Samsung PCIe host controller driver");
|
|
MODULE_LICENSE("GPL v2");
|