123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754 |
- /*
- * Designware application register space functions for Keystone PCI controller
- *
- * Copyright (C) 2013-2014 Texas Instruments., Ltd.
- * http://www.ti.com
- *
- * Author: Murali Karicheri <m-karicheri2@ti.com>
- *
- *
- * This program is free software; you can redistribute it and/or modify
- * it under the terms of the GNU General Public License version 2 as
- * published by the Free Software Foundation.
- */
- #include <linux/delay.h>
- #include <linux/irq.h>
- #include <linux/irqdomain.h>
- #include <linux/irqreturn.h>
- #include <linux/module.h>
- #include <linux/of.h>
- #include <linux/of_pci.h>
- #include <linux/pci.h>
- #include <linux/platform_device.h>
- #include "pcie-designware.h"
- #include "pci-keystone.h"
- /* Application register defines */
- #define LTSSM_EN_VAL 1
- #define LTSSM_STATE_MASK 0x1f
- #define LTSSM_STATE_L0 0x11
- #define DBI_CS2_EN_VAL 0x20
- #define OB_XLAT_EN_VAL 2
- #define IB_XLAT_EN_VAL 4
- /* Application registers */
- #define CMD_STATUS 0x004
- #define CFG_SETUP 0x008
- #define OB_SIZE 0x030
- #define CFG_PCIM_WIN_SZ_IDX 3
- #define CFG_PCIM_WIN_CNT 32
- #define SPACE0_REMOTE_CFG_OFFSET 0x1000
- #define OB_OFFSET_INDEX(n) (0x200 + (8 * n))
- #define OB_OFFSET_HI(n) (0x204 + (8 * n))
- #define IB_BAR(n) (0x300 + (0x10 * (n)))
- #define IB_START_LO(n) (0x304 + (0x10 * (n)))
- #define IB_START_HI(n) (0x308 + (0x10 * (n)))
- #define IB_OFFSET(n) (0x30c + (0x10 * (n)))
- /* IRQ register defines */
- #define IRQ_EOI 0x050
- #define IRQ_STATUS 0x184
- #define IRQ_ENABLE_SET 0x188
- #define IRQ_ENABLE_CLR 0x18c
- #define MSI_IRQ 0x054
- #define MSI0_IRQ_STATUS 0x104
- #define MSI0_IRQ_ENABLE_SET 0x108
- #define MSI0_IRQ_ENABLE_CLR 0x10c
- #define IRQ_STATUS 0x184
- #define MSI_IRQ_OFFSET 4
- /* Error IRQ bits */
- #define ERR_AER BIT(5) /* ECRC error */
- #define ERR_AXI BIT(4) /* AXI tag lookup fatal error */
- #define ERR_CORR BIT(3) /* Correctable error */
- #define ERR_NONFATAL BIT(2) /* Non-fatal error */
- #define ERR_FATAL BIT(1) /* Fatal error */
- #define ERR_SYS BIT(0) /* System (fatal, non-fatal, or correctable) */
- #define ERR_IRQ_ALL (ERR_AER | ERR_AXI | ERR_CORR | \
- ERR_NONFATAL | ERR_FATAL | ERR_SYS)
- #define ERR_FATAL_IRQ (ERR_FATAL | ERR_AXI)
- #define ERR_IRQ_STATUS_RAW 0x1c0
- #define ERR_IRQ_STATUS 0x1c4
- #define ERR_IRQ_ENABLE_SET 0x1c8
- #define ERR_IRQ_ENABLE_CLR 0x1cc
- /* Config space registers */
- #define DEBUG0 0x728
- #define APP_ADDR_SPACE_0 (16 * SZ_1K)
- #define WIN_INDEX_MASK 0x1f
- #define WIN_INDEX_SHIFT 20
- #define WIN_SIZE SZ_1M
- #define PCIE_LEGACY_IRQ_ENABLE_SET(n) (0x188 + (0x10 * ((n) - 1)))
- #define PCIE_LEGACY_IRQ_ENABLE_CLR(n) (0x18c + (0x10 * ((n) - 1)))
- #define PCIE_EP_IRQ_SET 0x64
- #define PCIE_EP_IRQ_CLR 0x68
- #define OB_ENABLEN 0x1
- #define INT_ENABLE 0x1
- #define to_keystone_pcie(x) dev_get_drvdata((x)->dev)
- static void ks_dw_app_writel(struct keystone_pcie *ks_pcie, u32 offset,
- u32 val);
- static u32 ks_dw_app_readl(struct keystone_pcie *ks_pcie, u32 offset);
- static void ks_dw_pcie_set_dbi_mode(struct keystone_pcie *ks_pcie);
- static void ks_dw_pcie_clear_dbi_mode(struct keystone_pcie *ks_pcie);
- void ks_dw_pcie_ep_init(struct dw_pcie_ep *ep)
- {
- struct dw_pcie *pci = to_dw_pcie_from_ep(ep);
- struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
- int flags;
- u32 val;
- ep->page_size = WIN_SIZE;
- flags = PCI_BASE_ADDRESS_SPACE_MEMORY | PCI_BASE_ADDRESS_MEM_TYPE_32;
- dw_pcie_writel_dbi2(pci, PCI_BASE_ADDRESS_0, APP_ADDR_SPACE_0 - 1);
- dw_pcie_writel_dbi(pci, PCI_BASE_ADDRESS_0, flags);
- ks_dw_app_writel(ks_pcie, OB_SIZE, 0x0);
- val = ks_dw_app_readl(ks_pcie, CMD_STATUS);
- ks_dw_app_writel(ks_pcie, CMD_STATUS, val | OB_XLAT_EN_VAL |
- IB_XLAT_EN_VAL);
- }
- void ks_dw_pcie_disable_outbound_atu(struct keystone_pcie *ks_pcie,
- phys_addr_t addr)
- {
- u8 index;
- u8 regions;
- u64 cpu_addr;
- index = (addr >> WIN_INDEX_SHIFT) & WIN_INDEX_MASK;
- regions = ks_pcie->ob_win[index].no_of_regions;
- cpu_addr = ks_pcie->ob_win[index].cpu_addr;
- WARN_ON(cpu_addr != addr);
- while (regions--) {
- ks_dw_app_writel(ks_pcie, OB_OFFSET_INDEX(index), 0x0);
- clear_bit(index++, &ks_pcie->ob_window_map);
- }
- }
- void ks_dw_pcie_disable_inbound_atu(struct keystone_pcie *ks_pcie, int index)
- {
- ks_dw_app_writel(ks_pcie, IB_BAR(index), 0x0);
- ks_dw_app_writel(ks_pcie, IB_START_LO(index), 0x0);
- ks_dw_app_writel(ks_pcie, IB_START_HI(index), 0x0);
- ks_dw_app_writel(ks_pcie, IB_OFFSET(index), 0x0);
- }
- void ks_dw_pcie_disable_atu(struct dw_pcie *pci, phys_addr_t addr, int index,
- enum dw_pcie_region_type type)
- {
- struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
- switch (type) {
- case DW_PCIE_REGION_INBOUND:
- ks_dw_pcie_disable_inbound_atu(ks_pcie, index);
- break;
- case DW_PCIE_REGION_OUTBOUND:
- ks_dw_pcie_disable_outbound_atu(ks_pcie, addr);
- break;
- default:
- return;
- }
- }
- int ks_dw_pcie_inbound_atu(struct dw_pcie *pci, u32 index,
- enum pci_barno bar, dma_addr_t cpu_addr)
- {
- struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
- struct device *dev = pci->dev;
- if (bar == BAR_0) {
- dev_err(dev, "BAR_0 is reserved\n");
- return -EINVAL;
- }
- ks_dw_app_writel(ks_pcie, IB_BAR(index), bar);
- ks_dw_app_writel(ks_pcie, IB_OFFSET(index), lower_32_bits(cpu_addr));
- return 0;
- }
- static int ks_dw_pcie_check_free(struct keystone_pcie *ks_pcie, u8 index,
- u8 regions)
- {
- while (regions--) {
- if (test_bit(index++, &ks_pcie->ob_window_map))
- return false;
- }
- return true;
- }
- int ks_dw_pcie_outbound_atu(struct dw_pcie *pci, u64 cpu_addr, u64 pci_addr,
- size_t size)
- {
- u8 index;
- u8 regions;
- struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
- index = (cpu_addr >> WIN_INDEX_SHIFT) & WIN_INDEX_MASK;
- regions = ((size - 1) >> WIN_INDEX_SHIFT) + 1;
- if (!ks_dw_pcie_check_free(ks_pcie, index, regions))
- return -ENOMEM;
- if (index + regions > OUTBOUND_WINDOWS)
- return -ENOMEM;
- ks_pcie->ob_win[index].cpu_addr = cpu_addr;
- ks_pcie->ob_win[index].no_of_regions = regions;
- while (regions--) {
- ks_dw_app_writel(ks_pcie, OB_OFFSET_INDEX(index),
- lower_32_bits(pci_addr) | OB_ENABLEN);
- ks_dw_app_writel(ks_pcie, OB_OFFSET_HI(index),
- upper_32_bits(pci_addr));
- set_bit(index++, &ks_pcie->ob_window_map);
- pci_addr += WIN_SIZE;
- }
- return 0;
- }
- void ks_dw_pcie_raise_legacy_irq(struct keystone_pcie *ks_pcie)
- {
- struct dw_pcie *pci = ks_pcie->pci;
- u8 int_pin;
- int_pin = dw_pcie_readb_dbi(pci, PCI_INTERRUPT_PIN);
- if (int_pin == 0 || int_pin > 4)
- return;
- ks_dw_app_writel(ks_pcie, PCIE_LEGACY_IRQ_ENABLE_SET(int_pin),
- INT_ENABLE);
- ks_dw_app_writel(ks_pcie, PCIE_EP_IRQ_SET, INT_ENABLE);
- mdelay(1);
- ks_dw_app_writel(ks_pcie, PCIE_EP_IRQ_CLR, INT_ENABLE);
- ks_dw_app_writel(ks_pcie, PCIE_LEGACY_IRQ_ENABLE_CLR(int_pin),
- INT_ENABLE);
- }
- u32 ks_dw_pcie_read_dbi2(struct dw_pcie *pci, void __iomem *base,
- u32 reg, size_t size)
- {
- struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
- u32 val;
- ks_dw_pcie_set_dbi_mode(ks_pcie);
- dw_pcie_read(base + reg, size, &val);
- ks_dw_pcie_clear_dbi_mode(ks_pcie);
- return val;
- }
- void ks_dw_pcie_write_dbi2(struct dw_pcie *pci, void __iomem *base,
- u32 reg, size_t size, u32 val)
- {
- struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
- ks_dw_pcie_set_dbi_mode(ks_pcie);
- dw_pcie_write(base + reg, size, val);
- ks_dw_pcie_clear_dbi_mode(ks_pcie);
- }
- static inline void update_reg_offset_bit_pos(u32 offset, u32 *reg_offset,
- u32 *bit_pos)
- {
- *reg_offset = offset % 8;
- *bit_pos = offset >> 3;
- }
- phys_addr_t ks_dw_pcie_get_msi_addr(struct pcie_port *pp)
- {
- struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
- struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
- return ks_pcie->app.start + MSI_IRQ;
- }
- static u32 ks_dw_app_readl(struct keystone_pcie *ks_pcie, u32 offset)
- {
- return readl(ks_pcie->va_app_base + offset);
- }
- static void ks_dw_app_writel(struct keystone_pcie *ks_pcie, u32 offset, u32 val)
- {
- writel(val, ks_pcie->va_app_base + offset);
- }
- void ks_dw_pcie_handle_msi_irq(struct keystone_pcie *ks_pcie, int offset)
- {
- struct dw_pcie *pci = ks_pcie->pci;
- struct pcie_port *pp = &pci->pp;
- struct device *dev = pci->dev;
- u32 pending, vector;
- int src, virq;
- pending = ks_dw_app_readl(ks_pcie, MSI0_IRQ_STATUS + (offset << 4));
- /*
- * MSI0 status bit 0-3 shows vectors 0, 8, 16, 24, MSI1 status bit
- * shows 1, 9, 17, 25 and so forth
- */
- for (src = 0; src < 4; src++) {
- if (BIT(src) & pending) {
- vector = offset + (src << 3);
- virq = irq_linear_revmap(pp->irq_domain, vector);
- dev_dbg(dev, "irq: bit %d, vector %d, virq %d\n",
- src, vector, virq);
- generic_handle_irq(virq);
- }
- }
- }
- static void ks_dw_pcie_msi_irq_ack(struct irq_data *d)
- {
- u32 offset, reg_offset, bit_pos;
- struct keystone_pcie *ks_pcie;
- struct msi_desc *msi;
- struct pcie_port *pp;
- struct dw_pcie *pci;
- msi = irq_data_get_msi_desc(d);
- pp = (struct pcie_port *) msi_desc_to_pci_sysdata(msi);
- pci = to_dw_pcie_from_pp(pp);
- ks_pcie = to_keystone_pcie(pci);
- offset = d->irq - irq_linear_revmap(pp->irq_domain, 0);
- update_reg_offset_bit_pos(offset, ®_offset, &bit_pos);
- ks_dw_app_writel(ks_pcie, MSI0_IRQ_STATUS + (reg_offset << 4),
- BIT(bit_pos));
- ks_dw_app_writel(ks_pcie, IRQ_EOI, reg_offset + MSI_IRQ_OFFSET);
- }
- void ks_dw_pcie_msi_set_irq(struct pcie_port *pp, int irq)
- {
- u32 reg_offset, bit_pos;
- struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
- struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
- update_reg_offset_bit_pos(irq, ®_offset, &bit_pos);
- ks_dw_app_writel(ks_pcie, MSI0_IRQ_ENABLE_SET + (reg_offset << 4),
- BIT(bit_pos));
- }
- void ks_dw_pcie_msi_clear_irq(struct pcie_port *pp, int irq)
- {
- u32 reg_offset, bit_pos;
- struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
- struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
- update_reg_offset_bit_pos(irq, ®_offset, &bit_pos);
- ks_dw_app_writel(ks_pcie, MSI0_IRQ_ENABLE_CLR + (reg_offset << 4),
- BIT(bit_pos));
- }
- static void ks_dw_pcie_msi_irq_mask(struct irq_data *d)
- {
- struct keystone_pcie *ks_pcie;
- struct msi_desc *msi;
- struct pcie_port *pp;
- struct dw_pcie *pci;
- u32 offset;
- msi = irq_data_get_msi_desc(d);
- pp = (struct pcie_port *) msi_desc_to_pci_sysdata(msi);
- pci = to_dw_pcie_from_pp(pp);
- ks_pcie = to_keystone_pcie(pci);
- offset = d->irq - irq_linear_revmap(pp->irq_domain, 0);
- /* Mask the end point if PVM implemented */
- if (IS_ENABLED(CONFIG_PCI_MSI)) {
- if (msi->msi_attrib.maskbit)
- pci_msi_mask_irq(d);
- }
- ks_dw_pcie_msi_clear_irq(pp, offset);
- }
- static void ks_dw_pcie_msi_irq_unmask(struct irq_data *d)
- {
- struct keystone_pcie *ks_pcie;
- struct msi_desc *msi;
- struct pcie_port *pp;
- struct dw_pcie *pci;
- u32 offset;
- msi = irq_data_get_msi_desc(d);
- pp = (struct pcie_port *) msi_desc_to_pci_sysdata(msi);
- pci = to_dw_pcie_from_pp(pp);
- ks_pcie = to_keystone_pcie(pci);
- offset = d->irq - irq_linear_revmap(pp->irq_domain, 0);
- /* Mask the end point if PVM implemented */
- if (IS_ENABLED(CONFIG_PCI_MSI)) {
- if (msi->msi_attrib.maskbit)
- pci_msi_unmask_irq(d);
- }
- ks_dw_pcie_msi_set_irq(pp, offset);
- }
- static struct irq_chip ks_dw_pcie_msi_irq_chip = {
- .name = "Keystone-PCIe-MSI-IRQ",
- .irq_ack = ks_dw_pcie_msi_irq_ack,
- .irq_mask = ks_dw_pcie_msi_irq_mask,
- .irq_unmask = ks_dw_pcie_msi_irq_unmask,
- };
- static int ks_dw_pcie_msi_map(struct irq_domain *domain, unsigned int irq,
- irq_hw_number_t hwirq)
- {
- irq_set_chip_and_handler(irq, &ks_dw_pcie_msi_irq_chip,
- handle_level_irq);
- irq_set_chip_data(irq, domain->host_data);
- return 0;
- }
- static const struct irq_domain_ops ks_dw_pcie_msi_domain_ops = {
- .map = ks_dw_pcie_msi_map,
- };
- int ks_dw_pcie_msi_host_init(struct pcie_port *pp, struct msi_controller *chip)
- {
- struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
- struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
- struct device *dev = pci->dev;
- int i;
- pp->irq_domain = irq_domain_add_linear(ks_pcie->msi_intc_np,
- MAX_MSI_IRQS,
- &ks_dw_pcie_msi_domain_ops,
- chip);
- if (!pp->irq_domain) {
- dev_err(dev, "irq domain init failed\n");
- return -ENXIO;
- }
- for (i = 0; i < MAX_MSI_IRQS; i++)
- irq_create_mapping(pp->irq_domain, i);
- return 0;
- }
- void ks_dw_pcie_enable_legacy_irqs(struct keystone_pcie *ks_pcie)
- {
- int i;
- for (i = 0; i < MAX_LEGACY_IRQS; i++)
- ks_dw_app_writel(ks_pcie, IRQ_ENABLE_SET + (i << 4), 0x1);
- }
- void ks_dw_pcie_handle_legacy_irq(struct keystone_pcie *ks_pcie, int offset)
- {
- struct dw_pcie *pci = ks_pcie->pci;
- struct device *dev = pci->dev;
- u32 pending;
- int virq;
- pending = ks_dw_app_readl(ks_pcie, IRQ_STATUS + (offset << 4));
- if (BIT(0) & pending) {
- virq = irq_linear_revmap(ks_pcie->legacy_irq_domain, offset);
- dev_dbg(dev, ": irq: irq_offset %d, virq %d\n", offset, virq);
- generic_handle_irq(virq);
- }
- /* EOI the INTx interrupt */
- ks_dw_app_writel(ks_pcie, IRQ_EOI, offset);
- }
- void ks_dw_pcie_enable_error_irq(struct keystone_pcie *ks_pcie)
- {
- ks_dw_app_writel(ks_pcie, ERR_IRQ_ENABLE_SET, ERR_IRQ_ALL);
- }
- irqreturn_t ks_dw_pcie_handle_error_irq(struct keystone_pcie *ks_pcie)
- {
- u32 status;
- status = ks_dw_app_readl(ks_pcie, ERR_IRQ_STATUS_RAW) & ERR_IRQ_ALL;
- if (!status)
- return IRQ_NONE;
- if (status & ERR_FATAL_IRQ)
- dev_err(ks_pcie->pci->dev, "fatal error (status %#010x)\n",
- status);
- /* Ack the IRQ; status bits are RW1C */
- ks_dw_app_writel(ks_pcie, ERR_IRQ_STATUS, status);
- return IRQ_HANDLED;
- }
- static void ks_dw_pcie_ack_legacy_irq(struct irq_data *d)
- {
- }
- static void ks_dw_pcie_mask_legacy_irq(struct irq_data *d)
- {
- }
- static void ks_dw_pcie_unmask_legacy_irq(struct irq_data *d)
- {
- }
- static struct irq_chip ks_dw_pcie_legacy_irq_chip = {
- .name = "Keystone-PCI-Legacy-IRQ",
- .irq_ack = ks_dw_pcie_ack_legacy_irq,
- .irq_mask = ks_dw_pcie_mask_legacy_irq,
- .irq_unmask = ks_dw_pcie_unmask_legacy_irq,
- };
- static int ks_dw_pcie_init_legacy_irq_map(struct irq_domain *d,
- unsigned int irq, irq_hw_number_t hw_irq)
- {
- irq_set_chip_and_handler(irq, &ks_dw_pcie_legacy_irq_chip,
- handle_level_irq);
- irq_set_chip_data(irq, d->host_data);
- return 0;
- }
- static const struct irq_domain_ops ks_dw_pcie_legacy_irq_domain_ops = {
- .map = ks_dw_pcie_init_legacy_irq_map,
- .xlate = irq_domain_xlate_onetwocell,
- };
- /**
- * ks_dw_pcie_set_dbi_mode() - Set DBI mode to access overlaid BAR mask
- * registers
- *
- * Since modification of dbi_cs2 involves different clock domain, read the
- * status back to ensure the transition is complete.
- */
- static void ks_dw_pcie_set_dbi_mode(struct keystone_pcie *ks_pcie)
- {
- u32 val;
- val = ks_dw_app_readl(ks_pcie, CMD_STATUS);
- ks_dw_app_writel(ks_pcie, CMD_STATUS, DBI_CS2_EN_VAL | val);
- do {
- val = ks_dw_app_readl(ks_pcie, CMD_STATUS);
- } while (!(val & DBI_CS2_EN_VAL));
- }
- /**
- * ks_dw_pcie_clear_dbi_mode() - Disable DBI mode
- *
- * Since modification of dbi_cs2 involves different clock domain, read the
- * status back to ensure the transition is complete.
- */
- static void ks_dw_pcie_clear_dbi_mode(struct keystone_pcie *ks_pcie)
- {
- u32 val;
- val = ks_dw_app_readl(ks_pcie, CMD_STATUS);
- ks_dw_app_writel(ks_pcie, CMD_STATUS, ~DBI_CS2_EN_VAL & val);
- do {
- val = ks_dw_app_readl(ks_pcie, CMD_STATUS);
- } while (val & DBI_CS2_EN_VAL);
- }
- void ks_dw_pcie_setup_rc_app_regs(struct keystone_pcie *ks_pcie)
- {
- struct dw_pcie *pci = ks_pcie->pci;
- struct pcie_port *pp = &pci->pp;
- u32 start = pp->mem->start, end = pp->mem->end;
- int i, tr_size;
- u32 val;
- /* Disable BARs for inbound access */
- ks_dw_pcie_set_dbi_mode(ks_pcie);
- dw_pcie_writel_dbi(pci, PCI_BASE_ADDRESS_0, 0);
- dw_pcie_writel_dbi(pci, PCI_BASE_ADDRESS_1, 0);
- ks_dw_pcie_clear_dbi_mode(ks_pcie);
- /* Set outbound translation size per window division */
- ks_dw_app_writel(ks_pcie, OB_SIZE, CFG_PCIM_WIN_SZ_IDX & 0x7);
- tr_size = (1 << (CFG_PCIM_WIN_SZ_IDX & 0x7)) * SZ_1M;
- /* Using Direct 1:1 mapping of RC <-> PCI memory space */
- for (i = 0; (i < CFG_PCIM_WIN_CNT) && (start < end); i++) {
- ks_dw_app_writel(ks_pcie, OB_OFFSET_INDEX(i), start | 1);
- ks_dw_app_writel(ks_pcie, OB_OFFSET_HI(i), 0);
- start += tr_size;
- }
- /* Enable OB translation */
- val = ks_dw_app_readl(ks_pcie, CMD_STATUS);
- ks_dw_app_writel(ks_pcie, CMD_STATUS, OB_XLAT_EN_VAL | val);
- }
- /**
- * ks_pcie_cfg_setup() - Set up configuration space address for a device
- *
- * @ks_pcie: ptr to keystone_pcie structure
- * @bus: Bus number the device is residing on
- * @devfn: device, function number info
- *
- * Forms and returns the address of configuration space mapped in PCIESS
- * address space 0. Also configures CFG_SETUP for remote configuration space
- * access.
- *
- * The address space has two regions to access configuration - local and remote.
- * We access local region for bus 0 (as RC is attached on bus 0) and remote
- * region for others with TYPE 1 access when bus > 1. As for device on bus = 1,
- * we will do TYPE 0 access as it will be on our secondary bus (logical).
- * CFG_SETUP is needed only for remote configuration access.
- */
- static void __iomem *ks_pcie_cfg_setup(struct keystone_pcie *ks_pcie, u8 bus,
- unsigned int devfn)
- {
- u8 device = PCI_SLOT(devfn), function = PCI_FUNC(devfn);
- struct dw_pcie *pci = ks_pcie->pci;
- struct pcie_port *pp = &pci->pp;
- u32 regval;
- if (bus == 0)
- return pci->dbi_base;
- regval = (bus << 16) | (device << 8) | function;
- /*
- * Since Bus#1 will be a virtual bus, we need to have TYPE0
- * access only.
- * TYPE 1
- */
- if (bus != 1)
- regval |= BIT(24);
- ks_dw_app_writel(ks_pcie, CFG_SETUP, regval);
- return pp->va_cfg0_base;
- }
- int ks_dw_pcie_rd_other_conf(struct pcie_port *pp, struct pci_bus *bus,
- unsigned int devfn, int where, int size, u32 *val)
- {
- struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
- struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
- u8 bus_num = bus->number;
- void __iomem *addr;
- addr = ks_pcie_cfg_setup(ks_pcie, bus_num, devfn);
- return dw_pcie_read(addr + where, size, val);
- }
- int ks_dw_pcie_wr_other_conf(struct pcie_port *pp, struct pci_bus *bus,
- unsigned int devfn, int where, int size, u32 val)
- {
- struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
- struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
- u8 bus_num = bus->number;
- void __iomem *addr;
- addr = ks_pcie_cfg_setup(ks_pcie, bus_num, devfn);
- return dw_pcie_write(addr + where, size, val);
- }
- /**
- * ks_dw_pcie_v3_65_scan_bus() - keystone scan_bus post initialization
- *
- * This sets BAR0 to enable inbound access for MSI_IRQ register
- */
- void ks_dw_pcie_v3_65_scan_bus(struct pcie_port *pp)
- {
- struct dw_pcie *pci = to_dw_pcie_from_pp(pp);
- struct keystone_pcie *ks_pcie = to_keystone_pcie(pci);
- /* Configure and set up BAR0 */
- ks_dw_pcie_set_dbi_mode(ks_pcie);
- /* Enable BAR0 */
- dw_pcie_writel_dbi(pci, PCI_BASE_ADDRESS_0, 1);
- dw_pcie_writel_dbi(pci, PCI_BASE_ADDRESS_0, SZ_4K - 1);
- ks_dw_pcie_clear_dbi_mode(ks_pcie);
- /*
- * For BAR0, just setting bus address for inbound writes (MSI) should
- * be sufficient. Use physical address to avoid any conflicts.
- */
- dw_pcie_writel_dbi(pci, PCI_BASE_ADDRESS_0, ks_pcie->app.start);
- }
- /**
- * ks_dw_pcie_link_up() - Check if link up
- */
- int ks_dw_pcie_link_up(struct dw_pcie *pci)
- {
- u32 val;
- val = dw_pcie_readl_dbi(pci, DEBUG0);
- return (val & LTSSM_STATE_MASK) == LTSSM_STATE_L0;
- }
- void ks_dw_pcie_initiate_link_train(struct keystone_pcie *ks_pcie)
- {
- u32 val;
- /* Disable Link training */
- val = ks_dw_app_readl(ks_pcie, CMD_STATUS);
- val &= ~LTSSM_EN_VAL;
- ks_dw_app_writel(ks_pcie, CMD_STATUS, LTSSM_EN_VAL | val);
- /* Initiate Link Training */
- val = ks_dw_app_readl(ks_pcie, CMD_STATUS);
- ks_dw_app_writel(ks_pcie, CMD_STATUS, LTSSM_EN_VAL | val);
- }
- /**
- * ks_dw_pcie_host_init() - initialize host for v3_65 dw hardware
- *
- * Ioremap the register resources, initialize legacy irq domain
- * and call dw_pcie_v3_65_host_init() API to initialize the Keystone
- * PCI host controller.
- */
- int __init ks_dw_pcie_host_init(struct keystone_pcie *ks_pcie,
- struct device_node *msi_intc_np)
- {
- struct dw_pcie *pci = ks_pcie->pci;
- struct pcie_port *pp = &pci->pp;
- struct device *dev = pci->dev;
- struct platform_device *pdev = to_platform_device(dev);
- struct resource *res;
- /* Index 0 is the config reg. space address */
- res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
- pci->dbi_base = devm_ioremap_resource(dev, res);
- if (IS_ERR(pci->dbi_base))
- return PTR_ERR(pci->dbi_base);
- /*
- * We set these same and is used in pcie rd/wr_other_conf
- * functions
- */
- pp->va_cfg0_base = pci->dbi_base + SPACE0_REMOTE_CFG_OFFSET;
- pp->va_cfg1_base = pp->va_cfg0_base;
- /* Create legacy IRQ domain */
- ks_pcie->legacy_irq_domain =
- irq_domain_add_linear(ks_pcie->legacy_intc_np,
- MAX_LEGACY_IRQS,
- &ks_dw_pcie_legacy_irq_domain_ops,
- NULL);
- if (!ks_pcie->legacy_irq_domain) {
- dev_err(dev, "Failed to add irq domain for legacy irqs\n");
- return -EINVAL;
- }
- return dw_pcie_host_init(pp);
- }
|