2019-05-27 14:55:01 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-or-later
|
2010-01-18 21:42:34 +08:00
|
|
|
/*
|
|
|
|
* Contains common pci routines for ALL ppc platform
|
|
|
|
* (based on pci_32.c and pci_64.c)
|
|
|
|
*
|
|
|
|
* Port for PPC64 David Engebretsen, IBM Corp.
|
|
|
|
* Contains common pci routines for ppc64 platform, pSeries and iSeries brands.
|
|
|
|
*
|
|
|
|
* Copyright (C) 2003 Anton Blanchard <anton@au.ibm.com>, IBM
|
|
|
|
* Rework, based on alpha PCI code.
|
|
|
|
*
|
|
|
|
* Common pmac/prep/chrp pci routines. -- Cort
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/pci.h>
|
|
|
|
#include <linux/string.h>
|
|
|
|
#include <linux/init.h>
|
2018-10-31 06:09:49 +08:00
|
|
|
#include <linux/memblock.h>
|
2010-01-18 21:42:34 +08:00
|
|
|
#include <linux/mm.h>
|
2017-02-25 06:59:36 +08:00
|
|
|
#include <linux/shmem_fs.h>
|
2010-01-18 21:42:34 +08:00
|
|
|
#include <linux/list.h>
|
|
|
|
#include <linux/syscalls.h>
|
|
|
|
#include <linux/irq.h>
|
|
|
|
#include <linux/vmalloc.h>
|
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 16:04:11 +08:00
|
|
|
#include <linux/slab.h>
|
2010-08-17 13:44:49 +08:00
|
|
|
#include <linux/of.h>
|
|
|
|
#include <linux/of_address.h>
|
2013-09-08 03:05:10 +08:00
|
|
|
#include <linux/of_irq.h>
|
2011-01-24 12:28:55 +08:00
|
|
|
#include <linux/of_pci.h>
|
2011-09-22 23:22:55 +08:00
|
|
|
#include <linux/export.h>
|
2010-01-18 21:42:34 +08:00
|
|
|
|
|
|
|
#include <asm/processor.h>
|
2012-12-27 17:40:38 +08:00
|
|
|
#include <linux/io.h>
|
2010-01-18 21:42:34 +08:00
|
|
|
#include <asm/pci-bridge.h>
|
|
|
|
#include <asm/byteorder.h>
|
|
|
|
|
|
|
|
static DEFINE_SPINLOCK(hose_spinlock);
|
|
|
|
LIST_HEAD(hose_list);
|
|
|
|
|
|
|
|
/* XXX kill that some day ... */
|
|
|
|
static int global_phb_number; /* Global phb counter */
|
|
|
|
|
|
|
|
/* ISA Memory physical address */
|
|
|
|
resource_size_t isa_mem_base;
|
|
|
|
|
2011-04-11 09:17:26 +08:00
|
|
|
unsigned long isa_io_base;
|
2016-03-17 10:36:28 +08:00
|
|
|
EXPORT_SYMBOL(isa_io_base);
|
|
|
|
|
2011-04-11 09:17:26 +08:00
|
|
|
static int pci_bus_count;
|
|
|
|
|
2010-01-18 21:42:34 +08:00
|
|
|
struct pci_controller *pcibios_alloc_controller(struct device_node *dev)
|
|
|
|
{
|
|
|
|
struct pci_controller *phb;
|
|
|
|
|
|
|
|
phb = zalloc_maybe_bootmem(sizeof(struct pci_controller), GFP_KERNEL);
|
|
|
|
if (!phb)
|
|
|
|
return NULL;
|
|
|
|
spin_lock(&hose_spinlock);
|
|
|
|
phb->global_number = global_phb_number++;
|
|
|
|
list_add_tail(&phb->list_node, &hose_list);
|
|
|
|
spin_unlock(&hose_spinlock);
|
|
|
|
phb->dn = dev;
|
|
|
|
phb->is_dynamic = mem_init_done;
|
|
|
|
return phb;
|
|
|
|
}
|
|
|
|
|
|
|
|
void pcibios_free_controller(struct pci_controller *phb)
|
|
|
|
{
|
|
|
|
spin_lock(&hose_spinlock);
|
|
|
|
list_del(&phb->list_node);
|
|
|
|
spin_unlock(&hose_spinlock);
|
|
|
|
|
|
|
|
if (phb->is_dynamic)
|
|
|
|
kfree(phb);
|
|
|
|
}
|
|
|
|
|
|
|
|
static resource_size_t pcibios_io_size(const struct pci_controller *hose)
|
|
|
|
{
|
2011-06-10 00:13:32 +08:00
|
|
|
return resource_size(&hose->io_resource);
|
2010-01-18 21:42:34 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
int pcibios_vaddr_is_ioport(void __iomem *address)
|
|
|
|
{
|
|
|
|
int ret = 0;
|
|
|
|
struct pci_controller *hose;
|
|
|
|
resource_size_t size;
|
|
|
|
|
|
|
|
spin_lock(&hose_spinlock);
|
|
|
|
list_for_each_entry(hose, &hose_list, list_node) {
|
|
|
|
size = pcibios_io_size(hose);
|
|
|
|
if (address >= hose->io_base_virt &&
|
|
|
|
address < (hose->io_base_virt + size)) {
|
|
|
|
ret = 1;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
spin_unlock(&hose_spinlock);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned long pci_address_to_pio(phys_addr_t address)
|
|
|
|
{
|
|
|
|
struct pci_controller *hose;
|
|
|
|
resource_size_t size;
|
|
|
|
unsigned long ret = ~0;
|
|
|
|
|
|
|
|
spin_lock(&hose_spinlock);
|
|
|
|
list_for_each_entry(hose, &hose_list, list_node) {
|
|
|
|
size = pcibios_io_size(hose);
|
|
|
|
if (address >= hose->io_base_phys &&
|
|
|
|
address < (hose->io_base_phys + size)) {
|
|
|
|
unsigned long base =
|
|
|
|
(unsigned long)hose->io_base_virt - _IO_BASE;
|
|
|
|
ret = base + (address - hose->io_base_phys);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
spin_unlock(&hose_spinlock);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(pci_address_to_pio);
|
|
|
|
|
|
|
|
/* This routine is meant to be used early during boot, when the
|
|
|
|
* PCI bus numbers have not yet been assigned, and you need to
|
|
|
|
* issue PCI config cycles to an OF device.
|
|
|
|
* It could also be used to "fix" RTAS config cycles if you want
|
|
|
|
* to set pci_assign_all_buses to 1 and still use RTAS for PCI
|
|
|
|
* config cycles.
|
|
|
|
*/
|
|
|
|
struct pci_controller *pci_find_hose_for_OF_device(struct device_node *node)
|
|
|
|
{
|
|
|
|
while (node) {
|
|
|
|
struct pci_controller *hose, *tmp;
|
|
|
|
list_for_each_entry_safe(hose, tmp, &hose_list, list_node)
|
|
|
|
if (hose->dn == node)
|
|
|
|
return hose;
|
|
|
|
node = node->parent;
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2011-10-29 05:47:56 +08:00
|
|
|
void pcibios_set_master(struct pci_dev *dev)
|
|
|
|
{
|
|
|
|
/* No special bus mastering setup handling */
|
|
|
|
}
|
|
|
|
|
2010-01-18 21:42:34 +08:00
|
|
|
/*
|
2018-02-19 21:01:23 +08:00
|
|
|
* Platform support for /proc/bus/pci/X/Y mmap()s.
|
2010-01-18 21:42:34 +08:00
|
|
|
*/
|
|
|
|
|
2018-02-19 21:01:23 +08:00
|
|
|
int pci_iobar_pfn(struct pci_dev *pdev, int bar, struct vm_area_struct *vma)
|
2010-01-18 21:42:34 +08:00
|
|
|
{
|
2018-02-19 21:01:23 +08:00
|
|
|
struct pci_controller *hose = pci_bus_to_host(pdev->bus);
|
|
|
|
resource_size_t ioaddr = pci_resource_start(pdev, bar);
|
2010-01-18 21:42:34 +08:00
|
|
|
|
2013-01-04 16:14:46 +08:00
|
|
|
if (!hose)
|
2018-02-19 21:01:23 +08:00
|
|
|
return -EINVAL; /* should never happen */
|
2010-01-18 21:42:34 +08:00
|
|
|
|
2018-02-19 21:01:23 +08:00
|
|
|
/* Convert to an offset within this PCI controller */
|
|
|
|
ioaddr -= (unsigned long)hose->io_base_virt - _IO_BASE;
|
2010-01-18 21:42:34 +08:00
|
|
|
|
2018-02-19 21:01:23 +08:00
|
|
|
vma->vm_pgoff += (ioaddr + hose->io_base_phys) >> PAGE_SHIFT;
|
|
|
|
return 0;
|
2010-01-18 21:42:34 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This one is used by /dev/mem and fbdev who have no clue about the
|
|
|
|
* PCI device, it tries to find the PCI device first and calls the
|
|
|
|
* above routine
|
|
|
|
*/
|
|
|
|
pgprot_t pci_phys_mem_access_prot(struct file *file,
|
|
|
|
unsigned long pfn,
|
|
|
|
unsigned long size,
|
|
|
|
pgprot_t prot)
|
|
|
|
{
|
|
|
|
struct pci_dev *pdev = NULL;
|
|
|
|
struct resource *found = NULL;
|
|
|
|
resource_size_t offset = ((resource_size_t)pfn) << PAGE_SHIFT;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (page_is_ram(pfn))
|
|
|
|
return prot;
|
|
|
|
|
|
|
|
prot = pgprot_noncached(prot);
|
|
|
|
for_each_pci_dev(pdev) {
|
|
|
|
for (i = 0; i <= PCI_ROM_RESOURCE; i++) {
|
|
|
|
struct resource *rp = &pdev->resource[i];
|
|
|
|
int flags = rp->flags;
|
|
|
|
|
|
|
|
/* Active and same type? */
|
|
|
|
if ((flags & IORESOURCE_MEM) == 0)
|
|
|
|
continue;
|
|
|
|
/* In the range of this resource? */
|
|
|
|
if (offset < (rp->start & PAGE_MASK) ||
|
|
|
|
offset > rp->end)
|
|
|
|
continue;
|
|
|
|
found = rp;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (found)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (found) {
|
|
|
|
if (found->flags & IORESOURCE_PREFETCH)
|
|
|
|
prot = pgprot_noncached_wc(prot);
|
|
|
|
pci_dev_put(pdev);
|
|
|
|
}
|
|
|
|
|
|
|
|
pr_debug("PCI: Non-PCI map for %llx, prot: %lx\n",
|
|
|
|
(unsigned long long)offset, pgprot_val(prot));
|
|
|
|
|
|
|
|
return prot;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* This provides legacy IO read access on a bus */
|
|
|
|
int pci_legacy_read(struct pci_bus *bus, loff_t port, u32 *val, size_t size)
|
|
|
|
{
|
|
|
|
unsigned long offset;
|
|
|
|
struct pci_controller *hose = pci_bus_to_host(bus);
|
|
|
|
struct resource *rp = &hose->io_resource;
|
|
|
|
void __iomem *addr;
|
|
|
|
|
|
|
|
/* Check if port can be supported by that bus. We only check
|
|
|
|
* the ranges of the PHB though, not the bus itself as the rules
|
|
|
|
* for forwarding legacy cycles down bridges are not our problem
|
|
|
|
* here. So if the host bridge supports it, we do it.
|
|
|
|
*/
|
|
|
|
offset = (unsigned long)hose->io_base_virt - _IO_BASE;
|
|
|
|
offset += port;
|
|
|
|
|
|
|
|
if (!(rp->flags & IORESOURCE_IO))
|
|
|
|
return -ENXIO;
|
|
|
|
if (offset < rp->start || (offset + size) > rp->end)
|
|
|
|
return -ENXIO;
|
|
|
|
addr = hose->io_base_virt + port;
|
|
|
|
|
|
|
|
switch (size) {
|
|
|
|
case 1:
|
|
|
|
*((u8 *)val) = in_8(addr);
|
|
|
|
return 1;
|
|
|
|
case 2:
|
|
|
|
if (port & 1)
|
|
|
|
return -EINVAL;
|
|
|
|
*((u16 *)val) = in_le16(addr);
|
|
|
|
return 2;
|
|
|
|
case 4:
|
|
|
|
if (port & 3)
|
|
|
|
return -EINVAL;
|
|
|
|
*((u32 *)val) = in_le32(addr);
|
|
|
|
return 4;
|
|
|
|
}
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* This provides legacy IO write access on a bus */
|
|
|
|
int pci_legacy_write(struct pci_bus *bus, loff_t port, u32 val, size_t size)
|
|
|
|
{
|
|
|
|
unsigned long offset;
|
|
|
|
struct pci_controller *hose = pci_bus_to_host(bus);
|
|
|
|
struct resource *rp = &hose->io_resource;
|
|
|
|
void __iomem *addr;
|
|
|
|
|
|
|
|
/* Check if port can be supported by that bus. We only check
|
|
|
|
* the ranges of the PHB though, not the bus itself as the rules
|
|
|
|
* for forwarding legacy cycles down bridges are not our problem
|
|
|
|
* here. So if the host bridge supports it, we do it.
|
|
|
|
*/
|
|
|
|
offset = (unsigned long)hose->io_base_virt - _IO_BASE;
|
|
|
|
offset += port;
|
|
|
|
|
|
|
|
if (!(rp->flags & IORESOURCE_IO))
|
|
|
|
return -ENXIO;
|
|
|
|
if (offset < rp->start || (offset + size) > rp->end)
|
|
|
|
return -ENXIO;
|
|
|
|
addr = hose->io_base_virt + port;
|
|
|
|
|
|
|
|
/* WARNING: The generic code is idiotic. It gets passed a pointer
|
|
|
|
* to what can be a 1, 2 or 4 byte quantity and always reads that
|
|
|
|
* as a u32, which means that we have to correct the location of
|
|
|
|
* the data read within those 32 bits for size 1 and 2
|
|
|
|
*/
|
|
|
|
switch (size) {
|
|
|
|
case 1:
|
|
|
|
out_8(addr, val >> 24);
|
|
|
|
return 1;
|
|
|
|
case 2:
|
|
|
|
if (port & 1)
|
|
|
|
return -EINVAL;
|
|
|
|
out_le16(addr, val >> 16);
|
|
|
|
return 2;
|
|
|
|
case 4:
|
|
|
|
if (port & 3)
|
|
|
|
return -EINVAL;
|
|
|
|
out_le32(addr, val);
|
|
|
|
return 4;
|
|
|
|
}
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* This provides legacy IO or memory mmap access on a bus */
|
|
|
|
int pci_mmap_legacy_page_range(struct pci_bus *bus,
|
|
|
|
struct vm_area_struct *vma,
|
|
|
|
enum pci_mmap_state mmap_state)
|
|
|
|
{
|
|
|
|
struct pci_controller *hose = pci_bus_to_host(bus);
|
|
|
|
resource_size_t offset =
|
|
|
|
((resource_size_t)vma->vm_pgoff) << PAGE_SHIFT;
|
|
|
|
resource_size_t size = vma->vm_end - vma->vm_start;
|
|
|
|
struct resource *rp;
|
|
|
|
|
|
|
|
pr_debug("pci_mmap_legacy_page_range(%04x:%02x, %s @%llx..%llx)\n",
|
|
|
|
pci_domain_nr(bus), bus->number,
|
|
|
|
mmap_state == pci_mmap_mem ? "MEM" : "IO",
|
|
|
|
(unsigned long long)offset,
|
|
|
|
(unsigned long long)(offset + size - 1));
|
|
|
|
|
|
|
|
if (mmap_state == pci_mmap_mem) {
|
|
|
|
/* Hack alert !
|
|
|
|
*
|
|
|
|
* Because X is lame and can fail starting if it gets an error
|
|
|
|
* trying to mmap legacy_mem (instead of just moving on without
|
|
|
|
* legacy memory access) we fake it here by giving it anonymous
|
|
|
|
* memory, effectively behaving just like /dev/zero
|
|
|
|
*/
|
|
|
|
if ((offset + size) > hose->isa_mem_size) {
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_debug("Process %s (pid:%d) mapped non-existing PCI",
|
|
|
|
current->comm, current->pid);
|
|
|
|
pr_debug("legacy memory for 0%04x:%02x\n",
|
|
|
|
pci_domain_nr(bus), bus->number);
|
2010-01-18 21:42:34 +08:00
|
|
|
if (vma->vm_flags & VM_SHARED)
|
|
|
|
return shmem_zero_setup(vma);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
offset += hose->isa_mem_phys;
|
|
|
|
} else {
|
2012-12-27 17:40:38 +08:00
|
|
|
unsigned long io_offset = (unsigned long)hose->io_base_virt -
|
2010-01-18 21:42:34 +08:00
|
|
|
_IO_BASE;
|
|
|
|
unsigned long roffset = offset + io_offset;
|
|
|
|
rp = &hose->io_resource;
|
|
|
|
if (!(rp->flags & IORESOURCE_IO))
|
|
|
|
return -ENXIO;
|
|
|
|
if (roffset < rp->start || (roffset + size) > rp->end)
|
|
|
|
return -ENXIO;
|
|
|
|
offset += hose->io_base_phys;
|
|
|
|
}
|
|
|
|
pr_debug(" -> mapping phys %llx\n", (unsigned long long)offset);
|
|
|
|
|
|
|
|
vma->vm_pgoff = offset >> PAGE_SHIFT;
|
|
|
|
vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
|
|
|
|
return remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff,
|
|
|
|
vma->vm_end - vma->vm_start,
|
|
|
|
vma->vm_page_prot);
|
|
|
|
}
|
|
|
|
|
|
|
|
void pci_resource_to_user(const struct pci_dev *dev, int bar,
|
|
|
|
const struct resource *rsrc,
|
|
|
|
resource_size_t *start, resource_size_t *end)
|
|
|
|
{
|
2016-06-18 03:43:34 +08:00
|
|
|
struct pci_bus_region region;
|
2010-01-18 21:42:34 +08:00
|
|
|
|
2016-06-18 03:43:34 +08:00
|
|
|
if (rsrc->flags & IORESOURCE_IO) {
|
|
|
|
pcibios_resource_to_bus(dev->bus, ®ion,
|
|
|
|
(struct resource *) rsrc);
|
|
|
|
*start = region.start;
|
|
|
|
*end = region.end;
|
2010-01-18 21:42:34 +08:00
|
|
|
return;
|
2016-06-18 03:43:34 +08:00
|
|
|
}
|
2010-01-18 21:42:34 +08:00
|
|
|
|
2016-06-18 03:43:34 +08:00
|
|
|
/* We pass a CPU physical address to userland for MMIO instead of a
|
|
|
|
* BAR value because X is lame and expects to be able to use that
|
|
|
|
* to pass to /dev/mem!
|
2010-01-18 21:42:34 +08:00
|
|
|
*
|
2016-06-18 03:43:34 +08:00
|
|
|
* That means we may have 64-bit values where some apps only expect
|
|
|
|
* 32 (like X itself since it thinks only Sparc has 64-bit MMIO).
|
2010-01-18 21:42:34 +08:00
|
|
|
*/
|
2016-06-18 03:43:34 +08:00
|
|
|
*start = rsrc->start;
|
|
|
|
*end = rsrc->end;
|
2010-01-18 21:42:34 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* pci_process_bridge_OF_ranges - Parse PCI bridge resources from device tree
|
|
|
|
* @hose: newly allocated pci_controller to be setup
|
|
|
|
* @dev: device node of the host bridge
|
|
|
|
* @primary: set if primary bus (32 bits only, soon to be deprecated)
|
|
|
|
*
|
|
|
|
* This function will parse the "ranges" property of a PCI host bridge device
|
|
|
|
* node and setup the resource mapping of a pci controller based on its
|
|
|
|
* content.
|
|
|
|
*
|
|
|
|
* Life would be boring if it wasn't for a few issues that we have to deal
|
|
|
|
* with here:
|
|
|
|
*
|
|
|
|
* - We can only cope with one IO space range and up to 3 Memory space
|
|
|
|
* ranges. However, some machines (thanks Apple !) tend to split their
|
|
|
|
* space into lots of small contiguous ranges. So we have to coalesce.
|
|
|
|
*
|
|
|
|
* - We can only cope with all memory ranges having the same offset
|
|
|
|
* between CPU addresses and PCI addresses. Unfortunately, some bridges
|
|
|
|
* are setup for a large 1:1 mapping along with a small "window" which
|
|
|
|
* maps PCI address 0 to some arbitrary high address of the CPU space in
|
|
|
|
* order to give access to the ISA memory hole.
|
|
|
|
* The way out of here that I've chosen for now is to always set the
|
|
|
|
* offset based on the first resource found, then override it if we
|
|
|
|
* have a different offset and the previous was set by an ISA hole.
|
|
|
|
*
|
|
|
|
* - Some busses have IO space not starting at 0, which causes trouble with
|
|
|
|
* the way we do our IO resource renumbering. The code somewhat deals with
|
|
|
|
* it for 64 bits but I would expect problems on 32 bits.
|
|
|
|
*
|
|
|
|
* - Some 32 bits platforms such as 4xx can have physical space larger than
|
|
|
|
* 32 bits so we need to use 64 bits values for the parsing
|
|
|
|
*/
|
2012-12-22 06:06:37 +08:00
|
|
|
void pci_process_bridge_OF_ranges(struct pci_controller *hose,
|
|
|
|
struct device_node *dev, int primary)
|
2010-01-18 21:42:34 +08:00
|
|
|
{
|
|
|
|
int memno = 0, isa_hole = -1;
|
|
|
|
unsigned long long isa_mb = 0;
|
|
|
|
struct resource *res;
|
2013-07-28 03:01:22 +08:00
|
|
|
struct of_pci_range range;
|
|
|
|
struct of_pci_range_parser parser;
|
2010-01-18 21:42:34 +08:00
|
|
|
|
2017-06-07 04:57:36 +08:00
|
|
|
pr_info("PCI host bridge %pOF %s ranges:\n",
|
|
|
|
dev, primary ? "(primary)" : "");
|
2010-01-18 21:42:34 +08:00
|
|
|
|
2013-07-28 03:01:22 +08:00
|
|
|
/* Check for ranges property */
|
|
|
|
if (of_pci_range_parser_init(&parser, dev))
|
2010-01-18 21:42:34 +08:00
|
|
|
return;
|
|
|
|
|
|
|
|
pr_debug("Parsing ranges property...\n");
|
2013-07-28 03:01:22 +08:00
|
|
|
for_each_of_pci_range(&parser, &range) {
|
2010-01-18 21:42:34 +08:00
|
|
|
/* Read next ranges element */
|
|
|
|
|
|
|
|
/* If we failed translation or got a zero-sized region
|
|
|
|
* (some FW try to feed us with non sensical zero sized regions
|
|
|
|
* such as power3 which look like some kind of attempt
|
|
|
|
* at exposing the VGA memory hole)
|
|
|
|
*/
|
2013-07-28 03:01:22 +08:00
|
|
|
if (range.cpu_addr == OF_BAD_ADDR || range.size == 0)
|
2010-01-18 21:42:34 +08:00
|
|
|
continue;
|
|
|
|
|
|
|
|
/* Act based on address space type */
|
|
|
|
res = NULL;
|
2013-07-28 03:01:22 +08:00
|
|
|
switch (range.flags & IORESOURCE_TYPE_BITS) {
|
|
|
|
case IORESOURCE_IO:
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_info(" IO 0x%016llx..0x%016llx -> 0x%016llx\n",
|
2013-07-28 03:01:22 +08:00
|
|
|
range.cpu_addr, range.cpu_addr + range.size - 1,
|
|
|
|
range.pci_addr);
|
2010-01-18 21:42:34 +08:00
|
|
|
|
|
|
|
/* We support only one IO range */
|
|
|
|
if (hose->pci_io_size) {
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_info(" \\--> Skipped (too many) !\n");
|
2010-01-18 21:42:34 +08:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
/* On 32 bits, limit I/O space to 16MB */
|
2013-07-28 03:01:22 +08:00
|
|
|
if (range.size > 0x01000000)
|
|
|
|
range.size = 0x01000000;
|
2010-01-18 21:42:34 +08:00
|
|
|
|
|
|
|
/* 32 bits needs to map IOs here */
|
2013-07-28 03:01:22 +08:00
|
|
|
hose->io_base_virt = ioremap(range.cpu_addr,
|
|
|
|
range.size);
|
2010-01-18 21:42:34 +08:00
|
|
|
|
|
|
|
/* Expect trouble if pci_addr is not 0 */
|
|
|
|
if (primary)
|
|
|
|
isa_io_base =
|
|
|
|
(unsigned long)hose->io_base_virt;
|
|
|
|
/* pci_io_size and io_base_phys always represent IO
|
|
|
|
* space starting at 0 so we factor in pci_addr
|
|
|
|
*/
|
2013-07-28 03:01:22 +08:00
|
|
|
hose->pci_io_size = range.pci_addr + range.size;
|
|
|
|
hose->io_base_phys = range.cpu_addr - range.pci_addr;
|
2010-01-18 21:42:34 +08:00
|
|
|
|
|
|
|
/* Build resource */
|
|
|
|
res = &hose->io_resource;
|
2013-07-28 03:01:22 +08:00
|
|
|
range.cpu_addr = range.pci_addr;
|
|
|
|
|
2010-01-18 21:42:34 +08:00
|
|
|
break;
|
2013-07-28 03:01:22 +08:00
|
|
|
case IORESOURCE_MEM:
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_info(" MEM 0x%016llx..0x%016llx -> 0x%016llx %s\n",
|
2013-07-28 03:01:22 +08:00
|
|
|
range.cpu_addr, range.cpu_addr + range.size - 1,
|
|
|
|
range.pci_addr,
|
2020-02-14 07:18:59 +08:00
|
|
|
(range.flags & IORESOURCE_PREFETCH) ?
|
2013-07-28 03:01:22 +08:00
|
|
|
"Prefetch" : "");
|
2010-01-18 21:42:34 +08:00
|
|
|
|
|
|
|
/* We support only 3 memory ranges */
|
|
|
|
if (memno >= 3) {
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_info(" \\--> Skipped (too many) !\n");
|
2010-01-18 21:42:34 +08:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
/* Handles ISA memory hole space here */
|
2013-07-28 03:01:22 +08:00
|
|
|
if (range.pci_addr == 0) {
|
|
|
|
isa_mb = range.cpu_addr;
|
2010-01-18 21:42:34 +08:00
|
|
|
isa_hole = memno;
|
|
|
|
if (primary || isa_mem_base == 0)
|
2013-07-28 03:01:22 +08:00
|
|
|
isa_mem_base = range.cpu_addr;
|
|
|
|
hose->isa_mem_phys = range.cpu_addr;
|
|
|
|
hose->isa_mem_size = range.size;
|
2010-01-18 21:42:34 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/* We get the PCI/Mem offset from the first range or
|
|
|
|
* the, current one if the offset came from an ISA
|
|
|
|
* hole. If they don't match, bugger.
|
|
|
|
*/
|
|
|
|
if (memno == 0 ||
|
2013-07-28 03:01:22 +08:00
|
|
|
(isa_hole >= 0 && range.pci_addr != 0 &&
|
2010-01-18 21:42:34 +08:00
|
|
|
hose->pci_mem_offset == isa_mb))
|
2013-07-28 03:01:22 +08:00
|
|
|
hose->pci_mem_offset = range.cpu_addr -
|
|
|
|
range.pci_addr;
|
|
|
|
else if (range.pci_addr != 0 &&
|
|
|
|
hose->pci_mem_offset != range.cpu_addr -
|
|
|
|
range.pci_addr) {
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_info(" \\--> Skipped (offset mismatch) !\n");
|
2010-01-18 21:42:34 +08:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Build resource */
|
|
|
|
res = &hose->mem_resources[memno++];
|
|
|
|
break;
|
|
|
|
}
|
2014-10-27 15:15:25 +08:00
|
|
|
if (res != NULL) {
|
|
|
|
res->name = dev->full_name;
|
|
|
|
res->flags = range.flags;
|
|
|
|
res->start = range.cpu_addr;
|
|
|
|
res->end = range.cpu_addr + range.size - 1;
|
|
|
|
res->parent = res->child = res->sibling = NULL;
|
|
|
|
}
|
2010-01-18 21:42:34 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/* If there's an ISA hole and the pci_mem_offset is -not- matching
|
|
|
|
* the ISA hole offset, then we need to remove the ISA hole from
|
|
|
|
* the resource list for that brige
|
|
|
|
*/
|
|
|
|
if (isa_hole >= 0 && hose->pci_mem_offset != isa_mb) {
|
|
|
|
unsigned int next = isa_hole + 1;
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_info(" Removing ISA hole at 0x%016llx\n", isa_mb);
|
2010-01-18 21:42:34 +08:00
|
|
|
if (next < memno)
|
|
|
|
memmove(&hose->mem_resources[isa_hole],
|
|
|
|
&hose->mem_resources[next],
|
|
|
|
sizeof(struct resource) * (memno - next));
|
|
|
|
hose->mem_resources[--memno].flags = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-09-01 18:14:46 +08:00
|
|
|
/* Display the domain number in /proc */
|
2010-01-18 21:42:34 +08:00
|
|
|
int pci_proc_domain(struct pci_bus *bus)
|
|
|
|
{
|
2016-09-01 18:14:46 +08:00
|
|
|
return pci_domain_nr(bus);
|
2010-01-18 21:42:34 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/* This header fixup will do the resource fixup for all devices as they are
|
|
|
|
* probed, but not for bridge ranges
|
|
|
|
*/
|
2012-12-22 06:06:37 +08:00
|
|
|
static void pcibios_fixup_resources(struct pci_dev *dev)
|
2010-01-18 21:42:34 +08:00
|
|
|
{
|
|
|
|
struct pci_controller *hose = pci_bus_to_host(dev->bus);
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (!hose) {
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_err("No host bridge for PCI dev %s !\n",
|
2010-01-18 21:42:34 +08:00
|
|
|
pci_name(dev));
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
for (i = 0; i < DEVICE_COUNT_RESOURCE; i++) {
|
|
|
|
struct resource *res = dev->resource + i;
|
|
|
|
if (!res->flags)
|
|
|
|
continue;
|
2012-02-24 11:18:57 +08:00
|
|
|
if (res->start == 0) {
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_debug("PCI:%s Resource %d %016llx-%016llx [%x]",
|
2010-01-18 21:42:34 +08:00
|
|
|
pci_name(dev), i,
|
|
|
|
(unsigned long long)res->start,
|
|
|
|
(unsigned long long)res->end,
|
|
|
|
(unsigned int)res->flags);
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_debug("is unassigned\n");
|
2010-01-18 21:42:34 +08:00
|
|
|
res->end -= res->start;
|
|
|
|
res->start = 0;
|
|
|
|
res->flags |= IORESOURCE_UNSET;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
2012-02-24 11:19:02 +08:00
|
|
|
pr_debug("PCI:%s Resource %d %016llx-%016llx [%x]\n",
|
2010-01-18 21:42:34 +08:00
|
|
|
pci_name(dev), i,
|
2012-12-27 17:40:38 +08:00
|
|
|
(unsigned long long)res->start,
|
2010-01-18 21:42:34 +08:00
|
|
|
(unsigned long long)res->end,
|
|
|
|
(unsigned int)res->flags);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_ANY_ID, PCI_ANY_ID, pcibios_fixup_resources);
|
|
|
|
|
2021-09-13 23:27:08 +08:00
|
|
|
int pcibios_device_add(struct pci_dev *dev)
|
2016-02-12 00:28:11 +08:00
|
|
|
{
|
|
|
|
dev->irq = of_irq_parse_and_map_pci(dev, 0, 0);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2010-01-18 21:42:34 +08:00
|
|
|
/*
|
|
|
|
* Reparent resource children of pr that conflict with res
|
|
|
|
* under res, and make res replace those children.
|
|
|
|
*/
|
|
|
|
static int __init reparent_resources(struct resource *parent,
|
|
|
|
struct resource *res)
|
|
|
|
{
|
|
|
|
struct resource *p, **pp;
|
|
|
|
struct resource **firstpp = NULL;
|
|
|
|
|
|
|
|
for (pp = &parent->child; (p = *pp) != NULL; pp = &p->sibling) {
|
|
|
|
if (p->end < res->start)
|
|
|
|
continue;
|
|
|
|
if (res->end < p->start)
|
|
|
|
break;
|
|
|
|
if (p->start < res->start || p->end > res->end)
|
|
|
|
return -1; /* not completely contained */
|
|
|
|
if (firstpp == NULL)
|
|
|
|
firstpp = pp;
|
|
|
|
}
|
|
|
|
if (firstpp == NULL)
|
|
|
|
return -1; /* didn't find any conflicting entries? */
|
|
|
|
res->parent = parent;
|
|
|
|
res->child = *firstpp;
|
|
|
|
res->sibling = *pp;
|
|
|
|
*firstpp = res;
|
|
|
|
*pp = NULL;
|
|
|
|
for (p = res->child; p != NULL; p = p->sibling) {
|
|
|
|
p->parent = res;
|
|
|
|
pr_debug("PCI: Reparented %s [%llx..%llx] under %s\n",
|
|
|
|
p->name,
|
|
|
|
(unsigned long long)p->start,
|
|
|
|
(unsigned long long)p->end, res->name);
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Handle resources of PCI devices. If the world were perfect, we could
|
|
|
|
* just allocate all the resource regions and do nothing more. It isn't.
|
|
|
|
* On the other hand, we cannot just re-allocate all devices, as it would
|
|
|
|
* require us to know lots of host bridge internals. So we attempt to
|
|
|
|
* keep as much of the original configuration as possible, but tweak it
|
|
|
|
* when it's found to be wrong.
|
|
|
|
*
|
|
|
|
* Known BIOS problems we have to work around:
|
|
|
|
* - I/O or memory regions not configured
|
|
|
|
* - regions configured, but not enabled in the command register
|
|
|
|
* - bogus I/O addresses above 64K used
|
|
|
|
* - expansion ROMs left enabled (this may sound harmless, but given
|
|
|
|
* the fact the PCI specs explicitly allow address decoders to be
|
|
|
|
* shared between expansion ROMs and other resource regions, it's
|
|
|
|
* at least dangerous)
|
|
|
|
*
|
|
|
|
* Our solution:
|
|
|
|
* (1) Allocate resources for all buses behind PCI-to-PCI bridges.
|
|
|
|
* This gives us fixed barriers on where we can allocate.
|
|
|
|
* (2) Allocate resources for all enabled devices. If there is
|
|
|
|
* a collision, just mark the resource as unallocated. Also
|
|
|
|
* disable expansion ROMs during this step.
|
|
|
|
* (3) Try to allocate resources for disabled devices. If the
|
|
|
|
* resources were assigned correctly, everything goes well,
|
|
|
|
* if they weren't, they won't disturb allocation of other
|
|
|
|
* resources.
|
|
|
|
* (4) Assign new addresses to resources which were either
|
|
|
|
* not configured at all or misconfigured. If explicitly
|
|
|
|
* requested by the user, configure expansion ROM address
|
|
|
|
* as well.
|
|
|
|
*/
|
|
|
|
|
2013-01-04 16:14:46 +08:00
|
|
|
static void pcibios_allocate_bus_resources(struct pci_bus *bus)
|
2010-01-18 21:42:34 +08:00
|
|
|
{
|
|
|
|
struct pci_bus *b;
|
|
|
|
int i;
|
|
|
|
struct resource *res, *pr;
|
|
|
|
|
|
|
|
pr_debug("PCI: Allocating bus resources for %04x:%02x...\n",
|
|
|
|
pci_domain_nr(bus), bus->number);
|
|
|
|
|
2010-04-16 15:03:00 +08:00
|
|
|
pci_bus_for_each_resource(bus, res, i) {
|
2010-01-18 21:42:34 +08:00
|
|
|
if (!res || !res->flags
|
|
|
|
|| res->start > res->end || res->parent)
|
|
|
|
continue;
|
|
|
|
if (bus->parent == NULL)
|
|
|
|
pr = (res->flags & IORESOURCE_IO) ?
|
|
|
|
&ioport_resource : &iomem_resource;
|
|
|
|
else {
|
|
|
|
/* Don't bother with non-root busses when
|
|
|
|
* re-assigning all resources. We clear the
|
|
|
|
* resource flags as if they were colliding
|
|
|
|
* and as such ensure proper re-allocation
|
|
|
|
* later.
|
|
|
|
*/
|
|
|
|
pr = pci_find_parent_resource(bus->self, res);
|
|
|
|
if (pr == res) {
|
|
|
|
/* this happens when the generic PCI
|
|
|
|
* code (wrongly) decides that this
|
|
|
|
* bridge is transparent -- paulus
|
|
|
|
*/
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_debug("PCI: %s (bus %d) bridge rsrc %d: %016llx-%016llx ",
|
2010-01-18 21:42:34 +08:00
|
|
|
bus->self ? pci_name(bus->self) : "PHB",
|
|
|
|
bus->number, i,
|
|
|
|
(unsigned long long)res->start,
|
2012-12-27 17:40:38 +08:00
|
|
|
(unsigned long long)res->end);
|
|
|
|
pr_debug("[0x%x], parent %p (%s)\n",
|
2010-01-18 21:42:34 +08:00
|
|
|
(unsigned int)res->flags,
|
|
|
|
pr, (pr && pr->name) ? pr->name : "nil");
|
|
|
|
|
|
|
|
if (pr && !(pr->flags & IORESOURCE_UNSET)) {
|
2015-01-16 06:21:50 +08:00
|
|
|
struct pci_dev *dev = bus->self;
|
|
|
|
|
2010-01-18 21:42:34 +08:00
|
|
|
if (request_resource(pr, res) == 0)
|
|
|
|
continue;
|
|
|
|
/*
|
|
|
|
* Must be a conflict with an existing entry.
|
|
|
|
* Move that entry (or entries) under the
|
|
|
|
* bridge resource and try again.
|
|
|
|
*/
|
|
|
|
if (reparent_resources(pr, res) == 0)
|
|
|
|
continue;
|
2015-01-16 06:21:50 +08:00
|
|
|
|
|
|
|
if (dev && i < PCI_BRIDGE_RESOURCE_NUM &&
|
|
|
|
pci_claim_bridge_resource(dev,
|
|
|
|
i + PCI_BRIDGE_RESOURCES) == 0)
|
|
|
|
continue;
|
|
|
|
|
2010-01-18 21:42:34 +08:00
|
|
|
}
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_warn("PCI: Cannot allocate resource region ");
|
|
|
|
pr_cont("%d of PCI bridge %d, will remap\n", i, bus->number);
|
2010-06-04 04:43:03 +08:00
|
|
|
res->start = res->end = 0;
|
2010-01-18 21:42:34 +08:00
|
|
|
res->flags = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
list_for_each_entry(b, &bus->children, node)
|
|
|
|
pcibios_allocate_bus_resources(b);
|
|
|
|
}
|
|
|
|
|
2012-12-22 06:06:37 +08:00
|
|
|
static inline void alloc_resource(struct pci_dev *dev, int idx)
|
2010-01-18 21:42:34 +08:00
|
|
|
{
|
|
|
|
struct resource *pr, *r = &dev->resource[idx];
|
|
|
|
|
|
|
|
pr_debug("PCI: Allocating %s: Resource %d: %016llx..%016llx [%x]\n",
|
|
|
|
pci_name(dev), idx,
|
|
|
|
(unsigned long long)r->start,
|
|
|
|
(unsigned long long)r->end,
|
|
|
|
(unsigned int)r->flags);
|
|
|
|
|
|
|
|
pr = pci_find_parent_resource(dev, r);
|
|
|
|
if (!pr || (pr->flags & IORESOURCE_UNSET) ||
|
|
|
|
request_resource(pr, r) < 0) {
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_warn("PCI: Cannot allocate resource region %d ", idx);
|
|
|
|
pr_cont("of device %s, will remap\n", pci_name(dev));
|
2010-01-18 21:42:34 +08:00
|
|
|
if (pr)
|
|
|
|
pr_debug("PCI: parent is %p: %016llx-%016llx [%x]\n",
|
|
|
|
pr,
|
|
|
|
(unsigned long long)pr->start,
|
|
|
|
(unsigned long long)pr->end,
|
|
|
|
(unsigned int)pr->flags);
|
|
|
|
/* We'll assign a new address later */
|
|
|
|
r->flags |= IORESOURCE_UNSET;
|
|
|
|
r->end -= r->start;
|
|
|
|
r->start = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __init pcibios_allocate_resources(int pass)
|
|
|
|
{
|
|
|
|
struct pci_dev *dev = NULL;
|
|
|
|
int idx, disabled;
|
|
|
|
u16 command;
|
|
|
|
struct resource *r;
|
|
|
|
|
|
|
|
for_each_pci_dev(dev) {
|
|
|
|
pci_read_config_word(dev, PCI_COMMAND, &command);
|
|
|
|
for (idx = 0; idx <= PCI_ROM_RESOURCE; idx++) {
|
|
|
|
r = &dev->resource[idx];
|
|
|
|
if (r->parent) /* Already allocated */
|
|
|
|
continue;
|
|
|
|
if (!r->flags || (r->flags & IORESOURCE_UNSET))
|
|
|
|
continue; /* Not assigned at all */
|
|
|
|
/* We only allocate ROMs on pass 1 just in case they
|
|
|
|
* have been screwed up by firmware
|
|
|
|
*/
|
|
|
|
if (idx == PCI_ROM_RESOURCE)
|
|
|
|
disabled = 1;
|
|
|
|
if (r->flags & IORESOURCE_IO)
|
|
|
|
disabled = !(command & PCI_COMMAND_IO);
|
|
|
|
else
|
|
|
|
disabled = !(command & PCI_COMMAND_MEMORY);
|
|
|
|
if (pass == disabled)
|
|
|
|
alloc_resource(dev, idx);
|
|
|
|
}
|
|
|
|
if (pass)
|
|
|
|
continue;
|
|
|
|
r = &dev->resource[PCI_ROM_RESOURCE];
|
|
|
|
if (r->flags) {
|
|
|
|
/* Turn the ROM off, leave the resource region,
|
|
|
|
* but keep it unregistered.
|
|
|
|
*/
|
|
|
|
u32 reg;
|
|
|
|
pci_read_config_dword(dev, dev->rom_base_reg, ®);
|
|
|
|
if (reg & PCI_ROM_ADDRESS_ENABLE) {
|
|
|
|
pr_debug("PCI: Switching off ROM of %s\n",
|
|
|
|
pci_name(dev));
|
|
|
|
r->flags &= ~IORESOURCE_ROM_ENABLE;
|
|
|
|
pci_write_config_dword(dev, dev->rom_base_reg,
|
|
|
|
reg & ~PCI_ROM_ADDRESS_ENABLE);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __init pcibios_reserve_legacy_regions(struct pci_bus *bus)
|
|
|
|
{
|
|
|
|
struct pci_controller *hose = pci_bus_to_host(bus);
|
|
|
|
resource_size_t offset;
|
|
|
|
struct resource *res, *pres;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
pr_debug("Reserving legacy ranges for domain %04x\n",
|
|
|
|
pci_domain_nr(bus));
|
|
|
|
|
|
|
|
/* Check for IO */
|
|
|
|
if (!(hose->io_resource.flags & IORESOURCE_IO))
|
|
|
|
goto no_io;
|
|
|
|
offset = (unsigned long)hose->io_base_virt - _IO_BASE;
|
|
|
|
res = kzalloc(sizeof(struct resource), GFP_KERNEL);
|
|
|
|
BUG_ON(res == NULL);
|
|
|
|
res->name = "Legacy IO";
|
|
|
|
res->flags = IORESOURCE_IO;
|
|
|
|
res->start = offset;
|
|
|
|
res->end = (offset + 0xfff) & 0xfffffffful;
|
|
|
|
pr_debug("Candidate legacy IO: %pR\n", res);
|
|
|
|
if (request_resource(&hose->io_resource, res)) {
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_debug("PCI %04x:%02x Cannot reserve Legacy IO %pR\n",
|
2010-01-18 21:42:34 +08:00
|
|
|
pci_domain_nr(bus), bus->number, res);
|
|
|
|
kfree(res);
|
|
|
|
}
|
|
|
|
|
|
|
|
no_io:
|
|
|
|
/* Check for memory */
|
|
|
|
offset = hose->pci_mem_offset;
|
|
|
|
pr_debug("hose mem offset: %016llx\n", (unsigned long long)offset);
|
|
|
|
for (i = 0; i < 3; i++) {
|
|
|
|
pres = &hose->mem_resources[i];
|
|
|
|
if (!(pres->flags & IORESOURCE_MEM))
|
|
|
|
continue;
|
|
|
|
pr_debug("hose mem res: %pR\n", pres);
|
|
|
|
if ((pres->start - offset) <= 0xa0000 &&
|
|
|
|
(pres->end - offset) >= 0xbffff)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (i >= 3)
|
|
|
|
return;
|
|
|
|
res = kzalloc(sizeof(struct resource), GFP_KERNEL);
|
|
|
|
BUG_ON(res == NULL);
|
|
|
|
res->name = "Legacy VGA memory";
|
|
|
|
res->flags = IORESOURCE_MEM;
|
|
|
|
res->start = 0xa0000 + offset;
|
|
|
|
res->end = 0xbffff + offset;
|
|
|
|
pr_debug("Candidate VGA memory: %pR\n", res);
|
|
|
|
if (request_resource(pres, res)) {
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_debug("PCI %04x:%02x Cannot reserve VGA memory %pR\n",
|
2010-01-18 21:42:34 +08:00
|
|
|
pci_domain_nr(bus), bus->number, res);
|
|
|
|
kfree(res);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void __init pcibios_resource_survey(void)
|
|
|
|
{
|
|
|
|
struct pci_bus *b;
|
|
|
|
|
|
|
|
/* Allocate and assign resources. If we re-assign everything, then
|
|
|
|
* we skip the allocate phase
|
|
|
|
*/
|
|
|
|
list_for_each_entry(b, &pci_root_buses, node)
|
|
|
|
pcibios_allocate_bus_resources(b);
|
|
|
|
|
2012-02-24 11:18:57 +08:00
|
|
|
pcibios_allocate_resources(0);
|
|
|
|
pcibios_allocate_resources(1);
|
2010-01-18 21:42:34 +08:00
|
|
|
|
|
|
|
/* Before we start assigning unassigned resource, we try to reserve
|
|
|
|
* the low IO area and the VGA memory area if they intersect the
|
|
|
|
* bus available resources to avoid allocating things on top of them
|
|
|
|
*/
|
2012-02-24 11:18:57 +08:00
|
|
|
list_for_each_entry(b, &pci_root_buses, node)
|
|
|
|
pcibios_reserve_legacy_regions(b);
|
2010-01-18 21:42:34 +08:00
|
|
|
|
2012-02-24 11:18:57 +08:00
|
|
|
/* Now proceed to assigning things that were left unassigned */
|
|
|
|
pr_debug("PCI: Assigning unassigned resources...\n");
|
|
|
|
pci_assign_unassigned_resources();
|
2010-01-18 21:42:34 +08:00
|
|
|
}
|
|
|
|
|
2012-12-22 06:06:37 +08:00
|
|
|
static void pcibios_setup_phb_resources(struct pci_controller *hose,
|
|
|
|
struct list_head *resources)
|
2010-01-18 21:42:34 +08:00
|
|
|
{
|
2012-05-16 07:03:25 +08:00
|
|
|
unsigned long io_offset;
|
2010-01-18 21:42:34 +08:00
|
|
|
struct resource *res;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
/* Hookup PHB IO resource */
|
2011-10-29 06:26:46 +08:00
|
|
|
res = &hose->io_resource;
|
|
|
|
|
|
|
|
/* Fixup IO space offset */
|
|
|
|
io_offset = (unsigned long)hose->io_base_virt - isa_io_base;
|
|
|
|
res->start = (res->start + io_offset) & 0xffffffffu;
|
|
|
|
res->end = (res->end + io_offset) & 0xffffffffu;
|
2010-01-18 21:42:34 +08:00
|
|
|
|
|
|
|
if (!res->flags) {
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_warn("PCI: I/O resource not set for host ");
|
2017-06-07 04:57:36 +08:00
|
|
|
pr_cont("bridge %pOF (domain %d)\n",
|
|
|
|
hose->dn, hose->global_number);
|
2010-01-18 21:42:34 +08:00
|
|
|
/* Workaround for lack of IO resource only on 32-bit */
|
|
|
|
res->start = (unsigned long)hose->io_base_virt - isa_io_base;
|
|
|
|
res->end = res->start + IO_SPACE_LIMIT;
|
|
|
|
res->flags = IORESOURCE_IO;
|
|
|
|
}
|
2013-01-04 16:14:46 +08:00
|
|
|
pci_add_resource_offset(resources, res,
|
|
|
|
(__force resource_size_t)(hose->io_base_virt - _IO_BASE));
|
2010-01-18 21:42:34 +08:00
|
|
|
|
|
|
|
pr_debug("PCI: PHB IO resource = %016llx-%016llx [%lx]\n",
|
|
|
|
(unsigned long long)res->start,
|
|
|
|
(unsigned long long)res->end,
|
|
|
|
(unsigned long)res->flags);
|
|
|
|
|
|
|
|
/* Hookup PHB Memory resources */
|
|
|
|
for (i = 0; i < 3; ++i) {
|
|
|
|
res = &hose->mem_resources[i];
|
|
|
|
if (!res->flags) {
|
|
|
|
if (i > 0)
|
|
|
|
continue;
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_err("PCI: Memory resource 0 not set for ");
|
2017-06-07 04:57:36 +08:00
|
|
|
pr_cont("host bridge %pOF (domain %d)\n",
|
|
|
|
hose->dn, hose->global_number);
|
2010-01-18 21:42:34 +08:00
|
|
|
|
|
|
|
/* Workaround for lack of MEM resource only on 32-bit */
|
|
|
|
res->start = hose->pci_mem_offset;
|
|
|
|
res->end = (resource_size_t)-1LL;
|
|
|
|
res->flags = IORESOURCE_MEM;
|
|
|
|
|
|
|
|
}
|
2012-02-24 11:19:02 +08:00
|
|
|
pci_add_resource_offset(resources, res, hose->pci_mem_offset);
|
2010-01-18 21:42:34 +08:00
|
|
|
|
|
|
|
pr_debug("PCI: PHB MEM resource %d = %016llx-%016llx [%lx]\n",
|
|
|
|
i, (unsigned long long)res->start,
|
|
|
|
(unsigned long long)res->end,
|
|
|
|
(unsigned long)res->flags);
|
|
|
|
}
|
|
|
|
|
|
|
|
pr_debug("PCI: PHB MEM offset = %016llx\n",
|
|
|
|
(unsigned long long)hose->pci_mem_offset);
|
|
|
|
pr_debug("PCI: PHB IO offset = %08lx\n",
|
|
|
|
(unsigned long)hose->io_base_virt - _IO_BASE);
|
|
|
|
}
|
|
|
|
|
2012-12-22 06:06:37 +08:00
|
|
|
static void pcibios_scan_phb(struct pci_controller *hose)
|
2011-04-11 09:17:26 +08:00
|
|
|
{
|
2011-10-29 06:26:46 +08:00
|
|
|
LIST_HEAD(resources);
|
2011-04-11 09:17:26 +08:00
|
|
|
struct pci_bus *bus;
|
|
|
|
struct device_node *node = hose->dn;
|
|
|
|
|
2017-06-07 04:57:36 +08:00
|
|
|
pr_debug("PCI: Scanning PHB %pOF\n", node);
|
2011-04-11 09:17:26 +08:00
|
|
|
|
2011-10-29 06:26:46 +08:00
|
|
|
pcibios_setup_phb_resources(hose, &resources);
|
|
|
|
|
2011-10-29 06:26:52 +08:00
|
|
|
bus = pci_scan_root_bus(hose->parent, hose->first_busno,
|
|
|
|
hose->ops, hose, &resources);
|
2011-04-11 09:17:26 +08:00
|
|
|
if (bus == NULL) {
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_err("Failed to create bus for PCI domain %04x\n",
|
2011-04-11 09:17:26 +08:00
|
|
|
hose->global_number);
|
2011-10-29 06:26:46 +08:00
|
|
|
pci_free_resource_list(&resources);
|
2011-04-11 09:17:26 +08:00
|
|
|
return;
|
|
|
|
}
|
2012-05-18 09:51:11 +08:00
|
|
|
bus->busn_res.start = hose->first_busno;
|
2011-04-11 09:17:26 +08:00
|
|
|
hose->bus = bus;
|
|
|
|
|
2012-05-18 09:51:11 +08:00
|
|
|
hose->last_busno = bus->busn_res.end;
|
2011-04-11 09:17:26 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static int __init pcibios_init(void)
|
|
|
|
{
|
|
|
|
struct pci_controller *hose, *tmp;
|
|
|
|
int next_busno = 0;
|
|
|
|
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_info("PCI: Probing PCI hardware\n");
|
2011-04-11 09:17:26 +08:00
|
|
|
|
|
|
|
/* Scan all of the recorded PCI controllers. */
|
|
|
|
list_for_each_entry_safe(hose, tmp, &hose_list, list_node) {
|
|
|
|
hose->last_busno = 0xff;
|
|
|
|
pcibios_scan_phb(hose);
|
|
|
|
if (next_busno <= hose->last_busno)
|
|
|
|
next_busno = hose->last_busno + 1;
|
|
|
|
}
|
|
|
|
pci_bus_count = next_busno;
|
|
|
|
|
|
|
|
/* Call common code to handle resource allocation */
|
|
|
|
pcibios_resource_survey();
|
PCI: Assign resources before drivers claim devices (pci_scan_root_bus())
Previously, pci_scan_root_bus() created a root PCI bus, enumerated the
devices on it, and called pci_bus_add_devices(), which made the devices
available for drivers to claim them.
Most callers assigned resources to devices after pci_scan_root_bus()
returns, which may be after drivers have claimed the devices. This is
incorrect; the PCI core should not change device resources while a driver
is managing the device.
Remove pci_bus_add_devices() from pci_scan_root_bus() and do it after any
resource assignment in the callers.
Note that ARM's pci_common_init_dev() already called pci_bus_add_devices()
after pci_scan_root_bus(), so we only need to remove the first call:
pci_common_init_dev
pcibios_init_hw
pci_scan_root_bus
pci_bus_add_devices # first call
pci_bus_assign_resources
pci_bus_add_devices # second call
[bhelgaas: changelog, drop "root_bus" var in alpha common_init_pci(),
return failure earlier in mn10300, add "return" in x86 pcibios_scan_root(),
return early if xtensa platform_pcibios_fixup() fails]
Signed-off-by: Yijing Wang <wangyijing@huawei.com>
Signed-off-by: Bjorn Helgaas <bhelgaas@google.com>
CC: Richard Henderson <rth@twiddle.net>
CC: Ivan Kokshaysky <ink@jurassic.park.msu.ru>
CC: Matt Turner <mattst88@gmail.com>
CC: David Howells <dhowells@redhat.com>
CC: Tony Luck <tony.luck@intel.com>
CC: Michal Simek <monstr@monstr.eu>
CC: Ralf Baechle <ralf@linux-mips.org>
CC: Koichi Yasutake <yasutake.koichi@jp.panasonic.com>
CC: Sebastian Ott <sebott@linux.vnet.ibm.com>
CC: "David S. Miller" <davem@davemloft.net>
CC: Chris Metcalf <cmetcalf@ezchip.com>
CC: Chris Zankel <chris@zankel.net>
CC: Max Filippov <jcmvbkbc@gmail.com>
CC: Thomas Gleixner <tglx@linutronix.de>
2015-03-16 11:18:56 +08:00
|
|
|
list_for_each_entry_safe(hose, tmp, &hose_list, list_node) {
|
|
|
|
if (hose->bus)
|
|
|
|
pci_bus_add_devices(hose->bus);
|
|
|
|
}
|
2011-04-11 09:17:26 +08:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
subsys_initcall(pcibios_init);
|
|
|
|
|
|
|
|
static struct pci_controller *pci_bus_to_hose(int bus)
|
|
|
|
{
|
|
|
|
struct pci_controller *hose, *tmp;
|
|
|
|
|
|
|
|
list_for_each_entry_safe(hose, tmp, &hose_list, list_node)
|
|
|
|
if (bus >= hose->first_busno && bus <= hose->last_busno)
|
|
|
|
return hose;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Provide information on locations of various I/O regions in physical
|
|
|
|
* memory. Do this on a per-card basis so that we choose the right
|
|
|
|
* root bridge.
|
|
|
|
* Note that the returned IO or memory base is a physical address
|
|
|
|
*/
|
|
|
|
|
|
|
|
long sys_pciconfig_iobase(long which, unsigned long bus, unsigned long devfn)
|
|
|
|
{
|
|
|
|
struct pci_controller *hose;
|
|
|
|
long result = -EOPNOTSUPP;
|
|
|
|
|
|
|
|
hose = pci_bus_to_hose(bus);
|
|
|
|
if (!hose)
|
|
|
|
return -ENODEV;
|
|
|
|
|
|
|
|
switch (which) {
|
|
|
|
case IOBASE_BRIDGE_NUMBER:
|
|
|
|
return (long)hose->first_busno;
|
|
|
|
case IOBASE_MEMORY:
|
|
|
|
return (long)hose->pci_mem_offset;
|
|
|
|
case IOBASE_IO:
|
|
|
|
return (long)hose->io_base_phys;
|
|
|
|
case IOBASE_ISA_IO:
|
|
|
|
return (long)isa_io_base;
|
|
|
|
case IOBASE_ISA_MEM:
|
|
|
|
return (long)isa_mem_base;
|
|
|
|
}
|
|
|
|
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2010-01-18 21:42:34 +08:00
|
|
|
/*
|
|
|
|
* Null PCI config access functions, for the case when we can't
|
|
|
|
* find a hose.
|
|
|
|
*/
|
|
|
|
#define NULL_PCI_OP(rw, size, type) \
|
|
|
|
static int \
|
|
|
|
null_##rw##_config_##size(struct pci_dev *dev, int offset, type val) \
|
|
|
|
{ \
|
|
|
|
return PCIBIOS_DEVICE_NOT_FOUND; \
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
null_read_config(struct pci_bus *bus, unsigned int devfn, int offset,
|
|
|
|
int len, u32 *val)
|
|
|
|
{
|
|
|
|
return PCIBIOS_DEVICE_NOT_FOUND;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
null_write_config(struct pci_bus *bus, unsigned int devfn, int offset,
|
|
|
|
int len, u32 val)
|
|
|
|
{
|
|
|
|
return PCIBIOS_DEVICE_NOT_FOUND;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct pci_ops null_pci_ops = {
|
|
|
|
.read = null_read_config,
|
|
|
|
.write = null_write_config,
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* These functions are used early on before PCI scanning is done
|
|
|
|
* and all of the pci_dev and pci_bus structures have been created.
|
|
|
|
*/
|
|
|
|
static struct pci_bus *
|
|
|
|
fake_pci_bus(struct pci_controller *hose, int busnr)
|
|
|
|
{
|
|
|
|
static struct pci_bus bus;
|
|
|
|
|
|
|
|
if (!hose)
|
2012-12-27 17:40:38 +08:00
|
|
|
pr_err("Can't find hose for PCI bus %d!\n", busnr);
|
2010-01-18 21:42:34 +08:00
|
|
|
|
|
|
|
bus.number = busnr;
|
|
|
|
bus.sysdata = hose;
|
|
|
|
bus.ops = hose ? hose->ops : &null_pci_ops;
|
|
|
|
return &bus;
|
|
|
|
}
|
|
|
|
|
|
|
|
#define EARLY_PCI_OP(rw, size, type) \
|
|
|
|
int early_##rw##_config_##size(struct pci_controller *hose, int bus, \
|
|
|
|
int devfn, int offset, type value) \
|
|
|
|
{ \
|
|
|
|
return pci_bus_##rw##_config_##size(fake_pci_bus(hose, bus), \
|
|
|
|
devfn, offset, value); \
|
|
|
|
}
|
|
|
|
|
|
|
|
EARLY_PCI_OP(read, byte, u8 *)
|
|
|
|
EARLY_PCI_OP(read, word, u16 *)
|
|
|
|
EARLY_PCI_OP(read, dword, u32 *)
|
|
|
|
EARLY_PCI_OP(write, byte, u8)
|
|
|
|
EARLY_PCI_OP(write, word, u16)
|
|
|
|
EARLY_PCI_OP(write, dword, u32)
|
|
|
|
|
|
|
|
int early_find_capability(struct pci_controller *hose, int bus, int devfn,
|
|
|
|
int cap)
|
|
|
|
{
|
|
|
|
return pci_bus_find_capability(fake_pci_bus(hose, bus), devfn, cap);
|
|
|
|
}
|