Merge branch 'stable/xen-pcifront-0.8.2' of git://git.kernel.org/pub/scm/linux/kernel/git/konrad/xen
and branch 'for-linus' of git://xenbits.xen.org/people/sstabellini/linux-pvhvm * 'for-linus' of git://xenbits.xen.org/people/sstabellini/linux-pvhvm: xen: register xen pci notifier xen: initialize cpu masks for pv guests in xen_smp_init xen: add a missing #include to arch/x86/pci/xen.c xen: mask the MTRR feature from the cpuid xen: make hvc_xen console work for dom0. xen: add the direct mapping area for ISA bus access xen: Initialize xenbus for dom0. xen: use vcpu_ops to setup cpu masks xen: map a dummy page for local apic and ioapic in xen_set_fixmap xen: remap MSIs into pirqs when running as initial domain xen: remap GSIs as pirqs when running as initial domain xen: introduce XEN_DOM0 as a silent option xen: map MSIs into pirqs xen: support GSI -> pirq remapping in PV on HVM guests xen: add xen hvm acpi_register_gsi variant acpi: use indirect call to register gsi in different modes xen: implement xen_hvm_register_pirq xen: get the maximum number of pirqs from xen xen: support pirq != irq * 'stable/xen-pcifront-0.8.2' of git://git.kernel.org/pub/scm/linux/kernel/git/konrad/xen: (27 commits) X86/PCI: Remove the dependency on isapnp_disable. xen: Update Makefile with CONFIG_BLOCK dependency for biomerge.c MAINTAINERS: Add myself to the Xen Hypervisor Interface and remove Chris Wright. x86: xen: Sanitse irq handling (part two) swiotlb-xen: On x86-32 builts, select SWIOTLB instead of depending on it. MAINTAINERS: Add myself for Xen PCI and Xen SWIOTLB maintainer. xen/pci: Request ACS when Xen-SWIOTLB is activated. xen-pcifront: Xen PCI frontend driver. xenbus: prevent warnings on unhandled enumeration values xenbus: Xen paravirtualised PCI hotplug support. xen/x86/PCI: Add support for the Xen PCI subsystem x86: Introduce x86_msi_ops msi: Introduce default_[teardown|setup]_msi_irqs with fallback. x86/PCI: Export pci_walk_bus function. x86/PCI: make sure _PAGE_IOMAP it set on pci mappings x86/PCI: Clean up pci_cache_line_size xen: fix shared irq device passthrough xen: Provide a variant of xen_poll_irq with timeout. xen: Find an unbound irq number in reverse order (high to low). xen: statically initialize cpu_evtchn_mask_p ... Fix up trivial conflicts in drivers/pci/Makefile
This commit is contained in:
commit
18cb657ca1
22
MAINTAINERS
22
MAINTAINERS
|
@ -6595,12 +6595,26 @@ T: git git://git.kernel.org/pub/scm/linux/kernel/git/mjg59/platform-drivers-x86.
|
|||
S: Maintained
|
||||
F: drivers/platform/x86
|
||||
|
||||
XEN HYPERVISOR INTERFACE
|
||||
M: Jeremy Fitzhardinge <jeremy@xensource.com>
|
||||
M: Chris Wright <chrisw@sous-sol.org>
|
||||
L: virtualization@lists.osdl.org
|
||||
XEN PCI SUBSYSTEM
|
||||
M: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com>
|
||||
L: xen-devel@lists.xensource.com
|
||||
S: Supported
|
||||
F: arch/x86/pci/*xen*
|
||||
F: drivers/pci/*xen*
|
||||
|
||||
XEN SWIOTLB SUBSYSTEM
|
||||
M: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com>
|
||||
L: xen-devel@lists.xensource.com
|
||||
S: Supported
|
||||
F: arch/x86/xen/*swiotlb*
|
||||
F: drivers/xen/*swiotlb*
|
||||
|
||||
XEN HYPERVISOR INTERFACE
|
||||
M: Jeremy Fitzhardinge <jeremy.fitzhardinge@citrix.com>
|
||||
M: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com>
|
||||
L: xen-devel@lists.xen.org
|
||||
L: virtualization@lists.osdl.org
|
||||
S: Supported
|
||||
F: arch/x86/xen/
|
||||
F: drivers/*/xen-*front.c
|
||||
F: drivers/xen/
|
||||
|
|
|
@ -1893,6 +1893,11 @@ config PCI_OLPC
|
|||
def_bool y
|
||||
depends on PCI && OLPC && (PCI_GOOLPC || PCI_GOANY)
|
||||
|
||||
config PCI_XEN
|
||||
def_bool y
|
||||
depends on PCI && XEN
|
||||
select SWIOTLB_XEN
|
||||
|
||||
config PCI_DOMAINS
|
||||
def_bool y
|
||||
depends on PCI
|
||||
|
|
|
@ -93,6 +93,9 @@ extern u8 acpi_sci_flags;
|
|||
extern int acpi_sci_override_gsi;
|
||||
void acpi_pic_sci_set_trigger(unsigned int, u16);
|
||||
|
||||
extern int (*__acpi_register_gsi)(struct device *dev, u32 gsi,
|
||||
int trigger, int polarity);
|
||||
|
||||
static inline void disable_acpi(void)
|
||||
{
|
||||
acpi_disabled = 1;
|
||||
|
|
|
@ -41,6 +41,8 @@
|
|||
#include <asm-generic/int-ll64.h>
|
||||
#include <asm/page.h>
|
||||
|
||||
#include <xen/xen.h>
|
||||
|
||||
#define build_mmio_read(name, size, type, reg, barrier) \
|
||||
static inline type name(const volatile void __iomem *addr) \
|
||||
{ type ret; asm volatile("mov" size " %1,%0":reg (ret) \
|
||||
|
@ -351,6 +353,17 @@ extern void early_iounmap(void __iomem *addr, unsigned long size);
|
|||
extern void fixup_early_ioremap(void);
|
||||
extern bool is_early_ioremap_ptep(pte_t *ptep);
|
||||
|
||||
#ifdef CONFIG_XEN
|
||||
struct bio_vec;
|
||||
|
||||
extern bool xen_biovec_phys_mergeable(const struct bio_vec *vec1,
|
||||
const struct bio_vec *vec2);
|
||||
|
||||
#define BIOVEC_PHYS_MERGEABLE(vec1, vec2) \
|
||||
(__BIOVEC_PHYS_MERGEABLE(vec1, vec2) && \
|
||||
(!xen_domain() || xen_biovec_phys_mergeable(vec1, vec2)))
|
||||
#endif /* CONFIG_XEN */
|
||||
|
||||
#define IO_SPACE_LIMIT 0xffff
|
||||
|
||||
#endif /* _ASM_X86_IO_H */
|
||||
|
|
|
@ -169,6 +169,7 @@ extern void mask_IO_APIC_setup(struct IO_APIC_route_entry **ioapic_entries);
|
|||
extern int restore_IO_APIC_setup(struct IO_APIC_route_entry **ioapic_entries);
|
||||
|
||||
extern void probe_nr_irqs_gsi(void);
|
||||
extern int get_nr_irqs_gsi(void);
|
||||
|
||||
extern void setup_ioapic_ids_from_mpc(void);
|
||||
|
||||
|
|
|
@ -7,6 +7,7 @@
|
|||
#include <linux/string.h>
|
||||
#include <asm/scatterlist.h>
|
||||
#include <asm/io.h>
|
||||
#include <asm/x86_init.h>
|
||||
|
||||
#ifdef __KERNEL__
|
||||
|
||||
|
@ -94,8 +95,36 @@ static inline void early_quirks(void) { }
|
|||
|
||||
extern void pci_iommu_alloc(void);
|
||||
|
||||
/* MSI arch hook */
|
||||
#define arch_setup_msi_irqs arch_setup_msi_irqs
|
||||
#ifdef CONFIG_PCI_MSI
|
||||
/* MSI arch specific hooks */
|
||||
static inline int x86_setup_msi_irqs(struct pci_dev *dev, int nvec, int type)
|
||||
{
|
||||
return x86_msi.setup_msi_irqs(dev, nvec, type);
|
||||
}
|
||||
|
||||
static inline void x86_teardown_msi_irqs(struct pci_dev *dev)
|
||||
{
|
||||
x86_msi.teardown_msi_irqs(dev);
|
||||
}
|
||||
|
||||
static inline void x86_teardown_msi_irq(unsigned int irq)
|
||||
{
|
||||
x86_msi.teardown_msi_irq(irq);
|
||||
}
|
||||
#define arch_setup_msi_irqs x86_setup_msi_irqs
|
||||
#define arch_teardown_msi_irqs x86_teardown_msi_irqs
|
||||
#define arch_teardown_msi_irq x86_teardown_msi_irq
|
||||
/* implemented in arch/x86/kernel/apic/io_apic. */
|
||||
int native_setup_msi_irqs(struct pci_dev *dev, int nvec, int type);
|
||||
void native_teardown_msi_irq(unsigned int irq);
|
||||
/* default to the implementation in drivers/lib/msi.c */
|
||||
#define HAVE_DEFAULT_MSI_TEARDOWN_IRQS
|
||||
void default_teardown_msi_irqs(struct pci_dev *dev);
|
||||
#else
|
||||
#define native_setup_msi_irqs NULL
|
||||
#define native_teardown_msi_irq NULL
|
||||
#define default_teardown_msi_irqs NULL
|
||||
#endif
|
||||
|
||||
#define PCI_DMA_BUS_IS_PHYS (dma_ops->is_phys)
|
||||
|
||||
|
|
|
@ -47,6 +47,7 @@ enum pci_bf_sort_state {
|
|||
extern unsigned int pcibios_max_latency;
|
||||
|
||||
void pcibios_resource_survey(void);
|
||||
void pcibios_set_cache_line_size(void);
|
||||
|
||||
/* pci-pc.c */
|
||||
|
||||
|
|
|
@ -154,9 +154,18 @@ struct x86_platform_ops {
|
|||
int (*i8042_detect)(void);
|
||||
};
|
||||
|
||||
struct pci_dev;
|
||||
|
||||
struct x86_msi_ops {
|
||||
int (*setup_msi_irqs)(struct pci_dev *dev, int nvec, int type);
|
||||
void (*teardown_msi_irq)(unsigned int irq);
|
||||
void (*teardown_msi_irqs)(struct pci_dev *dev);
|
||||
};
|
||||
|
||||
extern struct x86_init_ops x86_init;
|
||||
extern struct x86_cpuinit_ops x86_cpuinit;
|
||||
extern struct x86_platform_ops x86_platform;
|
||||
extern struct x86_msi_ops x86_msi;
|
||||
|
||||
extern void x86_init_noop(void);
|
||||
extern void x86_init_uint_noop(unsigned int unused);
|
||||
|
|
|
@ -0,0 +1,65 @@
|
|||
#ifndef _ASM_X86_XEN_PCI_H
|
||||
#define _ASM_X86_XEN_PCI_H
|
||||
|
||||
#if defined(CONFIG_PCI_XEN)
|
||||
extern int __init pci_xen_init(void);
|
||||
extern int __init pci_xen_hvm_init(void);
|
||||
#define pci_xen 1
|
||||
#else
|
||||
#define pci_xen 0
|
||||
#define pci_xen_init (0)
|
||||
static inline int pci_xen_hvm_init(void)
|
||||
{
|
||||
return -1;
|
||||
}
|
||||
#endif
|
||||
#if defined(CONFIG_XEN_DOM0)
|
||||
void __init xen_setup_pirqs(void);
|
||||
#else
|
||||
static inline void __init xen_setup_pirqs(void)
|
||||
{
|
||||
}
|
||||
#endif
|
||||
|
||||
#if defined(CONFIG_PCI_MSI)
|
||||
#if defined(CONFIG_PCI_XEN)
|
||||
/* The drivers/pci/xen-pcifront.c sets this structure to
|
||||
* its own functions.
|
||||
*/
|
||||
struct xen_pci_frontend_ops {
|
||||
int (*enable_msi)(struct pci_dev *dev, int **vectors);
|
||||
void (*disable_msi)(struct pci_dev *dev);
|
||||
int (*enable_msix)(struct pci_dev *dev, int **vectors, int nvec);
|
||||
void (*disable_msix)(struct pci_dev *dev);
|
||||
};
|
||||
|
||||
extern struct xen_pci_frontend_ops *xen_pci_frontend;
|
||||
|
||||
static inline int xen_pci_frontend_enable_msi(struct pci_dev *dev,
|
||||
int **vectors)
|
||||
{
|
||||
if (xen_pci_frontend && xen_pci_frontend->enable_msi)
|
||||
return xen_pci_frontend->enable_msi(dev, vectors);
|
||||
return -ENODEV;
|
||||
}
|
||||
static inline void xen_pci_frontend_disable_msi(struct pci_dev *dev)
|
||||
{
|
||||
if (xen_pci_frontend && xen_pci_frontend->disable_msi)
|
||||
xen_pci_frontend->disable_msi(dev);
|
||||
}
|
||||
static inline int xen_pci_frontend_enable_msix(struct pci_dev *dev,
|
||||
int **vectors, int nvec)
|
||||
{
|
||||
if (xen_pci_frontend && xen_pci_frontend->enable_msix)
|
||||
return xen_pci_frontend->enable_msix(dev, vectors, nvec);
|
||||
return -ENODEV;
|
||||
}
|
||||
static inline void xen_pci_frontend_disable_msix(struct pci_dev *dev)
|
||||
{
|
||||
if (xen_pci_frontend && xen_pci_frontend->disable_msix)
|
||||
xen_pci_frontend->disable_msix(dev);
|
||||
}
|
||||
#endif /* CONFIG_PCI_XEN */
|
||||
#endif /* CONFIG_PCI_MSI */
|
||||
|
||||
#endif /* _ASM_X86_XEN_PCI_H */
|
|
@ -513,6 +513,33 @@ int acpi_isa_irq_to_gsi(unsigned isa_irq, u32 *gsi)
|
|||
return 0;
|
||||
}
|
||||
|
||||
static int acpi_register_gsi_pic(struct device *dev, u32 gsi,
|
||||
int trigger, int polarity)
|
||||
{
|
||||
#ifdef CONFIG_PCI
|
||||
/*
|
||||
* Make sure all (legacy) PCI IRQs are set as level-triggered.
|
||||
*/
|
||||
if (trigger == ACPI_LEVEL_SENSITIVE)
|
||||
eisa_set_level_irq(gsi);
|
||||
#endif
|
||||
|
||||
return gsi;
|
||||
}
|
||||
|
||||
static int acpi_register_gsi_ioapic(struct device *dev, u32 gsi,
|
||||
int trigger, int polarity)
|
||||
{
|
||||
#ifdef CONFIG_X86_IO_APIC
|
||||
gsi = mp_register_gsi(dev, gsi, trigger, polarity);
|
||||
#endif
|
||||
|
||||
return gsi;
|
||||
}
|
||||
|
||||
int (*__acpi_register_gsi)(struct device *dev, u32 gsi,
|
||||
int trigger, int polarity) = acpi_register_gsi_pic;
|
||||
|
||||
/*
|
||||
* success: return IRQ number (>=0)
|
||||
* failure: return < 0
|
||||
|
@ -522,26 +549,26 @@ int acpi_register_gsi(struct device *dev, u32 gsi, int trigger, int polarity)
|
|||
unsigned int irq;
|
||||
unsigned int plat_gsi = gsi;
|
||||
|
||||
#ifdef CONFIG_PCI
|
||||
/*
|
||||
* Make sure all (legacy) PCI IRQs are set as level-triggered.
|
||||
*/
|
||||
if (acpi_irq_model == ACPI_IRQ_MODEL_PIC) {
|
||||
if (trigger == ACPI_LEVEL_SENSITIVE)
|
||||
eisa_set_level_irq(gsi);
|
||||
}
|
||||
#endif
|
||||
|
||||
#ifdef CONFIG_X86_IO_APIC
|
||||
if (acpi_irq_model == ACPI_IRQ_MODEL_IOAPIC) {
|
||||
plat_gsi = mp_register_gsi(dev, gsi, trigger, polarity);
|
||||
}
|
||||
#endif
|
||||
plat_gsi = (*__acpi_register_gsi)(dev, gsi, trigger, polarity);
|
||||
irq = gsi_to_irq(plat_gsi);
|
||||
|
||||
return irq;
|
||||
}
|
||||
|
||||
void __init acpi_set_irq_model_pic(void)
|
||||
{
|
||||
acpi_irq_model = ACPI_IRQ_MODEL_PIC;
|
||||
__acpi_register_gsi = acpi_register_gsi_pic;
|
||||
acpi_ioapic = 0;
|
||||
}
|
||||
|
||||
void __init acpi_set_irq_model_ioapic(void)
|
||||
{
|
||||
acpi_irq_model = ACPI_IRQ_MODEL_IOAPIC;
|
||||
__acpi_register_gsi = acpi_register_gsi_ioapic;
|
||||
acpi_ioapic = 1;
|
||||
}
|
||||
|
||||
/*
|
||||
* ACPI based hotplug support for CPU
|
||||
*/
|
||||
|
@ -1259,8 +1286,7 @@ static void __init acpi_process_madt(void)
|
|||
*/
|
||||
error = acpi_parse_madt_ioapic_entries();
|
||||
if (!error) {
|
||||
acpi_irq_model = ACPI_IRQ_MODEL_IOAPIC;
|
||||
acpi_ioapic = 1;
|
||||
acpi_set_irq_model_ioapic();
|
||||
|
||||
smp_found_config = 1;
|
||||
}
|
||||
|
|
|
@ -3331,7 +3331,7 @@ static int setup_msi_irq(struct pci_dev *dev, struct msi_desc *msidesc, int irq)
|
|||
return 0;
|
||||
}
|
||||
|
||||
int arch_setup_msi_irqs(struct pci_dev *dev, int nvec, int type)
|
||||
int native_setup_msi_irqs(struct pci_dev *dev, int nvec, int type)
|
||||
{
|
||||
int node, ret, sub_handle, index = 0;
|
||||
unsigned int irq, irq_want;
|
||||
|
@ -3389,7 +3389,7 @@ error:
|
|||
return ret;
|
||||
}
|
||||
|
||||
void arch_teardown_msi_irq(unsigned int irq)
|
||||
void native_teardown_msi_irq(unsigned int irq)
|
||||
{
|
||||
destroy_irq(irq);
|
||||
}
|
||||
|
@ -3650,6 +3650,11 @@ void __init probe_nr_irqs_gsi(void)
|
|||
printk(KERN_DEBUG "nr_irqs_gsi: %d\n", nr_irqs_gsi);
|
||||
}
|
||||
|
||||
int get_nr_irqs_gsi(void)
|
||||
{
|
||||
return nr_irqs_gsi;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_SPARSE_IRQ
|
||||
int __init arch_probe_nr_irqs(void)
|
||||
{
|
||||
|
|
|
@ -6,10 +6,12 @@
|
|||
#include <linux/init.h>
|
||||
#include <linux/ioport.h>
|
||||
#include <linux/module.h>
|
||||
#include <linux/pci.h>
|
||||
|
||||
#include <asm/bios_ebda.h>
|
||||
#include <asm/paravirt.h>
|
||||
#include <asm/pci_x86.h>
|
||||
#include <asm/pci.h>
|
||||
#include <asm/mpspec.h>
|
||||
#include <asm/setup.h>
|
||||
#include <asm/apic.h>
|
||||
|
@ -99,3 +101,8 @@ struct x86_platform_ops x86_platform = {
|
|||
};
|
||||
|
||||
EXPORT_SYMBOL_GPL(x86_platform);
|
||||
struct x86_msi_ops x86_msi = {
|
||||
.setup_msi_irqs = native_setup_msi_irqs,
|
||||
.teardown_msi_irq = native_teardown_msi_irq,
|
||||
.teardown_msi_irqs = default_teardown_msi_irqs,
|
||||
};
|
||||
|
|
|
@ -4,6 +4,7 @@ obj-$(CONFIG_PCI_BIOS) += pcbios.o
|
|||
obj-$(CONFIG_PCI_MMCONFIG) += mmconfig_$(BITS).o direct.o mmconfig-shared.o
|
||||
obj-$(CONFIG_PCI_DIRECT) += direct.o
|
||||
obj-$(CONFIG_PCI_OLPC) += olpc.o
|
||||
obj-$(CONFIG_PCI_XEN) += xen.o
|
||||
|
||||
obj-y += fixup.o
|
||||
obj-$(CONFIG_ACPI) += acpi.o
|
||||
|
|
|
@ -421,16 +421,10 @@ struct pci_bus * __devinit pcibios_scan_root(int busnum)
|
|||
|
||||
return bus;
|
||||
}
|
||||
|
||||
int __init pcibios_init(void)
|
||||
void __init pcibios_set_cache_line_size(void)
|
||||
{
|
||||
struct cpuinfo_x86 *c = &boot_cpu_data;
|
||||
|
||||
if (!raw_pci_ops) {
|
||||
printk(KERN_WARNING "PCI: System does not support PCI\n");
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* Set PCI cacheline size to that of the CPU if the CPU has reported it.
|
||||
* (For older CPUs that don't support cpuid, we se it to 32 bytes
|
||||
|
@ -445,7 +439,16 @@ int __init pcibios_init(void)
|
|||
pci_dfl_cache_line_size = 32 >> 2;
|
||||
printk(KERN_DEBUG "PCI: Unknown cacheline size. Setting to 32 bytes\n");
|
||||
}
|
||||
}
|
||||
|
||||
int __init pcibios_init(void)
|
||||
{
|
||||
if (!raw_pci_ops) {
|
||||
printk(KERN_WARNING "PCI: System does not support PCI\n");
|
||||
return 0;
|
||||
}
|
||||
|
||||
pcibios_set_cache_line_size();
|
||||
pcibios_resource_survey();
|
||||
|
||||
if (pci_bf_sort >= pci_force_bf)
|
||||
|
|
|
@ -316,6 +316,8 @@ int pci_mmap_page_range(struct pci_dev *dev, struct vm_area_struct *vma,
|
|||
*/
|
||||
prot |= _PAGE_CACHE_UC_MINUS;
|
||||
|
||||
prot |= _PAGE_IOMAP; /* creating a mapping for IO */
|
||||
|
||||
vma->vm_page_prot = __pgprot(prot);
|
||||
|
||||
if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff,
|
||||
|
|
|
@ -0,0 +1,414 @@
|
|||
/*
|
||||
* Xen PCI Frontend Stub - puts some "dummy" functions in to the Linux
|
||||
* x86 PCI core to support the Xen PCI Frontend
|
||||
*
|
||||
* Author: Ryan Wilson <hap9@epoch.ncsc.mil>
|
||||
*/
|
||||
#include <linux/module.h>
|
||||
#include <linux/init.h>
|
||||
#include <linux/pci.h>
|
||||
#include <linux/acpi.h>
|
||||
|
||||
#include <linux/io.h>
|
||||
#include <asm/io_apic.h>
|
||||
#include <asm/pci_x86.h>
|
||||
|
||||
#include <asm/xen/hypervisor.h>
|
||||
|
||||
#include <xen/features.h>
|
||||
#include <xen/events.h>
|
||||
#include <asm/xen/pci.h>
|
||||
|
||||
#ifdef CONFIG_ACPI
|
||||
static int xen_hvm_register_pirq(u32 gsi, int triggering)
|
||||
{
|
||||
int rc, irq;
|
||||
struct physdev_map_pirq map_irq;
|
||||
int shareable = 0;
|
||||
char *name;
|
||||
|
||||
if (!xen_hvm_domain())
|
||||
return -1;
|
||||
|
||||
map_irq.domid = DOMID_SELF;
|
||||
map_irq.type = MAP_PIRQ_TYPE_GSI;
|
||||
map_irq.index = gsi;
|
||||
map_irq.pirq = -1;
|
||||
|
||||
rc = HYPERVISOR_physdev_op(PHYSDEVOP_map_pirq, &map_irq);
|
||||
if (rc) {
|
||||
printk(KERN_WARNING "xen map irq failed %d\n", rc);
|
||||
return -1;
|
||||
}
|
||||
|
||||
if (triggering == ACPI_EDGE_SENSITIVE) {
|
||||
shareable = 0;
|
||||
name = "ioapic-edge";
|
||||
} else {
|
||||
shareable = 1;
|
||||
name = "ioapic-level";
|
||||
}
|
||||
|
||||
irq = xen_map_pirq_gsi(map_irq.pirq, gsi, shareable, name);
|
||||
|
||||
printk(KERN_DEBUG "xen: --> irq=%d, pirq=%d\n", irq, map_irq.pirq);
|
||||
|
||||
return irq;
|
||||
}
|
||||
|
||||
static int acpi_register_gsi_xen_hvm(struct device *dev, u32 gsi,
|
||||
int trigger, int polarity)
|
||||
{
|
||||
return xen_hvm_register_pirq(gsi, trigger);
|
||||
}
|
||||
#endif
|
||||
|
||||
#if defined(CONFIG_PCI_MSI)
|
||||
#include <linux/msi.h>
|
||||
#include <asm/msidef.h>
|
||||
|
||||
struct xen_pci_frontend_ops *xen_pci_frontend;
|
||||
EXPORT_SYMBOL_GPL(xen_pci_frontend);
|
||||
|
||||
static void xen_msi_compose_msg(struct pci_dev *pdev, unsigned int pirq,
|
||||
struct msi_msg *msg)
|
||||
{
|
||||
/* We set vector == 0 to tell the hypervisor we don't care about it,
|
||||
* but we want a pirq setup instead.
|
||||
* We use the dest_id field to pass the pirq that we want. */
|
||||
msg->address_hi = MSI_ADDR_BASE_HI | MSI_ADDR_EXT_DEST_ID(pirq);
|
||||
msg->address_lo =
|
||||
MSI_ADDR_BASE_LO |
|
||||
MSI_ADDR_DEST_MODE_PHYSICAL |
|
||||
MSI_ADDR_REDIRECTION_CPU |
|
||||
MSI_ADDR_DEST_ID(pirq);
|
||||
|
||||
msg->data =
|
||||
MSI_DATA_TRIGGER_EDGE |
|
||||
MSI_DATA_LEVEL_ASSERT |
|
||||
/* delivery mode reserved */
|
||||
(3 << 8) |
|
||||
MSI_DATA_VECTOR(0);
|
||||
}
|
||||
|
||||
static int xen_hvm_setup_msi_irqs(struct pci_dev *dev, int nvec, int type)
|
||||
{
|
||||
int irq, pirq, ret = 0;
|
||||
struct msi_desc *msidesc;
|
||||
struct msi_msg msg;
|
||||
|
||||
list_for_each_entry(msidesc, &dev->msi_list, list) {
|
||||
xen_allocate_pirq_msi((type == PCI_CAP_ID_MSIX) ?
|
||||
"msi-x" : "msi", &irq, &pirq);
|
||||
if (irq < 0 || pirq < 0)
|
||||
goto error;
|
||||
printk(KERN_DEBUG "xen: msi --> irq=%d, pirq=%d\n", irq, pirq);
|
||||
xen_msi_compose_msg(dev, pirq, &msg);
|
||||
ret = set_irq_msi(irq, msidesc);
|
||||
if (ret < 0)
|
||||
goto error_while;
|
||||
write_msi_msg(irq, &msg);
|
||||
}
|
||||
return 0;
|
||||
|
||||
error_while:
|
||||
unbind_from_irqhandler(irq, NULL);
|
||||
error:
|
||||
if (ret == -ENODEV)
|
||||
dev_err(&dev->dev, "Xen PCI frontend has not registered" \
|
||||
" MSI/MSI-X support!\n");
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
/*
|
||||
* For MSI interrupts we have to use drivers/xen/event.s functions to
|
||||
* allocate an irq_desc and setup the right */
|
||||
|
||||
|
||||
static int xen_setup_msi_irqs(struct pci_dev *dev, int nvec, int type)
|
||||
{
|
||||
int irq, ret, i;
|
||||
struct msi_desc *msidesc;
|
||||
int *v;
|
||||
|
||||
v = kzalloc(sizeof(int) * max(1, nvec), GFP_KERNEL);
|
||||
if (!v)
|
||||
return -ENOMEM;
|
||||
|
||||
if (type == PCI_CAP_ID_MSIX)
|
||||
ret = xen_pci_frontend_enable_msix(dev, &v, nvec);
|
||||
else
|
||||
ret = xen_pci_frontend_enable_msi(dev, &v);
|
||||
if (ret)
|
||||
goto error;
|
||||
i = 0;
|
||||
list_for_each_entry(msidesc, &dev->msi_list, list) {
|
||||
irq = xen_allocate_pirq(v[i], 0, /* not sharable */
|
||||
(type == PCI_CAP_ID_MSIX) ?
|
||||
"pcifront-msi-x" : "pcifront-msi");
|
||||
if (irq < 0)
|
||||
return -1;
|
||||
|
||||
ret = set_irq_msi(irq, msidesc);
|
||||
if (ret)
|
||||
goto error_while;
|
||||
i++;
|
||||
}
|
||||
kfree(v);
|
||||
return 0;
|
||||
|
||||
error_while:
|
||||
unbind_from_irqhandler(irq, NULL);
|
||||
error:
|
||||
if (ret == -ENODEV)
|
||||
dev_err(&dev->dev, "Xen PCI frontend has not registered" \
|
||||
" MSI/MSI-X support!\n");
|
||||
|
||||
kfree(v);
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void xen_teardown_msi_irqs(struct pci_dev *dev)
|
||||
{
|
||||
struct msi_desc *msidesc;
|
||||
|
||||
msidesc = list_entry(dev->msi_list.next, struct msi_desc, list);
|
||||
if (msidesc->msi_attrib.is_msix)
|
||||
xen_pci_frontend_disable_msix(dev);
|
||||
else
|
||||
xen_pci_frontend_disable_msi(dev);
|
||||
}
|
||||
|
||||
static void xen_teardown_msi_irq(unsigned int irq)
|
||||
{
|
||||
xen_destroy_irq(irq);
|
||||
}
|
||||
|
||||
static int xen_initdom_setup_msi_irqs(struct pci_dev *dev, int nvec, int type)
|
||||
{
|
||||
int irq, ret;
|
||||
struct msi_desc *msidesc;
|
||||
|
||||
list_for_each_entry(msidesc, &dev->msi_list, list) {
|
||||
irq = xen_create_msi_irq(dev, msidesc, type);
|
||||
if (irq < 0)
|
||||
return -1;
|
||||
|
||||
ret = set_irq_msi(irq, msidesc);
|
||||
if (ret)
|
||||
goto error;
|
||||
}
|
||||
return 0;
|
||||
|
||||
error:
|
||||
xen_destroy_irq(irq);
|
||||
return ret;
|
||||
}
|
||||
#endif
|
||||
|
||||
static int xen_pcifront_enable_irq(struct pci_dev *dev)
|
||||
{
|
||||
int rc;
|
||||
int share = 1;
|
||||
|
||||
dev_info(&dev->dev, "Xen PCI enabling IRQ: %d\n", dev->irq);
|
||||
|
||||
if (dev->irq < 0)
|
||||
return -EINVAL;
|
||||
|
||||
if (dev->irq < NR_IRQS_LEGACY)
|
||||
share = 0;
|
||||
|
||||
rc = xen_allocate_pirq(dev->irq, share, "pcifront");
|
||||
if (rc < 0) {
|
||||
dev_warn(&dev->dev, "Xen PCI IRQ: %d, failed to register:%d\n",
|
||||
dev->irq, rc);
|
||||
return rc;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
int __init pci_xen_init(void)
|
||||
{
|
||||
if (!xen_pv_domain() || xen_initial_domain())
|
||||
return -ENODEV;
|
||||
|
||||
printk(KERN_INFO "PCI: setting up Xen PCI frontend stub\n");
|
||||
|
||||
pcibios_set_cache_line_size();
|
||||
|
||||
pcibios_enable_irq = xen_pcifront_enable_irq;
|
||||
pcibios_disable_irq = NULL;
|
||||
|
||||
#ifdef CONFIG_ACPI
|
||||
/* Keep ACPI out of the picture */
|
||||
acpi_noirq = 1;
|
||||
#endif
|
||||
|
||||
#ifdef CONFIG_PCI_MSI
|
||||
x86_msi.setup_msi_irqs = xen_setup_msi_irqs;
|
||||
x86_msi.teardown_msi_irq = xen_teardown_msi_irq;
|
||||
x86_msi.teardown_msi_irqs = xen_teardown_msi_irqs;
|
||||
#endif
|
||||
return 0;
|
||||
}
|
||||
|
||||
int __init pci_xen_hvm_init(void)
|
||||
{
|
||||
if (!xen_feature(XENFEAT_hvm_pirqs))
|
||||
return 0;
|
||||
|
||||
#ifdef CONFIG_ACPI
|
||||
/*
|
||||
* We don't want to change the actual ACPI delivery model,
|
||||
* just how GSIs get registered.
|
||||
*/
|
||||
__acpi_register_gsi = acpi_register_gsi_xen_hvm;
|
||||
#endif
|
||||
|
||||
#ifdef CONFIG_PCI_MSI
|
||||
x86_msi.setup_msi_irqs = xen_hvm_setup_msi_irqs;
|
||||
x86_msi.teardown_msi_irq = xen_teardown_msi_irq;
|
||||
#endif
|
||||
return 0;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_XEN_DOM0
|
||||
static int xen_register_pirq(u32 gsi, int triggering)
|
||||
{
|
||||
int rc, irq;
|
||||
struct physdev_map_pirq map_irq;
|
||||
int shareable = 0;
|
||||
char *name;
|
||||
|
||||
if (!xen_pv_domain())
|
||||
return -1;
|
||||
|
||||
if (triggering == ACPI_EDGE_SENSITIVE) {
|
||||
shareable = 0;
|
||||
name = "ioapic-edge";
|
||||
} else {
|
||||
shareable = 1;
|
||||
name = "ioapic-level";
|
||||
}
|
||||
|
||||
irq = xen_allocate_pirq(gsi, shareable, name);
|
||||
|
||||
printk(KERN_DEBUG "xen: --> irq=%d\n", irq);
|
||||
|
||||
if (irq < 0)
|
||||
goto out;
|
||||
|
||||
map_irq.domid = DOMID_SELF;
|
||||
map_irq.type = MAP_PIRQ_TYPE_GSI;
|
||||
map_irq.index = gsi;
|
||||
map_irq.pirq = irq;
|
||||
|
||||
rc = HYPERVISOR_physdev_op(PHYSDEVOP_map_pirq, &map_irq);
|
||||
if (rc) {
|
||||
printk(KERN_WARNING "xen map irq failed %d\n", rc);
|
||||
return -1;
|
||||
}
|
||||
|
||||
out:
|
||||
return irq;
|
||||
}
|
||||
|
||||
static int xen_register_gsi(u32 gsi, int triggering, int polarity)
|
||||
{
|
||||
int rc, irq;
|
||||
struct physdev_setup_gsi setup_gsi;
|
||||
|
||||
if (!xen_pv_domain())
|
||||
return -1;
|
||||
|
||||
printk(KERN_DEBUG "xen: registering gsi %u triggering %d polarity %d\n",
|
||||
gsi, triggering, polarity);
|
||||
|
||||
irq = xen_register_pirq(gsi, triggering);
|
||||
|
||||
setup_gsi.gsi = gsi;
|
||||
setup_gsi.triggering = (triggering == ACPI_EDGE_SENSITIVE ? 0 : 1);
|
||||
setup_gsi.polarity = (polarity == ACPI_ACTIVE_HIGH ? 0 : 1);
|
||||
|
||||
rc = HYPERVISOR_physdev_op(PHYSDEVOP_setup_gsi, &setup_gsi);
|
||||
if (rc == -EEXIST)
|
||||
printk(KERN_INFO "Already setup the GSI :%d\n", gsi);
|
||||
else if (rc) {
|
||||
printk(KERN_ERR "Failed to setup GSI :%d, err_code:%d\n",
|
||||
gsi, rc);
|
||||
}
|
||||
|
||||
return irq;
|
||||
}
|
||||
|
||||
static __init void xen_setup_acpi_sci(void)
|
||||
{
|
||||
int rc;
|
||||
int trigger, polarity;
|
||||
int gsi = acpi_sci_override_gsi;
|
||||
|
||||
if (!gsi)
|
||||
return;
|
||||
|
||||
rc = acpi_get_override_irq(gsi, &trigger, &polarity);
|
||||
if (rc) {
|
||||
printk(KERN_WARNING "xen: acpi_get_override_irq failed for acpi"
|
||||
" sci, rc=%d\n", rc);
|
||||
return;
|
||||
}
|
||||
trigger = trigger ? ACPI_LEVEL_SENSITIVE : ACPI_EDGE_SENSITIVE;
|
||||
polarity = polarity ? ACPI_ACTIVE_LOW : ACPI_ACTIVE_HIGH;
|
||||
|
||||
printk(KERN_INFO "xen: sci override: global_irq=%d trigger=%d "
|
||||
"polarity=%d\n", gsi, trigger, polarity);
|
||||
|
||||
gsi = xen_register_gsi(gsi, trigger, polarity);
|
||||
printk(KERN_INFO "xen: acpi sci %d\n", gsi);
|
||||
|
||||
return;
|
||||
}
|
||||
|
||||
static int acpi_register_gsi_xen(struct device *dev, u32 gsi,
|
||||
int trigger, int polarity)
|
||||
{
|
||||
return xen_register_gsi(gsi, trigger, polarity);
|
||||
}
|
||||
|
||||
static int __init pci_xen_initial_domain(void)
|
||||
{
|
||||
#ifdef CONFIG_PCI_MSI
|
||||
x86_msi.setup_msi_irqs = xen_initdom_setup_msi_irqs;
|
||||
x86_msi.teardown_msi_irq = xen_teardown_msi_irq;
|
||||
#endif
|
||||
xen_setup_acpi_sci();
|
||||
__acpi_register_gsi = acpi_register_gsi_xen;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
void __init xen_setup_pirqs(void)
|
||||
{
|
||||
int irq;
|
||||
|
||||
pci_xen_initial_domain();
|
||||
|
||||
if (0 == nr_ioapics) {
|
||||
for (irq = 0; irq < NR_IRQS_LEGACY; irq++)
|
||||
xen_allocate_pirq(irq, 0, "xt-pic");
|
||||
return;
|
||||
}
|
||||
|
||||
/* Pre-allocate legacy irqs */
|
||||
for (irq = 0; irq < NR_IRQS_LEGACY; irq++) {
|
||||
int trigger, polarity;
|
||||
|
||||
if (acpi_get_override_irq(irq, &trigger, &polarity) == -1)
|
||||
continue;
|
||||
|
||||
xen_register_pirq(irq,
|
||||
trigger ? ACPI_LEVEL_SENSITIVE : ACPI_EDGE_SENSITIVE);
|
||||
}
|
||||
}
|
||||
#endif
|
|
@ -13,6 +13,16 @@ config XEN
|
|||
kernel to boot in a paravirtualized environment under the
|
||||
Xen hypervisor.
|
||||
|
||||
config XEN_DOM0
|
||||
def_bool y
|
||||
depends on XEN && PCI_XEN && SWIOTLB_XEN
|
||||
depends on X86_LOCAL_APIC && X86_IO_APIC && ACPI && PCI
|
||||
|
||||
# Dummy symbol since people have come to rely on the PRIVILEGED_GUEST
|
||||
# name in tools.
|
||||
config XEN_PRIVILEGED_GUEST
|
||||
def_bool XEN_DOM0
|
||||
|
||||
config XEN_PVHVM
|
||||
def_bool y
|
||||
depends on XEN
|
||||
|
|
|
@ -46,6 +46,7 @@
|
|||
#include <asm/paravirt.h>
|
||||
#include <asm/apic.h>
|
||||
#include <asm/page.h>
|
||||
#include <asm/xen/pci.h>
|
||||
#include <asm/xen/hypercall.h>
|
||||
#include <asm/xen/hypervisor.h>
|
||||
#include <asm/fixmap.h>
|
||||
|
@ -236,6 +237,7 @@ static __init void xen_init_cpuid_mask(void)
|
|||
cpuid_leaf1_edx_mask =
|
||||
~((1 << X86_FEATURE_MCE) | /* disable MCE */
|
||||
(1 << X86_FEATURE_MCA) | /* disable MCA */
|
||||
(1 << X86_FEATURE_MTRR) | /* disable MTRR */
|
||||
(1 << X86_FEATURE_ACC)); /* thermal monitoring */
|
||||
|
||||
if (!xen_initial_domain())
|
||||
|
@ -1184,6 +1186,7 @@ asmlinkage void __init xen_start_kernel(void)
|
|||
|
||||
xen_raw_console_write("mapping kernel into physical memory\n");
|
||||
pgd = xen_setup_kernel_pagetable(pgd, xen_start_info->nr_pages);
|
||||
xen_ident_map_ISA();
|
||||
|
||||
/* Allocate and initialize top and mid mfn levels for p2m structure */
|
||||
xen_build_mfn_list_list();
|
||||
|
@ -1222,6 +1225,8 @@ asmlinkage void __init xen_start_kernel(void)
|
|||
add_preferred_console("xenboot", 0, NULL);
|
||||
add_preferred_console("tty", 0, NULL);
|
||||
add_preferred_console("hvc", 0, NULL);
|
||||
if (pci_xen)
|
||||
x86_init.pci.arch_init = pci_xen_init;
|
||||
} else {
|
||||
/* Make sure ACS will be enabled */
|
||||
pci_request_acs();
|
||||
|
|
|
@ -1975,6 +1975,7 @@ static void *m2v(phys_addr_t maddr)
|
|||
return __ka(m2p(maddr));
|
||||
}
|
||||
|
||||
/* Set the page permissions on an identity-mapped pages */
|
||||
static void set_page_prot(void *addr, pgprot_t prot)
|
||||
{
|
||||
unsigned long pfn = __pa(addr) >> PAGE_SHIFT;
|
||||
|
@ -2159,6 +2160,8 @@ __init pgd_t *xen_setup_kernel_pagetable(pgd_t *pgd,
|
|||
}
|
||||
#endif /* CONFIG_X86_64 */
|
||||
|
||||
static unsigned char dummy_mapping[PAGE_SIZE] __page_aligned_bss;
|
||||
|
||||
static void xen_set_fixmap(unsigned idx, phys_addr_t phys, pgprot_t prot)
|
||||
{
|
||||
pte_t pte;
|
||||
|
@ -2178,9 +2181,6 @@ static void xen_set_fixmap(unsigned idx, phys_addr_t phys, pgprot_t prot)
|
|||
# endif
|
||||
#else
|
||||
case VSYSCALL_LAST_PAGE ... VSYSCALL_FIRST_PAGE:
|
||||
#endif
|
||||
#ifdef CONFIG_X86_LOCAL_APIC
|
||||
case FIX_APIC_BASE: /* maps dummy local APIC */
|
||||
#endif
|
||||
case FIX_TEXT_POKE0:
|
||||
case FIX_TEXT_POKE1:
|
||||
|
@ -2188,6 +2188,22 @@ static void xen_set_fixmap(unsigned idx, phys_addr_t phys, pgprot_t prot)
|
|||
pte = pfn_pte(phys, prot);
|
||||
break;
|
||||
|
||||
#ifdef CONFIG_X86_LOCAL_APIC
|
||||
case FIX_APIC_BASE: /* maps dummy local APIC */
|
||||
pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL);
|
||||
break;
|
||||
#endif
|
||||
|
||||
#ifdef CONFIG_X86_IO_APIC
|
||||
case FIX_IO_APIC_BASE_0 ... FIX_IO_APIC_BASE_END:
|
||||
/*
|
||||
* We just don't map the IO APIC - all access is via
|
||||
* hypercalls. Keep the address in the pte for reference.
|
||||
*/
|
||||
pte = pfn_pte(PFN_DOWN(__pa(dummy_mapping)), PAGE_KERNEL);
|
||||
break;
|
||||
#endif
|
||||
|
||||
case FIX_PARAVIRT_BOOTMAP:
|
||||
/* This is an MFN, but it isn't an IO mapping from the
|
||||
IO domain */
|
||||
|
@ -2212,6 +2228,29 @@ static void xen_set_fixmap(unsigned idx, phys_addr_t phys, pgprot_t prot)
|
|||
#endif
|
||||
}
|
||||
|
||||
__init void xen_ident_map_ISA(void)
|
||||
{
|
||||
unsigned long pa;
|
||||
|
||||
/*
|
||||
* If we're dom0, then linear map the ISA machine addresses into
|
||||
* the kernel's address space.
|
||||
*/
|
||||
if (!xen_initial_domain())
|
||||
return;
|
||||
|
||||
xen_raw_printk("Xen: setup ISA identity maps\n");
|
||||
|
||||
for (pa = ISA_START_ADDRESS; pa < ISA_END_ADDRESS; pa += PAGE_SIZE) {
|
||||
pte_t pte = mfn_pte(PFN_DOWN(pa), PAGE_KERNEL_IO);
|
||||
|
||||
if (HYPERVISOR_update_va_mapping(PAGE_OFFSET + pa, pte, 0))
|
||||
BUG();
|
||||
}
|
||||
|
||||
xen_flush_tlb();
|
||||
}
|
||||
|
||||
static __init void xen_post_allocator_init(void)
|
||||
{
|
||||
pv_mmu_ops.set_pte = xen_set_pte;
|
||||
|
@ -2320,6 +2359,8 @@ void __init xen_init_mmu_ops(void)
|
|||
pv_mmu_ops = xen_mmu_ops;
|
||||
|
||||
vmap_lazy_unmap = false;
|
||||
|
||||
memset(dummy_mapping, 0xff, PAGE_SIZE);
|
||||
}
|
||||
|
||||
/* Protected by xen_reservation_lock. */
|
||||
|
|
|
@ -1,6 +1,7 @@
|
|||
/* Glue code to lib/swiotlb-xen.c */
|
||||
|
||||
#include <linux/dma-mapping.h>
|
||||
#include <linux/pci.h>
|
||||
#include <xen/swiotlb-xen.h>
|
||||
|
||||
#include <asm/xen/hypervisor.h>
|
||||
|
@ -55,6 +56,9 @@ void __init pci_xen_swiotlb_init(void)
|
|||
if (xen_swiotlb) {
|
||||
xen_swiotlb_init(1);
|
||||
dma_ops = &xen_swiotlb_dma_ops;
|
||||
|
||||
/* Make sure ACS will be enabled */
|
||||
pci_request_acs();
|
||||
}
|
||||
}
|
||||
IOMMU_INIT_FINISH(pci_xen_swiotlb_detect,
|
||||
|
|
|
@ -204,6 +204,9 @@ char * __init xen_memory_setup(void)
|
|||
* Even though this is normal, usable memory under Xen, reserve
|
||||
* ISA memory anyway because too many things think they can poke
|
||||
* about in there.
|
||||
*
|
||||
* In a dom0 kernel, this region is identity mapped with the
|
||||
* hardware ISA area, so it really is out of bounds.
|
||||
*/
|
||||
e820_add_region(ISA_START_ADDRESS, ISA_END_ADDRESS - ISA_START_ADDRESS,
|
||||
E820_RESERVED);
|
||||
|
@ -367,7 +370,5 @@ void __init xen_arch_setup(void)
|
|||
|
||||
pm_idle = xen_idle;
|
||||
|
||||
paravirt_disable_iospace();
|
||||
|
||||
fiddle_vdso();
|
||||
}
|
||||
|
|
|
@ -28,6 +28,7 @@
|
|||
#include <asm/xen/interface.h>
|
||||
#include <asm/xen/hypercall.h>
|
||||
|
||||
#include <xen/xen.h>
|
||||
#include <xen/page.h>
|
||||
#include <xen/events.h>
|
||||
|
||||
|
@ -156,6 +157,9 @@ static void __init xen_fill_possible_map(void)
|
|||
{
|
||||
int i, rc;
|
||||
|
||||
if (xen_initial_domain())
|
||||
return;
|
||||
|
||||
for (i = 0; i < nr_cpu_ids; i++) {
|
||||
rc = HYPERVISOR_vcpu_op(VCPUOP_is_up, i, NULL);
|
||||
if (rc >= 0) {
|
||||
|
@ -165,6 +169,27 @@ static void __init xen_fill_possible_map(void)
|
|||
}
|
||||
}
|
||||
|
||||
static void __init xen_filter_cpu_maps(void)
|
||||
{
|
||||
int i, rc;
|
||||
|
||||
if (!xen_initial_domain())
|
||||
return;
|
||||
|
||||
num_processors = 0;
|
||||
disabled_cpus = 0;
|
||||
for (i = 0; i < nr_cpu_ids; i++) {
|
||||
rc = HYPERVISOR_vcpu_op(VCPUOP_is_up, i, NULL);
|
||||
if (rc >= 0) {
|
||||
num_processors++;
|
||||
set_cpu_possible(i, true);
|
||||
} else {
|
||||
set_cpu_possible(i, false);
|
||||
set_cpu_present(i, false);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
static void __init xen_smp_prepare_boot_cpu(void)
|
||||
{
|
||||
BUG_ON(smp_processor_id() != 0);
|
||||
|
@ -174,6 +199,7 @@ static void __init xen_smp_prepare_boot_cpu(void)
|
|||
old memory can be recycled */
|
||||
make_lowmem_page_readwrite(xen_initial_gdt);
|
||||
|
||||
xen_filter_cpu_maps();
|
||||
xen_setup_vcpu_info_placement();
|
||||
}
|
||||
|
||||
|
|
|
@ -1112,6 +1112,8 @@ static void blkback_changed(struct xenbus_device *dev,
|
|||
case XenbusStateInitialising:
|
||||
case XenbusStateInitWait:
|
||||
case XenbusStateInitialised:
|
||||
case XenbusStateReconfiguring:
|
||||
case XenbusStateReconfigured:
|
||||
case XenbusStateUnknown:
|
||||
case XenbusStateClosed:
|
||||
break;
|
||||
|
|
|
@ -79,7 +79,7 @@ static int __write_console(const char *data, int len)
|
|||
return sent;
|
||||
}
|
||||
|
||||
static int write_console(uint32_t vtermno, const char *data, int len)
|
||||
static int domU_write_console(uint32_t vtermno, const char *data, int len)
|
||||
{
|
||||
int ret = len;
|
||||
|
||||
|
@ -102,7 +102,7 @@ static int write_console(uint32_t vtermno, const char *data, int len)
|
|||
return ret;
|
||||
}
|
||||
|
||||
static int read_console(uint32_t vtermno, char *buf, int len)
|
||||
static int domU_read_console(uint32_t vtermno, char *buf, int len)
|
||||
{
|
||||
struct xencons_interface *intf = xencons_interface();
|
||||
XENCONS_RING_IDX cons, prod;
|
||||
|
@ -123,28 +123,62 @@ static int read_console(uint32_t vtermno, char *buf, int len)
|
|||
return recv;
|
||||
}
|
||||
|
||||
static const struct hv_ops hvc_ops = {
|
||||
.get_chars = read_console,
|
||||
.put_chars = write_console,
|
||||
static struct hv_ops domU_hvc_ops = {
|
||||
.get_chars = domU_read_console,
|
||||
.put_chars = domU_write_console,
|
||||
.notifier_add = notifier_add_irq,
|
||||
.notifier_del = notifier_del_irq,
|
||||
.notifier_hangup = notifier_hangup_irq,
|
||||
};
|
||||
|
||||
static int __init xen_init(void)
|
||||
static int dom0_read_console(uint32_t vtermno, char *buf, int len)
|
||||
{
|
||||
return HYPERVISOR_console_io(CONSOLEIO_read, len, buf);
|
||||
}
|
||||
|
||||
/*
|
||||
* Either for a dom0 to write to the system console, or a domU with a
|
||||
* debug version of Xen
|
||||
*/
|
||||
static int dom0_write_console(uint32_t vtermno, const char *str, int len)
|
||||
{
|
||||
int rc = HYPERVISOR_console_io(CONSOLEIO_write, len, (char *)str);
|
||||
if (rc < 0)
|
||||
return 0;
|
||||
|
||||
return len;
|
||||
}
|
||||
|
||||
static struct hv_ops dom0_hvc_ops = {
|
||||
.get_chars = dom0_read_console,
|
||||
.put_chars = dom0_write_console,
|
||||
.notifier_add = notifier_add_irq,
|
||||
.notifier_del = notifier_del_irq,
|
||||
.notifier_hangup = notifier_hangup_irq,
|
||||
};
|
||||
|
||||
static int __init xen_hvc_init(void)
|
||||
{
|
||||
struct hvc_struct *hp;
|
||||
struct hv_ops *ops;
|
||||
|
||||
if (!xen_pv_domain() ||
|
||||
xen_initial_domain() ||
|
||||
!xen_start_info->console.domU.evtchn)
|
||||
if (!xen_pv_domain())
|
||||
return -ENODEV;
|
||||
|
||||
xencons_irq = bind_evtchn_to_irq(xen_start_info->console.domU.evtchn);
|
||||
if (xen_initial_domain()) {
|
||||
ops = &dom0_hvc_ops;
|
||||
xencons_irq = bind_virq_to_irq(VIRQ_CONSOLE, 0);
|
||||
} else {
|
||||
if (!xen_start_info->console.domU.evtchn)
|
||||
return -ENODEV;
|
||||
|
||||
ops = &domU_hvc_ops;
|
||||
xencons_irq = bind_evtchn_to_irq(xen_start_info->console.domU.evtchn);
|
||||
}
|
||||
if (xencons_irq < 0)
|
||||
xencons_irq = 0; /* NO_IRQ */
|
||||
|
||||
hp = hvc_alloc(HVC_COOKIE, xencons_irq, &hvc_ops, 256);
|
||||
hp = hvc_alloc(HVC_COOKIE, xencons_irq, ops, 256);
|
||||
if (IS_ERR(hp))
|
||||
return PTR_ERR(hp);
|
||||
|
||||
|
@ -161,7 +195,7 @@ void xen_console_resume(void)
|
|||
rebind_evtchn_irq(xen_start_info->console.domU.evtchn, xencons_irq);
|
||||
}
|
||||
|
||||
static void __exit xen_fini(void)
|
||||
static void __exit xen_hvc_fini(void)
|
||||
{
|
||||
if (hvc)
|
||||
hvc_remove(hvc);
|
||||
|
@ -169,29 +203,24 @@ static void __exit xen_fini(void)
|
|||
|
||||
static int xen_cons_init(void)
|
||||
{
|
||||
struct hv_ops *ops;
|
||||
|
||||
if (!xen_pv_domain())
|
||||
return 0;
|
||||
|
||||
hvc_instantiate(HVC_COOKIE, 0, &hvc_ops);
|
||||
if (xen_initial_domain())
|
||||
ops = &dom0_hvc_ops;
|
||||
else
|
||||
ops = &domU_hvc_ops;
|
||||
|
||||
hvc_instantiate(HVC_COOKIE, 0, ops);
|
||||
return 0;
|
||||
}
|
||||
|
||||
module_init(xen_init);
|
||||
module_exit(xen_fini);
|
||||
module_init(xen_hvc_init);
|
||||
module_exit(xen_hvc_fini);
|
||||
console_initcall(xen_cons_init);
|
||||
|
||||
static void raw_console_write(const char *str, int len)
|
||||
{
|
||||
while(len > 0) {
|
||||
int rc = HYPERVISOR_console_io(CONSOLEIO_write, len, (char *)str);
|
||||
if (rc <= 0)
|
||||
break;
|
||||
|
||||
str += rc;
|
||||
len -= rc;
|
||||
}
|
||||
}
|
||||
|
||||
#ifdef CONFIG_EARLY_PRINTK
|
||||
static void xenboot_write_console(struct console *console, const char *string,
|
||||
unsigned len)
|
||||
|
@ -199,19 +228,22 @@ static void xenboot_write_console(struct console *console, const char *string,
|
|||
unsigned int linelen, off = 0;
|
||||
const char *pos;
|
||||
|
||||
raw_console_write(string, len);
|
||||
dom0_write_console(0, string, len);
|
||||
|
||||
write_console(0, "(early) ", 8);
|
||||
if (xen_initial_domain())
|
||||
return;
|
||||
|
||||
domU_write_console(0, "(early) ", 8);
|
||||
while (off < len && NULL != (pos = strchr(string+off, '\n'))) {
|
||||
linelen = pos-string+off;
|
||||
if (off + linelen > len)
|
||||
break;
|
||||
write_console(0, string+off, linelen);
|
||||
write_console(0, "\r\n", 2);
|
||||
domU_write_console(0, string+off, linelen);
|
||||
domU_write_console(0, "\r\n", 2);
|
||||
off += linelen + 1;
|
||||
}
|
||||
if (off < len)
|
||||
write_console(0, string+off, len-off);
|
||||
domU_write_console(0, string+off, len-off);
|
||||
}
|
||||
|
||||
struct console xenboot_console = {
|
||||
|
@ -223,7 +255,7 @@ struct console xenboot_console = {
|
|||
|
||||
void xen_raw_console_write(const char *str)
|
||||
{
|
||||
raw_console_write(str, strlen(str));
|
||||
dom0_write_console(0, str, strlen(str));
|
||||
}
|
||||
|
||||
void xen_raw_printk(const char *fmt, ...)
|
||||
|
|
|
@ -276,6 +276,8 @@ static void xenkbd_backend_changed(struct xenbus_device *dev,
|
|||
switch (backend_state) {
|
||||
case XenbusStateInitialising:
|
||||
case XenbusStateInitialised:
|
||||
case XenbusStateReconfiguring:
|
||||
case XenbusStateReconfigured:
|
||||
case XenbusStateUnknown:
|
||||
case XenbusStateClosed:
|
||||
break;
|
||||
|
|
|
@ -1610,6 +1610,8 @@ static void netback_changed(struct xenbus_device *dev,
|
|||
switch (backend_state) {
|
||||
case XenbusStateInitialising:
|
||||
case XenbusStateInitialised:
|
||||
case XenbusStateReconfiguring:
|
||||
case XenbusStateReconfigured:
|
||||
case XenbusStateConnected:
|
||||
case XenbusStateUnknown:
|
||||
case XenbusStateClosed:
|
||||
|
|
|
@ -40,6 +40,27 @@ config PCI_STUB
|
|||
|
||||
When in doubt, say N.
|
||||
|
||||
config XEN_PCIDEV_FRONTEND
|
||||
tristate "Xen PCI Frontend"
|
||||
depends on PCI && X86 && XEN
|
||||
select HOTPLUG
|
||||
select PCI_XEN
|
||||
default y
|
||||
help
|
||||
The PCI device frontend driver allows the kernel to import arbitrary
|
||||
PCI devices from a PCI backend to support PCI driver domains.
|
||||
|
||||
config XEN_PCIDEV_FE_DEBUG
|
||||
bool "Xen PCI Frontend debugging"
|
||||
depends on XEN_PCIDEV_FRONTEND && PCI_DEBUG
|
||||
help
|
||||
Say Y here if you want the Xen PCI frontend to produce a bunch of debug
|
||||
messages to the system log. Select this if you are having a
|
||||
problem with Xen PCI frontend support and want to see more of what is
|
||||
going on.
|
||||
|
||||
When in doubt, say N.
|
||||
|
||||
config HT_IRQ
|
||||
bool "Interrupts on hypertransport devices"
|
||||
default y
|
||||
|
|
|
@ -65,4 +65,6 @@ obj-$(CONFIG_PCI_SYSCALL) += syscall.o
|
|||
|
||||
obj-$(CONFIG_PCI_STUB) += pci-stub.o
|
||||
|
||||
obj-$(CONFIG_XEN_PCIDEV_FRONTEND) += xen-pcifront.o
|
||||
|
||||
ccflags-$(CONFIG_PCI_DEBUG) := -DDEBUG
|
||||
|
|
|
@ -342,6 +342,7 @@ void pci_walk_bus(struct pci_bus *top, int (*cb)(struct pci_dev *, void *),
|
|||
}
|
||||
up_read(&pci_bus_sem);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(pci_walk_bus);
|
||||
|
||||
EXPORT_SYMBOL(pci_bus_alloc_resource);
|
||||
EXPORT_SYMBOL_GPL(pci_bus_add_device);
|
||||
|
|
|
@ -35,7 +35,12 @@ int arch_msi_check_device(struct pci_dev *dev, int nvec, int type)
|
|||
#endif
|
||||
|
||||
#ifndef arch_setup_msi_irqs
|
||||
int arch_setup_msi_irqs(struct pci_dev *dev, int nvec, int type)
|
||||
# define arch_setup_msi_irqs default_setup_msi_irqs
|
||||
# define HAVE_DEFAULT_MSI_SETUP_IRQS
|
||||
#endif
|
||||
|
||||
#ifdef HAVE_DEFAULT_MSI_SETUP_IRQS
|
||||
int default_setup_msi_irqs(struct pci_dev *dev, int nvec, int type)
|
||||
{
|
||||
struct msi_desc *entry;
|
||||
int ret;
|
||||
|
@ -60,7 +65,12 @@ int arch_setup_msi_irqs(struct pci_dev *dev, int nvec, int type)
|
|||
#endif
|
||||
|
||||
#ifndef arch_teardown_msi_irqs
|
||||
void arch_teardown_msi_irqs(struct pci_dev *dev)
|
||||
# define arch_teardown_msi_irqs default_teardown_msi_irqs
|
||||
# define HAVE_DEFAULT_MSI_TEARDOWN_IRQS
|
||||
#endif
|
||||
|
||||
#ifdef HAVE_DEFAULT_MSI_TEARDOWN_IRQS
|
||||
void default_teardown_msi_irqs(struct pci_dev *dev)
|
||||
{
|
||||
struct msi_desc *entry;
|
||||
|
||||
|
|
File diff suppressed because it is too large
Load Diff
|
@ -631,6 +631,8 @@ static void xenfb_backend_changed(struct xenbus_device *dev,
|
|||
switch (backend_state) {
|
||||
case XenbusStateInitialising:
|
||||
case XenbusStateInitialised:
|
||||
case XenbusStateReconfiguring:
|
||||
case XenbusStateReconfigured:
|
||||
case XenbusStateUnknown:
|
||||
case XenbusStateClosed:
|
||||
break;
|
||||
|
|
|
@ -74,6 +74,7 @@ config XEN_PLATFORM_PCI
|
|||
|
||||
config SWIOTLB_XEN
|
||||
def_bool y
|
||||
depends on SWIOTLB
|
||||
depends on PCI
|
||||
select SWIOTLB
|
||||
|
||||
endmenu
|
||||
|
|
|
@ -4,6 +4,7 @@ obj-y += xenbus/
|
|||
nostackp := $(call cc-option, -fno-stack-protector)
|
||||
CFLAGS_features.o := $(nostackp)
|
||||
|
||||
obj-$(CONFIG_BLOCK) += biomerge.o
|
||||
obj-$(CONFIG_HOTPLUG_CPU) += cpu_hotplug.o
|
||||
obj-$(CONFIG_XEN_XENCOMM) += xencomm.o
|
||||
obj-$(CONFIG_XEN_BALLOON) += balloon.o
|
||||
|
@ -12,3 +13,4 @@ obj-$(CONFIG_XENFS) += xenfs/
|
|||
obj-$(CONFIG_XEN_SYS_HYPERVISOR) += sys-hypervisor.o
|
||||
obj-$(CONFIG_XEN_PLATFORM_PCI) += platform-pci.o
|
||||
obj-$(CONFIG_SWIOTLB_XEN) += swiotlb-xen.o
|
||||
obj-$(CONFIG_XEN_DOM0) += pci.o
|
||||
|
|
|
@ -0,0 +1,13 @@
|
|||
#include <linux/bio.h>
|
||||
#include <linux/io.h>
|
||||
#include <xen/page.h>
|
||||
|
||||
bool xen_biovec_phys_mergeable(const struct bio_vec *vec1,
|
||||
const struct bio_vec *vec2)
|
||||
{
|
||||
unsigned long mfn1 = pfn_to_mfn(page_to_pfn(vec1->bv_page));
|
||||
unsigned long mfn2 = pfn_to_mfn(page_to_pfn(vec2->bv_page));
|
||||
|
||||
return __BIOVEC_PHYS_MERGEABLE(vec1, vec2) &&
|
||||
((mfn1 == mfn2) || ((mfn1+1) == mfn2));
|
||||
}
|
|
@ -16,7 +16,7 @@
|
|||
* (typically dom0).
|
||||
* 2. VIRQs, typically used for timers. These are per-cpu events.
|
||||
* 3. IPIs.
|
||||
* 4. Hardware interrupts. Not supported at present.
|
||||
* 4. PIRQs - Hardware interrupts.
|
||||
*
|
||||
* Jeremy Fitzhardinge <jeremy@xensource.com>, XenSource Inc, 2007
|
||||
*/
|
||||
|
@ -28,12 +28,16 @@
|
|||
#include <linux/string.h>
|
||||
#include <linux/bootmem.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/irqnr.h>
|
||||
#include <linux/pci.h>
|
||||
|
||||
#include <asm/desc.h>
|
||||
#include <asm/ptrace.h>
|
||||
#include <asm/irq.h>
|
||||
#include <asm/idle.h>
|
||||
#include <asm/io_apic.h>
|
||||
#include <asm/sync_bitops.h>
|
||||
#include <asm/xen/pci.h>
|
||||
#include <asm/xen/hypercall.h>
|
||||
#include <asm/xen/hypervisor.h>
|
||||
|
||||
|
@ -73,7 +77,8 @@ enum xen_irq_type {
|
|||
* event channel - irq->event channel mapping
|
||||
* cpu - cpu this event channel is bound to
|
||||
* index - type-specific information:
|
||||
* PIRQ - vector, with MSB being "needs EIO"
|
||||
* PIRQ - vector, with MSB being "needs EIO", or physical IRQ of the HVM
|
||||
* guest, or GSI (real passthrough IRQ) of the device.
|
||||
* VIRQ - virq number
|
||||
* IPI - IPI vector
|
||||
* EVTCHN -
|
||||
|
@ -88,21 +93,30 @@ struct irq_info
|
|||
unsigned short virq;
|
||||
enum ipi_vector ipi;
|
||||
struct {
|
||||
unsigned short pirq;
|
||||
unsigned short gsi;
|
||||
unsigned short vector;
|
||||
unsigned char vector;
|
||||
unsigned char flags;
|
||||
} pirq;
|
||||
} u;
|
||||
};
|
||||
#define PIRQ_NEEDS_EOI (1 << 0)
|
||||
#define PIRQ_SHAREABLE (1 << 1)
|
||||
|
||||
static struct irq_info irq_info[NR_IRQS];
|
||||
static struct irq_info *irq_info;
|
||||
static int *pirq_to_irq;
|
||||
static int nr_pirqs;
|
||||
|
||||
static int evtchn_to_irq[NR_EVENT_CHANNELS] = {
|
||||
[0 ... NR_EVENT_CHANNELS-1] = -1
|
||||
};
|
||||
static int *evtchn_to_irq;
|
||||
struct cpu_evtchn_s {
|
||||
unsigned long bits[NR_EVENT_CHANNELS/BITS_PER_LONG];
|
||||
};
|
||||
static struct cpu_evtchn_s *cpu_evtchn_mask_p;
|
||||
|
||||
static __initdata struct cpu_evtchn_s init_evtchn_mask = {
|
||||
.bits[0 ... (NR_EVENT_CHANNELS/BITS_PER_LONG)-1] = ~0ul,
|
||||
};
|
||||
static struct cpu_evtchn_s *cpu_evtchn_mask_p = &init_evtchn_mask;
|
||||
|
||||
static inline unsigned long *cpu_evtchn_mask(int cpu)
|
||||
{
|
||||
return cpu_evtchn_mask_p[cpu].bits;
|
||||
|
@ -113,6 +127,7 @@ static inline unsigned long *cpu_evtchn_mask(int cpu)
|
|||
|
||||
static struct irq_chip xen_dynamic_chip;
|
||||
static struct irq_chip xen_percpu_chip;
|
||||
static struct irq_chip xen_pirq_chip;
|
||||
|
||||
/* Constructor for packed IRQ information. */
|
||||
static struct irq_info mk_unbound_info(void)
|
||||
|
@ -138,11 +153,12 @@ static struct irq_info mk_virq_info(unsigned short evtchn, unsigned short virq)
|
|||
.cpu = 0, .u.virq = virq };
|
||||
}
|
||||
|
||||
static struct irq_info mk_pirq_info(unsigned short evtchn,
|
||||
static struct irq_info mk_pirq_info(unsigned short evtchn, unsigned short pirq,
|
||||
unsigned short gsi, unsigned short vector)
|
||||
{
|
||||
return (struct irq_info) { .type = IRQT_PIRQ, .evtchn = evtchn,
|
||||
.cpu = 0, .u.pirq = { .gsi = gsi, .vector = vector } };
|
||||
.cpu = 0,
|
||||
.u.pirq = { .pirq = pirq, .gsi = gsi, .vector = vector } };
|
||||
}
|
||||
|
||||
/*
|
||||
|
@ -184,6 +200,16 @@ static unsigned virq_from_irq(unsigned irq)
|
|||
return info->u.virq;
|
||||
}
|
||||
|
||||
static unsigned pirq_from_irq(unsigned irq)
|
||||
{
|
||||
struct irq_info *info = info_for_irq(irq);
|
||||
|
||||
BUG_ON(info == NULL);
|
||||
BUG_ON(info->type != IRQT_PIRQ);
|
||||
|
||||
return info->u.pirq.pirq;
|
||||
}
|
||||
|
||||
static unsigned gsi_from_irq(unsigned irq)
|
||||
{
|
||||
struct irq_info *info = info_for_irq(irq);
|
||||
|
@ -225,6 +251,15 @@ static unsigned int cpu_from_evtchn(unsigned int evtchn)
|
|||
return ret;
|
||||
}
|
||||
|
||||
static bool pirq_needs_eoi(unsigned irq)
|
||||
{
|
||||
struct irq_info *info = info_for_irq(irq);
|
||||
|
||||
BUG_ON(info->type != IRQT_PIRQ);
|
||||
|
||||
return info->u.pirq.flags & PIRQ_NEEDS_EOI;
|
||||
}
|
||||
|
||||
static inline unsigned long active_evtchns(unsigned int cpu,
|
||||
struct shared_info *sh,
|
||||
unsigned int idx)
|
||||
|
@ -336,12 +371,40 @@ static void unmask_evtchn(int port)
|
|||
put_cpu();
|
||||
}
|
||||
|
||||
static int get_nr_hw_irqs(void)
|
||||
{
|
||||
int ret = 1;
|
||||
|
||||
#ifdef CONFIG_X86_IO_APIC
|
||||
ret = get_nr_irqs_gsi();
|
||||
#endif
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
/* callers of this function should make sure that PHYSDEVOP_get_nr_pirqs
|
||||
* succeeded otherwise nr_pirqs won't hold the right value */
|
||||
static int find_unbound_pirq(void)
|
||||
{
|
||||
int i;
|
||||
for (i = nr_pirqs-1; i >= 0; i--) {
|
||||
if (pirq_to_irq[i] < 0)
|
||||
return i;
|
||||
}
|
||||
return -1;
|
||||
}
|
||||
|
||||
static int find_unbound_irq(void)
|
||||
{
|
||||
struct irq_data *data;
|
||||
int irq, res;
|
||||
int start = get_nr_hw_irqs();
|
||||
|
||||
for (irq = 0; irq < nr_irqs; irq++) {
|
||||
if (start == nr_irqs)
|
||||
goto no_irqs;
|
||||
|
||||
/* nr_irqs is a magic value. Must not use it.*/
|
||||
for (irq = nr_irqs-1; irq > start; irq--) {
|
||||
data = irq_get_irq_data(irq);
|
||||
/* only 0->15 have init'd desc; handle irq > 16 */
|
||||
if (!data)
|
||||
|
@ -354,8 +417,8 @@ static int find_unbound_irq(void)
|
|||
return irq;
|
||||
}
|
||||
|
||||
if (irq == nr_irqs)
|
||||
panic("No available IRQ to bind to: increase nr_irqs!\n");
|
||||
if (irq == start)
|
||||
goto no_irqs;
|
||||
|
||||
res = irq_alloc_desc_at(irq, 0);
|
||||
|
||||
|
@ -363,6 +426,357 @@ static int find_unbound_irq(void)
|
|||
return -1;
|
||||
|
||||
return irq;
|
||||
|
||||
no_irqs:
|
||||
panic("No available IRQ to bind to: increase nr_irqs!\n");
|
||||
}
|
||||
|
||||
static bool identity_mapped_irq(unsigned irq)
|
||||
{
|
||||
/* identity map all the hardware irqs */
|
||||
return irq < get_nr_hw_irqs();
|
||||
}
|
||||
|
||||
static void pirq_unmask_notify(int irq)
|
||||
{
|
||||
struct physdev_eoi eoi = { .irq = pirq_from_irq(irq) };
|
||||
|
||||
if (unlikely(pirq_needs_eoi(irq))) {
|
||||
int rc = HYPERVISOR_physdev_op(PHYSDEVOP_eoi, &eoi);
|
||||
WARN_ON(rc);
|
||||
}
|
||||
}
|
||||
|
||||
static void pirq_query_unmask(int irq)
|
||||
{
|
||||
struct physdev_irq_status_query irq_status;
|
||||
struct irq_info *info = info_for_irq(irq);
|
||||
|
||||
BUG_ON(info->type != IRQT_PIRQ);
|
||||
|
||||
irq_status.irq = pirq_from_irq(irq);
|
||||
if (HYPERVISOR_physdev_op(PHYSDEVOP_irq_status_query, &irq_status))
|
||||
irq_status.flags = 0;
|
||||
|
||||
info->u.pirq.flags &= ~PIRQ_NEEDS_EOI;
|
||||
if (irq_status.flags & XENIRQSTAT_needs_eoi)
|
||||
info->u.pirq.flags |= PIRQ_NEEDS_EOI;
|
||||
}
|
||||
|
||||
static bool probing_irq(int irq)
|
||||
{
|
||||
struct irq_desc *desc = irq_to_desc(irq);
|
||||
|
||||
return desc && desc->action == NULL;
|
||||
}
|
||||
|
||||
static unsigned int startup_pirq(unsigned int irq)
|
||||
{
|
||||
struct evtchn_bind_pirq bind_pirq;
|
||||
struct irq_info *info = info_for_irq(irq);
|
||||
int evtchn = evtchn_from_irq(irq);
|
||||
int rc;
|
||||
|
||||
BUG_ON(info->type != IRQT_PIRQ);
|
||||
|
||||
if (VALID_EVTCHN(evtchn))
|
||||
goto out;
|
||||
|
||||
bind_pirq.pirq = pirq_from_irq(irq);
|
||||
/* NB. We are happy to share unless we are probing. */
|
||||
bind_pirq.flags = info->u.pirq.flags & PIRQ_SHAREABLE ?
|
||||
BIND_PIRQ__WILL_SHARE : 0;
|
||||
rc = HYPERVISOR_event_channel_op(EVTCHNOP_bind_pirq, &bind_pirq);
|
||||
if (rc != 0) {
|
||||
if (!probing_irq(irq))
|
||||
printk(KERN_INFO "Failed to obtain physical IRQ %d\n",
|
||||
irq);
|
||||
return 0;
|
||||
}
|
||||
evtchn = bind_pirq.port;
|
||||
|
||||
pirq_query_unmask(irq);
|
||||
|
||||
evtchn_to_irq[evtchn] = irq;
|
||||
bind_evtchn_to_cpu(evtchn, 0);
|
||||
info->evtchn = evtchn;
|
||||
|
||||
out:
|
||||
unmask_evtchn(evtchn);
|
||||
pirq_unmask_notify(irq);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void shutdown_pirq(unsigned int irq)
|
||||
{
|
||||
struct evtchn_close close;
|
||||
struct irq_info *info = info_for_irq(irq);
|
||||
int evtchn = evtchn_from_irq(irq);
|
||||
|
||||
BUG_ON(info->type != IRQT_PIRQ);
|
||||
|
||||
if (!VALID_EVTCHN(evtchn))
|
||||
return;
|
||||
|
||||
mask_evtchn(evtchn);
|
||||
|
||||
close.port = evtchn;
|
||||
if (HYPERVISOR_event_channel_op(EVTCHNOP_close, &close) != 0)
|
||||
BUG();
|
||||
|
||||
bind_evtchn_to_cpu(evtchn, 0);
|
||||
evtchn_to_irq[evtchn] = -1;
|
||||
info->evtchn = 0;
|
||||
}
|
||||
|
||||
static void enable_pirq(unsigned int irq)
|
||||
{
|
||||
startup_pirq(irq);
|
||||
}
|
||||
|
||||
static void disable_pirq(unsigned int irq)
|
||||
{
|
||||
}
|
||||
|
||||
static void ack_pirq(unsigned int irq)
|
||||
{
|
||||
int evtchn = evtchn_from_irq(irq);
|
||||
|
||||
move_native_irq(irq);
|
||||
|
||||
if (VALID_EVTCHN(evtchn)) {
|
||||
mask_evtchn(evtchn);
|
||||
clear_evtchn(evtchn);
|
||||
}
|
||||
}
|
||||
|
||||
static void end_pirq(unsigned int irq)
|
||||
{
|
||||
int evtchn = evtchn_from_irq(irq);
|
||||
struct irq_desc *desc = irq_to_desc(irq);
|
||||
|
||||
if (WARN_ON(!desc))
|
||||
return;
|
||||
|
||||
if ((desc->status & (IRQ_DISABLED|IRQ_PENDING)) ==
|
||||
(IRQ_DISABLED|IRQ_PENDING)) {
|
||||
shutdown_pirq(irq);
|
||||
} else if (VALID_EVTCHN(evtchn)) {
|
||||
unmask_evtchn(evtchn);
|
||||
pirq_unmask_notify(irq);
|
||||
}
|
||||
}
|
||||
|
||||
static int find_irq_by_gsi(unsigned gsi)
|
||||
{
|
||||
int irq;
|
||||
|
||||
for (irq = 0; irq < nr_irqs; irq++) {
|
||||
struct irq_info *info = info_for_irq(irq);
|
||||
|
||||
if (info == NULL || info->type != IRQT_PIRQ)
|
||||
continue;
|
||||
|
||||
if (gsi_from_irq(irq) == gsi)
|
||||
return irq;
|
||||
}
|
||||
|
||||
return -1;
|
||||
}
|
||||
|
||||
int xen_allocate_pirq(unsigned gsi, int shareable, char *name)
|
||||
{
|
||||
return xen_map_pirq_gsi(gsi, gsi, shareable, name);
|
||||
}
|
||||
|
||||
/* xen_map_pirq_gsi might allocate irqs from the top down, as a
|
||||
* consequence don't assume that the irq number returned has a low value
|
||||
* or can be used as a pirq number unless you know otherwise.
|
||||
*
|
||||
* One notable exception is when xen_map_pirq_gsi is called passing an
|
||||
* hardware gsi as argument, in that case the irq number returned
|
||||
* matches the gsi number passed as second argument.
|
||||
*
|
||||
* Note: We don't assign an event channel until the irq actually started
|
||||
* up. Return an existing irq if we've already got one for the gsi.
|
||||
*/
|
||||
int xen_map_pirq_gsi(unsigned pirq, unsigned gsi, int shareable, char *name)
|
||||
{
|
||||
int irq = 0;
|
||||
struct physdev_irq irq_op;
|
||||
|
||||
spin_lock(&irq_mapping_update_lock);
|
||||
|
||||
if ((pirq > nr_pirqs) || (gsi > nr_irqs)) {
|
||||
printk(KERN_WARNING "xen_map_pirq_gsi: %s %s is incorrect!\n",
|
||||
pirq > nr_pirqs ? "nr_pirqs" :"",
|
||||
gsi > nr_irqs ? "nr_irqs" : "");
|
||||
goto out;
|
||||
}
|
||||
|
||||
irq = find_irq_by_gsi(gsi);
|
||||
if (irq != -1) {
|
||||
printk(KERN_INFO "xen_map_pirq_gsi: returning irq %d for gsi %u\n",
|
||||
irq, gsi);
|
||||
goto out; /* XXX need refcount? */
|
||||
}
|
||||
|
||||
/* If we are a PV guest, we don't have GSIs (no ACPI passed). Therefore
|
||||
* we are using the !xen_initial_domain() to drop in the function.*/
|
||||
if (identity_mapped_irq(gsi) || (!xen_initial_domain() &&
|
||||
xen_pv_domain())) {
|
||||
irq = gsi;
|
||||
irq_alloc_desc_at(irq, 0);
|
||||
} else
|
||||
irq = find_unbound_irq();
|
||||
|
||||
set_irq_chip_and_handler_name(irq, &xen_pirq_chip,
|
||||
handle_level_irq, name);
|
||||
|
||||
irq_op.irq = irq;
|
||||
irq_op.vector = 0;
|
||||
|
||||
/* Only the privileged domain can do this. For non-priv, the pcifront
|
||||
* driver provides a PCI bus that does the call to do exactly
|
||||
* this in the priv domain. */
|
||||
if (xen_initial_domain() &&
|
||||
HYPERVISOR_physdev_op(PHYSDEVOP_alloc_irq_vector, &irq_op)) {
|
||||
irq_free_desc(irq);
|
||||
irq = -ENOSPC;
|
||||
goto out;
|
||||
}
|
||||
|
||||
irq_info[irq] = mk_pirq_info(0, pirq, gsi, irq_op.vector);
|
||||
irq_info[irq].u.pirq.flags |= shareable ? PIRQ_SHAREABLE : 0;
|
||||
pirq_to_irq[pirq] = irq;
|
||||
|
||||
out:
|
||||
spin_unlock(&irq_mapping_update_lock);
|
||||
|
||||
return irq;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_PCI_MSI
|
||||
#include <linux/msi.h>
|
||||
#include "../pci/msi.h"
|
||||
|
||||
void xen_allocate_pirq_msi(char *name, int *irq, int *pirq)
|
||||
{
|
||||
spin_lock(&irq_mapping_update_lock);
|
||||
|
||||
*irq = find_unbound_irq();
|
||||
if (*irq == -1)
|
||||
goto out;
|
||||
|
||||
*pirq = find_unbound_pirq();
|
||||
if (*pirq == -1)
|
||||
goto out;
|
||||
|
||||
set_irq_chip_and_handler_name(*irq, &xen_pirq_chip,
|
||||
handle_level_irq, name);
|
||||
|
||||
irq_info[*irq] = mk_pirq_info(0, *pirq, 0, 0);
|
||||
pirq_to_irq[*pirq] = *irq;
|
||||
|
||||
out:
|
||||
spin_unlock(&irq_mapping_update_lock);
|
||||
}
|
||||
|
||||
int xen_create_msi_irq(struct pci_dev *dev, struct msi_desc *msidesc, int type)
|
||||
{
|
||||
int irq = -1;
|
||||
struct physdev_map_pirq map_irq;
|
||||
int rc;
|
||||
int pos;
|
||||
u32 table_offset, bir;
|
||||
|
||||
memset(&map_irq, 0, sizeof(map_irq));
|
||||
map_irq.domid = DOMID_SELF;
|
||||
map_irq.type = MAP_PIRQ_TYPE_MSI;
|
||||
map_irq.index = -1;
|
||||
map_irq.pirq = -1;
|
||||
map_irq.bus = dev->bus->number;
|
||||
map_irq.devfn = dev->devfn;
|
||||
|
||||
if (type == PCI_CAP_ID_MSIX) {
|
||||
pos = pci_find_capability(dev, PCI_CAP_ID_MSIX);
|
||||
|
||||
pci_read_config_dword(dev, msix_table_offset_reg(pos),
|
||||
&table_offset);
|
||||
bir = (u8)(table_offset & PCI_MSIX_FLAGS_BIRMASK);
|
||||
|
||||
map_irq.table_base = pci_resource_start(dev, bir);
|
||||
map_irq.entry_nr = msidesc->msi_attrib.entry_nr;
|
||||
}
|
||||
|
||||
spin_lock(&irq_mapping_update_lock);
|
||||
|
||||
irq = find_unbound_irq();
|
||||
|
||||
if (irq == -1)
|
||||
goto out;
|
||||
|
||||
rc = HYPERVISOR_physdev_op(PHYSDEVOP_map_pirq, &map_irq);
|
||||
if (rc) {
|
||||
printk(KERN_WARNING "xen map irq failed %d\n", rc);
|
||||
|
||||
irq_free_desc(irq);
|
||||
|
||||
irq = -1;
|
||||
goto out;
|
||||
}
|
||||
irq_info[irq] = mk_pirq_info(0, map_irq.pirq, 0, map_irq.index);
|
||||
|
||||
set_irq_chip_and_handler_name(irq, &xen_pirq_chip,
|
||||
handle_level_irq,
|
||||
(type == PCI_CAP_ID_MSIX) ? "msi-x":"msi");
|
||||
|
||||
out:
|
||||
spin_unlock(&irq_mapping_update_lock);
|
||||
return irq;
|
||||
}
|
||||
#endif
|
||||
|
||||
int xen_destroy_irq(int irq)
|
||||
{
|
||||
struct irq_desc *desc;
|
||||
struct physdev_unmap_pirq unmap_irq;
|
||||
struct irq_info *info = info_for_irq(irq);
|
||||
int rc = -ENOENT;
|
||||
|
||||
spin_lock(&irq_mapping_update_lock);
|
||||
|
||||
desc = irq_to_desc(irq);
|
||||
if (!desc)
|
||||
goto out;
|
||||
|
||||
if (xen_initial_domain()) {
|
||||
unmap_irq.pirq = info->u.pirq.gsi;
|
||||
unmap_irq.domid = DOMID_SELF;
|
||||
rc = HYPERVISOR_physdev_op(PHYSDEVOP_unmap_pirq, &unmap_irq);
|
||||
if (rc) {
|
||||
printk(KERN_WARNING "unmap irq failed %d\n", rc);
|
||||
goto out;
|
||||
}
|
||||
}
|
||||
irq_info[irq] = mk_unbound_info();
|
||||
|
||||
irq_free_desc(irq);
|
||||
|
||||
out:
|
||||
spin_unlock(&irq_mapping_update_lock);
|
||||
return rc;
|
||||
}
|
||||
|
||||
int xen_vector_from_irq(unsigned irq)
|
||||
{
|
||||
return vector_from_irq(irq);
|
||||
}
|
||||
|
||||
int xen_gsi_from_irq(unsigned irq)
|
||||
{
|
||||
return gsi_from_irq(irq);
|
||||
}
|
||||
|
||||
int bind_evtchn_to_irq(unsigned int evtchn)
|
||||
|
@ -425,7 +839,7 @@ static int bind_ipi_to_irq(unsigned int ipi, unsigned int cpu)
|
|||
}
|
||||
|
||||
|
||||
static int bind_virq_to_irq(unsigned int virq, unsigned int cpu)
|
||||
int bind_virq_to_irq(unsigned int virq, unsigned int cpu)
|
||||
{
|
||||
struct evtchn_bind_virq bind_virq;
|
||||
int evtchn, irq;
|
||||
|
@ -928,7 +1342,7 @@ void xen_clear_irq_pending(int irq)
|
|||
if (VALID_EVTCHN(evtchn))
|
||||
clear_evtchn(evtchn);
|
||||
}
|
||||
|
||||
EXPORT_SYMBOL(xen_clear_irq_pending);
|
||||
void xen_set_irq_pending(int irq)
|
||||
{
|
||||
int evtchn = evtchn_from_irq(irq);
|
||||
|
@ -948,9 +1362,9 @@ bool xen_test_irq_pending(int irq)
|
|||
return ret;
|
||||
}
|
||||
|
||||
/* Poll waiting for an irq to become pending. In the usual case, the
|
||||
irq will be disabled so it won't deliver an interrupt. */
|
||||
void xen_poll_irq(int irq)
|
||||
/* Poll waiting for an irq to become pending with timeout. In the usual case,
|
||||
* the irq will be disabled so it won't deliver an interrupt. */
|
||||
void xen_poll_irq_timeout(int irq, u64 timeout)
|
||||
{
|
||||
evtchn_port_t evtchn = evtchn_from_irq(irq);
|
||||
|
||||
|
@ -958,13 +1372,20 @@ void xen_poll_irq(int irq)
|
|||
struct sched_poll poll;
|
||||
|
||||
poll.nr_ports = 1;
|
||||
poll.timeout = 0;
|
||||
poll.timeout = timeout;
|
||||
set_xen_guest_handle(poll.ports, &evtchn);
|
||||
|
||||
if (HYPERVISOR_sched_op(SCHEDOP_poll, &poll) != 0)
|
||||
BUG();
|
||||
}
|
||||
}
|
||||
EXPORT_SYMBOL(xen_poll_irq_timeout);
|
||||
/* Poll waiting for an irq to become pending. In the usual case, the
|
||||
* irq will be disabled so it won't deliver an interrupt. */
|
||||
void xen_poll_irq(int irq)
|
||||
{
|
||||
xen_poll_irq_timeout(irq, 0 /* no timeout */);
|
||||
}
|
||||
|
||||
void xen_irq_resume(void)
|
||||
{
|
||||
|
@ -1001,6 +1422,26 @@ static struct irq_chip xen_dynamic_chip __read_mostly = {
|
|||
.retrigger = retrigger_dynirq,
|
||||
};
|
||||
|
||||
static struct irq_chip xen_pirq_chip __read_mostly = {
|
||||
.name = "xen-pirq",
|
||||
|
||||
.startup = startup_pirq,
|
||||
.shutdown = shutdown_pirq,
|
||||
|
||||
.enable = enable_pirq,
|
||||
.unmask = enable_pirq,
|
||||
|
||||
.disable = disable_pirq,
|
||||
.mask = disable_pirq,
|
||||
|
||||
.ack = ack_pirq,
|
||||
.end = end_pirq,
|
||||
|
||||
.set_affinity = set_affinity_irq,
|
||||
|
||||
.retrigger = retrigger_dynirq,
|
||||
};
|
||||
|
||||
static struct irq_chip xen_percpu_chip __read_mostly = {
|
||||
.name = "xen-percpu",
|
||||
|
||||
|
@ -1051,11 +1492,32 @@ void xen_callback_vector(void) {}
|
|||
|
||||
void __init xen_init_IRQ(void)
|
||||
{
|
||||
int i;
|
||||
int i, rc;
|
||||
struct physdev_nr_pirqs op_nr_pirqs;
|
||||
|
||||
cpu_evtchn_mask_p = kcalloc(nr_cpu_ids, sizeof(struct cpu_evtchn_s),
|
||||
GFP_KERNEL);
|
||||
BUG_ON(cpu_evtchn_mask_p == NULL);
|
||||
irq_info = kcalloc(nr_irqs, sizeof(*irq_info), GFP_KERNEL);
|
||||
|
||||
rc = HYPERVISOR_physdev_op(PHYSDEVOP_get_nr_pirqs, &op_nr_pirqs);
|
||||
if (rc < 0) {
|
||||
nr_pirqs = nr_irqs;
|
||||
if (rc != -ENOSYS)
|
||||
printk(KERN_WARNING "PHYSDEVOP_get_nr_pirqs returned rc=%d\n", rc);
|
||||
} else {
|
||||
if (xen_pv_domain() && !xen_initial_domain())
|
||||
nr_pirqs = max((int)op_nr_pirqs.nr_pirqs, nr_irqs);
|
||||
else
|
||||
nr_pirqs = op_nr_pirqs.nr_pirqs;
|
||||
}
|
||||
pirq_to_irq = kcalloc(nr_pirqs, sizeof(*pirq_to_irq), GFP_KERNEL);
|
||||
for (i = 0; i < nr_pirqs; i++)
|
||||
pirq_to_irq[i] = -1;
|
||||
|
||||
evtchn_to_irq = kcalloc(NR_EVENT_CHANNELS, sizeof(*evtchn_to_irq),
|
||||
GFP_KERNEL);
|
||||
for (i = 0; i < NR_EVENT_CHANNELS; i++)
|
||||
evtchn_to_irq[i] = -1;
|
||||
|
||||
init_evtchn_cpu_bindings();
|
||||
|
||||
|
@ -1066,7 +1528,12 @@ void __init xen_init_IRQ(void)
|
|||
if (xen_hvm_domain()) {
|
||||
xen_callback_vector();
|
||||
native_init_IRQ();
|
||||
/* pci_xen_hvm_init must be called after native_init_IRQ so that
|
||||
* __acpi_register_gsi can point at the right function */
|
||||
pci_xen_hvm_init();
|
||||
} else {
|
||||
irq_ctx_init(smp_processor_id());
|
||||
if (xen_initial_domain())
|
||||
xen_setup_pirqs();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -0,0 +1,117 @@
|
|||
/*
|
||||
* Copyright (c) 2009, Intel Corporation.
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or modify it
|
||||
* under the terms and conditions of the GNU General Public License,
|
||||
* version 2, as published by the Free Software Foundation.
|
||||
*
|
||||
* This program is distributed in the hope it will be useful, but WITHOUT
|
||||
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
||||
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
||||
* more details.
|
||||
*
|
||||
* You should have received a copy of the GNU General Public License along with
|
||||
* this program; if not, write to the Free Software Foundation, Inc., 59 Temple
|
||||
* Place - Suite 330, Boston, MA 02111-1307 USA.
|
||||
*
|
||||
* Author: Weidong Han <weidong.han@intel.com>
|
||||
*/
|
||||
|
||||
#include <linux/pci.h>
|
||||
#include <xen/xen.h>
|
||||
#include <xen/interface/physdev.h>
|
||||
#include <xen/interface/xen.h>
|
||||
|
||||
#include <asm/xen/hypervisor.h>
|
||||
#include <asm/xen/hypercall.h>
|
||||
#include "../pci/pci.h"
|
||||
|
||||
static int xen_add_device(struct device *dev)
|
||||
{
|
||||
int r;
|
||||
struct pci_dev *pci_dev = to_pci_dev(dev);
|
||||
|
||||
#ifdef CONFIG_PCI_IOV
|
||||
if (pci_dev->is_virtfn) {
|
||||
struct physdev_manage_pci_ext manage_pci_ext = {
|
||||
.bus = pci_dev->bus->number,
|
||||
.devfn = pci_dev->devfn,
|
||||
.is_virtfn = 1,
|
||||
.physfn.bus = pci_dev->physfn->bus->number,
|
||||
.physfn.devfn = pci_dev->physfn->devfn,
|
||||
};
|
||||
|
||||
r = HYPERVISOR_physdev_op(PHYSDEVOP_manage_pci_add_ext,
|
||||
&manage_pci_ext);
|
||||
} else
|
||||
#endif
|
||||
if (pci_ari_enabled(pci_dev->bus) && PCI_SLOT(pci_dev->devfn)) {
|
||||
struct physdev_manage_pci_ext manage_pci_ext = {
|
||||
.bus = pci_dev->bus->number,
|
||||
.devfn = pci_dev->devfn,
|
||||
.is_extfn = 1,
|
||||
};
|
||||
|
||||
r = HYPERVISOR_physdev_op(PHYSDEVOP_manage_pci_add_ext,
|
||||
&manage_pci_ext);
|
||||
} else {
|
||||
struct physdev_manage_pci manage_pci = {
|
||||
.bus = pci_dev->bus->number,
|
||||
.devfn = pci_dev->devfn,
|
||||
};
|
||||
|
||||
r = HYPERVISOR_physdev_op(PHYSDEVOP_manage_pci_add,
|
||||
&manage_pci);
|
||||
}
|
||||
|
||||
return r;
|
||||
}
|
||||
|
||||
static int xen_remove_device(struct device *dev)
|
||||
{
|
||||
int r;
|
||||
struct pci_dev *pci_dev = to_pci_dev(dev);
|
||||
struct physdev_manage_pci manage_pci;
|
||||
|
||||
manage_pci.bus = pci_dev->bus->number;
|
||||
manage_pci.devfn = pci_dev->devfn;
|
||||
|
||||
r = HYPERVISOR_physdev_op(PHYSDEVOP_manage_pci_remove,
|
||||
&manage_pci);
|
||||
|
||||
return r;
|
||||
}
|
||||
|
||||
static int xen_pci_notifier(struct notifier_block *nb,
|
||||
unsigned long action, void *data)
|
||||
{
|
||||
struct device *dev = data;
|
||||
int r = 0;
|
||||
|
||||
switch (action) {
|
||||
case BUS_NOTIFY_ADD_DEVICE:
|
||||
r = xen_add_device(dev);
|
||||
break;
|
||||
case BUS_NOTIFY_DEL_DEVICE:
|
||||
r = xen_remove_device(dev);
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
}
|
||||
|
||||
return r;
|
||||
}
|
||||
|
||||
struct notifier_block device_nb = {
|
||||
.notifier_call = xen_pci_notifier,
|
||||
};
|
||||
|
||||
static int __init register_xen_pci_notifier(void)
|
||||
{
|
||||
if (!xen_initial_domain())
|
||||
return 0;
|
||||
|
||||
return bus_register_notifier(&pci_bus_type, &device_nb);
|
||||
}
|
||||
|
||||
arch_initcall(register_xen_pci_notifier);
|
|
@ -50,6 +50,8 @@ const char *xenbus_strstate(enum xenbus_state state)
|
|||
[ XenbusStateConnected ] = "Connected",
|
||||
[ XenbusStateClosing ] = "Closing",
|
||||
[ XenbusStateClosed ] = "Closed",
|
||||
[XenbusStateReconfiguring] = "Reconfiguring",
|
||||
[XenbusStateReconfigured] = "Reconfigured",
|
||||
};
|
||||
return (state < ARRAY_SIZE(name)) ? name[state] : "INVALID";
|
||||
}
|
||||
|
|
|
@ -803,6 +803,7 @@ device_initcall(xenbus_probe_initcall);
|
|||
static int __init xenbus_init(void)
|
||||
{
|
||||
int err = 0;
|
||||
unsigned long page = 0;
|
||||
|
||||
DPRINTK("");
|
||||
|
||||
|
@ -823,7 +824,31 @@ static int __init xenbus_init(void)
|
|||
* Domain0 doesn't have a store_evtchn or store_mfn yet.
|
||||
*/
|
||||
if (xen_initial_domain()) {
|
||||
/* dom0 not yet supported */
|
||||
struct evtchn_alloc_unbound alloc_unbound;
|
||||
|
||||
/* Allocate Xenstore page */
|
||||
page = get_zeroed_page(GFP_KERNEL);
|
||||
if (!page)
|
||||
goto out_error;
|
||||
|
||||
xen_store_mfn = xen_start_info->store_mfn =
|
||||
pfn_to_mfn(virt_to_phys((void *)page) >>
|
||||
PAGE_SHIFT);
|
||||
|
||||
/* Next allocate a local port which xenstored can bind to */
|
||||
alloc_unbound.dom = DOMID_SELF;
|
||||
alloc_unbound.remote_dom = 0;
|
||||
|
||||
err = HYPERVISOR_event_channel_op(EVTCHNOP_alloc_unbound,
|
||||
&alloc_unbound);
|
||||
if (err == -ENOSYS)
|
||||
goto out_error;
|
||||
|
||||
BUG_ON(err);
|
||||
xen_store_evtchn = xen_start_info->store_evtchn =
|
||||
alloc_unbound.port;
|
||||
|
||||
xen_store_interface = mfn_to_virt(xen_store_mfn);
|
||||
} else {
|
||||
if (xen_hvm_domain()) {
|
||||
uint64_t v = 0;
|
||||
|
@ -869,6 +894,8 @@ static int __init xenbus_init(void)
|
|||
bus_unregister(&xenbus_frontend.bus);
|
||||
|
||||
out_error:
|
||||
if (page != 0)
|
||||
free_page(page);
|
||||
return err;
|
||||
}
|
||||
|
||||
|
|
|
@ -12,6 +12,7 @@ int bind_evtchn_to_irqhandler(unsigned int evtchn,
|
|||
irq_handler_t handler,
|
||||
unsigned long irqflags, const char *devname,
|
||||
void *dev_id);
|
||||
int bind_virq_to_irq(unsigned int virq, unsigned int cpu);
|
||||
int bind_virq_to_irqhandler(unsigned int virq, unsigned int cpu,
|
||||
irq_handler_t handler,
|
||||
unsigned long irqflags, const char *devname,
|
||||
|
@ -53,6 +54,10 @@ bool xen_test_irq_pending(int irq);
|
|||
irq will be disabled so it won't deliver an interrupt. */
|
||||
void xen_poll_irq(int irq);
|
||||
|
||||
/* Poll waiting for an irq to become pending with a timeout. In the usual case,
|
||||
* the irq will be disabled so it won't deliver an interrupt. */
|
||||
void xen_poll_irq_timeout(int irq, u64 timeout);
|
||||
|
||||
/* Determine the IRQ which is bound to an event channel */
|
||||
unsigned irq_from_evtchn(unsigned int evtchn);
|
||||
|
||||
|
@ -63,4 +68,25 @@ int xen_set_callback_via(uint64_t via);
|
|||
void xen_evtchn_do_upcall(struct pt_regs *regs);
|
||||
void xen_hvm_evtchn_do_upcall(void);
|
||||
|
||||
/* Allocate an irq for a physical interrupt, given a gsi. "Legacy"
|
||||
* GSIs are identity mapped; others are dynamically allocated as
|
||||
* usual. */
|
||||
int xen_allocate_pirq(unsigned gsi, int shareable, char *name);
|
||||
int xen_map_pirq_gsi(unsigned pirq, unsigned gsi, int shareable, char *name);
|
||||
|
||||
#ifdef CONFIG_PCI_MSI
|
||||
/* Allocate an irq and a pirq to be used with MSIs. */
|
||||
void xen_allocate_pirq_msi(char *name, int *irq, int *pirq);
|
||||
int xen_create_msi_irq(struct pci_dev *dev, struct msi_desc *msidesc, int type);
|
||||
#endif
|
||||
|
||||
/* De-allocates the above mentioned physical interrupt. */
|
||||
int xen_destroy_irq(int irq);
|
||||
|
||||
/* Return vector allocated to pirq */
|
||||
int xen_vector_from_irq(unsigned pirq);
|
||||
|
||||
/* Return gsi allocated to pirq */
|
||||
int xen_gsi_from_irq(unsigned pirq);
|
||||
|
||||
#endif /* _XEN_EVENTS_H */
|
||||
|
|
|
@ -47,6 +47,9 @@
|
|||
/* x86: pvclock algorithm is safe to use on HVM */
|
||||
#define XENFEAT_hvm_safe_pvclock 9
|
||||
|
||||
/* x86: pirq can be used by HVM guests */
|
||||
#define XENFEAT_hvm_pirqs 10
|
||||
|
||||
#define XENFEAT_NR_SUBMAPS 1
|
||||
|
||||
#endif /* __XEN_PUBLIC_FEATURES_H__ */
|
||||
|
|
|
@ -0,0 +1,112 @@
|
|||
/*
|
||||
* PCI Backend/Frontend Common Data Structures & Macros
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
||||
* of this software and associated documentation files (the "Software"), to
|
||||
* deal in the Software without restriction, including without limitation the
|
||||
* rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
|
||||
* sell copies of the Software, and to permit persons to whom the Software is
|
||||
* furnished to do so, subject to the following conditions:
|
||||
*
|
||||
* The above copyright notice and this permission notice shall be included in
|
||||
* all copies or substantial portions of the Software.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
||||
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
||||
* DEALINGS IN THE SOFTWARE.
|
||||
*
|
||||
* Author: Ryan Wilson <hap9@epoch.ncsc.mil>
|
||||
*/
|
||||
#ifndef __XEN_PCI_COMMON_H__
|
||||
#define __XEN_PCI_COMMON_H__
|
||||
|
||||
/* Be sure to bump this number if you change this file */
|
||||
#define XEN_PCI_MAGIC "7"
|
||||
|
||||
/* xen_pci_sharedinfo flags */
|
||||
#define _XEN_PCIF_active (0)
|
||||
#define XEN_PCIF_active (1<<_XEN_PCIF_active)
|
||||
#define _XEN_PCIB_AERHANDLER (1)
|
||||
#define XEN_PCIB_AERHANDLER (1<<_XEN_PCIB_AERHANDLER)
|
||||
#define _XEN_PCIB_active (2)
|
||||
#define XEN_PCIB_active (1<<_XEN_PCIB_active)
|
||||
|
||||
/* xen_pci_op commands */
|
||||
#define XEN_PCI_OP_conf_read (0)
|
||||
#define XEN_PCI_OP_conf_write (1)
|
||||
#define XEN_PCI_OP_enable_msi (2)
|
||||
#define XEN_PCI_OP_disable_msi (3)
|
||||
#define XEN_PCI_OP_enable_msix (4)
|
||||
#define XEN_PCI_OP_disable_msix (5)
|
||||
#define XEN_PCI_OP_aer_detected (6)
|
||||
#define XEN_PCI_OP_aer_resume (7)
|
||||
#define XEN_PCI_OP_aer_mmio (8)
|
||||
#define XEN_PCI_OP_aer_slotreset (9)
|
||||
|
||||
/* xen_pci_op error numbers */
|
||||
#define XEN_PCI_ERR_success (0)
|
||||
#define XEN_PCI_ERR_dev_not_found (-1)
|
||||
#define XEN_PCI_ERR_invalid_offset (-2)
|
||||
#define XEN_PCI_ERR_access_denied (-3)
|
||||
#define XEN_PCI_ERR_not_implemented (-4)
|
||||
/* XEN_PCI_ERR_op_failed - backend failed to complete the operation */
|
||||
#define XEN_PCI_ERR_op_failed (-5)
|
||||
|
||||
/*
|
||||
* it should be PAGE_SIZE-sizeof(struct xen_pci_op))/sizeof(struct msix_entry))
|
||||
* Should not exceed 128
|
||||
*/
|
||||
#define SH_INFO_MAX_VEC 128
|
||||
|
||||
struct xen_msix_entry {
|
||||
uint16_t vector;
|
||||
uint16_t entry;
|
||||
};
|
||||
struct xen_pci_op {
|
||||
/* IN: what action to perform: XEN_PCI_OP_* */
|
||||
uint32_t cmd;
|
||||
|
||||
/* OUT: will contain an error number (if any) from errno.h */
|
||||
int32_t err;
|
||||
|
||||
/* IN: which device to touch */
|
||||
uint32_t domain; /* PCI Domain/Segment */
|
||||
uint32_t bus;
|
||||
uint32_t devfn;
|
||||
|
||||
/* IN: which configuration registers to touch */
|
||||
int32_t offset;
|
||||
int32_t size;
|
||||
|
||||
/* IN/OUT: Contains the result after a READ or the value to WRITE */
|
||||
uint32_t value;
|
||||
/* IN: Contains extra infor for this operation */
|
||||
uint32_t info;
|
||||
/*IN: param for msi-x */
|
||||
struct xen_msix_entry msix_entries[SH_INFO_MAX_VEC];
|
||||
};
|
||||
|
||||
/*used for pcie aer handling*/
|
||||
struct xen_pcie_aer_op {
|
||||
/* IN: what action to perform: XEN_PCI_OP_* */
|
||||
uint32_t cmd;
|
||||
/*IN/OUT: return aer_op result or carry error_detected state as input*/
|
||||
int32_t err;
|
||||
|
||||
/* IN: which device to touch */
|
||||
uint32_t domain; /* PCI Domain/Segment*/
|
||||
uint32_t bus;
|
||||
uint32_t devfn;
|
||||
};
|
||||
struct xen_pci_sharedinfo {
|
||||
/* flags - XEN_PCIF_* */
|
||||
uint32_t flags;
|
||||
struct xen_pci_op op;
|
||||
struct xen_pcie_aer_op aer_op;
|
||||
};
|
||||
|
||||
#endif /* __XEN_PCI_COMMON_H__ */
|
|
@ -27,8 +27,14 @@ enum xenbus_state
|
|||
XenbusStateClosing = 5, /* The device is being closed
|
||||
due to an error or an unplug
|
||||
event. */
|
||||
XenbusStateClosed = 6
|
||||
XenbusStateClosed = 6,
|
||||
|
||||
/*
|
||||
* Reconfiguring: The device is being reconfigured.
|
||||
*/
|
||||
XenbusStateReconfiguring = 7,
|
||||
|
||||
XenbusStateReconfigured = 8
|
||||
};
|
||||
|
||||
#endif /* _XEN_PUBLIC_IO_XENBUS_H */
|
||||
|
|
|
@ -106,6 +106,57 @@ struct physdev_irq {
|
|||
uint32_t vector;
|
||||
};
|
||||
|
||||
#define MAP_PIRQ_TYPE_MSI 0x0
|
||||
#define MAP_PIRQ_TYPE_GSI 0x1
|
||||
#define MAP_PIRQ_TYPE_UNKNOWN 0x2
|
||||
|
||||
#define PHYSDEVOP_map_pirq 13
|
||||
struct physdev_map_pirq {
|
||||
domid_t domid;
|
||||
/* IN */
|
||||
int type;
|
||||
/* IN */
|
||||
int index;
|
||||
/* IN or OUT */
|
||||
int pirq;
|
||||
/* IN */
|
||||
int bus;
|
||||
/* IN */
|
||||
int devfn;
|
||||
/* IN */
|
||||
int entry_nr;
|
||||
/* IN */
|
||||
uint64_t table_base;
|
||||
};
|
||||
|
||||
#define PHYSDEVOP_unmap_pirq 14
|
||||
struct physdev_unmap_pirq {
|
||||
domid_t domid;
|
||||
/* IN */
|
||||
int pirq;
|
||||
};
|
||||
|
||||
#define PHYSDEVOP_manage_pci_add 15
|
||||
#define PHYSDEVOP_manage_pci_remove 16
|
||||
struct physdev_manage_pci {
|
||||
/* IN */
|
||||
uint8_t bus;
|
||||
uint8_t devfn;
|
||||
};
|
||||
|
||||
#define PHYSDEVOP_manage_pci_add_ext 20
|
||||
struct physdev_manage_pci_ext {
|
||||
/* IN */
|
||||
uint8_t bus;
|
||||
uint8_t devfn;
|
||||
unsigned is_extfn;
|
||||
unsigned is_virtfn;
|
||||
struct {
|
||||
uint8_t bus;
|
||||
uint8_t devfn;
|
||||
} physfn;
|
||||
};
|
||||
|
||||
/*
|
||||
* Argument to physdev_op_compat() hypercall. Superceded by new physdev_op()
|
||||
* hypercall since 0x00030202.
|
||||
|
@ -121,6 +172,22 @@ struct physdev_op {
|
|||
} u;
|
||||
};
|
||||
|
||||
#define PHYSDEVOP_setup_gsi 21
|
||||
struct physdev_setup_gsi {
|
||||
int gsi;
|
||||
/* IN */
|
||||
uint8_t triggering;
|
||||
/* IN */
|
||||
uint8_t polarity;
|
||||
/* IN */
|
||||
};
|
||||
|
||||
#define PHYSDEVOP_get_nr_pirqs 22
|
||||
struct physdev_nr_pirqs {
|
||||
/* OUT */
|
||||
uint32_t nr_pirqs;
|
||||
};
|
||||
|
||||
/*
|
||||
* Notify that some PIRQ-bound event channels have been unmasked.
|
||||
* ** This command is obsolete since interface version 0x00030202 and is **
|
||||
|
|
Loading…
Reference in New Issue