2019-06-04 16:11:33 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-only
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* linux/arch/arm/common/amba.c
|
|
|
|
*
|
|
|
|
* Copyright (C) 2003 Deep Blue Solutions Ltd, All Rights Reserved.
|
|
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/device.h>
|
2005-10-31 07:03:48 +08:00
|
|
|
#include <linux/string.h>
|
|
|
|
#include <linux/slab.h>
|
2009-01-08 17:58:51 +08:00
|
|
|
#include <linux/io.h>
|
2011-02-23 11:33:17 +08:00
|
|
|
#include <linux/pm.h>
|
|
|
|
#include <linux/pm_runtime.h>
|
2014-09-29 19:58:47 +08:00
|
|
|
#include <linux/pm_domain.h>
|
2006-01-07 21:52:45 +08:00
|
|
|
#include <linux/amba/bus.h>
|
2012-06-24 19:46:16 +08:00
|
|
|
#include <linux/sizes.h>
|
2015-01-06 18:15:11 +08:00
|
|
|
#include <linux/limits.h>
|
2016-08-11 03:17:34 +08:00
|
|
|
#include <linux/clk/clk-conf.h>
|
2018-04-28 10:51:58 +08:00
|
|
|
#include <linux/platform_device.h>
|
2019-09-04 09:13:08 +08:00
|
|
|
#include <linux/reset.h>
|
2021-08-23 17:41:43 +08:00
|
|
|
#include <linux/of_irq.h>
|
2022-04-18 08:49:52 +08:00
|
|
|
#include <linux/of_device.h>
|
|
|
|
#include <linux/acpi.h>
|
bus: platform,amba,fsl-mc,PCI: Add device DMA ownership management
The devices on platform/amba/fsl-mc/PCI buses could be bound to drivers
with the device DMA managed by kernel drivers or user-space applications.
Unfortunately, multiple devices may be placed in the same IOMMU group
because they cannot be isolated from each other. The DMA on these devices
must either be entirely under kernel control or userspace control, never
a mixture. Otherwise the driver integrity is not guaranteed because they
could access each other through the peer-to-peer accesses which by-pass
the IOMMU protection.
This checks and sets the default DMA mode during driver binding, and
cleanups during driver unbinding. In the default mode, the device DMA is
managed by the device driver which handles DMA operations through the
kernel DMA APIs (see Documentation/core-api/dma-api.rst).
For cases where the devices are assigned for userspace control through the
userspace driver framework(i.e. VFIO), the drivers(for example, vfio_pci/
vfio_platfrom etc.) may set a new flag (driver_managed_dma) to skip this
default setting in the assumption that the drivers know what they are
doing with the device DMA.
Calling iommu_device_use_default_domain() before {of,acpi}_dma_configure
is currently a problem. As things stand, the IOMMU driver ignored the
initial iommu_probe_device() call when the device was added, since at
that point it had no fwspec yet. In this situation,
{of,acpi}_iommu_configure() are retriggering iommu_probe_device() after
the IOMMU driver has seen the firmware data via .of_xlate to learn that
it actually responsible for the given device. As the result, before
that gets fixed, iommu_use_default_domain() goes at the end, and calls
arch_teardown_dma_ops() if it fails.
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Bjorn Helgaas <bhelgaas@google.com>
Cc: Stuart Yoder <stuyoder@gmail.com>
Cc: Laurentiu Tudor <laurentiu.tudor@nxp.com>
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
Reviewed-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
Reviewed-by: Robin Murphy <robin.murphy@arm.com>
Tested-by: Eric Auger <eric.auger@redhat.com>
Link: https://lore.kernel.org/r/20220418005000.897664-5-baolu.lu@linux.intel.com
Signed-off-by: Joerg Roedel <jroedel@suse.de>
2022-04-18 08:49:53 +08:00
|
|
|
#include <linux/iommu.h>
|
|
|
|
#include <linux/dma-map-ops.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
#define to_amba_driver(d) container_of(d, struct amba_driver, drv)
|
|
|
|
|
2019-02-13 21:41:50 +08:00
|
|
|
/* called on periphid match and class 0x9 coresight device. */
|
|
|
|
static int
|
|
|
|
amba_cs_uci_id_match(const struct amba_id *table, struct amba_device *dev)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
int ret = 0;
|
2019-02-13 21:41:50 +08:00
|
|
|
struct amba_cs_uci_id *uci;
|
|
|
|
|
|
|
|
uci = table->data;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2019-02-13 21:41:50 +08:00
|
|
|
/* no table data or zero mask - return match on periphid */
|
|
|
|
if (!uci || (uci->devarch_mask == 0))
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
/* test against read devtype and masked devarch value */
|
|
|
|
ret = (dev->uci.devtype == uci->devtype) &&
|
|
|
|
((dev->uci.devarch & uci->devarch_mask) == uci->devarch);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct amba_id *
|
|
|
|
amba_lookup(const struct amba_id *table, struct amba_device *dev)
|
|
|
|
{
|
2005-04-17 06:20:36 +08:00
|
|
|
while (table->mask) {
|
2019-02-13 21:41:50 +08:00
|
|
|
if (((dev->periphid & table->mask) == table->id) &&
|
|
|
|
((dev->cid != CORESIGHT_CID) ||
|
|
|
|
(amba_cs_uci_id_match(table, dev))))
|
|
|
|
return table;
|
2005-04-17 06:20:36 +08:00
|
|
|
table++;
|
|
|
|
}
|
2019-02-13 21:41:50 +08:00
|
|
|
return NULL;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2021-01-27 00:58:32 +08:00
|
|
|
static int amba_get_enable_pclk(struct amba_device *pcdev)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2021-01-27 00:58:32 +08:00
|
|
|
int ret;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2021-01-27 00:58:32 +08:00
|
|
|
pcdev->pclk = clk_get(&pcdev->dev, "apb_pclk");
|
|
|
|
if (IS_ERR(pcdev->pclk))
|
|
|
|
return PTR_ERR(pcdev->pclk);
|
2015-01-06 18:15:11 +08:00
|
|
|
|
2021-01-27 00:58:32 +08:00
|
|
|
ret = clk_prepare_enable(pcdev->pclk);
|
|
|
|
if (ret)
|
|
|
|
clk_put(pcdev->pclk);
|
|
|
|
|
|
|
|
return ret;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2021-01-27 00:58:32 +08:00
|
|
|
static void amba_put_disable_pclk(struct amba_device *pcdev)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2021-01-27 00:58:32 +08:00
|
|
|
clk_disable_unprepare(pcdev->pclk);
|
|
|
|
clk_put(pcdev->pclk);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2021-01-27 00:58:32 +08:00
|
|
|
|
2015-01-06 18:15:11 +08:00
|
|
|
static ssize_t driver_override_show(struct device *_dev,
|
|
|
|
struct device_attribute *attr, char *buf)
|
|
|
|
{
|
|
|
|
struct amba_device *dev = to_amba_device(_dev);
|
2018-04-10 21:21:44 +08:00
|
|
|
ssize_t len;
|
2015-01-06 18:15:11 +08:00
|
|
|
|
2018-04-10 21:21:44 +08:00
|
|
|
device_lock(_dev);
|
|
|
|
len = sprintf(buf, "%s\n", dev->driver_override);
|
|
|
|
device_unlock(_dev);
|
|
|
|
return len;
|
2015-01-06 18:15:11 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t driver_override_store(struct device *_dev,
|
|
|
|
struct device_attribute *attr,
|
|
|
|
const char *buf, size_t count)
|
|
|
|
{
|
|
|
|
struct amba_device *dev = to_amba_device(_dev);
|
2022-04-19 19:34:25 +08:00
|
|
|
int ret;
|
2015-01-06 18:15:11 +08:00
|
|
|
|
2022-04-19 19:34:25 +08:00
|
|
|
ret = driver_set_override(_dev, &dev->driver_override, buf, count);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
2015-01-06 18:15:11 +08:00
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
2017-06-06 20:16:49 +08:00
|
|
|
static DEVICE_ATTR_RW(driver_override);
|
2015-01-06 18:15:11 +08:00
|
|
|
|
2006-11-30 22:04:49 +08:00
|
|
|
#define amba_attr_func(name,fmt,arg...) \
|
|
|
|
static ssize_t name##_show(struct device *_dev, \
|
|
|
|
struct device_attribute *attr, char *buf) \
|
|
|
|
{ \
|
|
|
|
struct amba_device *dev = to_amba_device(_dev); \
|
|
|
|
return sprintf(buf, fmt, arg); \
|
2017-06-06 20:16:49 +08:00
|
|
|
} \
|
|
|
|
static DEVICE_ATTR_RO(name)
|
2006-11-30 22:04:49 +08:00
|
|
|
|
|
|
|
amba_attr_func(id, "%08x\n", dev->periphid);
|
|
|
|
amba_attr_func(resource, "\t%016llx\t%016llx\t%016lx\n",
|
|
|
|
(unsigned long long)dev->res.start, (unsigned long long)dev->res.end,
|
|
|
|
dev->res.flags);
|
|
|
|
|
2017-06-06 20:16:49 +08:00
|
|
|
static struct attribute *amba_dev_attrs[] = {
|
|
|
|
&dev_attr_id.attr,
|
|
|
|
&dev_attr_resource.attr,
|
|
|
|
&dev_attr_driver_override.attr,
|
|
|
|
NULL,
|
2006-11-30 22:04:49 +08:00
|
|
|
};
|
2017-06-06 20:16:49 +08:00
|
|
|
ATTRIBUTE_GROUPS(amba_dev);
|
2006-11-30 22:04:49 +08:00
|
|
|
|
2021-01-27 00:58:32 +08:00
|
|
|
static int amba_match(struct device *dev, struct device_driver *drv)
|
|
|
|
{
|
|
|
|
struct amba_device *pcdev = to_amba_device(dev);
|
|
|
|
struct amba_driver *pcdrv = to_amba_driver(drv);
|
|
|
|
|
|
|
|
/* When driver_override is set, only bind to the matching driver */
|
|
|
|
if (pcdev->driver_override)
|
|
|
|
return !strcmp(pcdev->driver_override, drv->name);
|
|
|
|
|
|
|
|
return amba_lookup(pcdrv->id_table, pcdev) != NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int amba_uevent(struct device *dev, struct kobj_uevent_env *env)
|
|
|
|
{
|
|
|
|
struct amba_device *pcdev = to_amba_device(dev);
|
|
|
|
int retval = 0;
|
|
|
|
|
|
|
|
retval = add_uevent_var(env, "AMBA_ID=%08x", pcdev->periphid);
|
|
|
|
if (retval)
|
|
|
|
return retval;
|
|
|
|
|
|
|
|
retval = add_uevent_var(env, "MODALIAS=amba:d%08X", pcdev->periphid);
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
2021-12-03 17:25:21 +08:00
|
|
|
static int of_amba_device_decode_irq(struct amba_device *dev)
|
|
|
|
{
|
|
|
|
struct device_node *node = dev->dev.of_node;
|
|
|
|
int i, irq = 0;
|
|
|
|
|
|
|
|
if (IS_ENABLED(CONFIG_OF_IRQ) && node) {
|
|
|
|
/* Decode the IRQs and address ranges */
|
|
|
|
for (i = 0; i < AMBA_NR_IRQS; i++) {
|
|
|
|
irq = of_irq_get(node, i);
|
|
|
|
if (irq < 0) {
|
|
|
|
if (irq == -EPROBE_DEFER)
|
|
|
|
return irq;
|
|
|
|
irq = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
dev->irq[i] = irq;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2021-01-27 00:58:35 +08:00
|
|
|
/*
|
|
|
|
* These are the device model conversion veneers; they convert the
|
|
|
|
* device model structures to our more specific structures.
|
|
|
|
*/
|
|
|
|
static int amba_probe(struct device *dev)
|
|
|
|
{
|
|
|
|
struct amba_device *pcdev = to_amba_device(dev);
|
|
|
|
struct amba_driver *pcdrv = to_amba_driver(dev->driver);
|
|
|
|
const struct amba_id *id = amba_lookup(pcdrv->id_table, pcdev);
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
do {
|
2021-12-03 17:25:21 +08:00
|
|
|
ret = of_amba_device_decode_irq(pcdev);
|
|
|
|
if (ret)
|
|
|
|
break;
|
|
|
|
|
2021-01-27 00:58:35 +08:00
|
|
|
ret = of_clk_set_defaults(dev->of_node, false);
|
|
|
|
if (ret < 0)
|
|
|
|
break;
|
|
|
|
|
|
|
|
ret = dev_pm_domain_attach(dev, true);
|
|
|
|
if (ret)
|
|
|
|
break;
|
|
|
|
|
|
|
|
ret = amba_get_enable_pclk(pcdev);
|
|
|
|
if (ret) {
|
|
|
|
dev_pm_domain_detach(dev, true);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
pm_runtime_get_noresume(dev);
|
|
|
|
pm_runtime_set_active(dev);
|
|
|
|
pm_runtime_enable(dev);
|
|
|
|
|
|
|
|
ret = pcdrv->probe(pcdev, id);
|
|
|
|
if (ret == 0)
|
|
|
|
break;
|
|
|
|
|
|
|
|
pm_runtime_disable(dev);
|
|
|
|
pm_runtime_set_suspended(dev);
|
|
|
|
pm_runtime_put_noidle(dev);
|
|
|
|
|
|
|
|
amba_put_disable_pclk(pcdev);
|
|
|
|
dev_pm_domain_detach(dev, true);
|
|
|
|
} while (0);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2021-07-14 03:35:22 +08:00
|
|
|
static void amba_remove(struct device *dev)
|
2021-01-27 00:58:35 +08:00
|
|
|
{
|
|
|
|
struct amba_device *pcdev = to_amba_device(dev);
|
|
|
|
struct amba_driver *drv = to_amba_driver(dev->driver);
|
|
|
|
|
|
|
|
pm_runtime_get_sync(dev);
|
|
|
|
if (drv->remove)
|
|
|
|
drv->remove(pcdev);
|
|
|
|
pm_runtime_put_noidle(dev);
|
|
|
|
|
|
|
|
/* Undo the runtime PM settings in amba_probe() */
|
|
|
|
pm_runtime_disable(dev);
|
|
|
|
pm_runtime_set_suspended(dev);
|
|
|
|
pm_runtime_put_noidle(dev);
|
|
|
|
|
|
|
|
amba_put_disable_pclk(pcdev);
|
|
|
|
dev_pm_domain_detach(dev, true);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void amba_shutdown(struct device *dev)
|
|
|
|
{
|
|
|
|
struct amba_driver *drv;
|
|
|
|
|
|
|
|
if (!dev->driver)
|
|
|
|
return;
|
|
|
|
|
|
|
|
drv = to_amba_driver(dev->driver);
|
|
|
|
if (drv->shutdown)
|
|
|
|
drv->shutdown(to_amba_device(dev));
|
|
|
|
}
|
|
|
|
|
2022-04-18 08:49:52 +08:00
|
|
|
static int amba_dma_configure(struct device *dev)
|
|
|
|
{
|
bus: platform,amba,fsl-mc,PCI: Add device DMA ownership management
The devices on platform/amba/fsl-mc/PCI buses could be bound to drivers
with the device DMA managed by kernel drivers or user-space applications.
Unfortunately, multiple devices may be placed in the same IOMMU group
because they cannot be isolated from each other. The DMA on these devices
must either be entirely under kernel control or userspace control, never
a mixture. Otherwise the driver integrity is not guaranteed because they
could access each other through the peer-to-peer accesses which by-pass
the IOMMU protection.
This checks and sets the default DMA mode during driver binding, and
cleanups during driver unbinding. In the default mode, the device DMA is
managed by the device driver which handles DMA operations through the
kernel DMA APIs (see Documentation/core-api/dma-api.rst).
For cases where the devices are assigned for userspace control through the
userspace driver framework(i.e. VFIO), the drivers(for example, vfio_pci/
vfio_platfrom etc.) may set a new flag (driver_managed_dma) to skip this
default setting in the assumption that the drivers know what they are
doing with the device DMA.
Calling iommu_device_use_default_domain() before {of,acpi}_dma_configure
is currently a problem. As things stand, the IOMMU driver ignored the
initial iommu_probe_device() call when the device was added, since at
that point it had no fwspec yet. In this situation,
{of,acpi}_iommu_configure() are retriggering iommu_probe_device() after
the IOMMU driver has seen the firmware data via .of_xlate to learn that
it actually responsible for the given device. As the result, before
that gets fixed, iommu_use_default_domain() goes at the end, and calls
arch_teardown_dma_ops() if it fails.
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Bjorn Helgaas <bhelgaas@google.com>
Cc: Stuart Yoder <stuyoder@gmail.com>
Cc: Laurentiu Tudor <laurentiu.tudor@nxp.com>
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
Reviewed-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
Reviewed-by: Robin Murphy <robin.murphy@arm.com>
Tested-by: Eric Auger <eric.auger@redhat.com>
Link: https://lore.kernel.org/r/20220418005000.897664-5-baolu.lu@linux.intel.com
Signed-off-by: Joerg Roedel <jroedel@suse.de>
2022-04-18 08:49:53 +08:00
|
|
|
struct amba_driver *drv = to_amba_driver(dev->driver);
|
2022-04-18 08:49:52 +08:00
|
|
|
enum dev_dma_attr attr;
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
if (dev->of_node) {
|
|
|
|
ret = of_dma_configure(dev, dev->of_node, true);
|
|
|
|
} else if (has_acpi_companion(dev)) {
|
|
|
|
attr = acpi_get_dma_attr(to_acpi_device_node(dev->fwnode));
|
|
|
|
ret = acpi_dma_configure(dev, attr);
|
|
|
|
}
|
|
|
|
|
bus: platform,amba,fsl-mc,PCI: Add device DMA ownership management
The devices on platform/amba/fsl-mc/PCI buses could be bound to drivers
with the device DMA managed by kernel drivers or user-space applications.
Unfortunately, multiple devices may be placed in the same IOMMU group
because they cannot be isolated from each other. The DMA on these devices
must either be entirely under kernel control or userspace control, never
a mixture. Otherwise the driver integrity is not guaranteed because they
could access each other through the peer-to-peer accesses which by-pass
the IOMMU protection.
This checks and sets the default DMA mode during driver binding, and
cleanups during driver unbinding. In the default mode, the device DMA is
managed by the device driver which handles DMA operations through the
kernel DMA APIs (see Documentation/core-api/dma-api.rst).
For cases where the devices are assigned for userspace control through the
userspace driver framework(i.e. VFIO), the drivers(for example, vfio_pci/
vfio_platfrom etc.) may set a new flag (driver_managed_dma) to skip this
default setting in the assumption that the drivers know what they are
doing with the device DMA.
Calling iommu_device_use_default_domain() before {of,acpi}_dma_configure
is currently a problem. As things stand, the IOMMU driver ignored the
initial iommu_probe_device() call when the device was added, since at
that point it had no fwspec yet. In this situation,
{of,acpi}_iommu_configure() are retriggering iommu_probe_device() after
the IOMMU driver has seen the firmware data via .of_xlate to learn that
it actually responsible for the given device. As the result, before
that gets fixed, iommu_use_default_domain() goes at the end, and calls
arch_teardown_dma_ops() if it fails.
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Bjorn Helgaas <bhelgaas@google.com>
Cc: Stuart Yoder <stuyoder@gmail.com>
Cc: Laurentiu Tudor <laurentiu.tudor@nxp.com>
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
Reviewed-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
Reviewed-by: Robin Murphy <robin.murphy@arm.com>
Tested-by: Eric Auger <eric.auger@redhat.com>
Link: https://lore.kernel.org/r/20220418005000.897664-5-baolu.lu@linux.intel.com
Signed-off-by: Joerg Roedel <jroedel@suse.de>
2022-04-18 08:49:53 +08:00
|
|
|
if (!ret && !drv->driver_managed_dma) {
|
|
|
|
ret = iommu_device_use_default_domain(dev);
|
|
|
|
if (ret)
|
|
|
|
arch_teardown_dma_ops(dev);
|
|
|
|
}
|
|
|
|
|
2022-04-18 08:49:52 +08:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
bus: platform,amba,fsl-mc,PCI: Add device DMA ownership management
The devices on platform/amba/fsl-mc/PCI buses could be bound to drivers
with the device DMA managed by kernel drivers or user-space applications.
Unfortunately, multiple devices may be placed in the same IOMMU group
because they cannot be isolated from each other. The DMA on these devices
must either be entirely under kernel control or userspace control, never
a mixture. Otherwise the driver integrity is not guaranteed because they
could access each other through the peer-to-peer accesses which by-pass
the IOMMU protection.
This checks and sets the default DMA mode during driver binding, and
cleanups during driver unbinding. In the default mode, the device DMA is
managed by the device driver which handles DMA operations through the
kernel DMA APIs (see Documentation/core-api/dma-api.rst).
For cases where the devices are assigned for userspace control through the
userspace driver framework(i.e. VFIO), the drivers(for example, vfio_pci/
vfio_platfrom etc.) may set a new flag (driver_managed_dma) to skip this
default setting in the assumption that the drivers know what they are
doing with the device DMA.
Calling iommu_device_use_default_domain() before {of,acpi}_dma_configure
is currently a problem. As things stand, the IOMMU driver ignored the
initial iommu_probe_device() call when the device was added, since at
that point it had no fwspec yet. In this situation,
{of,acpi}_iommu_configure() are retriggering iommu_probe_device() after
the IOMMU driver has seen the firmware data via .of_xlate to learn that
it actually responsible for the given device. As the result, before
that gets fixed, iommu_use_default_domain() goes at the end, and calls
arch_teardown_dma_ops() if it fails.
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Bjorn Helgaas <bhelgaas@google.com>
Cc: Stuart Yoder <stuyoder@gmail.com>
Cc: Laurentiu Tudor <laurentiu.tudor@nxp.com>
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
Reviewed-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
Reviewed-by: Robin Murphy <robin.murphy@arm.com>
Tested-by: Eric Auger <eric.auger@redhat.com>
Link: https://lore.kernel.org/r/20220418005000.897664-5-baolu.lu@linux.intel.com
Signed-off-by: Joerg Roedel <jroedel@suse.de>
2022-04-18 08:49:53 +08:00
|
|
|
static void amba_dma_cleanup(struct device *dev)
|
|
|
|
{
|
|
|
|
struct amba_driver *drv = to_amba_driver(dev->driver);
|
|
|
|
|
|
|
|
if (!drv->driver_managed_dma)
|
|
|
|
iommu_device_unuse_default_domain(dev);
|
|
|
|
}
|
|
|
|
|
2014-02-12 21:06:43 +08:00
|
|
|
#ifdef CONFIG_PM
|
2011-08-14 16:13:48 +08:00
|
|
|
/*
|
|
|
|
* Hooks to provide runtime PM of the pclk (bus clock). It is safe to
|
|
|
|
* enable/disable the bus clock at runtime PM suspend/resume as this
|
2012-04-13 20:11:50 +08:00
|
|
|
* does not result in loss of context.
|
2011-08-14 16:13:48 +08:00
|
|
|
*/
|
|
|
|
static int amba_pm_runtime_suspend(struct device *dev)
|
|
|
|
{
|
|
|
|
struct amba_device *pcdev = to_amba_device(dev);
|
|
|
|
int ret = pm_generic_runtime_suspend(dev);
|
|
|
|
|
2014-11-14 16:48:27 +08:00
|
|
|
if (ret == 0 && dev->driver) {
|
|
|
|
if (pm_runtime_is_irq_safe(dev))
|
|
|
|
clk_disable(pcdev->pclk);
|
|
|
|
else
|
|
|
|
clk_disable_unprepare(pcdev->pclk);
|
|
|
|
}
|
2011-08-14 16:13:48 +08:00
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int amba_pm_runtime_resume(struct device *dev)
|
|
|
|
{
|
|
|
|
struct amba_device *pcdev = to_amba_device(dev);
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (dev->driver) {
|
2014-11-14 16:48:27 +08:00
|
|
|
if (pm_runtime_is_irq_safe(dev))
|
|
|
|
ret = clk_enable(pcdev->pclk);
|
|
|
|
else
|
|
|
|
ret = clk_prepare_enable(pcdev->pclk);
|
2011-08-14 16:13:48 +08:00
|
|
|
/* Failure is probably fatal to the system, but... */
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
return pm_generic_runtime_resume(dev);
|
|
|
|
}
|
2014-11-14 16:48:27 +08:00
|
|
|
#endif /* CONFIG_PM */
|
2011-08-14 16:13:48 +08:00
|
|
|
|
2011-02-23 11:33:17 +08:00
|
|
|
static const struct dev_pm_ops amba_pm = {
|
2013-12-09 17:38:20 +08:00
|
|
|
.suspend = pm_generic_suspend,
|
|
|
|
.resume = pm_generic_resume,
|
|
|
|
.freeze = pm_generic_freeze,
|
|
|
|
.thaw = pm_generic_thaw,
|
|
|
|
.poweroff = pm_generic_poweroff,
|
|
|
|
.restore = pm_generic_restore,
|
2014-12-04 07:34:11 +08:00
|
|
|
SET_RUNTIME_PM_OPS(
|
2011-08-14 16:13:48 +08:00
|
|
|
amba_pm_runtime_suspend,
|
|
|
|
amba_pm_runtime_resume,
|
PM / Runtime: Rework the "runtime idle" helper routine
The "runtime idle" helper routine, rpm_idle(), currently ignores
return values from .runtime_idle() callbacks executed by it.
However, it turns out that many subsystems use
pm_generic_runtime_idle() which checks the return value of the
driver's callback and executes pm_runtime_suspend() for the device
unless that value is not 0. If that logic is moved to rpm_idle()
instead, pm_generic_runtime_idle() can be dropped and its users
will not need any .runtime_idle() callbacks any more.
Moreover, the PCI, SCSI, and SATA subsystems' .runtime_idle()
routines, pci_pm_runtime_idle(), scsi_runtime_idle(), and
ata_port_runtime_idle(), respectively, as well as a few drivers'
ones may be simplified if rpm_idle() calls rpm_suspend() after 0 has
been returned by the .runtime_idle() callback executed by it.
To reduce overall code bloat, make the changes described above.
Tested-by: Mika Westerberg <mika.westerberg@linux.intel.com>
Tested-by: Kevin Hilman <khilman@linaro.org>
Signed-off-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
Acked-by: Kevin Hilman <khilman@linaro.org>
Reviewed-by: Ulf Hansson <ulf.hansson@linaro.org>
Acked-by: Alan Stern <stern@rowland.harvard.edu>
2013-06-04 03:49:52 +08:00
|
|
|
NULL
|
2011-02-23 11:33:17 +08:00
|
|
|
)
|
|
|
|
};
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Primecells are part of the Advanced Microcontroller Bus Architecture,
|
|
|
|
* so we call the bus "amba".
|
2018-04-28 10:51:58 +08:00
|
|
|
* DMA configuration for platform and AMBA bus is same. So here we reuse
|
|
|
|
* platform's DMA config routine.
|
2005-04-17 06:20:36 +08:00
|
|
|
*/
|
2011-02-12 22:58:25 +08:00
|
|
|
struct bus_type amba_bustype = {
|
2005-04-17 06:20:36 +08:00
|
|
|
.name = "amba",
|
2017-06-06 20:16:49 +08:00
|
|
|
.dev_groups = amba_dev_groups,
|
2005-04-17 06:20:36 +08:00
|
|
|
.match = amba_match,
|
2005-11-26 12:04:26 +08:00
|
|
|
.uevent = amba_uevent,
|
2021-01-27 00:58:35 +08:00
|
|
|
.probe = amba_probe,
|
|
|
|
.remove = amba_remove,
|
|
|
|
.shutdown = amba_shutdown,
|
2022-04-18 08:49:52 +08:00
|
|
|
.dma_configure = amba_dma_configure,
|
bus: platform,amba,fsl-mc,PCI: Add device DMA ownership management
The devices on platform/amba/fsl-mc/PCI buses could be bound to drivers
with the device DMA managed by kernel drivers or user-space applications.
Unfortunately, multiple devices may be placed in the same IOMMU group
because they cannot be isolated from each other. The DMA on these devices
must either be entirely under kernel control or userspace control, never
a mixture. Otherwise the driver integrity is not guaranteed because they
could access each other through the peer-to-peer accesses which by-pass
the IOMMU protection.
This checks and sets the default DMA mode during driver binding, and
cleanups during driver unbinding. In the default mode, the device DMA is
managed by the device driver which handles DMA operations through the
kernel DMA APIs (see Documentation/core-api/dma-api.rst).
For cases where the devices are assigned for userspace control through the
userspace driver framework(i.e. VFIO), the drivers(for example, vfio_pci/
vfio_platfrom etc.) may set a new flag (driver_managed_dma) to skip this
default setting in the assumption that the drivers know what they are
doing with the device DMA.
Calling iommu_device_use_default_domain() before {of,acpi}_dma_configure
is currently a problem. As things stand, the IOMMU driver ignored the
initial iommu_probe_device() call when the device was added, since at
that point it had no fwspec yet. In this situation,
{of,acpi}_iommu_configure() are retriggering iommu_probe_device() after
the IOMMU driver has seen the firmware data via .of_xlate to learn that
it actually responsible for the given device. As the result, before
that gets fixed, iommu_use_default_domain() goes at the end, and calls
arch_teardown_dma_ops() if it fails.
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Bjorn Helgaas <bhelgaas@google.com>
Cc: Stuart Yoder <stuyoder@gmail.com>
Cc: Laurentiu Tudor <laurentiu.tudor@nxp.com>
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
Reviewed-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
Reviewed-by: Robin Murphy <robin.murphy@arm.com>
Tested-by: Eric Auger <eric.auger@redhat.com>
Link: https://lore.kernel.org/r/20220418005000.897664-5-baolu.lu@linux.intel.com
Signed-off-by: Joerg Roedel <jroedel@suse.de>
2022-04-18 08:49:53 +08:00
|
|
|
.dma_cleanup = amba_dma_cleanup,
|
2013-12-09 17:38:20 +08:00
|
|
|
.pm = &amba_pm,
|
2005-04-17 06:20:36 +08:00
|
|
|
};
|
2018-05-15 04:51:55 +08:00
|
|
|
EXPORT_SYMBOL_GPL(amba_bustype);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
static int __init amba_init(void)
|
|
|
|
{
|
|
|
|
return bus_register(&amba_bustype);
|
|
|
|
}
|
|
|
|
|
|
|
|
postcore_initcall(amba_init);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* amba_driver_register - register an AMBA device driver
|
|
|
|
* @drv: amba device driver structure
|
|
|
|
*
|
|
|
|
* Register an AMBA device driver with the Linux device model
|
|
|
|
* core. If devices pre-exist, the drivers probe function will
|
|
|
|
* be called.
|
|
|
|
*/
|
|
|
|
int amba_driver_register(struct amba_driver *drv)
|
|
|
|
{
|
2021-01-27 00:58:31 +08:00
|
|
|
if (!drv->probe)
|
|
|
|
return -EINVAL;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2021-01-27 00:58:31 +08:00
|
|
|
drv->drv.bus = &amba_bustype;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
return driver_register(&drv->drv);
|
|
|
|
}
|
2021-12-18 16:30:42 +08:00
|
|
|
EXPORT_SYMBOL(amba_driver_register);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
/**
|
|
|
|
* amba_driver_unregister - remove an AMBA device driver
|
|
|
|
* @drv: AMBA device driver structure to remove
|
|
|
|
*
|
|
|
|
* Unregister an AMBA device driver from the Linux device
|
|
|
|
* model. The device model will call the drivers remove function
|
|
|
|
* for each device the device driver is currently handling.
|
|
|
|
*/
|
|
|
|
void amba_driver_unregister(struct amba_driver *drv)
|
|
|
|
{
|
|
|
|
driver_unregister(&drv->drv);
|
|
|
|
}
|
2021-12-18 16:30:42 +08:00
|
|
|
EXPORT_SYMBOL(amba_driver_unregister);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
static void amba_device_release(struct device *dev)
|
|
|
|
{
|
|
|
|
struct amba_device *d = to_amba_device(dev);
|
|
|
|
|
|
|
|
if (d->res.parent)
|
|
|
|
release_resource(&d->res);
|
|
|
|
kfree(d);
|
|
|
|
}
|
|
|
|
|
2022-04-18 14:07:13 +08:00
|
|
|
static int amba_read_periphid(struct amba_device *dev)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2022-04-18 14:07:13 +08:00
|
|
|
struct reset_control *rstc;
|
|
|
|
u32 size, pid, cid;
|
2005-04-17 06:20:36 +08:00
|
|
|
void __iomem *tmp;
|
|
|
|
int i, ret;
|
2021-08-23 17:41:43 +08:00
|
|
|
|
2022-04-18 14:07:13 +08:00
|
|
|
ret = dev_pm_domain_attach(&dev->dev, true);
|
2006-11-30 22:04:49 +08:00
|
|
|
if (ret)
|
|
|
|
goto err_out;
|
|
|
|
|
2022-04-18 14:07:13 +08:00
|
|
|
ret = amba_get_enable_pclk(dev);
|
|
|
|
if (ret)
|
|
|
|
goto err_pm;
|
2011-03-24 23:12:40 +08:00
|
|
|
|
2009-07-29 06:34:59 +08:00
|
|
|
/*
|
2022-04-18 14:07:13 +08:00
|
|
|
* Find reset control(s) of the amba bus and de-assert them.
|
2009-07-29 06:34:59 +08:00
|
|
|
*/
|
2022-04-18 14:07:13 +08:00
|
|
|
rstc = of_reset_control_array_get_optional_shared(dev->dev.of_node);
|
|
|
|
if (IS_ERR(rstc)) {
|
|
|
|
ret = PTR_ERR(rstc);
|
|
|
|
if (ret != -EPROBE_DEFER)
|
|
|
|
dev_err(&dev->dev, "can't get reset: %d\n", ret);
|
|
|
|
goto err_clk;
|
|
|
|
}
|
|
|
|
reset_control_deassert(rstc);
|
|
|
|
reset_control_put(rstc);
|
|
|
|
|
2009-07-29 06:34:59 +08:00
|
|
|
size = resource_size(&dev->res);
|
|
|
|
tmp = ioremap(dev->res.start, size);
|
2006-11-30 22:04:49 +08:00
|
|
|
if (!tmp) {
|
|
|
|
ret = -ENOMEM;
|
2022-04-18 14:07:13 +08:00
|
|
|
goto err_clk;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2006-11-30 22:04:49 +08:00
|
|
|
|
2022-04-18 14:07:13 +08:00
|
|
|
/*
|
|
|
|
* Read pid and cid based on size of resource
|
|
|
|
* they are located at end of region
|
|
|
|
*/
|
|
|
|
for (pid = 0, i = 0; i < 4; i++)
|
|
|
|
pid |= (readl(tmp + size - 0x20 + 4 * i) & 255) << (i * 8);
|
|
|
|
for (cid = 0, i = 0; i < 4; i++)
|
|
|
|
cid |= (readl(tmp + size - 0x10 + 4 * i) & 255) << (i * 8);
|
2019-02-13 21:41:50 +08:00
|
|
|
|
2022-04-18 14:07:13 +08:00
|
|
|
if (cid == CORESIGHT_CID) {
|
|
|
|
/* set the base to the start of the last 4k block */
|
|
|
|
void __iomem *csbase = tmp + size - 4096;
|
2006-11-30 22:04:49 +08:00
|
|
|
|
2022-04-18 14:07:13 +08:00
|
|
|
dev->uci.devarch = readl(csbase + UCI_REG_DEVARCH_OFFSET);
|
|
|
|
dev->uci.devtype = readl(csbase + UCI_REG_DEVTYPE_OFFSET) & 0xff;
|
|
|
|
}
|
2010-07-15 17:47:14 +08:00
|
|
|
|
2022-04-18 14:07:13 +08:00
|
|
|
if (cid == AMBA_CID || cid == CORESIGHT_CID) {
|
|
|
|
dev->periphid = pid;
|
|
|
|
dev->cid = cid;
|
2006-11-30 22:04:49 +08:00
|
|
|
}
|
|
|
|
|
2022-04-18 14:07:13 +08:00
|
|
|
if (!dev->periphid)
|
|
|
|
ret = -ENODEV;
|
|
|
|
|
2010-07-15 17:47:14 +08:00
|
|
|
iounmap(tmp);
|
2022-04-18 14:07:13 +08:00
|
|
|
|
|
|
|
err_clk:
|
|
|
|
amba_put_disable_pclk(dev);
|
|
|
|
err_pm:
|
2016-04-21 14:58:35 +08:00
|
|
|
dev_pm_domain_detach(&dev->dev, true);
|
2022-04-18 14:07:13 +08:00
|
|
|
err_out:
|
|
|
|
return ret;
|
|
|
|
}
|
2010-07-15 17:47:14 +08:00
|
|
|
|
2022-04-18 14:07:13 +08:00
|
|
|
static int amba_device_try_add(struct amba_device *dev, struct resource *parent)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = request_resource(parent, &dev->res);
|
2010-07-15 17:47:14 +08:00
|
|
|
if (ret)
|
2022-04-18 14:07:13 +08:00
|
|
|
goto err_out;
|
|
|
|
|
|
|
|
/* Hard-coded primecell ID instead of plug-n-play */
|
|
|
|
if (dev->periphid != 0)
|
|
|
|
goto skip_probe;
|
2010-07-15 17:47:14 +08:00
|
|
|
|
2022-04-18 14:07:13 +08:00
|
|
|
ret = amba_read_periphid(dev);
|
2022-05-24 16:03:46 +08:00
|
|
|
if (ret)
|
2022-04-18 14:07:13 +08:00
|
|
|
goto err_release;
|
2022-04-18 14:07:14 +08:00
|
|
|
|
2022-04-18 14:07:13 +08:00
|
|
|
skip_probe:
|
2009-07-06 05:39:08 +08:00
|
|
|
ret = device_add(&dev->dev);
|
2022-04-18 14:07:13 +08:00
|
|
|
err_release:
|
2021-12-03 17:25:20 +08:00
|
|
|
if (ret)
|
|
|
|
release_resource(&dev->res);
|
2022-04-18 14:07:13 +08:00
|
|
|
err_out:
|
2005-04-17 06:20:36 +08:00
|
|
|
return ret;
|
|
|
|
}
|
2016-04-21 14:58:35 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Registration of AMBA device require reading its pid and cid registers.
|
|
|
|
* To do this, the device must be turned on (if it is a part of power domain)
|
|
|
|
* and have clocks enabled. However in some cases those resources might not be
|
|
|
|
* yet available. Returning EPROBE_DEFER is not a solution in such case,
|
|
|
|
* because callers don't handle this special error code. Instead such devices
|
|
|
|
* are added to the special list and their registration is retried from
|
|
|
|
* periodic worker, until all resources are available and registration succeeds.
|
|
|
|
*/
|
|
|
|
struct deferred_device {
|
|
|
|
struct amba_device *dev;
|
|
|
|
struct resource *parent;
|
|
|
|
struct list_head node;
|
|
|
|
};
|
|
|
|
|
|
|
|
static LIST_HEAD(deferred_devices);
|
|
|
|
static DEFINE_MUTEX(deferred_devices_lock);
|
|
|
|
|
|
|
|
static void amba_deferred_retry_func(struct work_struct *dummy);
|
|
|
|
static DECLARE_DELAYED_WORK(deferred_retry_work, amba_deferred_retry_func);
|
|
|
|
|
|
|
|
#define DEFERRED_DEVICE_TIMEOUT (msecs_to_jiffies(5 * 1000))
|
|
|
|
|
2020-04-30 04:58:12 +08:00
|
|
|
static int amba_deferred_retry(void)
|
2016-04-21 14:58:35 +08:00
|
|
|
{
|
|
|
|
struct deferred_device *ddev, *tmp;
|
|
|
|
|
|
|
|
mutex_lock(&deferred_devices_lock);
|
|
|
|
|
|
|
|
list_for_each_entry_safe(ddev, tmp, &deferred_devices, node) {
|
|
|
|
int ret = amba_device_try_add(ddev->dev, ddev->parent);
|
|
|
|
|
|
|
|
if (ret == -EPROBE_DEFER)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
list_del_init(&ddev->node);
|
2022-05-24 16:03:46 +08:00
|
|
|
amba_device_put(ddev->dev);
|
2016-04-21 14:58:35 +08:00
|
|
|
kfree(ddev);
|
|
|
|
}
|
|
|
|
|
2020-04-30 04:58:12 +08:00
|
|
|
mutex_unlock(&deferred_devices_lock);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
late_initcall(amba_deferred_retry);
|
|
|
|
|
|
|
|
static void amba_deferred_retry_func(struct work_struct *dummy)
|
|
|
|
{
|
|
|
|
amba_deferred_retry();
|
|
|
|
|
2016-04-21 14:58:35 +08:00
|
|
|
if (!list_empty(&deferred_devices))
|
|
|
|
schedule_delayed_work(&deferred_retry_work,
|
|
|
|
DEFERRED_DEVICE_TIMEOUT);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* amba_device_add - add a previously allocated AMBA device structure
|
|
|
|
* @dev: AMBA device allocated by amba_device_alloc
|
|
|
|
* @parent: resource parent for this devices resources
|
|
|
|
*
|
|
|
|
* Claim the resource, and read the device cell ID if not already
|
|
|
|
* initialized. Register the AMBA device with the Linux device
|
|
|
|
* manager.
|
|
|
|
*/
|
|
|
|
int amba_device_add(struct amba_device *dev, struct resource *parent)
|
|
|
|
{
|
|
|
|
int ret = amba_device_try_add(dev, parent);
|
|
|
|
|
|
|
|
if (ret == -EPROBE_DEFER) {
|
|
|
|
struct deferred_device *ddev;
|
|
|
|
|
|
|
|
ddev = kmalloc(sizeof(*ddev), GFP_KERNEL);
|
|
|
|
if (!ddev)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
ddev->dev = dev;
|
|
|
|
ddev->parent = parent;
|
|
|
|
ret = 0;
|
|
|
|
|
|
|
|
mutex_lock(&deferred_devices_lock);
|
|
|
|
|
|
|
|
if (list_empty(&deferred_devices))
|
|
|
|
schedule_delayed_work(&deferred_retry_work,
|
|
|
|
DEFERRED_DEVICE_TIMEOUT);
|
|
|
|
list_add_tail(&ddev->node, &deferred_devices);
|
|
|
|
|
|
|
|
mutex_unlock(&deferred_devices_lock);
|
|
|
|
}
|
|
|
|
return ret;
|
|
|
|
}
|
2011-12-18 19:07:47 +08:00
|
|
|
EXPORT_SYMBOL_GPL(amba_device_add);
|
|
|
|
|
|
|
|
static void amba_device_initialize(struct amba_device *dev, const char *name)
|
|
|
|
{
|
|
|
|
device_initialize(&dev->dev);
|
|
|
|
if (name)
|
|
|
|
dev_set_name(&dev->dev, "%s", name);
|
|
|
|
dev->dev.release = amba_device_release;
|
|
|
|
dev->dev.bus = &amba_bustype;
|
2013-06-27 17:25:33 +08:00
|
|
|
dev->dev.dma_mask = &dev->dev.coherent_dma_mask;
|
2020-04-22 18:10:13 +08:00
|
|
|
dev->dev.dma_parms = &dev->dma_parms;
|
2011-12-18 19:07:47 +08:00
|
|
|
dev->res.name = dev_name(&dev->dev);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* amba_device_alloc - allocate an AMBA device
|
|
|
|
* @name: sysfs name of the AMBA device
|
|
|
|
* @base: base of AMBA device
|
|
|
|
* @size: size of AMBA device
|
|
|
|
*
|
|
|
|
* Allocate and initialize an AMBA device structure. Returns %NULL
|
|
|
|
* on failure.
|
|
|
|
*/
|
|
|
|
struct amba_device *amba_device_alloc(const char *name, resource_size_t base,
|
|
|
|
size_t size)
|
|
|
|
{
|
|
|
|
struct amba_device *dev;
|
|
|
|
|
|
|
|
dev = kzalloc(sizeof(*dev), GFP_KERNEL);
|
|
|
|
if (dev) {
|
|
|
|
amba_device_initialize(dev, name);
|
|
|
|
dev->res.start = base;
|
|
|
|
dev->res.end = base + size - 1;
|
|
|
|
dev->res.flags = IORESOURCE_MEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
return dev;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(amba_device_alloc);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* amba_device_register - register an AMBA device
|
|
|
|
* @dev: AMBA device to register
|
|
|
|
* @parent: parent memory resource
|
|
|
|
*
|
|
|
|
* Setup the AMBA device, reading the cell ID if present.
|
|
|
|
* Claim the resource, and register the AMBA device with
|
|
|
|
* the Linux device manager.
|
|
|
|
*/
|
|
|
|
int amba_device_register(struct amba_device *dev, struct resource *parent)
|
|
|
|
{
|
|
|
|
amba_device_initialize(dev, dev->dev.init_name);
|
|
|
|
dev->dev.init_name = NULL;
|
|
|
|
|
|
|
|
return amba_device_add(dev, parent);
|
|
|
|
}
|
2021-12-18 16:30:42 +08:00
|
|
|
EXPORT_SYMBOL(amba_device_register);
|
2011-12-18 19:07:47 +08:00
|
|
|
|
|
|
|
/**
|
|
|
|
* amba_device_put - put an AMBA device
|
|
|
|
* @dev: AMBA device to put
|
|
|
|
*/
|
|
|
|
void amba_device_put(struct amba_device *dev)
|
|
|
|
{
|
|
|
|
put_device(&dev->dev);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(amba_device_put);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
/**
|
|
|
|
* amba_device_unregister - unregister an AMBA device
|
|
|
|
* @dev: AMBA device to remove
|
|
|
|
*
|
|
|
|
* Remove the specified AMBA device from the Linux device
|
|
|
|
* manager. All files associated with this object will be
|
|
|
|
* destroyed, and device drivers notified that the device has
|
|
|
|
* been removed. The AMBA device's resources including
|
|
|
|
* the amba_device structure will be freed once all
|
|
|
|
* references to it have been dropped.
|
|
|
|
*/
|
|
|
|
void amba_device_unregister(struct amba_device *dev)
|
|
|
|
{
|
|
|
|
device_unregister(&dev->dev);
|
|
|
|
}
|
2021-12-18 16:30:42 +08:00
|
|
|
EXPORT_SYMBOL(amba_device_unregister);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
/**
|
|
|
|
* amba_request_regions - request all mem regions associated with device
|
|
|
|
* @dev: amba_device structure for device
|
|
|
|
* @name: name, or NULL to use driver name
|
|
|
|
*/
|
|
|
|
int amba_request_regions(struct amba_device *dev, const char *name)
|
|
|
|
{
|
|
|
|
int ret = 0;
|
2009-07-29 06:34:59 +08:00
|
|
|
u32 size;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
if (!name)
|
|
|
|
name = dev->dev.driver->name;
|
|
|
|
|
2009-07-29 06:34:59 +08:00
|
|
|
size = resource_size(&dev->res);
|
|
|
|
|
|
|
|
if (!request_mem_region(dev->res.start, size, name))
|
2005-04-17 06:20:36 +08:00
|
|
|
ret = -EBUSY;
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
2021-12-18 16:30:42 +08:00
|
|
|
EXPORT_SYMBOL(amba_request_regions);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
/**
|
2011-03-31 09:57:33 +08:00
|
|
|
* amba_release_regions - release mem regions associated with device
|
2005-04-17 06:20:36 +08:00
|
|
|
* @dev: amba_device structure for device
|
|
|
|
*
|
|
|
|
* Release regions claimed by a successful call to amba_request_regions.
|
|
|
|
*/
|
|
|
|
void amba_release_regions(struct amba_device *dev)
|
|
|
|
{
|
2009-07-29 06:34:59 +08:00
|
|
|
u32 size;
|
|
|
|
|
|
|
|
size = resource_size(&dev->res);
|
|
|
|
release_mem_region(dev->res.start, size);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(amba_release_regions);
|