cxl/pci: Reserve individual register block regions

Some hardware implementations mix component and device registers into
the same BAR and the driver stack is going to need independent mapping
implementations for those 2 cases.  Furthermore, it will be nice to have
finer grained mappings should user space want to map some register
blocks.

Now that individual register blocks are mapped; those blocks regions
should be reserved individually to fully separate the register blocks.

Release the 'global' memory reservation and create individual register
block region reservations through devm.

NOTE: pci_release_mem_regions() is still compatible with
pcim_enable_device() because it removes the automatic region release
when called.  So preserve the pcim_enable_device() so that the pcim
interface can be called if needed.

Reviewed-by: Jonathan Cameron <Jonathan.Cameron@huawei.com>
Signed-off-by: Ira Weiny <ira.weiny@intel.com>
Link: https://lore.kernel.org/r/20210604005316.4187340-1-ira.weiny@intel.com
Signed-off-by: Dan Williams <dan.j.williams@intel.com>
This commit is contained in:
Ira Weiny 2021-06-03 17:53:16 -07:00 committed by Dan Williams
parent 30af97296f
commit 9a016527dc
2 changed files with 34 additions and 4 deletions

View File

@ -79,11 +79,33 @@ void cxl_probe_device_regs(struct device *dev, void __iomem *base,
}
EXPORT_SYMBOL_GPL(cxl_probe_device_regs);
static void __iomem *devm_cxl_iomap_block(struct pci_dev *pdev,
resource_size_t addr,
resource_size_t length)
{
struct device *dev = &pdev->dev;
void __iomem *ret_val;
struct resource *res;
res = devm_request_mem_region(dev, addr, length, pci_name(pdev));
if (!res) {
resource_size_t end = addr + length - 1;
dev_err(dev, "Failed to request region %pa-%pa\n", &addr, &end);
return NULL;
}
ret_val = devm_ioremap(dev, addr, length);
if (!ret_val)
dev_err(dev, "Failed to map region %pr\n", res);
return ret_val;
}
int cxl_map_device_regs(struct pci_dev *pdev,
struct cxl_device_regs *regs,
struct cxl_register_map *map)
{
struct device *dev = &pdev->dev;
resource_size_t phys_addr;
phys_addr = pci_resource_start(pdev, map->barno);
@ -95,7 +117,9 @@ int cxl_map_device_regs(struct pci_dev *pdev,
addr = phys_addr + map->device_map.status.offset;
length = map->device_map.status.size;
regs->status = devm_ioremap(dev, addr, length);
regs->status = devm_cxl_iomap_block(pdev, addr, length);
if (!regs->status)
return -ENOMEM;
}
if (map->device_map.mbox.valid) {
@ -104,7 +128,9 @@ int cxl_map_device_regs(struct pci_dev *pdev,
addr = phys_addr + map->device_map.mbox.offset;
length = map->device_map.mbox.size;
regs->mbox = devm_ioremap(dev, addr, length);
regs->mbox = devm_cxl_iomap_block(pdev, addr, length);
if (!regs->mbox)
return -ENOMEM;
}
if (map->device_map.memdev.valid) {
@ -113,7 +139,9 @@ int cxl_map_device_regs(struct pci_dev *pdev,
addr = phys_addr + map->device_map.memdev.offset;
length = map->device_map.memdev.size;
regs->memdev = devm_ioremap(dev, addr, length);
regs->memdev = devm_cxl_iomap_block(pdev, addr, length);
if (!regs->memdev)
return -ENOMEM;
}
return 0;

View File

@ -1110,6 +1110,8 @@ static int cxl_mem_setup_regs(struct cxl_mem *cxlm)
goto free_maps;
}
pci_release_mem_regions(pdev);
list_for_each_entry(map, &register_maps, list) {
ret = cxl_map_regs(cxlm, map);
if (ret)