964 lines
23 KiB
C
964 lines
23 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/**
|
|
* Host side test driver to test endpoint functionality
|
|
*
|
|
* Copyright (C) 2017 Texas Instruments
|
|
* Author: Kishon Vijay Abraham I <kishon@ti.com>
|
|
*/
|
|
|
|
#include <linux/crc32.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/io.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/miscdevice.h>
|
|
#include <linux/module.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/random.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/uaccess.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/pci_ids.h>
|
|
|
|
#include <linux/pci_regs.h>
|
|
|
|
#include <uapi/linux/pcitest.h>
|
|
|
|
#define DRV_MODULE_NAME "pci-endpoint-test"
|
|
|
|
#define IRQ_TYPE_UNDEFINED -1
|
|
#define IRQ_TYPE_LEGACY 0
|
|
#define IRQ_TYPE_MSI 1
|
|
#define IRQ_TYPE_MSIX 2
|
|
|
|
#define PCI_ENDPOINT_TEST_MAGIC 0x0
|
|
|
|
#define PCI_ENDPOINT_TEST_COMMAND 0x4
|
|
#define COMMAND_RAISE_LEGACY_IRQ BIT(0)
|
|
#define COMMAND_RAISE_MSI_IRQ BIT(1)
|
|
#define COMMAND_RAISE_MSIX_IRQ BIT(2)
|
|
#define COMMAND_READ BIT(3)
|
|
#define COMMAND_WRITE BIT(4)
|
|
#define COMMAND_COPY BIT(5)
|
|
|
|
#define PCI_ENDPOINT_TEST_STATUS 0x8
|
|
#define STATUS_READ_SUCCESS BIT(0)
|
|
#define STATUS_READ_FAIL BIT(1)
|
|
#define STATUS_WRITE_SUCCESS BIT(2)
|
|
#define STATUS_WRITE_FAIL BIT(3)
|
|
#define STATUS_COPY_SUCCESS BIT(4)
|
|
#define STATUS_COPY_FAIL BIT(5)
|
|
#define STATUS_IRQ_RAISED BIT(6)
|
|
#define STATUS_SRC_ADDR_INVALID BIT(7)
|
|
#define STATUS_DST_ADDR_INVALID BIT(8)
|
|
|
|
#define PCI_ENDPOINT_TEST_LOWER_SRC_ADDR 0x0c
|
|
#define PCI_ENDPOINT_TEST_UPPER_SRC_ADDR 0x10
|
|
|
|
#define PCI_ENDPOINT_TEST_LOWER_DST_ADDR 0x14
|
|
#define PCI_ENDPOINT_TEST_UPPER_DST_ADDR 0x18
|
|
|
|
#define PCI_ENDPOINT_TEST_SIZE 0x1c
|
|
#define PCI_ENDPOINT_TEST_CHECKSUM 0x20
|
|
|
|
#define PCI_ENDPOINT_TEST_IRQ_TYPE 0x24
|
|
#define PCI_ENDPOINT_TEST_IRQ_NUMBER 0x28
|
|
|
|
#define PCI_ENDPOINT_TEST_FLAGS 0x2c
|
|
#define FLAG_USE_DMA BIT(0)
|
|
|
|
#define PCI_DEVICE_ID_TI_AM654 0xb00c
|
|
|
|
#define is_am654_pci_dev(pdev) \
|
|
((pdev)->device == PCI_DEVICE_ID_TI_AM654)
|
|
|
|
#define PCI_DEVICE_ID_RENESAS_R8A774C0 0x002d
|
|
|
|
static DEFINE_IDA(pci_endpoint_test_ida);
|
|
|
|
#define to_endpoint_test(priv) container_of((priv), struct pci_endpoint_test, \
|
|
miscdev)
|
|
|
|
static bool no_msi;
|
|
module_param(no_msi, bool, 0444);
|
|
MODULE_PARM_DESC(no_msi, "Disable MSI interrupt in pci_endpoint_test");
|
|
|
|
static int irq_type = IRQ_TYPE_MSI;
|
|
module_param(irq_type, int, 0444);
|
|
MODULE_PARM_DESC(irq_type, "IRQ mode selection in pci_endpoint_test (0 - Legacy, 1 - MSI, 2 - MSI-X)");
|
|
|
|
enum pci_barno {
|
|
BAR_0,
|
|
BAR_1,
|
|
BAR_2,
|
|
BAR_3,
|
|
BAR_4,
|
|
BAR_5,
|
|
};
|
|
|
|
struct pci_endpoint_test {
|
|
struct pci_dev *pdev;
|
|
void __iomem *base;
|
|
void __iomem *bar[PCI_STD_NUM_BARS];
|
|
struct completion irq_raised;
|
|
int last_irq;
|
|
int num_irqs;
|
|
int irq_type;
|
|
/* mutex to protect the ioctls */
|
|
struct mutex mutex;
|
|
struct miscdevice miscdev;
|
|
enum pci_barno test_reg_bar;
|
|
size_t alignment;
|
|
const char *name;
|
|
};
|
|
|
|
struct pci_endpoint_test_data {
|
|
enum pci_barno test_reg_bar;
|
|
size_t alignment;
|
|
int irq_type;
|
|
};
|
|
|
|
static inline u32 pci_endpoint_test_readl(struct pci_endpoint_test *test,
|
|
u32 offset)
|
|
{
|
|
return readl(test->base + offset);
|
|
}
|
|
|
|
static inline void pci_endpoint_test_writel(struct pci_endpoint_test *test,
|
|
u32 offset, u32 value)
|
|
{
|
|
writel(value, test->base + offset);
|
|
}
|
|
|
|
static inline u32 pci_endpoint_test_bar_readl(struct pci_endpoint_test *test,
|
|
int bar, int offset)
|
|
{
|
|
return readl(test->bar[bar] + offset);
|
|
}
|
|
|
|
static inline void pci_endpoint_test_bar_writel(struct pci_endpoint_test *test,
|
|
int bar, u32 offset, u32 value)
|
|
{
|
|
writel(value, test->bar[bar] + offset);
|
|
}
|
|
|
|
static irqreturn_t pci_endpoint_test_irqhandler(int irq, void *dev_id)
|
|
{
|
|
struct pci_endpoint_test *test = dev_id;
|
|
u32 reg;
|
|
|
|
reg = pci_endpoint_test_readl(test, PCI_ENDPOINT_TEST_STATUS);
|
|
if (reg & STATUS_IRQ_RAISED) {
|
|
test->last_irq = irq;
|
|
complete(&test->irq_raised);
|
|
reg &= ~STATUS_IRQ_RAISED;
|
|
}
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_STATUS,
|
|
reg);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static void pci_endpoint_test_free_irq_vectors(struct pci_endpoint_test *test)
|
|
{
|
|
struct pci_dev *pdev = test->pdev;
|
|
|
|
pci_free_irq_vectors(pdev);
|
|
test->irq_type = IRQ_TYPE_UNDEFINED;
|
|
}
|
|
|
|
static bool pci_endpoint_test_alloc_irq_vectors(struct pci_endpoint_test *test,
|
|
int type)
|
|
{
|
|
int irq = -1;
|
|
struct pci_dev *pdev = test->pdev;
|
|
struct device *dev = &pdev->dev;
|
|
bool res = true;
|
|
|
|
switch (type) {
|
|
case IRQ_TYPE_LEGACY:
|
|
irq = pci_alloc_irq_vectors(pdev, 1, 1, PCI_IRQ_LEGACY);
|
|
if (irq < 0)
|
|
dev_err(dev, "Failed to get Legacy interrupt\n");
|
|
break;
|
|
case IRQ_TYPE_MSI:
|
|
irq = pci_alloc_irq_vectors(pdev, 1, 32, PCI_IRQ_MSI);
|
|
if (irq < 0)
|
|
dev_err(dev, "Failed to get MSI interrupts\n");
|
|
break;
|
|
case IRQ_TYPE_MSIX:
|
|
irq = pci_alloc_irq_vectors(pdev, 1, 2048, PCI_IRQ_MSIX);
|
|
if (irq < 0)
|
|
dev_err(dev, "Failed to get MSI-X interrupts\n");
|
|
break;
|
|
default:
|
|
dev_err(dev, "Invalid IRQ type selected\n");
|
|
}
|
|
|
|
if (irq < 0) {
|
|
irq = 0;
|
|
res = false;
|
|
}
|
|
|
|
test->irq_type = type;
|
|
test->num_irqs = irq;
|
|
|
|
return res;
|
|
}
|
|
|
|
static void pci_endpoint_test_release_irq(struct pci_endpoint_test *test)
|
|
{
|
|
int i;
|
|
struct pci_dev *pdev = test->pdev;
|
|
struct device *dev = &pdev->dev;
|
|
|
|
for (i = 0; i < test->num_irqs; i++)
|
|
devm_free_irq(dev, pci_irq_vector(pdev, i), test);
|
|
|
|
test->num_irqs = 0;
|
|
}
|
|
|
|
static bool pci_endpoint_test_request_irq(struct pci_endpoint_test *test)
|
|
{
|
|
int i;
|
|
int err;
|
|
struct pci_dev *pdev = test->pdev;
|
|
struct device *dev = &pdev->dev;
|
|
|
|
for (i = 0; i < test->num_irqs; i++) {
|
|
err = devm_request_irq(dev, pci_irq_vector(pdev, i),
|
|
pci_endpoint_test_irqhandler,
|
|
IRQF_SHARED, test->name, test);
|
|
if (err)
|
|
goto fail;
|
|
}
|
|
|
|
return true;
|
|
|
|
fail:
|
|
switch (irq_type) {
|
|
case IRQ_TYPE_LEGACY:
|
|
dev_err(dev, "Failed to request IRQ %d for Legacy\n",
|
|
pci_irq_vector(pdev, i));
|
|
break;
|
|
case IRQ_TYPE_MSI:
|
|
dev_err(dev, "Failed to request IRQ %d for MSI %d\n",
|
|
pci_irq_vector(pdev, i),
|
|
i + 1);
|
|
break;
|
|
case IRQ_TYPE_MSIX:
|
|
dev_err(dev, "Failed to request IRQ %d for MSI-X %d\n",
|
|
pci_irq_vector(pdev, i),
|
|
i + 1);
|
|
break;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
static bool pci_endpoint_test_bar(struct pci_endpoint_test *test,
|
|
enum pci_barno barno)
|
|
{
|
|
int j;
|
|
u32 val;
|
|
int size;
|
|
struct pci_dev *pdev = test->pdev;
|
|
|
|
if (!test->bar[barno])
|
|
return false;
|
|
|
|
size = pci_resource_len(pdev, barno);
|
|
|
|
if (barno == test->test_reg_bar)
|
|
size = 0x4;
|
|
|
|
for (j = 0; j < size; j += 4)
|
|
pci_endpoint_test_bar_writel(test, barno, j, 0xA0A0A0A0);
|
|
|
|
for (j = 0; j < size; j += 4) {
|
|
val = pci_endpoint_test_bar_readl(test, barno, j);
|
|
if (val != 0xA0A0A0A0)
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool pci_endpoint_test_legacy_irq(struct pci_endpoint_test *test)
|
|
{
|
|
u32 val;
|
|
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE,
|
|
IRQ_TYPE_LEGACY);
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, 0);
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND,
|
|
COMMAND_RAISE_LEGACY_IRQ);
|
|
val = wait_for_completion_timeout(&test->irq_raised,
|
|
msecs_to_jiffies(1000));
|
|
if (!val)
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool pci_endpoint_test_msi_irq(struct pci_endpoint_test *test,
|
|
u16 msi_num, bool msix)
|
|
{
|
|
u32 val;
|
|
struct pci_dev *pdev = test->pdev;
|
|
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE,
|
|
msix == false ? IRQ_TYPE_MSI :
|
|
IRQ_TYPE_MSIX);
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, msi_num);
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND,
|
|
msix == false ? COMMAND_RAISE_MSI_IRQ :
|
|
COMMAND_RAISE_MSIX_IRQ);
|
|
val = wait_for_completion_timeout(&test->irq_raised,
|
|
msecs_to_jiffies(1000));
|
|
if (!val)
|
|
return false;
|
|
|
|
if (pci_irq_vector(pdev, msi_num - 1) == test->last_irq)
|
|
return true;
|
|
|
|
return false;
|
|
}
|
|
|
|
static bool pci_endpoint_test_copy(struct pci_endpoint_test *test,
|
|
unsigned long arg)
|
|
{
|
|
struct pci_endpoint_test_xfer_param param;
|
|
bool ret = false;
|
|
void *src_addr;
|
|
void *dst_addr;
|
|
u32 flags = 0;
|
|
bool use_dma;
|
|
size_t size;
|
|
dma_addr_t src_phys_addr;
|
|
dma_addr_t dst_phys_addr;
|
|
struct pci_dev *pdev = test->pdev;
|
|
struct device *dev = &pdev->dev;
|
|
void *orig_src_addr;
|
|
dma_addr_t orig_src_phys_addr;
|
|
void *orig_dst_addr;
|
|
dma_addr_t orig_dst_phys_addr;
|
|
size_t offset;
|
|
size_t alignment = test->alignment;
|
|
int irq_type = test->irq_type;
|
|
u32 src_crc32;
|
|
u32 dst_crc32;
|
|
int err;
|
|
|
|
err = copy_from_user(¶m, (void __user *)arg, sizeof(param));
|
|
if (err) {
|
|
dev_err(dev, "Failed to get transfer param\n");
|
|
return false;
|
|
}
|
|
|
|
size = param.size;
|
|
if (size > SIZE_MAX - alignment)
|
|
goto err;
|
|
|
|
use_dma = !!(param.flags & PCITEST_FLAGS_USE_DMA);
|
|
if (use_dma)
|
|
flags |= FLAG_USE_DMA;
|
|
|
|
if (irq_type < IRQ_TYPE_LEGACY || irq_type > IRQ_TYPE_MSIX) {
|
|
dev_err(dev, "Invalid IRQ type option\n");
|
|
goto err;
|
|
}
|
|
|
|
orig_src_addr = kzalloc(size + alignment, GFP_KERNEL);
|
|
if (!orig_src_addr) {
|
|
dev_err(dev, "Failed to allocate source buffer\n");
|
|
ret = false;
|
|
goto err;
|
|
}
|
|
|
|
get_random_bytes(orig_src_addr, size + alignment);
|
|
orig_src_phys_addr = dma_map_single(dev, orig_src_addr,
|
|
size + alignment, DMA_TO_DEVICE);
|
|
if (dma_mapping_error(dev, orig_src_phys_addr)) {
|
|
dev_err(dev, "failed to map source buffer address\n");
|
|
ret = false;
|
|
goto err_src_phys_addr;
|
|
}
|
|
|
|
if (alignment && !IS_ALIGNED(orig_src_phys_addr, alignment)) {
|
|
src_phys_addr = PTR_ALIGN(orig_src_phys_addr, alignment);
|
|
offset = src_phys_addr - orig_src_phys_addr;
|
|
src_addr = orig_src_addr + offset;
|
|
} else {
|
|
src_phys_addr = orig_src_phys_addr;
|
|
src_addr = orig_src_addr;
|
|
}
|
|
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_LOWER_SRC_ADDR,
|
|
lower_32_bits(src_phys_addr));
|
|
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_UPPER_SRC_ADDR,
|
|
upper_32_bits(src_phys_addr));
|
|
|
|
src_crc32 = crc32_le(~0, src_addr, size);
|
|
|
|
orig_dst_addr = kzalloc(size + alignment, GFP_KERNEL);
|
|
if (!orig_dst_addr) {
|
|
dev_err(dev, "Failed to allocate destination address\n");
|
|
ret = false;
|
|
goto err_dst_addr;
|
|
}
|
|
|
|
orig_dst_phys_addr = dma_map_single(dev, orig_dst_addr,
|
|
size + alignment, DMA_FROM_DEVICE);
|
|
if (dma_mapping_error(dev, orig_dst_phys_addr)) {
|
|
dev_err(dev, "failed to map destination buffer address\n");
|
|
ret = false;
|
|
goto err_dst_phys_addr;
|
|
}
|
|
|
|
if (alignment && !IS_ALIGNED(orig_dst_phys_addr, alignment)) {
|
|
dst_phys_addr = PTR_ALIGN(orig_dst_phys_addr, alignment);
|
|
offset = dst_phys_addr - orig_dst_phys_addr;
|
|
dst_addr = orig_dst_addr + offset;
|
|
} else {
|
|
dst_phys_addr = orig_dst_phys_addr;
|
|
dst_addr = orig_dst_addr;
|
|
}
|
|
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_LOWER_DST_ADDR,
|
|
lower_32_bits(dst_phys_addr));
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_UPPER_DST_ADDR,
|
|
upper_32_bits(dst_phys_addr));
|
|
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_SIZE,
|
|
size);
|
|
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_FLAGS, flags);
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE, irq_type);
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, 1);
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND,
|
|
COMMAND_COPY);
|
|
|
|
wait_for_completion(&test->irq_raised);
|
|
|
|
dma_unmap_single(dev, orig_dst_phys_addr, size + alignment,
|
|
DMA_FROM_DEVICE);
|
|
|
|
dst_crc32 = crc32_le(~0, dst_addr, size);
|
|
if (dst_crc32 == src_crc32)
|
|
ret = true;
|
|
|
|
err_dst_phys_addr:
|
|
kfree(orig_dst_addr);
|
|
|
|
err_dst_addr:
|
|
dma_unmap_single(dev, orig_src_phys_addr, size + alignment,
|
|
DMA_TO_DEVICE);
|
|
|
|
err_src_phys_addr:
|
|
kfree(orig_src_addr);
|
|
|
|
err:
|
|
return ret;
|
|
}
|
|
|
|
static bool pci_endpoint_test_write(struct pci_endpoint_test *test,
|
|
unsigned long arg)
|
|
{
|
|
struct pci_endpoint_test_xfer_param param;
|
|
bool ret = false;
|
|
u32 flags = 0;
|
|
bool use_dma;
|
|
u32 reg;
|
|
void *addr;
|
|
dma_addr_t phys_addr;
|
|
struct pci_dev *pdev = test->pdev;
|
|
struct device *dev = &pdev->dev;
|
|
void *orig_addr;
|
|
dma_addr_t orig_phys_addr;
|
|
size_t offset;
|
|
size_t alignment = test->alignment;
|
|
int irq_type = test->irq_type;
|
|
size_t size;
|
|
u32 crc32;
|
|
int err;
|
|
|
|
err = copy_from_user(¶m, (void __user *)arg, sizeof(param));
|
|
if (err != 0) {
|
|
dev_err(dev, "Failed to get transfer param\n");
|
|
return false;
|
|
}
|
|
|
|
size = param.size;
|
|
if (size > SIZE_MAX - alignment)
|
|
goto err;
|
|
|
|
use_dma = !!(param.flags & PCITEST_FLAGS_USE_DMA);
|
|
if (use_dma)
|
|
flags |= FLAG_USE_DMA;
|
|
|
|
if (irq_type < IRQ_TYPE_LEGACY || irq_type > IRQ_TYPE_MSIX) {
|
|
dev_err(dev, "Invalid IRQ type option\n");
|
|
goto err;
|
|
}
|
|
|
|
orig_addr = kzalloc(size + alignment, GFP_KERNEL);
|
|
if (!orig_addr) {
|
|
dev_err(dev, "Failed to allocate address\n");
|
|
ret = false;
|
|
goto err;
|
|
}
|
|
|
|
get_random_bytes(orig_addr, size + alignment);
|
|
|
|
orig_phys_addr = dma_map_single(dev, orig_addr, size + alignment,
|
|
DMA_TO_DEVICE);
|
|
if (dma_mapping_error(dev, orig_phys_addr)) {
|
|
dev_err(dev, "failed to map source buffer address\n");
|
|
ret = false;
|
|
goto err_phys_addr;
|
|
}
|
|
|
|
if (alignment && !IS_ALIGNED(orig_phys_addr, alignment)) {
|
|
phys_addr = PTR_ALIGN(orig_phys_addr, alignment);
|
|
offset = phys_addr - orig_phys_addr;
|
|
addr = orig_addr + offset;
|
|
} else {
|
|
phys_addr = orig_phys_addr;
|
|
addr = orig_addr;
|
|
}
|
|
|
|
crc32 = crc32_le(~0, addr, size);
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_CHECKSUM,
|
|
crc32);
|
|
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_LOWER_SRC_ADDR,
|
|
lower_32_bits(phys_addr));
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_UPPER_SRC_ADDR,
|
|
upper_32_bits(phys_addr));
|
|
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_SIZE, size);
|
|
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_FLAGS, flags);
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE, irq_type);
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, 1);
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND,
|
|
COMMAND_READ);
|
|
|
|
wait_for_completion(&test->irq_raised);
|
|
|
|
reg = pci_endpoint_test_readl(test, PCI_ENDPOINT_TEST_STATUS);
|
|
if (reg & STATUS_READ_SUCCESS)
|
|
ret = true;
|
|
|
|
dma_unmap_single(dev, orig_phys_addr, size + alignment,
|
|
DMA_TO_DEVICE);
|
|
|
|
err_phys_addr:
|
|
kfree(orig_addr);
|
|
|
|
err:
|
|
return ret;
|
|
}
|
|
|
|
static bool pci_endpoint_test_read(struct pci_endpoint_test *test,
|
|
unsigned long arg)
|
|
{
|
|
struct pci_endpoint_test_xfer_param param;
|
|
bool ret = false;
|
|
u32 flags = 0;
|
|
bool use_dma;
|
|
size_t size;
|
|
void *addr;
|
|
dma_addr_t phys_addr;
|
|
struct pci_dev *pdev = test->pdev;
|
|
struct device *dev = &pdev->dev;
|
|
void *orig_addr;
|
|
dma_addr_t orig_phys_addr;
|
|
size_t offset;
|
|
size_t alignment = test->alignment;
|
|
int irq_type = test->irq_type;
|
|
u32 crc32;
|
|
int err;
|
|
|
|
err = copy_from_user(¶m, (void __user *)arg, sizeof(param));
|
|
if (err) {
|
|
dev_err(dev, "Failed to get transfer param\n");
|
|
return false;
|
|
}
|
|
|
|
size = param.size;
|
|
if (size > SIZE_MAX - alignment)
|
|
goto err;
|
|
|
|
use_dma = !!(param.flags & PCITEST_FLAGS_USE_DMA);
|
|
if (use_dma)
|
|
flags |= FLAG_USE_DMA;
|
|
|
|
if (irq_type < IRQ_TYPE_LEGACY || irq_type > IRQ_TYPE_MSIX) {
|
|
dev_err(dev, "Invalid IRQ type option\n");
|
|
goto err;
|
|
}
|
|
|
|
orig_addr = kzalloc(size + alignment, GFP_KERNEL);
|
|
if (!orig_addr) {
|
|
dev_err(dev, "Failed to allocate destination address\n");
|
|
ret = false;
|
|
goto err;
|
|
}
|
|
|
|
orig_phys_addr = dma_map_single(dev, orig_addr, size + alignment,
|
|
DMA_FROM_DEVICE);
|
|
if (dma_mapping_error(dev, orig_phys_addr)) {
|
|
dev_err(dev, "failed to map source buffer address\n");
|
|
ret = false;
|
|
goto err_phys_addr;
|
|
}
|
|
|
|
if (alignment && !IS_ALIGNED(orig_phys_addr, alignment)) {
|
|
phys_addr = PTR_ALIGN(orig_phys_addr, alignment);
|
|
offset = phys_addr - orig_phys_addr;
|
|
addr = orig_addr + offset;
|
|
} else {
|
|
phys_addr = orig_phys_addr;
|
|
addr = orig_addr;
|
|
}
|
|
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_LOWER_DST_ADDR,
|
|
lower_32_bits(phys_addr));
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_UPPER_DST_ADDR,
|
|
upper_32_bits(phys_addr));
|
|
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_SIZE, size);
|
|
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_FLAGS, flags);
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_TYPE, irq_type);
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_IRQ_NUMBER, 1);
|
|
pci_endpoint_test_writel(test, PCI_ENDPOINT_TEST_COMMAND,
|
|
COMMAND_WRITE);
|
|
|
|
wait_for_completion(&test->irq_raised);
|
|
|
|
dma_unmap_single(dev, orig_phys_addr, size + alignment,
|
|
DMA_FROM_DEVICE);
|
|
|
|
crc32 = crc32_le(~0, addr, size);
|
|
if (crc32 == pci_endpoint_test_readl(test, PCI_ENDPOINT_TEST_CHECKSUM))
|
|
ret = true;
|
|
|
|
err_phys_addr:
|
|
kfree(orig_addr);
|
|
err:
|
|
return ret;
|
|
}
|
|
|
|
static bool pci_endpoint_test_clear_irq(struct pci_endpoint_test *test)
|
|
{
|
|
pci_endpoint_test_release_irq(test);
|
|
pci_endpoint_test_free_irq_vectors(test);
|
|
return true;
|
|
}
|
|
|
|
static bool pci_endpoint_test_set_irq(struct pci_endpoint_test *test,
|
|
int req_irq_type)
|
|
{
|
|
struct pci_dev *pdev = test->pdev;
|
|
struct device *dev = &pdev->dev;
|
|
|
|
if (req_irq_type < IRQ_TYPE_LEGACY || req_irq_type > IRQ_TYPE_MSIX) {
|
|
dev_err(dev, "Invalid IRQ type option\n");
|
|
return false;
|
|
}
|
|
|
|
if (test->irq_type == req_irq_type)
|
|
return true;
|
|
|
|
pci_endpoint_test_release_irq(test);
|
|
pci_endpoint_test_free_irq_vectors(test);
|
|
|
|
if (!pci_endpoint_test_alloc_irq_vectors(test, req_irq_type))
|
|
goto err;
|
|
|
|
if (!pci_endpoint_test_request_irq(test))
|
|
goto err;
|
|
|
|
return true;
|
|
|
|
err:
|
|
pci_endpoint_test_free_irq_vectors(test);
|
|
return false;
|
|
}
|
|
|
|
static long pci_endpoint_test_ioctl(struct file *file, unsigned int cmd,
|
|
unsigned long arg)
|
|
{
|
|
int ret = -EINVAL;
|
|
enum pci_barno bar;
|
|
struct pci_endpoint_test *test = to_endpoint_test(file->private_data);
|
|
struct pci_dev *pdev = test->pdev;
|
|
|
|
mutex_lock(&test->mutex);
|
|
switch (cmd) {
|
|
case PCITEST_BAR:
|
|
bar = arg;
|
|
if (bar < 0 || bar > 5)
|
|
goto ret;
|
|
if (is_am654_pci_dev(pdev) && bar == BAR_0)
|
|
goto ret;
|
|
ret = pci_endpoint_test_bar(test, bar);
|
|
break;
|
|
case PCITEST_LEGACY_IRQ:
|
|
ret = pci_endpoint_test_legacy_irq(test);
|
|
break;
|
|
case PCITEST_MSI:
|
|
case PCITEST_MSIX:
|
|
ret = pci_endpoint_test_msi_irq(test, arg, cmd == PCITEST_MSIX);
|
|
break;
|
|
case PCITEST_WRITE:
|
|
ret = pci_endpoint_test_write(test, arg);
|
|
break;
|
|
case PCITEST_READ:
|
|
ret = pci_endpoint_test_read(test, arg);
|
|
break;
|
|
case PCITEST_COPY:
|
|
ret = pci_endpoint_test_copy(test, arg);
|
|
break;
|
|
case PCITEST_SET_IRQTYPE:
|
|
ret = pci_endpoint_test_set_irq(test, arg);
|
|
break;
|
|
case PCITEST_GET_IRQTYPE:
|
|
ret = irq_type;
|
|
break;
|
|
case PCITEST_CLEAR_IRQ:
|
|
ret = pci_endpoint_test_clear_irq(test);
|
|
break;
|
|
}
|
|
|
|
ret:
|
|
mutex_unlock(&test->mutex);
|
|
return ret;
|
|
}
|
|
|
|
static const struct file_operations pci_endpoint_test_fops = {
|
|
.owner = THIS_MODULE,
|
|
.unlocked_ioctl = pci_endpoint_test_ioctl,
|
|
};
|
|
|
|
static int pci_endpoint_test_probe(struct pci_dev *pdev,
|
|
const struct pci_device_id *ent)
|
|
{
|
|
int err;
|
|
int id;
|
|
char name[24];
|
|
enum pci_barno bar;
|
|
void __iomem *base;
|
|
struct device *dev = &pdev->dev;
|
|
struct pci_endpoint_test *test;
|
|
struct pci_endpoint_test_data *data;
|
|
enum pci_barno test_reg_bar = BAR_0;
|
|
struct miscdevice *misc_device;
|
|
|
|
if (pci_is_bridge(pdev))
|
|
return -ENODEV;
|
|
|
|
test = devm_kzalloc(dev, sizeof(*test), GFP_KERNEL);
|
|
if (!test)
|
|
return -ENOMEM;
|
|
|
|
test->test_reg_bar = 0;
|
|
test->alignment = 0;
|
|
test->pdev = pdev;
|
|
test->irq_type = IRQ_TYPE_UNDEFINED;
|
|
|
|
if (no_msi)
|
|
irq_type = IRQ_TYPE_LEGACY;
|
|
|
|
data = (struct pci_endpoint_test_data *)ent->driver_data;
|
|
if (data) {
|
|
test_reg_bar = data->test_reg_bar;
|
|
test->test_reg_bar = test_reg_bar;
|
|
test->alignment = data->alignment;
|
|
irq_type = data->irq_type;
|
|
}
|
|
|
|
init_completion(&test->irq_raised);
|
|
mutex_init(&test->mutex);
|
|
|
|
if ((dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(48)) != 0) &&
|
|
dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32)) != 0) {
|
|
dev_err(dev, "Cannot set DMA mask\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
err = pci_enable_device(pdev);
|
|
if (err) {
|
|
dev_err(dev, "Cannot enable PCI device\n");
|
|
return err;
|
|
}
|
|
|
|
err = pci_request_regions(pdev, DRV_MODULE_NAME);
|
|
if (err) {
|
|
dev_err(dev, "Cannot obtain PCI resources\n");
|
|
goto err_disable_pdev;
|
|
}
|
|
|
|
pci_set_master(pdev);
|
|
|
|
if (!pci_endpoint_test_alloc_irq_vectors(test, irq_type))
|
|
goto err_disable_irq;
|
|
|
|
for (bar = 0; bar < PCI_STD_NUM_BARS; bar++) {
|
|
if (pci_resource_flags(pdev, bar) & IORESOURCE_MEM) {
|
|
base = pci_ioremap_bar(pdev, bar);
|
|
if (!base) {
|
|
dev_err(dev, "Failed to read BAR%d\n", bar);
|
|
WARN_ON(bar == test_reg_bar);
|
|
}
|
|
test->bar[bar] = base;
|
|
}
|
|
}
|
|
|
|
test->base = test->bar[test_reg_bar];
|
|
if (!test->base) {
|
|
err = -ENOMEM;
|
|
dev_err(dev, "Cannot perform PCI test without BAR%d\n",
|
|
test_reg_bar);
|
|
goto err_iounmap;
|
|
}
|
|
|
|
pci_set_drvdata(pdev, test);
|
|
|
|
id = ida_simple_get(&pci_endpoint_test_ida, 0, 0, GFP_KERNEL);
|
|
if (id < 0) {
|
|
err = id;
|
|
dev_err(dev, "Unable to get id\n");
|
|
goto err_iounmap;
|
|
}
|
|
|
|
snprintf(name, sizeof(name), DRV_MODULE_NAME ".%d", id);
|
|
test->name = kstrdup(name, GFP_KERNEL);
|
|
if (!test->name) {
|
|
err = -ENOMEM;
|
|
goto err_ida_remove;
|
|
}
|
|
|
|
if (!pci_endpoint_test_request_irq(test))
|
|
goto err_kfree_test_name;
|
|
|
|
misc_device = &test->miscdev;
|
|
misc_device->minor = MISC_DYNAMIC_MINOR;
|
|
misc_device->name = kstrdup(name, GFP_KERNEL);
|
|
if (!misc_device->name) {
|
|
err = -ENOMEM;
|
|
goto err_release_irq;
|
|
}
|
|
misc_device->fops = &pci_endpoint_test_fops,
|
|
|
|
err = misc_register(misc_device);
|
|
if (err) {
|
|
dev_err(dev, "Failed to register device\n");
|
|
goto err_kfree_name;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_kfree_name:
|
|
kfree(misc_device->name);
|
|
|
|
err_release_irq:
|
|
pci_endpoint_test_release_irq(test);
|
|
|
|
err_kfree_test_name:
|
|
kfree(test->name);
|
|
|
|
err_ida_remove:
|
|
ida_simple_remove(&pci_endpoint_test_ida, id);
|
|
|
|
err_iounmap:
|
|
for (bar = 0; bar < PCI_STD_NUM_BARS; bar++) {
|
|
if (test->bar[bar])
|
|
pci_iounmap(pdev, test->bar[bar]);
|
|
}
|
|
|
|
err_disable_irq:
|
|
pci_endpoint_test_free_irq_vectors(test);
|
|
pci_release_regions(pdev);
|
|
|
|
err_disable_pdev:
|
|
pci_disable_device(pdev);
|
|
|
|
return err;
|
|
}
|
|
|
|
static void pci_endpoint_test_remove(struct pci_dev *pdev)
|
|
{
|
|
int id;
|
|
enum pci_barno bar;
|
|
struct pci_endpoint_test *test = pci_get_drvdata(pdev);
|
|
struct miscdevice *misc_device = &test->miscdev;
|
|
|
|
if (sscanf(misc_device->name, DRV_MODULE_NAME ".%d", &id) != 1)
|
|
return;
|
|
if (id < 0)
|
|
return;
|
|
|
|
misc_deregister(&test->miscdev);
|
|
kfree(misc_device->name);
|
|
kfree(test->name);
|
|
ida_simple_remove(&pci_endpoint_test_ida, id);
|
|
for (bar = 0; bar < PCI_STD_NUM_BARS; bar++) {
|
|
if (test->bar[bar])
|
|
pci_iounmap(pdev, test->bar[bar]);
|
|
}
|
|
|
|
pci_endpoint_test_release_irq(test);
|
|
pci_endpoint_test_free_irq_vectors(test);
|
|
|
|
pci_release_regions(pdev);
|
|
pci_disable_device(pdev);
|
|
}
|
|
|
|
static const struct pci_endpoint_test_data default_data = {
|
|
.test_reg_bar = BAR_0,
|
|
.alignment = SZ_4K,
|
|
.irq_type = IRQ_TYPE_MSI,
|
|
};
|
|
|
|
static const struct pci_endpoint_test_data am654_data = {
|
|
.test_reg_bar = BAR_2,
|
|
.alignment = SZ_64K,
|
|
.irq_type = IRQ_TYPE_MSI,
|
|
};
|
|
|
|
static const struct pci_device_id pci_endpoint_test_tbl[] = {
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_TI, PCI_DEVICE_ID_TI_DRA74x),
|
|
.driver_data = (kernel_ulong_t)&default_data,
|
|
},
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_TI, PCI_DEVICE_ID_TI_DRA72x),
|
|
.driver_data = (kernel_ulong_t)&default_data,
|
|
},
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_FREESCALE, 0x81c0) },
|
|
{ PCI_DEVICE_DATA(SYNOPSYS, EDDA, NULL) },
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_TI, PCI_DEVICE_ID_TI_AM654),
|
|
.driver_data = (kernel_ulong_t)&am654_data
|
|
},
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_RENESAS, PCI_DEVICE_ID_RENESAS_R8A774C0),
|
|
},
|
|
{ }
|
|
};
|
|
MODULE_DEVICE_TABLE(pci, pci_endpoint_test_tbl);
|
|
|
|
static struct pci_driver pci_endpoint_test_driver = {
|
|
.name = DRV_MODULE_NAME,
|
|
.id_table = pci_endpoint_test_tbl,
|
|
.probe = pci_endpoint_test_probe,
|
|
.remove = pci_endpoint_test_remove,
|
|
};
|
|
module_pci_driver(pci_endpoint_test_driver);
|
|
|
|
MODULE_DESCRIPTION("PCI ENDPOINT TEST HOST DRIVER");
|
|
MODULE_AUTHOR("Kishon Vijay Abraham I <kishon@ti.com>");
|
|
MODULE_LICENSE("GPL v2");
|