2005-04-17 06:20:36 +08:00
|
|
|
/* include this file if the platform implements the dma_ DMA Mapping API
|
|
|
|
* and wants to provide the pci_ DMA Mapping API in terms of it */
|
|
|
|
|
|
|
|
#ifndef _ASM_GENERIC_PCI_DMA_COMPAT_H
|
|
|
|
#define _ASM_GENERIC_PCI_DMA_COMPAT_H
|
|
|
|
|
|
|
|
#include <linux/dma-mapping.h>
|
|
|
|
|
2016-03-08 01:39:16 +08:00
|
|
|
/* This defines the direction arg to the DMA mapping routines. */
|
|
|
|
#define PCI_DMA_BIDIRECTIONAL 0
|
|
|
|
#define PCI_DMA_TODEVICE 1
|
|
|
|
#define PCI_DMA_FROMDEVICE 2
|
|
|
|
#define PCI_DMA_NONE 3
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
static inline void *
|
|
|
|
pci_alloc_consistent(struct pci_dev *hwdev, size_t size,
|
|
|
|
dma_addr_t *dma_handle)
|
|
|
|
{
|
|
|
|
return dma_alloc_coherent(hwdev == NULL ? NULL : &hwdev->dev, size, dma_handle, GFP_ATOMIC);
|
|
|
|
}
|
|
|
|
|
2014-08-09 05:24:08 +08:00
|
|
|
static inline void *
|
|
|
|
pci_zalloc_consistent(struct pci_dev *hwdev, size_t size,
|
|
|
|
dma_addr_t *dma_handle)
|
|
|
|
{
|
|
|
|
return dma_zalloc_coherent(hwdev == NULL ? NULL : &hwdev->dev,
|
|
|
|
size, dma_handle, GFP_ATOMIC);
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
static inline void
|
|
|
|
pci_free_consistent(struct pci_dev *hwdev, size_t size,
|
|
|
|
void *vaddr, dma_addr_t dma_handle)
|
|
|
|
{
|
|
|
|
dma_free_coherent(hwdev == NULL ? NULL : &hwdev->dev, size, vaddr, dma_handle);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline dma_addr_t
|
|
|
|
pci_map_single(struct pci_dev *hwdev, void *ptr, size_t size, int direction)
|
|
|
|
{
|
|
|
|
return dma_map_single(hwdev == NULL ? NULL : &hwdev->dev, ptr, size, (enum dma_data_direction)direction);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
pci_unmap_single(struct pci_dev *hwdev, dma_addr_t dma_addr,
|
|
|
|
size_t size, int direction)
|
|
|
|
{
|
|
|
|
dma_unmap_single(hwdev == NULL ? NULL : &hwdev->dev, dma_addr, size, (enum dma_data_direction)direction);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline dma_addr_t
|
|
|
|
pci_map_page(struct pci_dev *hwdev, struct page *page,
|
|
|
|
unsigned long offset, size_t size, int direction)
|
|
|
|
{
|
|
|
|
return dma_map_page(hwdev == NULL ? NULL : &hwdev->dev, page, offset, size, (enum dma_data_direction)direction);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
pci_unmap_page(struct pci_dev *hwdev, dma_addr_t dma_address,
|
|
|
|
size_t size, int direction)
|
|
|
|
{
|
|
|
|
dma_unmap_page(hwdev == NULL ? NULL : &hwdev->dev, dma_address, size, (enum dma_data_direction)direction);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int
|
|
|
|
pci_map_sg(struct pci_dev *hwdev, struct scatterlist *sg,
|
|
|
|
int nents, int direction)
|
|
|
|
{
|
|
|
|
return dma_map_sg(hwdev == NULL ? NULL : &hwdev->dev, sg, nents, (enum dma_data_direction)direction);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
pci_unmap_sg(struct pci_dev *hwdev, struct scatterlist *sg,
|
|
|
|
int nents, int direction)
|
|
|
|
{
|
|
|
|
dma_unmap_sg(hwdev == NULL ? NULL : &hwdev->dev, sg, nents, (enum dma_data_direction)direction);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
pci_dma_sync_single_for_cpu(struct pci_dev *hwdev, dma_addr_t dma_handle,
|
|
|
|
size_t size, int direction)
|
|
|
|
{
|
|
|
|
dma_sync_single_for_cpu(hwdev == NULL ? NULL : &hwdev->dev, dma_handle, size, (enum dma_data_direction)direction);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
pci_dma_sync_single_for_device(struct pci_dev *hwdev, dma_addr_t dma_handle,
|
|
|
|
size_t size, int direction)
|
|
|
|
{
|
|
|
|
dma_sync_single_for_device(hwdev == NULL ? NULL : &hwdev->dev, dma_handle, size, (enum dma_data_direction)direction);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
pci_dma_sync_sg_for_cpu(struct pci_dev *hwdev, struct scatterlist *sg,
|
|
|
|
int nelems, int direction)
|
|
|
|
{
|
|
|
|
dma_sync_sg_for_cpu(hwdev == NULL ? NULL : &hwdev->dev, sg, nelems, (enum dma_data_direction)direction);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
pci_dma_sync_sg_for_device(struct pci_dev *hwdev, struct scatterlist *sg,
|
|
|
|
int nelems, int direction)
|
|
|
|
{
|
|
|
|
dma_sync_sg_for_device(hwdev == NULL ? NULL : &hwdev->dev, sg, nelems, (enum dma_data_direction)direction);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int
|
2008-07-26 10:44:49 +08:00
|
|
|
pci_dma_mapping_error(struct pci_dev *pdev, dma_addr_t dma_addr)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2008-07-26 10:44:49 +08:00
|
|
|
return dma_mapping_error(&pdev->dev, dma_addr);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2010-03-11 07:23:41 +08:00
|
|
|
#ifdef CONFIG_PCI
|
|
|
|
static inline int pci_set_dma_mask(struct pci_dev *dev, u64 mask)
|
|
|
|
{
|
|
|
|
return dma_set_mask(&dev->dev, mask);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int pci_set_consistent_dma_mask(struct pci_dev *dev, u64 mask)
|
|
|
|
{
|
|
|
|
return dma_set_coherent_mask(&dev->dev, mask);
|
|
|
|
}
|
2016-03-08 01:39:16 +08:00
|
|
|
|
|
|
|
static inline int pci_set_dma_max_seg_size(struct pci_dev *dev,
|
|
|
|
unsigned int size)
|
|
|
|
{
|
|
|
|
return dma_set_max_seg_size(&dev->dev, size);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int pci_set_dma_seg_boundary(struct pci_dev *dev,
|
|
|
|
unsigned long mask)
|
|
|
|
{
|
|
|
|
return dma_set_seg_boundary(&dev->dev, mask);
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
static inline int pci_set_dma_mask(struct pci_dev *dev, u64 mask)
|
|
|
|
{ return -EIO; }
|
|
|
|
static inline int pci_set_consistent_dma_mask(struct pci_dev *dev, u64 mask)
|
|
|
|
{ return -EIO; }
|
|
|
|
static inline int pci_set_dma_max_seg_size(struct pci_dev *dev,
|
|
|
|
unsigned int size)
|
|
|
|
{ return -EIO; }
|
|
|
|
static inline int pci_set_dma_seg_boundary(struct pci_dev *dev,
|
|
|
|
unsigned long mask)
|
|
|
|
{ return -EIO; }
|
2010-03-11 07:23:41 +08:00
|
|
|
#endif
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
#endif
|