qtnfmac: switch from 'pci_' to 'dma_' API
The wrappers in include/linux/pci-dma-compat.h should go away. The patch has been generated with the coccinelle script below. It has been compile tested. @@ @@ - PCI_DMA_BIDIRECTIONAL + DMA_BIDIRECTIONAL @@ @@ - PCI_DMA_TODEVICE + DMA_TO_DEVICE @@ @@ - PCI_DMA_FROMDEVICE + DMA_FROM_DEVICE @@ @@ - PCI_DMA_NONE + DMA_NONE @@ expression e1, e2, e3; @@ - pci_alloc_consistent(e1, e2, e3) + dma_alloc_coherent(&e1->dev, e2, e3, GFP_) @@ expression e1, e2, e3; @@ - pci_zalloc_consistent(e1, e2, e3) + dma_alloc_coherent(&e1->dev, e2, e3, GFP_) @@ expression e1, e2, e3, e4; @@ - pci_free_consistent(e1, e2, e3, e4) + dma_free_coherent(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_map_single(e1, e2, e3, e4) + dma_map_single(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_unmap_single(e1, e2, e3, e4) + dma_unmap_single(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4, e5; @@ - pci_map_page(e1, e2, e3, e4, e5) + dma_map_page(&e1->dev, e2, e3, e4, e5) @@ expression e1, e2, e3, e4; @@ - pci_unmap_page(e1, e2, e3, e4) + dma_unmap_page(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_map_sg(e1, e2, e3, e4) + dma_map_sg(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_unmap_sg(e1, e2, e3, e4) + dma_unmap_sg(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_dma_sync_single_for_cpu(e1, e2, e3, e4) + dma_sync_single_for_cpu(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_dma_sync_single_for_device(e1, e2, e3, e4) + dma_sync_single_for_device(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_dma_sync_sg_for_cpu(e1, e2, e3, e4) + dma_sync_sg_for_cpu(&e1->dev, e2, e3, e4) @@ expression e1, e2, e3, e4; @@ - pci_dma_sync_sg_for_device(e1, e2, e3, e4) + dma_sync_sg_for_device(&e1->dev, e2, e3, e4) @@ expression e1, e2; @@ - pci_dma_mapping_error(e1, e2) + dma_mapping_error(&e1->dev, e2) @@ expression e1, e2; @@ - pci_set_dma_mask(e1, e2) + dma_set_mask(&e1->dev, e2) @@ expression e1, e2; @@ - pci_set_consistent_dma_mask(e1, e2) + dma_set_coherent_mask(&e1->dev, e2) Signed-off-by: Christophe JAILLET <christophe.jaillet@wanadoo.fr> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
e5c88bc91b
commit
06e1359cc8
|
@ -295,9 +295,9 @@ static int pearl_skb2rbd_attach(struct qtnf_pcie_pearl_state *ps, u16 index)
|
|||
priv->rx_skb[index] = skb;
|
||||
rxbd = &ps->rx_bd_vbase[index];
|
||||
|
||||
paddr = pci_map_single(priv->pdev, skb->data,
|
||||
SKB_BUF_SIZE, PCI_DMA_FROMDEVICE);
|
||||
if (pci_dma_mapping_error(priv->pdev, paddr)) {
|
||||
paddr = dma_map_single(&priv->pdev->dev, skb->data, SKB_BUF_SIZE,
|
||||
DMA_FROM_DEVICE);
|
||||
if (dma_mapping_error(&priv->pdev->dev, paddr)) {
|
||||
pr_err("skb DMA mapping error: %pad\n", &paddr);
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
@ -357,8 +357,8 @@ static void qtnf_pearl_free_xfer_buffers(struct qtnf_pcie_pearl_state *ps)
|
|||
skb = priv->rx_skb[i];
|
||||
paddr = QTN_HOST_ADDR(le32_to_cpu(rxbd->addr_h),
|
||||
le32_to_cpu(rxbd->addr));
|
||||
pci_unmap_single(priv->pdev, paddr, SKB_BUF_SIZE,
|
||||
PCI_DMA_FROMDEVICE);
|
||||
dma_unmap_single(&priv->pdev->dev, paddr,
|
||||
SKB_BUF_SIZE, DMA_FROM_DEVICE);
|
||||
dev_kfree_skb_any(skb);
|
||||
priv->rx_skb[i] = NULL;
|
||||
}
|
||||
|
@ -371,8 +371,8 @@ static void qtnf_pearl_free_xfer_buffers(struct qtnf_pcie_pearl_state *ps)
|
|||
skb = priv->tx_skb[i];
|
||||
paddr = QTN_HOST_ADDR(le32_to_cpu(txbd->addr_h),
|
||||
le32_to_cpu(txbd->addr));
|
||||
pci_unmap_single(priv->pdev, paddr, skb->len,
|
||||
PCI_DMA_TODEVICE);
|
||||
dma_unmap_single(&priv->pdev->dev, paddr, skb->len,
|
||||
DMA_TO_DEVICE);
|
||||
dev_kfree_skb_any(skb);
|
||||
priv->tx_skb[i] = NULL;
|
||||
}
|
||||
|
@ -485,8 +485,8 @@ static void qtnf_pearl_data_tx_reclaim(struct qtnf_pcie_pearl_state *ps)
|
|||
txbd = &ps->tx_bd_vbase[i];
|
||||
paddr = QTN_HOST_ADDR(le32_to_cpu(txbd->addr_h),
|
||||
le32_to_cpu(txbd->addr));
|
||||
pci_unmap_single(priv->pdev, paddr, skb->len,
|
||||
PCI_DMA_TODEVICE);
|
||||
dma_unmap_single(&priv->pdev->dev, paddr, skb->len,
|
||||
DMA_TO_DEVICE);
|
||||
|
||||
if (skb->dev) {
|
||||
dev_sw_netstats_tx_add(skb->dev, 1, skb->len);
|
||||
|
@ -559,9 +559,9 @@ static int qtnf_pcie_skb_send(struct qtnf_bus *bus, struct sk_buff *skb)
|
|||
priv->tx_skb[i] = skb;
|
||||
len = skb->len;
|
||||
|
||||
skb_paddr = pci_map_single(priv->pdev, skb->data,
|
||||
skb->len, PCI_DMA_TODEVICE);
|
||||
if (pci_dma_mapping_error(priv->pdev, skb_paddr)) {
|
||||
skb_paddr = dma_map_single(&priv->pdev->dev, skb->data, skb->len,
|
||||
DMA_TO_DEVICE);
|
||||
if (dma_mapping_error(&priv->pdev->dev, skb_paddr)) {
|
||||
pr_err("skb DMA mapping error: %pad\n", &skb_paddr);
|
||||
ret = -ENOMEM;
|
||||
goto tx_done;
|
||||
|
@ -748,8 +748,8 @@ static int qtnf_pcie_pearl_rx_poll(struct napi_struct *napi, int budget)
|
|||
if (skb) {
|
||||
skb_paddr = QTN_HOST_ADDR(le32_to_cpu(rxbd->addr_h),
|
||||
le32_to_cpu(rxbd->addr));
|
||||
pci_unmap_single(priv->pdev, skb_paddr, SKB_BUF_SIZE,
|
||||
PCI_DMA_FROMDEVICE);
|
||||
dma_unmap_single(&priv->pdev->dev, skb_paddr,
|
||||
SKB_BUF_SIZE, DMA_FROM_DEVICE);
|
||||
}
|
||||
|
||||
if (consume) {
|
||||
|
|
|
@ -255,9 +255,9 @@ topaz_skb2rbd_attach(struct qtnf_pcie_topaz_state *ts, u16 index, u32 wrap)
|
|||
|
||||
ts->base.rx_skb[index] = skb;
|
||||
|
||||
paddr = pci_map_single(ts->base.pdev, skb->data,
|
||||
SKB_BUF_SIZE, PCI_DMA_FROMDEVICE);
|
||||
if (pci_dma_mapping_error(ts->base.pdev, paddr)) {
|
||||
paddr = dma_map_single(&ts->base.pdev->dev, skb->data, SKB_BUF_SIZE,
|
||||
DMA_FROM_DEVICE);
|
||||
if (dma_mapping_error(&ts->base.pdev->dev, paddr)) {
|
||||
pr_err("skb mapping error: %pad\n", &paddr);
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
@ -306,8 +306,8 @@ static void qtnf_topaz_free_xfer_buffers(struct qtnf_pcie_topaz_state *ts)
|
|||
rxbd = &ts->rx_bd_vbase[i];
|
||||
skb = priv->rx_skb[i];
|
||||
paddr = QTN_HOST_ADDR(0x0, le32_to_cpu(rxbd->addr));
|
||||
pci_unmap_single(priv->pdev, paddr, SKB_BUF_SIZE,
|
||||
PCI_DMA_FROMDEVICE);
|
||||
dma_unmap_single(&priv->pdev->dev, paddr,
|
||||
SKB_BUF_SIZE, DMA_FROM_DEVICE);
|
||||
dev_kfree_skb_any(skb);
|
||||
priv->rx_skb[i] = NULL;
|
||||
rxbd->addr = 0;
|
||||
|
@ -321,8 +321,8 @@ static void qtnf_topaz_free_xfer_buffers(struct qtnf_pcie_topaz_state *ts)
|
|||
txbd = &ts->tx_bd_vbase[i];
|
||||
skb = priv->tx_skb[i];
|
||||
paddr = QTN_HOST_ADDR(0x0, le32_to_cpu(txbd->addr));
|
||||
pci_unmap_single(priv->pdev, paddr, SKB_BUF_SIZE,
|
||||
PCI_DMA_TODEVICE);
|
||||
dma_unmap_single(&priv->pdev->dev, paddr,
|
||||
SKB_BUF_SIZE, DMA_TO_DEVICE);
|
||||
dev_kfree_skb_any(skb);
|
||||
priv->tx_skb[i] = NULL;
|
||||
txbd->addr = 0;
|
||||
|
@ -414,8 +414,8 @@ static void qtnf_topaz_data_tx_reclaim(struct qtnf_pcie_topaz_state *ts)
|
|||
if (likely(skb)) {
|
||||
txbd = &ts->tx_bd_vbase[i];
|
||||
paddr = QTN_HOST_ADDR(0x0, le32_to_cpu(txbd->addr));
|
||||
pci_unmap_single(priv->pdev, paddr, skb->len,
|
||||
PCI_DMA_TODEVICE);
|
||||
dma_unmap_single(&priv->pdev->dev, paddr, skb->len,
|
||||
DMA_TO_DEVICE);
|
||||
|
||||
if (skb->dev) {
|
||||
dev_sw_netstats_tx_add(skb->dev, 1, skb->len);
|
||||
|
@ -522,9 +522,9 @@ static int qtnf_pcie_data_tx(struct qtnf_bus *bus, struct sk_buff *skb,
|
|||
priv->tx_skb[i] = skb;
|
||||
len = skb->len;
|
||||
|
||||
skb_paddr = pci_map_single(priv->pdev, skb->data,
|
||||
skb->len, PCI_DMA_TODEVICE);
|
||||
if (pci_dma_mapping_error(priv->pdev, skb_paddr)) {
|
||||
skb_paddr = dma_map_single(&priv->pdev->dev, skb->data, skb->len,
|
||||
DMA_TO_DEVICE);
|
||||
if (dma_mapping_error(&priv->pdev->dev, skb_paddr)) {
|
||||
ret = -ENOMEM;
|
||||
goto tx_done;
|
||||
}
|
||||
|
@ -653,8 +653,8 @@ static int qtnf_topaz_rx_poll(struct napi_struct *napi, int budget)
|
|||
|
||||
if (skb) {
|
||||
skb_paddr = QTN_HOST_ADDR(0x0, le32_to_cpu(rxbd->addr));
|
||||
pci_unmap_single(priv->pdev, skb_paddr, SKB_BUF_SIZE,
|
||||
PCI_DMA_FROMDEVICE);
|
||||
dma_unmap_single(&priv->pdev->dev, skb_paddr,
|
||||
SKB_BUF_SIZE, DMA_FROM_DEVICE);
|
||||
}
|
||||
|
||||
if (consume) {
|
||||
|
|
Loading…
Reference in New Issue