stmmac: add length field to dma data
Currently, the code pulls out the length field when unmapping a buffer directly from the descriptor. This will result in an uncached read to a dma_alloc_coherent() region. There is no need to do this, so this patch simply puts the value directly into a data structure which will hit the cache. Signed-off-by: Fabrice Gasnier <fabrice.gasnier@st.com> Signed-off-by: Giuseppe Cavallaro <peppe.cavallaro@st.com> Signed-off-by: Alexandre TORGUE <alexandre.torgue@st.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
e3ad57c967
commit
553e2ab313
|
@ -49,6 +49,7 @@ static int stmmac_jumbo_frm(void *p, struct sk_buff *skb, int csum)
|
||||||
if (dma_mapping_error(priv->device, desc->des2))
|
if (dma_mapping_error(priv->device, desc->des2))
|
||||||
return -1;
|
return -1;
|
||||||
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
||||||
|
priv->tx_skbuff_dma[entry].len = bmax;
|
||||||
priv->hw->desc->prepare_tx_desc(desc, 1, bmax, csum, STMMAC_CHAIN_MODE);
|
priv->hw->desc->prepare_tx_desc(desc, 1, bmax, csum, STMMAC_CHAIN_MODE);
|
||||||
|
|
||||||
while (len != 0) {
|
while (len != 0) {
|
||||||
|
@ -63,6 +64,7 @@ static int stmmac_jumbo_frm(void *p, struct sk_buff *skb, int csum)
|
||||||
if (dma_mapping_error(priv->device, desc->des2))
|
if (dma_mapping_error(priv->device, desc->des2))
|
||||||
return -1;
|
return -1;
|
||||||
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
||||||
|
priv->tx_skbuff_dma[entry].len = bmax;
|
||||||
priv->hw->desc->prepare_tx_desc(desc, 0, bmax, csum,
|
priv->hw->desc->prepare_tx_desc(desc, 0, bmax, csum,
|
||||||
STMMAC_CHAIN_MODE);
|
STMMAC_CHAIN_MODE);
|
||||||
priv->hw->desc->set_tx_owner(desc);
|
priv->hw->desc->set_tx_owner(desc);
|
||||||
|
@ -75,6 +77,7 @@ static int stmmac_jumbo_frm(void *p, struct sk_buff *skb, int csum)
|
||||||
if (dma_mapping_error(priv->device, desc->des2))
|
if (dma_mapping_error(priv->device, desc->des2))
|
||||||
return -1;
|
return -1;
|
||||||
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
||||||
|
priv->tx_skbuff_dma[entry].len = len;
|
||||||
priv->hw->desc->prepare_tx_desc(desc, 0, len, csum,
|
priv->hw->desc->prepare_tx_desc(desc, 0, len, csum,
|
||||||
STMMAC_CHAIN_MODE);
|
STMMAC_CHAIN_MODE);
|
||||||
priv->hw->desc->set_tx_owner(desc);
|
priv->hw->desc->set_tx_owner(desc);
|
||||||
|
|
|
@ -56,6 +56,8 @@ static int stmmac_jumbo_frm(void *p, struct sk_buff *skb, int csum)
|
||||||
return -1;
|
return -1;
|
||||||
|
|
||||||
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
||||||
|
priv->tx_skbuff_dma[entry].len = bmax;
|
||||||
|
|
||||||
desc->des3 = desc->des2 + BUF_SIZE_4KiB;
|
desc->des3 = desc->des2 + BUF_SIZE_4KiB;
|
||||||
priv->hw->desc->prepare_tx_desc(desc, 1, bmax, csum,
|
priv->hw->desc->prepare_tx_desc(desc, 1, bmax, csum,
|
||||||
STMMAC_RING_MODE);
|
STMMAC_RING_MODE);
|
||||||
|
@ -73,6 +75,8 @@ static int stmmac_jumbo_frm(void *p, struct sk_buff *skb, int csum)
|
||||||
if (dma_mapping_error(priv->device, desc->des2))
|
if (dma_mapping_error(priv->device, desc->des2))
|
||||||
return -1;
|
return -1;
|
||||||
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
||||||
|
priv->tx_skbuff_dma[entry].len = len;
|
||||||
|
|
||||||
desc->des3 = desc->des2 + BUF_SIZE_4KiB;
|
desc->des3 = desc->des2 + BUF_SIZE_4KiB;
|
||||||
priv->hw->desc->prepare_tx_desc(desc, 0, len, csum,
|
priv->hw->desc->prepare_tx_desc(desc, 0, len, csum,
|
||||||
STMMAC_RING_MODE);
|
STMMAC_RING_MODE);
|
||||||
|
@ -84,6 +88,7 @@ static int stmmac_jumbo_frm(void *p, struct sk_buff *skb, int csum)
|
||||||
if (dma_mapping_error(priv->device, desc->des2))
|
if (dma_mapping_error(priv->device, desc->des2))
|
||||||
return -1;
|
return -1;
|
||||||
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
||||||
|
priv->tx_skbuff_dma[entry].len = nopaged_len;
|
||||||
desc->des3 = desc->des2 + BUF_SIZE_4KiB;
|
desc->des3 = desc->des2 + BUF_SIZE_4KiB;
|
||||||
priv->hw->desc->prepare_tx_desc(desc, 1, nopaged_len, csum,
|
priv->hw->desc->prepare_tx_desc(desc, 1, nopaged_len, csum,
|
||||||
STMMAC_RING_MODE);
|
STMMAC_RING_MODE);
|
||||||
|
|
|
@ -45,6 +45,7 @@ struct stmmac_resources {
|
||||||
struct stmmac_tx_info {
|
struct stmmac_tx_info {
|
||||||
dma_addr_t buf;
|
dma_addr_t buf;
|
||||||
bool map_as_page;
|
bool map_as_page;
|
||||||
|
unsigned len;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct stmmac_priv {
|
struct stmmac_priv {
|
||||||
|
|
|
@ -1093,6 +1093,7 @@ static int init_dma_desc_rings(struct net_device *dev, gfp_t flags)
|
||||||
p->des2 = 0;
|
p->des2 = 0;
|
||||||
priv->tx_skbuff_dma[i].buf = 0;
|
priv->tx_skbuff_dma[i].buf = 0;
|
||||||
priv->tx_skbuff_dma[i].map_as_page = false;
|
priv->tx_skbuff_dma[i].map_as_page = false;
|
||||||
|
priv->tx_skbuff_dma[i].len = 0;
|
||||||
priv->tx_skbuff[i] = NULL;
|
priv->tx_skbuff[i] = NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1136,12 +1137,12 @@ static void dma_free_tx_skbufs(struct stmmac_priv *priv)
|
||||||
if (priv->tx_skbuff_dma[i].map_as_page)
|
if (priv->tx_skbuff_dma[i].map_as_page)
|
||||||
dma_unmap_page(priv->device,
|
dma_unmap_page(priv->device,
|
||||||
priv->tx_skbuff_dma[i].buf,
|
priv->tx_skbuff_dma[i].buf,
|
||||||
priv->hw->desc->get_tx_len(p),
|
priv->tx_skbuff_dma[i].len,
|
||||||
DMA_TO_DEVICE);
|
DMA_TO_DEVICE);
|
||||||
else
|
else
|
||||||
dma_unmap_single(priv->device,
|
dma_unmap_single(priv->device,
|
||||||
priv->tx_skbuff_dma[i].buf,
|
priv->tx_skbuff_dma[i].buf,
|
||||||
priv->hw->desc->get_tx_len(p),
|
priv->tx_skbuff_dma[i].len,
|
||||||
DMA_TO_DEVICE);
|
DMA_TO_DEVICE);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1347,12 +1348,12 @@ static void stmmac_tx_clean(struct stmmac_priv *priv)
|
||||||
if (priv->tx_skbuff_dma[entry].map_as_page)
|
if (priv->tx_skbuff_dma[entry].map_as_page)
|
||||||
dma_unmap_page(priv->device,
|
dma_unmap_page(priv->device,
|
||||||
priv->tx_skbuff_dma[entry].buf,
|
priv->tx_skbuff_dma[entry].buf,
|
||||||
priv->hw->desc->get_tx_len(p),
|
priv->tx_skbuff_dma[entry].len,
|
||||||
DMA_TO_DEVICE);
|
DMA_TO_DEVICE);
|
||||||
else
|
else
|
||||||
dma_unmap_single(priv->device,
|
dma_unmap_single(priv->device,
|
||||||
priv->tx_skbuff_dma[entry].buf,
|
priv->tx_skbuff_dma[entry].buf,
|
||||||
priv->hw->desc->get_tx_len(p),
|
priv->tx_skbuff_dma[entry].len,
|
||||||
DMA_TO_DEVICE);
|
DMA_TO_DEVICE);
|
||||||
priv->tx_skbuff_dma[entry].buf = 0;
|
priv->tx_skbuff_dma[entry].buf = 0;
|
||||||
priv->tx_skbuff_dma[entry].map_as_page = false;
|
priv->tx_skbuff_dma[entry].map_as_page = false;
|
||||||
|
@ -1986,6 +1987,7 @@ static netdev_tx_t stmmac_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
if (dma_mapping_error(priv->device, desc->des2))
|
if (dma_mapping_error(priv->device, desc->des2))
|
||||||
goto dma_map_err;
|
goto dma_map_err;
|
||||||
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
||||||
|
priv->tx_skbuff_dma[entry].len = nopaged_len;
|
||||||
priv->hw->desc->prepare_tx_desc(desc, 1, nopaged_len,
|
priv->hw->desc->prepare_tx_desc(desc, 1, nopaged_len,
|
||||||
csum_insertion, priv->mode);
|
csum_insertion, priv->mode);
|
||||||
} else {
|
} else {
|
||||||
|
@ -2014,6 +2016,7 @@ static netdev_tx_t stmmac_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||||
|
|
||||||
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
priv->tx_skbuff_dma[entry].buf = desc->des2;
|
||||||
priv->tx_skbuff_dma[entry].map_as_page = true;
|
priv->tx_skbuff_dma[entry].map_as_page = true;
|
||||||
|
priv->tx_skbuff_dma[entry].len = len;
|
||||||
priv->hw->desc->prepare_tx_desc(desc, 0, len, csum_insertion,
|
priv->hw->desc->prepare_tx_desc(desc, 0, len, csum_insertion,
|
||||||
priv->mode);
|
priv->mode);
|
||||||
wmb();
|
wmb();
|
||||||
|
|
Loading…
Reference in New Issue