tsnep: Prepare RX buffer for XDP support
Always reserve XDP_PACKET_HEADROOM in front of RX buffer. Similar DMA direction is always set to DMA_BIDIRECTIONAL. This eliminates the need for RX queue reconfiguration during BPF program setup. The RX queue is always prepared for XDP. No negative impact of DMA_BIDIRECTIONAL was measured. Signed-off-by: Gerhard Engleder <gerhard@engleder-embedded.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
59d562aa19
commit
cc3e254f94
|
@ -26,9 +26,10 @@
|
|||
#include <linux/etherdevice.h>
|
||||
#include <linux/phy.h>
|
||||
#include <linux/iopoll.h>
|
||||
#include <linux/bpf.h>
|
||||
|
||||
#define TSNEP_SKB_PAD (NET_SKB_PAD + NET_IP_ALIGN)
|
||||
#define TSNEP_HEADROOM ALIGN(TSNEP_SKB_PAD, 4)
|
||||
#define TSNEP_RX_OFFSET (max(NET_SKB_PAD, XDP_PACKET_HEADROOM) + NET_IP_ALIGN)
|
||||
#define TSNEP_HEADROOM ALIGN(TSNEP_RX_OFFSET, 4)
|
||||
#define TSNEP_MAX_RX_BUF_SIZE (PAGE_SIZE - TSNEP_HEADROOM - \
|
||||
SKB_DATA_ALIGN(sizeof(struct skb_shared_info)))
|
||||
|
||||
|
@ -806,9 +807,9 @@ static int tsnep_rx_ring_init(struct tsnep_rx *rx)
|
|||
pp_params.pool_size = TSNEP_RING_SIZE;
|
||||
pp_params.nid = dev_to_node(dmadev);
|
||||
pp_params.dev = dmadev;
|
||||
pp_params.dma_dir = DMA_FROM_DEVICE;
|
||||
pp_params.dma_dir = DMA_BIDIRECTIONAL;
|
||||
pp_params.max_len = TSNEP_MAX_RX_BUF_SIZE;
|
||||
pp_params.offset = TSNEP_SKB_PAD;
|
||||
pp_params.offset = TSNEP_RX_OFFSET;
|
||||
rx->page_pool = page_pool_create(&pp_params);
|
||||
if (IS_ERR(rx->page_pool)) {
|
||||
retval = PTR_ERR(rx->page_pool);
|
||||
|
@ -843,7 +844,7 @@ static void tsnep_rx_set_page(struct tsnep_rx *rx, struct tsnep_rx_entry *entry,
|
|||
entry->page = page;
|
||||
entry->len = TSNEP_MAX_RX_BUF_SIZE;
|
||||
entry->dma = page_pool_get_dma_addr(entry->page);
|
||||
entry->desc->rx = __cpu_to_le64(entry->dma + TSNEP_SKB_PAD);
|
||||
entry->desc->rx = __cpu_to_le64(entry->dma + TSNEP_RX_OFFSET);
|
||||
}
|
||||
|
||||
static int tsnep_rx_alloc_buffer(struct tsnep_rx *rx, int index)
|
||||
|
@ -947,14 +948,14 @@ static struct sk_buff *tsnep_build_skb(struct tsnep_rx *rx, struct page *page,
|
|||
return NULL;
|
||||
|
||||
/* update pointers within the skb to store the data */
|
||||
skb_reserve(skb, TSNEP_SKB_PAD + TSNEP_RX_INLINE_METADATA_SIZE);
|
||||
skb_reserve(skb, TSNEP_RX_OFFSET + TSNEP_RX_INLINE_METADATA_SIZE);
|
||||
__skb_put(skb, length - ETH_FCS_LEN);
|
||||
|
||||
if (rx->adapter->hwtstamp_config.rx_filter == HWTSTAMP_FILTER_ALL) {
|
||||
struct skb_shared_hwtstamps *hwtstamps = skb_hwtstamps(skb);
|
||||
struct tsnep_rx_inline *rx_inline =
|
||||
(struct tsnep_rx_inline *)(page_address(page) +
|
||||
TSNEP_SKB_PAD);
|
||||
TSNEP_RX_OFFSET);
|
||||
|
||||
skb_shinfo(skb)->tx_flags |=
|
||||
SKBTX_HW_TSTAMP_NETDEV;
|
||||
|
@ -1014,11 +1015,11 @@ static int tsnep_rx_poll(struct tsnep_rx *rx, struct napi_struct *napi,
|
|||
*/
|
||||
dma_rmb();
|
||||
|
||||
prefetch(page_address(entry->page) + TSNEP_SKB_PAD);
|
||||
prefetch(page_address(entry->page) + TSNEP_RX_OFFSET);
|
||||
length = __le32_to_cpu(entry->desc_wb->properties) &
|
||||
TSNEP_DESC_LENGTH_MASK;
|
||||
dma_sync_single_range_for_cpu(dmadev, entry->dma, TSNEP_SKB_PAD,
|
||||
length, dma_dir);
|
||||
dma_sync_single_range_for_cpu(dmadev, entry->dma,
|
||||
TSNEP_RX_OFFSET, length, dma_dir);
|
||||
|
||||
/* RX metadata with timestamps is in front of actual data,
|
||||
* subtract metadata size to get length of actual data and
|
||||
|
|
Loading…
Reference in New Issue