hv_netvsc: Add support for XDP_REDIRECT

Handle XDP_REDIRECT action in netvsc driver.
Also, transparently pass ndo_xdp_xmit to VF when available.

Signed-off-by: Haiyang Zhang <haiyangz@microsoft.com>
Link: https://lore.kernel.org/r/1649362894-20077-1-git-send-email-haiyangz@microsoft.com
Signed-off-by: Jakub Kicinski <kuba@kernel.org>
This commit is contained in:
Haiyang Zhang 2022-04-07 13:21:34 -07:00 committed by Jakub Kicinski
parent 2e36437f44
commit 1cb9d3b618
4 changed files with 228 additions and 94 deletions

View File

@ -15,6 +15,7 @@
#include <linux/list.h> #include <linux/list.h>
#include <linux/hyperv.h> #include <linux/hyperv.h>
#include <linux/rndis.h> #include <linux/rndis.h>
#include <linux/jhash.h>
/* RSS related */ /* RSS related */
#define OID_GEN_RECEIVE_SCALE_CAPABILITIES 0x00010203 /* query only */ #define OID_GEN_RECEIVE_SCALE_CAPABILITIES 0x00010203 /* query only */
@ -237,6 +238,7 @@ int netvsc_recv_callback(struct net_device *net,
void netvsc_channel_cb(void *context); void netvsc_channel_cb(void *context);
int netvsc_poll(struct napi_struct *napi, int budget); int netvsc_poll(struct napi_struct *napi, int budget);
void netvsc_xdp_xmit(struct sk_buff *skb, struct net_device *ndev);
u32 netvsc_run_xdp(struct net_device *ndev, struct netvsc_channel *nvchan, u32 netvsc_run_xdp(struct net_device *ndev, struct netvsc_channel *nvchan,
struct xdp_buff *xdp); struct xdp_buff *xdp);
unsigned int netvsc_xdp_fraglen(unsigned int len); unsigned int netvsc_xdp_fraglen(unsigned int len);
@ -246,6 +248,8 @@ int netvsc_xdp_set(struct net_device *dev, struct bpf_prog *prog,
struct netvsc_device *nvdev); struct netvsc_device *nvdev);
int netvsc_vf_setxdp(struct net_device *vf_netdev, struct bpf_prog *prog); int netvsc_vf_setxdp(struct net_device *vf_netdev, struct bpf_prog *prog);
int netvsc_bpf(struct net_device *dev, struct netdev_bpf *bpf); int netvsc_bpf(struct net_device *dev, struct netdev_bpf *bpf);
int netvsc_ndoxdp_xmit(struct net_device *ndev, int n,
struct xdp_frame **frames, u32 flags);
int rndis_set_subchannel(struct net_device *ndev, int rndis_set_subchannel(struct net_device *ndev,
struct netvsc_device *nvdev, struct netvsc_device *nvdev,
@ -942,12 +946,21 @@ struct nvsc_rsc {
#define NVSC_RSC_CSUM_INFO BIT(1) /* valid/present bit for 'csum_info' */ #define NVSC_RSC_CSUM_INFO BIT(1) /* valid/present bit for 'csum_info' */
#define NVSC_RSC_HASH_INFO BIT(2) /* valid/present bit for 'hash_info' */ #define NVSC_RSC_HASH_INFO BIT(2) /* valid/present bit for 'hash_info' */
struct netvsc_stats { struct netvsc_stats_tx {
u64 packets;
u64 bytes;
u64 xdp_xmit;
struct u64_stats_sync syncp;
};
struct netvsc_stats_rx {
u64 packets; u64 packets;
u64 bytes; u64 bytes;
u64 broadcast; u64 broadcast;
u64 multicast; u64 multicast;
u64 xdp_drop; u64 xdp_drop;
u64 xdp_redirect;
u64 xdp_tx;
struct u64_stats_sync syncp; struct u64_stats_sync syncp;
}; };
@ -1046,6 +1059,55 @@ struct net_device_context {
struct netvsc_device_info *saved_netvsc_dev_info; struct netvsc_device_info *saved_netvsc_dev_info;
}; };
/* Azure hosts don't support non-TCP port numbers in hashing for fragmented
* packets. We can use ethtool to change UDP hash level when necessary.
*/
static inline u32 netvsc_get_hash(struct sk_buff *skb,
const struct net_device_context *ndc)
{
struct flow_keys flow;
u32 hash, pkt_proto = 0;
static u32 hashrnd __read_mostly;
net_get_random_once(&hashrnd, sizeof(hashrnd));
if (!skb_flow_dissect_flow_keys(skb, &flow, 0))
return 0;
switch (flow.basic.ip_proto) {
case IPPROTO_TCP:
if (flow.basic.n_proto == htons(ETH_P_IP))
pkt_proto = HV_TCP4_L4HASH;
else if (flow.basic.n_proto == htons(ETH_P_IPV6))
pkt_proto = HV_TCP6_L4HASH;
break;
case IPPROTO_UDP:
if (flow.basic.n_proto == htons(ETH_P_IP))
pkt_proto = HV_UDP4_L4HASH;
else if (flow.basic.n_proto == htons(ETH_P_IPV6))
pkt_proto = HV_UDP6_L4HASH;
break;
}
if (pkt_proto & ndc->l4_hash) {
return skb_get_hash(skb);
} else {
if (flow.basic.n_proto == htons(ETH_P_IP))
hash = jhash2((u32 *)&flow.addrs.v4addrs, 2, hashrnd);
else if (flow.basic.n_proto == htons(ETH_P_IPV6))
hash = jhash2((u32 *)&flow.addrs.v6addrs, 8, hashrnd);
else
return 0;
__skb_set_sw_hash(skb, hash, false);
}
return hash;
}
/* Per channel data */ /* Per channel data */
struct netvsc_channel { struct netvsc_channel {
struct vmbus_channel *channel; struct vmbus_channel *channel;
@ -1060,9 +1122,10 @@ struct netvsc_channel {
struct bpf_prog __rcu *bpf_prog; struct bpf_prog __rcu *bpf_prog;
struct xdp_rxq_info xdp_rxq; struct xdp_rxq_info xdp_rxq;
bool xdp_flush;
struct netvsc_stats tx_stats; struct netvsc_stats_tx tx_stats;
struct netvsc_stats rx_stats; struct netvsc_stats_rx rx_stats;
}; };
/* Per netvsc device */ /* Per netvsc device */

View File

@ -20,6 +20,7 @@
#include <linux/vmalloc.h> #include <linux/vmalloc.h>
#include <linux/rtnetlink.h> #include <linux/rtnetlink.h>
#include <linux/prefetch.h> #include <linux/prefetch.h>
#include <linux/filter.h>
#include <asm/sync_bitops.h> #include <asm/sync_bitops.h>
#include <asm/mshyperv.h> #include <asm/mshyperv.h>
@ -805,7 +806,7 @@ static void netvsc_send_tx_complete(struct net_device *ndev,
struct hv_netvsc_packet *packet struct hv_netvsc_packet *packet
= (struct hv_netvsc_packet *)skb->cb; = (struct hv_netvsc_packet *)skb->cb;
u32 send_index = packet->send_buf_index; u32 send_index = packet->send_buf_index;
struct netvsc_stats *tx_stats; struct netvsc_stats_tx *tx_stats;
if (send_index != NETVSC_INVALID_INDEX) if (send_index != NETVSC_INVALID_INDEX)
netvsc_free_send_slot(net_device, send_index); netvsc_free_send_slot(net_device, send_index);
@ -1670,12 +1671,17 @@ int netvsc_poll(struct napi_struct *napi, int budget)
if (!nvchan->desc) if (!nvchan->desc)
nvchan->desc = hv_pkt_iter_first(channel); nvchan->desc = hv_pkt_iter_first(channel);
nvchan->xdp_flush = false;
while (nvchan->desc && work_done < budget) { while (nvchan->desc && work_done < budget) {
work_done += netvsc_process_raw_pkt(device, nvchan, net_device, work_done += netvsc_process_raw_pkt(device, nvchan, net_device,
ndev, nvchan->desc, budget); ndev, nvchan->desc, budget);
nvchan->desc = hv_pkt_iter_next(channel, nvchan->desc); nvchan->desc = hv_pkt_iter_next(channel, nvchan->desc);
} }
if (nvchan->xdp_flush)
xdp_do_flush();
/* Send any pending receive completions */ /* Send any pending receive completions */
ret = send_recv_completions(ndev, net_device, nvchan); ret = send_recv_completions(ndev, net_device, nvchan);

View File

@ -10,6 +10,7 @@
#include <linux/netdevice.h> #include <linux/netdevice.h>
#include <linux/etherdevice.h> #include <linux/etherdevice.h>
#include <linux/ethtool.h> #include <linux/ethtool.h>
#include <linux/netpoll.h>
#include <linux/bpf.h> #include <linux/bpf.h>
#include <linux/bpf_trace.h> #include <linux/bpf_trace.h>
#include <linux/kernel.h> #include <linux/kernel.h>
@ -23,11 +24,13 @@
u32 netvsc_run_xdp(struct net_device *ndev, struct netvsc_channel *nvchan, u32 netvsc_run_xdp(struct net_device *ndev, struct netvsc_channel *nvchan,
struct xdp_buff *xdp) struct xdp_buff *xdp)
{ {
struct netvsc_stats_rx *rx_stats = &nvchan->rx_stats;
void *data = nvchan->rsc.data[0]; void *data = nvchan->rsc.data[0];
u32 len = nvchan->rsc.len[0]; u32 len = nvchan->rsc.len[0];
struct page *page = NULL; struct page *page = NULL;
struct bpf_prog *prog; struct bpf_prog *prog;
u32 act = XDP_PASS; u32 act = XDP_PASS;
bool drop = true;
xdp->data_hard_start = NULL; xdp->data_hard_start = NULL;
@ -60,9 +63,34 @@ u32 netvsc_run_xdp(struct net_device *ndev, struct netvsc_channel *nvchan,
switch (act) { switch (act) {
case XDP_PASS: case XDP_PASS:
case XDP_TX: case XDP_TX:
drop = false;
break;
case XDP_DROP: case XDP_DROP:
break; break;
case XDP_REDIRECT:
if (!xdp_do_redirect(ndev, xdp, prog)) {
nvchan->xdp_flush = true;
drop = false;
u64_stats_update_begin(&rx_stats->syncp);
rx_stats->xdp_redirect++;
rx_stats->packets++;
rx_stats->bytes += nvchan->rsc.pktlen;
u64_stats_update_end(&rx_stats->syncp);
break;
} else {
u64_stats_update_begin(&rx_stats->syncp);
rx_stats->xdp_drop++;
u64_stats_update_end(&rx_stats->syncp);
}
fallthrough;
case XDP_ABORTED: case XDP_ABORTED:
trace_xdp_exception(ndev, prog, act); trace_xdp_exception(ndev, prog, act);
break; break;
@ -74,7 +102,7 @@ u32 netvsc_run_xdp(struct net_device *ndev, struct netvsc_channel *nvchan,
out: out:
rcu_read_unlock(); rcu_read_unlock();
if (page && act != XDP_PASS && act != XDP_TX) { if (page && drop) {
__free_page(page); __free_page(page);
xdp->data_hard_start = NULL; xdp->data_hard_start = NULL;
} }
@ -197,3 +225,68 @@ int netvsc_bpf(struct net_device *dev, struct netdev_bpf *bpf)
return -EINVAL; return -EINVAL;
} }
} }
static int netvsc_ndoxdp_xmit_fm(struct net_device *ndev,
struct xdp_frame *frame, u16 q_idx)
{
struct sk_buff *skb;
skb = xdp_build_skb_from_frame(frame, ndev);
if (unlikely(!skb))
return -ENOMEM;
netvsc_get_hash(skb, netdev_priv(ndev));
skb_record_rx_queue(skb, q_idx);
netvsc_xdp_xmit(skb, ndev);
return 0;
}
int netvsc_ndoxdp_xmit(struct net_device *ndev, int n,
struct xdp_frame **frames, u32 flags)
{
struct net_device_context *ndev_ctx = netdev_priv(ndev);
const struct net_device_ops *vf_ops;
struct netvsc_stats_tx *tx_stats;
struct netvsc_device *nvsc_dev;
struct net_device *vf_netdev;
int i, count = 0;
u16 q_idx;
/* Don't transmit if netvsc_device is gone */
nvsc_dev = rcu_dereference_bh(ndev_ctx->nvdev);
if (unlikely(!nvsc_dev || nvsc_dev->destroy))
return 0;
/* If VF is present and up then redirect packets to it.
* Skip the VF if it is marked down or has no carrier.
* If netpoll is in uses, then VF can not be used either.
*/
vf_netdev = rcu_dereference_bh(ndev_ctx->vf_netdev);
if (vf_netdev && netif_running(vf_netdev) &&
netif_carrier_ok(vf_netdev) && !netpoll_tx_running(ndev) &&
vf_netdev->netdev_ops->ndo_xdp_xmit &&
ndev_ctx->data_path_is_vf) {
vf_ops = vf_netdev->netdev_ops;
return vf_ops->ndo_xdp_xmit(vf_netdev, n, frames, flags);
}
q_idx = smp_processor_id() % ndev->real_num_tx_queues;
for (i = 0; i < n; i++) {
if (netvsc_ndoxdp_xmit_fm(ndev, frames[i], q_idx))
break;
count++;
}
tx_stats = &nvsc_dev->chan_table[q_idx].tx_stats;
u64_stats_update_begin(&tx_stats->syncp);
tx_stats->xdp_xmit += count;
u64_stats_update_end(&tx_stats->syncp);
return count;
}

View File

@ -242,56 +242,6 @@ static inline void *init_ppi_data(struct rndis_message *msg,
return ppi + 1; return ppi + 1;
} }
/* Azure hosts don't support non-TCP port numbers in hashing for fragmented
* packets. We can use ethtool to change UDP hash level when necessary.
*/
static inline u32 netvsc_get_hash(
struct sk_buff *skb,
const struct net_device_context *ndc)
{
struct flow_keys flow;
u32 hash, pkt_proto = 0;
static u32 hashrnd __read_mostly;
net_get_random_once(&hashrnd, sizeof(hashrnd));
if (!skb_flow_dissect_flow_keys(skb, &flow, 0))
return 0;
switch (flow.basic.ip_proto) {
case IPPROTO_TCP:
if (flow.basic.n_proto == htons(ETH_P_IP))
pkt_proto = HV_TCP4_L4HASH;
else if (flow.basic.n_proto == htons(ETH_P_IPV6))
pkt_proto = HV_TCP6_L4HASH;
break;
case IPPROTO_UDP:
if (flow.basic.n_proto == htons(ETH_P_IP))
pkt_proto = HV_UDP4_L4HASH;
else if (flow.basic.n_proto == htons(ETH_P_IPV6))
pkt_proto = HV_UDP6_L4HASH;
break;
}
if (pkt_proto & ndc->l4_hash) {
return skb_get_hash(skb);
} else {
if (flow.basic.n_proto == htons(ETH_P_IP))
hash = jhash2((u32 *)&flow.addrs.v4addrs, 2, hashrnd);
else if (flow.basic.n_proto == htons(ETH_P_IPV6))
hash = jhash2((u32 *)&flow.addrs.v6addrs, 8, hashrnd);
else
return 0;
__skb_set_sw_hash(skb, hash, false);
}
return hash;
}
static inline int netvsc_get_tx_queue(struct net_device *ndev, static inline int netvsc_get_tx_queue(struct net_device *ndev,
struct sk_buff *skb, int old_idx) struct sk_buff *skb, int old_idx)
{ {
@ -804,7 +754,7 @@ void netvsc_linkstatus_callback(struct net_device *net,
} }
/* This function should only be called after skb_record_rx_queue() */ /* This function should only be called after skb_record_rx_queue() */
static void netvsc_xdp_xmit(struct sk_buff *skb, struct net_device *ndev) void netvsc_xdp_xmit(struct sk_buff *skb, struct net_device *ndev)
{ {
int rc; int rc;
@ -925,7 +875,7 @@ int netvsc_recv_callback(struct net_device *net,
struct vmbus_channel *channel = nvchan->channel; struct vmbus_channel *channel = nvchan->channel;
u16 q_idx = channel->offermsg.offer.sub_channel_index; u16 q_idx = channel->offermsg.offer.sub_channel_index;
struct sk_buff *skb; struct sk_buff *skb;
struct netvsc_stats *rx_stats = &nvchan->rx_stats; struct netvsc_stats_rx *rx_stats = &nvchan->rx_stats;
struct xdp_buff xdp; struct xdp_buff xdp;
u32 act; u32 act;
@ -934,6 +884,9 @@ int netvsc_recv_callback(struct net_device *net,
act = netvsc_run_xdp(net, nvchan, &xdp); act = netvsc_run_xdp(net, nvchan, &xdp);
if (act == XDP_REDIRECT)
return NVSP_STAT_SUCCESS;
if (act != XDP_PASS && act != XDP_TX) { if (act != XDP_PASS && act != XDP_TX) {
u64_stats_update_begin(&rx_stats->syncp); u64_stats_update_begin(&rx_stats->syncp);
rx_stats->xdp_drop++; rx_stats->xdp_drop++;
@ -958,6 +911,9 @@ int netvsc_recv_callback(struct net_device *net,
* statistics will not work correctly. * statistics will not work correctly.
*/ */
u64_stats_update_begin(&rx_stats->syncp); u64_stats_update_begin(&rx_stats->syncp);
if (act == XDP_TX)
rx_stats->xdp_tx++;
rx_stats->packets++; rx_stats->packets++;
rx_stats->bytes += nvchan->rsc.pktlen; rx_stats->bytes += nvchan->rsc.pktlen;
@ -1353,28 +1309,29 @@ static void netvsc_get_pcpu_stats(struct net_device *net,
/* fetch percpu stats of netvsc */ /* fetch percpu stats of netvsc */
for (i = 0; i < nvdev->num_chn; i++) { for (i = 0; i < nvdev->num_chn; i++) {
const struct netvsc_channel *nvchan = &nvdev->chan_table[i]; const struct netvsc_channel *nvchan = &nvdev->chan_table[i];
const struct netvsc_stats *stats; const struct netvsc_stats_tx *tx_stats;
const struct netvsc_stats_rx *rx_stats;
struct netvsc_ethtool_pcpu_stats *this_tot = struct netvsc_ethtool_pcpu_stats *this_tot =
&pcpu_tot[nvchan->channel->target_cpu]; &pcpu_tot[nvchan->channel->target_cpu];
u64 packets, bytes; u64 packets, bytes;
unsigned int start; unsigned int start;
stats = &nvchan->tx_stats; tx_stats = &nvchan->tx_stats;
do { do {
start = u64_stats_fetch_begin_irq(&stats->syncp); start = u64_stats_fetch_begin_irq(&tx_stats->syncp);
packets = stats->packets; packets = tx_stats->packets;
bytes = stats->bytes; bytes = tx_stats->bytes;
} while (u64_stats_fetch_retry_irq(&stats->syncp, start)); } while (u64_stats_fetch_retry_irq(&tx_stats->syncp, start));
this_tot->tx_bytes += bytes; this_tot->tx_bytes += bytes;
this_tot->tx_packets += packets; this_tot->tx_packets += packets;
stats = &nvchan->rx_stats; rx_stats = &nvchan->rx_stats;
do { do {
start = u64_stats_fetch_begin_irq(&stats->syncp); start = u64_stats_fetch_begin_irq(&rx_stats->syncp);
packets = stats->packets; packets = rx_stats->packets;
bytes = stats->bytes; bytes = rx_stats->bytes;
} while (u64_stats_fetch_retry_irq(&stats->syncp, start)); } while (u64_stats_fetch_retry_irq(&rx_stats->syncp, start));
this_tot->rx_bytes += bytes; this_tot->rx_bytes += bytes;
this_tot->rx_packets += packets; this_tot->rx_packets += packets;
@ -1406,27 +1363,28 @@ static void netvsc_get_stats64(struct net_device *net,
for (i = 0; i < nvdev->num_chn; i++) { for (i = 0; i < nvdev->num_chn; i++) {
const struct netvsc_channel *nvchan = &nvdev->chan_table[i]; const struct netvsc_channel *nvchan = &nvdev->chan_table[i];
const struct netvsc_stats *stats; const struct netvsc_stats_tx *tx_stats;
const struct netvsc_stats_rx *rx_stats;
u64 packets, bytes, multicast; u64 packets, bytes, multicast;
unsigned int start; unsigned int start;
stats = &nvchan->tx_stats; tx_stats = &nvchan->tx_stats;
do { do {
start = u64_stats_fetch_begin_irq(&stats->syncp); start = u64_stats_fetch_begin_irq(&tx_stats->syncp);
packets = stats->packets; packets = tx_stats->packets;
bytes = stats->bytes; bytes = tx_stats->bytes;
} while (u64_stats_fetch_retry_irq(&stats->syncp, start)); } while (u64_stats_fetch_retry_irq(&tx_stats->syncp, start));
t->tx_bytes += bytes; t->tx_bytes += bytes;
t->tx_packets += packets; t->tx_packets += packets;
stats = &nvchan->rx_stats; rx_stats = &nvchan->rx_stats;
do { do {
start = u64_stats_fetch_begin_irq(&stats->syncp); start = u64_stats_fetch_begin_irq(&rx_stats->syncp);
packets = stats->packets; packets = rx_stats->packets;
bytes = stats->bytes; bytes = rx_stats->bytes;
multicast = stats->multicast + stats->broadcast; multicast = rx_stats->multicast + rx_stats->broadcast;
} while (u64_stats_fetch_retry_irq(&stats->syncp, start)); } while (u64_stats_fetch_retry_irq(&rx_stats->syncp, start));
t->rx_bytes += bytes; t->rx_bytes += bytes;
t->rx_packets += packets; t->rx_packets += packets;
@ -1515,8 +1473,8 @@ static const struct {
/* statistics per queue (rx/tx packets/bytes) */ /* statistics per queue (rx/tx packets/bytes) */
#define NETVSC_PCPU_STATS_LEN (num_present_cpus() * ARRAY_SIZE(pcpu_stats)) #define NETVSC_PCPU_STATS_LEN (num_present_cpus() * ARRAY_SIZE(pcpu_stats))
/* 5 statistics per queue (rx/tx packets/bytes, rx xdp_drop) */ /* 8 statistics per queue (rx/tx packets/bytes, XDP actions) */
#define NETVSC_QUEUE_STATS_LEN(dev) ((dev)->num_chn * 5) #define NETVSC_QUEUE_STATS_LEN(dev) ((dev)->num_chn * 8)
static int netvsc_get_sset_count(struct net_device *dev, int string_set) static int netvsc_get_sset_count(struct net_device *dev, int string_set)
{ {
@ -1543,12 +1501,16 @@ static void netvsc_get_ethtool_stats(struct net_device *dev,
struct net_device_context *ndc = netdev_priv(dev); struct net_device_context *ndc = netdev_priv(dev);
struct netvsc_device *nvdev = rtnl_dereference(ndc->nvdev); struct netvsc_device *nvdev = rtnl_dereference(ndc->nvdev);
const void *nds = &ndc->eth_stats; const void *nds = &ndc->eth_stats;
const struct netvsc_stats *qstats; const struct netvsc_stats_tx *tx_stats;
const struct netvsc_stats_rx *rx_stats;
struct netvsc_vf_pcpu_stats sum; struct netvsc_vf_pcpu_stats sum;
struct netvsc_ethtool_pcpu_stats *pcpu_sum; struct netvsc_ethtool_pcpu_stats *pcpu_sum;
unsigned int start; unsigned int start;
u64 packets, bytes; u64 packets, bytes;
u64 xdp_drop; u64 xdp_drop;
u64 xdp_redirect;
u64 xdp_tx;
u64 xdp_xmit;
int i, j, cpu; int i, j, cpu;
if (!nvdev) if (!nvdev)
@ -1562,26 +1524,32 @@ static void netvsc_get_ethtool_stats(struct net_device *dev,
data[i++] = *(u64 *)((void *)&sum + vf_stats[j].offset); data[i++] = *(u64 *)((void *)&sum + vf_stats[j].offset);
for (j = 0; j < nvdev->num_chn; j++) { for (j = 0; j < nvdev->num_chn; j++) {
qstats = &nvdev->chan_table[j].tx_stats; tx_stats = &nvdev->chan_table[j].tx_stats;
do { do {
start = u64_stats_fetch_begin_irq(&qstats->syncp); start = u64_stats_fetch_begin_irq(&tx_stats->syncp);
packets = qstats->packets; packets = tx_stats->packets;
bytes = qstats->bytes; bytes = tx_stats->bytes;
} while (u64_stats_fetch_retry_irq(&qstats->syncp, start)); xdp_xmit = tx_stats->xdp_xmit;
} while (u64_stats_fetch_retry_irq(&tx_stats->syncp, start));
data[i++] = packets; data[i++] = packets;
data[i++] = bytes; data[i++] = bytes;
data[i++] = xdp_xmit;
qstats = &nvdev->chan_table[j].rx_stats; rx_stats = &nvdev->chan_table[j].rx_stats;
do { do {
start = u64_stats_fetch_begin_irq(&qstats->syncp); start = u64_stats_fetch_begin_irq(&rx_stats->syncp);
packets = qstats->packets; packets = rx_stats->packets;
bytes = qstats->bytes; bytes = rx_stats->bytes;
xdp_drop = qstats->xdp_drop; xdp_drop = rx_stats->xdp_drop;
} while (u64_stats_fetch_retry_irq(&qstats->syncp, start)); xdp_redirect = rx_stats->xdp_redirect;
xdp_tx = rx_stats->xdp_tx;
} while (u64_stats_fetch_retry_irq(&rx_stats->syncp, start));
data[i++] = packets; data[i++] = packets;
data[i++] = bytes; data[i++] = bytes;
data[i++] = xdp_drop; data[i++] = xdp_drop;
data[i++] = xdp_redirect;
data[i++] = xdp_tx;
} }
pcpu_sum = kvmalloc_array(num_possible_cpus(), pcpu_sum = kvmalloc_array(num_possible_cpus(),
@ -1622,9 +1590,12 @@ static void netvsc_get_strings(struct net_device *dev, u32 stringset, u8 *data)
for (i = 0; i < nvdev->num_chn; i++) { for (i = 0; i < nvdev->num_chn; i++) {
ethtool_sprintf(&p, "tx_queue_%u_packets", i); ethtool_sprintf(&p, "tx_queue_%u_packets", i);
ethtool_sprintf(&p, "tx_queue_%u_bytes", i); ethtool_sprintf(&p, "tx_queue_%u_bytes", i);
ethtool_sprintf(&p, "tx_queue_%u_xdp_xmit", i);
ethtool_sprintf(&p, "rx_queue_%u_packets", i); ethtool_sprintf(&p, "rx_queue_%u_packets", i);
ethtool_sprintf(&p, "rx_queue_%u_bytes", i); ethtool_sprintf(&p, "rx_queue_%u_bytes", i);
ethtool_sprintf(&p, "rx_queue_%u_xdp_drop", i); ethtool_sprintf(&p, "rx_queue_%u_xdp_drop", i);
ethtool_sprintf(&p, "rx_queue_%u_xdp_redirect", i);
ethtool_sprintf(&p, "rx_queue_%u_xdp_tx", i);
} }
for_each_present_cpu(cpu) { for_each_present_cpu(cpu) {
@ -2057,6 +2028,7 @@ static const struct net_device_ops device_ops = {
.ndo_select_queue = netvsc_select_queue, .ndo_select_queue = netvsc_select_queue,
.ndo_get_stats64 = netvsc_get_stats64, .ndo_get_stats64 = netvsc_get_stats64,
.ndo_bpf = netvsc_bpf, .ndo_bpf = netvsc_bpf,
.ndo_xdp_xmit = netvsc_ndoxdp_xmit,
}; };
/* /*