1089 lines
25 KiB
C
1089 lines
25 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/* Copyright(c) 2020 Intel Corporation. */
|
|
|
|
#define _GNU_SOURCE
|
|
#include <poll.h>
|
|
#include <pthread.h>
|
|
#include <signal.h>
|
|
#include <sched.h>
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
#include <sys/mman.h>
|
|
#include <sys/resource.h>
|
|
#include <sys/socket.h>
|
|
#include <sys/types.h>
|
|
#include <time.h>
|
|
#include <unistd.h>
|
|
#include <getopt.h>
|
|
#include <netinet/ether.h>
|
|
#include <net/if.h>
|
|
|
|
#include <linux/bpf.h>
|
|
#include <linux/if_link.h>
|
|
#include <linux/if_xdp.h>
|
|
|
|
#include <bpf/libbpf.h>
|
|
#include <bpf/xsk.h>
|
|
#include <bpf/bpf.h>
|
|
|
|
/* libbpf APIs for AF_XDP are deprecated starting from v0.7 */
|
|
#pragma GCC diagnostic ignored "-Wdeprecated-declarations"
|
|
|
|
#define ARRAY_SIZE(x) (sizeof(x) / sizeof((x)[0]))
|
|
|
|
typedef __u64 u64;
|
|
typedef __u32 u32;
|
|
typedef __u16 u16;
|
|
typedef __u8 u8;
|
|
|
|
/* This program illustrates the packet forwarding between multiple AF_XDP
|
|
* sockets in multi-threaded environment. All threads are sharing a common
|
|
* buffer pool, with each socket having its own private buffer cache.
|
|
*
|
|
* Example 1: Single thread handling two sockets. The packets received by socket
|
|
* A (interface IFA, queue QA) are forwarded to socket B (interface IFB, queue
|
|
* QB), while the packets received by socket B are forwarded to socket A. The
|
|
* thread is running on CPU core X:
|
|
*
|
|
* ./xsk_fwd -i IFA -q QA -i IFB -q QB -c X
|
|
*
|
|
* Example 2: Two threads, each handling two sockets. The thread running on CPU
|
|
* core X forwards all the packets received by socket A to socket B, and all the
|
|
* packets received by socket B to socket A. The thread running on CPU core Y is
|
|
* performing the same packet forwarding between sockets C and D:
|
|
*
|
|
* ./xsk_fwd -i IFA -q QA -i IFB -q QB -i IFC -q QC -i IFD -q QD
|
|
* -c CX -c CY
|
|
*/
|
|
|
|
/*
|
|
* Buffer pool and buffer cache
|
|
*
|
|
* For packet forwarding, the packet buffers are typically allocated from the
|
|
* pool for packet reception and freed back to the pool for further reuse once
|
|
* the packet transmission is completed.
|
|
*
|
|
* The buffer pool is shared between multiple threads. In order to minimize the
|
|
* access latency to the shared buffer pool, each thread creates one (or
|
|
* several) buffer caches, which, unlike the buffer pool, are private to the
|
|
* thread that creates them and therefore cannot be shared with other threads.
|
|
* The access to the shared pool is only needed either (A) when the cache gets
|
|
* empty due to repeated buffer allocations and it needs to be replenished from
|
|
* the pool, or (B) when the cache gets full due to repeated buffer free and it
|
|
* needs to be flushed back to the pull.
|
|
*
|
|
* In a packet forwarding system, a packet received on any input port can
|
|
* potentially be transmitted on any output port, depending on the forwarding
|
|
* configuration. For AF_XDP sockets, for this to work with zero-copy of the
|
|
* packet buffers when, it is required that the buffer pool memory fits into the
|
|
* UMEM area shared by all the sockets.
|
|
*/
|
|
|
|
struct bpool_params {
|
|
u32 n_buffers;
|
|
u32 buffer_size;
|
|
int mmap_flags;
|
|
|
|
u32 n_users_max;
|
|
u32 n_buffers_per_slab;
|
|
};
|
|
|
|
/* This buffer pool implementation organizes the buffers into equally sized
|
|
* slabs of *n_buffers_per_slab*. Initially, there are *n_slabs* slabs in the
|
|
* pool that are completely filled with buffer pointers (full slabs).
|
|
*
|
|
* Each buffer cache has a slab for buffer allocation and a slab for buffer
|
|
* free, with both of these slabs initially empty. When the cache's allocation
|
|
* slab goes empty, it is swapped with one of the available full slabs from the
|
|
* pool, if any is available. When the cache's free slab goes full, it is
|
|
* swapped for one of the empty slabs from the pool, which is guaranteed to
|
|
* succeed.
|
|
*
|
|
* Partially filled slabs never get traded between the cache and the pool
|
|
* (except when the cache itself is destroyed), which enables fast operation
|
|
* through pointer swapping.
|
|
*/
|
|
struct bpool {
|
|
struct bpool_params params;
|
|
pthread_mutex_t lock;
|
|
void *addr;
|
|
|
|
u64 **slabs;
|
|
u64 **slabs_reserved;
|
|
u64 *buffers;
|
|
u64 *buffers_reserved;
|
|
|
|
u64 n_slabs;
|
|
u64 n_slabs_reserved;
|
|
u64 n_buffers;
|
|
|
|
u64 n_slabs_available;
|
|
u64 n_slabs_reserved_available;
|
|
|
|
struct xsk_umem_config umem_cfg;
|
|
struct xsk_ring_prod umem_fq;
|
|
struct xsk_ring_cons umem_cq;
|
|
struct xsk_umem *umem;
|
|
};
|
|
|
|
static struct bpool *
|
|
bpool_init(struct bpool_params *params,
|
|
struct xsk_umem_config *umem_cfg)
|
|
{
|
|
struct rlimit r = {RLIM_INFINITY, RLIM_INFINITY};
|
|
u64 n_slabs, n_slabs_reserved, n_buffers, n_buffers_reserved;
|
|
u64 slabs_size, slabs_reserved_size;
|
|
u64 buffers_size, buffers_reserved_size;
|
|
u64 total_size, i;
|
|
struct bpool *bp;
|
|
u8 *p;
|
|
int status;
|
|
|
|
/* mmap prep. */
|
|
if (setrlimit(RLIMIT_MEMLOCK, &r))
|
|
return NULL;
|
|
|
|
/* bpool internals dimensioning. */
|
|
n_slabs = (params->n_buffers + params->n_buffers_per_slab - 1) /
|
|
params->n_buffers_per_slab;
|
|
n_slabs_reserved = params->n_users_max * 2;
|
|
n_buffers = n_slabs * params->n_buffers_per_slab;
|
|
n_buffers_reserved = n_slabs_reserved * params->n_buffers_per_slab;
|
|
|
|
slabs_size = n_slabs * sizeof(u64 *);
|
|
slabs_reserved_size = n_slabs_reserved * sizeof(u64 *);
|
|
buffers_size = n_buffers * sizeof(u64);
|
|
buffers_reserved_size = n_buffers_reserved * sizeof(u64);
|
|
|
|
total_size = sizeof(struct bpool) +
|
|
slabs_size + slabs_reserved_size +
|
|
buffers_size + buffers_reserved_size;
|
|
|
|
/* bpool memory allocation. */
|
|
p = calloc(total_size, sizeof(u8));
|
|
if (!p)
|
|
return NULL;
|
|
|
|
/* bpool memory initialization. */
|
|
bp = (struct bpool *)p;
|
|
memcpy(&bp->params, params, sizeof(*params));
|
|
bp->params.n_buffers = n_buffers;
|
|
|
|
bp->slabs = (u64 **)&p[sizeof(struct bpool)];
|
|
bp->slabs_reserved = (u64 **)&p[sizeof(struct bpool) +
|
|
slabs_size];
|
|
bp->buffers = (u64 *)&p[sizeof(struct bpool) +
|
|
slabs_size + slabs_reserved_size];
|
|
bp->buffers_reserved = (u64 *)&p[sizeof(struct bpool) +
|
|
slabs_size + slabs_reserved_size + buffers_size];
|
|
|
|
bp->n_slabs = n_slabs;
|
|
bp->n_slabs_reserved = n_slabs_reserved;
|
|
bp->n_buffers = n_buffers;
|
|
|
|
for (i = 0; i < n_slabs; i++)
|
|
bp->slabs[i] = &bp->buffers[i * params->n_buffers_per_slab];
|
|
bp->n_slabs_available = n_slabs;
|
|
|
|
for (i = 0; i < n_slabs_reserved; i++)
|
|
bp->slabs_reserved[i] = &bp->buffers_reserved[i *
|
|
params->n_buffers_per_slab];
|
|
bp->n_slabs_reserved_available = n_slabs_reserved;
|
|
|
|
for (i = 0; i < n_buffers; i++)
|
|
bp->buffers[i] = i * params->buffer_size;
|
|
|
|
/* lock. */
|
|
status = pthread_mutex_init(&bp->lock, NULL);
|
|
if (status) {
|
|
free(p);
|
|
return NULL;
|
|
}
|
|
|
|
/* mmap. */
|
|
bp->addr = mmap(NULL,
|
|
n_buffers * params->buffer_size,
|
|
PROT_READ | PROT_WRITE,
|
|
MAP_PRIVATE | MAP_ANONYMOUS | params->mmap_flags,
|
|
-1,
|
|
0);
|
|
if (bp->addr == MAP_FAILED) {
|
|
pthread_mutex_destroy(&bp->lock);
|
|
free(p);
|
|
return NULL;
|
|
}
|
|
|
|
/* umem. */
|
|
status = xsk_umem__create(&bp->umem,
|
|
bp->addr,
|
|
bp->params.n_buffers * bp->params.buffer_size,
|
|
&bp->umem_fq,
|
|
&bp->umem_cq,
|
|
umem_cfg);
|
|
if (status) {
|
|
munmap(bp->addr, bp->params.n_buffers * bp->params.buffer_size);
|
|
pthread_mutex_destroy(&bp->lock);
|
|
free(p);
|
|
return NULL;
|
|
}
|
|
memcpy(&bp->umem_cfg, umem_cfg, sizeof(*umem_cfg));
|
|
|
|
return bp;
|
|
}
|
|
|
|
static void
|
|
bpool_free(struct bpool *bp)
|
|
{
|
|
if (!bp)
|
|
return;
|
|
|
|
xsk_umem__delete(bp->umem);
|
|
munmap(bp->addr, bp->params.n_buffers * bp->params.buffer_size);
|
|
pthread_mutex_destroy(&bp->lock);
|
|
free(bp);
|
|
}
|
|
|
|
struct bcache {
|
|
struct bpool *bp;
|
|
|
|
u64 *slab_cons;
|
|
u64 *slab_prod;
|
|
|
|
u64 n_buffers_cons;
|
|
u64 n_buffers_prod;
|
|
};
|
|
|
|
static u32
|
|
bcache_slab_size(struct bcache *bc)
|
|
{
|
|
struct bpool *bp = bc->bp;
|
|
|
|
return bp->params.n_buffers_per_slab;
|
|
}
|
|
|
|
static struct bcache *
|
|
bcache_init(struct bpool *bp)
|
|
{
|
|
struct bcache *bc;
|
|
|
|
bc = calloc(1, sizeof(struct bcache));
|
|
if (!bc)
|
|
return NULL;
|
|
|
|
bc->bp = bp;
|
|
bc->n_buffers_cons = 0;
|
|
bc->n_buffers_prod = 0;
|
|
|
|
pthread_mutex_lock(&bp->lock);
|
|
if (bp->n_slabs_reserved_available == 0) {
|
|
pthread_mutex_unlock(&bp->lock);
|
|
free(bc);
|
|
return NULL;
|
|
}
|
|
|
|
bc->slab_cons = bp->slabs_reserved[bp->n_slabs_reserved_available - 1];
|
|
bc->slab_prod = bp->slabs_reserved[bp->n_slabs_reserved_available - 2];
|
|
bp->n_slabs_reserved_available -= 2;
|
|
pthread_mutex_unlock(&bp->lock);
|
|
|
|
return bc;
|
|
}
|
|
|
|
static void
|
|
bcache_free(struct bcache *bc)
|
|
{
|
|
struct bpool *bp;
|
|
|
|
if (!bc)
|
|
return;
|
|
|
|
/* In order to keep this example simple, the case of freeing any
|
|
* existing buffers from the cache back to the pool is ignored.
|
|
*/
|
|
|
|
bp = bc->bp;
|
|
pthread_mutex_lock(&bp->lock);
|
|
bp->slabs_reserved[bp->n_slabs_reserved_available] = bc->slab_prod;
|
|
bp->slabs_reserved[bp->n_slabs_reserved_available + 1] = bc->slab_cons;
|
|
bp->n_slabs_reserved_available += 2;
|
|
pthread_mutex_unlock(&bp->lock);
|
|
|
|
free(bc);
|
|
}
|
|
|
|
/* To work correctly, the implementation requires that the *n_buffers* input
|
|
* argument is never greater than the buffer pool's *n_buffers_per_slab*. This
|
|
* is typically the case, with one exception taking place when large number of
|
|
* buffers are allocated at init time (e.g. for the UMEM fill queue setup).
|
|
*/
|
|
static inline u32
|
|
bcache_cons_check(struct bcache *bc, u32 n_buffers)
|
|
{
|
|
struct bpool *bp = bc->bp;
|
|
u64 n_buffers_per_slab = bp->params.n_buffers_per_slab;
|
|
u64 n_buffers_cons = bc->n_buffers_cons;
|
|
u64 n_slabs_available;
|
|
u64 *slab_full;
|
|
|
|
/*
|
|
* Consumer slab is not empty: Use what's available locally. Do not
|
|
* look for more buffers from the pool when the ask can only be
|
|
* partially satisfied.
|
|
*/
|
|
if (n_buffers_cons)
|
|
return (n_buffers_cons < n_buffers) ?
|
|
n_buffers_cons :
|
|
n_buffers;
|
|
|
|
/*
|
|
* Consumer slab is empty: look to trade the current consumer slab
|
|
* (full) for a full slab from the pool, if any is available.
|
|
*/
|
|
pthread_mutex_lock(&bp->lock);
|
|
n_slabs_available = bp->n_slabs_available;
|
|
if (!n_slabs_available) {
|
|
pthread_mutex_unlock(&bp->lock);
|
|
return 0;
|
|
}
|
|
|
|
n_slabs_available--;
|
|
slab_full = bp->slabs[n_slabs_available];
|
|
bp->slabs[n_slabs_available] = bc->slab_cons;
|
|
bp->n_slabs_available = n_slabs_available;
|
|
pthread_mutex_unlock(&bp->lock);
|
|
|
|
bc->slab_cons = slab_full;
|
|
bc->n_buffers_cons = n_buffers_per_slab;
|
|
return n_buffers;
|
|
}
|
|
|
|
static inline u64
|
|
bcache_cons(struct bcache *bc)
|
|
{
|
|
u64 n_buffers_cons = bc->n_buffers_cons - 1;
|
|
u64 buffer;
|
|
|
|
buffer = bc->slab_cons[n_buffers_cons];
|
|
bc->n_buffers_cons = n_buffers_cons;
|
|
return buffer;
|
|
}
|
|
|
|
static inline void
|
|
bcache_prod(struct bcache *bc, u64 buffer)
|
|
{
|
|
struct bpool *bp = bc->bp;
|
|
u64 n_buffers_per_slab = bp->params.n_buffers_per_slab;
|
|
u64 n_buffers_prod = bc->n_buffers_prod;
|
|
u64 n_slabs_available;
|
|
u64 *slab_empty;
|
|
|
|
/*
|
|
* Producer slab is not yet full: store the current buffer to it.
|
|
*/
|
|
if (n_buffers_prod < n_buffers_per_slab) {
|
|
bc->slab_prod[n_buffers_prod] = buffer;
|
|
bc->n_buffers_prod = n_buffers_prod + 1;
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Producer slab is full: trade the cache's current producer slab
|
|
* (full) for an empty slab from the pool, then store the current
|
|
* buffer to the new producer slab. As one full slab exists in the
|
|
* cache, it is guaranteed that there is at least one empty slab
|
|
* available in the pool.
|
|
*/
|
|
pthread_mutex_lock(&bp->lock);
|
|
n_slabs_available = bp->n_slabs_available;
|
|
slab_empty = bp->slabs[n_slabs_available];
|
|
bp->slabs[n_slabs_available] = bc->slab_prod;
|
|
bp->n_slabs_available = n_slabs_available + 1;
|
|
pthread_mutex_unlock(&bp->lock);
|
|
|
|
slab_empty[0] = buffer;
|
|
bc->slab_prod = slab_empty;
|
|
bc->n_buffers_prod = 1;
|
|
}
|
|
|
|
/*
|
|
* Port
|
|
*
|
|
* Each of the forwarding ports sits on top of an AF_XDP socket. In order for
|
|
* packet forwarding to happen with no packet buffer copy, all the sockets need
|
|
* to share the same UMEM area, which is used as the buffer pool memory.
|
|
*/
|
|
#ifndef MAX_BURST_RX
|
|
#define MAX_BURST_RX 64
|
|
#endif
|
|
|
|
#ifndef MAX_BURST_TX
|
|
#define MAX_BURST_TX 64
|
|
#endif
|
|
|
|
struct burst_rx {
|
|
u64 addr[MAX_BURST_RX];
|
|
u32 len[MAX_BURST_RX];
|
|
};
|
|
|
|
struct burst_tx {
|
|
u64 addr[MAX_BURST_TX];
|
|
u32 len[MAX_BURST_TX];
|
|
u32 n_pkts;
|
|
};
|
|
|
|
struct port_params {
|
|
struct xsk_socket_config xsk_cfg;
|
|
struct bpool *bp;
|
|
const char *iface;
|
|
u32 iface_queue;
|
|
};
|
|
|
|
struct port {
|
|
struct port_params params;
|
|
|
|
struct bcache *bc;
|
|
|
|
struct xsk_ring_cons rxq;
|
|
struct xsk_ring_prod txq;
|
|
struct xsk_ring_prod umem_fq;
|
|
struct xsk_ring_cons umem_cq;
|
|
struct xsk_socket *xsk;
|
|
int umem_fq_initialized;
|
|
|
|
u64 n_pkts_rx;
|
|
u64 n_pkts_tx;
|
|
};
|
|
|
|
static void
|
|
port_free(struct port *p)
|
|
{
|
|
if (!p)
|
|
return;
|
|
|
|
/* To keep this example simple, the code to free the buffers from the
|
|
* socket's receive and transmit queues, as well as from the UMEM fill
|
|
* and completion queues, is not included.
|
|
*/
|
|
|
|
if (p->xsk)
|
|
xsk_socket__delete(p->xsk);
|
|
|
|
bcache_free(p->bc);
|
|
|
|
free(p);
|
|
}
|
|
|
|
static struct port *
|
|
port_init(struct port_params *params)
|
|
{
|
|
struct port *p;
|
|
u32 umem_fq_size, pos = 0;
|
|
int status, i;
|
|
|
|
/* Memory allocation and initialization. */
|
|
p = calloc(sizeof(struct port), 1);
|
|
if (!p)
|
|
return NULL;
|
|
|
|
memcpy(&p->params, params, sizeof(p->params));
|
|
umem_fq_size = params->bp->umem_cfg.fill_size;
|
|
|
|
/* bcache. */
|
|
p->bc = bcache_init(params->bp);
|
|
if (!p->bc ||
|
|
(bcache_slab_size(p->bc) < umem_fq_size) ||
|
|
(bcache_cons_check(p->bc, umem_fq_size) < umem_fq_size)) {
|
|
port_free(p);
|
|
return NULL;
|
|
}
|
|
|
|
/* xsk socket. */
|
|
status = xsk_socket__create_shared(&p->xsk,
|
|
params->iface,
|
|
params->iface_queue,
|
|
params->bp->umem,
|
|
&p->rxq,
|
|
&p->txq,
|
|
&p->umem_fq,
|
|
&p->umem_cq,
|
|
¶ms->xsk_cfg);
|
|
if (status) {
|
|
port_free(p);
|
|
return NULL;
|
|
}
|
|
|
|
/* umem fq. */
|
|
xsk_ring_prod__reserve(&p->umem_fq, umem_fq_size, &pos);
|
|
|
|
for (i = 0; i < umem_fq_size; i++)
|
|
*xsk_ring_prod__fill_addr(&p->umem_fq, pos + i) =
|
|
bcache_cons(p->bc);
|
|
|
|
xsk_ring_prod__submit(&p->umem_fq, umem_fq_size);
|
|
p->umem_fq_initialized = 1;
|
|
|
|
return p;
|
|
}
|
|
|
|
static inline u32
|
|
port_rx_burst(struct port *p, struct burst_rx *b)
|
|
{
|
|
u32 n_pkts, pos, i;
|
|
|
|
/* Free buffers for FQ replenish. */
|
|
n_pkts = ARRAY_SIZE(b->addr);
|
|
|
|
n_pkts = bcache_cons_check(p->bc, n_pkts);
|
|
if (!n_pkts)
|
|
return 0;
|
|
|
|
/* RXQ. */
|
|
n_pkts = xsk_ring_cons__peek(&p->rxq, n_pkts, &pos);
|
|
if (!n_pkts) {
|
|
if (xsk_ring_prod__needs_wakeup(&p->umem_fq)) {
|
|
struct pollfd pollfd = {
|
|
.fd = xsk_socket__fd(p->xsk),
|
|
.events = POLLIN,
|
|
};
|
|
|
|
poll(&pollfd, 1, 0);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
for (i = 0; i < n_pkts; i++) {
|
|
b->addr[i] = xsk_ring_cons__rx_desc(&p->rxq, pos + i)->addr;
|
|
b->len[i] = xsk_ring_cons__rx_desc(&p->rxq, pos + i)->len;
|
|
}
|
|
|
|
xsk_ring_cons__release(&p->rxq, n_pkts);
|
|
p->n_pkts_rx += n_pkts;
|
|
|
|
/* UMEM FQ. */
|
|
for ( ; ; ) {
|
|
int status;
|
|
|
|
status = xsk_ring_prod__reserve(&p->umem_fq, n_pkts, &pos);
|
|
if (status == n_pkts)
|
|
break;
|
|
|
|
if (xsk_ring_prod__needs_wakeup(&p->umem_fq)) {
|
|
struct pollfd pollfd = {
|
|
.fd = xsk_socket__fd(p->xsk),
|
|
.events = POLLIN,
|
|
};
|
|
|
|
poll(&pollfd, 1, 0);
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < n_pkts; i++)
|
|
*xsk_ring_prod__fill_addr(&p->umem_fq, pos + i) =
|
|
bcache_cons(p->bc);
|
|
|
|
xsk_ring_prod__submit(&p->umem_fq, n_pkts);
|
|
|
|
return n_pkts;
|
|
}
|
|
|
|
static inline void
|
|
port_tx_burst(struct port *p, struct burst_tx *b)
|
|
{
|
|
u32 n_pkts, pos, i;
|
|
int status;
|
|
|
|
/* UMEM CQ. */
|
|
n_pkts = p->params.bp->umem_cfg.comp_size;
|
|
|
|
n_pkts = xsk_ring_cons__peek(&p->umem_cq, n_pkts, &pos);
|
|
|
|
for (i = 0; i < n_pkts; i++) {
|
|
u64 addr = *xsk_ring_cons__comp_addr(&p->umem_cq, pos + i);
|
|
|
|
bcache_prod(p->bc, addr);
|
|
}
|
|
|
|
xsk_ring_cons__release(&p->umem_cq, n_pkts);
|
|
|
|
/* TXQ. */
|
|
n_pkts = b->n_pkts;
|
|
|
|
for ( ; ; ) {
|
|
status = xsk_ring_prod__reserve(&p->txq, n_pkts, &pos);
|
|
if (status == n_pkts)
|
|
break;
|
|
|
|
if (xsk_ring_prod__needs_wakeup(&p->txq))
|
|
sendto(xsk_socket__fd(p->xsk), NULL, 0, MSG_DONTWAIT,
|
|
NULL, 0);
|
|
}
|
|
|
|
for (i = 0; i < n_pkts; i++) {
|
|
xsk_ring_prod__tx_desc(&p->txq, pos + i)->addr = b->addr[i];
|
|
xsk_ring_prod__tx_desc(&p->txq, pos + i)->len = b->len[i];
|
|
}
|
|
|
|
xsk_ring_prod__submit(&p->txq, n_pkts);
|
|
if (xsk_ring_prod__needs_wakeup(&p->txq))
|
|
sendto(xsk_socket__fd(p->xsk), NULL, 0, MSG_DONTWAIT, NULL, 0);
|
|
p->n_pkts_tx += n_pkts;
|
|
}
|
|
|
|
/*
|
|
* Thread
|
|
*
|
|
* Packet forwarding threads.
|
|
*/
|
|
#ifndef MAX_PORTS_PER_THREAD
|
|
#define MAX_PORTS_PER_THREAD 16
|
|
#endif
|
|
|
|
struct thread_data {
|
|
struct port *ports_rx[MAX_PORTS_PER_THREAD];
|
|
struct port *ports_tx[MAX_PORTS_PER_THREAD];
|
|
u32 n_ports_rx;
|
|
struct burst_rx burst_rx;
|
|
struct burst_tx burst_tx[MAX_PORTS_PER_THREAD];
|
|
u32 cpu_core_id;
|
|
int quit;
|
|
};
|
|
|
|
static void swap_mac_addresses(void *data)
|
|
{
|
|
struct ether_header *eth = (struct ether_header *)data;
|
|
struct ether_addr *src_addr = (struct ether_addr *)ð->ether_shost;
|
|
struct ether_addr *dst_addr = (struct ether_addr *)ð->ether_dhost;
|
|
struct ether_addr tmp;
|
|
|
|
tmp = *src_addr;
|
|
*src_addr = *dst_addr;
|
|
*dst_addr = tmp;
|
|
}
|
|
|
|
static void *
|
|
thread_func(void *arg)
|
|
{
|
|
struct thread_data *t = arg;
|
|
cpu_set_t cpu_cores;
|
|
u32 i;
|
|
|
|
CPU_ZERO(&cpu_cores);
|
|
CPU_SET(t->cpu_core_id, &cpu_cores);
|
|
pthread_setaffinity_np(pthread_self(), sizeof(cpu_set_t), &cpu_cores);
|
|
|
|
for (i = 0; !t->quit; i = (i + 1) & (t->n_ports_rx - 1)) {
|
|
struct port *port_rx = t->ports_rx[i];
|
|
struct port *port_tx = t->ports_tx[i];
|
|
struct burst_rx *brx = &t->burst_rx;
|
|
struct burst_tx *btx = &t->burst_tx[i];
|
|
u32 n_pkts, j;
|
|
|
|
/* RX. */
|
|
n_pkts = port_rx_burst(port_rx, brx);
|
|
if (!n_pkts)
|
|
continue;
|
|
|
|
/* Process & TX. */
|
|
for (j = 0; j < n_pkts; j++) {
|
|
u64 addr = xsk_umem__add_offset_to_addr(brx->addr[j]);
|
|
u8 *pkt = xsk_umem__get_data(port_rx->params.bp->addr,
|
|
addr);
|
|
|
|
swap_mac_addresses(pkt);
|
|
|
|
btx->addr[btx->n_pkts] = brx->addr[j];
|
|
btx->len[btx->n_pkts] = brx->len[j];
|
|
btx->n_pkts++;
|
|
|
|
if (btx->n_pkts == MAX_BURST_TX) {
|
|
port_tx_burst(port_tx, btx);
|
|
btx->n_pkts = 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/*
|
|
* Process
|
|
*/
|
|
static const struct bpool_params bpool_params_default = {
|
|
.n_buffers = 64 * 1024,
|
|
.buffer_size = XSK_UMEM__DEFAULT_FRAME_SIZE,
|
|
.mmap_flags = 0,
|
|
|
|
.n_users_max = 16,
|
|
.n_buffers_per_slab = XSK_RING_PROD__DEFAULT_NUM_DESCS * 2,
|
|
};
|
|
|
|
static const struct xsk_umem_config umem_cfg_default = {
|
|
.fill_size = XSK_RING_PROD__DEFAULT_NUM_DESCS * 2,
|
|
.comp_size = XSK_RING_CONS__DEFAULT_NUM_DESCS,
|
|
.frame_size = XSK_UMEM__DEFAULT_FRAME_SIZE,
|
|
.frame_headroom = XSK_UMEM__DEFAULT_FRAME_HEADROOM,
|
|
.flags = 0,
|
|
};
|
|
|
|
static const struct port_params port_params_default = {
|
|
.xsk_cfg = {
|
|
.rx_size = XSK_RING_CONS__DEFAULT_NUM_DESCS,
|
|
.tx_size = XSK_RING_PROD__DEFAULT_NUM_DESCS,
|
|
.libbpf_flags = 0,
|
|
.xdp_flags = XDP_FLAGS_DRV_MODE,
|
|
.bind_flags = XDP_USE_NEED_WAKEUP | XDP_ZEROCOPY,
|
|
},
|
|
|
|
.bp = NULL,
|
|
.iface = NULL,
|
|
.iface_queue = 0,
|
|
};
|
|
|
|
#ifndef MAX_PORTS
|
|
#define MAX_PORTS 64
|
|
#endif
|
|
|
|
#ifndef MAX_THREADS
|
|
#define MAX_THREADS 64
|
|
#endif
|
|
|
|
static struct bpool_params bpool_params;
|
|
static struct xsk_umem_config umem_cfg;
|
|
static struct bpool *bp;
|
|
|
|
static struct port_params port_params[MAX_PORTS];
|
|
static struct port *ports[MAX_PORTS];
|
|
static u64 n_pkts_rx[MAX_PORTS];
|
|
static u64 n_pkts_tx[MAX_PORTS];
|
|
static int n_ports;
|
|
|
|
static pthread_t threads[MAX_THREADS];
|
|
static struct thread_data thread_data[MAX_THREADS];
|
|
static int n_threads;
|
|
|
|
static void
|
|
print_usage(char *prog_name)
|
|
{
|
|
const char *usage =
|
|
"Usage:\n"
|
|
"\t%s [ -b SIZE ] -c CORE -i INTERFACE [ -q QUEUE ]\n"
|
|
"\n"
|
|
"-c CORE CPU core to run a packet forwarding thread\n"
|
|
" on. May be invoked multiple times.\n"
|
|
"\n"
|
|
"-b SIZE Number of buffers in the buffer pool shared\n"
|
|
" by all the forwarding threads. Default: %u.\n"
|
|
"\n"
|
|
"-i INTERFACE Network interface. Each (INTERFACE, QUEUE)\n"
|
|
" pair specifies one forwarding port. May be\n"
|
|
" invoked multiple times.\n"
|
|
"\n"
|
|
"-q QUEUE Network interface queue for RX and TX. Each\n"
|
|
" (INTERFACE, QUEUE) pair specified one\n"
|
|
" forwarding port. Default: %u. May be invoked\n"
|
|
" multiple times.\n"
|
|
"\n";
|
|
printf(usage,
|
|
prog_name,
|
|
bpool_params_default.n_buffers,
|
|
port_params_default.iface_queue);
|
|
}
|
|
|
|
static int
|
|
parse_args(int argc, char **argv)
|
|
{
|
|
struct option lgopts[] = {
|
|
{ NULL, 0, 0, 0 }
|
|
};
|
|
int opt, option_index;
|
|
|
|
/* Parse the input arguments. */
|
|
for ( ; ;) {
|
|
opt = getopt_long(argc, argv, "c:i:q:", lgopts, &option_index);
|
|
if (opt == EOF)
|
|
break;
|
|
|
|
switch (opt) {
|
|
case 'b':
|
|
bpool_params.n_buffers = atoi(optarg);
|
|
break;
|
|
|
|
case 'c':
|
|
if (n_threads == MAX_THREADS) {
|
|
printf("Max number of threads (%d) reached.\n",
|
|
MAX_THREADS);
|
|
return -1;
|
|
}
|
|
|
|
thread_data[n_threads].cpu_core_id = atoi(optarg);
|
|
n_threads++;
|
|
break;
|
|
|
|
case 'i':
|
|
if (n_ports == MAX_PORTS) {
|
|
printf("Max number of ports (%d) reached.\n",
|
|
MAX_PORTS);
|
|
return -1;
|
|
}
|
|
|
|
port_params[n_ports].iface = optarg;
|
|
port_params[n_ports].iface_queue = 0;
|
|
n_ports++;
|
|
break;
|
|
|
|
case 'q':
|
|
if (n_ports == 0) {
|
|
printf("No port specified for queue.\n");
|
|
return -1;
|
|
}
|
|
port_params[n_ports - 1].iface_queue = atoi(optarg);
|
|
break;
|
|
|
|
default:
|
|
printf("Illegal argument.\n");
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
optind = 1; /* reset getopt lib */
|
|
|
|
/* Check the input arguments. */
|
|
if (!n_ports) {
|
|
printf("No ports specified.\n");
|
|
return -1;
|
|
}
|
|
|
|
if (!n_threads) {
|
|
printf("No threads specified.\n");
|
|
return -1;
|
|
}
|
|
|
|
if (n_ports % n_threads) {
|
|
printf("Ports cannot be evenly distributed to threads.\n");
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
print_port(u32 port_id)
|
|
{
|
|
struct port *port = ports[port_id];
|
|
|
|
printf("Port %u: interface = %s, queue = %u\n",
|
|
port_id, port->params.iface, port->params.iface_queue);
|
|
}
|
|
|
|
static void
|
|
print_thread(u32 thread_id)
|
|
{
|
|
struct thread_data *t = &thread_data[thread_id];
|
|
u32 i;
|
|
|
|
printf("Thread %u (CPU core %u): ",
|
|
thread_id, t->cpu_core_id);
|
|
|
|
for (i = 0; i < t->n_ports_rx; i++) {
|
|
struct port *port_rx = t->ports_rx[i];
|
|
struct port *port_tx = t->ports_tx[i];
|
|
|
|
printf("(%s, %u) -> (%s, %u), ",
|
|
port_rx->params.iface,
|
|
port_rx->params.iface_queue,
|
|
port_tx->params.iface,
|
|
port_tx->params.iface_queue);
|
|
}
|
|
|
|
printf("\n");
|
|
}
|
|
|
|
static void
|
|
print_port_stats_separator(void)
|
|
{
|
|
printf("+-%4s-+-%12s-+-%13s-+-%12s-+-%13s-+\n",
|
|
"----",
|
|
"------------",
|
|
"-------------",
|
|
"------------",
|
|
"-------------");
|
|
}
|
|
|
|
static void
|
|
print_port_stats_header(void)
|
|
{
|
|
print_port_stats_separator();
|
|
printf("| %4s | %12s | %13s | %12s | %13s |\n",
|
|
"Port",
|
|
"RX packets",
|
|
"RX rate (pps)",
|
|
"TX packets",
|
|
"TX_rate (pps)");
|
|
print_port_stats_separator();
|
|
}
|
|
|
|
static void
|
|
print_port_stats_trailer(void)
|
|
{
|
|
print_port_stats_separator();
|
|
printf("\n");
|
|
}
|
|
|
|
static void
|
|
print_port_stats(int port_id, u64 ns_diff)
|
|
{
|
|
struct port *p = ports[port_id];
|
|
double rx_pps, tx_pps;
|
|
|
|
rx_pps = (p->n_pkts_rx - n_pkts_rx[port_id]) * 1000000000. / ns_diff;
|
|
tx_pps = (p->n_pkts_tx - n_pkts_tx[port_id]) * 1000000000. / ns_diff;
|
|
|
|
printf("| %4d | %12llu | %13.0f | %12llu | %13.0f |\n",
|
|
port_id,
|
|
p->n_pkts_rx,
|
|
rx_pps,
|
|
p->n_pkts_tx,
|
|
tx_pps);
|
|
|
|
n_pkts_rx[port_id] = p->n_pkts_rx;
|
|
n_pkts_tx[port_id] = p->n_pkts_tx;
|
|
}
|
|
|
|
static void
|
|
print_port_stats_all(u64 ns_diff)
|
|
{
|
|
int i;
|
|
|
|
print_port_stats_header();
|
|
for (i = 0; i < n_ports; i++)
|
|
print_port_stats(i, ns_diff);
|
|
print_port_stats_trailer();
|
|
}
|
|
|
|
static int quit;
|
|
|
|
static void
|
|
signal_handler(int sig)
|
|
{
|
|
quit = 1;
|
|
}
|
|
|
|
static void remove_xdp_program(void)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0 ; i < n_ports; i++)
|
|
bpf_set_link_xdp_fd(if_nametoindex(port_params[i].iface), -1,
|
|
port_params[i].xsk_cfg.xdp_flags);
|
|
}
|
|
|
|
int main(int argc, char **argv)
|
|
{
|
|
struct timespec time;
|
|
u64 ns0;
|
|
int i;
|
|
|
|
/* Parse args. */
|
|
memcpy(&bpool_params, &bpool_params_default,
|
|
sizeof(struct bpool_params));
|
|
memcpy(&umem_cfg, &umem_cfg_default,
|
|
sizeof(struct xsk_umem_config));
|
|
for (i = 0; i < MAX_PORTS; i++)
|
|
memcpy(&port_params[i], &port_params_default,
|
|
sizeof(struct port_params));
|
|
|
|
if (parse_args(argc, argv)) {
|
|
print_usage(argv[0]);
|
|
return -1;
|
|
}
|
|
|
|
/* Buffer pool initialization. */
|
|
bp = bpool_init(&bpool_params, &umem_cfg);
|
|
if (!bp) {
|
|
printf("Buffer pool initialization failed.\n");
|
|
return -1;
|
|
}
|
|
printf("Buffer pool created successfully.\n");
|
|
|
|
/* Ports initialization. */
|
|
for (i = 0; i < MAX_PORTS; i++)
|
|
port_params[i].bp = bp;
|
|
|
|
for (i = 0; i < n_ports; i++) {
|
|
ports[i] = port_init(&port_params[i]);
|
|
if (!ports[i]) {
|
|
printf("Port %d initialization failed.\n", i);
|
|
return -1;
|
|
}
|
|
print_port(i);
|
|
}
|
|
printf("All ports created successfully.\n");
|
|
|
|
/* Threads. */
|
|
for (i = 0; i < n_threads; i++) {
|
|
struct thread_data *t = &thread_data[i];
|
|
u32 n_ports_per_thread = n_ports / n_threads, j;
|
|
|
|
for (j = 0; j < n_ports_per_thread; j++) {
|
|
t->ports_rx[j] = ports[i * n_ports_per_thread + j];
|
|
t->ports_tx[j] = ports[i * n_ports_per_thread +
|
|
(j + 1) % n_ports_per_thread];
|
|
}
|
|
|
|
t->n_ports_rx = n_ports_per_thread;
|
|
|
|
print_thread(i);
|
|
}
|
|
|
|
for (i = 0; i < n_threads; i++) {
|
|
int status;
|
|
|
|
status = pthread_create(&threads[i],
|
|
NULL,
|
|
thread_func,
|
|
&thread_data[i]);
|
|
if (status) {
|
|
printf("Thread %d creation failed.\n", i);
|
|
return -1;
|
|
}
|
|
}
|
|
printf("All threads created successfully.\n");
|
|
|
|
/* Print statistics. */
|
|
signal(SIGINT, signal_handler);
|
|
signal(SIGTERM, signal_handler);
|
|
signal(SIGABRT, signal_handler);
|
|
|
|
clock_gettime(CLOCK_MONOTONIC, &time);
|
|
ns0 = time.tv_sec * 1000000000UL + time.tv_nsec;
|
|
for ( ; !quit; ) {
|
|
u64 ns1, ns_diff;
|
|
|
|
sleep(1);
|
|
clock_gettime(CLOCK_MONOTONIC, &time);
|
|
ns1 = time.tv_sec * 1000000000UL + time.tv_nsec;
|
|
ns_diff = ns1 - ns0;
|
|
ns0 = ns1;
|
|
|
|
print_port_stats_all(ns_diff);
|
|
}
|
|
|
|
/* Threads completion. */
|
|
printf("Quit.\n");
|
|
for (i = 0; i < n_threads; i++)
|
|
thread_data[i].quit = 1;
|
|
|
|
for (i = 0; i < n_threads; i++)
|
|
pthread_join(threads[i], NULL);
|
|
|
|
for (i = 0; i < n_ports; i++)
|
|
port_free(ports[i]);
|
|
|
|
bpool_free(bp);
|
|
|
|
remove_xdp_program();
|
|
|
|
return 0;
|
|
}
|