2019-05-27 14:55:01 +08:00
|
|
|
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* INET An implementation of the TCP/IP protocol suite for the LINUX
|
|
|
|
* operating system. INET is implemented using the BSD Socket
|
|
|
|
* interface as the means of communication with the user level.
|
|
|
|
*
|
|
|
|
* Definitions for the IP router.
|
|
|
|
*
|
|
|
|
* Version: @(#)route.h 1.0.4 05/27/93
|
|
|
|
*
|
2005-05-06 07:16:16 +08:00
|
|
|
* Authors: Ross Biro
|
2005-04-17 06:20:36 +08:00
|
|
|
* Fred N. van Kempen, <waltje@uWalt.NL.Mugnet.ORG>
|
|
|
|
* Fixes:
|
|
|
|
* Alan Cox : Reformatted. Added ip_rt_local()
|
|
|
|
* Alan Cox : Support for TCP parameters.
|
|
|
|
* Alexey Kuznetsov: Major changes for new routing code.
|
|
|
|
* Mike McLagan : Routing by source
|
|
|
|
* Robert Olsson : Added rt_cache statistics
|
|
|
|
*/
|
|
|
|
#ifndef _ROUTE_H
|
|
|
|
#define _ROUTE_H
|
|
|
|
|
|
|
|
#include <net/dst.h>
|
|
|
|
#include <net/inetpeer.h>
|
|
|
|
#include <net/flow.h>
|
2008-10-01 22:35:39 +08:00
|
|
|
#include <net/inet_sock.h>
|
2015-09-30 16:12:22 +08:00
|
|
|
#include <net/ip_fib.h>
|
ipv4: Add helpers for neigh lookup for nexthop
A common theme in the output path is looking up a neigh entry for a
nexthop, either the gateway in an rtable or a fallback to the daddr
in the skb:
nexthop = (__force u32)rt_nexthop(rt, ip_hdr(skb)->daddr);
neigh = __ipv4_neigh_lookup_noref(dev, nexthop);
if (unlikely(!neigh))
neigh = __neigh_create(&arp_tbl, &nexthop, dev, false);
To allow the nexthop to be an IPv6 address we need to consider the
family of the nexthop and then call __ipv{4,6}_neigh_lookup_noref based
on it.
To make this simpler, add a ip_neigh_gw4 helper similar to ip_neigh_gw6
added in an earlier patch which handles:
neigh = __ipv4_neigh_lookup_noref(dev, nexthop);
if (unlikely(!neigh))
neigh = __neigh_create(&arp_tbl, &nexthop, dev, false);
And then add a second one, ip_neigh_for_gw, that calls either
ip_neigh_gw4 or ip_neigh_gw6 based on the address family of the gateway.
Update the output paths in the VRF driver and core v4 code to use
ip_neigh_for_gw simplifying the family based lookup and making both
ready for a v6 nexthop.
ipv4_neigh_lookup has a different need - the potential to resolve a
passed in address in addition to any gateway in the rtable or skb. Since
this is a one-off, add ip_neigh_gw4 and ip_neigh_gw6 diectly. The
difference between __neigh_create used by the helpers and neigh_create
called by ipv4_neigh_lookup is taking a refcount, so add rcu_read_lock_bh
and bump the refcnt on the neigh entry.
Signed-off-by: David Ahern <dsahern@gmail.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2019-04-06 07:30:34 +08:00
|
|
|
#include <net/arp.h>
|
|
|
|
#include <net/ndisc.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
#include <linux/in_route.h>
|
|
|
|
#include <linux/rtnetlink.h>
|
2012-07-26 19:14:38 +08:00
|
|
|
#include <linux/rcupdate.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
#include <linux/route.h>
|
|
|
|
#include <linux/ip.h>
|
|
|
|
#include <linux/cache.h>
|
2006-08-05 14:12:42 +08:00
|
|
|
#include <linux/security.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2014-01-09 17:01:15 +08:00
|
|
|
/* IPv4 datagram length is stored into 16bit field (tot_len) */
|
|
|
|
#define IP_MAX_MTU 0xFFFFU
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
#define RTO_ONLINK 0x01
|
|
|
|
|
|
|
|
#define RT_CONN_FLAGS(sk) (RT_TOS(inet_sk(sk)->tos) | sock_flag(sk, SOCK_LOCALROUTE))
|
2013-09-24 21:43:09 +08:00
|
|
|
#define RT_CONN_FLAGS_TOS(sk,tos) (RT_TOS(tos) | sock_flag(sk, SOCK_LOCALROUTE))
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
struct fib_nh;
|
net: Implement read-only protection and COW'ing of metrics.
Routing metrics are now copy-on-write.
Initially a route entry points it's metrics at a read-only location.
If a routing table entry exists, it will point there. Else it will
point at the all zero metric place-holder called 'dst_default_metrics'.
The writeability state of the metrics is stored in the low bits of the
metrics pointer, we have two bits left to spare if we want to store
more states.
For the initial implementation, COW is implemented simply via kmalloc.
However future enhancements will change this to place the writable
metrics somewhere else, in order to increase sharing. Very likely
this "somewhere else" will be the inetpeer cache.
Note also that this means that metrics updates may transiently fail
if we cannot COW the metrics successfully.
But even by itself, this patch should decrease memory usage and
increase cache locality especially for routing workloads. In those
cases the read-only metric copies stay in place and never get written
to.
TCP workloads where metrics get updated, and those rare cases where
PMTU triggers occur, will take a very slight performance hit. But
that hit will be alleviated when the long-term writable metrics
move to a more sharable location.
Since the metrics storage went from a u32 array of RTAX_MAX entries to
what is essentially a pointer, some retooling of the dst_entry layout
was necessary.
Most importantly, we need to preserve the alignment of the reference
count so that it doesn't share cache lines with the read-mostly state,
as per Eric Dumazet's alignment assertion checks.
The only non-trivial bit here is the move of the 'flags' member into
the writeable cacheline. This is OK since we are always accessing the
flags around the same moment when we made a modification to the
reference count.
Signed-off-by: David S. Miller <davem@davemloft.net>
2011-01-27 12:51:05 +08:00
|
|
|
struct fib_info;
|
2015-01-15 07:17:06 +08:00
|
|
|
struct uncached_list;
|
2009-11-03 11:26:03 +08:00
|
|
|
struct rtable {
|
2010-06-11 14:31:35 +08:00
|
|
|
struct dst_entry dst;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
[IPV4] route cache: Introduce rt_genid for smooth cache invalidation
Current ip route cache implementation is not suited to large caches.
We can consume a lot of CPU when cache must be invalidated, since we
currently need to evict all cache entries, and this eviction is
sometimes asynchronous. min_delay & max_delay can somewhat control this
asynchronism behavior, but whole thing is a kludge, regularly triggering
infamous soft lockup messages. When entries are still in use, this also
consumes a lot of ram, filling dst_garbage.list.
A better scheme is to use a generation identifier on each entry,
so that cache invalidation can be performed by changing the table
identifier, without having to scan all entries.
No more delayed flushing, no more stalling when secret_interval expires.
Invalidated entries will then be freed at GC time (controled by
ip_rt_gc_timeout or stress), or when an invalidated entry is found
in a chain when an insert is done.
Thus we keep a normal equilibrium.
This patch :
- renames rt_hash_rnd to rt_genid (and makes it an atomic_t)
- Adds a new rt_genid field to 'struct rtable' (filling a hole on 64bit)
- Checks entry->rt_genid at appropriate places :
2008-02-01 09:05:09 +08:00
|
|
|
int rt_genid;
|
2012-04-15 13:58:06 +08:00
|
|
|
unsigned int rt_flags;
|
2005-04-17 06:20:36 +08:00
|
|
|
__u16 rt_type;
|
2012-10-08 19:41:18 +08:00
|
|
|
__u8 rt_is_input;
|
2019-09-18 01:39:49 +08:00
|
|
|
__u8 rt_uses_gateway;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
int rt_iif;
|
|
|
|
|
2019-09-18 01:39:49 +08:00
|
|
|
u8 rt_gw_family;
|
2005-04-17 06:20:36 +08:00
|
|
|
/* Info on neighbour */
|
2019-04-06 07:30:29 +08:00
|
|
|
union {
|
|
|
|
__be32 rt_gw4;
|
|
|
|
struct in6_addr rt_gw6;
|
|
|
|
};
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
/* Miscellaneous cached information */
|
2018-03-14 17:21:14 +08:00
|
|
|
u32 rt_mtu_locked:1,
|
|
|
|
rt_pmtu:31;
|
2012-08-01 06:06:50 +08:00
|
|
|
|
|
|
|
struct list_head rt_uncached;
|
2015-01-15 07:17:06 +08:00
|
|
|
struct uncached_list *rt_uncached_list;
|
2005-04-17 06:20:36 +08:00
|
|
|
};
|
|
|
|
|
2011-11-23 10:14:15 +08:00
|
|
|
static inline bool rt_is_input_route(const struct rtable *rt)
|
2010-11-12 09:07:48 +08:00
|
|
|
{
|
2012-07-18 05:44:26 +08:00
|
|
|
return rt->rt_is_input != 0;
|
2010-11-12 09:07:48 +08:00
|
|
|
}
|
|
|
|
|
2011-11-23 10:14:15 +08:00
|
|
|
static inline bool rt_is_output_route(const struct rtable *rt)
|
2010-11-12 09:07:48 +08:00
|
|
|
{
|
2012-07-18 05:44:26 +08:00
|
|
|
return rt->rt_is_input == 0;
|
2010-11-12 09:07:48 +08:00
|
|
|
}
|
|
|
|
|
2012-07-13 20:03:45 +08:00
|
|
|
static inline __be32 rt_nexthop(const struct rtable *rt, __be32 daddr)
|
|
|
|
{
|
2019-04-06 07:30:27 +08:00
|
|
|
if (rt->rt_gw_family == AF_INET)
|
|
|
|
return rt->rt_gw4;
|
2012-07-13 20:03:45 +08:00
|
|
|
return daddr;
|
|
|
|
}
|
|
|
|
|
2009-11-03 11:26:03 +08:00
|
|
|
struct ip_rt_acct {
|
2005-04-17 06:20:36 +08:00
|
|
|
__u32 o_bytes;
|
|
|
|
__u32 o_packets;
|
|
|
|
__u32 i_bytes;
|
|
|
|
__u32 i_packets;
|
|
|
|
};
|
|
|
|
|
2009-11-03 11:26:03 +08:00
|
|
|
struct rt_cache_stat {
|
2005-04-17 06:20:36 +08:00
|
|
|
unsigned int in_slow_tot;
|
|
|
|
unsigned int in_slow_mc;
|
|
|
|
unsigned int in_no_route;
|
|
|
|
unsigned int in_brd;
|
|
|
|
unsigned int in_martian_dst;
|
|
|
|
unsigned int in_martian_src;
|
|
|
|
unsigned int out_slow_tot;
|
|
|
|
unsigned int out_slow_mc;
|
|
|
|
};
|
|
|
|
|
2010-02-16 23:20:26 +08:00
|
|
|
extern struct ip_rt_acct __percpu *ip_rt_acct;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
struct in_device;
|
2013-09-23 01:32:22 +08:00
|
|
|
|
|
|
|
int ip_rt_init(void);
|
|
|
|
void rt_cache_flush(struct net *net);
|
|
|
|
void rt_flush_dev(struct net_device *dev);
|
2017-05-26 01:42:33 +08:00
|
|
|
struct rtable *ip_route_output_key_hash(struct net *net, struct flowi4 *flp,
|
|
|
|
const struct sk_buff *skb);
|
|
|
|
struct rtable *ip_route_output_key_hash_rcu(struct net *net, struct flowi4 *flp,
|
|
|
|
struct fib_result *res,
|
|
|
|
const struct sk_buff *skb);
|
2015-09-30 16:12:22 +08:00
|
|
|
|
|
|
|
static inline struct rtable *__ip_route_output_key(struct net *net,
|
|
|
|
struct flowi4 *flp)
|
|
|
|
{
|
2017-05-26 01:42:33 +08:00
|
|
|
return ip_route_output_key_hash(net, flp, NULL);
|
2015-09-30 16:12:22 +08:00
|
|
|
}
|
|
|
|
|
2013-09-23 01:32:22 +08:00
|
|
|
struct rtable *ip_route_output_flow(struct net *, struct flowi4 *flp,
|
2015-09-25 22:39:10 +08:00
|
|
|
const struct sock *sk);
|
2013-09-23 01:32:22 +08:00
|
|
|
struct dst_entry *ipv4_blackhole_route(struct net *net,
|
|
|
|
struct dst_entry *dst_orig);
|
2010-05-10 19:32:55 +08:00
|
|
|
|
2011-03-12 14:12:47 +08:00
|
|
|
static inline struct rtable *ip_route_output_key(struct net *net, struct flowi4 *flp)
|
2011-03-03 06:56:30 +08:00
|
|
|
{
|
|
|
|
return ip_route_output_flow(net, flp, NULL);
|
|
|
|
}
|
|
|
|
|
2011-03-12 13:00:52 +08:00
|
|
|
static inline struct rtable *ip_route_output(struct net *net, __be32 daddr,
|
|
|
|
__be32 saddr, u8 tos, int oif)
|
|
|
|
{
|
2011-03-12 14:12:47 +08:00
|
|
|
struct flowi4 fl4 = {
|
|
|
|
.flowi4_oif = oif,
|
2012-06-11 04:05:24 +08:00
|
|
|
.flowi4_tos = tos,
|
2011-03-12 14:12:47 +08:00
|
|
|
.daddr = daddr,
|
|
|
|
.saddr = saddr,
|
2011-03-12 13:00:52 +08:00
|
|
|
};
|
2011-03-12 14:12:47 +08:00
|
|
|
return ip_route_output_key(net, &fl4);
|
2011-03-12 13:00:52 +08:00
|
|
|
}
|
|
|
|
|
2011-05-04 11:25:42 +08:00
|
|
|
static inline struct rtable *ip_route_output_ports(struct net *net, struct flowi4 *fl4,
|
|
|
|
struct sock *sk,
|
2011-03-12 13:00:52 +08:00
|
|
|
__be32 daddr, __be32 saddr,
|
|
|
|
__be16 dport, __be16 sport,
|
|
|
|
__u8 proto, __u8 tos, int oif)
|
|
|
|
{
|
2011-05-04 11:25:42 +08:00
|
|
|
flowi4_init_output(fl4, oif, sk ? sk->sk_mark : 0, tos,
|
2011-03-31 19:52:59 +08:00
|
|
|
RT_SCOPE_UNIVERSE, proto,
|
|
|
|
sk ? inet_sk_flowi_flags(sk) : 0,
|
2016-11-04 01:23:43 +08:00
|
|
|
daddr, saddr, dport, sport, sock_net_uid(net, sk));
|
2011-03-12 13:00:52 +08:00
|
|
|
if (sk)
|
2011-05-04 11:25:42 +08:00
|
|
|
security_sk_classify_flow(sk, flowi4_to_flowi(fl4));
|
|
|
|
return ip_route_output_flow(net, fl4, sk);
|
2011-03-12 13:00:52 +08:00
|
|
|
}
|
|
|
|
|
2011-05-05 03:33:34 +08:00
|
|
|
static inline struct rtable *ip_route_output_gre(struct net *net, struct flowi4 *fl4,
|
2011-03-12 13:00:52 +08:00
|
|
|
__be32 daddr, __be32 saddr,
|
|
|
|
__be32 gre_key, __u8 tos, int oif)
|
|
|
|
{
|
2011-05-05 03:33:34 +08:00
|
|
|
memset(fl4, 0, sizeof(*fl4));
|
|
|
|
fl4->flowi4_oif = oif;
|
|
|
|
fl4->daddr = daddr;
|
|
|
|
fl4->saddr = saddr;
|
|
|
|
fl4->flowi4_tos = tos;
|
|
|
|
fl4->flowi4_proto = IPPROTO_GRE;
|
|
|
|
fl4->fl4_gre_key = gre_key;
|
|
|
|
return ip_route_output_key(net, fl4);
|
2011-03-12 13:00:52 +08:00
|
|
|
}
|
2017-09-28 21:51:37 +08:00
|
|
|
int ip_mc_validate_source(struct sk_buff *skb, __be32 daddr, __be32 saddr,
|
|
|
|
u8 tos, struct net_device *dev,
|
|
|
|
struct in_device *in_dev, u32 *itag);
|
2013-09-23 01:32:22 +08:00
|
|
|
int ip_route_input_noref(struct sk_buff *skb, __be32 dst, __be32 src,
|
|
|
|
u8 tos, struct net_device *devin);
|
2017-05-26 01:42:34 +08:00
|
|
|
int ip_route_input_rcu(struct sk_buff *skb, __be32 dst, __be32 src,
|
|
|
|
u8 tos, struct net_device *devin,
|
|
|
|
struct fib_result *res);
|
2012-07-26 19:14:38 +08:00
|
|
|
|
|
|
|
static inline int ip_route_input(struct sk_buff *skb, __be32 dst, __be32 src,
|
|
|
|
u8 tos, struct net_device *devin)
|
|
|
|
{
|
|
|
|
int err;
|
|
|
|
|
|
|
|
rcu_read_lock();
|
|
|
|
err = ip_route_input_noref(skb, dst, src, tos, devin);
|
2017-09-01 00:11:41 +08:00
|
|
|
if (!err) {
|
2017-09-22 00:15:46 +08:00
|
|
|
skb_dst_force(skb);
|
2017-09-01 00:11:41 +08:00
|
|
|
if (!skb_dst(skb))
|
|
|
|
err = -EINVAL;
|
|
|
|
}
|
2012-07-26 19:14:38 +08:00
|
|
|
rcu_read_unlock();
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
2010-05-10 19:32:55 +08:00
|
|
|
|
2013-09-23 01:32:22 +08:00
|
|
|
void ipv4_update_pmtu(struct sk_buff *skb, struct net *net, u32 mtu, int oif,
|
2018-09-26 11:56:26 +08:00
|
|
|
u8 protocol);
|
2013-09-23 01:32:22 +08:00
|
|
|
void ipv4_sk_update_pmtu(struct sk_buff *skb, struct sock *sk, u32 mtu);
|
2018-09-26 11:56:27 +08:00
|
|
|
void ipv4_redirect(struct sk_buff *skb, struct net *net, int oif, u8 protocol);
|
2013-09-23 01:32:22 +08:00
|
|
|
void ipv4_sk_redirect(struct sk_buff *skb, struct sock *sk);
|
|
|
|
void ip_rt_send_redirect(struct sk_buff *skb);
|
|
|
|
|
|
|
|
unsigned int inet_addr_type(struct net *net, __be32 addr);
|
2015-09-02 04:26:35 +08:00
|
|
|
unsigned int inet_addr_type_table(struct net *net, __be32 addr, u32 tb_id);
|
2013-09-23 01:32:22 +08:00
|
|
|
unsigned int inet_dev_addr_type(struct net *net, const struct net_device *dev,
|
|
|
|
__be32 addr);
|
2015-08-14 04:59:05 +08:00
|
|
|
unsigned int inet_addr_type_dev_table(struct net *net,
|
|
|
|
const struct net_device *dev,
|
|
|
|
__be32 addr);
|
2013-09-23 01:32:22 +08:00
|
|
|
void ip_rt_multicast_event(struct in_device *);
|
2017-07-01 20:03:10 +08:00
|
|
|
int ip_rt_ioctl(struct net *, unsigned int cmd, struct rtentry *rt);
|
2013-09-23 01:32:22 +08:00
|
|
|
void ip_rt_get_source(u8 *src, struct sk_buff *skb, struct rtable *rt);
|
2016-04-08 02:10:06 +08:00
|
|
|
struct rtable *rt_dst_alloc(struct net_device *dev,
|
|
|
|
unsigned int flags, u16 type,
|
|
|
|
bool nopolicy, bool noxfrm, bool will_cache);
|
2019-06-26 14:21:16 +08:00
|
|
|
struct rtable *rt_dst_clone(struct net_device *dev, struct rtable *rt);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2005-11-23 06:47:37 +08:00
|
|
|
struct in_ifaddr;
|
2013-09-23 01:32:22 +08:00
|
|
|
void fib_add_ifaddr(struct in_ifaddr *);
|
|
|
|
void fib_del_ifaddr(struct in_ifaddr *, struct in_ifaddr *);
|
2018-05-27 23:09:57 +08:00
|
|
|
void fib_modify_prefix_metric(struct in_ifaddr *ifa, u32 new_metric);
|
2005-11-23 06:47:37 +08:00
|
|
|
|
xfrm: reuse uncached_list to track xdsts
In early time, when freeing a xdst, it would be inserted into
dst_garbage.list first. Then if it's refcnt was still held
somewhere, later it would be put into dst_busy_list in
dst_gc_task().
When one dev was being unregistered, the dev of these dsts in
dst_busy_list would be set with loopback_dev and put this dev.
So that this dev's removal wouldn't get blocked, and avoid the
kmsg warning:
kernel:unregister_netdevice: waiting for veth0 to become \
free. Usage count = 2
However after Commit 52df157f17e5 ("xfrm: take refcnt of dst
when creating struct xfrm_dst bundle"), the xdst will not be
freed with dst gc, and this warning happens.
To fix it, we need to find these xdsts that are still held by
others when removing the dev, and free xdst's dev and set it
with loopback_dev.
But unfortunately after flow_cache for xfrm was deleted, no
list tracks them anymore. So we need to save these xdsts
somewhere to release the xdst's dev later.
To make this easier, this patch is to reuse uncached_list to
track xdsts, so that the dev refcnt can be released in the
event NETDEV_UNREGISTER process of fib_netdev_notifier.
Thanks to Florian, we could move forward this fix quickly.
Fixes: 52df157f17e5 ("xfrm: take refcnt of dst when creating struct xfrm_dst bundle")
Reported-by: Jianlin Shi <jishi@redhat.com>
Reported-by: Hangbin Liu <liuhangbin@gmail.com>
Tested-by: Eyal Birger <eyal.birger@gmail.com>
Signed-off-by: Xin Long <lucien.xin@gmail.com>
Signed-off-by: Steffen Klassert <steffen.klassert@secunet.com>
2018-02-14 19:06:02 +08:00
|
|
|
void rt_add_uncached_list(struct rtable *rt);
|
|
|
|
void rt_del_uncached_list(struct rtable *rt);
|
|
|
|
|
ipv4: Dump route exceptions if requested
Since commit 4895c771c7f0 ("ipv4: Add FIB nexthop exceptions."), cached
exception routes are stored as a separate entity, so they are not dumped
on a FIB dump, even if the RTM_F_CLONED flag is passed.
This implies that the command 'ip route list cache' doesn't return any
result anymore.
If the RTM_F_CLONED is passed, and strict checking requested, retrieve
nexthop exception routes and dump them. If no strict checking is
requested, filtering can't be performed consistently: dump everything in
that case.
With this, we need to add an argument to the netlink callback in order to
track how many entries were already dumped for the last leaf included in
a partial netlink dump.
A single additional argument is sufficient, even if we traverse logically
nested structures (nexthop objects, hash table buckets, bucket chains): it
doesn't matter if we stop in the middle of any of those, because they are
always traversed the same way. As an example, s_i values in [], s_fa
values in ():
node (fa) #1 [1]
nexthop #1
bucket #1 -> #0 in chain (1)
bucket #2 -> #0 in chain (2) -> #1 in chain (3) -> #2 in chain (4)
bucket #3 -> #0 in chain (5) -> #1 in chain (6)
nexthop #2
bucket #1 -> #0 in chain (7) -> #1 in chain (8)
bucket #2 -> #0 in chain (9)
--
node (fa) #2 [2]
nexthop #1
bucket #1 -> #0 in chain (1) -> #1 in chain (2)
bucket #2 -> #0 in chain (3)
it doesn't matter if we stop at (3), (4), (7) for "node #1", or at (2)
for "node #2": walking flattens all that.
It would even be possible to drop the distinction between the in-tree
(s_i) and in-node (s_fa) counter, but a further improvement might
advise against this. This is only as accurate as the existing tracking
mechanism for leaves: if a partial dump is restarted after exceptions
are removed or expired, we might skip some non-dumped entries.
To improve this, we could attach a 'sernum' attribute (similar to the
one used for IPv6) to nexthop entities, and bump this counter whenever
exceptions change: having a distinction between the two counters would
make this more convenient.
Listing of exception routes (modified routes pre-3.5) was tested against
these versions of kernel and iproute2:
iproute2
kernel 4.14.0 4.15.0 4.19.0 5.0.0 5.1.0
3.5-rc4 + + + + +
4.4
4.9
4.14
4.15
4.19
5.0
5.1
fixed + + + + +
v7:
- Move loop over nexthop objects to route.c, and pass struct fib_info
and table ID to it, not a struct fib_alias (suggested by David Ahern)
- While at it, note that the NULL check on fa->fa_info is redundant,
and the check on RTNH_F_DEAD is also not consistent with what's done
with regular route listing: just keep it for nhc_flags
- Rename entry point function for dumping exceptions to
fib_dump_info_fnhe(), and rearrange arguments for consistency with
fib_dump_info()
- Rename fnhe_dump_buckets() to fnhe_dump_bucket() and make it handle
one bucket at a time
- Expand commit message to describe why we can have a single "skip"
counter for all exceptions stored in bucket chains in nexthop objects
(suggested by David Ahern)
v6:
- Rebased onto net-next
- Loop over nexthop paths too. Move loop over fnhe buckets to route.c,
avoids need to export rt_fill_info() and to touch exceptions from
fib_trie.c. Pass NULL as flow to rt_fill_info(), it now allows that
(suggested by David Ahern)
Fixes: 4895c771c7f0 ("ipv4: Add FIB nexthop exceptions.")
Signed-off-by: Stefano Brivio <sbrivio@redhat.com>
Reviewed-by: David Ahern <dsahern@gmail.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2019-06-21 23:45:23 +08:00
|
|
|
int fib_dump_info_fnhe(struct sk_buff *skb, struct netlink_callback *cb,
|
|
|
|
u32 table_id, struct fib_info *fi,
|
2019-08-24 08:11:38 +08:00
|
|
|
int *fa_index, int fa_start, unsigned int flags);
|
ipv4: Dump route exceptions if requested
Since commit 4895c771c7f0 ("ipv4: Add FIB nexthop exceptions."), cached
exception routes are stored as a separate entity, so they are not dumped
on a FIB dump, even if the RTM_F_CLONED flag is passed.
This implies that the command 'ip route list cache' doesn't return any
result anymore.
If the RTM_F_CLONED is passed, and strict checking requested, retrieve
nexthop exception routes and dump them. If no strict checking is
requested, filtering can't be performed consistently: dump everything in
that case.
With this, we need to add an argument to the netlink callback in order to
track how many entries were already dumped for the last leaf included in
a partial netlink dump.
A single additional argument is sufficient, even if we traverse logically
nested structures (nexthop objects, hash table buckets, bucket chains): it
doesn't matter if we stop in the middle of any of those, because they are
always traversed the same way. As an example, s_i values in [], s_fa
values in ():
node (fa) #1 [1]
nexthop #1
bucket #1 -> #0 in chain (1)
bucket #2 -> #0 in chain (2) -> #1 in chain (3) -> #2 in chain (4)
bucket #3 -> #0 in chain (5) -> #1 in chain (6)
nexthop #2
bucket #1 -> #0 in chain (7) -> #1 in chain (8)
bucket #2 -> #0 in chain (9)
--
node (fa) #2 [2]
nexthop #1
bucket #1 -> #0 in chain (1) -> #1 in chain (2)
bucket #2 -> #0 in chain (3)
it doesn't matter if we stop at (3), (4), (7) for "node #1", or at (2)
for "node #2": walking flattens all that.
It would even be possible to drop the distinction between the in-tree
(s_i) and in-node (s_fa) counter, but a further improvement might
advise against this. This is only as accurate as the existing tracking
mechanism for leaves: if a partial dump is restarted after exceptions
are removed or expired, we might skip some non-dumped entries.
To improve this, we could attach a 'sernum' attribute (similar to the
one used for IPv6) to nexthop entities, and bump this counter whenever
exceptions change: having a distinction between the two counters would
make this more convenient.
Listing of exception routes (modified routes pre-3.5) was tested against
these versions of kernel and iproute2:
iproute2
kernel 4.14.0 4.15.0 4.19.0 5.0.0 5.1.0
3.5-rc4 + + + + +
4.4
4.9
4.14
4.15
4.19
5.0
5.1
fixed + + + + +
v7:
- Move loop over nexthop objects to route.c, and pass struct fib_info
and table ID to it, not a struct fib_alias (suggested by David Ahern)
- While at it, note that the NULL check on fa->fa_info is redundant,
and the check on RTNH_F_DEAD is also not consistent with what's done
with regular route listing: just keep it for nhc_flags
- Rename entry point function for dumping exceptions to
fib_dump_info_fnhe(), and rearrange arguments for consistency with
fib_dump_info()
- Rename fnhe_dump_buckets() to fnhe_dump_bucket() and make it handle
one bucket at a time
- Expand commit message to describe why we can have a single "skip"
counter for all exceptions stored in bucket chains in nexthop objects
(suggested by David Ahern)
v6:
- Rebased onto net-next
- Loop over nexthop paths too. Move loop over fnhe buckets to route.c,
avoids need to export rt_fill_info() and to touch exceptions from
fib_trie.c. Pass NULL as flow to rt_fill_info(), it now allows that
(suggested by David Ahern)
Fixes: 4895c771c7f0 ("ipv4: Add FIB nexthop exceptions.")
Signed-off-by: Stefano Brivio <sbrivio@redhat.com>
Reviewed-by: David Ahern <dsahern@gmail.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2019-06-21 23:45:23 +08:00
|
|
|
|
2012-10-29 06:33:23 +08:00
|
|
|
static inline void ip_rt_put(struct rtable *rt)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2012-10-29 06:33:23 +08:00
|
|
|
/* dst_release() accepts a NULL parameter.
|
|
|
|
* We rely on dst being first structure in struct rtable
|
|
|
|
*/
|
|
|
|
BUILD_BUG_ON(offsetof(struct rtable, dst) != 0);
|
|
|
|
dst_release(&rt->dst);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
#define IPTOS_RT_MASK (IPTOS_TOS_MASK & ~3)
|
|
|
|
|
2007-07-10 06:32:57 +08:00
|
|
|
extern const __u8 ip_tos2prio[16];
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
static inline char rt_tos2priority(u8 tos)
|
|
|
|
{
|
|
|
|
return ip_tos2prio[IPTOS_TOS(tos)>>1];
|
|
|
|
}
|
|
|
|
|
2011-04-27 04:28:44 +08:00
|
|
|
/* ip_route_connect() and ip_route_newports() work in tandem whilst
|
|
|
|
* binding a socket for a new outgoing connection.
|
|
|
|
*
|
|
|
|
* In order to use IPSEC properly, we must, in the end, have a
|
|
|
|
* route that was looked up using all available keys including source
|
|
|
|
* and destination ports.
|
|
|
|
*
|
|
|
|
* However, if a source port needs to be allocated (the user specified
|
|
|
|
* a wildcard source port) we need to obtain addressing information
|
|
|
|
* in order to perform that allocation.
|
|
|
|
*
|
|
|
|
* So ip_route_connect() looks up a route using wildcarded source and
|
|
|
|
* destination ports in the key, simply so that we can get a pair of
|
|
|
|
* addresses to use for port allocation.
|
|
|
|
*
|
|
|
|
* Later, once the ports are allocated, ip_route_newports() will make
|
|
|
|
* another route lookup if needed to make sure we catch any IPSEC
|
|
|
|
* rules keyed on the port information.
|
|
|
|
*
|
|
|
|
* The callers allocate the flow key on their stack, and must pass in
|
|
|
|
* the same flowi4 object to both the ip_route_connect() and the
|
|
|
|
* ip_route_newports() calls.
|
|
|
|
*/
|
|
|
|
|
|
|
|
static inline void ip_route_connect_init(struct flowi4 *fl4, __be32 dst, __be32 src,
|
|
|
|
u32 tos, int oif, u8 protocol,
|
|
|
|
__be16 sport, __be16 dport,
|
2013-08-28 14:04:14 +08:00
|
|
|
struct sock *sk)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-04-27 04:28:44 +08:00
|
|
|
__u8 flow_flags = 0;
|
2008-10-01 22:35:39 +08:00
|
|
|
|
|
|
|
if (inet_sk(sk)->transparent)
|
2011-03-31 19:52:59 +08:00
|
|
|
flow_flags |= FLOWI_FLAG_ANYSRC;
|
|
|
|
|
2011-04-27 04:28:44 +08:00
|
|
|
flowi4_init_output(fl4, oif, sk->sk_mark, tos, RT_SCOPE_UNIVERSE,
|
2016-11-04 01:23:43 +08:00
|
|
|
protocol, flow_flags, dst, src, dport, sport,
|
|
|
|
sk->sk_uid);
|
2011-04-27 04:28:44 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct rtable *ip_route_connect(struct flowi4 *fl4,
|
|
|
|
__be32 dst, __be32 src, u32 tos,
|
|
|
|
int oif, u8 protocol,
|
|
|
|
__be16 sport, __be16 dport,
|
2013-08-28 14:04:14 +08:00
|
|
|
struct sock *sk)
|
2011-04-27 04:28:44 +08:00
|
|
|
{
|
|
|
|
struct net *net = sock_net(sk);
|
|
|
|
struct rtable *rt;
|
|
|
|
|
|
|
|
ip_route_connect_init(fl4, dst, src, tos, oif, protocol,
|
2013-08-28 14:04:14 +08:00
|
|
|
sport, dport, sk);
|
2008-10-01 22:35:39 +08:00
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
if (!dst || !src) {
|
2011-04-27 04:28:44 +08:00
|
|
|
rt = __ip_route_output_key(net, fl4);
|
2011-03-03 06:31:35 +08:00
|
|
|
if (IS_ERR(rt))
|
|
|
|
return rt;
|
|
|
|
ip_rt_put(rt);
|
2012-02-04 21:04:46 +08:00
|
|
|
flowi4_update_output(fl4, oif, tos, fl4->daddr, fl4->saddr);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2011-04-27 04:28:44 +08:00
|
|
|
security_sk_classify_flow(sk, flowi4_to_flowi(fl4));
|
|
|
|
return ip_route_output_flow(net, fl4, sk);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-04-27 04:28:44 +08:00
|
|
|
static inline struct rtable *ip_route_newports(struct flowi4 *fl4, struct rtable *rt,
|
|
|
|
__be16 orig_sport, __be16 orig_dport,
|
|
|
|
__be16 sport, __be16 dport,
|
|
|
|
struct sock *sk)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-02-25 05:38:12 +08:00
|
|
|
if (sport != orig_sport || dport != orig_dport) {
|
2011-04-27 04:28:44 +08:00
|
|
|
fl4->fl4_dport = dport;
|
|
|
|
fl4->fl4_sport = sport;
|
2011-03-03 06:31:35 +08:00
|
|
|
ip_rt_put(rt);
|
2012-02-04 21:04:46 +08:00
|
|
|
flowi4_update_output(fl4, sk->sk_bound_dev_if,
|
|
|
|
RT_CONN_FLAGS(sk), fl4->daddr,
|
|
|
|
fl4->saddr);
|
2011-04-27 04:28:44 +08:00
|
|
|
security_sk_classify_flow(sk, flowi4_to_flowi(fl4));
|
|
|
|
return ip_route_output_flow(sock_net(sk), fl4, sk);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2011-03-03 06:31:35 +08:00
|
|
|
return rt;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2008-10-01 22:33:10 +08:00
|
|
|
static inline int inet_iif(const struct sk_buff *skb)
|
|
|
|
{
|
2016-04-05 23:22:49 +08:00
|
|
|
struct rtable *rt = skb_rtable(skb);
|
|
|
|
|
|
|
|
if (rt && rt->rt_iif)
|
|
|
|
return rt->rt_iif;
|
2012-07-24 04:57:45 +08:00
|
|
|
|
|
|
|
return skb->skb_iif;
|
2008-10-01 22:33:10 +08:00
|
|
|
}
|
|
|
|
|
2010-12-13 13:55:08 +08:00
|
|
|
static inline int ip4_dst_hoplimit(const struct dst_entry *dst)
|
|
|
|
{
|
|
|
|
int hoplimit = dst_metric_raw(dst, RTAX_HOPLIMIT);
|
2016-02-15 18:11:27 +08:00
|
|
|
struct net *net = dev_net(dst->dev);
|
2010-12-13 13:55:08 +08:00
|
|
|
|
|
|
|
if (hoplimit == 0)
|
2016-02-15 18:11:27 +08:00
|
|
|
hoplimit = net->ipv4.sysctl_ip_default_ttl;
|
2010-12-13 13:55:08 +08:00
|
|
|
return hoplimit;
|
|
|
|
}
|
|
|
|
|
ipv4: Add helpers for neigh lookup for nexthop
A common theme in the output path is looking up a neigh entry for a
nexthop, either the gateway in an rtable or a fallback to the daddr
in the skb:
nexthop = (__force u32)rt_nexthop(rt, ip_hdr(skb)->daddr);
neigh = __ipv4_neigh_lookup_noref(dev, nexthop);
if (unlikely(!neigh))
neigh = __neigh_create(&arp_tbl, &nexthop, dev, false);
To allow the nexthop to be an IPv6 address we need to consider the
family of the nexthop and then call __ipv{4,6}_neigh_lookup_noref based
on it.
To make this simpler, add a ip_neigh_gw4 helper similar to ip_neigh_gw6
added in an earlier patch which handles:
neigh = __ipv4_neigh_lookup_noref(dev, nexthop);
if (unlikely(!neigh))
neigh = __neigh_create(&arp_tbl, &nexthop, dev, false);
And then add a second one, ip_neigh_for_gw, that calls either
ip_neigh_gw4 or ip_neigh_gw6 based on the address family of the gateway.
Update the output paths in the VRF driver and core v4 code to use
ip_neigh_for_gw simplifying the family based lookup and making both
ready for a v6 nexthop.
ipv4_neigh_lookup has a different need - the potential to resolve a
passed in address in addition to any gateway in the rtable or skb. Since
this is a one-off, add ip_neigh_gw4 and ip_neigh_gw6 diectly. The
difference between __neigh_create used by the helpers and neigh_create
called by ipv4_neigh_lookup is taking a refcount, so add rcu_read_lock_bh
and bump the refcnt on the neigh entry.
Signed-off-by: David Ahern <dsahern@gmail.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2019-04-06 07:30:34 +08:00
|
|
|
static inline struct neighbour *ip_neigh_gw4(struct net_device *dev,
|
|
|
|
__be32 daddr)
|
|
|
|
{
|
|
|
|
struct neighbour *neigh;
|
|
|
|
|
|
|
|
neigh = __ipv4_neigh_lookup_noref(dev, daddr);
|
|
|
|
if (unlikely(!neigh))
|
|
|
|
neigh = __neigh_create(&arp_tbl, &daddr, dev, false);
|
|
|
|
|
|
|
|
return neigh;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct neighbour *ip_neigh_for_gw(struct rtable *rt,
|
|
|
|
struct sk_buff *skb,
|
|
|
|
bool *is_v6gw)
|
|
|
|
{
|
|
|
|
struct net_device *dev = rt->dst.dev;
|
|
|
|
struct neighbour *neigh;
|
|
|
|
|
|
|
|
if (likely(rt->rt_gw_family == AF_INET)) {
|
|
|
|
neigh = ip_neigh_gw4(dev, rt->rt_gw4);
|
|
|
|
} else if (rt->rt_gw_family == AF_INET6) {
|
|
|
|
neigh = ip_neigh_gw6(dev, &rt->rt_gw6);
|
|
|
|
*is_v6gw = true;
|
|
|
|
} else {
|
|
|
|
neigh = ip_neigh_gw4(dev, ip_hdr(skb)->daddr);
|
|
|
|
}
|
|
|
|
return neigh;
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
#endif /* _ROUTE_H */
|