ipv6: udp: Optimise multicast reception
IPV6 UDP multicast rx path is a bit complex and can hold a spinlock for a long time. Using a small (32 or 64 entries) stack of socket pointers can help to perform expensive operations (skb_clone(), udp_queue_rcv_skb()) outside of the lock, in most cases. Signed-off-by: Eric Dumazet <eric.dumazet@gmail.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
1240d1373c
commit
a1ab77f97e
|
@ -569,6 +569,27 @@ static struct sock *udp_v6_mcast_next(struct net *net, struct sock *sk,
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void flush_stack(struct sock **stack, unsigned int count,
|
||||||
|
struct sk_buff *skb, unsigned int final)
|
||||||
|
{
|
||||||
|
unsigned int i;
|
||||||
|
struct sock *sk;
|
||||||
|
struct sk_buff *skb1;
|
||||||
|
|
||||||
|
for (i = 0; i < count; i++) {
|
||||||
|
skb1 = (i == final) ? skb : skb_clone(skb, GFP_ATOMIC);
|
||||||
|
|
||||||
|
if (skb1) {
|
||||||
|
sk = stack[i];
|
||||||
|
bh_lock_sock(sk);
|
||||||
|
if (!sock_owned_by_user(sk))
|
||||||
|
udpv6_queue_rcv_skb(sk, skb1);
|
||||||
|
else
|
||||||
|
sk_add_backlog(sk, skb1);
|
||||||
|
bh_unlock_sock(sk);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
/*
|
/*
|
||||||
* Note: called only from the BH handler context,
|
* Note: called only from the BH handler context,
|
||||||
* so we don't need to lock the hashes.
|
* so we don't need to lock the hashes.
|
||||||
|
@ -577,41 +598,43 @@ static int __udp6_lib_mcast_deliver(struct net *net, struct sk_buff *skb,
|
||||||
struct in6_addr *saddr, struct in6_addr *daddr,
|
struct in6_addr *saddr, struct in6_addr *daddr,
|
||||||
struct udp_table *udptable)
|
struct udp_table *udptable)
|
||||||
{
|
{
|
||||||
struct sock *sk, *sk2;
|
struct sock *sk, *stack[256 / sizeof(struct sock *)];
|
||||||
const struct udphdr *uh = udp_hdr(skb);
|
const struct udphdr *uh = udp_hdr(skb);
|
||||||
struct udp_hslot *hslot = udp_hashslot(udptable, net, ntohs(uh->dest));
|
struct udp_hslot *hslot = udp_hashslot(udptable, net, ntohs(uh->dest));
|
||||||
int dif;
|
int dif;
|
||||||
|
unsigned int i, count = 0;
|
||||||
|
|
||||||
spin_lock(&hslot->lock);
|
spin_lock(&hslot->lock);
|
||||||
sk = sk_nulls_head(&hslot->head);
|
sk = sk_nulls_head(&hslot->head);
|
||||||
dif = inet6_iif(skb);
|
dif = inet6_iif(skb);
|
||||||
sk = udp_v6_mcast_next(net, sk, uh->dest, daddr, uh->source, saddr, dif);
|
sk = udp_v6_mcast_next(net, sk, uh->dest, daddr, uh->source, saddr, dif);
|
||||||
if (!sk) {
|
while (sk) {
|
||||||
kfree_skb(skb);
|
stack[count++] = sk;
|
||||||
goto out;
|
sk = udp_v6_mcast_next(net, sk_nulls_next(sk), uh->dest, daddr,
|
||||||
}
|
uh->source, saddr, dif);
|
||||||
|
if (unlikely(count == ARRAY_SIZE(stack))) {
|
||||||
sk2 = sk;
|
if (!sk)
|
||||||
while ((sk2 = udp_v6_mcast_next(net, sk_nulls_next(sk2), uh->dest, daddr,
|
break;
|
||||||
uh->source, saddr, dif))) {
|
flush_stack(stack, count, skb, ~0);
|
||||||
struct sk_buff *buff = skb_clone(skb, GFP_ATOMIC);
|
count = 0;
|
||||||
if (buff) {
|
|
||||||
bh_lock_sock(sk2);
|
|
||||||
if (!sock_owned_by_user(sk2))
|
|
||||||
udpv6_queue_rcv_skb(sk2, buff);
|
|
||||||
else
|
|
||||||
sk_add_backlog(sk2, buff);
|
|
||||||
bh_unlock_sock(sk2);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
bh_lock_sock(sk);
|
/*
|
||||||
if (!sock_owned_by_user(sk))
|
* before releasing the lock, we must take reference on sockets
|
||||||
udpv6_queue_rcv_skb(sk, skb);
|
*/
|
||||||
else
|
for (i = 0; i < count; i++)
|
||||||
sk_add_backlog(sk, skb);
|
sock_hold(stack[i]);
|
||||||
bh_unlock_sock(sk);
|
|
||||||
out:
|
|
||||||
spin_unlock(&hslot->lock);
|
spin_unlock(&hslot->lock);
|
||||||
|
|
||||||
|
if (count) {
|
||||||
|
flush_stack(stack, count, skb, count - 1);
|
||||||
|
|
||||||
|
for (i = 0; i < count; i++)
|
||||||
|
sock_put(stack[i]);
|
||||||
|
} else {
|
||||||
|
kfree_skb(skb);
|
||||||
|
}
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue