Merge git://git.kernel.org/pub/scm/linux/kernel/git/davem/net
Conflicts: net/core/rtnetlink.c net/core/skbuff.c Both conflicts were very simple overlapping changes. Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
commit
902455e007
|
@ -83,9 +83,9 @@ static void bpf_flush_icache(void *start_, void *end_)
|
|||
#define BNE (F2(0, 2) | CONDNE)
|
||||
|
||||
#ifdef CONFIG_SPARC64
|
||||
#define BNE_PTR (F2(0, 1) | CONDNE | (2 << 20))
|
||||
#define BE_PTR (F2(0, 1) | CONDE | (2 << 20))
|
||||
#else
|
||||
#define BNE_PTR BNE
|
||||
#define BE_PTR BE
|
||||
#endif
|
||||
|
||||
#define SETHI(K, REG) \
|
||||
|
@ -592,7 +592,7 @@ void bpf_jit_compile(struct sk_filter *fp)
|
|||
case BPF_ANC | SKF_AD_IFINDEX:
|
||||
emit_skb_loadptr(dev, r_A);
|
||||
emit_cmpi(r_A, 0);
|
||||
emit_branch(BNE_PTR, cleanup_addr + 4);
|
||||
emit_branch(BE_PTR, cleanup_addr + 4);
|
||||
emit_nop();
|
||||
emit_load32(r_A, struct net_device, ifindex, r_A);
|
||||
break;
|
||||
|
@ -605,7 +605,7 @@ void bpf_jit_compile(struct sk_filter *fp)
|
|||
case BPF_ANC | SKF_AD_HATYPE:
|
||||
emit_skb_loadptr(dev, r_A);
|
||||
emit_cmpi(r_A, 0);
|
||||
emit_branch(BNE_PTR, cleanup_addr + 4);
|
||||
emit_branch(BE_PTR, cleanup_addr + 4);
|
||||
emit_nop();
|
||||
emit_load16(r_A, struct net_device, type, r_A);
|
||||
break;
|
||||
|
|
|
@ -66,10 +66,17 @@
|
|||
#define EFX_USE_QWORD_IO 1
|
||||
#endif
|
||||
|
||||
/* Hardware issue requires that only 64-bit naturally aligned writes
|
||||
* are seen by hardware. Its not strictly necessary to restrict to
|
||||
* x86_64 arch, but done for safety since unusual write combining behaviour
|
||||
* can break PIO.
|
||||
*/
|
||||
#ifdef CONFIG_X86_64
|
||||
/* PIO is a win only if write-combining is possible */
|
||||
#ifdef ARCH_HAS_IOREMAP_WC
|
||||
#define EFX_USE_PIO 1
|
||||
#endif
|
||||
#endif
|
||||
|
||||
#ifdef EFX_USE_QWORD_IO
|
||||
static inline void _efx_writeq(struct efx_nic *efx, __le64 value,
|
||||
|
|
|
@ -189,6 +189,18 @@ struct efx_short_copy_buffer {
|
|||
u8 buf[L1_CACHE_BYTES];
|
||||
};
|
||||
|
||||
/* Copy in explicit 64-bit writes. */
|
||||
static void efx_memcpy_64(void __iomem *dest, void *src, size_t len)
|
||||
{
|
||||
u64 *src64 = src;
|
||||
u64 __iomem *dest64 = dest;
|
||||
size_t l64 = len / 8;
|
||||
size_t i;
|
||||
|
||||
for (i = 0; i < l64; i++)
|
||||
writeq(src64[i], &dest64[i]);
|
||||
}
|
||||
|
||||
/* Copy to PIO, respecting that writes to PIO buffers must be dword aligned.
|
||||
* Advances piobuf pointer. Leaves additional data in the copy buffer.
|
||||
*/
|
||||
|
@ -198,7 +210,7 @@ static void efx_memcpy_toio_aligned(struct efx_nic *efx, u8 __iomem **piobuf,
|
|||
{
|
||||
int block_len = len & ~(sizeof(copy_buf->buf) - 1);
|
||||
|
||||
memcpy_toio(*piobuf, data, block_len);
|
||||
efx_memcpy_64(*piobuf, data, block_len);
|
||||
*piobuf += block_len;
|
||||
len -= block_len;
|
||||
|
||||
|
@ -230,7 +242,7 @@ static void efx_memcpy_toio_aligned_cb(struct efx_nic *efx, u8 __iomem **piobuf,
|
|||
if (copy_buf->used < sizeof(copy_buf->buf))
|
||||
return;
|
||||
|
||||
memcpy_toio(*piobuf, copy_buf->buf, sizeof(copy_buf->buf));
|
||||
efx_memcpy_64(*piobuf, copy_buf->buf, sizeof(copy_buf->buf));
|
||||
*piobuf += sizeof(copy_buf->buf);
|
||||
data += copy_to_buf;
|
||||
len -= copy_to_buf;
|
||||
|
@ -245,7 +257,7 @@ static void efx_flush_copy_buffer(struct efx_nic *efx, u8 __iomem *piobuf,
|
|||
{
|
||||
/* if there's anything in it, write the whole buffer, including junk */
|
||||
if (copy_buf->used)
|
||||
memcpy_toio(piobuf, copy_buf->buf, sizeof(copy_buf->buf));
|
||||
efx_memcpy_64(piobuf, copy_buf->buf, sizeof(copy_buf->buf));
|
||||
}
|
||||
|
||||
/* Traverse skb structure and copy fragments in to PIO buffer.
|
||||
|
@ -304,8 +316,8 @@ efx_enqueue_skb_pio(struct efx_tx_queue *tx_queue, struct sk_buff *skb)
|
|||
*/
|
||||
BUILD_BUG_ON(L1_CACHE_BYTES >
|
||||
SKB_DATA_ALIGN(sizeof(struct skb_shared_info)));
|
||||
memcpy_toio(tx_queue->piobuf, skb->data,
|
||||
ALIGN(skb->len, L1_CACHE_BYTES));
|
||||
efx_memcpy_64(tx_queue->piobuf, skb->data,
|
||||
ALIGN(skb->len, L1_CACHE_BYTES));
|
||||
}
|
||||
|
||||
EFX_POPULATE_QWORD_5(buffer->option,
|
||||
|
|
|
@ -1204,7 +1204,6 @@ static int macvlan_device_event(struct notifier_block *unused,
|
|||
list_for_each_entry_safe(vlan, next, &port->vlans, list)
|
||||
vlan->dev->rtnl_link_ops->dellink(vlan->dev, &list_kill);
|
||||
unregister_netdevice_many(&list_kill);
|
||||
list_del(&list_kill);
|
||||
break;
|
||||
case NETDEV_PRE_TYPE_CHANGE:
|
||||
/* Forbid underlaying device to change its type. */
|
||||
|
|
|
@ -763,7 +763,12 @@ static const struct usb_device_id products[] = {
|
|||
{QMI_FIXED_INTF(0x2357, 0x9000, 4)}, /* TP-LINK MA260 */
|
||||
{QMI_FIXED_INTF(0x1bc7, 0x1200, 5)}, /* Telit LE920 */
|
||||
{QMI_FIXED_INTF(0x1bc7, 0x1201, 2)}, /* Telit LE920 */
|
||||
{QMI_FIXED_INTF(0x0b3c, 0xc005, 6)}, /* Olivetti Olicard 200 */
|
||||
{QMI_FIXED_INTF(0x0b3c, 0xc000, 4)}, /* Olivetti Olicard 100 */
|
||||
{QMI_FIXED_INTF(0x0b3c, 0xc001, 4)}, /* Olivetti Olicard 120 */
|
||||
{QMI_FIXED_INTF(0x0b3c, 0xc002, 4)}, /* Olivetti Olicard 140 */
|
||||
{QMI_FIXED_INTF(0x0b3c, 0xc004, 6)}, /* Olivetti Olicard 155 */
|
||||
{QMI_FIXED_INTF(0x0b3c, 0xc005, 6)}, /* Olivetti Olicard 200 */
|
||||
{QMI_FIXED_INTF(0x0b3c, 0xc00a, 6)}, /* Olivetti Olicard 160 */
|
||||
{QMI_FIXED_INTF(0x0b3c, 0xc00b, 4)}, /* Olivetti Olicard 500 */
|
||||
{QMI_FIXED_INTF(0x1e2d, 0x0060, 4)}, /* Cinterion PLxx */
|
||||
{QMI_FIXED_INTF(0x1e2d, 0x0053, 4)}, /* Cinterion PHxx,PXxx */
|
||||
|
|
|
@ -6634,6 +6634,9 @@ EXPORT_SYMBOL(unregister_netdevice_queue);
|
|||
/**
|
||||
* unregister_netdevice_many - unregister many devices
|
||||
* @head: list of devices
|
||||
*
|
||||
* Note: As most callers use a stack allocated list_head,
|
||||
* we force a list_del() to make sure stack wont be corrupted later.
|
||||
*/
|
||||
void unregister_netdevice_many(struct list_head *head)
|
||||
{
|
||||
|
@ -6643,6 +6646,7 @@ void unregister_netdevice_many(struct list_head *head)
|
|||
rollback_registered_many(head);
|
||||
list_for_each_entry(dev, head, unreg_list)
|
||||
net_set_todo(dev);
|
||||
list_del(head);
|
||||
}
|
||||
}
|
||||
EXPORT_SYMBOL(unregister_netdevice_many);
|
||||
|
@ -7098,7 +7102,6 @@ static void __net_exit default_device_exit_batch(struct list_head *net_list)
|
|||
}
|
||||
}
|
||||
unregister_netdevice_many(&dev_kill_list);
|
||||
list_del(&dev_kill_list);
|
||||
rtnl_unlock();
|
||||
}
|
||||
|
||||
|
|
|
@ -1216,6 +1216,8 @@ static const struct nla_policy ifla_vf_policy[IFLA_VF_MAX+1] = {
|
|||
.len = sizeof(struct ifla_vf_spoofchk) },
|
||||
[IFLA_VF_RATE] = { .type = NLA_BINARY,
|
||||
.len = sizeof(struct ifla_vf_rate) },
|
||||
[IFLA_VF_LINK_STATE] = { .type = NLA_BINARY,
|
||||
.len = sizeof(struct ifla_vf_link_state) },
|
||||
};
|
||||
|
||||
static const struct nla_policy ifla_port_policy[IFLA_PORT_MAX+1] = {
|
||||
|
@ -1770,7 +1772,6 @@ static int rtnl_dellink(struct sk_buff *skb, struct nlmsghdr *nlh)
|
|||
|
||||
ops->dellink(dev, &list_kill);
|
||||
unregister_netdevice_many(&list_kill);
|
||||
list_del(&list_kill);
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
|
|
@ -2885,6 +2885,7 @@ struct sk_buff *skb_segment(struct sk_buff *head_skb,
|
|||
int pos;
|
||||
int dummy;
|
||||
|
||||
__skb_push(head_skb, doffset);
|
||||
proto = skb_network_protocol(head_skb, &dummy);
|
||||
if (unlikely(!proto))
|
||||
return ERR_PTR(-EINVAL);
|
||||
|
@ -2892,7 +2893,6 @@ struct sk_buff *skb_segment(struct sk_buff *head_skb,
|
|||
csum = !head_skb->encap_hdr_csum &&
|
||||
!!can_checksum_protocol(features, proto);
|
||||
|
||||
__skb_push(head_skb, doffset);
|
||||
headroom = skb_headroom(head_skb);
|
||||
pos = skb_headlen(head_skb);
|
||||
|
||||
|
|
|
@ -149,7 +149,9 @@ int dns_query(const char *type, const char *name, size_t namelen,
|
|||
if (!*_result)
|
||||
goto put;
|
||||
|
||||
memcpy(*_result, upayload->data, len + 1);
|
||||
memcpy(*_result, upayload->data, len);
|
||||
*_result[len] = '\0';
|
||||
|
||||
if (_expiry)
|
||||
*_expiry = rkey->expiry;
|
||||
|
||||
|
|
|
@ -268,6 +268,7 @@ static struct ip_tunnel *ip_tunnel_find(struct ip_tunnel_net *itn,
|
|||
__be32 remote = parms->iph.daddr;
|
||||
__be32 local = parms->iph.saddr;
|
||||
__be32 key = parms->i_key;
|
||||
__be16 flags = parms->i_flags;
|
||||
int link = parms->link;
|
||||
struct ip_tunnel *t = NULL;
|
||||
struct hlist_head *head = ip_bucket(itn, parms);
|
||||
|
@ -275,9 +276,9 @@ static struct ip_tunnel *ip_tunnel_find(struct ip_tunnel_net *itn,
|
|||
hlist_for_each_entry_rcu(t, head, hash_node) {
|
||||
if (local == t->parms.iph.saddr &&
|
||||
remote == t->parms.iph.daddr &&
|
||||
key == t->parms.i_key &&
|
||||
link == t->parms.link &&
|
||||
type == t->dev->type)
|
||||
type == t->dev->type &&
|
||||
ip_tunnel_key_match(&t->parms, flags, key))
|
||||
break;
|
||||
}
|
||||
return t;
|
||||
|
@ -667,6 +668,7 @@ void ip_tunnel_xmit(struct sk_buff *skb, struct net_device *dev,
|
|||
dev->needed_headroom = max_headroom;
|
||||
|
||||
if (skb_cow_head(skb, dev->needed_headroom)) {
|
||||
ip_rt_put(rt);
|
||||
dev->stats.tx_dropped++;
|
||||
kfree_skb(skb);
|
||||
return;
|
||||
|
@ -746,10 +748,12 @@ int ip_tunnel_ioctl(struct net_device *dev, struct ip_tunnel_parm *p, int cmd)
|
|||
goto done;
|
||||
if (p->iph.ttl)
|
||||
p->iph.frag_off |= htons(IP_DF);
|
||||
if (!(p->i_flags&TUNNEL_KEY))
|
||||
p->i_key = 0;
|
||||
if (!(p->o_flags&TUNNEL_KEY))
|
||||
p->o_key = 0;
|
||||
if (!(p->i_flags & VTI_ISVTI)) {
|
||||
if (!(p->i_flags & TUNNEL_KEY))
|
||||
p->i_key = 0;
|
||||
if (!(p->o_flags & TUNNEL_KEY))
|
||||
p->o_key = 0;
|
||||
}
|
||||
|
||||
t = ip_tunnel_find(itn, p, itn->fb_tunnel_dev->type);
|
||||
|
||||
|
|
|
@ -313,7 +313,13 @@ vti_tunnel_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
|
|||
return -EINVAL;
|
||||
}
|
||||
|
||||
p.i_flags |= VTI_ISVTI;
|
||||
if (!(p.i_flags & GRE_KEY))
|
||||
p.i_key = 0;
|
||||
if (!(p.o_flags & GRE_KEY))
|
||||
p.o_key = 0;
|
||||
|
||||
p.i_flags = VTI_ISVTI;
|
||||
|
||||
err = ip_tunnel_ioctl(dev, &p, cmd);
|
||||
if (err)
|
||||
return err;
|
||||
|
|
|
@ -149,13 +149,13 @@ static int ipip_err(struct sk_buff *skb, u32 info)
|
|||
|
||||
if (type == ICMP_DEST_UNREACH && code == ICMP_FRAG_NEEDED) {
|
||||
ipv4_update_pmtu(skb, dev_net(skb->dev), info,
|
||||
t->dev->ifindex, 0, IPPROTO_IPIP, 0);
|
||||
t->parms.link, 0, IPPROTO_IPIP, 0);
|
||||
err = 0;
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (type == ICMP_REDIRECT) {
|
||||
ipv4_redirect(skb, dev_net(skb->dev), t->dev->ifindex, 0,
|
||||
ipv4_redirect(skb, dev_net(skb->dev), t->parms.link, 0,
|
||||
IPPROTO_IPIP, 0);
|
||||
err = 0;
|
||||
goto out;
|
||||
|
|
|
@ -78,6 +78,7 @@ int __ip6_local_out(struct sk_buff *skb)
|
|||
if (len > IPV6_MAXPLEN)
|
||||
len = 0;
|
||||
ipv6_hdr(skb)->payload_len = htons(len);
|
||||
IP6CB(skb)->nhoff = offsetof(struct ipv6hdr, nexthdr);
|
||||
|
||||
return nf_hook(NFPROTO_IPV6, NF_INET_LOCAL_OUT, skb, NULL,
|
||||
skb_dst(skb)->dev, dst_output);
|
||||
|
|
|
@ -560,12 +560,12 @@ static int ipip6_err(struct sk_buff *skb, u32 info)
|
|||
|
||||
if (type == ICMP_DEST_UNREACH && code == ICMP_FRAG_NEEDED) {
|
||||
ipv4_update_pmtu(skb, dev_net(skb->dev), info,
|
||||
t->dev->ifindex, 0, IPPROTO_IPV6, 0);
|
||||
t->parms.link, 0, IPPROTO_IPV6, 0);
|
||||
err = 0;
|
||||
goto out;
|
||||
}
|
||||
if (type == ICMP_REDIRECT) {
|
||||
ipv4_redirect(skb, dev_net(skb->dev), t->dev->ifindex, 0,
|
||||
ipv4_redirect(skb, dev_net(skb->dev), t->parms.link, 0,
|
||||
IPPROTO_IPV6, 0);
|
||||
err = 0;
|
||||
goto out;
|
||||
|
|
|
@ -1780,7 +1780,6 @@ void ieee80211_remove_interfaces(struct ieee80211_local *local)
|
|||
}
|
||||
mutex_unlock(&local->iflist_mtx);
|
||||
unregister_netdevice_many(&unreg_list);
|
||||
list_del(&unreg_list);
|
||||
|
||||
list_for_each_entry_safe(sdata, tmp, &wdev_list, list) {
|
||||
list_del(&sdata->list);
|
||||
|
|
Loading…
Reference in New Issue