Merge git://git.kernel.org/pub/scm/linux/kernel/git/davem/net

Conflicts:
	net/core/rtnetlink.c
	net/core/skbuff.c

Both conflicts were very simple overlapping changes.

Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
David S. Miller 2014-06-11 16:02:55 -07:00
commit 902455e007
15 changed files with 66 additions and 27 deletions

View File

@ -83,9 +83,9 @@ static void bpf_flush_icache(void *start_, void *end_)
#define BNE (F2(0, 2) | CONDNE) #define BNE (F2(0, 2) | CONDNE)
#ifdef CONFIG_SPARC64 #ifdef CONFIG_SPARC64
#define BNE_PTR (F2(0, 1) | CONDNE | (2 << 20)) #define BE_PTR (F2(0, 1) | CONDE | (2 << 20))
#else #else
#define BNE_PTR BNE #define BE_PTR BE
#endif #endif
#define SETHI(K, REG) \ #define SETHI(K, REG) \
@ -592,7 +592,7 @@ void bpf_jit_compile(struct sk_filter *fp)
case BPF_ANC | SKF_AD_IFINDEX: case BPF_ANC | SKF_AD_IFINDEX:
emit_skb_loadptr(dev, r_A); emit_skb_loadptr(dev, r_A);
emit_cmpi(r_A, 0); emit_cmpi(r_A, 0);
emit_branch(BNE_PTR, cleanup_addr + 4); emit_branch(BE_PTR, cleanup_addr + 4);
emit_nop(); emit_nop();
emit_load32(r_A, struct net_device, ifindex, r_A); emit_load32(r_A, struct net_device, ifindex, r_A);
break; break;
@ -605,7 +605,7 @@ void bpf_jit_compile(struct sk_filter *fp)
case BPF_ANC | SKF_AD_HATYPE: case BPF_ANC | SKF_AD_HATYPE:
emit_skb_loadptr(dev, r_A); emit_skb_loadptr(dev, r_A);
emit_cmpi(r_A, 0); emit_cmpi(r_A, 0);
emit_branch(BNE_PTR, cleanup_addr + 4); emit_branch(BE_PTR, cleanup_addr + 4);
emit_nop(); emit_nop();
emit_load16(r_A, struct net_device, type, r_A); emit_load16(r_A, struct net_device, type, r_A);
break; break;

View File

@ -66,10 +66,17 @@
#define EFX_USE_QWORD_IO 1 #define EFX_USE_QWORD_IO 1
#endif #endif
/* Hardware issue requires that only 64-bit naturally aligned writes
* are seen by hardware. Its not strictly necessary to restrict to
* x86_64 arch, but done for safety since unusual write combining behaviour
* can break PIO.
*/
#ifdef CONFIG_X86_64
/* PIO is a win only if write-combining is possible */ /* PIO is a win only if write-combining is possible */
#ifdef ARCH_HAS_IOREMAP_WC #ifdef ARCH_HAS_IOREMAP_WC
#define EFX_USE_PIO 1 #define EFX_USE_PIO 1
#endif #endif
#endif
#ifdef EFX_USE_QWORD_IO #ifdef EFX_USE_QWORD_IO
static inline void _efx_writeq(struct efx_nic *efx, __le64 value, static inline void _efx_writeq(struct efx_nic *efx, __le64 value,

View File

@ -189,6 +189,18 @@ struct efx_short_copy_buffer {
u8 buf[L1_CACHE_BYTES]; u8 buf[L1_CACHE_BYTES];
}; };
/* Copy in explicit 64-bit writes. */
static void efx_memcpy_64(void __iomem *dest, void *src, size_t len)
{
u64 *src64 = src;
u64 __iomem *dest64 = dest;
size_t l64 = len / 8;
size_t i;
for (i = 0; i < l64; i++)
writeq(src64[i], &dest64[i]);
}
/* Copy to PIO, respecting that writes to PIO buffers must be dword aligned. /* Copy to PIO, respecting that writes to PIO buffers must be dword aligned.
* Advances piobuf pointer. Leaves additional data in the copy buffer. * Advances piobuf pointer. Leaves additional data in the copy buffer.
*/ */
@ -198,7 +210,7 @@ static void efx_memcpy_toio_aligned(struct efx_nic *efx, u8 __iomem **piobuf,
{ {
int block_len = len & ~(sizeof(copy_buf->buf) - 1); int block_len = len & ~(sizeof(copy_buf->buf) - 1);
memcpy_toio(*piobuf, data, block_len); efx_memcpy_64(*piobuf, data, block_len);
*piobuf += block_len; *piobuf += block_len;
len -= block_len; len -= block_len;
@ -230,7 +242,7 @@ static void efx_memcpy_toio_aligned_cb(struct efx_nic *efx, u8 __iomem **piobuf,
if (copy_buf->used < sizeof(copy_buf->buf)) if (copy_buf->used < sizeof(copy_buf->buf))
return; return;
memcpy_toio(*piobuf, copy_buf->buf, sizeof(copy_buf->buf)); efx_memcpy_64(*piobuf, copy_buf->buf, sizeof(copy_buf->buf));
*piobuf += sizeof(copy_buf->buf); *piobuf += sizeof(copy_buf->buf);
data += copy_to_buf; data += copy_to_buf;
len -= copy_to_buf; len -= copy_to_buf;
@ -245,7 +257,7 @@ static void efx_flush_copy_buffer(struct efx_nic *efx, u8 __iomem *piobuf,
{ {
/* if there's anything in it, write the whole buffer, including junk */ /* if there's anything in it, write the whole buffer, including junk */
if (copy_buf->used) if (copy_buf->used)
memcpy_toio(piobuf, copy_buf->buf, sizeof(copy_buf->buf)); efx_memcpy_64(piobuf, copy_buf->buf, sizeof(copy_buf->buf));
} }
/* Traverse skb structure and copy fragments in to PIO buffer. /* Traverse skb structure and copy fragments in to PIO buffer.
@ -304,8 +316,8 @@ efx_enqueue_skb_pio(struct efx_tx_queue *tx_queue, struct sk_buff *skb)
*/ */
BUILD_BUG_ON(L1_CACHE_BYTES > BUILD_BUG_ON(L1_CACHE_BYTES >
SKB_DATA_ALIGN(sizeof(struct skb_shared_info))); SKB_DATA_ALIGN(sizeof(struct skb_shared_info)));
memcpy_toio(tx_queue->piobuf, skb->data, efx_memcpy_64(tx_queue->piobuf, skb->data,
ALIGN(skb->len, L1_CACHE_BYTES)); ALIGN(skb->len, L1_CACHE_BYTES));
} }
EFX_POPULATE_QWORD_5(buffer->option, EFX_POPULATE_QWORD_5(buffer->option,

View File

@ -1204,7 +1204,6 @@ static int macvlan_device_event(struct notifier_block *unused,
list_for_each_entry_safe(vlan, next, &port->vlans, list) list_for_each_entry_safe(vlan, next, &port->vlans, list)
vlan->dev->rtnl_link_ops->dellink(vlan->dev, &list_kill); vlan->dev->rtnl_link_ops->dellink(vlan->dev, &list_kill);
unregister_netdevice_many(&list_kill); unregister_netdevice_many(&list_kill);
list_del(&list_kill);
break; break;
case NETDEV_PRE_TYPE_CHANGE: case NETDEV_PRE_TYPE_CHANGE:
/* Forbid underlaying device to change its type. */ /* Forbid underlaying device to change its type. */

View File

@ -763,7 +763,12 @@ static const struct usb_device_id products[] = {
{QMI_FIXED_INTF(0x2357, 0x9000, 4)}, /* TP-LINK MA260 */ {QMI_FIXED_INTF(0x2357, 0x9000, 4)}, /* TP-LINK MA260 */
{QMI_FIXED_INTF(0x1bc7, 0x1200, 5)}, /* Telit LE920 */ {QMI_FIXED_INTF(0x1bc7, 0x1200, 5)}, /* Telit LE920 */
{QMI_FIXED_INTF(0x1bc7, 0x1201, 2)}, /* Telit LE920 */ {QMI_FIXED_INTF(0x1bc7, 0x1201, 2)}, /* Telit LE920 */
{QMI_FIXED_INTF(0x0b3c, 0xc005, 6)}, /* Olivetti Olicard 200 */ {QMI_FIXED_INTF(0x0b3c, 0xc000, 4)}, /* Olivetti Olicard 100 */
{QMI_FIXED_INTF(0x0b3c, 0xc001, 4)}, /* Olivetti Olicard 120 */
{QMI_FIXED_INTF(0x0b3c, 0xc002, 4)}, /* Olivetti Olicard 140 */
{QMI_FIXED_INTF(0x0b3c, 0xc004, 6)}, /* Olivetti Olicard 155 */
{QMI_FIXED_INTF(0x0b3c, 0xc005, 6)}, /* Olivetti Olicard 200 */
{QMI_FIXED_INTF(0x0b3c, 0xc00a, 6)}, /* Olivetti Olicard 160 */
{QMI_FIXED_INTF(0x0b3c, 0xc00b, 4)}, /* Olivetti Olicard 500 */ {QMI_FIXED_INTF(0x0b3c, 0xc00b, 4)}, /* Olivetti Olicard 500 */
{QMI_FIXED_INTF(0x1e2d, 0x0060, 4)}, /* Cinterion PLxx */ {QMI_FIXED_INTF(0x1e2d, 0x0060, 4)}, /* Cinterion PLxx */
{QMI_FIXED_INTF(0x1e2d, 0x0053, 4)}, /* Cinterion PHxx,PXxx */ {QMI_FIXED_INTF(0x1e2d, 0x0053, 4)}, /* Cinterion PHxx,PXxx */

View File

@ -6634,6 +6634,9 @@ EXPORT_SYMBOL(unregister_netdevice_queue);
/** /**
* unregister_netdevice_many - unregister many devices * unregister_netdevice_many - unregister many devices
* @head: list of devices * @head: list of devices
*
* Note: As most callers use a stack allocated list_head,
* we force a list_del() to make sure stack wont be corrupted later.
*/ */
void unregister_netdevice_many(struct list_head *head) void unregister_netdevice_many(struct list_head *head)
{ {
@ -6643,6 +6646,7 @@ void unregister_netdevice_many(struct list_head *head)
rollback_registered_many(head); rollback_registered_many(head);
list_for_each_entry(dev, head, unreg_list) list_for_each_entry(dev, head, unreg_list)
net_set_todo(dev); net_set_todo(dev);
list_del(head);
} }
} }
EXPORT_SYMBOL(unregister_netdevice_many); EXPORT_SYMBOL(unregister_netdevice_many);
@ -7098,7 +7102,6 @@ static void __net_exit default_device_exit_batch(struct list_head *net_list)
} }
} }
unregister_netdevice_many(&dev_kill_list); unregister_netdevice_many(&dev_kill_list);
list_del(&dev_kill_list);
rtnl_unlock(); rtnl_unlock();
} }

View File

@ -1216,6 +1216,8 @@ static const struct nla_policy ifla_vf_policy[IFLA_VF_MAX+1] = {
.len = sizeof(struct ifla_vf_spoofchk) }, .len = sizeof(struct ifla_vf_spoofchk) },
[IFLA_VF_RATE] = { .type = NLA_BINARY, [IFLA_VF_RATE] = { .type = NLA_BINARY,
.len = sizeof(struct ifla_vf_rate) }, .len = sizeof(struct ifla_vf_rate) },
[IFLA_VF_LINK_STATE] = { .type = NLA_BINARY,
.len = sizeof(struct ifla_vf_link_state) },
}; };
static const struct nla_policy ifla_port_policy[IFLA_PORT_MAX+1] = { static const struct nla_policy ifla_port_policy[IFLA_PORT_MAX+1] = {
@ -1770,7 +1772,6 @@ static int rtnl_dellink(struct sk_buff *skb, struct nlmsghdr *nlh)
ops->dellink(dev, &list_kill); ops->dellink(dev, &list_kill);
unregister_netdevice_many(&list_kill); unregister_netdevice_many(&list_kill);
list_del(&list_kill);
return 0; return 0;
} }

View File

@ -2885,6 +2885,7 @@ struct sk_buff *skb_segment(struct sk_buff *head_skb,
int pos; int pos;
int dummy; int dummy;
__skb_push(head_skb, doffset);
proto = skb_network_protocol(head_skb, &dummy); proto = skb_network_protocol(head_skb, &dummy);
if (unlikely(!proto)) if (unlikely(!proto))
return ERR_PTR(-EINVAL); return ERR_PTR(-EINVAL);
@ -2892,7 +2893,6 @@ struct sk_buff *skb_segment(struct sk_buff *head_skb,
csum = !head_skb->encap_hdr_csum && csum = !head_skb->encap_hdr_csum &&
!!can_checksum_protocol(features, proto); !!can_checksum_protocol(features, proto);
__skb_push(head_skb, doffset);
headroom = skb_headroom(head_skb); headroom = skb_headroom(head_skb);
pos = skb_headlen(head_skb); pos = skb_headlen(head_skb);

View File

@ -149,7 +149,9 @@ int dns_query(const char *type, const char *name, size_t namelen,
if (!*_result) if (!*_result)
goto put; goto put;
memcpy(*_result, upayload->data, len + 1); memcpy(*_result, upayload->data, len);
*_result[len] = '\0';
if (_expiry) if (_expiry)
*_expiry = rkey->expiry; *_expiry = rkey->expiry;

View File

@ -268,6 +268,7 @@ static struct ip_tunnel *ip_tunnel_find(struct ip_tunnel_net *itn,
__be32 remote = parms->iph.daddr; __be32 remote = parms->iph.daddr;
__be32 local = parms->iph.saddr; __be32 local = parms->iph.saddr;
__be32 key = parms->i_key; __be32 key = parms->i_key;
__be16 flags = parms->i_flags;
int link = parms->link; int link = parms->link;
struct ip_tunnel *t = NULL; struct ip_tunnel *t = NULL;
struct hlist_head *head = ip_bucket(itn, parms); struct hlist_head *head = ip_bucket(itn, parms);
@ -275,9 +276,9 @@ static struct ip_tunnel *ip_tunnel_find(struct ip_tunnel_net *itn,
hlist_for_each_entry_rcu(t, head, hash_node) { hlist_for_each_entry_rcu(t, head, hash_node) {
if (local == t->parms.iph.saddr && if (local == t->parms.iph.saddr &&
remote == t->parms.iph.daddr && remote == t->parms.iph.daddr &&
key == t->parms.i_key &&
link == t->parms.link && link == t->parms.link &&
type == t->dev->type) type == t->dev->type &&
ip_tunnel_key_match(&t->parms, flags, key))
break; break;
} }
return t; return t;
@ -667,6 +668,7 @@ void ip_tunnel_xmit(struct sk_buff *skb, struct net_device *dev,
dev->needed_headroom = max_headroom; dev->needed_headroom = max_headroom;
if (skb_cow_head(skb, dev->needed_headroom)) { if (skb_cow_head(skb, dev->needed_headroom)) {
ip_rt_put(rt);
dev->stats.tx_dropped++; dev->stats.tx_dropped++;
kfree_skb(skb); kfree_skb(skb);
return; return;
@ -746,10 +748,12 @@ int ip_tunnel_ioctl(struct net_device *dev, struct ip_tunnel_parm *p, int cmd)
goto done; goto done;
if (p->iph.ttl) if (p->iph.ttl)
p->iph.frag_off |= htons(IP_DF); p->iph.frag_off |= htons(IP_DF);
if (!(p->i_flags&TUNNEL_KEY)) if (!(p->i_flags & VTI_ISVTI)) {
p->i_key = 0; if (!(p->i_flags & TUNNEL_KEY))
if (!(p->o_flags&TUNNEL_KEY)) p->i_key = 0;
p->o_key = 0; if (!(p->o_flags & TUNNEL_KEY))
p->o_key = 0;
}
t = ip_tunnel_find(itn, p, itn->fb_tunnel_dev->type); t = ip_tunnel_find(itn, p, itn->fb_tunnel_dev->type);

View File

@ -313,7 +313,13 @@ vti_tunnel_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
return -EINVAL; return -EINVAL;
} }
p.i_flags |= VTI_ISVTI; if (!(p.i_flags & GRE_KEY))
p.i_key = 0;
if (!(p.o_flags & GRE_KEY))
p.o_key = 0;
p.i_flags = VTI_ISVTI;
err = ip_tunnel_ioctl(dev, &p, cmd); err = ip_tunnel_ioctl(dev, &p, cmd);
if (err) if (err)
return err; return err;

View File

@ -149,13 +149,13 @@ static int ipip_err(struct sk_buff *skb, u32 info)
if (type == ICMP_DEST_UNREACH && code == ICMP_FRAG_NEEDED) { if (type == ICMP_DEST_UNREACH && code == ICMP_FRAG_NEEDED) {
ipv4_update_pmtu(skb, dev_net(skb->dev), info, ipv4_update_pmtu(skb, dev_net(skb->dev), info,
t->dev->ifindex, 0, IPPROTO_IPIP, 0); t->parms.link, 0, IPPROTO_IPIP, 0);
err = 0; err = 0;
goto out; goto out;
} }
if (type == ICMP_REDIRECT) { if (type == ICMP_REDIRECT) {
ipv4_redirect(skb, dev_net(skb->dev), t->dev->ifindex, 0, ipv4_redirect(skb, dev_net(skb->dev), t->parms.link, 0,
IPPROTO_IPIP, 0); IPPROTO_IPIP, 0);
err = 0; err = 0;
goto out; goto out;

View File

@ -78,6 +78,7 @@ int __ip6_local_out(struct sk_buff *skb)
if (len > IPV6_MAXPLEN) if (len > IPV6_MAXPLEN)
len = 0; len = 0;
ipv6_hdr(skb)->payload_len = htons(len); ipv6_hdr(skb)->payload_len = htons(len);
IP6CB(skb)->nhoff = offsetof(struct ipv6hdr, nexthdr);
return nf_hook(NFPROTO_IPV6, NF_INET_LOCAL_OUT, skb, NULL, return nf_hook(NFPROTO_IPV6, NF_INET_LOCAL_OUT, skb, NULL,
skb_dst(skb)->dev, dst_output); skb_dst(skb)->dev, dst_output);

View File

@ -560,12 +560,12 @@ static int ipip6_err(struct sk_buff *skb, u32 info)
if (type == ICMP_DEST_UNREACH && code == ICMP_FRAG_NEEDED) { if (type == ICMP_DEST_UNREACH && code == ICMP_FRAG_NEEDED) {
ipv4_update_pmtu(skb, dev_net(skb->dev), info, ipv4_update_pmtu(skb, dev_net(skb->dev), info,
t->dev->ifindex, 0, IPPROTO_IPV6, 0); t->parms.link, 0, IPPROTO_IPV6, 0);
err = 0; err = 0;
goto out; goto out;
} }
if (type == ICMP_REDIRECT) { if (type == ICMP_REDIRECT) {
ipv4_redirect(skb, dev_net(skb->dev), t->dev->ifindex, 0, ipv4_redirect(skb, dev_net(skb->dev), t->parms.link, 0,
IPPROTO_IPV6, 0); IPPROTO_IPV6, 0);
err = 0; err = 0;
goto out; goto out;

View File

@ -1780,7 +1780,6 @@ void ieee80211_remove_interfaces(struct ieee80211_local *local)
} }
mutex_unlock(&local->iflist_mtx); mutex_unlock(&local->iflist_mtx);
unregister_netdevice_many(&unreg_list); unregister_netdevice_many(&unreg_list);
list_del(&unreg_list);
list_for_each_entry_safe(sdata, tmp, &wdev_list, list) { list_for_each_entry_safe(sdata, tmp, &wdev_list, list) {
list_del(&sdata->list); list_del(&sdata->list);