2010-02-28 02:41:45 +07:00
|
|
|
/*
|
|
|
|
* Bridge multicast support.
|
|
|
|
*
|
|
|
|
* Copyright (c) 2010 Herbert Xu <herbert@gondor.apana.org.au>
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
|
|
* under the terms of the GNU General Public License as published by the Free
|
|
|
|
* Software Foundation; either version 2 of the License, or (at your option)
|
|
|
|
* any later version.
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/err.h>
|
|
|
|
#include <linux/if_ether.h>
|
|
|
|
#include <linux/igmp.h>
|
|
|
|
#include <linux/jhash.h>
|
|
|
|
#include <linux/kernel.h>
|
2010-02-28 02:41:51 +07:00
|
|
|
#include <linux/log2.h>
|
2010-02-28 02:41:45 +07:00
|
|
|
#include <linux/netdevice.h>
|
|
|
|
#include <linux/netfilter_bridge.h>
|
|
|
|
#include <linux/random.h>
|
|
|
|
#include <linux/rculist.h>
|
|
|
|
#include <linux/skbuff.h>
|
|
|
|
#include <linux/slab.h>
|
|
|
|
#include <linux/timer.h>
|
2013-05-22 04:52:54 +07:00
|
|
|
#include <linux/inetdevice.h>
|
2010-02-28 02:41:45 +07:00
|
|
|
#include <net/ip.h>
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2010-04-22 23:54:22 +07:00
|
|
|
#include <net/ipv6.h>
|
|
|
|
#include <net/mld.h>
|
2010-04-28 00:16:54 +07:00
|
|
|
#include <net/ip6_checksum.h>
|
2013-09-04 07:13:39 +07:00
|
|
|
#include <net/addrconf.h>
|
2010-04-22 23:54:22 +07:00
|
|
|
#endif
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
#include "br_private.h"
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
static void br_multicast_start_querier(struct net_bridge *br,
|
|
|
|
struct bridge_mcast_query *query);
|
2012-12-10 09:15:35 +07:00
|
|
|
unsigned int br_mdb_rehash_seq;
|
2012-04-13 09:37:42 +07:00
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
static inline int br_ip_equal(const struct br_ip *a, const struct br_ip *b)
|
|
|
|
{
|
|
|
|
if (a->proto != b->proto)
|
|
|
|
return 0;
|
2013-02-13 19:00:17 +07:00
|
|
|
if (a->vid != b->vid)
|
|
|
|
return 0;
|
2010-04-18 10:42:07 +07:00
|
|
|
switch (a->proto) {
|
|
|
|
case htons(ETH_P_IP):
|
|
|
|
return a->u.ip4 == b->u.ip4;
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2010-04-22 23:54:22 +07:00
|
|
|
case htons(ETH_P_IPV6):
|
|
|
|
return ipv6_addr_equal(&a->u.ip6, &b->u.ip6);
|
|
|
|
#endif
|
2010-04-18 10:42:07 +07:00
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-02-13 19:00:17 +07:00
|
|
|
static inline int __br_ip4_hash(struct net_bridge_mdb_htable *mdb, __be32 ip,
|
|
|
|
__u16 vid)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
2013-02-13 19:00:17 +07:00
|
|
|
return jhash_2words((__force u32)ip, vid, mdb->secret) & (mdb->max - 1);
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2010-04-22 23:54:22 +07:00
|
|
|
static inline int __br_ip6_hash(struct net_bridge_mdb_htable *mdb,
|
2013-02-13 19:00:17 +07:00
|
|
|
const struct in6_addr *ip,
|
|
|
|
__u16 vid)
|
2010-04-22 23:54:22 +07:00
|
|
|
{
|
2013-02-13 19:00:17 +07:00
|
|
|
return jhash_2words(ipv6_addr_hash(ip), vid,
|
|
|
|
mdb->secret) & (mdb->max - 1);
|
2010-04-22 23:54:22 +07:00
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
static inline int br_ip_hash(struct net_bridge_mdb_htable *mdb,
|
|
|
|
struct br_ip *ip)
|
|
|
|
{
|
|
|
|
switch (ip->proto) {
|
|
|
|
case htons(ETH_P_IP):
|
2013-02-13 19:00:17 +07:00
|
|
|
return __br_ip4_hash(mdb, ip->u.ip4, ip->vid);
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2010-04-22 23:54:22 +07:00
|
|
|
case htons(ETH_P_IPV6):
|
2013-02-13 19:00:17 +07:00
|
|
|
return __br_ip6_hash(mdb, &ip->u.ip6, ip->vid);
|
2010-04-22 23:54:22 +07:00
|
|
|
#endif
|
2010-04-18 10:42:07 +07:00
|
|
|
}
|
|
|
|
return 0;
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static struct net_bridge_mdb_entry *__br_mdb_ip_get(
|
2010-04-18 10:42:07 +07:00
|
|
|
struct net_bridge_mdb_htable *mdb, struct br_ip *dst, int hash)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
|
|
|
struct net_bridge_mdb_entry *mp;
|
|
|
|
|
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 08:06:00 +07:00
|
|
|
hlist_for_each_entry_rcu(mp, &mdb->mhash[hash], hlist[mdb->ver]) {
|
2010-04-18 10:42:07 +07:00
|
|
|
if (br_ip_equal(&mp->addr, dst))
|
2010-02-28 02:41:45 +07:00
|
|
|
return mp;
|
|
|
|
}
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2012-12-12 05:23:08 +07:00
|
|
|
struct net_bridge_mdb_entry *br_mdb_ip_get(struct net_bridge_mdb_htable *mdb,
|
|
|
|
struct br_ip *dst)
|
2010-07-05 21:50:08 +07:00
|
|
|
{
|
|
|
|
if (!mdb)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
return __br_mdb_ip_get(mdb, dst, br_ip_hash(mdb, dst));
|
|
|
|
}
|
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
static struct net_bridge_mdb_entry *br_mdb_ip4_get(
|
2013-02-13 19:00:17 +07:00
|
|
|
struct net_bridge_mdb_htable *mdb, __be32 dst, __u16 vid)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
2010-04-18 10:42:07 +07:00
|
|
|
struct br_ip br_dst;
|
|
|
|
|
|
|
|
br_dst.u.ip4 = dst;
|
|
|
|
br_dst.proto = htons(ETH_P_IP);
|
2013-02-13 19:00:17 +07:00
|
|
|
br_dst.vid = vid;
|
2010-03-16 10:38:25 +07:00
|
|
|
|
2010-07-05 21:50:08 +07:00
|
|
|
return br_mdb_ip_get(mdb, &br_dst);
|
2010-04-18 10:42:07 +07:00
|
|
|
}
|
|
|
|
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2010-04-22 23:54:22 +07:00
|
|
|
static struct net_bridge_mdb_entry *br_mdb_ip6_get(
|
2013-02-13 19:00:17 +07:00
|
|
|
struct net_bridge_mdb_htable *mdb, const struct in6_addr *dst,
|
|
|
|
__u16 vid)
|
2010-04-22 23:54:22 +07:00
|
|
|
{
|
|
|
|
struct br_ip br_dst;
|
2010-03-16 10:38:25 +07:00
|
|
|
|
2011-11-21 10:39:03 +07:00
|
|
|
br_dst.u.ip6 = *dst;
|
2010-04-22 23:54:22 +07:00
|
|
|
br_dst.proto = htons(ETH_P_IPV6);
|
2013-02-13 19:00:17 +07:00
|
|
|
br_dst.vid = vid;
|
2010-04-22 23:54:22 +07:00
|
|
|
|
2010-07-05 21:50:08 +07:00
|
|
|
return br_mdb_ip_get(mdb, &br_dst);
|
2010-04-22 23:54:22 +07:00
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
struct net_bridge_mdb_entry *br_mdb_get(struct net_bridge *br,
|
2013-03-07 10:05:33 +07:00
|
|
|
struct sk_buff *skb, u16 vid)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
2010-11-15 13:38:10 +07:00
|
|
|
struct net_bridge_mdb_htable *mdb = rcu_dereference(br->mdb);
|
2010-04-18 10:42:07 +07:00
|
|
|
struct br_ip ip;
|
|
|
|
|
2010-07-05 21:50:08 +07:00
|
|
|
if (br->multicast_disabled)
|
2010-02-28 02:41:45 +07:00
|
|
|
return NULL;
|
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
if (BR_INPUT_SKB_CB(skb)->igmp)
|
2010-02-28 02:41:45 +07:00
|
|
|
return NULL;
|
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
ip.proto = skb->protocol;
|
2013-03-07 10:05:33 +07:00
|
|
|
ip.vid = vid;
|
2010-04-18 10:42:07 +07:00
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
switch (skb->protocol) {
|
|
|
|
case htons(ETH_P_IP):
|
2010-04-18 10:42:07 +07:00
|
|
|
ip.u.ip4 = ip_hdr(skb)->daddr;
|
|
|
|
break;
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2010-04-22 23:54:22 +07:00
|
|
|
case htons(ETH_P_IPV6):
|
2011-11-21 10:39:03 +07:00
|
|
|
ip.u.ip6 = ipv6_hdr(skb)->daddr;
|
2010-04-22 23:54:22 +07:00
|
|
|
break;
|
|
|
|
#endif
|
2010-04-18 10:42:07 +07:00
|
|
|
default:
|
|
|
|
return NULL;
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
return br_mdb_ip_get(mdb, &ip);
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static void br_mdb_free(struct rcu_head *head)
|
|
|
|
{
|
|
|
|
struct net_bridge_mdb_htable *mdb =
|
|
|
|
container_of(head, struct net_bridge_mdb_htable, rcu);
|
|
|
|
struct net_bridge_mdb_htable *old = mdb->old;
|
|
|
|
|
|
|
|
mdb->old = NULL;
|
|
|
|
kfree(old->mhash);
|
|
|
|
kfree(old);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int br_mdb_copy(struct net_bridge_mdb_htable *new,
|
|
|
|
struct net_bridge_mdb_htable *old,
|
|
|
|
int elasticity)
|
|
|
|
{
|
|
|
|
struct net_bridge_mdb_entry *mp;
|
|
|
|
int maxlen;
|
|
|
|
int len;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < old->max; i++)
|
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 08:06:00 +07:00
|
|
|
hlist_for_each_entry(mp, &old->mhash[i], hlist[old->ver])
|
2010-02-28 02:41:45 +07:00
|
|
|
hlist_add_head(&mp->hlist[new->ver],
|
2010-04-18 10:42:07 +07:00
|
|
|
&new->mhash[br_ip_hash(new, &mp->addr)]);
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
if (!elasticity)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
maxlen = 0;
|
|
|
|
for (i = 0; i < new->max; i++) {
|
|
|
|
len = 0;
|
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 08:06:00 +07:00
|
|
|
hlist_for_each_entry(mp, &new->mhash[i], hlist[new->ver])
|
2010-02-28 02:41:45 +07:00
|
|
|
len++;
|
|
|
|
if (len > maxlen)
|
|
|
|
maxlen = len;
|
|
|
|
}
|
|
|
|
|
|
|
|
return maxlen > elasticity ? -EINVAL : 0;
|
|
|
|
}
|
|
|
|
|
2012-12-12 05:23:08 +07:00
|
|
|
void br_multicast_free_pg(struct rcu_head *head)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
|
|
|
struct net_bridge_port_group *p =
|
|
|
|
container_of(head, struct net_bridge_port_group, rcu);
|
|
|
|
|
|
|
|
kfree(p);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void br_multicast_free_group(struct rcu_head *head)
|
|
|
|
{
|
|
|
|
struct net_bridge_mdb_entry *mp =
|
|
|
|
container_of(head, struct net_bridge_mdb_entry, rcu);
|
|
|
|
|
|
|
|
kfree(mp);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void br_multicast_group_expired(unsigned long data)
|
|
|
|
{
|
|
|
|
struct net_bridge_mdb_entry *mp = (void *)data;
|
|
|
|
struct net_bridge *br = mp->br;
|
|
|
|
struct net_bridge_mdb_htable *mdb;
|
|
|
|
|
|
|
|
spin_lock(&br->multicast_lock);
|
|
|
|
if (!netif_running(br->dev) || timer_pending(&mp->timer))
|
|
|
|
goto out;
|
|
|
|
|
2011-02-12 16:05:42 +07:00
|
|
|
mp->mglist = false;
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
if (mp->ports)
|
|
|
|
goto out;
|
|
|
|
|
2010-11-15 13:38:10 +07:00
|
|
|
mdb = mlock_dereference(br->mdb, br);
|
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
hlist_del_rcu(&mp->hlist[mdb->ver]);
|
|
|
|
mdb->size--;
|
|
|
|
|
|
|
|
call_rcu_bh(&mp->rcu, br_multicast_free_group);
|
|
|
|
|
|
|
|
out:
|
|
|
|
spin_unlock(&br->multicast_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void br_multicast_del_pg(struct net_bridge *br,
|
|
|
|
struct net_bridge_port_group *pg)
|
|
|
|
{
|
2010-11-15 13:38:10 +07:00
|
|
|
struct net_bridge_mdb_htable *mdb;
|
2010-02-28 02:41:45 +07:00
|
|
|
struct net_bridge_mdb_entry *mp;
|
|
|
|
struct net_bridge_port_group *p;
|
2010-11-15 13:38:10 +07:00
|
|
|
struct net_bridge_port_group __rcu **pp;
|
|
|
|
|
|
|
|
mdb = mlock_dereference(br->mdb, br);
|
2010-02-28 02:41:45 +07:00
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
mp = br_mdb_ip_get(mdb, &pg->addr);
|
2010-02-28 02:41:45 +07:00
|
|
|
if (WARN_ON(!mp))
|
|
|
|
return;
|
|
|
|
|
2010-11-15 13:38:10 +07:00
|
|
|
for (pp = &mp->ports;
|
|
|
|
(p = mlock_dereference(*pp, br)) != NULL;
|
|
|
|
pp = &p->next) {
|
2010-02-28 02:41:45 +07:00
|
|
|
if (p != pg)
|
|
|
|
continue;
|
|
|
|
|
2010-04-27 22:01:06 +07:00
|
|
|
rcu_assign_pointer(*pp, p->next);
|
2010-02-28 02:41:45 +07:00
|
|
|
hlist_del_init(&p->mglist);
|
|
|
|
del_timer(&p->timer);
|
|
|
|
call_rcu_bh(&p->rcu, br_multicast_free_pg);
|
|
|
|
|
2013-10-20 05:58:57 +07:00
|
|
|
if (!mp->ports && !mp->mglist &&
|
2010-02-28 02:41:45 +07:00
|
|
|
netif_running(br->dev))
|
|
|
|
mod_timer(&mp->timer, jiffies);
|
|
|
|
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
WARN_ON(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void br_multicast_port_group_expired(unsigned long data)
|
|
|
|
{
|
|
|
|
struct net_bridge_port_group *pg = (void *)data;
|
|
|
|
struct net_bridge *br = pg->port->br;
|
|
|
|
|
|
|
|
spin_lock(&br->multicast_lock);
|
|
|
|
if (!netif_running(br->dev) || timer_pending(&pg->timer) ||
|
2012-12-15 05:09:51 +07:00
|
|
|
hlist_unhashed(&pg->mglist) || pg->state & MDB_PERMANENT)
|
2010-02-28 02:41:45 +07:00
|
|
|
goto out;
|
|
|
|
|
|
|
|
br_multicast_del_pg(br, pg);
|
|
|
|
|
|
|
|
out:
|
|
|
|
spin_unlock(&br->multicast_lock);
|
|
|
|
}
|
|
|
|
|
2010-11-15 13:38:10 +07:00
|
|
|
static int br_mdb_rehash(struct net_bridge_mdb_htable __rcu **mdbp, int max,
|
2010-02-28 02:41:45 +07:00
|
|
|
int elasticity)
|
|
|
|
{
|
2010-11-15 13:38:10 +07:00
|
|
|
struct net_bridge_mdb_htable *old = rcu_dereference_protected(*mdbp, 1);
|
2010-02-28 02:41:45 +07:00
|
|
|
struct net_bridge_mdb_htable *mdb;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
mdb = kmalloc(sizeof(*mdb), GFP_ATOMIC);
|
|
|
|
if (!mdb)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
mdb->max = max;
|
|
|
|
mdb->old = old;
|
|
|
|
|
|
|
|
mdb->mhash = kzalloc(max * sizeof(*mdb->mhash), GFP_ATOMIC);
|
|
|
|
if (!mdb->mhash) {
|
|
|
|
kfree(mdb);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
mdb->size = old ? old->size : 0;
|
|
|
|
mdb->ver = old ? old->ver ^ 1 : 0;
|
|
|
|
|
|
|
|
if (!old || elasticity)
|
|
|
|
get_random_bytes(&mdb->secret, sizeof(mdb->secret));
|
|
|
|
else
|
|
|
|
mdb->secret = old->secret;
|
|
|
|
|
|
|
|
if (!old)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
err = br_mdb_copy(mdb, old, elasticity);
|
|
|
|
if (err) {
|
|
|
|
kfree(mdb->mhash);
|
|
|
|
kfree(mdb);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2012-12-10 09:15:35 +07:00
|
|
|
br_mdb_rehash_seq++;
|
2010-02-28 02:41:45 +07:00
|
|
|
call_rcu_bh(&mdb->rcu, br_mdb_free);
|
|
|
|
|
|
|
|
out:
|
|
|
|
rcu_assign_pointer(*mdbp, mdb);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
static struct sk_buff *br_ip4_multicast_alloc_query(struct net_bridge *br,
|
|
|
|
__be32 group)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
|
|
|
struct sk_buff *skb;
|
|
|
|
struct igmphdr *ih;
|
|
|
|
struct ethhdr *eth;
|
|
|
|
struct iphdr *iph;
|
|
|
|
|
|
|
|
skb = netdev_alloc_skb_ip_align(br->dev, sizeof(*eth) + sizeof(*iph) +
|
|
|
|
sizeof(*ih) + 4);
|
|
|
|
if (!skb)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
skb->protocol = htons(ETH_P_IP);
|
|
|
|
|
|
|
|
skb_reset_mac_header(skb);
|
|
|
|
eth = eth_hdr(skb);
|
|
|
|
|
2013-10-02 09:04:40 +07:00
|
|
|
memcpy(eth->h_source, br->dev->dev_addr, ETH_ALEN);
|
2010-02-28 02:41:45 +07:00
|
|
|
eth->h_dest[0] = 1;
|
|
|
|
eth->h_dest[1] = 0;
|
|
|
|
eth->h_dest[2] = 0x5e;
|
|
|
|
eth->h_dest[3] = 0;
|
|
|
|
eth->h_dest[4] = 0;
|
|
|
|
eth->h_dest[5] = 1;
|
|
|
|
eth->h_proto = htons(ETH_P_IP);
|
|
|
|
skb_put(skb, sizeof(*eth));
|
|
|
|
|
|
|
|
skb_set_network_header(skb, skb->len);
|
|
|
|
iph = ip_hdr(skb);
|
|
|
|
|
|
|
|
iph->version = 4;
|
|
|
|
iph->ihl = 6;
|
|
|
|
iph->tos = 0xc0;
|
|
|
|
iph->tot_len = htons(sizeof(*iph) + sizeof(*ih) + 4);
|
|
|
|
iph->id = 0;
|
|
|
|
iph->frag_off = htons(IP_DF);
|
|
|
|
iph->ttl = 1;
|
|
|
|
iph->protocol = IPPROTO_IGMP;
|
2013-05-22 04:52:54 +07:00
|
|
|
iph->saddr = br->multicast_query_use_ifaddr ?
|
|
|
|
inet_select_addr(br->dev, 0, RT_SCOPE_LINK) : 0;
|
2010-02-28 02:41:45 +07:00
|
|
|
iph->daddr = htonl(INADDR_ALLHOSTS_GROUP);
|
|
|
|
((u8 *)&iph[1])[0] = IPOPT_RA;
|
|
|
|
((u8 *)&iph[1])[1] = 4;
|
|
|
|
((u8 *)&iph[1])[2] = 0;
|
|
|
|
((u8 *)&iph[1])[3] = 0;
|
|
|
|
ip_send_check(iph);
|
|
|
|
skb_put(skb, 24);
|
|
|
|
|
|
|
|
skb_set_transport_header(skb, skb->len);
|
|
|
|
ih = igmp_hdr(skb);
|
|
|
|
ih->type = IGMP_HOST_MEMBERSHIP_QUERY;
|
|
|
|
ih->code = (group ? br->multicast_last_member_interval :
|
|
|
|
br->multicast_query_response_interval) /
|
|
|
|
(HZ / IGMP_TIMER_SCALE);
|
|
|
|
ih->group = group;
|
|
|
|
ih->csum = 0;
|
|
|
|
ih->csum = ip_compute_csum((void *)ih, sizeof(struct igmphdr));
|
|
|
|
skb_put(skb, sizeof(*ih));
|
|
|
|
|
|
|
|
__skb_pull(skb, sizeof(*eth));
|
|
|
|
|
|
|
|
out:
|
|
|
|
return skb;
|
|
|
|
}
|
|
|
|
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2010-04-22 23:54:22 +07:00
|
|
|
static struct sk_buff *br_ip6_multicast_alloc_query(struct net_bridge *br,
|
2011-04-22 11:53:02 +07:00
|
|
|
const struct in6_addr *group)
|
2010-04-22 23:54:22 +07:00
|
|
|
{
|
|
|
|
struct sk_buff *skb;
|
|
|
|
struct ipv6hdr *ip6h;
|
|
|
|
struct mld_msg *mldq;
|
|
|
|
struct ethhdr *eth;
|
|
|
|
u8 *hopopt;
|
|
|
|
unsigned long interval;
|
|
|
|
|
|
|
|
skb = netdev_alloc_skb_ip_align(br->dev, sizeof(*eth) + sizeof(*ip6h) +
|
|
|
|
8 + sizeof(*mldq));
|
|
|
|
if (!skb)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
skb->protocol = htons(ETH_P_IPV6);
|
|
|
|
|
|
|
|
/* Ethernet header */
|
|
|
|
skb_reset_mac_header(skb);
|
|
|
|
eth = eth_hdr(skb);
|
|
|
|
|
2013-10-02 09:04:40 +07:00
|
|
|
memcpy(eth->h_source, br->dev->dev_addr, ETH_ALEN);
|
2010-04-22 23:54:22 +07:00
|
|
|
eth->h_proto = htons(ETH_P_IPV6);
|
|
|
|
skb_put(skb, sizeof(*eth));
|
|
|
|
|
|
|
|
/* IPv6 header + HbH option */
|
|
|
|
skb_set_network_header(skb, skb->len);
|
|
|
|
ip6h = ipv6_hdr(skb);
|
|
|
|
|
|
|
|
*(__force __be32 *)ip6h = htonl(0x60000000);
|
2010-12-14 15:42:16 +07:00
|
|
|
ip6h->payload_len = htons(8 + sizeof(*mldq));
|
2010-04-22 23:54:22 +07:00
|
|
|
ip6h->nexthdr = IPPROTO_HOPOPTS;
|
|
|
|
ip6h->hop_limit = 1;
|
2011-03-22 18:40:32 +07:00
|
|
|
ipv6_addr_set(&ip6h->daddr, htonl(0xff020000), 0, 0, htonl(1));
|
2012-03-05 11:52:44 +07:00
|
|
|
if (ipv6_dev_get_saddr(dev_net(br->dev), br->dev, &ip6h->daddr, 0,
|
|
|
|
&ip6h->saddr)) {
|
|
|
|
kfree_skb(skb);
|
|
|
|
return NULL;
|
|
|
|
}
|
2011-02-17 15:17:51 +07:00
|
|
|
ipv6_eth_mc_map(&ip6h->daddr, eth->h_dest);
|
2010-04-22 23:54:22 +07:00
|
|
|
|
|
|
|
hopopt = (u8 *)(ip6h + 1);
|
|
|
|
hopopt[0] = IPPROTO_ICMPV6; /* next hdr */
|
|
|
|
hopopt[1] = 0; /* length of HbH */
|
|
|
|
hopopt[2] = IPV6_TLV_ROUTERALERT; /* Router Alert */
|
|
|
|
hopopt[3] = 2; /* Length of RA Option */
|
|
|
|
hopopt[4] = 0; /* Type = 0x0000 (MLD) */
|
|
|
|
hopopt[5] = 0;
|
2012-05-17 13:00:25 +07:00
|
|
|
hopopt[6] = IPV6_TLV_PAD1; /* Pad1 */
|
|
|
|
hopopt[7] = IPV6_TLV_PAD1; /* Pad1 */
|
2010-04-22 23:54:22 +07:00
|
|
|
|
|
|
|
skb_put(skb, sizeof(*ip6h) + 8);
|
|
|
|
|
|
|
|
/* ICMPv6 */
|
|
|
|
skb_set_transport_header(skb, skb->len);
|
|
|
|
mldq = (struct mld_msg *) icmp6_hdr(skb);
|
|
|
|
|
2013-06-17 04:20:34 +07:00
|
|
|
interval = ipv6_addr_any(group) ?
|
|
|
|
br->multicast_query_response_interval :
|
|
|
|
br->multicast_last_member_interval;
|
2010-04-22 23:54:22 +07:00
|
|
|
|
|
|
|
mldq->mld_type = ICMPV6_MGM_QUERY;
|
|
|
|
mldq->mld_code = 0;
|
|
|
|
mldq->mld_cksum = 0;
|
|
|
|
mldq->mld_maxdelay = htons((u16)jiffies_to_msecs(interval));
|
|
|
|
mldq->mld_reserved = 0;
|
2011-11-21 10:39:03 +07:00
|
|
|
mldq->mld_mca = *group;
|
2010-04-22 23:54:22 +07:00
|
|
|
|
|
|
|
/* checksum */
|
|
|
|
mldq->mld_cksum = csum_ipv6_magic(&ip6h->saddr, &ip6h->daddr,
|
|
|
|
sizeof(*mldq), IPPROTO_ICMPV6,
|
|
|
|
csum_partial(mldq,
|
|
|
|
sizeof(*mldq), 0));
|
|
|
|
skb_put(skb, sizeof(*mldq));
|
|
|
|
|
|
|
|
__skb_pull(skb, sizeof(*eth));
|
|
|
|
|
|
|
|
out:
|
|
|
|
return skb;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
static struct sk_buff *br_multicast_alloc_query(struct net_bridge *br,
|
|
|
|
struct br_ip *addr)
|
|
|
|
{
|
|
|
|
switch (addr->proto) {
|
|
|
|
case htons(ETH_P_IP):
|
|
|
|
return br_ip4_multicast_alloc_query(br, addr->u.ip4);
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2010-04-22 23:54:22 +07:00
|
|
|
case htons(ETH_P_IPV6):
|
|
|
|
return br_ip6_multicast_alloc_query(br, &addr->u.ip6);
|
|
|
|
#endif
|
2010-04-18 10:42:07 +07:00
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
static struct net_bridge_mdb_entry *br_multicast_get_group(
|
2010-04-18 10:42:07 +07:00
|
|
|
struct net_bridge *br, struct net_bridge_port *port,
|
|
|
|
struct br_ip *group, int hash)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
2010-11-15 13:38:10 +07:00
|
|
|
struct net_bridge_mdb_htable *mdb;
|
2010-02-28 02:41:45 +07:00
|
|
|
struct net_bridge_mdb_entry *mp;
|
2012-04-15 12:58:06 +07:00
|
|
|
unsigned int count = 0;
|
|
|
|
unsigned int max;
|
2010-02-28 02:41:45 +07:00
|
|
|
int elasticity;
|
|
|
|
int err;
|
|
|
|
|
2010-11-15 13:38:10 +07:00
|
|
|
mdb = rcu_dereference_protected(br->mdb, 1);
|
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 08:06:00 +07:00
|
|
|
hlist_for_each_entry(mp, &mdb->mhash[hash], hlist[mdb->ver]) {
|
2010-02-28 02:41:45 +07:00
|
|
|
count++;
|
2010-04-18 10:42:07 +07:00
|
|
|
if (unlikely(br_ip_equal(group, &mp->addr)))
|
2010-02-28 02:41:45 +07:00
|
|
|
return mp;
|
|
|
|
}
|
|
|
|
|
|
|
|
elasticity = 0;
|
|
|
|
max = mdb->max;
|
|
|
|
|
|
|
|
if (unlikely(count > br->hash_elasticity && count)) {
|
|
|
|
if (net_ratelimit())
|
2010-05-10 16:31:09 +07:00
|
|
|
br_info(br, "Multicast hash table "
|
|
|
|
"chain limit reached: %s\n",
|
|
|
|
port ? port->dev->name : br->dev->name);
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
elasticity = br->hash_elasticity;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (mdb->size >= max) {
|
|
|
|
max *= 2;
|
2012-07-11 05:29:19 +07:00
|
|
|
if (unlikely(max > br->hash_max)) {
|
|
|
|
br_warn(br, "Multicast hash table maximum of %d "
|
|
|
|
"reached, disabling snooping: %s\n",
|
|
|
|
br->hash_max,
|
|
|
|
port ? port->dev->name : br->dev->name);
|
2010-02-28 02:41:45 +07:00
|
|
|
err = -E2BIG;
|
|
|
|
disable:
|
|
|
|
br->multicast_disabled = 1;
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (max > mdb->max || elasticity) {
|
|
|
|
if (mdb->old) {
|
|
|
|
if (net_ratelimit())
|
2010-05-10 16:31:09 +07:00
|
|
|
br_info(br, "Multicast hash table "
|
|
|
|
"on fire: %s\n",
|
|
|
|
port ? port->dev->name : br->dev->name);
|
2010-02-28 02:41:45 +07:00
|
|
|
err = -EEXIST;
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
|
|
|
|
err = br_mdb_rehash(&br->mdb, max, elasticity);
|
|
|
|
if (err) {
|
2010-05-10 16:31:09 +07:00
|
|
|
br_warn(br, "Cannot rehash multicast "
|
|
|
|
"hash table, disabling snooping: %s, %d, %d\n",
|
|
|
|
port ? port->dev->name : br->dev->name,
|
|
|
|
mdb->size, err);
|
2010-02-28 02:41:45 +07:00
|
|
|
goto disable;
|
|
|
|
}
|
|
|
|
|
|
|
|
err = -EAGAIN;
|
|
|
|
goto err;
|
|
|
|
}
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
err:
|
|
|
|
mp = ERR_PTR(err);
|
|
|
|
return mp;
|
|
|
|
}
|
|
|
|
|
2012-12-12 05:23:08 +07:00
|
|
|
struct net_bridge_mdb_entry *br_multicast_new_group(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port, struct br_ip *group)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
2010-11-15 13:38:10 +07:00
|
|
|
struct net_bridge_mdb_htable *mdb;
|
2010-02-28 02:41:45 +07:00
|
|
|
struct net_bridge_mdb_entry *mp;
|
|
|
|
int hash;
|
2010-12-10 10:18:04 +07:00
|
|
|
int err;
|
2010-02-28 02:41:45 +07:00
|
|
|
|
2010-11-15 13:38:10 +07:00
|
|
|
mdb = rcu_dereference_protected(br->mdb, 1);
|
2010-02-28 02:41:45 +07:00
|
|
|
if (!mdb) {
|
2010-12-10 10:18:04 +07:00
|
|
|
err = br_mdb_rehash(&br->mdb, BR_HASH_SIZE, 0);
|
|
|
|
if (err)
|
|
|
|
return ERR_PTR(err);
|
2010-02-28 02:41:45 +07:00
|
|
|
goto rehash;
|
|
|
|
}
|
|
|
|
|
|
|
|
hash = br_ip_hash(mdb, group);
|
|
|
|
mp = br_multicast_get_group(br, port, group, hash);
|
|
|
|
switch (PTR_ERR(mp)) {
|
|
|
|
case 0:
|
|
|
|
break;
|
|
|
|
|
|
|
|
case -EAGAIN:
|
|
|
|
rehash:
|
2010-11-15 13:38:10 +07:00
|
|
|
mdb = rcu_dereference_protected(br->mdb, 1);
|
2010-02-28 02:41:45 +07:00
|
|
|
hash = br_ip_hash(mdb, group);
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
mp = kzalloc(sizeof(*mp), GFP_ATOMIC);
|
|
|
|
if (unlikely(!mp))
|
2010-12-10 10:18:04 +07:00
|
|
|
return ERR_PTR(-ENOMEM);
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
mp->br = br;
|
2010-04-18 10:42:07 +07:00
|
|
|
mp->addr = *group;
|
2013-07-20 10:07:16 +07:00
|
|
|
setup_timer(&mp->timer, br_multicast_group_expired,
|
|
|
|
(unsigned long)mp);
|
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
hlist_add_head_rcu(&mp->hlist[mdb->ver], &mdb->mhash[hash]);
|
|
|
|
mdb->size++;
|
|
|
|
|
|
|
|
out:
|
|
|
|
return mp;
|
|
|
|
}
|
|
|
|
|
2012-12-12 05:23:08 +07:00
|
|
|
struct net_bridge_port_group *br_multicast_new_port_group(
|
|
|
|
struct net_bridge_port *port,
|
|
|
|
struct br_ip *group,
|
2012-12-15 05:09:51 +07:00
|
|
|
struct net_bridge_port_group __rcu *next,
|
|
|
|
unsigned char state)
|
2012-12-12 05:23:08 +07:00
|
|
|
{
|
|
|
|
struct net_bridge_port_group *p;
|
|
|
|
|
|
|
|
p = kzalloc(sizeof(*p), GFP_ATOMIC);
|
|
|
|
if (unlikely(!p))
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
p->addr = *group;
|
|
|
|
p->port = port;
|
2012-12-15 05:09:51 +07:00
|
|
|
p->state = state;
|
2012-12-13 13:51:28 +07:00
|
|
|
rcu_assign_pointer(p->next, next);
|
2012-12-12 05:23:08 +07:00
|
|
|
hlist_add_head(&p->mglist, &port->mglist);
|
|
|
|
setup_timer(&p->timer, br_multicast_port_group_expired,
|
|
|
|
(unsigned long)p);
|
|
|
|
return p;
|
|
|
|
}
|
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
static int br_multicast_add_group(struct net_bridge *br,
|
2010-04-18 10:42:07 +07:00
|
|
|
struct net_bridge_port *port,
|
|
|
|
struct br_ip *group)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
|
|
|
struct net_bridge_mdb_entry *mp;
|
|
|
|
struct net_bridge_port_group *p;
|
2010-11-15 13:38:10 +07:00
|
|
|
struct net_bridge_port_group __rcu **pp;
|
2013-10-20 05:58:57 +07:00
|
|
|
unsigned long now = jiffies;
|
2010-02-28 02:41:45 +07:00
|
|
|
int err;
|
|
|
|
|
|
|
|
spin_lock(&br->multicast_lock);
|
|
|
|
if (!netif_running(br->dev) ||
|
|
|
|
(port && port->state == BR_STATE_DISABLED))
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
mp = br_multicast_new_group(br, port, group);
|
|
|
|
err = PTR_ERR(mp);
|
2010-12-10 10:18:04 +07:00
|
|
|
if (IS_ERR(mp))
|
2010-02-28 02:41:45 +07:00
|
|
|
goto err;
|
|
|
|
|
|
|
|
if (!port) {
|
2011-02-12 16:05:42 +07:00
|
|
|
mp->mglist = true;
|
2013-10-20 05:58:57 +07:00
|
|
|
mod_timer(&mp->timer, now + br->multicast_membership_interval);
|
2010-02-28 02:41:45 +07:00
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2010-11-15 13:38:10 +07:00
|
|
|
for (pp = &mp->ports;
|
|
|
|
(p = mlock_dereference(*pp, br)) != NULL;
|
|
|
|
pp = &p->next) {
|
2010-02-28 02:41:45 +07:00
|
|
|
if (p->port == port)
|
2013-10-20 05:58:57 +07:00
|
|
|
goto found;
|
2010-02-28 02:41:45 +07:00
|
|
|
if ((unsigned long)p->port < (unsigned long)port)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2012-12-15 05:09:51 +07:00
|
|
|
p = br_multicast_new_port_group(port, group, *pp, MDB_TEMPORARY);
|
2010-02-28 02:41:45 +07:00
|
|
|
if (unlikely(!p))
|
|
|
|
goto err;
|
|
|
|
rcu_assign_pointer(*pp, p);
|
2012-12-12 05:23:07 +07:00
|
|
|
br_mdb_notify(br->dev, port, group, RTM_NEWMDB);
|
2010-02-28 02:41:45 +07:00
|
|
|
|
2013-10-20 05:58:57 +07:00
|
|
|
found:
|
|
|
|
mod_timer(&p->timer, now + br->multicast_membership_interval);
|
2010-02-28 02:41:45 +07:00
|
|
|
out:
|
|
|
|
err = 0;
|
|
|
|
|
|
|
|
err:
|
|
|
|
spin_unlock(&br->multicast_lock);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
static int br_ip4_multicast_add_group(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port,
|
2013-02-13 19:00:17 +07:00
|
|
|
__be32 group,
|
|
|
|
__u16 vid)
|
2010-04-18 10:42:07 +07:00
|
|
|
{
|
|
|
|
struct br_ip br_group;
|
|
|
|
|
|
|
|
if (ipv4_is_local_multicast(group))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
br_group.u.ip4 = group;
|
|
|
|
br_group.proto = htons(ETH_P_IP);
|
2013-02-13 19:00:17 +07:00
|
|
|
br_group.vid = vid;
|
2010-04-18 10:42:07 +07:00
|
|
|
|
|
|
|
return br_multicast_add_group(br, port, &br_group);
|
|
|
|
}
|
|
|
|
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2010-04-22 23:54:22 +07:00
|
|
|
static int br_ip6_multicast_add_group(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port,
|
2013-02-13 19:00:17 +07:00
|
|
|
const struct in6_addr *group,
|
|
|
|
__u16 vid)
|
2010-04-22 23:54:22 +07:00
|
|
|
{
|
|
|
|
struct br_ip br_group;
|
|
|
|
|
2013-09-04 07:13:39 +07:00
|
|
|
if (ipv6_addr_is_ll_all_nodes(group))
|
2010-04-22 23:54:22 +07:00
|
|
|
return 0;
|
|
|
|
|
2011-11-21 10:39:03 +07:00
|
|
|
br_group.u.ip6 = *group;
|
2011-02-15 20:19:17 +07:00
|
|
|
br_group.proto = htons(ETH_P_IPV6);
|
2013-02-13 19:00:17 +07:00
|
|
|
br_group.vid = vid;
|
2010-04-22 23:54:22 +07:00
|
|
|
|
|
|
|
return br_multicast_add_group(br, port, &br_group);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
static void br_multicast_router_expired(unsigned long data)
|
|
|
|
{
|
|
|
|
struct net_bridge_port *port = (void *)data;
|
|
|
|
struct net_bridge *br = port->br;
|
|
|
|
|
|
|
|
spin_lock(&br->multicast_lock);
|
|
|
|
if (port->multicast_router != 1 ||
|
|
|
|
timer_pending(&port->multicast_router_timer) ||
|
|
|
|
hlist_unhashed(&port->rlist))
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
hlist_del_init_rcu(&port->rlist);
|
|
|
|
|
|
|
|
out:
|
|
|
|
spin_unlock(&br->multicast_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void br_multicast_local_router_expired(unsigned long data)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
static void br_multicast_querier_expired(struct net_bridge *br,
|
|
|
|
struct bridge_mcast_query *query)
|
2012-04-13 09:37:42 +07:00
|
|
|
{
|
|
|
|
spin_lock(&br->multicast_lock);
|
|
|
|
if (!netif_running(br->dev) || br->multicast_disabled)
|
|
|
|
goto out;
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
br_multicast_start_querier(br, query);
|
2012-04-13 09:37:42 +07:00
|
|
|
|
|
|
|
out:
|
|
|
|
spin_unlock(&br->multicast_lock);
|
|
|
|
}
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
static void br_ip4_multicast_querier_expired(unsigned long data)
|
|
|
|
{
|
|
|
|
struct net_bridge *br = (void *)data;
|
|
|
|
|
|
|
|
br_multicast_querier_expired(br, &br->ip4_query);
|
|
|
|
}
|
|
|
|
|
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
|
|
static void br_ip6_multicast_querier_expired(unsigned long data)
|
|
|
|
{
|
|
|
|
struct net_bridge *br = (void *)data;
|
|
|
|
|
|
|
|
br_multicast_querier_expired(br, &br->ip6_query);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
static void __br_multicast_send_query(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port,
|
|
|
|
struct br_ip *ip)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
|
|
|
struct sk_buff *skb;
|
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
skb = br_multicast_alloc_query(br, ip);
|
2010-02-28 02:41:45 +07:00
|
|
|
if (!skb)
|
2010-04-18 10:42:07 +07:00
|
|
|
return;
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
if (port) {
|
|
|
|
__skb_push(skb, sizeof(struct ethhdr));
|
|
|
|
skb->dev = port->dev;
|
2010-03-23 10:07:21 +07:00
|
|
|
NF_HOOK(NFPROTO_BRIDGE, NF_BR_LOCAL_OUT, skb, NULL, skb->dev,
|
2010-02-28 02:41:45 +07:00
|
|
|
dev_queue_xmit);
|
|
|
|
} else
|
|
|
|
netif_rx(skb);
|
2010-04-18 10:42:07 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static void br_multicast_send_query(struct net_bridge *br,
|
2013-08-30 22:28:17 +07:00
|
|
|
struct net_bridge_port *port,
|
|
|
|
struct bridge_mcast_query *query)
|
2010-04-18 10:42:07 +07:00
|
|
|
{
|
|
|
|
unsigned long time;
|
|
|
|
struct br_ip br_group;
|
2013-08-30 22:28:17 +07:00
|
|
|
struct bridge_mcast_querier *querier = NULL;
|
2010-04-18 10:42:07 +07:00
|
|
|
|
|
|
|
if (!netif_running(br->dev) || br->multicast_disabled ||
|
2013-08-30 22:28:17 +07:00
|
|
|
!br->multicast_querier)
|
2010-04-18 10:42:07 +07:00
|
|
|
return;
|
|
|
|
|
2010-04-22 23:54:22 +07:00
|
|
|
memset(&br_group.u, 0, sizeof(br_group.u));
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
if (port ? (query == &port->ip4_query) :
|
|
|
|
(query == &br->ip4_query)) {
|
|
|
|
querier = &br->ip4_querier;
|
|
|
|
br_group.proto = htons(ETH_P_IP);
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2013-08-30 22:28:17 +07:00
|
|
|
} else {
|
|
|
|
querier = &br->ip6_querier;
|
|
|
|
br_group.proto = htons(ETH_P_IPV6);
|
2010-04-22 23:54:22 +07:00
|
|
|
#endif
|
2013-08-30 22:28:17 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
if (!querier || timer_pending(&querier->timer))
|
|
|
|
return;
|
|
|
|
|
|
|
|
__br_multicast_send_query(br, port, &br_group);
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
time = jiffies;
|
2013-08-30 22:28:17 +07:00
|
|
|
time += query->startup_sent < br->multicast_startup_query_count ?
|
2010-02-28 02:41:45 +07:00
|
|
|
br->multicast_startup_query_interval :
|
|
|
|
br->multicast_query_interval;
|
2013-08-30 22:28:17 +07:00
|
|
|
mod_timer(&query->timer, time);
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
static void br_multicast_port_query_expired(struct net_bridge_port *port,
|
|
|
|
struct bridge_mcast_query *query)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
|
|
|
struct net_bridge *br = port->br;
|
|
|
|
|
|
|
|
spin_lock(&br->multicast_lock);
|
2010-03-06 08:14:09 +07:00
|
|
|
if (port->state == BR_STATE_DISABLED ||
|
|
|
|
port->state == BR_STATE_BLOCKING)
|
2010-02-28 02:41:45 +07:00
|
|
|
goto out;
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
if (query->startup_sent < br->multicast_startup_query_count)
|
|
|
|
query->startup_sent++;
|
2010-02-28 02:41:45 +07:00
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
br_multicast_send_query(port->br, port, query);
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
out:
|
|
|
|
spin_unlock(&br->multicast_lock);
|
|
|
|
}
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
static void br_ip4_multicast_port_query_expired(unsigned long data)
|
|
|
|
{
|
|
|
|
struct net_bridge_port *port = (void *)data;
|
|
|
|
|
|
|
|
br_multicast_port_query_expired(port, &port->ip4_query);
|
|
|
|
}
|
|
|
|
|
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
|
|
static void br_ip6_multicast_port_query_expired(unsigned long data)
|
|
|
|
{
|
|
|
|
struct net_bridge_port *port = (void *)data;
|
|
|
|
|
|
|
|
br_multicast_port_query_expired(port, &port->ip6_query);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
void br_multicast_add_port(struct net_bridge_port *port)
|
|
|
|
{
|
|
|
|
port->multicast_router = 1;
|
|
|
|
|
|
|
|
setup_timer(&port->multicast_router_timer, br_multicast_router_expired,
|
|
|
|
(unsigned long)port);
|
2013-08-30 22:28:17 +07:00
|
|
|
setup_timer(&port->ip4_query.timer, br_ip4_multicast_port_query_expired,
|
|
|
|
(unsigned long)port);
|
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
|
|
setup_timer(&port->ip6_query.timer, br_ip6_multicast_port_query_expired,
|
|
|
|
(unsigned long)port);
|
|
|
|
#endif
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
void br_multicast_del_port(struct net_bridge_port *port)
|
|
|
|
{
|
|
|
|
del_timer_sync(&port->multicast_router_timer);
|
|
|
|
}
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
static void br_multicast_enable(struct bridge_mcast_query *query)
|
2010-02-28 02:41:50 +07:00
|
|
|
{
|
2013-08-30 22:28:17 +07:00
|
|
|
query->startup_sent = 0;
|
2010-02-28 02:41:50 +07:00
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
if (try_to_del_timer_sync(&query->timer) >= 0 ||
|
|
|
|
del_timer(&query->timer))
|
|
|
|
mod_timer(&query->timer, jiffies);
|
2010-02-28 02:41:50 +07:00
|
|
|
}
|
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
void br_multicast_enable_port(struct net_bridge_port *port)
|
|
|
|
{
|
|
|
|
struct net_bridge *br = port->br;
|
|
|
|
|
|
|
|
spin_lock(&br->multicast_lock);
|
|
|
|
if (br->multicast_disabled || !netif_running(br->dev))
|
|
|
|
goto out;
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
br_multicast_enable(&port->ip4_query);
|
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
|
|
br_multicast_enable(&port->ip6_query);
|
|
|
|
#endif
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
out:
|
|
|
|
spin_unlock(&br->multicast_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
void br_multicast_disable_port(struct net_bridge_port *port)
|
|
|
|
{
|
|
|
|
struct net_bridge *br = port->br;
|
|
|
|
struct net_bridge_port_group *pg;
|
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 08:06:00 +07:00
|
|
|
struct hlist_node *n;
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
spin_lock(&br->multicast_lock);
|
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 08:06:00 +07:00
|
|
|
hlist_for_each_entry_safe(pg, n, &port->mglist, mglist)
|
2010-02-28 02:41:45 +07:00
|
|
|
br_multicast_del_pg(br, pg);
|
|
|
|
|
|
|
|
if (!hlist_unhashed(&port->rlist))
|
|
|
|
hlist_del_init_rcu(&port->rlist);
|
|
|
|
del_timer(&port->multicast_router_timer);
|
2013-08-30 22:28:17 +07:00
|
|
|
del_timer(&port->ip4_query.timer);
|
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
|
|
del_timer(&port->ip6_query.timer);
|
|
|
|
#endif
|
2010-02-28 02:41:45 +07:00
|
|
|
spin_unlock(&br->multicast_lock);
|
|
|
|
}
|
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
static int br_ip4_multicast_igmp3_report(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port,
|
2013-10-29 02:45:07 +07:00
|
|
|
struct sk_buff *skb,
|
|
|
|
u16 vid)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
|
|
|
struct igmpv3_report *ih;
|
|
|
|
struct igmpv3_grec *grec;
|
|
|
|
int i;
|
|
|
|
int len;
|
|
|
|
int num;
|
|
|
|
int type;
|
|
|
|
int err = 0;
|
|
|
|
__be32 group;
|
|
|
|
|
|
|
|
if (!pskb_may_pull(skb, sizeof(*ih)))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
ih = igmpv3_report_hdr(skb);
|
|
|
|
num = ntohs(ih->ngrec);
|
|
|
|
len = sizeof(*ih);
|
|
|
|
|
|
|
|
for (i = 0; i < num; i++) {
|
|
|
|
len += sizeof(*grec);
|
|
|
|
if (!pskb_may_pull(skb, len))
|
|
|
|
return -EINVAL;
|
|
|
|
|
2010-04-08 11:20:47 +07:00
|
|
|
grec = (void *)(skb->data + len - sizeof(*grec));
|
2010-02-28 02:41:45 +07:00
|
|
|
group = grec->grec_mca;
|
|
|
|
type = grec->grec_type;
|
|
|
|
|
2010-04-20 10:20:05 +07:00
|
|
|
len += ntohs(grec->grec_nsrcs) * 4;
|
2010-02-28 02:41:45 +07:00
|
|
|
if (!pskb_may_pull(skb, len))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
/* We treat this as an IGMPv2 report for now. */
|
|
|
|
switch (type) {
|
|
|
|
case IGMPV3_MODE_IS_INCLUDE:
|
|
|
|
case IGMPV3_MODE_IS_EXCLUDE:
|
|
|
|
case IGMPV3_CHANGE_TO_INCLUDE:
|
|
|
|
case IGMPV3_CHANGE_TO_EXCLUDE:
|
|
|
|
case IGMPV3_ALLOW_NEW_SOURCES:
|
|
|
|
case IGMPV3_BLOCK_OLD_SOURCES:
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
2013-02-13 19:00:17 +07:00
|
|
|
err = br_ip4_multicast_add_group(br, port, group, vid);
|
2010-02-28 02:41:45 +07:00
|
|
|
if (err)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2010-04-22 23:54:22 +07:00
|
|
|
static int br_ip6_multicast_mld2_report(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port,
|
2013-10-29 02:45:07 +07:00
|
|
|
struct sk_buff *skb,
|
|
|
|
u16 vid)
|
2010-04-22 23:54:22 +07:00
|
|
|
{
|
|
|
|
struct icmp6hdr *icmp6h;
|
|
|
|
struct mld2_grec *grec;
|
|
|
|
int i;
|
|
|
|
int len;
|
|
|
|
int num;
|
|
|
|
int err = 0;
|
|
|
|
|
|
|
|
if (!pskb_may_pull(skb, sizeof(*icmp6h)))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
icmp6h = icmp6_hdr(skb);
|
|
|
|
num = ntohs(icmp6h->icmp6_dataun.un_data16[1]);
|
|
|
|
len = sizeof(*icmp6h);
|
|
|
|
|
|
|
|
for (i = 0; i < num; i++) {
|
|
|
|
__be16 *nsrcs, _nsrcs;
|
|
|
|
|
|
|
|
nsrcs = skb_header_pointer(skb,
|
|
|
|
len + offsetof(struct mld2_grec,
|
2011-02-15 20:19:18 +07:00
|
|
|
grec_nsrcs),
|
2010-04-22 23:54:22 +07:00
|
|
|
sizeof(_nsrcs), &_nsrcs);
|
|
|
|
if (!nsrcs)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (!pskb_may_pull(skb,
|
|
|
|
len + sizeof(*grec) +
|
2011-02-15 20:19:19 +07:00
|
|
|
sizeof(struct in6_addr) * ntohs(*nsrcs)))
|
2010-04-22 23:54:22 +07:00
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
grec = (struct mld2_grec *)(skb->data + len);
|
2011-02-15 20:19:19 +07:00
|
|
|
len += sizeof(*grec) +
|
|
|
|
sizeof(struct in6_addr) * ntohs(*nsrcs);
|
2010-04-22 23:54:22 +07:00
|
|
|
|
|
|
|
/* We treat these as MLDv1 reports for now. */
|
|
|
|
switch (grec->grec_type) {
|
|
|
|
case MLD2_MODE_IS_INCLUDE:
|
|
|
|
case MLD2_MODE_IS_EXCLUDE:
|
|
|
|
case MLD2_CHANGE_TO_INCLUDE:
|
|
|
|
case MLD2_CHANGE_TO_EXCLUDE:
|
|
|
|
case MLD2_ALLOW_NEW_SOURCES:
|
|
|
|
case MLD2_BLOCK_OLD_SOURCES:
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
2013-02-13 19:00:17 +07:00
|
|
|
err = br_ip6_multicast_add_group(br, port, &grec->grec_mca,
|
|
|
|
vid);
|
2010-04-22 23:54:22 +07:00
|
|
|
if (!err)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
static void
|
|
|
|
br_multicast_update_querier_timer(struct net_bridge *br,
|
|
|
|
struct bridge_mcast_querier *querier,
|
|
|
|
unsigned long max_delay)
|
2013-08-01 06:06:20 +07:00
|
|
|
{
|
2013-08-30 22:28:17 +07:00
|
|
|
if (!timer_pending(&querier->timer))
|
|
|
|
querier->delay_time = jiffies + max_delay;
|
2013-08-01 06:06:20 +07:00
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
mod_timer(&querier->timer, jiffies + br->multicast_querier_interval);
|
2013-08-01 06:06:20 +07:00
|
|
|
}
|
|
|
|
|
2010-04-27 22:01:04 +07:00
|
|
|
/*
|
2013-06-21 14:37:25 +07:00
|
|
|
* Add port to router_list
|
2010-04-27 22:01:04 +07:00
|
|
|
* list is maintained ordered by pointer value
|
|
|
|
* and locked by br->multicast_lock and RCU
|
|
|
|
*/
|
2010-02-28 02:41:49 +07:00
|
|
|
static void br_multicast_add_router(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port)
|
|
|
|
{
|
2010-04-27 14:13:11 +07:00
|
|
|
struct net_bridge_port *p;
|
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 08:06:00 +07:00
|
|
|
struct hlist_node *slot = NULL;
|
2010-04-27 14:13:11 +07:00
|
|
|
|
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 08:06:00 +07:00
|
|
|
hlist_for_each_entry(p, &br->router_list, rlist) {
|
2010-04-27 22:01:04 +07:00
|
|
|
if ((unsigned long) port >= (unsigned long) p)
|
|
|
|
break;
|
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 08:06:00 +07:00
|
|
|
slot = &p->rlist;
|
2010-04-27 14:13:11 +07:00
|
|
|
}
|
|
|
|
|
2010-04-27 22:01:04 +07:00
|
|
|
if (slot)
|
|
|
|
hlist_add_after_rcu(slot, &port->rlist);
|
2010-04-27 14:13:11 +07:00
|
|
|
else
|
|
|
|
hlist_add_head_rcu(&port->rlist, &br->router_list);
|
2010-02-28 02:41:49 +07:00
|
|
|
}
|
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
static void br_multicast_mark_router(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port)
|
|
|
|
{
|
|
|
|
unsigned long now = jiffies;
|
|
|
|
|
|
|
|
if (!port) {
|
|
|
|
if (br->multicast_router == 1)
|
|
|
|
mod_timer(&br->multicast_router_timer,
|
|
|
|
now + br->multicast_querier_interval);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (port->multicast_router != 1)
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (!hlist_unhashed(&port->rlist))
|
|
|
|
goto timer;
|
|
|
|
|
2010-02-28 02:41:49 +07:00
|
|
|
br_multicast_add_router(br, port);
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
timer:
|
|
|
|
mod_timer(&port->multicast_router_timer,
|
|
|
|
now + br->multicast_querier_interval);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void br_multicast_query_received(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port,
|
2013-08-30 22:28:17 +07:00
|
|
|
struct bridge_mcast_querier *querier,
|
2013-08-01 06:06:20 +07:00
|
|
|
int saddr,
|
|
|
|
unsigned long max_delay)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
|
|
|
if (saddr)
|
2013-08-30 22:28:17 +07:00
|
|
|
br_multicast_update_querier_timer(br, querier, max_delay);
|
|
|
|
else if (timer_pending(&querier->timer))
|
2010-02-28 02:41:45 +07:00
|
|
|
return;
|
|
|
|
|
|
|
|
br_multicast_mark_router(br, port);
|
|
|
|
}
|
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
static int br_ip4_multicast_query(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port,
|
2013-10-29 02:45:07 +07:00
|
|
|
struct sk_buff *skb,
|
|
|
|
u16 vid)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
2011-04-22 11:53:02 +07:00
|
|
|
const struct iphdr *iph = ip_hdr(skb);
|
2010-02-28 02:41:45 +07:00
|
|
|
struct igmphdr *ih = igmp_hdr(skb);
|
|
|
|
struct net_bridge_mdb_entry *mp;
|
|
|
|
struct igmpv3_query *ih3;
|
|
|
|
struct net_bridge_port_group *p;
|
2010-11-15 13:38:10 +07:00
|
|
|
struct net_bridge_port_group __rcu **pp;
|
2010-02-28 02:41:45 +07:00
|
|
|
unsigned long max_delay;
|
|
|
|
unsigned long now = jiffies;
|
|
|
|
__be32 group;
|
2010-03-14 03:27:21 +07:00
|
|
|
int err = 0;
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
spin_lock(&br->multicast_lock);
|
|
|
|
if (!netif_running(br->dev) ||
|
|
|
|
(port && port->state == BR_STATE_DISABLED))
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
group = ih->group;
|
|
|
|
|
|
|
|
if (skb->len == sizeof(*ih)) {
|
|
|
|
max_delay = ih->code * (HZ / IGMP_TIMER_SCALE);
|
|
|
|
|
|
|
|
if (!max_delay) {
|
|
|
|
max_delay = 10 * HZ;
|
|
|
|
group = 0;
|
|
|
|
}
|
|
|
|
} else {
|
2010-03-14 03:27:21 +07:00
|
|
|
if (!pskb_may_pull(skb, sizeof(struct igmpv3_query))) {
|
|
|
|
err = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
ih3 = igmpv3_query_hdr(skb);
|
|
|
|
if (ih3->nsrcs)
|
2010-03-14 03:27:21 +07:00
|
|
|
goto out;
|
2010-02-28 02:41:45 +07:00
|
|
|
|
2010-03-16 02:27:00 +07:00
|
|
|
max_delay = ih3->code ?
|
|
|
|
IGMPV3_MRC(ih3->code) * (HZ / IGMP_TIMER_SCALE) : 1;
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
br_multicast_query_received(br, port, &br->ip4_querier, !!iph->saddr,
|
|
|
|
max_delay);
|
2013-08-01 06:06:20 +07:00
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
if (!group)
|
|
|
|
goto out;
|
|
|
|
|
2013-02-13 19:00:17 +07:00
|
|
|
mp = br_mdb_ip4_get(mlock_dereference(br->mdb, br), group, vid);
|
2010-02-28 02:41:45 +07:00
|
|
|
if (!mp)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
max_delay *= br->multicast_last_member_count;
|
|
|
|
|
2011-02-12 16:05:42 +07:00
|
|
|
if (mp->mglist &&
|
2010-02-28 02:41:45 +07:00
|
|
|
(timer_pending(&mp->timer) ?
|
|
|
|
time_after(mp->timer.expires, now + max_delay) :
|
|
|
|
try_to_del_timer_sync(&mp->timer) >= 0))
|
|
|
|
mod_timer(&mp->timer, now + max_delay);
|
|
|
|
|
2010-11-15 13:38:10 +07:00
|
|
|
for (pp = &mp->ports;
|
|
|
|
(p = mlock_dereference(*pp, br)) != NULL;
|
|
|
|
pp = &p->next) {
|
2010-02-28 02:41:45 +07:00
|
|
|
if (timer_pending(&p->timer) ?
|
|
|
|
time_after(p->timer.expires, now + max_delay) :
|
|
|
|
try_to_del_timer_sync(&p->timer) >= 0)
|
2011-02-11 19:42:07 +07:00
|
|
|
mod_timer(&p->timer, now + max_delay);
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
spin_unlock(&br->multicast_lock);
|
2010-03-14 03:27:21 +07:00
|
|
|
return err;
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2010-04-22 23:54:22 +07:00
|
|
|
static int br_ip6_multicast_query(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port,
|
2013-10-29 02:45:07 +07:00
|
|
|
struct sk_buff *skb,
|
|
|
|
u16 vid)
|
2010-04-22 23:54:22 +07:00
|
|
|
{
|
2011-04-22 11:53:02 +07:00
|
|
|
const struct ipv6hdr *ip6h = ipv6_hdr(skb);
|
2012-12-13 13:51:28 +07:00
|
|
|
struct mld_msg *mld;
|
2010-04-22 23:54:22 +07:00
|
|
|
struct net_bridge_mdb_entry *mp;
|
|
|
|
struct mld2_query *mld2q;
|
2010-11-15 13:38:10 +07:00
|
|
|
struct net_bridge_port_group *p;
|
|
|
|
struct net_bridge_port_group __rcu **pp;
|
2010-04-22 23:54:22 +07:00
|
|
|
unsigned long max_delay;
|
|
|
|
unsigned long now = jiffies;
|
2011-04-22 11:53:02 +07:00
|
|
|
const struct in6_addr *group = NULL;
|
2010-04-22 23:54:22 +07:00
|
|
|
int err = 0;
|
|
|
|
|
|
|
|
spin_lock(&br->multicast_lock);
|
|
|
|
if (!netif_running(br->dev) ||
|
|
|
|
(port && port->state == BR_STATE_DISABLED))
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
if (skb->len == sizeof(*mld)) {
|
|
|
|
if (!pskb_may_pull(skb, sizeof(*mld))) {
|
|
|
|
err = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
mld = (struct mld_msg *) icmp6_hdr(skb);
|
2012-07-10 06:56:12 +07:00
|
|
|
max_delay = msecs_to_jiffies(ntohs(mld->mld_maxdelay));
|
2010-04-22 23:54:22 +07:00
|
|
|
if (max_delay)
|
|
|
|
group = &mld->mld_mca;
|
2013-08-06 05:32:05 +07:00
|
|
|
} else {
|
2010-04-22 23:54:22 +07:00
|
|
|
if (!pskb_may_pull(skb, sizeof(*mld2q))) {
|
|
|
|
err = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
mld2q = (struct mld2_query *)icmp6_hdr(skb);
|
|
|
|
if (!mld2q->mld2q_nsrcs)
|
|
|
|
group = &mld2q->mld2q_mca;
|
2013-09-04 05:19:39 +07:00
|
|
|
|
|
|
|
max_delay = max(msecs_to_jiffies(mldv2_mrc(mld2q)), 1UL);
|
2010-04-22 23:54:22 +07:00
|
|
|
}
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
br_multicast_query_received(br, port, &br->ip6_querier,
|
|
|
|
!ipv6_addr_any(&ip6h->saddr), max_delay);
|
2013-08-01 06:06:20 +07:00
|
|
|
|
2010-04-22 23:54:22 +07:00
|
|
|
if (!group)
|
|
|
|
goto out;
|
|
|
|
|
2013-02-13 19:00:17 +07:00
|
|
|
mp = br_mdb_ip6_get(mlock_dereference(br->mdb, br), group, vid);
|
2010-04-22 23:54:22 +07:00
|
|
|
if (!mp)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
max_delay *= br->multicast_last_member_count;
|
2011-02-12 16:05:42 +07:00
|
|
|
if (mp->mglist &&
|
2010-04-22 23:54:22 +07:00
|
|
|
(timer_pending(&mp->timer) ?
|
|
|
|
time_after(mp->timer.expires, now + max_delay) :
|
|
|
|
try_to_del_timer_sync(&mp->timer) >= 0))
|
|
|
|
mod_timer(&mp->timer, now + max_delay);
|
|
|
|
|
2010-11-15 13:38:10 +07:00
|
|
|
for (pp = &mp->ports;
|
|
|
|
(p = mlock_dereference(*pp, br)) != NULL;
|
|
|
|
pp = &p->next) {
|
2010-04-22 23:54:22 +07:00
|
|
|
if (timer_pending(&p->timer) ?
|
|
|
|
time_after(p->timer.expires, now + max_delay) :
|
|
|
|
try_to_del_timer_sync(&p->timer) >= 0)
|
2011-02-11 19:42:07 +07:00
|
|
|
mod_timer(&p->timer, now + max_delay);
|
2010-04-22 23:54:22 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
spin_unlock(&br->multicast_lock);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
static void br_multicast_leave_group(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port,
|
2013-08-30 22:28:17 +07:00
|
|
|
struct br_ip *group,
|
|
|
|
struct bridge_mcast_querier *querier,
|
|
|
|
struct bridge_mcast_query *query)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
|
|
|
struct net_bridge_mdb_htable *mdb;
|
|
|
|
struct net_bridge_mdb_entry *mp;
|
|
|
|
struct net_bridge_port_group *p;
|
|
|
|
unsigned long now;
|
|
|
|
unsigned long time;
|
|
|
|
|
|
|
|
spin_lock(&br->multicast_lock);
|
|
|
|
if (!netif_running(br->dev) ||
|
|
|
|
(port && port->state == BR_STATE_DISABLED) ||
|
2013-08-30 22:28:17 +07:00
|
|
|
timer_pending(&querier->timer))
|
2010-02-28 02:41:45 +07:00
|
|
|
goto out;
|
|
|
|
|
2010-11-15 13:38:10 +07:00
|
|
|
mdb = mlock_dereference(br->mdb, br);
|
2010-02-28 02:41:45 +07:00
|
|
|
mp = br_mdb_ip_get(mdb, group);
|
|
|
|
if (!mp)
|
|
|
|
goto out;
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
if (br->multicast_querier) {
|
2013-05-22 04:52:56 +07:00
|
|
|
__br_multicast_send_query(br, port, &mp->addr);
|
|
|
|
|
|
|
|
time = jiffies + br->multicast_last_member_count *
|
|
|
|
br->multicast_last_member_interval;
|
2013-08-30 22:28:17 +07:00
|
|
|
|
|
|
|
mod_timer(&query->timer, time);
|
2013-05-22 04:52:56 +07:00
|
|
|
|
|
|
|
for (p = mlock_dereference(mp->ports, br);
|
|
|
|
p != NULL;
|
|
|
|
p = mlock_dereference(p->next, br)) {
|
|
|
|
if (p->port != port)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (!hlist_unhashed(&p->mglist) &&
|
|
|
|
(timer_pending(&p->timer) ?
|
|
|
|
time_after(p->timer.expires, time) :
|
|
|
|
try_to_del_timer_sync(&p->timer) >= 0)) {
|
|
|
|
mod_timer(&p->timer, time);
|
|
|
|
}
|
|
|
|
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-12-06 04:24:45 +07:00
|
|
|
if (port && (port->flags & BR_MULTICAST_FAST_LEAVE)) {
|
2012-12-04 06:56:40 +07:00
|
|
|
struct net_bridge_port_group __rcu **pp;
|
|
|
|
|
|
|
|
for (pp = &mp->ports;
|
|
|
|
(p = mlock_dereference(*pp, br)) != NULL;
|
|
|
|
pp = &p->next) {
|
|
|
|
if (p->port != port)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
rcu_assign_pointer(*pp, p->next);
|
|
|
|
hlist_del_init(&p->mglist);
|
|
|
|
del_timer(&p->timer);
|
|
|
|
call_rcu_bh(&p->rcu, br_multicast_free_pg);
|
2012-12-12 05:23:07 +07:00
|
|
|
br_mdb_notify(br->dev, port, group, RTM_DELMDB);
|
2012-12-04 06:56:40 +07:00
|
|
|
|
2013-10-20 05:58:57 +07:00
|
|
|
if (!mp->ports && !mp->mglist &&
|
2012-12-04 06:56:40 +07:00
|
|
|
netif_running(br->dev))
|
|
|
|
mod_timer(&mp->timer, jiffies);
|
|
|
|
}
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
now = jiffies;
|
|
|
|
time = now + br->multicast_last_member_count *
|
|
|
|
br->multicast_last_member_interval;
|
|
|
|
|
|
|
|
if (!port) {
|
2013-10-20 05:58:57 +07:00
|
|
|
if (mp->mglist &&
|
2010-02-28 02:41:45 +07:00
|
|
|
(timer_pending(&mp->timer) ?
|
|
|
|
time_after(mp->timer.expires, time) :
|
|
|
|
try_to_del_timer_sync(&mp->timer) >= 0)) {
|
|
|
|
mod_timer(&mp->timer, time);
|
|
|
|
}
|
2013-10-20 05:58:57 +07:00
|
|
|
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (p = mlock_dereference(mp->ports, br);
|
|
|
|
p != NULL;
|
|
|
|
p = mlock_dereference(p->next, br)) {
|
|
|
|
if (p->port != port)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (!hlist_unhashed(&p->mglist) &&
|
|
|
|
(timer_pending(&p->timer) ?
|
|
|
|
time_after(p->timer.expires, time) :
|
|
|
|
try_to_del_timer_sync(&p->timer) >= 0)) {
|
|
|
|
mod_timer(&p->timer, time);
|
|
|
|
}
|
|
|
|
|
|
|
|
break;
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
out:
|
|
|
|
spin_unlock(&br->multicast_lock);
|
|
|
|
}
|
|
|
|
|
2010-04-18 10:42:07 +07:00
|
|
|
static void br_ip4_multicast_leave_group(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port,
|
2013-02-13 19:00:17 +07:00
|
|
|
__be32 group,
|
|
|
|
__u16 vid)
|
2010-04-18 10:42:07 +07:00
|
|
|
{
|
|
|
|
struct br_ip br_group;
|
2013-08-30 22:28:17 +07:00
|
|
|
struct bridge_mcast_query *query = port ? &port->ip4_query :
|
|
|
|
&br->ip4_query;
|
2010-04-18 10:42:07 +07:00
|
|
|
|
|
|
|
if (ipv4_is_local_multicast(group))
|
|
|
|
return;
|
|
|
|
|
|
|
|
br_group.u.ip4 = group;
|
|
|
|
br_group.proto = htons(ETH_P_IP);
|
2013-02-13 19:00:17 +07:00
|
|
|
br_group.vid = vid;
|
2010-04-18 10:42:07 +07:00
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
br_multicast_leave_group(br, port, &br_group, &br->ip4_querier, query);
|
2010-04-18 10:42:07 +07:00
|
|
|
}
|
|
|
|
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2010-04-22 23:54:22 +07:00
|
|
|
static void br_ip6_multicast_leave_group(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port,
|
2013-02-13 19:00:17 +07:00
|
|
|
const struct in6_addr *group,
|
|
|
|
__u16 vid)
|
2010-04-22 23:54:22 +07:00
|
|
|
{
|
|
|
|
struct br_ip br_group;
|
2013-08-30 22:28:17 +07:00
|
|
|
struct bridge_mcast_query *query = port ? &port->ip6_query :
|
|
|
|
&br->ip6_query;
|
|
|
|
|
2010-04-22 23:54:22 +07:00
|
|
|
|
2013-09-04 07:13:39 +07:00
|
|
|
if (ipv6_addr_is_ll_all_nodes(group))
|
2010-04-22 23:54:22 +07:00
|
|
|
return;
|
|
|
|
|
2011-11-21 10:39:03 +07:00
|
|
|
br_group.u.ip6 = *group;
|
2010-04-22 23:54:22 +07:00
|
|
|
br_group.proto = htons(ETH_P_IPV6);
|
2013-02-13 19:00:17 +07:00
|
|
|
br_group.vid = vid;
|
2010-04-22 23:54:22 +07:00
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
br_multicast_leave_group(br, port, &br_group, &br->ip6_querier, query);
|
2010-04-22 23:54:22 +07:00
|
|
|
}
|
|
|
|
#endif
|
2010-04-18 10:42:07 +07:00
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
static int br_multicast_ipv4_rcv(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port,
|
2013-10-29 02:45:07 +07:00
|
|
|
struct sk_buff *skb,
|
|
|
|
u16 vid)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
|
|
|
struct sk_buff *skb2 = skb;
|
2011-04-22 11:53:02 +07:00
|
|
|
const struct iphdr *iph;
|
2010-02-28 02:41:45 +07:00
|
|
|
struct igmphdr *ih;
|
2012-04-15 12:58:06 +07:00
|
|
|
unsigned int len;
|
|
|
|
unsigned int offset;
|
2010-02-28 02:41:45 +07:00
|
|
|
int err;
|
|
|
|
|
|
|
|
/* We treat OOM as packet loss for now. */
|
|
|
|
if (!pskb_may_pull(skb, sizeof(*iph)))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
iph = ip_hdr(skb);
|
|
|
|
|
|
|
|
if (iph->ihl < 5 || iph->version != 4)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (!pskb_may_pull(skb, ip_hdrlen(skb)))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
iph = ip_hdr(skb);
|
|
|
|
|
|
|
|
if (unlikely(ip_fast_csum((u8 *)iph, iph->ihl)))
|
|
|
|
return -EINVAL;
|
|
|
|
|
2011-06-23 09:39:12 +07:00
|
|
|
if (iph->protocol != IPPROTO_IGMP) {
|
2013-03-07 09:32:26 +07:00
|
|
|
if (!ipv4_is_local_multicast(iph->daddr))
|
2011-06-23 09:39:12 +07:00
|
|
|
BR_INPUT_SKB_CB(skb)->mrouters_only = 1;
|
2010-02-28 02:41:45 +07:00
|
|
|
return 0;
|
2011-06-23 09:39:12 +07:00
|
|
|
}
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
len = ntohs(iph->tot_len);
|
|
|
|
if (skb->len < len || len < ip_hdrlen(skb))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (skb->len > len) {
|
|
|
|
skb2 = skb_clone(skb, GFP_ATOMIC);
|
|
|
|
if (!skb2)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
err = pskb_trim_rcsum(skb2, len);
|
|
|
|
if (err)
|
2010-03-16 02:26:56 +07:00
|
|
|
goto err_out;
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
len -= ip_hdrlen(skb2);
|
|
|
|
offset = skb_network_offset(skb2) + ip_hdrlen(skb2);
|
|
|
|
__skb_pull(skb2, offset);
|
|
|
|
skb_reset_transport_header(skb2);
|
|
|
|
|
|
|
|
err = -EINVAL;
|
|
|
|
if (!pskb_may_pull(skb2, sizeof(*ih)))
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
switch (skb2->ip_summed) {
|
|
|
|
case CHECKSUM_COMPLETE:
|
|
|
|
if (!csum_fold(skb2->csum))
|
|
|
|
break;
|
|
|
|
/* fall through */
|
|
|
|
case CHECKSUM_NONE:
|
|
|
|
skb2->csum = 0;
|
|
|
|
if (skb_checksum_complete(skb2))
|
2010-03-16 02:26:56 +07:00
|
|
|
goto out;
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
err = 0;
|
|
|
|
|
|
|
|
BR_INPUT_SKB_CB(skb)->igmp = 1;
|
|
|
|
ih = igmp_hdr(skb2);
|
|
|
|
|
|
|
|
switch (ih->type) {
|
|
|
|
case IGMP_HOST_MEMBERSHIP_REPORT:
|
|
|
|
case IGMPV2_HOST_MEMBERSHIP_REPORT:
|
2011-06-13 22:04:43 +07:00
|
|
|
BR_INPUT_SKB_CB(skb)->mrouters_only = 1;
|
2013-02-13 19:00:17 +07:00
|
|
|
err = br_ip4_multicast_add_group(br, port, ih->group, vid);
|
2010-02-28 02:41:45 +07:00
|
|
|
break;
|
|
|
|
case IGMPV3_HOST_MEMBERSHIP_REPORT:
|
2013-10-29 02:45:07 +07:00
|
|
|
err = br_ip4_multicast_igmp3_report(br, port, skb2, vid);
|
2010-02-28 02:41:45 +07:00
|
|
|
break;
|
|
|
|
case IGMP_HOST_MEMBERSHIP_QUERY:
|
2013-10-29 02:45:07 +07:00
|
|
|
err = br_ip4_multicast_query(br, port, skb2, vid);
|
2010-02-28 02:41:45 +07:00
|
|
|
break;
|
|
|
|
case IGMP_HOST_LEAVE_MESSAGE:
|
2013-02-13 19:00:17 +07:00
|
|
|
br_ip4_multicast_leave_group(br, port, ih->group, vid);
|
2010-02-28 02:41:45 +07:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
__skb_push(skb2, offset);
|
2010-03-16 02:26:56 +07:00
|
|
|
err_out:
|
2010-02-28 02:41:45 +07:00
|
|
|
if (skb2 != skb)
|
|
|
|
kfree_skb(skb2);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2010-04-22 23:54:22 +07:00
|
|
|
static int br_multicast_ipv6_rcv(struct net_bridge *br,
|
|
|
|
struct net_bridge_port *port,
|
2013-10-29 02:45:07 +07:00
|
|
|
struct sk_buff *skb,
|
|
|
|
u16 vid)
|
2010-04-22 23:54:22 +07:00
|
|
|
{
|
2011-01-04 02:26:08 +07:00
|
|
|
struct sk_buff *skb2;
|
2011-04-22 11:53:02 +07:00
|
|
|
const struct ipv6hdr *ip6h;
|
2011-08-24 02:57:05 +07:00
|
|
|
u8 icmp6_type;
|
2010-04-22 23:54:22 +07:00
|
|
|
u8 nexthdr;
|
2011-12-01 08:05:51 +07:00
|
|
|
__be16 frag_off;
|
2012-04-15 12:58:06 +07:00
|
|
|
unsigned int len;
|
2010-07-15 15:47:33 +07:00
|
|
|
int offset;
|
2010-04-22 23:54:22 +07:00
|
|
|
int err;
|
|
|
|
|
|
|
|
if (!pskb_may_pull(skb, sizeof(*ip6h)))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
ip6h = ipv6_hdr(skb);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We're interested in MLD messages only.
|
|
|
|
* - Version is 6
|
|
|
|
* - MLD has always Router Alert hop-by-hop option
|
|
|
|
* - But we do not support jumbrograms.
|
|
|
|
*/
|
2013-09-04 07:13:38 +07:00
|
|
|
if (ip6h->version != 6)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* Prevent flooding this packet if there is no listener present */
|
2013-09-04 07:13:39 +07:00
|
|
|
if (!ipv6_addr_is_ll_all_nodes(&ip6h->daddr))
|
2013-09-04 07:13:38 +07:00
|
|
|
BR_INPUT_SKB_CB(skb)->mrouters_only = 1;
|
|
|
|
|
|
|
|
if (ip6h->nexthdr != IPPROTO_HOPOPTS ||
|
2010-04-22 23:54:22 +07:00
|
|
|
ip6h->payload_len == 0)
|
|
|
|
return 0;
|
|
|
|
|
2011-03-27 03:27:24 +07:00
|
|
|
len = ntohs(ip6h->payload_len) + sizeof(*ip6h);
|
2010-04-22 23:54:22 +07:00
|
|
|
if (skb->len < len)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
nexthdr = ip6h->nexthdr;
|
2011-12-01 08:05:51 +07:00
|
|
|
offset = ipv6_skip_exthdr(skb, sizeof(*ip6h), &nexthdr, &frag_off);
|
2010-04-22 23:54:22 +07:00
|
|
|
|
|
|
|
if (offset < 0 || nexthdr != IPPROTO_ICMPV6)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* Okay, we found ICMPv6 header */
|
|
|
|
skb2 = skb_clone(skb, GFP_ATOMIC);
|
|
|
|
if (!skb2)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2011-01-04 02:26:08 +07:00
|
|
|
err = -EINVAL;
|
|
|
|
if (!pskb_may_pull(skb2, offset + sizeof(struct icmp6hdr)))
|
|
|
|
goto out;
|
|
|
|
|
2010-04-22 23:54:22 +07:00
|
|
|
len -= offset - skb_network_offset(skb2);
|
|
|
|
|
|
|
|
__skb_pull(skb2, offset);
|
|
|
|
skb_reset_transport_header(skb2);
|
2011-11-15 15:09:14 +07:00
|
|
|
skb_postpull_rcsum(skb2, skb_network_header(skb2),
|
|
|
|
skb_network_header_len(skb2));
|
2010-04-22 23:54:22 +07:00
|
|
|
|
2011-08-24 02:57:05 +07:00
|
|
|
icmp6_type = icmp6_hdr(skb2)->icmp6_type;
|
2010-04-22 23:54:22 +07:00
|
|
|
|
2011-08-24 02:57:05 +07:00
|
|
|
switch (icmp6_type) {
|
2010-04-22 23:54:22 +07:00
|
|
|
case ICMPV6_MGM_QUERY:
|
|
|
|
case ICMPV6_MGM_REPORT:
|
|
|
|
case ICMPV6_MGM_REDUCTION:
|
|
|
|
case ICMPV6_MLD2_REPORT:
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
err = 0;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Okay, we found MLD message. Check further. */
|
|
|
|
if (skb2->len > len) {
|
|
|
|
err = pskb_trim_rcsum(skb2, len);
|
|
|
|
if (err)
|
|
|
|
goto out;
|
2011-08-24 05:54:33 +07:00
|
|
|
err = -EINVAL;
|
2010-04-22 23:54:22 +07:00
|
|
|
}
|
|
|
|
|
2011-08-24 05:54:33 +07:00
|
|
|
ip6h = ipv6_hdr(skb2);
|
|
|
|
|
2010-04-22 23:54:22 +07:00
|
|
|
switch (skb2->ip_summed) {
|
|
|
|
case CHECKSUM_COMPLETE:
|
2011-08-24 05:54:33 +07:00
|
|
|
if (!csum_ipv6_magic(&ip6h->saddr, &ip6h->daddr, skb2->len,
|
|
|
|
IPPROTO_ICMPV6, skb2->csum))
|
2010-04-22 23:54:22 +07:00
|
|
|
break;
|
|
|
|
/*FALLTHROUGH*/
|
|
|
|
case CHECKSUM_NONE:
|
2011-08-24 05:54:33 +07:00
|
|
|
skb2->csum = ~csum_unfold(csum_ipv6_magic(&ip6h->saddr,
|
|
|
|
&ip6h->daddr,
|
|
|
|
skb2->len,
|
|
|
|
IPPROTO_ICMPV6, 0));
|
|
|
|
if (__skb_checksum_complete(skb2))
|
2010-04-22 23:54:22 +07:00
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
err = 0;
|
|
|
|
|
|
|
|
BR_INPUT_SKB_CB(skb)->igmp = 1;
|
|
|
|
|
2011-08-24 02:57:05 +07:00
|
|
|
switch (icmp6_type) {
|
2010-04-22 23:54:22 +07:00
|
|
|
case ICMPV6_MGM_REPORT:
|
|
|
|
{
|
2011-01-04 02:26:08 +07:00
|
|
|
struct mld_msg *mld;
|
|
|
|
if (!pskb_may_pull(skb2, sizeof(*mld))) {
|
|
|
|
err = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
mld = (struct mld_msg *)skb_transport_header(skb2);
|
2011-06-13 22:06:58 +07:00
|
|
|
BR_INPUT_SKB_CB(skb)->mrouters_only = 1;
|
2013-02-13 19:00:17 +07:00
|
|
|
err = br_ip6_multicast_add_group(br, port, &mld->mld_mca, vid);
|
2010-04-22 23:54:22 +07:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
case ICMPV6_MLD2_REPORT:
|
2013-10-29 02:45:07 +07:00
|
|
|
err = br_ip6_multicast_mld2_report(br, port, skb2, vid);
|
2010-04-22 23:54:22 +07:00
|
|
|
break;
|
|
|
|
case ICMPV6_MGM_QUERY:
|
2013-10-29 02:45:07 +07:00
|
|
|
err = br_ip6_multicast_query(br, port, skb2, vid);
|
2010-04-22 23:54:22 +07:00
|
|
|
break;
|
|
|
|
case ICMPV6_MGM_REDUCTION:
|
|
|
|
{
|
2011-01-04 02:26:08 +07:00
|
|
|
struct mld_msg *mld;
|
|
|
|
if (!pskb_may_pull(skb2, sizeof(*mld))) {
|
|
|
|
err = -EINVAL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
mld = (struct mld_msg *)skb_transport_header(skb2);
|
2013-02-13 19:00:17 +07:00
|
|
|
br_ip6_multicast_leave_group(br, port, &mld->mld_mca, vid);
|
2010-04-22 23:54:22 +07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
2011-01-04 02:26:08 +07:00
|
|
|
kfree_skb(skb2);
|
2010-04-22 23:54:22 +07:00
|
|
|
return err;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
int br_multicast_rcv(struct net_bridge *br, struct net_bridge_port *port,
|
2013-10-29 02:45:07 +07:00
|
|
|
struct sk_buff *skb, u16 vid)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
2010-04-25 15:06:40 +07:00
|
|
|
BR_INPUT_SKB_CB(skb)->igmp = 0;
|
|
|
|
BR_INPUT_SKB_CB(skb)->mrouters_only = 0;
|
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
if (br->multicast_disabled)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
switch (skb->protocol) {
|
|
|
|
case htons(ETH_P_IP):
|
2013-10-29 02:45:07 +07:00
|
|
|
return br_multicast_ipv4_rcv(br, port, skb, vid);
|
2011-12-10 16:48:31 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
2010-04-22 23:54:22 +07:00
|
|
|
case htons(ETH_P_IPV6):
|
2013-10-29 02:45:07 +07:00
|
|
|
return br_multicast_ipv6_rcv(br, port, skb, vid);
|
2010-04-22 23:54:22 +07:00
|
|
|
#endif
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
static void br_multicast_query_expired(struct net_bridge *br,
|
|
|
|
struct bridge_mcast_query *query)
|
|
|
|
{
|
|
|
|
spin_lock(&br->multicast_lock);
|
|
|
|
if (query->startup_sent < br->multicast_startup_query_count)
|
|
|
|
query->startup_sent++;
|
|
|
|
|
|
|
|
br_multicast_send_query(br, NULL, query);
|
|
|
|
spin_unlock(&br->multicast_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void br_ip4_multicast_query_expired(unsigned long data)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
|
|
|
struct net_bridge *br = (void *)data;
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
br_multicast_query_expired(br, &br->ip4_query);
|
|
|
|
}
|
2010-02-28 02:41:45 +07:00
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
|
|
static void br_ip6_multicast_query_expired(unsigned long data)
|
|
|
|
{
|
|
|
|
struct net_bridge *br = (void *)data;
|
2010-02-28 02:41:45 +07:00
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
br_multicast_query_expired(br, &br->ip6_query);
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
2013-08-30 22:28:17 +07:00
|
|
|
#endif
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
void br_multicast_init(struct net_bridge *br)
|
|
|
|
{
|
|
|
|
br->hash_elasticity = 4;
|
|
|
|
br->hash_max = 512;
|
|
|
|
|
|
|
|
br->multicast_router = 1;
|
2012-04-13 09:37:42 +07:00
|
|
|
br->multicast_querier = 0;
|
2013-05-22 04:52:54 +07:00
|
|
|
br->multicast_query_use_ifaddr = 0;
|
2010-02-28 02:41:45 +07:00
|
|
|
br->multicast_last_member_count = 2;
|
|
|
|
br->multicast_startup_query_count = 2;
|
|
|
|
|
|
|
|
br->multicast_last_member_interval = HZ;
|
|
|
|
br->multicast_query_response_interval = 10 * HZ;
|
|
|
|
br->multicast_startup_query_interval = 125 * HZ / 4;
|
|
|
|
br->multicast_query_interval = 125 * HZ;
|
|
|
|
br->multicast_querier_interval = 255 * HZ;
|
|
|
|
br->multicast_membership_interval = 260 * HZ;
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
br->ip4_querier.delay_time = 0;
|
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
|
|
br->ip6_querier.delay_time = 0;
|
|
|
|
#endif
|
2013-08-01 06:06:20 +07:00
|
|
|
|
2010-02-28 02:41:45 +07:00
|
|
|
spin_lock_init(&br->multicast_lock);
|
|
|
|
setup_timer(&br->multicast_router_timer,
|
|
|
|
br_multicast_local_router_expired, 0);
|
2013-08-30 22:28:17 +07:00
|
|
|
setup_timer(&br->ip4_querier.timer, br_ip4_multicast_querier_expired,
|
|
|
|
(unsigned long)br);
|
|
|
|
setup_timer(&br->ip4_query.timer, br_ip4_multicast_query_expired,
|
2010-02-28 02:41:45 +07:00
|
|
|
(unsigned long)br);
|
2013-08-30 22:28:17 +07:00
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
|
|
setup_timer(&br->ip6_querier.timer, br_ip6_multicast_querier_expired,
|
|
|
|
(unsigned long)br);
|
|
|
|
setup_timer(&br->ip6_query.timer, br_ip6_multicast_query_expired,
|
|
|
|
(unsigned long)br);
|
|
|
|
#endif
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
static void __br_multicast_open(struct net_bridge *br,
|
|
|
|
struct bridge_mcast_query *query)
|
2010-02-28 02:41:45 +07:00
|
|
|
{
|
2013-08-30 22:28:17 +07:00
|
|
|
query->startup_sent = 0;
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
if (br->multicast_disabled)
|
|
|
|
return;
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
mod_timer(&query->timer, jiffies);
|
|
|
|
}
|
|
|
|
|
|
|
|
void br_multicast_open(struct net_bridge *br)
|
|
|
|
{
|
|
|
|
__br_multicast_open(br, &br->ip4_query);
|
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
|
|
__br_multicast_open(br, &br->ip6_query);
|
|
|
|
#endif
|
2010-02-28 02:41:45 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
void br_multicast_stop(struct net_bridge *br)
|
|
|
|
{
|
|
|
|
struct net_bridge_mdb_htable *mdb;
|
|
|
|
struct net_bridge_mdb_entry *mp;
|
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 08:06:00 +07:00
|
|
|
struct hlist_node *n;
|
2010-02-28 02:41:45 +07:00
|
|
|
u32 ver;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
del_timer_sync(&br->multicast_router_timer);
|
2013-08-30 22:28:17 +07:00
|
|
|
del_timer_sync(&br->ip4_querier.timer);
|
|
|
|
del_timer_sync(&br->ip4_query.timer);
|
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
|
|
del_timer_sync(&br->ip6_querier.timer);
|
|
|
|
del_timer_sync(&br->ip6_query.timer);
|
|
|
|
#endif
|
2010-02-28 02:41:45 +07:00
|
|
|
|
|
|
|
spin_lock_bh(&br->multicast_lock);
|
2010-11-15 13:38:10 +07:00
|
|
|
mdb = mlock_dereference(br->mdb, br);
|
2010-02-28 02:41:45 +07:00
|
|
|
if (!mdb)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
br->mdb = NULL;
|
|
|
|
|
|
|
|
ver = mdb->ver;
|
|
|
|
for (i = 0; i < mdb->max; i++) {
|
hlist: drop the node parameter from iterators
I'm not sure why, but the hlist for each entry iterators were conceived
list_for_each_entry(pos, head, member)
The hlist ones were greedy and wanted an extra parameter:
hlist_for_each_entry(tpos, pos, head, member)
Why did they need an extra pos parameter? I'm not quite sure. Not only
they don't really need it, it also prevents the iterator from looking
exactly like the list iterator, which is unfortunate.
Besides the semantic patch, there was some manual work required:
- Fix up the actual hlist iterators in linux/list.h
- Fix up the declaration of other iterators based on the hlist ones.
- A very small amount of places were using the 'node' parameter, this
was modified to use 'obj->member' instead.
- Coccinelle didn't handle the hlist_for_each_entry_safe iterator
properly, so those had to be fixed up manually.
The semantic patch which is mostly the work of Peter Senna Tschudin is here:
@@
iterator name hlist_for_each_entry, hlist_for_each_entry_continue, hlist_for_each_entry_from, hlist_for_each_entry_rcu, hlist_for_each_entry_rcu_bh, hlist_for_each_entry_continue_rcu_bh, for_each_busy_worker, ax25_uid_for_each, ax25_for_each, inet_bind_bucket_for_each, sctp_for_each_hentry, sk_for_each, sk_for_each_rcu, sk_for_each_from, sk_for_each_safe, sk_for_each_bound, hlist_for_each_entry_safe, hlist_for_each_entry_continue_rcu, nr_neigh_for_each, nr_neigh_for_each_safe, nr_node_for_each, nr_node_for_each_safe, for_each_gfn_indirect_valid_sp, for_each_gfn_sp, for_each_host;
type T;
expression a,c,d,e;
identifier b;
statement S;
@@
-T b;
<+... when != b
(
hlist_for_each_entry(a,
- b,
c, d) S
|
hlist_for_each_entry_continue(a,
- b,
c) S
|
hlist_for_each_entry_from(a,
- b,
c) S
|
hlist_for_each_entry_rcu(a,
- b,
c, d) S
|
hlist_for_each_entry_rcu_bh(a,
- b,
c, d) S
|
hlist_for_each_entry_continue_rcu_bh(a,
- b,
c) S
|
for_each_busy_worker(a, c,
- b,
d) S
|
ax25_uid_for_each(a,
- b,
c) S
|
ax25_for_each(a,
- b,
c) S
|
inet_bind_bucket_for_each(a,
- b,
c) S
|
sctp_for_each_hentry(a,
- b,
c) S
|
sk_for_each(a,
- b,
c) S
|
sk_for_each_rcu(a,
- b,
c) S
|
sk_for_each_from
-(a, b)
+(a)
S
+ sk_for_each_from(a) S
|
sk_for_each_safe(a,
- b,
c, d) S
|
sk_for_each_bound(a,
- b,
c) S
|
hlist_for_each_entry_safe(a,
- b,
c, d, e) S
|
hlist_for_each_entry_continue_rcu(a,
- b,
c) S
|
nr_neigh_for_each(a,
- b,
c) S
|
nr_neigh_for_each_safe(a,
- b,
c, d) S
|
nr_node_for_each(a,
- b,
c) S
|
nr_node_for_each_safe(a,
- b,
c, d) S
|
- for_each_gfn_sp(a, c, d, b) S
+ for_each_gfn_sp(a, c, d) S
|
- for_each_gfn_indirect_valid_sp(a, c, d, b) S
+ for_each_gfn_indirect_valid_sp(a, c, d) S
|
for_each_host(a,
- b,
c) S
|
for_each_host_safe(a,
- b,
c, d) S
|
for_each_mesh_entry(a,
- b,
c, d) S
)
...+>
[akpm@linux-foundation.org: drop bogus change from net/ipv4/raw.c]
[akpm@linux-foundation.org: drop bogus hunk from net/ipv6/raw.c]
[akpm@linux-foundation.org: checkpatch fixes]
[akpm@linux-foundation.org: fix warnings]
[akpm@linux-foudnation.org: redo intrusive kvm changes]
Tested-by: Peter Senna Tschudin <peter.senna@gmail.com>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Signed-off-by: Sasha Levin <sasha.levin@oracle.com>
Cc: Wu Fengguang <fengguang.wu@intel.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Gleb Natapov <gleb@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2013-02-28 08:06:00 +07:00
|
|
|
hlist_for_each_entry_safe(mp, n, &mdb->mhash[i],
|
2010-02-28 02:41:45 +07:00
|
|
|
hlist[ver]) {
|
|
|
|
del_timer(&mp->timer);
|
|
|
|
call_rcu_bh(&mp->rcu, br_multicast_free_group);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (mdb->old) {
|
|
|
|
spin_unlock_bh(&br->multicast_lock);
|
2010-03-06 04:03:35 +07:00
|
|
|
rcu_barrier_bh();
|
2010-02-28 02:41:45 +07:00
|
|
|
spin_lock_bh(&br->multicast_lock);
|
|
|
|
WARN_ON(mdb->old);
|
|
|
|
}
|
|
|
|
|
|
|
|
mdb->old = mdb;
|
|
|
|
call_rcu_bh(&mdb->rcu, br_mdb_free);
|
|
|
|
|
|
|
|
out:
|
|
|
|
spin_unlock_bh(&br->multicast_lock);
|
|
|
|
}
|
2010-02-28 02:41:49 +07:00
|
|
|
|
|
|
|
int br_multicast_set_router(struct net_bridge *br, unsigned long val)
|
|
|
|
{
|
|
|
|
int err = -ENOENT;
|
|
|
|
|
|
|
|
spin_lock_bh(&br->multicast_lock);
|
|
|
|
if (!netif_running(br->dev))
|
|
|
|
goto unlock;
|
|
|
|
|
|
|
|
switch (val) {
|
|
|
|
case 0:
|
|
|
|
case 2:
|
|
|
|
del_timer(&br->multicast_router_timer);
|
|
|
|
/* fall through */
|
|
|
|
case 1:
|
|
|
|
br->multicast_router = val;
|
|
|
|
err = 0;
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
err = -EINVAL;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
unlock:
|
|
|
|
spin_unlock_bh(&br->multicast_lock);
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
int br_multicast_set_port_router(struct net_bridge_port *p, unsigned long val)
|
|
|
|
{
|
|
|
|
struct net_bridge *br = p->br;
|
|
|
|
int err = -ENOENT;
|
|
|
|
|
|
|
|
spin_lock(&br->multicast_lock);
|
|
|
|
if (!netif_running(br->dev) || p->state == BR_STATE_DISABLED)
|
|
|
|
goto unlock;
|
|
|
|
|
|
|
|
switch (val) {
|
|
|
|
case 0:
|
|
|
|
case 1:
|
|
|
|
case 2:
|
|
|
|
p->multicast_router = val;
|
|
|
|
err = 0;
|
|
|
|
|
|
|
|
if (val < 2 && !hlist_unhashed(&p->rlist))
|
|
|
|
hlist_del_init_rcu(&p->rlist);
|
|
|
|
|
|
|
|
if (val == 1)
|
|
|
|
break;
|
|
|
|
|
|
|
|
del_timer(&p->multicast_router_timer);
|
|
|
|
|
|
|
|
if (val == 0)
|
|
|
|
break;
|
|
|
|
|
|
|
|
br_multicast_add_router(br, p);
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
err = -EINVAL;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
unlock:
|
|
|
|
spin_unlock(&br->multicast_lock);
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
2010-02-28 02:41:50 +07:00
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
static void br_multicast_start_querier(struct net_bridge *br,
|
|
|
|
struct bridge_mcast_query *query)
|
2010-02-28 02:41:50 +07:00
|
|
|
{
|
|
|
|
struct net_bridge_port *port;
|
2012-04-13 09:37:42 +07:00
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
__br_multicast_open(br, query);
|
2012-04-13 09:37:42 +07:00
|
|
|
|
|
|
|
list_for_each_entry(port, &br->port_list, list) {
|
|
|
|
if (port->state == BR_STATE_DISABLED ||
|
|
|
|
port->state == BR_STATE_BLOCKING)
|
|
|
|
continue;
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
if (query == &br->ip4_query)
|
|
|
|
br_multicast_enable(&port->ip4_query);
|
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
|
|
else
|
|
|
|
br_multicast_enable(&port->ip6_query);
|
|
|
|
#endif
|
2012-04-13 09:37:42 +07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
int br_multicast_toggle(struct net_bridge *br, unsigned long val)
|
|
|
|
{
|
2010-07-29 07:45:30 +07:00
|
|
|
int err = 0;
|
2010-11-15 13:38:10 +07:00
|
|
|
struct net_bridge_mdb_htable *mdb;
|
2010-02-28 02:41:50 +07:00
|
|
|
|
2011-11-10 12:48:03 +07:00
|
|
|
spin_lock_bh(&br->multicast_lock);
|
2010-02-28 02:41:50 +07:00
|
|
|
if (br->multicast_disabled == !val)
|
|
|
|
goto unlock;
|
|
|
|
|
|
|
|
br->multicast_disabled = !val;
|
|
|
|
if (br->multicast_disabled)
|
|
|
|
goto unlock;
|
|
|
|
|
2010-07-29 07:45:30 +07:00
|
|
|
if (!netif_running(br->dev))
|
|
|
|
goto unlock;
|
|
|
|
|
2010-11-15 13:38:10 +07:00
|
|
|
mdb = mlock_dereference(br->mdb, br);
|
|
|
|
if (mdb) {
|
|
|
|
if (mdb->old) {
|
2010-02-28 02:41:50 +07:00
|
|
|
err = -EEXIST;
|
|
|
|
rollback:
|
|
|
|
br->multicast_disabled = !!val;
|
|
|
|
goto unlock;
|
|
|
|
}
|
|
|
|
|
2010-11-15 13:38:10 +07:00
|
|
|
err = br_mdb_rehash(&br->mdb, mdb->max,
|
2010-02-28 02:41:50 +07:00
|
|
|
br->hash_elasticity);
|
|
|
|
if (err)
|
|
|
|
goto rollback;
|
|
|
|
}
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
br_multicast_start_querier(br, &br->ip4_query);
|
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
|
|
br_multicast_start_querier(br, &br->ip6_query);
|
|
|
|
#endif
|
2010-02-28 02:41:50 +07:00
|
|
|
|
|
|
|
unlock:
|
2011-11-10 12:48:03 +07:00
|
|
|
spin_unlock_bh(&br->multicast_lock);
|
2010-02-28 02:41:50 +07:00
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
2010-02-28 02:41:51 +07:00
|
|
|
|
2012-04-13 09:37:42 +07:00
|
|
|
int br_multicast_set_querier(struct net_bridge *br, unsigned long val)
|
|
|
|
{
|
2013-08-01 06:06:20 +07:00
|
|
|
unsigned long max_delay;
|
|
|
|
|
2012-04-13 09:37:42 +07:00
|
|
|
val = !!val;
|
|
|
|
|
|
|
|
spin_lock_bh(&br->multicast_lock);
|
|
|
|
if (br->multicast_querier == val)
|
|
|
|
goto unlock;
|
|
|
|
|
|
|
|
br->multicast_querier = val;
|
2013-08-01 06:06:20 +07:00
|
|
|
if (!val)
|
|
|
|
goto unlock;
|
|
|
|
|
|
|
|
max_delay = br->multicast_query_response_interval;
|
|
|
|
|
2013-08-30 22:28:17 +07:00
|
|
|
if (!timer_pending(&br->ip4_querier.timer))
|
|
|
|
br->ip4_querier.delay_time = jiffies + max_delay;
|
|
|
|
|
|
|
|
br_multicast_start_querier(br, &br->ip4_query);
|
|
|
|
|
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
|
|
if (!timer_pending(&br->ip6_querier.timer))
|
|
|
|
br->ip6_querier.delay_time = jiffies + max_delay;
|
|
|
|
|
|
|
|
br_multicast_start_querier(br, &br->ip6_query);
|
|
|
|
#endif
|
2012-04-13 09:37:42 +07:00
|
|
|
|
|
|
|
unlock:
|
|
|
|
spin_unlock_bh(&br->multicast_lock);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2010-02-28 02:41:51 +07:00
|
|
|
int br_multicast_set_hash_max(struct net_bridge *br, unsigned long val)
|
|
|
|
{
|
|
|
|
int err = -ENOENT;
|
|
|
|
u32 old;
|
2010-11-15 13:38:10 +07:00
|
|
|
struct net_bridge_mdb_htable *mdb;
|
2010-02-28 02:41:51 +07:00
|
|
|
|
2014-01-07 02:00:32 +07:00
|
|
|
spin_lock_bh(&br->multicast_lock);
|
2010-02-28 02:41:51 +07:00
|
|
|
if (!netif_running(br->dev))
|
|
|
|
goto unlock;
|
|
|
|
|
|
|
|
err = -EINVAL;
|
|
|
|
if (!is_power_of_2(val))
|
|
|
|
goto unlock;
|
2010-11-15 13:38:10 +07:00
|
|
|
|
|
|
|
mdb = mlock_dereference(br->mdb, br);
|
|
|
|
if (mdb && val < mdb->size)
|
2010-02-28 02:41:51 +07:00
|
|
|
goto unlock;
|
|
|
|
|
|
|
|
err = 0;
|
|
|
|
|
|
|
|
old = br->hash_max;
|
|
|
|
br->hash_max = val;
|
|
|
|
|
2010-11-15 13:38:10 +07:00
|
|
|
if (mdb) {
|
|
|
|
if (mdb->old) {
|
2010-02-28 02:41:51 +07:00
|
|
|
err = -EEXIST;
|
|
|
|
rollback:
|
|
|
|
br->hash_max = old;
|
|
|
|
goto unlock;
|
|
|
|
}
|
|
|
|
|
|
|
|
err = br_mdb_rehash(&br->mdb, br->hash_max,
|
|
|
|
br->hash_elasticity);
|
|
|
|
if (err)
|
|
|
|
goto rollback;
|
|
|
|
}
|
|
|
|
|
|
|
|
unlock:
|
2014-01-07 02:00:32 +07:00
|
|
|
spin_unlock_bh(&br->multicast_lock);
|
2010-02-28 02:41:51 +07:00
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|