mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-12 22:36:39 +07:00
b63f6044d8
Pablo Neira Ayuso says: ==================== Netfilter updates for net-next The following patchset contains Netfilter updates for your net-next tree. Basically, updates to the conntrack core, enhancements for nf_tables, conversion of netfilter hooks from linked list to array to improve memory locality and asorted improvements for the Netfilter codebase. More specifically, they are: 1) Add expection to hashes after timer initialization to prevent access from another CPU that walks on the hashes and calls del_timer(), from Florian Westphal. 2) Don't update nf_tables chain counters from hot path, this is only used by the x_tables compatibility layer. 3) Get rid of nested rcu_read_lock() calls from netfilter hook path. Hooks are always guaranteed to run from rcu read side, so remove nested rcu_read_lock() where possible. Patch from Taehee Yoo. 4) nf_tables new ruleset generation notifications include PID and name of the process that has updated the ruleset, from Phil Sutter. 5) Use skb_header_pointer() from nft_fib, so we can reuse this code from the nf_family netdev family. Patch from Pablo M. Bermudo. 6) Add support for nft_fib in nf_tables netdev family, also from Pablo. 7) Use deferrable workqueue for conntrack garbage collection, to reduce power consumption, from Patch from Subash Abhinov Kasiviswanathan. 8) Add nf_ct_expect_iterate_net() helper and use it. From Florian Westphal. 9) Call nf_ct_unconfirmed_destroy only from cttimeout, from Florian. 10) Drop references on conntrack removal path when skbuffs has escaped via nfqueue, from Florian. 11) Don't queue packets to nfqueue with dying conntrack, from Florian. 12) Constify nf_hook_ops structure, from Florian. 13) Remove neededlessly branch in nf_tables trace code, from Phil Sutter. 14) Add nla_strdup(), from Phil Sutter. 15) Rise nf_tables objects name size up to 255 chars, people want to use DNS names, so increase this according to what RFC 1035 specifies. Patch series from Phil Sutter. 16) Kill nf_conntrack_default_on, it's broken. Default on conntrack hook registration on demand, suggested by Eric Dumazet, patch from Florian. 17) Remove unused variables in compat_copy_entry_from_user both in ip_tables and arp_tables code. Patch from Taehee Yoo. 18) Constify struct nf_conntrack_l4proto, from Julia Lawall. 19) Constify nf_loginfo structure, also from Julia. 20) Use a single rb root in connlimit, from Taehee Yoo. 21) Remove unused netfilter_queue_init() prototype, from Taehee Yoo. 22) Use audit_log() instead of open-coding it, from Geliang Tang. 23) Allow to mangle tcp options via nft_exthdr, from Florian. 24) Allow to fetch TCP MSS from nft_rt, from Florian. This includes a fix for a miscalculation of the minimal length. 25) Simplify branch logic in h323 helper, from Nick Desaulniers. 26) Calculate netlink attribute size for conntrack tuple at compile time, from Florian. 27) Remove protocol name field from nf_conntrack_{l3,l4}proto structure. From Florian. 28) Remove holes in nf_conntrack_l4proto structure, so it becomes smaller. From Florian. 29) Get rid of print_tuple() indirection for /proc conntrack listing. Place all the code in net/netfilter/nf_conntrack_standalone.c. Patch from Florian. 30) Do not built in print_conntrack() if CONFIG_NF_CONNTRACK_PROCFS is off. From Florian. 31) Constify most nf_conntrack_{l3,l4}proto helper functions, from Florian. 32) Fix broken indentation in ebtables extensions, from Colin Ian King. 33) Fix several harmless sparse warning, from Florian. 34) Convert netfilter hook infrastructure to use array for better memory locality, joint work done by Florian and Aaron Conole. Moreover, add some instrumentation to debug this. 35) Batch nf_unregister_net_hooks() calls, to call synchronize_net once per batch, from Florian. 36) Get rid of noisy logging in ICMPv6 conntrack helper, from Florian. 37) Get rid of obsolete NFDEBUG() instrumentation, from Varsha Rao. 38) Remove unused code in the generic protocol tracker, from Davide Caratti. I think I will have material for a second Netfilter batch in my queue if time allow to make it fit in this merge window. ==================== Signed-off-by: David S. Miller <davem@davemloft.net>
1018 lines
26 KiB
C
1018 lines
26 KiB
C
/* Copyright (c) 2014 Mahesh Bandewar <maheshb@google.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License as
|
|
* published by the Free Software Foundation; either version 2 of
|
|
* the License, or (at your option) any later version.
|
|
*
|
|
*/
|
|
|
|
#include "ipvlan.h"
|
|
|
|
static unsigned int ipvlan_netid __read_mostly;
|
|
|
|
struct ipvlan_netns {
|
|
unsigned int ipvl_nf_hook_refcnt;
|
|
};
|
|
|
|
static const struct nf_hook_ops ipvl_nfops[] = {
|
|
{
|
|
.hook = ipvlan_nf_input,
|
|
.pf = NFPROTO_IPV4,
|
|
.hooknum = NF_INET_LOCAL_IN,
|
|
.priority = INT_MAX,
|
|
},
|
|
{
|
|
.hook = ipvlan_nf_input,
|
|
.pf = NFPROTO_IPV6,
|
|
.hooknum = NF_INET_LOCAL_IN,
|
|
.priority = INT_MAX,
|
|
},
|
|
};
|
|
|
|
static const struct l3mdev_ops ipvl_l3mdev_ops = {
|
|
.l3mdev_l3_rcv = ipvlan_l3_rcv,
|
|
};
|
|
|
|
static void ipvlan_adjust_mtu(struct ipvl_dev *ipvlan, struct net_device *dev)
|
|
{
|
|
ipvlan->dev->mtu = dev->mtu;
|
|
}
|
|
|
|
static int ipvlan_register_nf_hook(struct net *net)
|
|
{
|
|
struct ipvlan_netns *vnet = net_generic(net, ipvlan_netid);
|
|
int err = 0;
|
|
|
|
if (!vnet->ipvl_nf_hook_refcnt) {
|
|
err = nf_register_net_hooks(net, ipvl_nfops,
|
|
ARRAY_SIZE(ipvl_nfops));
|
|
if (!err)
|
|
vnet->ipvl_nf_hook_refcnt = 1;
|
|
} else {
|
|
vnet->ipvl_nf_hook_refcnt++;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static void ipvlan_unregister_nf_hook(struct net *net)
|
|
{
|
|
struct ipvlan_netns *vnet = net_generic(net, ipvlan_netid);
|
|
|
|
if (WARN_ON(!vnet->ipvl_nf_hook_refcnt))
|
|
return;
|
|
|
|
vnet->ipvl_nf_hook_refcnt--;
|
|
if (!vnet->ipvl_nf_hook_refcnt)
|
|
nf_unregister_net_hooks(net, ipvl_nfops,
|
|
ARRAY_SIZE(ipvl_nfops));
|
|
}
|
|
|
|
static int ipvlan_set_port_mode(struct ipvl_port *port, u16 nval)
|
|
{
|
|
struct ipvl_dev *ipvlan;
|
|
struct net_device *mdev = port->dev;
|
|
int err = 0;
|
|
|
|
ASSERT_RTNL();
|
|
if (port->mode != nval) {
|
|
if (nval == IPVLAN_MODE_L3S) {
|
|
/* New mode is L3S */
|
|
err = ipvlan_register_nf_hook(read_pnet(&port->pnet));
|
|
if (!err) {
|
|
mdev->l3mdev_ops = &ipvl_l3mdev_ops;
|
|
mdev->priv_flags |= IFF_L3MDEV_MASTER;
|
|
} else
|
|
return err;
|
|
} else if (port->mode == IPVLAN_MODE_L3S) {
|
|
/* Old mode was L3S */
|
|
mdev->priv_flags &= ~IFF_L3MDEV_MASTER;
|
|
ipvlan_unregister_nf_hook(read_pnet(&port->pnet));
|
|
mdev->l3mdev_ops = NULL;
|
|
}
|
|
list_for_each_entry(ipvlan, &port->ipvlans, pnode) {
|
|
if (nval == IPVLAN_MODE_L3 || nval == IPVLAN_MODE_L3S)
|
|
ipvlan->dev->flags |= IFF_NOARP;
|
|
else
|
|
ipvlan->dev->flags &= ~IFF_NOARP;
|
|
}
|
|
port->mode = nval;
|
|
}
|
|
return err;
|
|
}
|
|
|
|
static int ipvlan_port_create(struct net_device *dev)
|
|
{
|
|
struct ipvl_port *port;
|
|
int err, idx;
|
|
|
|
if (dev->type != ARPHRD_ETHER || dev->flags & IFF_LOOPBACK) {
|
|
netdev_err(dev, "Master is either lo or non-ether device\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (netdev_is_rx_handler_busy(dev)) {
|
|
netdev_err(dev, "Device is already in use.\n");
|
|
return -EBUSY;
|
|
}
|
|
|
|
port = kzalloc(sizeof(struct ipvl_port), GFP_KERNEL);
|
|
if (!port)
|
|
return -ENOMEM;
|
|
|
|
write_pnet(&port->pnet, dev_net(dev));
|
|
port->dev = dev;
|
|
port->mode = IPVLAN_MODE_L3;
|
|
INIT_LIST_HEAD(&port->ipvlans);
|
|
for (idx = 0; idx < IPVLAN_HASH_SIZE; idx++)
|
|
INIT_HLIST_HEAD(&port->hlhead[idx]);
|
|
|
|
skb_queue_head_init(&port->backlog);
|
|
INIT_WORK(&port->wq, ipvlan_process_multicast);
|
|
ida_init(&port->ida);
|
|
port->dev_id_start = 1;
|
|
|
|
err = netdev_rx_handler_register(dev, ipvlan_handle_frame, port);
|
|
if (err)
|
|
goto err;
|
|
|
|
dev->priv_flags |= IFF_IPVLAN_MASTER;
|
|
return 0;
|
|
|
|
err:
|
|
kfree(port);
|
|
return err;
|
|
}
|
|
|
|
static void ipvlan_port_destroy(struct net_device *dev)
|
|
{
|
|
struct ipvl_port *port = ipvlan_port_get_rtnl(dev);
|
|
struct sk_buff *skb;
|
|
|
|
dev->priv_flags &= ~IFF_IPVLAN_MASTER;
|
|
if (port->mode == IPVLAN_MODE_L3S) {
|
|
dev->priv_flags &= ~IFF_L3MDEV_MASTER;
|
|
ipvlan_unregister_nf_hook(dev_net(dev));
|
|
dev->l3mdev_ops = NULL;
|
|
}
|
|
netdev_rx_handler_unregister(dev);
|
|
cancel_work_sync(&port->wq);
|
|
while ((skb = __skb_dequeue(&port->backlog)) != NULL) {
|
|
if (skb->dev)
|
|
dev_put(skb->dev);
|
|
kfree_skb(skb);
|
|
}
|
|
ida_destroy(&port->ida);
|
|
kfree(port);
|
|
}
|
|
|
|
#define IPVLAN_FEATURES \
|
|
(NETIF_F_SG | NETIF_F_HW_CSUM | NETIF_F_HIGHDMA | NETIF_F_FRAGLIST | \
|
|
NETIF_F_GSO | NETIF_F_TSO | NETIF_F_GSO_ROBUST | \
|
|
NETIF_F_TSO_ECN | NETIF_F_TSO6 | NETIF_F_GRO | NETIF_F_RXCSUM | \
|
|
NETIF_F_HW_VLAN_CTAG_FILTER | NETIF_F_HW_VLAN_STAG_FILTER)
|
|
|
|
#define IPVLAN_STATE_MASK \
|
|
((1<<__LINK_STATE_NOCARRIER) | (1<<__LINK_STATE_DORMANT))
|
|
|
|
static int ipvlan_init(struct net_device *dev)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
const struct net_device *phy_dev = ipvlan->phy_dev;
|
|
struct ipvl_port *port = ipvlan->port;
|
|
|
|
dev->state = (dev->state & ~IPVLAN_STATE_MASK) |
|
|
(phy_dev->state & IPVLAN_STATE_MASK);
|
|
dev->features = phy_dev->features & IPVLAN_FEATURES;
|
|
dev->features |= NETIF_F_LLTX;
|
|
dev->gso_max_size = phy_dev->gso_max_size;
|
|
dev->gso_max_segs = phy_dev->gso_max_segs;
|
|
dev->hard_header_len = phy_dev->hard_header_len;
|
|
|
|
netdev_lockdep_set_classes(dev);
|
|
|
|
ipvlan->pcpu_stats = netdev_alloc_pcpu_stats(struct ipvl_pcpu_stats);
|
|
if (!ipvlan->pcpu_stats)
|
|
return -ENOMEM;
|
|
|
|
port->count += 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ipvlan_uninit(struct net_device *dev)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
struct ipvl_port *port = ipvlan->port;
|
|
|
|
free_percpu(ipvlan->pcpu_stats);
|
|
|
|
port->count -= 1;
|
|
if (!port->count)
|
|
ipvlan_port_destroy(port->dev);
|
|
}
|
|
|
|
static int ipvlan_open(struct net_device *dev)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
struct net_device *phy_dev = ipvlan->phy_dev;
|
|
struct ipvl_addr *addr;
|
|
|
|
if (ipvlan->port->mode == IPVLAN_MODE_L3 ||
|
|
ipvlan->port->mode == IPVLAN_MODE_L3S)
|
|
dev->flags |= IFF_NOARP;
|
|
else
|
|
dev->flags &= ~IFF_NOARP;
|
|
|
|
list_for_each_entry(addr, &ipvlan->addrs, anode)
|
|
ipvlan_ht_addr_add(ipvlan, addr);
|
|
|
|
return dev_uc_add(phy_dev, phy_dev->dev_addr);
|
|
}
|
|
|
|
static int ipvlan_stop(struct net_device *dev)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
struct net_device *phy_dev = ipvlan->phy_dev;
|
|
struct ipvl_addr *addr;
|
|
|
|
dev_uc_unsync(phy_dev, dev);
|
|
dev_mc_unsync(phy_dev, dev);
|
|
|
|
dev_uc_del(phy_dev, phy_dev->dev_addr);
|
|
|
|
list_for_each_entry(addr, &ipvlan->addrs, anode)
|
|
ipvlan_ht_addr_del(addr);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static netdev_tx_t ipvlan_start_xmit(struct sk_buff *skb,
|
|
struct net_device *dev)
|
|
{
|
|
const struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
int skblen = skb->len;
|
|
int ret;
|
|
|
|
ret = ipvlan_queue_xmit(skb, dev);
|
|
if (likely(ret == NET_XMIT_SUCCESS || ret == NET_XMIT_CN)) {
|
|
struct ipvl_pcpu_stats *pcptr;
|
|
|
|
pcptr = this_cpu_ptr(ipvlan->pcpu_stats);
|
|
|
|
u64_stats_update_begin(&pcptr->syncp);
|
|
pcptr->tx_pkts++;
|
|
pcptr->tx_bytes += skblen;
|
|
u64_stats_update_end(&pcptr->syncp);
|
|
} else {
|
|
this_cpu_inc(ipvlan->pcpu_stats->tx_drps);
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static netdev_features_t ipvlan_fix_features(struct net_device *dev,
|
|
netdev_features_t features)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
|
|
return features & (ipvlan->sfeatures | ~IPVLAN_FEATURES);
|
|
}
|
|
|
|
static void ipvlan_change_rx_flags(struct net_device *dev, int change)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
struct net_device *phy_dev = ipvlan->phy_dev;
|
|
|
|
if (change & IFF_ALLMULTI)
|
|
dev_set_allmulti(phy_dev, dev->flags & IFF_ALLMULTI? 1 : -1);
|
|
}
|
|
|
|
static void ipvlan_set_multicast_mac_filter(struct net_device *dev)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
|
|
if (dev->flags & (IFF_PROMISC | IFF_ALLMULTI)) {
|
|
bitmap_fill(ipvlan->mac_filters, IPVLAN_MAC_FILTER_SIZE);
|
|
} else {
|
|
struct netdev_hw_addr *ha;
|
|
DECLARE_BITMAP(mc_filters, IPVLAN_MAC_FILTER_SIZE);
|
|
|
|
bitmap_zero(mc_filters, IPVLAN_MAC_FILTER_SIZE);
|
|
netdev_for_each_mc_addr(ha, dev)
|
|
__set_bit(ipvlan_mac_hash(ha->addr), mc_filters);
|
|
|
|
/* Turn-on broadcast bit irrespective of address family,
|
|
* since broadcast is deferred to a work-queue, hence no
|
|
* impact on fast-path processing.
|
|
*/
|
|
__set_bit(ipvlan_mac_hash(dev->broadcast), mc_filters);
|
|
|
|
bitmap_copy(ipvlan->mac_filters, mc_filters,
|
|
IPVLAN_MAC_FILTER_SIZE);
|
|
}
|
|
dev_uc_sync(ipvlan->phy_dev, dev);
|
|
dev_mc_sync(ipvlan->phy_dev, dev);
|
|
}
|
|
|
|
static void ipvlan_get_stats64(struct net_device *dev,
|
|
struct rtnl_link_stats64 *s)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
|
|
if (ipvlan->pcpu_stats) {
|
|
struct ipvl_pcpu_stats *pcptr;
|
|
u64 rx_pkts, rx_bytes, rx_mcast, tx_pkts, tx_bytes;
|
|
u32 rx_errs = 0, tx_drps = 0;
|
|
u32 strt;
|
|
int idx;
|
|
|
|
for_each_possible_cpu(idx) {
|
|
pcptr = per_cpu_ptr(ipvlan->pcpu_stats, idx);
|
|
do {
|
|
strt= u64_stats_fetch_begin_irq(&pcptr->syncp);
|
|
rx_pkts = pcptr->rx_pkts;
|
|
rx_bytes = pcptr->rx_bytes;
|
|
rx_mcast = pcptr->rx_mcast;
|
|
tx_pkts = pcptr->tx_pkts;
|
|
tx_bytes = pcptr->tx_bytes;
|
|
} while (u64_stats_fetch_retry_irq(&pcptr->syncp,
|
|
strt));
|
|
|
|
s->rx_packets += rx_pkts;
|
|
s->rx_bytes += rx_bytes;
|
|
s->multicast += rx_mcast;
|
|
s->tx_packets += tx_pkts;
|
|
s->tx_bytes += tx_bytes;
|
|
|
|
/* u32 values are updated without syncp protection. */
|
|
rx_errs += pcptr->rx_errs;
|
|
tx_drps += pcptr->tx_drps;
|
|
}
|
|
s->rx_errors = rx_errs;
|
|
s->rx_dropped = rx_errs;
|
|
s->tx_dropped = tx_drps;
|
|
}
|
|
}
|
|
|
|
static int ipvlan_vlan_rx_add_vid(struct net_device *dev, __be16 proto, u16 vid)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
struct net_device *phy_dev = ipvlan->phy_dev;
|
|
|
|
return vlan_vid_add(phy_dev, proto, vid);
|
|
}
|
|
|
|
static int ipvlan_vlan_rx_kill_vid(struct net_device *dev, __be16 proto,
|
|
u16 vid)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
struct net_device *phy_dev = ipvlan->phy_dev;
|
|
|
|
vlan_vid_del(phy_dev, proto, vid);
|
|
return 0;
|
|
}
|
|
|
|
static int ipvlan_get_iflink(const struct net_device *dev)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
|
|
return ipvlan->phy_dev->ifindex;
|
|
}
|
|
|
|
static const struct net_device_ops ipvlan_netdev_ops = {
|
|
.ndo_init = ipvlan_init,
|
|
.ndo_uninit = ipvlan_uninit,
|
|
.ndo_open = ipvlan_open,
|
|
.ndo_stop = ipvlan_stop,
|
|
.ndo_start_xmit = ipvlan_start_xmit,
|
|
.ndo_fix_features = ipvlan_fix_features,
|
|
.ndo_change_rx_flags = ipvlan_change_rx_flags,
|
|
.ndo_set_rx_mode = ipvlan_set_multicast_mac_filter,
|
|
.ndo_get_stats64 = ipvlan_get_stats64,
|
|
.ndo_vlan_rx_add_vid = ipvlan_vlan_rx_add_vid,
|
|
.ndo_vlan_rx_kill_vid = ipvlan_vlan_rx_kill_vid,
|
|
.ndo_get_iflink = ipvlan_get_iflink,
|
|
};
|
|
|
|
static int ipvlan_hard_header(struct sk_buff *skb, struct net_device *dev,
|
|
unsigned short type, const void *daddr,
|
|
const void *saddr, unsigned len)
|
|
{
|
|
const struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
struct net_device *phy_dev = ipvlan->phy_dev;
|
|
|
|
/* TODO Probably use a different field than dev_addr so that the
|
|
* mac-address on the virtual device is portable and can be carried
|
|
* while the packets use the mac-addr on the physical device.
|
|
*/
|
|
return dev_hard_header(skb, phy_dev, type, daddr,
|
|
saddr ? : dev->dev_addr, len);
|
|
}
|
|
|
|
static const struct header_ops ipvlan_header_ops = {
|
|
.create = ipvlan_hard_header,
|
|
.parse = eth_header_parse,
|
|
.cache = eth_header_cache,
|
|
.cache_update = eth_header_cache_update,
|
|
};
|
|
|
|
static int ipvlan_ethtool_get_link_ksettings(struct net_device *dev,
|
|
struct ethtool_link_ksettings *cmd)
|
|
{
|
|
const struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
|
|
return __ethtool_get_link_ksettings(ipvlan->phy_dev, cmd);
|
|
}
|
|
|
|
static void ipvlan_ethtool_get_drvinfo(struct net_device *dev,
|
|
struct ethtool_drvinfo *drvinfo)
|
|
{
|
|
strlcpy(drvinfo->driver, IPVLAN_DRV, sizeof(drvinfo->driver));
|
|
strlcpy(drvinfo->version, IPV_DRV_VER, sizeof(drvinfo->version));
|
|
}
|
|
|
|
static u32 ipvlan_ethtool_get_msglevel(struct net_device *dev)
|
|
{
|
|
const struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
|
|
return ipvlan->msg_enable;
|
|
}
|
|
|
|
static void ipvlan_ethtool_set_msglevel(struct net_device *dev, u32 value)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
|
|
ipvlan->msg_enable = value;
|
|
}
|
|
|
|
static const struct ethtool_ops ipvlan_ethtool_ops = {
|
|
.get_link = ethtool_op_get_link,
|
|
.get_link_ksettings = ipvlan_ethtool_get_link_ksettings,
|
|
.get_drvinfo = ipvlan_ethtool_get_drvinfo,
|
|
.get_msglevel = ipvlan_ethtool_get_msglevel,
|
|
.set_msglevel = ipvlan_ethtool_set_msglevel,
|
|
};
|
|
|
|
static int ipvlan_nl_changelink(struct net_device *dev,
|
|
struct nlattr *tb[], struct nlattr *data[],
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
struct ipvl_port *port = ipvlan_port_get_rtnl(ipvlan->phy_dev);
|
|
int err = 0;
|
|
|
|
if (data && data[IFLA_IPVLAN_MODE]) {
|
|
u16 nmode = nla_get_u16(data[IFLA_IPVLAN_MODE]);
|
|
|
|
err = ipvlan_set_port_mode(port, nmode);
|
|
}
|
|
return err;
|
|
}
|
|
|
|
static size_t ipvlan_nl_getsize(const struct net_device *dev)
|
|
{
|
|
return (0
|
|
+ nla_total_size(2) /* IFLA_IPVLAN_MODE */
|
|
);
|
|
}
|
|
|
|
static int ipvlan_nl_validate(struct nlattr *tb[], struct nlattr *data[],
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
if (data && data[IFLA_IPVLAN_MODE]) {
|
|
u16 mode = nla_get_u16(data[IFLA_IPVLAN_MODE]);
|
|
|
|
if (mode < IPVLAN_MODE_L2 || mode >= IPVLAN_MODE_MAX)
|
|
return -EINVAL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int ipvlan_nl_fillinfo(struct sk_buff *skb,
|
|
const struct net_device *dev)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
struct ipvl_port *port = ipvlan_port_get_rtnl(ipvlan->phy_dev);
|
|
int ret = -EINVAL;
|
|
|
|
if (!port)
|
|
goto err;
|
|
|
|
ret = -EMSGSIZE;
|
|
if (nla_put_u16(skb, IFLA_IPVLAN_MODE, port->mode))
|
|
goto err;
|
|
|
|
return 0;
|
|
|
|
err:
|
|
return ret;
|
|
}
|
|
|
|
int ipvlan_link_new(struct net *src_net, struct net_device *dev,
|
|
struct nlattr *tb[], struct nlattr *data[],
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
struct ipvl_port *port;
|
|
struct net_device *phy_dev;
|
|
int err;
|
|
u16 mode = IPVLAN_MODE_L3;
|
|
bool create = false;
|
|
|
|
if (!tb[IFLA_LINK])
|
|
return -EINVAL;
|
|
|
|
phy_dev = __dev_get_by_index(src_net, nla_get_u32(tb[IFLA_LINK]));
|
|
if (!phy_dev)
|
|
return -ENODEV;
|
|
|
|
if (netif_is_ipvlan(phy_dev)) {
|
|
struct ipvl_dev *tmp = netdev_priv(phy_dev);
|
|
|
|
phy_dev = tmp->phy_dev;
|
|
} else if (!netif_is_ipvlan_port(phy_dev)) {
|
|
err = ipvlan_port_create(phy_dev);
|
|
if (err < 0)
|
|
return err;
|
|
create = true;
|
|
}
|
|
|
|
if (data && data[IFLA_IPVLAN_MODE])
|
|
mode = nla_get_u16(data[IFLA_IPVLAN_MODE]);
|
|
|
|
port = ipvlan_port_get_rtnl(phy_dev);
|
|
ipvlan->phy_dev = phy_dev;
|
|
ipvlan->dev = dev;
|
|
ipvlan->port = port;
|
|
ipvlan->sfeatures = IPVLAN_FEATURES;
|
|
ipvlan_adjust_mtu(ipvlan, phy_dev);
|
|
INIT_LIST_HEAD(&ipvlan->addrs);
|
|
|
|
/* If the port-id base is at the MAX value, then wrap it around and
|
|
* begin from 0x1 again. This may be due to a busy system where lots
|
|
* of slaves are getting created and deleted.
|
|
*/
|
|
if (port->dev_id_start == 0xFFFE)
|
|
port->dev_id_start = 0x1;
|
|
|
|
/* Since L2 address is shared among all IPvlan slaves including
|
|
* master, use unique 16 bit dev-ids to diffentiate among them.
|
|
* Assign IDs between 0x1 and 0xFFFE (used by the master) to each
|
|
* slave link [see addrconf_ifid_eui48()].
|
|
*/
|
|
err = ida_simple_get(&port->ida, port->dev_id_start, 0xFFFE,
|
|
GFP_KERNEL);
|
|
if (err < 0)
|
|
err = ida_simple_get(&port->ida, 0x1, port->dev_id_start,
|
|
GFP_KERNEL);
|
|
if (err < 0)
|
|
goto destroy_ipvlan_port;
|
|
dev->dev_id = err;
|
|
/* Increment id-base to the next slot for the future assignment */
|
|
port->dev_id_start = err + 1;
|
|
|
|
/* TODO Probably put random address here to be presented to the
|
|
* world but keep using the physical-dev address for the outgoing
|
|
* packets.
|
|
*/
|
|
memcpy(dev->dev_addr, phy_dev->dev_addr, ETH_ALEN);
|
|
|
|
dev->priv_flags |= IFF_IPVLAN_SLAVE;
|
|
|
|
err = register_netdevice(dev);
|
|
if (err < 0)
|
|
goto remove_ida;
|
|
|
|
err = netdev_upper_dev_link(phy_dev, dev);
|
|
if (err) {
|
|
goto unregister_netdev;
|
|
}
|
|
err = ipvlan_set_port_mode(port, mode);
|
|
if (err) {
|
|
goto unlink_netdev;
|
|
}
|
|
|
|
list_add_tail_rcu(&ipvlan->pnode, &port->ipvlans);
|
|
netif_stacked_transfer_operstate(phy_dev, dev);
|
|
return 0;
|
|
|
|
unlink_netdev:
|
|
netdev_upper_dev_unlink(phy_dev, dev);
|
|
unregister_netdev:
|
|
unregister_netdevice(dev);
|
|
remove_ida:
|
|
ida_simple_remove(&port->ida, dev->dev_id);
|
|
destroy_ipvlan_port:
|
|
if (create)
|
|
ipvlan_port_destroy(phy_dev);
|
|
return err;
|
|
}
|
|
EXPORT_SYMBOL_GPL(ipvlan_link_new);
|
|
|
|
void ipvlan_link_delete(struct net_device *dev, struct list_head *head)
|
|
{
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
struct ipvl_addr *addr, *next;
|
|
|
|
list_for_each_entry_safe(addr, next, &ipvlan->addrs, anode) {
|
|
ipvlan_ht_addr_del(addr);
|
|
list_del(&addr->anode);
|
|
kfree_rcu(addr, rcu);
|
|
}
|
|
|
|
ida_simple_remove(&ipvlan->port->ida, dev->dev_id);
|
|
list_del_rcu(&ipvlan->pnode);
|
|
unregister_netdevice_queue(dev, head);
|
|
netdev_upper_dev_unlink(ipvlan->phy_dev, dev);
|
|
}
|
|
EXPORT_SYMBOL_GPL(ipvlan_link_delete);
|
|
|
|
void ipvlan_link_setup(struct net_device *dev)
|
|
{
|
|
ether_setup(dev);
|
|
|
|
dev->priv_flags &= ~(IFF_XMIT_DST_RELEASE | IFF_TX_SKB_SHARING);
|
|
dev->priv_flags |= IFF_UNICAST_FLT | IFF_NO_QUEUE;
|
|
dev->netdev_ops = &ipvlan_netdev_ops;
|
|
dev->needs_free_netdev = true;
|
|
dev->header_ops = &ipvlan_header_ops;
|
|
dev->ethtool_ops = &ipvlan_ethtool_ops;
|
|
}
|
|
EXPORT_SYMBOL_GPL(ipvlan_link_setup);
|
|
|
|
static const struct nla_policy ipvlan_nl_policy[IFLA_IPVLAN_MAX + 1] =
|
|
{
|
|
[IFLA_IPVLAN_MODE] = { .type = NLA_U16 },
|
|
};
|
|
|
|
static struct rtnl_link_ops ipvlan_link_ops = {
|
|
.kind = "ipvlan",
|
|
.priv_size = sizeof(struct ipvl_dev),
|
|
|
|
.setup = ipvlan_link_setup,
|
|
.newlink = ipvlan_link_new,
|
|
.dellink = ipvlan_link_delete,
|
|
};
|
|
|
|
int ipvlan_link_register(struct rtnl_link_ops *ops)
|
|
{
|
|
ops->get_size = ipvlan_nl_getsize;
|
|
ops->policy = ipvlan_nl_policy;
|
|
ops->validate = ipvlan_nl_validate;
|
|
ops->fill_info = ipvlan_nl_fillinfo;
|
|
ops->changelink = ipvlan_nl_changelink;
|
|
ops->maxtype = IFLA_IPVLAN_MAX;
|
|
return rtnl_link_register(ops);
|
|
}
|
|
EXPORT_SYMBOL_GPL(ipvlan_link_register);
|
|
|
|
static int ipvlan_device_event(struct notifier_block *unused,
|
|
unsigned long event, void *ptr)
|
|
{
|
|
struct net_device *dev = netdev_notifier_info_to_dev(ptr);
|
|
struct ipvl_dev *ipvlan, *next;
|
|
struct ipvl_port *port;
|
|
LIST_HEAD(lst_kill);
|
|
|
|
if (!netif_is_ipvlan_port(dev))
|
|
return NOTIFY_DONE;
|
|
|
|
port = ipvlan_port_get_rtnl(dev);
|
|
|
|
switch (event) {
|
|
case NETDEV_CHANGE:
|
|
list_for_each_entry(ipvlan, &port->ipvlans, pnode)
|
|
netif_stacked_transfer_operstate(ipvlan->phy_dev,
|
|
ipvlan->dev);
|
|
break;
|
|
|
|
case NETDEV_REGISTER: {
|
|
struct net *oldnet, *newnet = dev_net(dev);
|
|
struct ipvlan_netns *old_vnet;
|
|
|
|
oldnet = read_pnet(&port->pnet);
|
|
if (net_eq(newnet, oldnet))
|
|
break;
|
|
|
|
write_pnet(&port->pnet, newnet);
|
|
|
|
old_vnet = net_generic(oldnet, ipvlan_netid);
|
|
if (!old_vnet->ipvl_nf_hook_refcnt)
|
|
break;
|
|
|
|
ipvlan_register_nf_hook(newnet);
|
|
ipvlan_unregister_nf_hook(oldnet);
|
|
break;
|
|
}
|
|
case NETDEV_UNREGISTER:
|
|
if (dev->reg_state != NETREG_UNREGISTERING)
|
|
break;
|
|
|
|
list_for_each_entry_safe(ipvlan, next, &port->ipvlans,
|
|
pnode)
|
|
ipvlan->dev->rtnl_link_ops->dellink(ipvlan->dev,
|
|
&lst_kill);
|
|
unregister_netdevice_many(&lst_kill);
|
|
break;
|
|
|
|
case NETDEV_FEAT_CHANGE:
|
|
list_for_each_entry(ipvlan, &port->ipvlans, pnode) {
|
|
ipvlan->dev->features = dev->features & IPVLAN_FEATURES;
|
|
ipvlan->dev->gso_max_size = dev->gso_max_size;
|
|
ipvlan->dev->gso_max_segs = dev->gso_max_segs;
|
|
netdev_features_change(ipvlan->dev);
|
|
}
|
|
break;
|
|
|
|
case NETDEV_CHANGEMTU:
|
|
list_for_each_entry(ipvlan, &port->ipvlans, pnode)
|
|
ipvlan_adjust_mtu(ipvlan, dev);
|
|
break;
|
|
|
|
case NETDEV_PRE_TYPE_CHANGE:
|
|
/* Forbid underlying device to change its type. */
|
|
return NOTIFY_BAD;
|
|
}
|
|
return NOTIFY_DONE;
|
|
}
|
|
|
|
static int ipvlan_add_addr(struct ipvl_dev *ipvlan, void *iaddr, bool is_v6)
|
|
{
|
|
struct ipvl_addr *addr;
|
|
|
|
addr = kzalloc(sizeof(struct ipvl_addr), GFP_ATOMIC);
|
|
if (!addr)
|
|
return -ENOMEM;
|
|
|
|
addr->master = ipvlan;
|
|
if (is_v6) {
|
|
memcpy(&addr->ip6addr, iaddr, sizeof(struct in6_addr));
|
|
addr->atype = IPVL_IPV6;
|
|
} else {
|
|
memcpy(&addr->ip4addr, iaddr, sizeof(struct in_addr));
|
|
addr->atype = IPVL_IPV4;
|
|
}
|
|
list_add_tail(&addr->anode, &ipvlan->addrs);
|
|
|
|
/* If the interface is not up, the address will be added to the hash
|
|
* list by ipvlan_open.
|
|
*/
|
|
if (netif_running(ipvlan->dev))
|
|
ipvlan_ht_addr_add(ipvlan, addr);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void ipvlan_del_addr(struct ipvl_dev *ipvlan, void *iaddr, bool is_v6)
|
|
{
|
|
struct ipvl_addr *addr;
|
|
|
|
addr = ipvlan_find_addr(ipvlan, iaddr, is_v6);
|
|
if (!addr)
|
|
return;
|
|
|
|
ipvlan_ht_addr_del(addr);
|
|
list_del(&addr->anode);
|
|
kfree_rcu(addr, rcu);
|
|
|
|
return;
|
|
}
|
|
|
|
static int ipvlan_add_addr6(struct ipvl_dev *ipvlan, struct in6_addr *ip6_addr)
|
|
{
|
|
if (ipvlan_addr_busy(ipvlan->port, ip6_addr, true)) {
|
|
netif_err(ipvlan, ifup, ipvlan->dev,
|
|
"Failed to add IPv6=%pI6c addr for %s intf\n",
|
|
ip6_addr, ipvlan->dev->name);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return ipvlan_add_addr(ipvlan, ip6_addr, true);
|
|
}
|
|
|
|
static void ipvlan_del_addr6(struct ipvl_dev *ipvlan, struct in6_addr *ip6_addr)
|
|
{
|
|
return ipvlan_del_addr(ipvlan, ip6_addr, true);
|
|
}
|
|
|
|
static int ipvlan_addr6_event(struct notifier_block *unused,
|
|
unsigned long event, void *ptr)
|
|
{
|
|
struct inet6_ifaddr *if6 = (struct inet6_ifaddr *)ptr;
|
|
struct net_device *dev = (struct net_device *)if6->idev->dev;
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
|
|
/* FIXME IPv6 autoconf calls us from bh without RTNL */
|
|
if (in_softirq())
|
|
return NOTIFY_DONE;
|
|
|
|
if (!netif_is_ipvlan(dev))
|
|
return NOTIFY_DONE;
|
|
|
|
if (!ipvlan || !ipvlan->port)
|
|
return NOTIFY_DONE;
|
|
|
|
switch (event) {
|
|
case NETDEV_UP:
|
|
if (ipvlan_add_addr6(ipvlan, &if6->addr))
|
|
return NOTIFY_BAD;
|
|
break;
|
|
|
|
case NETDEV_DOWN:
|
|
ipvlan_del_addr6(ipvlan, &if6->addr);
|
|
break;
|
|
}
|
|
|
|
return NOTIFY_OK;
|
|
}
|
|
|
|
static int ipvlan_addr6_validator_event(struct notifier_block *unused,
|
|
unsigned long event, void *ptr)
|
|
{
|
|
struct in6_validator_info *i6vi = (struct in6_validator_info *)ptr;
|
|
struct net_device *dev = (struct net_device *)i6vi->i6vi_dev->dev;
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
|
|
/* FIXME IPv6 autoconf calls us from bh without RTNL */
|
|
if (in_softirq())
|
|
return NOTIFY_DONE;
|
|
|
|
if (!netif_is_ipvlan(dev))
|
|
return NOTIFY_DONE;
|
|
|
|
if (!ipvlan || !ipvlan->port)
|
|
return NOTIFY_DONE;
|
|
|
|
switch (event) {
|
|
case NETDEV_UP:
|
|
if (ipvlan_addr_busy(ipvlan->port, &i6vi->i6vi_addr, true))
|
|
return notifier_from_errno(-EADDRINUSE);
|
|
break;
|
|
}
|
|
|
|
return NOTIFY_OK;
|
|
}
|
|
|
|
static int ipvlan_add_addr4(struct ipvl_dev *ipvlan, struct in_addr *ip4_addr)
|
|
{
|
|
if (ipvlan_addr_busy(ipvlan->port, ip4_addr, false)) {
|
|
netif_err(ipvlan, ifup, ipvlan->dev,
|
|
"Failed to add IPv4=%pI4 on %s intf.\n",
|
|
ip4_addr, ipvlan->dev->name);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return ipvlan_add_addr(ipvlan, ip4_addr, false);
|
|
}
|
|
|
|
static void ipvlan_del_addr4(struct ipvl_dev *ipvlan, struct in_addr *ip4_addr)
|
|
{
|
|
return ipvlan_del_addr(ipvlan, ip4_addr, false);
|
|
}
|
|
|
|
static int ipvlan_addr4_event(struct notifier_block *unused,
|
|
unsigned long event, void *ptr)
|
|
{
|
|
struct in_ifaddr *if4 = (struct in_ifaddr *)ptr;
|
|
struct net_device *dev = (struct net_device *)if4->ifa_dev->dev;
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
struct in_addr ip4_addr;
|
|
|
|
if (!netif_is_ipvlan(dev))
|
|
return NOTIFY_DONE;
|
|
|
|
if (!ipvlan || !ipvlan->port)
|
|
return NOTIFY_DONE;
|
|
|
|
switch (event) {
|
|
case NETDEV_UP:
|
|
ip4_addr.s_addr = if4->ifa_address;
|
|
if (ipvlan_add_addr4(ipvlan, &ip4_addr))
|
|
return NOTIFY_BAD;
|
|
break;
|
|
|
|
case NETDEV_DOWN:
|
|
ip4_addr.s_addr = if4->ifa_address;
|
|
ipvlan_del_addr4(ipvlan, &ip4_addr);
|
|
break;
|
|
}
|
|
|
|
return NOTIFY_OK;
|
|
}
|
|
|
|
static int ipvlan_addr4_validator_event(struct notifier_block *unused,
|
|
unsigned long event, void *ptr)
|
|
{
|
|
struct in_validator_info *ivi = (struct in_validator_info *)ptr;
|
|
struct net_device *dev = (struct net_device *)ivi->ivi_dev->dev;
|
|
struct ipvl_dev *ipvlan = netdev_priv(dev);
|
|
|
|
if (!netif_is_ipvlan(dev))
|
|
return NOTIFY_DONE;
|
|
|
|
if (!ipvlan || !ipvlan->port)
|
|
return NOTIFY_DONE;
|
|
|
|
switch (event) {
|
|
case NETDEV_UP:
|
|
if (ipvlan_addr_busy(ipvlan->port, &ivi->ivi_addr, false))
|
|
return notifier_from_errno(-EADDRINUSE);
|
|
break;
|
|
}
|
|
|
|
return NOTIFY_OK;
|
|
}
|
|
|
|
static struct notifier_block ipvlan_addr4_notifier_block __read_mostly = {
|
|
.notifier_call = ipvlan_addr4_event,
|
|
};
|
|
|
|
static struct notifier_block ipvlan_addr4_vtor_notifier_block __read_mostly = {
|
|
.notifier_call = ipvlan_addr4_validator_event,
|
|
};
|
|
|
|
static struct notifier_block ipvlan_notifier_block __read_mostly = {
|
|
.notifier_call = ipvlan_device_event,
|
|
};
|
|
|
|
static struct notifier_block ipvlan_addr6_notifier_block __read_mostly = {
|
|
.notifier_call = ipvlan_addr6_event,
|
|
};
|
|
|
|
static struct notifier_block ipvlan_addr6_vtor_notifier_block __read_mostly = {
|
|
.notifier_call = ipvlan_addr6_validator_event,
|
|
};
|
|
|
|
static void ipvlan_ns_exit(struct net *net)
|
|
{
|
|
struct ipvlan_netns *vnet = net_generic(net, ipvlan_netid);
|
|
|
|
if (WARN_ON_ONCE(vnet->ipvl_nf_hook_refcnt)) {
|
|
vnet->ipvl_nf_hook_refcnt = 0;
|
|
nf_unregister_net_hooks(net, ipvl_nfops,
|
|
ARRAY_SIZE(ipvl_nfops));
|
|
}
|
|
}
|
|
|
|
static struct pernet_operations ipvlan_net_ops = {
|
|
.id = &ipvlan_netid,
|
|
.size = sizeof(struct ipvlan_netns),
|
|
.exit = ipvlan_ns_exit,
|
|
};
|
|
|
|
static int __init ipvlan_init_module(void)
|
|
{
|
|
int err;
|
|
|
|
ipvlan_init_secret();
|
|
register_netdevice_notifier(&ipvlan_notifier_block);
|
|
register_inet6addr_notifier(&ipvlan_addr6_notifier_block);
|
|
register_inet6addr_validator_notifier(
|
|
&ipvlan_addr6_vtor_notifier_block);
|
|
register_inetaddr_notifier(&ipvlan_addr4_notifier_block);
|
|
register_inetaddr_validator_notifier(&ipvlan_addr4_vtor_notifier_block);
|
|
|
|
err = register_pernet_subsys(&ipvlan_net_ops);
|
|
if (err < 0)
|
|
goto error;
|
|
|
|
err = ipvlan_link_register(&ipvlan_link_ops);
|
|
if (err < 0) {
|
|
unregister_pernet_subsys(&ipvlan_net_ops);
|
|
goto error;
|
|
}
|
|
|
|
return 0;
|
|
error:
|
|
unregister_inetaddr_notifier(&ipvlan_addr4_notifier_block);
|
|
unregister_inetaddr_validator_notifier(
|
|
&ipvlan_addr4_vtor_notifier_block);
|
|
unregister_inet6addr_notifier(&ipvlan_addr6_notifier_block);
|
|
unregister_inet6addr_validator_notifier(
|
|
&ipvlan_addr6_vtor_notifier_block);
|
|
unregister_netdevice_notifier(&ipvlan_notifier_block);
|
|
return err;
|
|
}
|
|
|
|
static void __exit ipvlan_cleanup_module(void)
|
|
{
|
|
rtnl_link_unregister(&ipvlan_link_ops);
|
|
unregister_pernet_subsys(&ipvlan_net_ops);
|
|
unregister_netdevice_notifier(&ipvlan_notifier_block);
|
|
unregister_inetaddr_notifier(&ipvlan_addr4_notifier_block);
|
|
unregister_inetaddr_validator_notifier(
|
|
&ipvlan_addr4_vtor_notifier_block);
|
|
unregister_inet6addr_notifier(&ipvlan_addr6_notifier_block);
|
|
unregister_inet6addr_validator_notifier(
|
|
&ipvlan_addr6_vtor_notifier_block);
|
|
}
|
|
|
|
module_init(ipvlan_init_module);
|
|
module_exit(ipvlan_cleanup_module);
|
|
|
|
MODULE_LICENSE("GPL");
|
|
MODULE_AUTHOR("Mahesh Bandewar <maheshb@google.com>");
|
|
MODULE_DESCRIPTION("Driver for L3 (IPv6/IPv4) based VLANs");
|
|
MODULE_ALIAS_RTNL_LINK("ipvlan");
|