mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-16 05:47:35 +07:00
23f4ffedb7
skb->cb may contain data from previous layers. In the observed scenario, the garbage data were misinterpreted as IP6CB(skb)->frag_max_size, so that small packets sent through the tunnel are mistakenly fragmented. This patch unconditionally clears the control buffer in ip6tunnel_xmit(), which affects ip6_tunnel, ip6_udp_tunnel and ip6_gre. Currently none of these tunnels set IP6CB(skb)->flags, otherwise it needs to be done earlier. Cc: stable@vger.kernel.org Signed-off-by: Eli Cooper <elicooper@gmx.com> Signed-off-by: David S. Miller <davem@davemloft.net>
158 lines
4.6 KiB
C
158 lines
4.6 KiB
C
#ifndef _NET_IP6_TUNNEL_H
|
|
#define _NET_IP6_TUNNEL_H
|
|
|
|
#include <linux/ipv6.h>
|
|
#include <linux/netdevice.h>
|
|
#include <linux/if_tunnel.h>
|
|
#include <linux/ip6_tunnel.h>
|
|
#include <net/ip_tunnels.h>
|
|
#include <net/dst_cache.h>
|
|
|
|
#define IP6TUNNEL_ERR_TIMEO (30*HZ)
|
|
|
|
/* capable of sending packets */
|
|
#define IP6_TNL_F_CAP_XMIT 0x10000
|
|
/* capable of receiving packets */
|
|
#define IP6_TNL_F_CAP_RCV 0x20000
|
|
/* determine capability on a per-packet basis */
|
|
#define IP6_TNL_F_CAP_PER_PACKET 0x40000
|
|
|
|
struct __ip6_tnl_parm {
|
|
char name[IFNAMSIZ]; /* name of tunnel device */
|
|
int link; /* ifindex of underlying L2 interface */
|
|
__u8 proto; /* tunnel protocol */
|
|
__u8 encap_limit; /* encapsulation limit for tunnel */
|
|
__u8 hop_limit; /* hop limit for tunnel */
|
|
bool collect_md;
|
|
__be32 flowinfo; /* traffic class and flowlabel for tunnel */
|
|
__u32 flags; /* tunnel flags */
|
|
struct in6_addr laddr; /* local tunnel end-point address */
|
|
struct in6_addr raddr; /* remote tunnel end-point address */
|
|
|
|
__be16 i_flags;
|
|
__be16 o_flags;
|
|
__be32 i_key;
|
|
__be32 o_key;
|
|
};
|
|
|
|
/* IPv6 tunnel */
|
|
struct ip6_tnl {
|
|
struct ip6_tnl __rcu *next; /* next tunnel in list */
|
|
struct net_device *dev; /* virtual device associated with tunnel */
|
|
struct net *net; /* netns for packet i/o */
|
|
struct __ip6_tnl_parm parms; /* tunnel configuration parameters */
|
|
struct flowi fl; /* flowi template for xmit */
|
|
struct dst_cache dst_cache; /* cached dst */
|
|
struct gro_cells gro_cells;
|
|
|
|
int err_count;
|
|
unsigned long err_time;
|
|
|
|
/* These fields used only by GRE */
|
|
__u32 i_seqno; /* The last seen seqno */
|
|
__u32 o_seqno; /* The last output seqno */
|
|
int hlen; /* tun_hlen + encap_hlen */
|
|
int tun_hlen; /* Precalculated header length */
|
|
int encap_hlen; /* Encap header length (FOU,GUE) */
|
|
struct ip_tunnel_encap encap;
|
|
int mlink;
|
|
};
|
|
|
|
struct ip6_tnl_encap_ops {
|
|
size_t (*encap_hlen)(struct ip_tunnel_encap *e);
|
|
int (*build_header)(struct sk_buff *skb, struct ip_tunnel_encap *e,
|
|
u8 *protocol, struct flowi6 *fl6);
|
|
};
|
|
|
|
#ifdef CONFIG_INET
|
|
|
|
extern const struct ip6_tnl_encap_ops __rcu *
|
|
ip6tun_encaps[MAX_IPTUN_ENCAP_OPS];
|
|
|
|
int ip6_tnl_encap_add_ops(const struct ip6_tnl_encap_ops *ops,
|
|
unsigned int num);
|
|
int ip6_tnl_encap_del_ops(const struct ip6_tnl_encap_ops *ops,
|
|
unsigned int num);
|
|
int ip6_tnl_encap_setup(struct ip6_tnl *t,
|
|
struct ip_tunnel_encap *ipencap);
|
|
|
|
static inline int ip6_encap_hlen(struct ip_tunnel_encap *e)
|
|
{
|
|
const struct ip6_tnl_encap_ops *ops;
|
|
int hlen = -EINVAL;
|
|
|
|
if (e->type == TUNNEL_ENCAP_NONE)
|
|
return 0;
|
|
|
|
if (e->type >= MAX_IPTUN_ENCAP_OPS)
|
|
return -EINVAL;
|
|
|
|
rcu_read_lock();
|
|
ops = rcu_dereference(ip6tun_encaps[e->type]);
|
|
if (likely(ops && ops->encap_hlen))
|
|
hlen = ops->encap_hlen(e);
|
|
rcu_read_unlock();
|
|
|
|
return hlen;
|
|
}
|
|
|
|
static inline int ip6_tnl_encap(struct sk_buff *skb, struct ip6_tnl *t,
|
|
u8 *protocol, struct flowi6 *fl6)
|
|
{
|
|
const struct ip6_tnl_encap_ops *ops;
|
|
int ret = -EINVAL;
|
|
|
|
if (t->encap.type == TUNNEL_ENCAP_NONE)
|
|
return 0;
|
|
|
|
if (t->encap.type >= MAX_IPTUN_ENCAP_OPS)
|
|
return -EINVAL;
|
|
|
|
rcu_read_lock();
|
|
ops = rcu_dereference(ip6tun_encaps[t->encap.type]);
|
|
if (likely(ops && ops->build_header))
|
|
ret = ops->build_header(skb, &t->encap, protocol, fl6);
|
|
rcu_read_unlock();
|
|
|
|
return ret;
|
|
}
|
|
|
|
/* Tunnel encapsulation limit destination sub-option */
|
|
|
|
struct ipv6_tlv_tnl_enc_lim {
|
|
__u8 type; /* type-code for option */
|
|
__u8 length; /* option length */
|
|
__u8 encap_limit; /* tunnel encapsulation limit */
|
|
} __packed;
|
|
|
|
int ip6_tnl_rcv_ctl(struct ip6_tnl *t, const struct in6_addr *laddr,
|
|
const struct in6_addr *raddr);
|
|
int ip6_tnl_rcv(struct ip6_tnl *tunnel, struct sk_buff *skb,
|
|
const struct tnl_ptk_info *tpi, struct metadata_dst *tun_dst,
|
|
bool log_ecn_error);
|
|
int ip6_tnl_xmit_ctl(struct ip6_tnl *t, const struct in6_addr *laddr,
|
|
const struct in6_addr *raddr);
|
|
int ip6_tnl_xmit(struct sk_buff *skb, struct net_device *dev, __u8 dsfield,
|
|
struct flowi6 *fl6, int encap_limit, __u32 *pmtu, __u8 proto);
|
|
__u16 ip6_tnl_parse_tlv_enc_lim(struct sk_buff *skb, __u8 *raw);
|
|
__u32 ip6_tnl_get_cap(struct ip6_tnl *t, const struct in6_addr *laddr,
|
|
const struct in6_addr *raddr);
|
|
struct net *ip6_tnl_get_link_net(const struct net_device *dev);
|
|
int ip6_tnl_get_iflink(const struct net_device *dev);
|
|
int ip6_tnl_change_mtu(struct net_device *dev, int new_mtu);
|
|
|
|
static inline void ip6tunnel_xmit(struct sock *sk, struct sk_buff *skb,
|
|
struct net_device *dev)
|
|
{
|
|
int pkt_len, err;
|
|
|
|
memset(skb->cb, 0, sizeof(struct inet6_skb_parm));
|
|
pkt_len = skb->len - skb_inner_network_offset(skb);
|
|
err = ip6_local_out(dev_net(skb_dst(skb)->dev), sk, skb);
|
|
if (unlikely(net_xmit_eval(err)))
|
|
pkt_len = -1;
|
|
iptunnel_xmit_stats(dev, pkt_len);
|
|
}
|
|
#endif
|
|
#endif
|