2018-07-14 06:14:01 +07:00
|
|
|
/* SPDX-License-Identifier: GPL-2.0 */
|
|
|
|
#ifndef _IPV6_FRAG_H
|
|
|
|
#define _IPV6_FRAG_H
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <net/addrconf.h>
|
|
|
|
#include <net/ipv6.h>
|
|
|
|
#include <net/inet_frag.h>
|
|
|
|
|
|
|
|
enum ip6_defrag_users {
|
|
|
|
IP6_DEFRAG_LOCAL_DELIVER,
|
|
|
|
IP6_DEFRAG_CONNTRACK_IN,
|
|
|
|
__IP6_DEFRAG_CONNTRACK_IN = IP6_DEFRAG_CONNTRACK_IN + USHRT_MAX,
|
|
|
|
IP6_DEFRAG_CONNTRACK_OUT,
|
|
|
|
__IP6_DEFRAG_CONNTRACK_OUT = IP6_DEFRAG_CONNTRACK_OUT + USHRT_MAX,
|
|
|
|
IP6_DEFRAG_CONNTRACK_BRIDGE_IN,
|
|
|
|
__IP6_DEFRAG_CONNTRACK_BRIDGE_IN = IP6_DEFRAG_CONNTRACK_BRIDGE_IN + USHRT_MAX,
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Equivalent of ipv4 struct ip
|
|
|
|
*/
|
|
|
|
struct frag_queue {
|
|
|
|
struct inet_frag_queue q;
|
|
|
|
|
|
|
|
int iif;
|
|
|
|
__u16 nhoffset;
|
|
|
|
u8 ecn;
|
|
|
|
};
|
|
|
|
|
|
|
|
#if IS_ENABLED(CONFIG_IPV6)
|
|
|
|
static inline void ip6frag_init(struct inet_frag_queue *q, const void *a)
|
|
|
|
{
|
|
|
|
struct frag_queue *fq = container_of(q, struct frag_queue, q);
|
|
|
|
const struct frag_v6_compare_key *key = a;
|
|
|
|
|
|
|
|
q->key.v6 = *key;
|
|
|
|
fq->ecn = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 ip6frag_key_hashfn(const void *data, u32 len, u32 seed)
|
|
|
|
{
|
|
|
|
return jhash2(data,
|
|
|
|
sizeof(struct frag_v6_compare_key) / sizeof(u32), seed);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 ip6frag_obj_hashfn(const void *data, u32 len, u32 seed)
|
|
|
|
{
|
|
|
|
const struct inet_frag_queue *fq = data;
|
|
|
|
|
|
|
|
return jhash2((const u32 *)&fq->key.v6,
|
|
|
|
sizeof(struct frag_v6_compare_key) / sizeof(u32), seed);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int
|
|
|
|
ip6frag_obj_cmpfn(struct rhashtable_compare_arg *arg, const void *ptr)
|
|
|
|
{
|
|
|
|
const struct frag_v6_compare_key *key = arg->key;
|
|
|
|
const struct inet_frag_queue *fq = ptr;
|
|
|
|
|
|
|
|
return !!memcmp(&fq->key, key, sizeof(*key));
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
ip6frag_expire_frag_queue(struct net *net, struct frag_queue *fq)
|
|
|
|
{
|
|
|
|
struct net_device *dev = NULL;
|
|
|
|
struct sk_buff *head;
|
|
|
|
|
|
|
|
rcu_read_lock();
|
2019-06-19 01:09:00 +07:00
|
|
|
if (fq->q.fqdir->dead)
|
|
|
|
goto out_rcu_unlock;
|
2018-07-14 06:14:01 +07:00
|
|
|
spin_lock(&fq->q.lock);
|
|
|
|
|
|
|
|
if (fq->q.flags & INET_FRAG_COMPLETE)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
inet_frag_kill(&fq->q);
|
|
|
|
|
|
|
|
dev = dev_get_by_index_rcu(net, fq->iif);
|
|
|
|
if (!dev)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
__IP6_INC_STATS(net, __in6_dev_get(dev), IPSTATS_MIB_REASMFAILS);
|
|
|
|
__IP6_INC_STATS(net, __in6_dev_get(dev), IPSTATS_MIB_REASMTIMEOUT);
|
|
|
|
|
|
|
|
/* Don't send error if the first segment did not arrive. */
|
2019-01-23 01:02:51 +07:00
|
|
|
if (!(fq->q.flags & INET_FRAG_FIRST_IN))
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
/* sk_buff::dev and sk_buff::rbnode are unionized. So we
|
|
|
|
* pull the head out of the tree in order to be able to
|
|
|
|
* deal with head->dev.
|
|
|
|
*/
|
|
|
|
head = inet_frag_pull_head(&fq->q);
|
|
|
|
if (!head)
|
2018-07-14 06:14:01 +07:00
|
|
|
goto out;
|
|
|
|
|
|
|
|
head->dev = dev;
|
|
|
|
spin_unlock(&fq->q.lock);
|
|
|
|
|
|
|
|
icmpv6_send(head, ICMPV6_TIME_EXCEED, ICMPV6_EXC_FRAGTIME, 0);
|
|
|
|
kfree_skb(head);
|
|
|
|
goto out_rcu_unlock;
|
|
|
|
|
|
|
|
out:
|
|
|
|
spin_unlock(&fq->q.lock);
|
|
|
|
out_rcu_unlock:
|
|
|
|
rcu_read_unlock();
|
|
|
|
inet_frag_put(&fq->q);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
#endif
|