2005-04-17 05:20:36 +07:00
|
|
|
/*
|
|
|
|
* net/sched/cls_api.c Packet classifier API.
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU General Public License
|
|
|
|
* as published by the Free Software Foundation; either version
|
|
|
|
* 2 of the License, or (at your option) any later version.
|
|
|
|
*
|
|
|
|
* Authors: Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru>
|
|
|
|
*
|
|
|
|
* Changes:
|
|
|
|
*
|
|
|
|
* Eduardo J. Blanco <ejbs@netlabs.com.uy> :990222: kmod support
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/types.h>
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/string.h>
|
|
|
|
#include <linux/errno.h>
|
2017-02-09 20:38:57 +07:00
|
|
|
#include <linux/err.h>
|
2005-04-17 05:20:36 +07:00
|
|
|
#include <linux/skbuff.h>
|
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/kmod.h>
|
2008-01-24 11:33:13 +07:00
|
|
|
#include <linux/err.h>
|
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 15:04:11 +07:00
|
|
|
#include <linux/slab.h>
|
2007-11-30 20:21:31 +07:00
|
|
|
#include <net/net_namespace.h>
|
|
|
|
#include <net/sock.h>
|
2007-03-26 13:06:12 +07:00
|
|
|
#include <net/netlink.h>
|
2005-04-17 05:20:36 +07:00
|
|
|
#include <net/pkt_sched.h>
|
|
|
|
#include <net/pkt_cls.h>
|
|
|
|
|
|
|
|
/* The list of all installed classifier types */
|
2013-12-16 11:15:11 +07:00
|
|
|
static LIST_HEAD(tcf_proto_base);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
/* Protects list of registered TC modules. It is pure SMP lock. */
|
|
|
|
static DEFINE_RWLOCK(cls_mod_lock);
|
|
|
|
|
|
|
|
/* Find classifier type by string name */
|
|
|
|
|
2017-02-09 20:38:57 +07:00
|
|
|
static const struct tcf_proto_ops *tcf_proto_lookup_ops(const char *kind)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
2013-12-21 01:04:18 +07:00
|
|
|
const struct tcf_proto_ops *t, *res = NULL;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
if (kind) {
|
|
|
|
read_lock(&cls_mod_lock);
|
2013-12-16 11:15:11 +07:00
|
|
|
list_for_each_entry(t, &tcf_proto_base, head) {
|
2017-02-09 20:38:57 +07:00
|
|
|
if (strcmp(kind, t->kind) == 0) {
|
2013-12-21 01:04:18 +07:00
|
|
|
if (try_module_get(t->owner))
|
|
|
|
res = t;
|
2005-04-17 05:20:36 +07:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
read_unlock(&cls_mod_lock);
|
|
|
|
}
|
2013-12-21 01:04:18 +07:00
|
|
|
return res;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Register(unregister) new classifier type */
|
|
|
|
|
|
|
|
int register_tcf_proto_ops(struct tcf_proto_ops *ops)
|
|
|
|
{
|
2013-12-16 11:15:11 +07:00
|
|
|
struct tcf_proto_ops *t;
|
2005-04-17 05:20:36 +07:00
|
|
|
int rc = -EEXIST;
|
|
|
|
|
|
|
|
write_lock(&cls_mod_lock);
|
2013-12-16 11:15:11 +07:00
|
|
|
list_for_each_entry(t, &tcf_proto_base, head)
|
2005-04-17 05:20:36 +07:00
|
|
|
if (!strcmp(ops->kind, t->kind))
|
|
|
|
goto out;
|
|
|
|
|
2013-12-16 11:15:11 +07:00
|
|
|
list_add_tail(&ops->head, &tcf_proto_base);
|
2005-04-17 05:20:36 +07:00
|
|
|
rc = 0;
|
|
|
|
out:
|
|
|
|
write_unlock(&cls_mod_lock);
|
|
|
|
return rc;
|
|
|
|
}
|
2008-01-21 17:26:41 +07:00
|
|
|
EXPORT_SYMBOL(register_tcf_proto_ops);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2017-10-27 08:24:28 +07:00
|
|
|
static struct workqueue_struct *tc_filter_wq;
|
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
int unregister_tcf_proto_ops(struct tcf_proto_ops *ops)
|
|
|
|
{
|
2013-12-16 11:15:11 +07:00
|
|
|
struct tcf_proto_ops *t;
|
2005-04-17 05:20:36 +07:00
|
|
|
int rc = -ENOENT;
|
|
|
|
|
net: sched: fix call_rcu() race on classifier module unloads
Vijay reported that a loop as simple as ...
while true; do
tc qdisc add dev foo root handle 1: prio
tc filter add dev foo parent 1: u32 match u32 0 0 flowid 1
tc qdisc del dev foo root
rmmod cls_u32
done
... will panic the kernel. Moreover, he bisected the change
apparently introducing it to 78fd1d0ab072 ("netlink: Re-add
locking to netlink_lookup() and seq walker").
The removal of synchronize_net() from the netlink socket
triggering the qdisc to be removed, seems to have uncovered
an RCU resp. module reference count race from the tc API.
Given that RCU conversion was done after e341694e3eb5 ("netlink:
Convert netlink_lookup() to use RCU protected hash table")
which added the synchronize_net() originally, occasion of
hitting the bug was less likely (not impossible though):
When qdiscs that i) support attaching classifiers and,
ii) have at least one of them attached, get deleted, they
invoke tcf_destroy_chain(), and thus call into ->destroy()
handler from a classifier module.
After RCU conversion, all classifier that have an internal
prio list, unlink them and initiate freeing via call_rcu()
deferral.
Meanhile, tcf_destroy() releases already reference to the
tp->ops->owner module before the queued RCU callback handler
has been invoked.
Subsequent rmmod on the classifier module is then not prevented
since all module references are already dropped.
By the time, the kernel invokes the RCU callback handler from
the module, that function address is then invalid.
One way to fix it would be to add an rcu_barrier() to
unregister_tcf_proto_ops() to wait for all pending call_rcu()s
to complete.
synchronize_rcu() is not appropriate as under heavy RCU
callback load, registered call_rcu()s could be deferred
longer than a grace period. In case we don't have any pending
call_rcu()s, the barrier is allowed to return immediately.
Since we came here via unregister_tcf_proto_ops(), there
are no users of a given classifier anymore. Further nested
call_rcu()s pointing into the module space are not being
done anywhere.
Only cls_bpf_delete_prog() may schedule a work item, to
unlock pages eventually, but that is not in the range/context
of cls_bpf anymore.
Fixes: 25d8c0d55f24 ("net: rcu-ify tcf_proto")
Fixes: 9888faefe132 ("net: sched: cls_basic use RCU")
Reported-by: Vijay Subramanian <subramanian.vijay@gmail.com>
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
Cc: John Fastabend <john.r.fastabend@intel.com>
Cc: Eric Dumazet <edumazet@google.com>
Cc: Thomas Graf <tgraf@suug.ch>
Cc: Jamal Hadi Salim <jhs@mojatatu.com>
Cc: Alexei Starovoitov <ast@plumgrid.com>
Tested-by: Vijay Subramanian <subramanian.vijay@gmail.com>
Acked-by: Alexei Starovoitov <ast@plumgrid.com>
Acked-by: Eric Dumazet <edumazet@google.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-05-20 22:13:33 +07:00
|
|
|
/* Wait for outstanding call_rcu()s, if any, from a
|
|
|
|
* tcf_proto_ops's destroy() handler.
|
|
|
|
*/
|
|
|
|
rcu_barrier();
|
2017-10-27 08:24:28 +07:00
|
|
|
flush_workqueue(tc_filter_wq);
|
net: sched: fix call_rcu() race on classifier module unloads
Vijay reported that a loop as simple as ...
while true; do
tc qdisc add dev foo root handle 1: prio
tc filter add dev foo parent 1: u32 match u32 0 0 flowid 1
tc qdisc del dev foo root
rmmod cls_u32
done
... will panic the kernel. Moreover, he bisected the change
apparently introducing it to 78fd1d0ab072 ("netlink: Re-add
locking to netlink_lookup() and seq walker").
The removal of synchronize_net() from the netlink socket
triggering the qdisc to be removed, seems to have uncovered
an RCU resp. module reference count race from the tc API.
Given that RCU conversion was done after e341694e3eb5 ("netlink:
Convert netlink_lookup() to use RCU protected hash table")
which added the synchronize_net() originally, occasion of
hitting the bug was less likely (not impossible though):
When qdiscs that i) support attaching classifiers and,
ii) have at least one of them attached, get deleted, they
invoke tcf_destroy_chain(), and thus call into ->destroy()
handler from a classifier module.
After RCU conversion, all classifier that have an internal
prio list, unlink them and initiate freeing via call_rcu()
deferral.
Meanhile, tcf_destroy() releases already reference to the
tp->ops->owner module before the queued RCU callback handler
has been invoked.
Subsequent rmmod on the classifier module is then not prevented
since all module references are already dropped.
By the time, the kernel invokes the RCU callback handler from
the module, that function address is then invalid.
One way to fix it would be to add an rcu_barrier() to
unregister_tcf_proto_ops() to wait for all pending call_rcu()s
to complete.
synchronize_rcu() is not appropriate as under heavy RCU
callback load, registered call_rcu()s could be deferred
longer than a grace period. In case we don't have any pending
call_rcu()s, the barrier is allowed to return immediately.
Since we came here via unregister_tcf_proto_ops(), there
are no users of a given classifier anymore. Further nested
call_rcu()s pointing into the module space are not being
done anywhere.
Only cls_bpf_delete_prog() may schedule a work item, to
unlock pages eventually, but that is not in the range/context
of cls_bpf anymore.
Fixes: 25d8c0d55f24 ("net: rcu-ify tcf_proto")
Fixes: 9888faefe132 ("net: sched: cls_basic use RCU")
Reported-by: Vijay Subramanian <subramanian.vijay@gmail.com>
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
Cc: John Fastabend <john.r.fastabend@intel.com>
Cc: Eric Dumazet <edumazet@google.com>
Cc: Thomas Graf <tgraf@suug.ch>
Cc: Jamal Hadi Salim <jhs@mojatatu.com>
Cc: Alexei Starovoitov <ast@plumgrid.com>
Tested-by: Vijay Subramanian <subramanian.vijay@gmail.com>
Acked-by: Alexei Starovoitov <ast@plumgrid.com>
Acked-by: Eric Dumazet <edumazet@google.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-05-20 22:13:33 +07:00
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
write_lock(&cls_mod_lock);
|
2013-12-21 01:04:18 +07:00
|
|
|
list_for_each_entry(t, &tcf_proto_base, head) {
|
|
|
|
if (t == ops) {
|
|
|
|
list_del(&t->head);
|
|
|
|
rc = 0;
|
2005-04-17 05:20:36 +07:00
|
|
|
break;
|
2013-12-21 01:04:18 +07:00
|
|
|
}
|
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
write_unlock(&cls_mod_lock);
|
|
|
|
return rc;
|
|
|
|
}
|
2008-01-21 17:26:41 +07:00
|
|
|
EXPORT_SYMBOL(unregister_tcf_proto_ops);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2017-10-27 08:24:28 +07:00
|
|
|
bool tcf_queue_work(struct work_struct *work)
|
|
|
|
{
|
|
|
|
return queue_work(tc_filter_wq, work);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(tcf_queue_work);
|
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
/* Select new prio value from the range, managed by kernel. */
|
|
|
|
|
2008-01-21 17:26:41 +07:00
|
|
|
static inline u32 tcf_auto_prio(struct tcf_proto *tp)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
2008-01-21 17:26:41 +07:00
|
|
|
u32 first = TC_H_MAKE(0xC0000000U, 0U);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
if (tp)
|
2011-01-20 02:26:56 +07:00
|
|
|
first = tp->prio - 1;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2017-05-17 16:07:58 +07:00
|
|
|
return TC_H_MAJ(first);
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
2017-02-09 20:38:57 +07:00
|
|
|
static struct tcf_proto *tcf_proto_create(const char *kind, u32 protocol,
|
2017-05-17 16:07:55 +07:00
|
|
|
u32 prio, u32 parent, struct Qdisc *q,
|
2017-05-17 16:08:01 +07:00
|
|
|
struct tcf_chain *chain)
|
2017-02-09 20:38:57 +07:00
|
|
|
{
|
|
|
|
struct tcf_proto *tp;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
tp = kzalloc(sizeof(*tp), GFP_KERNEL);
|
|
|
|
if (!tp)
|
|
|
|
return ERR_PTR(-ENOBUFS);
|
|
|
|
|
|
|
|
err = -ENOENT;
|
|
|
|
tp->ops = tcf_proto_lookup_ops(kind);
|
|
|
|
if (!tp->ops) {
|
|
|
|
#ifdef CONFIG_MODULES
|
|
|
|
rtnl_unlock();
|
|
|
|
request_module("cls_%s", kind);
|
|
|
|
rtnl_lock();
|
|
|
|
tp->ops = tcf_proto_lookup_ops(kind);
|
|
|
|
/* We dropped the RTNL semaphore in order to perform
|
|
|
|
* the module load. So, even if we succeeded in loading
|
|
|
|
* the module we have to replay the request. We indicate
|
|
|
|
* this using -EAGAIN.
|
|
|
|
*/
|
|
|
|
if (tp->ops) {
|
|
|
|
module_put(tp->ops->owner);
|
|
|
|
err = -EAGAIN;
|
|
|
|
} else {
|
|
|
|
err = -ENOENT;
|
|
|
|
}
|
|
|
|
goto errout;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
tp->classify = tp->ops->classify;
|
|
|
|
tp->protocol = protocol;
|
|
|
|
tp->prio = prio;
|
|
|
|
tp->classid = parent;
|
|
|
|
tp->q = q;
|
2017-05-17 16:08:01 +07:00
|
|
|
tp->chain = chain;
|
2017-02-09 20:38:57 +07:00
|
|
|
|
|
|
|
err = tp->ops->init(tp);
|
|
|
|
if (err) {
|
|
|
|
module_put(tp->ops->owner);
|
|
|
|
goto errout;
|
|
|
|
}
|
|
|
|
return tp;
|
|
|
|
|
|
|
|
errout:
|
|
|
|
kfree(tp);
|
|
|
|
return ERR_PTR(err);
|
|
|
|
}
|
|
|
|
|
2017-04-20 04:21:21 +07:00
|
|
|
static void tcf_proto_destroy(struct tcf_proto *tp)
|
2017-02-09 20:38:56 +07:00
|
|
|
{
|
2017-04-20 04:21:21 +07:00
|
|
|
tp->ops->destroy(tp);
|
|
|
|
module_put(tp->ops->owner);
|
|
|
|
kfree_rcu(tp, rcu);
|
2017-02-09 20:38:56 +07:00
|
|
|
}
|
|
|
|
|
2017-05-17 16:08:01 +07:00
|
|
|
static struct tcf_chain *tcf_chain_create(struct tcf_block *block,
|
|
|
|
u32 chain_index)
|
2017-05-17 16:07:59 +07:00
|
|
|
{
|
2017-05-17 16:08:01 +07:00
|
|
|
struct tcf_chain *chain;
|
|
|
|
|
|
|
|
chain = kzalloc(sizeof(*chain), GFP_KERNEL);
|
|
|
|
if (!chain)
|
|
|
|
return NULL;
|
|
|
|
list_add_tail(&chain->list, &block->chain_list);
|
|
|
|
chain->block = block;
|
|
|
|
chain->index = chain_index;
|
2017-09-12 06:33:31 +07:00
|
|
|
chain->refcnt = 1;
|
2017-05-17 16:08:01 +07:00
|
|
|
return chain;
|
2017-05-17 16:07:59 +07:00
|
|
|
}
|
|
|
|
|
2017-11-03 17:46:24 +07:00
|
|
|
static void tcf_chain_head_change(struct tcf_chain *chain,
|
|
|
|
struct tcf_proto *tp_head)
|
|
|
|
{
|
|
|
|
if (chain->chain_head_change)
|
|
|
|
chain->chain_head_change(tp_head,
|
|
|
|
chain->chain_head_change_priv);
|
|
|
|
}
|
|
|
|
|
2017-05-20 20:01:32 +07:00
|
|
|
static void tcf_chain_flush(struct tcf_chain *chain)
|
2017-02-09 20:38:56 +07:00
|
|
|
{
|
|
|
|
struct tcf_proto *tp;
|
|
|
|
|
2017-11-03 17:46:24 +07:00
|
|
|
tcf_chain_head_change(chain, NULL);
|
2017-05-17 16:07:59 +07:00
|
|
|
while ((tp = rtnl_dereference(chain->filter_chain)) != NULL) {
|
|
|
|
RCU_INIT_POINTER(chain->filter_chain, tp->next);
|
2017-09-12 06:33:31 +07:00
|
|
|
tcf_chain_put(chain);
|
2017-04-20 04:21:21 +07:00
|
|
|
tcf_proto_destroy(tp);
|
2017-02-09 20:38:56 +07:00
|
|
|
}
|
2017-05-20 20:01:32 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static void tcf_chain_destroy(struct tcf_chain *chain)
|
|
|
|
{
|
2017-09-12 06:33:31 +07:00
|
|
|
list_del(&chain->list);
|
|
|
|
kfree(chain);
|
|
|
|
}
|
2017-08-23 03:46:49 +07:00
|
|
|
|
2017-09-12 06:33:31 +07:00
|
|
|
static void tcf_chain_hold(struct tcf_chain *chain)
|
|
|
|
{
|
|
|
|
++chain->refcnt;
|
2017-05-17 16:07:59 +07:00
|
|
|
}
|
|
|
|
|
2017-05-23 23:42:37 +07:00
|
|
|
struct tcf_chain *tcf_chain_get(struct tcf_block *block, u32 chain_index,
|
|
|
|
bool create)
|
2017-05-17 16:08:01 +07:00
|
|
|
{
|
|
|
|
struct tcf_chain *chain;
|
|
|
|
|
|
|
|
list_for_each_entry(chain, &block->chain_list, list) {
|
2017-09-12 06:33:31 +07:00
|
|
|
if (chain->index == chain_index) {
|
|
|
|
tcf_chain_hold(chain);
|
|
|
|
return chain;
|
|
|
|
}
|
2017-05-17 16:08:01 +07:00
|
|
|
}
|
2017-09-06 18:14:19 +07:00
|
|
|
|
2017-09-12 06:33:31 +07:00
|
|
|
return create ? tcf_chain_create(block, chain_index) : NULL;
|
2017-05-17 16:08:01 +07:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(tcf_chain_get);
|
|
|
|
|
|
|
|
void tcf_chain_put(struct tcf_chain *chain)
|
|
|
|
{
|
2017-09-12 06:33:31 +07:00
|
|
|
if (--chain->refcnt == 0)
|
2017-05-17 16:08:01 +07:00
|
|
|
tcf_chain_destroy(chain);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(tcf_chain_put);
|
|
|
|
|
2017-10-19 20:50:29 +07:00
|
|
|
static void tcf_block_offload_cmd(struct tcf_block *block, struct Qdisc *q,
|
|
|
|
struct tcf_block_ext_info *ei,
|
|
|
|
enum tc_block_command command)
|
|
|
|
{
|
|
|
|
struct net_device *dev = q->dev_queue->dev;
|
|
|
|
struct tc_block_offload bo = {};
|
|
|
|
|
2017-11-01 17:47:39 +07:00
|
|
|
if (!dev->netdev_ops->ndo_setup_tc)
|
2017-10-19 20:50:29 +07:00
|
|
|
return;
|
|
|
|
bo.command = command;
|
|
|
|
bo.binder_type = ei->binder_type;
|
|
|
|
bo.block = block;
|
|
|
|
dev->netdev_ops->ndo_setup_tc(dev, TC_SETUP_BLOCK, &bo);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void tcf_block_offload_bind(struct tcf_block *block, struct Qdisc *q,
|
|
|
|
struct tcf_block_ext_info *ei)
|
|
|
|
{
|
|
|
|
tcf_block_offload_cmd(block, q, ei, TC_BLOCK_BIND);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void tcf_block_offload_unbind(struct tcf_block *block, struct Qdisc *q,
|
|
|
|
struct tcf_block_ext_info *ei)
|
|
|
|
{
|
|
|
|
tcf_block_offload_cmd(block, q, ei, TC_BLOCK_UNBIND);
|
|
|
|
}
|
|
|
|
|
2017-11-03 17:46:24 +07:00
|
|
|
int tcf_block_get_ext(struct tcf_block **p_block, struct Qdisc *q,
|
2017-10-19 20:50:29 +07:00
|
|
|
struct tcf_block_ext_info *ei)
|
2017-05-17 16:07:55 +07:00
|
|
|
{
|
|
|
|
struct tcf_block *block = kzalloc(sizeof(*block), GFP_KERNEL);
|
2017-05-17 16:08:01 +07:00
|
|
|
struct tcf_chain *chain;
|
2017-05-17 16:07:59 +07:00
|
|
|
int err;
|
2017-05-17 16:07:55 +07:00
|
|
|
|
|
|
|
if (!block)
|
|
|
|
return -ENOMEM;
|
2017-05-17 16:08:01 +07:00
|
|
|
INIT_LIST_HEAD(&block->chain_list);
|
2017-10-19 20:50:31 +07:00
|
|
|
INIT_LIST_HEAD(&block->cb_list);
|
|
|
|
|
2017-05-17 16:08:01 +07:00
|
|
|
/* Create chain 0 by default, it has to be always present. */
|
|
|
|
chain = tcf_chain_create(block, 0);
|
|
|
|
if (!chain) {
|
2017-05-17 16:07:59 +07:00
|
|
|
err = -ENOMEM;
|
|
|
|
goto err_chain_create;
|
|
|
|
}
|
2017-11-03 17:46:24 +07:00
|
|
|
WARN_ON(!ei->chain_head_change);
|
|
|
|
chain->chain_head_change = ei->chain_head_change;
|
|
|
|
chain->chain_head_change_priv = ei->chain_head_change_priv;
|
2017-10-13 19:00:58 +07:00
|
|
|
block->net = qdisc_net(q);
|
2017-10-13 19:00:57 +07:00
|
|
|
block->q = q;
|
2017-10-19 20:50:29 +07:00
|
|
|
tcf_block_offload_bind(block, q, ei);
|
2017-05-17 16:07:55 +07:00
|
|
|
*p_block = block;
|
|
|
|
return 0;
|
2017-05-17 16:07:59 +07:00
|
|
|
|
|
|
|
err_chain_create:
|
|
|
|
kfree(block);
|
|
|
|
return err;
|
2017-05-17 16:07:55 +07:00
|
|
|
}
|
2017-10-19 20:50:29 +07:00
|
|
|
EXPORT_SYMBOL(tcf_block_get_ext);
|
|
|
|
|
2017-11-03 17:46:24 +07:00
|
|
|
static void tcf_chain_head_change_dflt(struct tcf_proto *tp_head, void *priv)
|
|
|
|
{
|
|
|
|
struct tcf_proto __rcu **p_filter_chain = priv;
|
|
|
|
|
|
|
|
rcu_assign_pointer(*p_filter_chain, tp_head);
|
|
|
|
}
|
|
|
|
|
2017-10-19 20:50:29 +07:00
|
|
|
int tcf_block_get(struct tcf_block **p_block,
|
|
|
|
struct tcf_proto __rcu **p_filter_chain, struct Qdisc *q)
|
|
|
|
{
|
2017-11-03 17:46:24 +07:00
|
|
|
struct tcf_block_ext_info ei = {
|
|
|
|
.chain_head_change = tcf_chain_head_change_dflt,
|
|
|
|
.chain_head_change_priv = p_filter_chain,
|
|
|
|
};
|
2017-10-19 20:50:29 +07:00
|
|
|
|
2017-11-03 17:46:24 +07:00
|
|
|
WARN_ON(!p_filter_chain);
|
|
|
|
return tcf_block_get_ext(p_block, q, &ei);
|
2017-10-19 20:50:29 +07:00
|
|
|
}
|
2017-05-17 16:07:55 +07:00
|
|
|
EXPORT_SYMBOL(tcf_block_get);
|
|
|
|
|
2017-10-27 08:24:28 +07:00
|
|
|
static void tcf_block_put_final(struct work_struct *work)
|
2017-05-17 16:07:55 +07:00
|
|
|
{
|
2017-10-27 08:24:28 +07:00
|
|
|
struct tcf_block *block = container_of(work, struct tcf_block, work);
|
2017-05-17 16:08:01 +07:00
|
|
|
struct tcf_chain *chain, *tmp;
|
|
|
|
|
2017-10-27 08:24:28 +07:00
|
|
|
rtnl_lock();
|
2017-10-31 01:10:09 +07:00
|
|
|
/* Only chain 0 should be still here. */
|
2017-10-27 08:24:28 +07:00
|
|
|
list_for_each_entry_safe(chain, tmp, &block->chain_list, list)
|
|
|
|
tcf_chain_put(chain);
|
|
|
|
rtnl_unlock();
|
|
|
|
kfree(block);
|
|
|
|
}
|
2017-09-12 06:33:32 +07:00
|
|
|
|
2017-10-27 08:24:28 +07:00
|
|
|
/* XXX: Standalone actions are not allowed to jump to any chain, and bound
|
2017-10-31 01:10:09 +07:00
|
|
|
* actions should be all removed after flushing. However, filters are now
|
|
|
|
* destroyed in tc filter workqueue with RTNL lock, they can not race here.
|
2017-10-27 08:24:28 +07:00
|
|
|
*/
|
2017-11-03 17:46:24 +07:00
|
|
|
void tcf_block_put_ext(struct tcf_block *block, struct Qdisc *q,
|
2017-10-30 12:10:01 +07:00
|
|
|
struct tcf_block_ext_info *ei)
|
2017-10-27 08:24:28 +07:00
|
|
|
{
|
2017-10-31 01:10:09 +07:00
|
|
|
struct tcf_chain *chain, *tmp;
|
2017-09-12 06:33:32 +07:00
|
|
|
|
2017-10-31 01:10:09 +07:00
|
|
|
list_for_each_entry_safe(chain, tmp, &block->chain_list, list)
|
2017-08-23 03:46:50 +07:00
|
|
|
tcf_chain_flush(chain);
|
2017-09-12 06:33:31 +07:00
|
|
|
|
2017-11-02 21:07:01 +07:00
|
|
|
tcf_block_offload_unbind(block, q, ei);
|
|
|
|
|
2017-10-27 08:24:28 +07:00
|
|
|
INIT_WORK(&block->work, tcf_block_put_final);
|
|
|
|
/* Wait for existing RCU callbacks to cool down, make sure their works
|
|
|
|
* have been queued before this. We can not flush pending works here
|
|
|
|
* because we are holding the RTNL lock.
|
|
|
|
*/
|
|
|
|
rcu_barrier();
|
|
|
|
tcf_queue_work(&block->work);
|
2017-05-17 16:07:55 +07:00
|
|
|
}
|
2017-10-19 20:50:29 +07:00
|
|
|
EXPORT_SYMBOL(tcf_block_put_ext);
|
|
|
|
|
|
|
|
void tcf_block_put(struct tcf_block *block)
|
|
|
|
{
|
|
|
|
struct tcf_block_ext_info ei = {0, };
|
|
|
|
|
2017-11-03 07:32:08 +07:00
|
|
|
if (!block)
|
|
|
|
return;
|
2017-11-03 17:46:24 +07:00
|
|
|
tcf_block_put_ext(block, block->q, &ei);
|
2017-10-19 20:50:29 +07:00
|
|
|
}
|
2017-10-30 12:10:01 +07:00
|
|
|
|
2017-05-17 16:07:55 +07:00
|
|
|
EXPORT_SYMBOL(tcf_block_put);
|
2017-02-09 20:38:56 +07:00
|
|
|
|
2017-10-19 20:50:31 +07:00
|
|
|
struct tcf_block_cb {
|
|
|
|
struct list_head list;
|
|
|
|
tc_setup_cb_t *cb;
|
|
|
|
void *cb_ident;
|
|
|
|
void *cb_priv;
|
|
|
|
unsigned int refcnt;
|
|
|
|
};
|
|
|
|
|
|
|
|
void *tcf_block_cb_priv(struct tcf_block_cb *block_cb)
|
|
|
|
{
|
|
|
|
return block_cb->cb_priv;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(tcf_block_cb_priv);
|
|
|
|
|
|
|
|
struct tcf_block_cb *tcf_block_cb_lookup(struct tcf_block *block,
|
|
|
|
tc_setup_cb_t *cb, void *cb_ident)
|
|
|
|
{ struct tcf_block_cb *block_cb;
|
|
|
|
|
|
|
|
list_for_each_entry(block_cb, &block->cb_list, list)
|
|
|
|
if (block_cb->cb == cb && block_cb->cb_ident == cb_ident)
|
|
|
|
return block_cb;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(tcf_block_cb_lookup);
|
|
|
|
|
|
|
|
void tcf_block_cb_incref(struct tcf_block_cb *block_cb)
|
|
|
|
{
|
|
|
|
block_cb->refcnt++;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(tcf_block_cb_incref);
|
|
|
|
|
|
|
|
unsigned int tcf_block_cb_decref(struct tcf_block_cb *block_cb)
|
|
|
|
{
|
|
|
|
return --block_cb->refcnt;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(tcf_block_cb_decref);
|
|
|
|
|
|
|
|
struct tcf_block_cb *__tcf_block_cb_register(struct tcf_block *block,
|
|
|
|
tc_setup_cb_t *cb, void *cb_ident,
|
|
|
|
void *cb_priv)
|
|
|
|
{
|
|
|
|
struct tcf_block_cb *block_cb;
|
|
|
|
|
|
|
|
block_cb = kzalloc(sizeof(*block_cb), GFP_KERNEL);
|
|
|
|
if (!block_cb)
|
|
|
|
return NULL;
|
|
|
|
block_cb->cb = cb;
|
|
|
|
block_cb->cb_ident = cb_ident;
|
|
|
|
block_cb->cb_priv = cb_priv;
|
|
|
|
list_add(&block_cb->list, &block->cb_list);
|
|
|
|
return block_cb;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(__tcf_block_cb_register);
|
|
|
|
|
|
|
|
int tcf_block_cb_register(struct tcf_block *block,
|
|
|
|
tc_setup_cb_t *cb, void *cb_ident,
|
|
|
|
void *cb_priv)
|
|
|
|
{
|
|
|
|
struct tcf_block_cb *block_cb;
|
|
|
|
|
|
|
|
block_cb = __tcf_block_cb_register(block, cb, cb_ident, cb_priv);
|
|
|
|
return block_cb ? 0 : -ENOMEM;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(tcf_block_cb_register);
|
|
|
|
|
|
|
|
void __tcf_block_cb_unregister(struct tcf_block_cb *block_cb)
|
|
|
|
{
|
|
|
|
list_del(&block_cb->list);
|
|
|
|
kfree(block_cb);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(__tcf_block_cb_unregister);
|
|
|
|
|
|
|
|
void tcf_block_cb_unregister(struct tcf_block *block,
|
|
|
|
tc_setup_cb_t *cb, void *cb_ident)
|
|
|
|
{
|
|
|
|
struct tcf_block_cb *block_cb;
|
|
|
|
|
|
|
|
block_cb = tcf_block_cb_lookup(block, cb, cb_ident);
|
|
|
|
if (!block_cb)
|
|
|
|
return;
|
|
|
|
__tcf_block_cb_unregister(block_cb);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(tcf_block_cb_unregister);
|
|
|
|
|
|
|
|
static int tcf_block_cb_call(struct tcf_block *block, enum tc_setup_type type,
|
|
|
|
void *type_data, bool err_stop)
|
|
|
|
{
|
|
|
|
struct tcf_block_cb *block_cb;
|
|
|
|
int ok_count = 0;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
list_for_each_entry(block_cb, &block->cb_list, list) {
|
|
|
|
err = block_cb->cb(type, type_data, block_cb->cb_priv);
|
|
|
|
if (err) {
|
|
|
|
if (err_stop)
|
|
|
|
return err;
|
|
|
|
} else {
|
|
|
|
ok_count++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return ok_count;
|
|
|
|
}
|
|
|
|
|
2017-05-17 16:07:54 +07:00
|
|
|
/* Main classifier routine: scans classifier chain attached
|
|
|
|
* to this qdisc, (optionally) tests for protocol and asks
|
|
|
|
* specific classifiers.
|
|
|
|
*/
|
|
|
|
int tcf_classify(struct sk_buff *skb, const struct tcf_proto *tp,
|
|
|
|
struct tcf_result *res, bool compat_mode)
|
|
|
|
{
|
|
|
|
__be16 protocol = tc_skb_protocol(skb);
|
|
|
|
#ifdef CONFIG_NET_CLS_ACT
|
|
|
|
const int max_reclassify_loop = 4;
|
2017-05-23 14:11:59 +07:00
|
|
|
const struct tcf_proto *orig_tp = tp;
|
|
|
|
const struct tcf_proto *first_tp;
|
2017-05-17 16:07:54 +07:00
|
|
|
int limit = 0;
|
|
|
|
|
|
|
|
reclassify:
|
|
|
|
#endif
|
|
|
|
for (; tp; tp = rcu_dereference_bh(tp->next)) {
|
|
|
|
int err;
|
|
|
|
|
|
|
|
if (tp->protocol != protocol &&
|
|
|
|
tp->protocol != htons(ETH_P_ALL))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
err = tp->classify(skb, tp, res);
|
|
|
|
#ifdef CONFIG_NET_CLS_ACT
|
2017-05-17 16:08:03 +07:00
|
|
|
if (unlikely(err == TC_ACT_RECLASSIFY && !compat_mode)) {
|
2017-05-23 14:11:59 +07:00
|
|
|
first_tp = orig_tp;
|
2017-05-17 16:07:54 +07:00
|
|
|
goto reset;
|
2017-05-17 16:08:03 +07:00
|
|
|
} else if (unlikely(TC_ACT_EXT_CMP(err, TC_ACT_GOTO_CHAIN))) {
|
2017-05-23 14:11:59 +07:00
|
|
|
first_tp = res->goto_tp;
|
2017-05-17 16:08:03 +07:00
|
|
|
goto reset;
|
|
|
|
}
|
2017-05-17 16:07:54 +07:00
|
|
|
#endif
|
|
|
|
if (err >= 0)
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
return TC_ACT_UNSPEC; /* signal: continue lookup */
|
|
|
|
#ifdef CONFIG_NET_CLS_ACT
|
|
|
|
reset:
|
|
|
|
if (unlikely(limit++ >= max_reclassify_loop)) {
|
|
|
|
net_notice_ratelimited("%s: reclassify loop, rule prio %u, protocol %02x\n",
|
|
|
|
tp->q->ops->id, tp->prio & 0xffff,
|
|
|
|
ntohs(tp->protocol));
|
|
|
|
return TC_ACT_SHOT;
|
|
|
|
}
|
|
|
|
|
2017-05-23 14:11:59 +07:00
|
|
|
tp = first_tp;
|
2017-05-17 16:07:54 +07:00
|
|
|
protocol = tc_skb_protocol(skb);
|
|
|
|
goto reclassify;
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(tcf_classify);
|
|
|
|
|
2017-05-17 16:07:59 +07:00
|
|
|
struct tcf_chain_info {
|
|
|
|
struct tcf_proto __rcu **pprev;
|
|
|
|
struct tcf_proto __rcu *next;
|
|
|
|
};
|
|
|
|
|
|
|
|
static struct tcf_proto *tcf_chain_tp_prev(struct tcf_chain_info *chain_info)
|
|
|
|
{
|
|
|
|
return rtnl_dereference(*chain_info->pprev);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void tcf_chain_tp_insert(struct tcf_chain *chain,
|
|
|
|
struct tcf_chain_info *chain_info,
|
|
|
|
struct tcf_proto *tp)
|
|
|
|
{
|
2017-11-03 17:46:24 +07:00
|
|
|
if (*chain_info->pprev == chain->filter_chain)
|
|
|
|
tcf_chain_head_change(chain, tp);
|
2017-05-17 16:07:59 +07:00
|
|
|
RCU_INIT_POINTER(tp->next, tcf_chain_tp_prev(chain_info));
|
|
|
|
rcu_assign_pointer(*chain_info->pprev, tp);
|
2017-09-12 06:33:31 +07:00
|
|
|
tcf_chain_hold(chain);
|
2017-05-17 16:07:59 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static void tcf_chain_tp_remove(struct tcf_chain *chain,
|
|
|
|
struct tcf_chain_info *chain_info,
|
|
|
|
struct tcf_proto *tp)
|
|
|
|
{
|
|
|
|
struct tcf_proto *next = rtnl_dereference(chain_info->next);
|
|
|
|
|
2017-11-03 17:46:24 +07:00
|
|
|
if (tp == chain->filter_chain)
|
|
|
|
tcf_chain_head_change(chain, next);
|
2017-05-17 16:07:59 +07:00
|
|
|
RCU_INIT_POINTER(*chain_info->pprev, next);
|
2017-09-12 06:33:31 +07:00
|
|
|
tcf_chain_put(chain);
|
2017-05-17 16:07:59 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static struct tcf_proto *tcf_chain_tp_find(struct tcf_chain *chain,
|
|
|
|
struct tcf_chain_info *chain_info,
|
|
|
|
u32 protocol, u32 prio,
|
|
|
|
bool prio_allocate)
|
|
|
|
{
|
|
|
|
struct tcf_proto **pprev;
|
|
|
|
struct tcf_proto *tp;
|
|
|
|
|
|
|
|
/* Check the chain for existence of proto-tcf with this priority */
|
|
|
|
for (pprev = &chain->filter_chain;
|
|
|
|
(tp = rtnl_dereference(*pprev)); pprev = &tp->next) {
|
|
|
|
if (tp->prio >= prio) {
|
|
|
|
if (tp->prio == prio) {
|
|
|
|
if (prio_allocate ||
|
|
|
|
(tp->protocol != protocol && protocol))
|
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
} else {
|
|
|
|
tp = NULL;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
chain_info->pprev = pprev;
|
|
|
|
chain_info->next = tp ? tp->next : NULL;
|
|
|
|
return tp;
|
|
|
|
}
|
|
|
|
|
2017-08-08 05:26:50 +07:00
|
|
|
static int tcf_fill_node(struct net *net, struct sk_buff *skb,
|
2017-10-13 19:01:05 +07:00
|
|
|
struct tcf_proto *tp, struct Qdisc *q, u32 parent,
|
|
|
|
void *fh, u32 portid, u32 seq, u16 flags, int event)
|
2017-08-08 05:26:50 +07:00
|
|
|
{
|
|
|
|
struct tcmsg *tcm;
|
|
|
|
struct nlmsghdr *nlh;
|
|
|
|
unsigned char *b = skb_tail_pointer(skb);
|
|
|
|
|
|
|
|
nlh = nlmsg_put(skb, portid, seq, event, sizeof(*tcm), flags);
|
|
|
|
if (!nlh)
|
|
|
|
goto out_nlmsg_trim;
|
|
|
|
tcm = nlmsg_data(nlh);
|
|
|
|
tcm->tcm_family = AF_UNSPEC;
|
|
|
|
tcm->tcm__pad1 = 0;
|
|
|
|
tcm->tcm__pad2 = 0;
|
2017-10-13 19:01:05 +07:00
|
|
|
tcm->tcm_ifindex = qdisc_dev(q)->ifindex;
|
|
|
|
tcm->tcm_parent = parent;
|
2017-08-08 05:26:50 +07:00
|
|
|
tcm->tcm_info = TC_H_MAKE(tp->prio, tp->protocol);
|
|
|
|
if (nla_put_string(skb, TCA_KIND, tp->ops->kind))
|
|
|
|
goto nla_put_failure;
|
|
|
|
if (nla_put_u32(skb, TCA_CHAIN, tp->chain->index))
|
|
|
|
goto nla_put_failure;
|
|
|
|
if (!fh) {
|
|
|
|
tcm->tcm_handle = 0;
|
|
|
|
} else {
|
|
|
|
if (tp->ops->dump && tp->ops->dump(net, tp, fh, skb, tcm) < 0)
|
|
|
|
goto nla_put_failure;
|
|
|
|
}
|
|
|
|
nlh->nlmsg_len = skb_tail_pointer(skb) - b;
|
|
|
|
return skb->len;
|
|
|
|
|
|
|
|
out_nlmsg_trim:
|
|
|
|
nla_put_failure:
|
|
|
|
nlmsg_trim(skb, b);
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int tfilter_notify(struct net *net, struct sk_buff *oskb,
|
|
|
|
struct nlmsghdr *n, struct tcf_proto *tp,
|
2017-10-13 19:01:05 +07:00
|
|
|
struct Qdisc *q, u32 parent,
|
2017-08-08 05:26:50 +07:00
|
|
|
void *fh, int event, bool unicast)
|
|
|
|
{
|
|
|
|
struct sk_buff *skb;
|
|
|
|
u32 portid = oskb ? NETLINK_CB(oskb).portid : 0;
|
|
|
|
|
|
|
|
skb = alloc_skb(NLMSG_GOODSIZE, GFP_KERNEL);
|
|
|
|
if (!skb)
|
|
|
|
return -ENOBUFS;
|
|
|
|
|
2017-10-13 19:01:05 +07:00
|
|
|
if (tcf_fill_node(net, skb, tp, q, parent, fh, portid, n->nlmsg_seq,
|
2017-08-08 05:26:50 +07:00
|
|
|
n->nlmsg_flags, event) <= 0) {
|
|
|
|
kfree_skb(skb);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (unicast)
|
|
|
|
return netlink_unicast(net->rtnl, skb, portid, MSG_DONTWAIT);
|
|
|
|
|
|
|
|
return rtnetlink_send(skb, net, portid, RTNLGRP_TC,
|
|
|
|
n->nlmsg_flags & NLM_F_ECHO);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int tfilter_del_notify(struct net *net, struct sk_buff *oskb,
|
|
|
|
struct nlmsghdr *n, struct tcf_proto *tp,
|
2017-10-13 19:01:05 +07:00
|
|
|
struct Qdisc *q, u32 parent,
|
2017-08-08 05:26:50 +07:00
|
|
|
void *fh, bool unicast, bool *last)
|
|
|
|
{
|
|
|
|
struct sk_buff *skb;
|
|
|
|
u32 portid = oskb ? NETLINK_CB(oskb).portid : 0;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
skb = alloc_skb(NLMSG_GOODSIZE, GFP_KERNEL);
|
|
|
|
if (!skb)
|
|
|
|
return -ENOBUFS;
|
|
|
|
|
2017-10-13 19:01:05 +07:00
|
|
|
if (tcf_fill_node(net, skb, tp, q, parent, fh, portid, n->nlmsg_seq,
|
2017-08-08 05:26:50 +07:00
|
|
|
n->nlmsg_flags, RTM_DELTFILTER) <= 0) {
|
|
|
|
kfree_skb(skb);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
err = tp->ops->delete(tp, fh, last);
|
|
|
|
if (err) {
|
|
|
|
kfree_skb(skb);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (unicast)
|
|
|
|
return netlink_unicast(net->rtnl, skb, portid, MSG_DONTWAIT);
|
|
|
|
|
|
|
|
return rtnetlink_send(skb, net, portid, RTNLGRP_TC,
|
|
|
|
n->nlmsg_flags & NLM_F_ECHO);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void tfilter_notify_chain(struct net *net, struct sk_buff *oskb,
|
2017-10-13 19:01:05 +07:00
|
|
|
struct Qdisc *q, u32 parent,
|
2017-08-08 05:26:50 +07:00
|
|
|
struct nlmsghdr *n,
|
|
|
|
struct tcf_chain *chain, int event)
|
|
|
|
{
|
|
|
|
struct tcf_proto *tp;
|
|
|
|
|
|
|
|
for (tp = rtnl_dereference(chain->filter_chain);
|
|
|
|
tp; tp = rtnl_dereference(tp->next))
|
2017-10-13 19:01:05 +07:00
|
|
|
tfilter_notify(net, oskb, n, tp, q, parent, 0, event, false);
|
2017-08-08 05:26:50 +07:00
|
|
|
}
|
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
/* Add/change/delete/get a filter node */
|
|
|
|
|
2017-04-16 23:48:24 +07:00
|
|
|
static int tc_ctl_tfilter(struct sk_buff *skb, struct nlmsghdr *n,
|
|
|
|
struct netlink_ext_ack *extack)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
2008-03-26 00:26:21 +07:00
|
|
|
struct net *net = sock_net(skb->sk);
|
2008-01-23 13:11:33 +07:00
|
|
|
struct nlattr *tca[TCA_MAX + 1];
|
2005-04-17 05:20:36 +07:00
|
|
|
struct tcmsg *t;
|
|
|
|
u32 protocol;
|
|
|
|
u32 prio;
|
2017-05-17 16:07:57 +07:00
|
|
|
bool prio_allocate;
|
2005-04-17 05:20:36 +07:00
|
|
|
u32 parent;
|
2017-05-17 16:08:01 +07:00
|
|
|
u32 chain_index;
|
2005-04-17 05:20:36 +07:00
|
|
|
struct net_device *dev;
|
|
|
|
struct Qdisc *q;
|
2017-05-17 16:07:59 +07:00
|
|
|
struct tcf_chain_info chain_info;
|
2017-05-17 16:08:01 +07:00
|
|
|
struct tcf_chain *chain = NULL;
|
2017-05-17 16:07:55 +07:00
|
|
|
struct tcf_block *block;
|
2005-04-17 05:20:36 +07:00
|
|
|
struct tcf_proto *tp;
|
2007-11-14 16:44:41 +07:00
|
|
|
const struct Qdisc_class_ops *cops;
|
2005-04-17 05:20:36 +07:00
|
|
|
unsigned long cl;
|
2017-08-05 11:31:43 +07:00
|
|
|
void *fh;
|
2005-04-17 05:20:36 +07:00
|
|
|
int err;
|
net, sched: fix soft lockup in tc_classify
Shahar reported a soft lockup in tc_classify(), where we run into an
endless loop when walking the classifier chain due to tp->next == tp
which is a state we should never run into. The issue only seems to
trigger under load in the tc control path.
What happens is that in tc_ctl_tfilter(), thread A allocates a new
tp, initializes it, sets tp_created to 1, and calls into tp->ops->change()
with it. In that classifier callback we had to unlock/lock the rtnl
mutex and returned with -EAGAIN. One reason why we need to drop there
is, for example, that we need to request an action module to be loaded.
This happens via tcf_exts_validate() -> tcf_action_init/_1() meaning
after we loaded and found the requested action, we need to redo the
whole request so we don't race against others. While we had to unlock
rtnl in that time, thread B's request was processed next on that CPU.
Thread B added a new tp instance successfully to the classifier chain.
When thread A returned grabbing the rtnl mutex again, propagating -EAGAIN
and destroying its tp instance which never got linked, we goto replay
and redo A's request.
This time when walking the classifier chain in tc_ctl_tfilter() for
checking for existing tp instances we had a priority match and found
the tp instance that was created and linked by thread B. Now calling
again into tp->ops->change() with that tp was successful and returned
without error.
tp_created was never cleared in the second round, thus kernel thinks
that we need to link it into the classifier chain (once again). tp and
*back point to the same object due to the match we had earlier on. Thus
for thread B's already public tp, we reset tp->next to tp itself and
link it into the chain, which eventually causes the mentioned endless
loop in tc_classify() once a packet hits the data path.
Fix is to clear tp_created at the beginning of each request, also when
we replay it. On the paths that can cause -EAGAIN we already destroy
the original tp instance we had and on replay we really need to start
from scratch. It seems that this issue was first introduced in commit
12186be7d2e1 ("net_cls: fix unconfigured struct tcf_proto keeps chaining
and avoid kernel panic when we use cls_cgroup").
Fixes: 12186be7d2e1 ("net_cls: fix unconfigured struct tcf_proto keeps chaining and avoid kernel panic when we use cls_cgroup")
Reported-by: Shahar Klein <shahark@mellanox.com>
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
Cc: Cong Wang <xiyou.wangcong@gmail.com>
Acked-by: Eric Dumazet <edumazet@google.com>
Tested-by: Shahar Klein <shahark@mellanox.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-12-22 00:04:11 +07:00
|
|
|
int tp_created;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2014-04-30 22:25:43 +07:00
|
|
|
if ((n->nlmsg_type != RTM_GETTFILTER) &&
|
2014-05-13 00:19:14 +07:00
|
|
|
!netlink_ns_capable(skb, net->user_ns, CAP_NET_ADMIN))
|
2012-11-16 10:03:00 +07:00
|
|
|
return -EPERM;
|
2013-03-26 00:36:33 +07:00
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
replay:
|
net, sched: fix soft lockup in tc_classify
Shahar reported a soft lockup in tc_classify(), where we run into an
endless loop when walking the classifier chain due to tp->next == tp
which is a state we should never run into. The issue only seems to
trigger under load in the tc control path.
What happens is that in tc_ctl_tfilter(), thread A allocates a new
tp, initializes it, sets tp_created to 1, and calls into tp->ops->change()
with it. In that classifier callback we had to unlock/lock the rtnl
mutex and returned with -EAGAIN. One reason why we need to drop there
is, for example, that we need to request an action module to be loaded.
This happens via tcf_exts_validate() -> tcf_action_init/_1() meaning
after we loaded and found the requested action, we need to redo the
whole request so we don't race against others. While we had to unlock
rtnl in that time, thread B's request was processed next on that CPU.
Thread B added a new tp instance successfully to the classifier chain.
When thread A returned grabbing the rtnl mutex again, propagating -EAGAIN
and destroying its tp instance which never got linked, we goto replay
and redo A's request.
This time when walking the classifier chain in tc_ctl_tfilter() for
checking for existing tp instances we had a priority match and found
the tp instance that was created and linked by thread B. Now calling
again into tp->ops->change() with that tp was successful and returned
without error.
tp_created was never cleared in the second round, thus kernel thinks
that we need to link it into the classifier chain (once again). tp and
*back point to the same object due to the match we had earlier on. Thus
for thread B's already public tp, we reset tp->next to tp itself and
link it into the chain, which eventually causes the mentioned endless
loop in tc_classify() once a packet hits the data path.
Fix is to clear tp_created at the beginning of each request, also when
we replay it. On the paths that can cause -EAGAIN we already destroy
the original tp instance we had and on replay we really need to start
from scratch. It seems that this issue was first introduced in commit
12186be7d2e1 ("net_cls: fix unconfigured struct tcf_proto keeps chaining
and avoid kernel panic when we use cls_cgroup").
Fixes: 12186be7d2e1 ("net_cls: fix unconfigured struct tcf_proto keeps chaining and avoid kernel panic when we use cls_cgroup")
Reported-by: Shahar Klein <shahark@mellanox.com>
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
Cc: Cong Wang <xiyou.wangcong@gmail.com>
Acked-by: Eric Dumazet <edumazet@google.com>
Tested-by: Shahar Klein <shahark@mellanox.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-12-22 00:04:11 +07:00
|
|
|
tp_created = 0;
|
|
|
|
|
2017-04-16 23:48:24 +07:00
|
|
|
err = nlmsg_parse(n, sizeof(*t), tca, TCA_MAX, NULL, extack);
|
2013-03-26 00:36:33 +07:00
|
|
|
if (err < 0)
|
|
|
|
return err;
|
|
|
|
|
2012-06-27 11:48:50 +07:00
|
|
|
t = nlmsg_data(n);
|
2005-04-17 05:20:36 +07:00
|
|
|
protocol = TC_H_MIN(t->tcm_info);
|
|
|
|
prio = TC_H_MAJ(t->tcm_info);
|
2017-05-17 16:07:57 +07:00
|
|
|
prio_allocate = false;
|
2005-04-17 05:20:36 +07:00
|
|
|
parent = t->tcm_parent;
|
|
|
|
cl = 0;
|
|
|
|
|
|
|
|
if (prio == 0) {
|
net, cls: allow for deleting all filters for given parent
Add a possibility where the user can just specify the parent and
all filters under that parent are then being purged. Currently,
for example for scripting, one needs to specify pref/prio to have
a well-defined number for 'tc filter del' command for addressing
the previously created instance or additionally filter handle in
case of priorities being the same. Improve usage by allowing the
option for tc to specify the parent and removing the whole chain
for that given parent.
Example usage after patch, no tc changes required:
# tc qdisc replace dev foo clsact
# tc filter add dev foo egress bpf da obj ./bpf.o
# tc filter add dev foo egress bpf da obj ./bpf.o
# tc filter show dev foo egress
filter protocol all pref 49151 bpf
filter protocol all pref 49151 bpf handle 0x1 bpf.o:[classifier] direct-action
filter protocol all pref 49152 bpf
filter protocol all pref 49152 bpf handle 0x1 bpf.o:[classifier] direct-action
# tc filter del dev foo egress
# tc filter show dev foo egress
#
Previously, RTM_DELTFILTER requests with invalid prio of 0 were
rejected, so only netlink requests with RTM_NEWTFILTER and NLM_F_CREATE
flag were allowed where the kernel would auto-generate a pref/prio.
We can piggyback on that and use prio of 0 as a wildcard for
requests of RTM_DELTFILTER.
For notifying tc netlink monitoring users (e.g. libnl uses this
for caching), there are two options, that is, sending individual
tfilter_notify() notifications for each tcf_proto, or sending a
single one indicating wildcard removal. I tried both and there
are pros and cons for each, eventually I decided for sending
individual tfilter_notify(), so that user space can support this
seamlessly and there won't be a mess of changing each and every
application to make sure expectations from the kernel won't break
when they don't understand single notification. Since linear chains
don't really scale, I expect only a handful of classifiers to be
attached at max for a given parent anyway.
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Jamal Hadi Salim <jhs@mojatatu.com>
Acked-by: Alexei Starovoitov <ast@kernel.org>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-06-11 04:10:22 +07:00
|
|
|
switch (n->nlmsg_type) {
|
|
|
|
case RTM_DELTFILTER:
|
2016-06-17 04:19:29 +07:00
|
|
|
if (protocol || t->tcm_handle || tca[TCA_KIND])
|
net, cls: allow for deleting all filters for given parent
Add a possibility where the user can just specify the parent and
all filters under that parent are then being purged. Currently,
for example for scripting, one needs to specify pref/prio to have
a well-defined number for 'tc filter del' command for addressing
the previously created instance or additionally filter handle in
case of priorities being the same. Improve usage by allowing the
option for tc to specify the parent and removing the whole chain
for that given parent.
Example usage after patch, no tc changes required:
# tc qdisc replace dev foo clsact
# tc filter add dev foo egress bpf da obj ./bpf.o
# tc filter add dev foo egress bpf da obj ./bpf.o
# tc filter show dev foo egress
filter protocol all pref 49151 bpf
filter protocol all pref 49151 bpf handle 0x1 bpf.o:[classifier] direct-action
filter protocol all pref 49152 bpf
filter protocol all pref 49152 bpf handle 0x1 bpf.o:[classifier] direct-action
# tc filter del dev foo egress
# tc filter show dev foo egress
#
Previously, RTM_DELTFILTER requests with invalid prio of 0 were
rejected, so only netlink requests with RTM_NEWTFILTER and NLM_F_CREATE
flag were allowed where the kernel would auto-generate a pref/prio.
We can piggyback on that and use prio of 0 as a wildcard for
requests of RTM_DELTFILTER.
For notifying tc netlink monitoring users (e.g. libnl uses this
for caching), there are two options, that is, sending individual
tfilter_notify() notifications for each tcf_proto, or sending a
single one indicating wildcard removal. I tried both and there
are pros and cons for each, eventually I decided for sending
individual tfilter_notify(), so that user space can support this
seamlessly and there won't be a mess of changing each and every
application to make sure expectations from the kernel won't break
when they don't understand single notification. Since linear chains
don't really scale, I expect only a handful of classifiers to be
attached at max for a given parent anyway.
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Jamal Hadi Salim <jhs@mojatatu.com>
Acked-by: Alexei Starovoitov <ast@kernel.org>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-06-11 04:10:22 +07:00
|
|
|
return -ENOENT;
|
|
|
|
break;
|
|
|
|
case RTM_NEWTFILTER:
|
|
|
|
/* If no priority is provided by the user,
|
|
|
|
* we allocate one.
|
|
|
|
*/
|
|
|
|
if (n->nlmsg_flags & NLM_F_CREATE) {
|
|
|
|
prio = TC_H_MAKE(0x80000000U, 0U);
|
2017-05-17 16:07:57 +07:00
|
|
|
prio_allocate = true;
|
net, cls: allow for deleting all filters for given parent
Add a possibility where the user can just specify the parent and
all filters under that parent are then being purged. Currently,
for example for scripting, one needs to specify pref/prio to have
a well-defined number for 'tc filter del' command for addressing
the previously created instance or additionally filter handle in
case of priorities being the same. Improve usage by allowing the
option for tc to specify the parent and removing the whole chain
for that given parent.
Example usage after patch, no tc changes required:
# tc qdisc replace dev foo clsact
# tc filter add dev foo egress bpf da obj ./bpf.o
# tc filter add dev foo egress bpf da obj ./bpf.o
# tc filter show dev foo egress
filter protocol all pref 49151 bpf
filter protocol all pref 49151 bpf handle 0x1 bpf.o:[classifier] direct-action
filter protocol all pref 49152 bpf
filter protocol all pref 49152 bpf handle 0x1 bpf.o:[classifier] direct-action
# tc filter del dev foo egress
# tc filter show dev foo egress
#
Previously, RTM_DELTFILTER requests with invalid prio of 0 were
rejected, so only netlink requests with RTM_NEWTFILTER and NLM_F_CREATE
flag were allowed where the kernel would auto-generate a pref/prio.
We can piggyback on that and use prio of 0 as a wildcard for
requests of RTM_DELTFILTER.
For notifying tc netlink monitoring users (e.g. libnl uses this
for caching), there are two options, that is, sending individual
tfilter_notify() notifications for each tcf_proto, or sending a
single one indicating wildcard removal. I tried both and there
are pros and cons for each, eventually I decided for sending
individual tfilter_notify(), so that user space can support this
seamlessly and there won't be a mess of changing each and every
application to make sure expectations from the kernel won't break
when they don't understand single notification. Since linear chains
don't really scale, I expect only a handful of classifiers to be
attached at max for a given parent anyway.
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Jamal Hadi Salim <jhs@mojatatu.com>
Acked-by: Alexei Starovoitov <ast@kernel.org>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-06-11 04:10:22 +07:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
/* fall-through */
|
|
|
|
default:
|
2005-04-17 05:20:36 +07:00
|
|
|
return -ENOENT;
|
net, cls: allow for deleting all filters for given parent
Add a possibility where the user can just specify the parent and
all filters under that parent are then being purged. Currently,
for example for scripting, one needs to specify pref/prio to have
a well-defined number for 'tc filter del' command for addressing
the previously created instance or additionally filter handle in
case of priorities being the same. Improve usage by allowing the
option for tc to specify the parent and removing the whole chain
for that given parent.
Example usage after patch, no tc changes required:
# tc qdisc replace dev foo clsact
# tc filter add dev foo egress bpf da obj ./bpf.o
# tc filter add dev foo egress bpf da obj ./bpf.o
# tc filter show dev foo egress
filter protocol all pref 49151 bpf
filter protocol all pref 49151 bpf handle 0x1 bpf.o:[classifier] direct-action
filter protocol all pref 49152 bpf
filter protocol all pref 49152 bpf handle 0x1 bpf.o:[classifier] direct-action
# tc filter del dev foo egress
# tc filter show dev foo egress
#
Previously, RTM_DELTFILTER requests with invalid prio of 0 were
rejected, so only netlink requests with RTM_NEWTFILTER and NLM_F_CREATE
flag were allowed where the kernel would auto-generate a pref/prio.
We can piggyback on that and use prio of 0 as a wildcard for
requests of RTM_DELTFILTER.
For notifying tc netlink monitoring users (e.g. libnl uses this
for caching), there are two options, that is, sending individual
tfilter_notify() notifications for each tcf_proto, or sending a
single one indicating wildcard removal. I tried both and there
are pros and cons for each, eventually I decided for sending
individual tfilter_notify(), so that user space can support this
seamlessly and there won't be a mess of changing each and every
application to make sure expectations from the kernel won't break
when they don't understand single notification. Since linear chains
don't really scale, I expect only a handful of classifiers to be
attached at max for a given parent anyway.
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Jamal Hadi Salim <jhs@mojatatu.com>
Acked-by: Alexei Starovoitov <ast@kernel.org>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-06-11 04:10:22 +07:00
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Find head of filter chain. */
|
|
|
|
|
|
|
|
/* Find link */
|
2010-03-19 22:40:13 +07:00
|
|
|
dev = __dev_get_by_index(net, t->tcm_ifindex);
|
2008-01-21 17:26:41 +07:00
|
|
|
if (dev == NULL)
|
2005-04-17 05:20:36 +07:00
|
|
|
return -ENODEV;
|
|
|
|
|
|
|
|
/* Find qdisc */
|
|
|
|
if (!parent) {
|
2009-09-04 13:41:18 +07:00
|
|
|
q = dev->qdisc;
|
2005-04-17 05:20:36 +07:00
|
|
|
parent = q->handle;
|
2008-01-21 17:26:41 +07:00
|
|
|
} else {
|
|
|
|
q = qdisc_lookup(dev, TC_H_MAJ(t->tcm_parent));
|
|
|
|
if (q == NULL)
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
/* Is it classful? */
|
2011-01-20 02:26:56 +07:00
|
|
|
cops = q->ops->cl_ops;
|
|
|
|
if (!cops)
|
2005-04-17 05:20:36 +07:00
|
|
|
return -EINVAL;
|
|
|
|
|
2017-05-17 16:07:55 +07:00
|
|
|
if (!cops->tcf_block)
|
2009-09-04 13:41:15 +07:00
|
|
|
return -EOPNOTSUPP;
|
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
/* Do we search for filter, attached to class? */
|
|
|
|
if (TC_H_MIN(parent)) {
|
net_sched: remove tc class reference counting
For TC classes, their ->get() and ->put() are always paired, and the
reference counting is completely useless, because:
1) For class modification and dumping paths, we already hold RTNL lock,
so all of these ->get(),->change(),->put() are atomic.
2) For filter bindiing/unbinding, we use other reference counter than
this one, and they should have RTNL lock too.
3) For ->qlen_notify(), it is special because it is called on ->enqueue()
path, but we already hold qdisc tree lock there, and we hold this
tree lock when graft or delete the class too, so it should not be gone
or changed until we release the tree lock.
Therefore, this patch removes ->get() and ->put(), but:
1) Adds a new ->find() to find the pointer to a class by classid, no
refcnt.
2) Move the original class destroy upon the last refcnt into ->delete(),
right after releasing tree lock. This is fine because the class is
already removed from hash when holding the lock.
For those who also use ->put() as ->unbind(), just rename them to reflect
this change.
Cc: Jamal Hadi Salim <jhs@mojatatu.com>
Signed-off-by: Cong Wang <xiyou.wangcong@gmail.com>
Acked-by: Jiri Pirko <jiri@mellanox.com>
Acked-by: Jamal Hadi Salim <jhs@mojatatu.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-08-25 06:51:29 +07:00
|
|
|
cl = cops->find(q, parent);
|
2005-04-17 05:20:36 +07:00
|
|
|
if (cl == 0)
|
|
|
|
return -ENOENT;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* And the last stroke */
|
2017-05-17 16:07:55 +07:00
|
|
|
block = cops->tcf_block(q, cl);
|
|
|
|
if (!block) {
|
2017-02-09 20:38:58 +07:00
|
|
|
err = -EINVAL;
|
2005-04-17 05:20:36 +07:00
|
|
|
goto errout;
|
2017-02-09 20:38:58 +07:00
|
|
|
}
|
2017-05-17 16:08:01 +07:00
|
|
|
|
|
|
|
chain_index = tca[TCA_CHAIN] ? nla_get_u32(tca[TCA_CHAIN]) : 0;
|
|
|
|
if (chain_index > TC_ACT_EXT_VAL_MASK) {
|
|
|
|
err = -EINVAL;
|
|
|
|
goto errout;
|
|
|
|
}
|
2017-05-23 23:42:37 +07:00
|
|
|
chain = tcf_chain_get(block, chain_index,
|
|
|
|
n->nlmsg_type == RTM_NEWTFILTER);
|
2017-05-17 16:08:01 +07:00
|
|
|
if (!chain) {
|
2017-05-23 23:42:37 +07:00
|
|
|
err = n->nlmsg_type == RTM_NEWTFILTER ? -ENOMEM : -EINVAL;
|
2017-05-17 16:08:01 +07:00
|
|
|
goto errout;
|
|
|
|
}
|
2017-05-17 16:07:55 +07:00
|
|
|
|
net, cls: allow for deleting all filters for given parent
Add a possibility where the user can just specify the parent and
all filters under that parent are then being purged. Currently,
for example for scripting, one needs to specify pref/prio to have
a well-defined number for 'tc filter del' command for addressing
the previously created instance or additionally filter handle in
case of priorities being the same. Improve usage by allowing the
option for tc to specify the parent and removing the whole chain
for that given parent.
Example usage after patch, no tc changes required:
# tc qdisc replace dev foo clsact
# tc filter add dev foo egress bpf da obj ./bpf.o
# tc filter add dev foo egress bpf da obj ./bpf.o
# tc filter show dev foo egress
filter protocol all pref 49151 bpf
filter protocol all pref 49151 bpf handle 0x1 bpf.o:[classifier] direct-action
filter protocol all pref 49152 bpf
filter protocol all pref 49152 bpf handle 0x1 bpf.o:[classifier] direct-action
# tc filter del dev foo egress
# tc filter show dev foo egress
#
Previously, RTM_DELTFILTER requests with invalid prio of 0 were
rejected, so only netlink requests with RTM_NEWTFILTER and NLM_F_CREATE
flag were allowed where the kernel would auto-generate a pref/prio.
We can piggyback on that and use prio of 0 as a wildcard for
requests of RTM_DELTFILTER.
For notifying tc netlink monitoring users (e.g. libnl uses this
for caching), there are two options, that is, sending individual
tfilter_notify() notifications for each tcf_proto, or sending a
single one indicating wildcard removal. I tried both and there
are pros and cons for each, eventually I decided for sending
individual tfilter_notify(), so that user space can support this
seamlessly and there won't be a mess of changing each and every
application to make sure expectations from the kernel won't break
when they don't understand single notification. Since linear chains
don't really scale, I expect only a handful of classifiers to be
attached at max for a given parent anyway.
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Jamal Hadi Salim <jhs@mojatatu.com>
Acked-by: Alexei Starovoitov <ast@kernel.org>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-06-11 04:10:22 +07:00
|
|
|
if (n->nlmsg_type == RTM_DELTFILTER && prio == 0) {
|
2017-10-13 19:01:05 +07:00
|
|
|
tfilter_notify_chain(net, skb, q, parent, n,
|
|
|
|
chain, RTM_DELTFILTER);
|
2017-05-20 20:01:32 +07:00
|
|
|
tcf_chain_flush(chain);
|
net, cls: allow for deleting all filters for given parent
Add a possibility where the user can just specify the parent and
all filters under that parent are then being purged. Currently,
for example for scripting, one needs to specify pref/prio to have
a well-defined number for 'tc filter del' command for addressing
the previously created instance or additionally filter handle in
case of priorities being the same. Improve usage by allowing the
option for tc to specify the parent and removing the whole chain
for that given parent.
Example usage after patch, no tc changes required:
# tc qdisc replace dev foo clsact
# tc filter add dev foo egress bpf da obj ./bpf.o
# tc filter add dev foo egress bpf da obj ./bpf.o
# tc filter show dev foo egress
filter protocol all pref 49151 bpf
filter protocol all pref 49151 bpf handle 0x1 bpf.o:[classifier] direct-action
filter protocol all pref 49152 bpf
filter protocol all pref 49152 bpf handle 0x1 bpf.o:[classifier] direct-action
# tc filter del dev foo egress
# tc filter show dev foo egress
#
Previously, RTM_DELTFILTER requests with invalid prio of 0 were
rejected, so only netlink requests with RTM_NEWTFILTER and NLM_F_CREATE
flag were allowed where the kernel would auto-generate a pref/prio.
We can piggyback on that and use prio of 0 as a wildcard for
requests of RTM_DELTFILTER.
For notifying tc netlink monitoring users (e.g. libnl uses this
for caching), there are two options, that is, sending individual
tfilter_notify() notifications for each tcf_proto, or sending a
single one indicating wildcard removal. I tried both and there
are pros and cons for each, eventually I decided for sending
individual tfilter_notify(), so that user space can support this
seamlessly and there won't be a mess of changing each and every
application to make sure expectations from the kernel won't break
when they don't understand single notification. Since linear chains
don't really scale, I expect only a handful of classifiers to be
attached at max for a given parent anyway.
Signed-off-by: Daniel Borkmann <daniel@iogearbox.net>
Acked-by: Jamal Hadi Salim <jhs@mojatatu.com>
Acked-by: Alexei Starovoitov <ast@kernel.org>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-06-11 04:10:22 +07:00
|
|
|
err = 0;
|
|
|
|
goto errout;
|
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2017-05-17 16:07:59 +07:00
|
|
|
tp = tcf_chain_tp_find(chain, &chain_info, protocol,
|
|
|
|
prio, prio_allocate);
|
|
|
|
if (IS_ERR(tp)) {
|
|
|
|
err = PTR_ERR(tp);
|
|
|
|
goto errout;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
if (tp == NULL) {
|
|
|
|
/* Proto-tcf does not exist, create new one */
|
|
|
|
|
2017-02-09 20:38:58 +07:00
|
|
|
if (tca[TCA_KIND] == NULL || !protocol) {
|
|
|
|
err = -EINVAL;
|
2005-04-17 05:20:36 +07:00
|
|
|
goto errout;
|
2017-02-09 20:38:58 +07:00
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2011-01-20 02:26:56 +07:00
|
|
|
if (n->nlmsg_type != RTM_NEWTFILTER ||
|
2017-02-09 20:38:58 +07:00
|
|
|
!(n->nlmsg_flags & NLM_F_CREATE)) {
|
|
|
|
err = -ENOENT;
|
2005-04-17 05:20:36 +07:00
|
|
|
goto errout;
|
2017-02-09 20:38:58 +07:00
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2017-05-17 16:07:57 +07:00
|
|
|
if (prio_allocate)
|
2017-05-17 16:07:59 +07:00
|
|
|
prio = tcf_auto_prio(tcf_chain_tp_prev(&chain_info));
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2017-02-09 20:38:57 +07:00
|
|
|
tp = tcf_proto_create(nla_data(tca[TCA_KIND]),
|
2017-05-17 16:08:01 +07:00
|
|
|
protocol, prio, parent, q, chain);
|
2017-02-09 20:38:57 +07:00
|
|
|
if (IS_ERR(tp)) {
|
|
|
|
err = PTR_ERR(tp);
|
2005-04-17 05:20:36 +07:00
|
|
|
goto errout;
|
|
|
|
}
|
2009-06-02 16:17:34 +07:00
|
|
|
tp_created = 1;
|
2017-02-09 20:38:58 +07:00
|
|
|
} else if (tca[TCA_KIND] && nla_strcmp(tca[TCA_KIND], tp->ops->kind)) {
|
|
|
|
err = -EINVAL;
|
2005-04-17 05:20:36 +07:00
|
|
|
goto errout;
|
2017-02-09 20:38:58 +07:00
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
fh = tp->ops->get(tp, t->tcm_handle);
|
|
|
|
|
2017-08-05 11:31:43 +07:00
|
|
|
if (!fh) {
|
2005-04-17 05:20:36 +07:00
|
|
|
if (n->nlmsg_type == RTM_DELTFILTER && t->tcm_handle == 0) {
|
2017-05-17 16:07:59 +07:00
|
|
|
tcf_chain_tp_remove(chain, &chain_info, tp);
|
2017-10-13 19:01:05 +07:00
|
|
|
tfilter_notify(net, skb, n, tp, q, parent, fh,
|
2016-10-10 10:25:55 +07:00
|
|
|
RTM_DELTFILTER, false);
|
2017-04-20 04:21:21 +07:00
|
|
|
tcf_proto_destroy(tp);
|
2005-04-17 05:20:36 +07:00
|
|
|
err = 0;
|
|
|
|
goto errout;
|
|
|
|
}
|
|
|
|
|
2008-01-21 17:26:41 +07:00
|
|
|
if (n->nlmsg_type != RTM_NEWTFILTER ||
|
2017-02-09 20:38:58 +07:00
|
|
|
!(n->nlmsg_flags & NLM_F_CREATE)) {
|
|
|
|
err = -ENOENT;
|
2005-04-17 05:20:36 +07:00
|
|
|
goto errout;
|
2017-02-09 20:38:58 +07:00
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
} else {
|
2017-04-20 04:21:21 +07:00
|
|
|
bool last;
|
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
switch (n->nlmsg_type) {
|
2007-02-09 21:25:16 +07:00
|
|
|
case RTM_NEWTFILTER:
|
2009-06-02 16:17:34 +07:00
|
|
|
if (n->nlmsg_flags & NLM_F_EXCL) {
|
|
|
|
if (tp_created)
|
2017-04-20 04:21:21 +07:00
|
|
|
tcf_proto_destroy(tp);
|
2017-02-09 20:38:58 +07:00
|
|
|
err = -EEXIST;
|
2005-04-17 05:20:36 +07:00
|
|
|
goto errout;
|
2009-06-02 16:17:34 +07:00
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
break;
|
|
|
|
case RTM_DELTFILTER:
|
2017-10-13 19:01:05 +07:00
|
|
|
err = tfilter_del_notify(net, skb, n, tp, q, parent,
|
|
|
|
fh, false, &last);
|
2017-02-09 20:39:00 +07:00
|
|
|
if (err)
|
|
|
|
goto errout;
|
2017-04-20 04:21:21 +07:00
|
|
|
if (last) {
|
2017-05-17 16:07:59 +07:00
|
|
|
tcf_chain_tp_remove(chain, &chain_info, tp);
|
2017-04-20 04:21:21 +07:00
|
|
|
tcf_proto_destroy(tp);
|
|
|
|
}
|
2017-02-14 22:27:13 +07:00
|
|
|
goto errout;
|
2005-04-17 05:20:36 +07:00
|
|
|
case RTM_GETTFILTER:
|
2017-10-13 19:01:05 +07:00
|
|
|
err = tfilter_notify(net, skb, n, tp, q, parent, fh,
|
2016-10-10 10:25:55 +07:00
|
|
|
RTM_NEWTFILTER, true);
|
2005-04-17 05:20:36 +07:00
|
|
|
goto errout;
|
|
|
|
default:
|
|
|
|
err = -EINVAL;
|
|
|
|
goto errout;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-04-26 03:54:06 +07:00
|
|
|
err = tp->ops->change(net, skb, tp, cl, t->tcm_handle, tca, &fh,
|
|
|
|
n->nlmsg_flags & NLM_F_CREATE ? TCA_ACT_NOREPLACE : TCA_ACT_REPLACE);
|
2009-06-02 16:17:34 +07:00
|
|
|
if (err == 0) {
|
2017-05-17 16:07:59 +07:00
|
|
|
if (tp_created)
|
|
|
|
tcf_chain_tp_insert(chain, &chain_info, tp);
|
2017-10-13 19:01:05 +07:00
|
|
|
tfilter_notify(net, skb, n, tp, q, parent, fh,
|
|
|
|
RTM_NEWTFILTER, false);
|
2009-06-02 16:17:34 +07:00
|
|
|
} else {
|
|
|
|
if (tp_created)
|
2017-04-20 04:21:21 +07:00
|
|
|
tcf_proto_destroy(tp);
|
2009-06-02 16:17:34 +07:00
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
errout:
|
2017-05-17 16:08:01 +07:00
|
|
|
if (chain)
|
|
|
|
tcf_chain_put(chain);
|
2005-04-17 05:20:36 +07:00
|
|
|
if (err == -EAGAIN)
|
|
|
|
/* Replay the request. */
|
|
|
|
goto replay;
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2008-01-21 17:26:41 +07:00
|
|
|
struct tcf_dump_args {
|
2005-04-17 05:20:36 +07:00
|
|
|
struct tcf_walker w;
|
|
|
|
struct sk_buff *skb;
|
|
|
|
struct netlink_callback *cb;
|
2017-10-13 19:01:05 +07:00
|
|
|
struct Qdisc *q;
|
|
|
|
u32 parent;
|
2005-04-17 05:20:36 +07:00
|
|
|
};
|
|
|
|
|
2017-08-05 11:31:43 +07:00
|
|
|
static int tcf_node_dump(struct tcf_proto *tp, void *n, struct tcf_walker *arg)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
2008-01-21 17:26:41 +07:00
|
|
|
struct tcf_dump_args *a = (void *)arg;
|
2014-01-10 07:14:01 +07:00
|
|
|
struct net *net = sock_net(a->skb->sk);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2017-10-13 19:01:05 +07:00
|
|
|
return tcf_fill_node(net, a->skb, tp, a->q, a->parent,
|
|
|
|
n, NETLINK_CB(a->cb->skb).portid,
|
2016-09-18 19:45:33 +07:00
|
|
|
a->cb->nlh->nlmsg_seq, NLM_F_MULTI,
|
|
|
|
RTM_NEWTFILTER);
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
2017-10-13 19:01:05 +07:00
|
|
|
static bool tcf_chain_dump(struct tcf_chain *chain, struct Qdisc *q, u32 parent,
|
|
|
|
struct sk_buff *skb, struct netlink_callback *cb,
|
2017-05-17 16:08:00 +07:00
|
|
|
long index_start, long *p_index)
|
|
|
|
{
|
|
|
|
struct net *net = sock_net(skb->sk);
|
|
|
|
struct tcmsg *tcm = nlmsg_data(cb->nlh);
|
|
|
|
struct tcf_dump_args arg;
|
|
|
|
struct tcf_proto *tp;
|
|
|
|
|
|
|
|
for (tp = rtnl_dereference(chain->filter_chain);
|
|
|
|
tp; tp = rtnl_dereference(tp->next), (*p_index)++) {
|
|
|
|
if (*p_index < index_start)
|
|
|
|
continue;
|
|
|
|
if (TC_H_MAJ(tcm->tcm_info) &&
|
|
|
|
TC_H_MAJ(tcm->tcm_info) != tp->prio)
|
|
|
|
continue;
|
|
|
|
if (TC_H_MIN(tcm->tcm_info) &&
|
|
|
|
TC_H_MIN(tcm->tcm_info) != tp->protocol)
|
|
|
|
continue;
|
|
|
|
if (*p_index > index_start)
|
|
|
|
memset(&cb->args[1], 0,
|
|
|
|
sizeof(cb->args) - sizeof(cb->args[0]));
|
|
|
|
if (cb->args[1] == 0) {
|
2017-10-13 19:01:05 +07:00
|
|
|
if (tcf_fill_node(net, skb, tp, q, parent, 0,
|
2017-05-17 16:08:00 +07:00
|
|
|
NETLINK_CB(cb->skb).portid,
|
|
|
|
cb->nlh->nlmsg_seq, NLM_F_MULTI,
|
|
|
|
RTM_NEWTFILTER) <= 0)
|
2017-05-17 16:08:01 +07:00
|
|
|
return false;
|
2017-05-17 16:08:00 +07:00
|
|
|
|
|
|
|
cb->args[1] = 1;
|
|
|
|
}
|
|
|
|
if (!tp->ops->walk)
|
|
|
|
continue;
|
|
|
|
arg.w.fn = tcf_node_dump;
|
|
|
|
arg.skb = skb;
|
|
|
|
arg.cb = cb;
|
2017-10-13 19:01:05 +07:00
|
|
|
arg.q = q;
|
|
|
|
arg.parent = parent;
|
2017-05-17 16:08:00 +07:00
|
|
|
arg.w.stop = 0;
|
|
|
|
arg.w.skip = cb->args[1] - 1;
|
|
|
|
arg.w.count = 0;
|
|
|
|
tp->ops->walk(tp, &arg.w);
|
|
|
|
cb->args[1] = arg.w.count + 1;
|
|
|
|
if (arg.w.stop)
|
2017-05-17 16:08:01 +07:00
|
|
|
return false;
|
2017-05-17 16:08:00 +07:00
|
|
|
}
|
2017-05-17 16:08:01 +07:00
|
|
|
return true;
|
2017-05-17 16:08:00 +07:00
|
|
|
}
|
|
|
|
|
2009-11-06 11:57:26 +07:00
|
|
|
/* called with RTNL */
|
2005-04-17 05:20:36 +07:00
|
|
|
static int tc_dump_tfilter(struct sk_buff *skb, struct netlink_callback *cb)
|
|
|
|
{
|
2008-03-26 00:26:21 +07:00
|
|
|
struct net *net = sock_net(skb->sk);
|
2017-05-17 16:08:01 +07:00
|
|
|
struct nlattr *tca[TCA_MAX + 1];
|
2005-04-17 05:20:36 +07:00
|
|
|
struct net_device *dev;
|
|
|
|
struct Qdisc *q;
|
2017-05-17 16:07:55 +07:00
|
|
|
struct tcf_block *block;
|
2017-05-17 16:07:59 +07:00
|
|
|
struct tcf_chain *chain;
|
2012-06-27 11:48:50 +07:00
|
|
|
struct tcmsg *tcm = nlmsg_data(cb->nlh);
|
2005-04-17 05:20:36 +07:00
|
|
|
unsigned long cl = 0;
|
2007-11-14 16:44:41 +07:00
|
|
|
const struct Qdisc_class_ops *cops;
|
2017-05-17 16:08:00 +07:00
|
|
|
long index_start;
|
|
|
|
long index;
|
2017-10-13 19:01:05 +07:00
|
|
|
u32 parent;
|
2017-05-17 16:08:01 +07:00
|
|
|
int err;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2013-03-27 13:47:04 +07:00
|
|
|
if (nlmsg_len(cb->nlh) < sizeof(*tcm))
|
2005-04-17 05:20:36 +07:00
|
|
|
return skb->len;
|
2017-05-17 16:08:01 +07:00
|
|
|
|
|
|
|
err = nlmsg_parse(cb->nlh, sizeof(*tcm), tca, TCA_MAX, NULL, NULL);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
|
2011-01-20 02:26:56 +07:00
|
|
|
dev = __dev_get_by_index(net, tcm->tcm_ifindex);
|
|
|
|
if (!dev)
|
2005-04-17 05:20:36 +07:00
|
|
|
return skb->len;
|
|
|
|
|
2017-10-13 19:01:05 +07:00
|
|
|
parent = tcm->tcm_parent;
|
|
|
|
if (!parent) {
|
2009-09-04 13:41:18 +07:00
|
|
|
q = dev->qdisc;
|
2017-10-13 19:01:05 +07:00
|
|
|
parent = q->handle;
|
|
|
|
} else {
|
2005-04-17 05:20:36 +07:00
|
|
|
q = qdisc_lookup(dev, TC_H_MAJ(tcm->tcm_parent));
|
2017-10-13 19:01:05 +07:00
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
if (!q)
|
|
|
|
goto out;
|
2011-01-20 02:26:56 +07:00
|
|
|
cops = q->ops->cl_ops;
|
|
|
|
if (!cops)
|
net_sched: remove tc class reference counting
For TC classes, their ->get() and ->put() are always paired, and the
reference counting is completely useless, because:
1) For class modification and dumping paths, we already hold RTNL lock,
so all of these ->get(),->change(),->put() are atomic.
2) For filter bindiing/unbinding, we use other reference counter than
this one, and they should have RTNL lock too.
3) For ->qlen_notify(), it is special because it is called on ->enqueue()
path, but we already hold qdisc tree lock there, and we hold this
tree lock when graft or delete the class too, so it should not be gone
or changed until we release the tree lock.
Therefore, this patch removes ->get() and ->put(), but:
1) Adds a new ->find() to find the pointer to a class by classid, no
refcnt.
2) Move the original class destroy upon the last refcnt into ->delete(),
right after releasing tree lock. This is fine because the class is
already removed from hash when holding the lock.
For those who also use ->put() as ->unbind(), just rename them to reflect
this change.
Cc: Jamal Hadi Salim <jhs@mojatatu.com>
Signed-off-by: Cong Wang <xiyou.wangcong@gmail.com>
Acked-by: Jiri Pirko <jiri@mellanox.com>
Acked-by: Jamal Hadi Salim <jhs@mojatatu.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-08-25 06:51:29 +07:00
|
|
|
goto out;
|
2017-05-17 16:07:55 +07:00
|
|
|
if (!cops->tcf_block)
|
net_sched: remove tc class reference counting
For TC classes, their ->get() and ->put() are always paired, and the
reference counting is completely useless, because:
1) For class modification and dumping paths, we already hold RTNL lock,
so all of these ->get(),->change(),->put() are atomic.
2) For filter bindiing/unbinding, we use other reference counter than
this one, and they should have RTNL lock too.
3) For ->qlen_notify(), it is special because it is called on ->enqueue()
path, but we already hold qdisc tree lock there, and we hold this
tree lock when graft or delete the class too, so it should not be gone
or changed until we release the tree lock.
Therefore, this patch removes ->get() and ->put(), but:
1) Adds a new ->find() to find the pointer to a class by classid, no
refcnt.
2) Move the original class destroy upon the last refcnt into ->delete(),
right after releasing tree lock. This is fine because the class is
already removed from hash when holding the lock.
For those who also use ->put() as ->unbind(), just rename them to reflect
this change.
Cc: Jamal Hadi Salim <jhs@mojatatu.com>
Signed-off-by: Cong Wang <xiyou.wangcong@gmail.com>
Acked-by: Jiri Pirko <jiri@mellanox.com>
Acked-by: Jamal Hadi Salim <jhs@mojatatu.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-08-25 06:51:29 +07:00
|
|
|
goto out;
|
2005-04-17 05:20:36 +07:00
|
|
|
if (TC_H_MIN(tcm->tcm_parent)) {
|
net_sched: remove tc class reference counting
For TC classes, their ->get() and ->put() are always paired, and the
reference counting is completely useless, because:
1) For class modification and dumping paths, we already hold RTNL lock,
so all of these ->get(),->change(),->put() are atomic.
2) For filter bindiing/unbinding, we use other reference counter than
this one, and they should have RTNL lock too.
3) For ->qlen_notify(), it is special because it is called on ->enqueue()
path, but we already hold qdisc tree lock there, and we hold this
tree lock when graft or delete the class too, so it should not be gone
or changed until we release the tree lock.
Therefore, this patch removes ->get() and ->put(), but:
1) Adds a new ->find() to find the pointer to a class by classid, no
refcnt.
2) Move the original class destroy upon the last refcnt into ->delete(),
right after releasing tree lock. This is fine because the class is
already removed from hash when holding the lock.
For those who also use ->put() as ->unbind(), just rename them to reflect
this change.
Cc: Jamal Hadi Salim <jhs@mojatatu.com>
Signed-off-by: Cong Wang <xiyou.wangcong@gmail.com>
Acked-by: Jiri Pirko <jiri@mellanox.com>
Acked-by: Jamal Hadi Salim <jhs@mojatatu.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-08-25 06:51:29 +07:00
|
|
|
cl = cops->find(q, tcm->tcm_parent);
|
2005-04-17 05:20:36 +07:00
|
|
|
if (cl == 0)
|
net_sched: remove tc class reference counting
For TC classes, their ->get() and ->put() are always paired, and the
reference counting is completely useless, because:
1) For class modification and dumping paths, we already hold RTNL lock,
so all of these ->get(),->change(),->put() are atomic.
2) For filter bindiing/unbinding, we use other reference counter than
this one, and they should have RTNL lock too.
3) For ->qlen_notify(), it is special because it is called on ->enqueue()
path, but we already hold qdisc tree lock there, and we hold this
tree lock when graft or delete the class too, so it should not be gone
or changed until we release the tree lock.
Therefore, this patch removes ->get() and ->put(), but:
1) Adds a new ->find() to find the pointer to a class by classid, no
refcnt.
2) Move the original class destroy upon the last refcnt into ->delete(),
right after releasing tree lock. This is fine because the class is
already removed from hash when holding the lock.
For those who also use ->put() as ->unbind(), just rename them to reflect
this change.
Cc: Jamal Hadi Salim <jhs@mojatatu.com>
Signed-off-by: Cong Wang <xiyou.wangcong@gmail.com>
Acked-by: Jiri Pirko <jiri@mellanox.com>
Acked-by: Jamal Hadi Salim <jhs@mojatatu.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-08-25 06:51:29 +07:00
|
|
|
goto out;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
2017-05-17 16:07:55 +07:00
|
|
|
block = cops->tcf_block(q, cl);
|
|
|
|
if (!block)
|
net_sched: remove tc class reference counting
For TC classes, their ->get() and ->put() are always paired, and the
reference counting is completely useless, because:
1) For class modification and dumping paths, we already hold RTNL lock,
so all of these ->get(),->change(),->put() are atomic.
2) For filter bindiing/unbinding, we use other reference counter than
this one, and they should have RTNL lock too.
3) For ->qlen_notify(), it is special because it is called on ->enqueue()
path, but we already hold qdisc tree lock there, and we hold this
tree lock when graft or delete the class too, so it should not be gone
or changed until we release the tree lock.
Therefore, this patch removes ->get() and ->put(), but:
1) Adds a new ->find() to find the pointer to a class by classid, no
refcnt.
2) Move the original class destroy upon the last refcnt into ->delete(),
right after releasing tree lock. This is fine because the class is
already removed from hash when holding the lock.
For those who also use ->put() as ->unbind(), just rename them to reflect
this change.
Cc: Jamal Hadi Salim <jhs@mojatatu.com>
Signed-off-by: Cong Wang <xiyou.wangcong@gmail.com>
Acked-by: Jiri Pirko <jiri@mellanox.com>
Acked-by: Jamal Hadi Salim <jhs@mojatatu.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-08-25 06:51:29 +07:00
|
|
|
goto out;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2017-05-17 16:08:00 +07:00
|
|
|
index_start = cb->args[0];
|
|
|
|
index = 0;
|
2017-05-17 16:08:01 +07:00
|
|
|
|
|
|
|
list_for_each_entry(chain, &block->chain_list, list) {
|
|
|
|
if (tca[TCA_CHAIN] &&
|
|
|
|
nla_get_u32(tca[TCA_CHAIN]) != chain->index)
|
|
|
|
continue;
|
2017-10-13 19:01:05 +07:00
|
|
|
if (!tcf_chain_dump(chain, q, parent, skb, cb,
|
|
|
|
index_start, &index))
|
2017-05-17 16:08:01 +07:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2017-05-17 16:08:00 +07:00
|
|
|
cb->args[0] = index;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
out:
|
|
|
|
return skb->len;
|
|
|
|
}
|
|
|
|
|
2014-09-26 00:26:37 +07:00
|
|
|
void tcf_exts_destroy(struct tcf_exts *exts)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
|
|
|
#ifdef CONFIG_NET_CLS_ACT
|
2016-08-14 12:35:00 +07:00
|
|
|
LIST_HEAD(actions);
|
|
|
|
|
2017-10-27 08:24:40 +07:00
|
|
|
ASSERT_RTNL();
|
2016-08-14 12:35:00 +07:00
|
|
|
tcf_exts_to_list(exts, &actions);
|
|
|
|
tcf_action_destroy(&actions, TCA_ACT_UNBIND);
|
|
|
|
kfree(exts->actions);
|
|
|
|
exts->nr_actions = 0;
|
2005-04-17 05:20:36 +07:00
|
|
|
#endif
|
|
|
|
}
|
2008-01-21 17:26:41 +07:00
|
|
|
EXPORT_SYMBOL(tcf_exts_destroy);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2013-01-14 12:15:39 +07:00
|
|
|
int tcf_exts_validate(struct net *net, struct tcf_proto *tp, struct nlattr **tb,
|
2016-09-18 19:45:33 +07:00
|
|
|
struct nlattr *rate_tlv, struct tcf_exts *exts, bool ovr)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
|
|
|
#ifdef CONFIG_NET_CLS_ACT
|
|
|
|
{
|
|
|
|
struct tc_action *act;
|
|
|
|
|
2013-12-16 11:15:07 +07:00
|
|
|
if (exts->police && tb[exts->police]) {
|
2017-05-17 16:08:02 +07:00
|
|
|
act = tcf_action_init_1(net, tp, tb[exts->police],
|
|
|
|
rate_tlv, "police", ovr,
|
|
|
|
TCA_ACT_BIND);
|
2008-01-24 11:33:13 +07:00
|
|
|
if (IS_ERR(act))
|
|
|
|
return PTR_ERR(act);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2013-12-16 11:15:05 +07:00
|
|
|
act->type = exts->type = TCA_OLD_COMPAT;
|
2016-08-14 12:35:00 +07:00
|
|
|
exts->actions[0] = act;
|
|
|
|
exts->nr_actions = 1;
|
2013-12-16 11:15:07 +07:00
|
|
|
} else if (exts->action && tb[exts->action]) {
|
2016-08-14 12:35:00 +07:00
|
|
|
LIST_HEAD(actions);
|
|
|
|
int err, i = 0;
|
|
|
|
|
2017-05-17 16:08:02 +07:00
|
|
|
err = tcf_action_init(net, tp, tb[exts->action],
|
|
|
|
rate_tlv, NULL, ovr, TCA_ACT_BIND,
|
2016-09-18 19:45:33 +07:00
|
|
|
&actions);
|
2013-12-16 11:15:05 +07:00
|
|
|
if (err)
|
|
|
|
return err;
|
2016-08-14 12:35:00 +07:00
|
|
|
list_for_each_entry(act, &actions, list)
|
|
|
|
exts->actions[i++] = act;
|
|
|
|
exts->nr_actions = i;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
2017-11-07 04:47:19 +07:00
|
|
|
exts->net = net;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
#else
|
2013-12-16 11:15:07 +07:00
|
|
|
if ((exts->action && tb[exts->action]) ||
|
|
|
|
(exts->police && tb[exts->police]))
|
2005-04-17 05:20:36 +07:00
|
|
|
return -EOPNOTSUPP;
|
|
|
|
#endif
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
2008-01-21 17:26:41 +07:00
|
|
|
EXPORT_SYMBOL(tcf_exts_validate);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2017-08-04 19:29:15 +07:00
|
|
|
void tcf_exts_change(struct tcf_exts *dst, struct tcf_exts *src)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
|
|
|
#ifdef CONFIG_NET_CLS_ACT
|
2016-08-14 12:35:00 +07:00
|
|
|
struct tcf_exts old = *dst;
|
|
|
|
|
2017-08-04 19:29:15 +07:00
|
|
|
*dst = *src;
|
2016-08-14 12:35:00 +07:00
|
|
|
tcf_exts_destroy(&old);
|
2005-04-17 05:20:36 +07:00
|
|
|
#endif
|
|
|
|
}
|
2008-01-21 17:26:41 +07:00
|
|
|
EXPORT_SYMBOL(tcf_exts_change);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2016-08-14 12:35:00 +07:00
|
|
|
#ifdef CONFIG_NET_CLS_ACT
|
|
|
|
static struct tc_action *tcf_exts_first_act(struct tcf_exts *exts)
|
|
|
|
{
|
|
|
|
if (exts->nr_actions == 0)
|
|
|
|
return NULL;
|
|
|
|
else
|
|
|
|
return exts->actions[0];
|
|
|
|
}
|
|
|
|
#endif
|
2013-12-16 11:15:05 +07:00
|
|
|
|
2013-12-16 11:15:07 +07:00
|
|
|
int tcf_exts_dump(struct sk_buff *skb, struct tcf_exts *exts)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
|
|
|
#ifdef CONFIG_NET_CLS_ACT
|
2014-07-17 04:25:30 +07:00
|
|
|
struct nlattr *nest;
|
|
|
|
|
2017-08-04 19:29:03 +07:00
|
|
|
if (exts->action && tcf_exts_has_actions(exts)) {
|
2005-04-17 05:20:36 +07:00
|
|
|
/*
|
|
|
|
* again for backward compatible mode - we want
|
|
|
|
* to work with both old and new modes of entering
|
|
|
|
* tc data even if iproute2 was newer - jhs
|
|
|
|
*/
|
2013-12-16 11:15:05 +07:00
|
|
|
if (exts->type != TCA_OLD_COMPAT) {
|
2016-08-14 12:35:00 +07:00
|
|
|
LIST_HEAD(actions);
|
|
|
|
|
2013-12-16 11:15:07 +07:00
|
|
|
nest = nla_nest_start(skb, exts->action);
|
2008-01-24 11:34:11 +07:00
|
|
|
if (nest == NULL)
|
|
|
|
goto nla_put_failure;
|
2016-08-14 12:35:00 +07:00
|
|
|
|
|
|
|
tcf_exts_to_list(exts, &actions);
|
|
|
|
if (tcf_action_dump(skb, &actions, 0, 0) < 0)
|
2008-01-23 13:11:33 +07:00
|
|
|
goto nla_put_failure;
|
2008-01-24 11:34:11 +07:00
|
|
|
nla_nest_end(skb, nest);
|
2013-12-16 11:15:07 +07:00
|
|
|
} else if (exts->police) {
|
2013-12-16 11:15:05 +07:00
|
|
|
struct tc_action *act = tcf_exts_first_act(exts);
|
2013-12-16 11:15:07 +07:00
|
|
|
nest = nla_nest_start(skb, exts->police);
|
2013-12-23 20:02:12 +07:00
|
|
|
if (nest == NULL || !act)
|
2008-01-24 11:34:11 +07:00
|
|
|
goto nla_put_failure;
|
2013-12-16 11:15:05 +07:00
|
|
|
if (tcf_action_dump_old(skb, act, 0, 0) < 0)
|
2008-01-23 13:11:33 +07:00
|
|
|
goto nla_put_failure;
|
2008-01-24 11:34:11 +07:00
|
|
|
nla_nest_end(skb, nest);
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return 0;
|
2014-07-17 04:25:30 +07:00
|
|
|
|
|
|
|
nla_put_failure:
|
|
|
|
nla_nest_cancel(skb, nest);
|
2005-04-17 05:20:36 +07:00
|
|
|
return -1;
|
2014-07-17 04:25:30 +07:00
|
|
|
#else
|
|
|
|
return 0;
|
|
|
|
#endif
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
2008-01-21 17:26:41 +07:00
|
|
|
EXPORT_SYMBOL(tcf_exts_dump);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2008-01-21 17:26:41 +07:00
|
|
|
|
2013-12-16 11:15:07 +07:00
|
|
|
int tcf_exts_dump_stats(struct sk_buff *skb, struct tcf_exts *exts)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
|
|
|
#ifdef CONFIG_NET_CLS_ACT
|
2013-12-16 11:15:05 +07:00
|
|
|
struct tc_action *a = tcf_exts_first_act(exts);
|
2015-02-04 01:05:18 +07:00
|
|
|
if (a != NULL && tcf_action_copy_stats(skb, a, 1) < 0)
|
2013-12-16 11:15:05 +07:00
|
|
|
return -1;
|
2005-04-17 05:20:36 +07:00
|
|
|
#endif
|
|
|
|
return 0;
|
|
|
|
}
|
2008-01-21 17:26:41 +07:00
|
|
|
EXPORT_SYMBOL(tcf_exts_dump_stats);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2017-10-11 14:41:09 +07:00
|
|
|
static int tc_exts_setup_cb_egdev_call(struct tcf_exts *exts,
|
|
|
|
enum tc_setup_type type,
|
|
|
|
void *type_data, bool err_stop)
|
2017-10-11 14:41:08 +07:00
|
|
|
{
|
|
|
|
int ok_count = 0;
|
|
|
|
#ifdef CONFIG_NET_CLS_ACT
|
|
|
|
const struct tc_action *a;
|
|
|
|
struct net_device *dev;
|
2017-10-24 12:58:02 +07:00
|
|
|
int i, ret;
|
2017-10-11 14:41:08 +07:00
|
|
|
|
|
|
|
if (!tcf_exts_has_actions(exts))
|
|
|
|
return 0;
|
|
|
|
|
2017-10-24 12:58:02 +07:00
|
|
|
for (i = 0; i < exts->nr_actions; i++) {
|
|
|
|
a = exts->actions[i];
|
2017-10-11 14:41:08 +07:00
|
|
|
if (!a->ops->get_dev)
|
|
|
|
continue;
|
|
|
|
dev = a->ops->get_dev(a);
|
2017-11-01 17:47:40 +07:00
|
|
|
if (!dev)
|
2017-10-11 14:41:08 +07:00
|
|
|
continue;
|
|
|
|
ret = tc_setup_cb_egdev_call(dev, type, type_data, err_stop);
|
|
|
|
if (ret < 0)
|
|
|
|
return ret;
|
|
|
|
ok_count += ret;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
return ok_count;
|
|
|
|
}
|
2017-10-11 14:41:09 +07:00
|
|
|
|
2017-10-19 20:50:32 +07:00
|
|
|
int tc_setup_cb_call(struct tcf_block *block, struct tcf_exts *exts,
|
|
|
|
enum tc_setup_type type, void *type_data, bool err_stop)
|
2017-10-11 14:41:09 +07:00
|
|
|
{
|
2017-10-19 20:50:32 +07:00
|
|
|
int ok_count;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = tcf_block_cb_call(block, type, type_data, err_stop);
|
|
|
|
if (ret < 0)
|
|
|
|
return ret;
|
|
|
|
ok_count = ret;
|
|
|
|
|
|
|
|
if (!exts)
|
|
|
|
return ok_count;
|
|
|
|
ret = tc_exts_setup_cb_egdev_call(exts, type, type_data, err_stop);
|
|
|
|
if (ret < 0)
|
|
|
|
return ret;
|
|
|
|
ok_count += ret;
|
|
|
|
|
|
|
|
return ok_count;
|
2017-10-11 14:41:09 +07:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(tc_setup_cb_call);
|
2017-10-11 14:41:08 +07:00
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
static int __init tc_filter_init(void)
|
|
|
|
{
|
2017-10-27 08:24:28 +07:00
|
|
|
tc_filter_wq = alloc_ordered_workqueue("tc_filter_workqueue", 0);
|
|
|
|
if (!tc_filter_wq)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2017-08-10 01:41:48 +07:00
|
|
|
rtnl_register(PF_UNSPEC, RTM_NEWTFILTER, tc_ctl_tfilter, NULL, 0);
|
|
|
|
rtnl_register(PF_UNSPEC, RTM_DELTFILTER, tc_ctl_tfilter, NULL, 0);
|
2007-03-23 01:56:22 +07:00
|
|
|
rtnl_register(PF_UNSPEC, RTM_GETTFILTER, tc_ctl_tfilter,
|
2017-08-10 01:41:48 +07:00
|
|
|
tc_dump_tfilter, 0);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
subsys_initcall(tc_filter_init);
|