dsa: bcm_sf2: use flow_rule infrastructure

Update this driver to use the flow_rule infrastructure, hence we can use
the same code to populate hardware IR from ethtool_rx_flow and the
cls_flower interfaces.

Signed-off-by: Pablo Neira Ayuso <pablo@netfilter.org>
Acked-by: Jiri Pirko <jiri@mellanox.com>
Reviewed-by: Florian Fainelli <f.fainelli@gmail.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
Pablo Neira Ayuso 2019-02-02 12:50:52 +01:00 committed by David S. Miller
parent eca4205f9e
commit e4f7ef54cb

View File

@ -16,6 +16,7 @@
#include <linux/netdevice.h> #include <linux/netdevice.h>
#include <net/dsa.h> #include <net/dsa.h>
#include <linux/bitmap.h> #include <linux/bitmap.h>
#include <net/flow_offload.h>
#include "bcm_sf2.h" #include "bcm_sf2.h"
#include "bcm_sf2_regs.h" #include "bcm_sf2_regs.h"
@ -257,7 +258,8 @@ static int bcm_sf2_cfp_act_pol_set(struct bcm_sf2_priv *priv,
} }
static void bcm_sf2_cfp_slice_ipv4(struct bcm_sf2_priv *priv, static void bcm_sf2_cfp_slice_ipv4(struct bcm_sf2_priv *priv,
struct ethtool_tcpip4_spec *v4_spec, struct flow_dissector_key_ipv4_addrs *addrs,
struct flow_dissector_key_ports *ports,
unsigned int slice_num, unsigned int slice_num,
bool mask) bool mask)
{ {
@ -278,7 +280,7 @@ static void bcm_sf2_cfp_slice_ipv4(struct bcm_sf2_priv *priv,
* UDF_n_A6 [23:8] * UDF_n_A6 [23:8]
* UDF_n_A5 [7:0] * UDF_n_A5 [7:0]
*/ */
reg = be16_to_cpu(v4_spec->pdst) >> 8; reg = be16_to_cpu(ports->dst) >> 8;
if (mask) if (mask)
offset = CORE_CFP_MASK_PORT(3); offset = CORE_CFP_MASK_PORT(3);
else else
@ -289,9 +291,9 @@ static void bcm_sf2_cfp_slice_ipv4(struct bcm_sf2_priv *priv,
* UDF_n_A4 [23:8] * UDF_n_A4 [23:8]
* UDF_n_A3 [7:0] * UDF_n_A3 [7:0]
*/ */
reg = (be16_to_cpu(v4_spec->pdst) & 0xff) << 24 | reg = (be16_to_cpu(ports->dst) & 0xff) << 24 |
(u32)be16_to_cpu(v4_spec->psrc) << 8 | (u32)be16_to_cpu(ports->src) << 8 |
(be32_to_cpu(v4_spec->ip4dst) & 0x0000ff00) >> 8; (be32_to_cpu(addrs->dst) & 0x0000ff00) >> 8;
if (mask) if (mask)
offset = CORE_CFP_MASK_PORT(2); offset = CORE_CFP_MASK_PORT(2);
else else
@ -302,9 +304,9 @@ static void bcm_sf2_cfp_slice_ipv4(struct bcm_sf2_priv *priv,
* UDF_n_A2 [23:8] * UDF_n_A2 [23:8]
* UDF_n_A1 [7:0] * UDF_n_A1 [7:0]
*/ */
reg = (u32)(be32_to_cpu(v4_spec->ip4dst) & 0xff) << 24 | reg = (u32)(be32_to_cpu(addrs->dst) & 0xff) << 24 |
(u32)(be32_to_cpu(v4_spec->ip4dst) >> 16) << 8 | (u32)(be32_to_cpu(addrs->dst) >> 16) << 8 |
(be32_to_cpu(v4_spec->ip4src) & 0x0000ff00) >> 8; (be32_to_cpu(addrs->src) & 0x0000ff00) >> 8;
if (mask) if (mask)
offset = CORE_CFP_MASK_PORT(1); offset = CORE_CFP_MASK_PORT(1);
else else
@ -317,8 +319,8 @@ static void bcm_sf2_cfp_slice_ipv4(struct bcm_sf2_priv *priv,
* Slice ID [3:2] * Slice ID [3:2]
* Slice valid [1:0] * Slice valid [1:0]
*/ */
reg = (u32)(be32_to_cpu(v4_spec->ip4src) & 0xff) << 24 | reg = (u32)(be32_to_cpu(addrs->src) & 0xff) << 24 |
(u32)(be32_to_cpu(v4_spec->ip4src) >> 16) << 8 | (u32)(be32_to_cpu(addrs->src) >> 16) << 8 |
SLICE_NUM(slice_num) | SLICE_VALID; SLICE_NUM(slice_num) | SLICE_VALID;
if (mask) if (mask)
offset = CORE_CFP_MASK_PORT(0); offset = CORE_CFP_MASK_PORT(0);
@ -332,9 +334,13 @@ static int bcm_sf2_cfp_ipv4_rule_set(struct bcm_sf2_priv *priv, int port,
unsigned int queue_num, unsigned int queue_num,
struct ethtool_rx_flow_spec *fs) struct ethtool_rx_flow_spec *fs)
{ {
struct ethtool_tcpip4_spec *v4_spec, *v4_m_spec; struct ethtool_rx_flow_spec_input input = {};
const struct cfp_udf_layout *layout; const struct cfp_udf_layout *layout;
unsigned int slice_num, rule_index; unsigned int slice_num, rule_index;
struct ethtool_rx_flow_rule *flow;
struct flow_match_ipv4_addrs ipv4;
struct flow_match_ports ports;
struct flow_match_ip ip;
u8 ip_proto, ip_frag; u8 ip_proto, ip_frag;
u8 num_udf; u8 num_udf;
u32 reg; u32 reg;
@ -343,13 +349,9 @@ static int bcm_sf2_cfp_ipv4_rule_set(struct bcm_sf2_priv *priv, int port,
switch (fs->flow_type & ~FLOW_EXT) { switch (fs->flow_type & ~FLOW_EXT) {
case TCP_V4_FLOW: case TCP_V4_FLOW:
ip_proto = IPPROTO_TCP; ip_proto = IPPROTO_TCP;
v4_spec = &fs->h_u.tcp_ip4_spec;
v4_m_spec = &fs->m_u.tcp_ip4_spec;
break; break;
case UDP_V4_FLOW: case UDP_V4_FLOW:
ip_proto = IPPROTO_UDP; ip_proto = IPPROTO_UDP;
v4_spec = &fs->h_u.udp_ip4_spec;
v4_m_spec = &fs->m_u.udp_ip4_spec;
break; break;
default: default:
return -EINVAL; return -EINVAL;
@ -367,11 +369,22 @@ static int bcm_sf2_cfp_ipv4_rule_set(struct bcm_sf2_priv *priv, int port,
if (rule_index > bcm_sf2_cfp_rule_size(priv)) if (rule_index > bcm_sf2_cfp_rule_size(priv))
return -ENOSPC; return -ENOSPC;
input.fs = fs;
flow = ethtool_rx_flow_rule_create(&input);
if (IS_ERR(flow))
return PTR_ERR(flow);
flow_rule_match_ipv4_addrs(flow->rule, &ipv4);
flow_rule_match_ports(flow->rule, &ports);
flow_rule_match_ip(flow->rule, &ip);
layout = &udf_tcpip4_layout; layout = &udf_tcpip4_layout;
/* We only use one UDF slice for now */ /* We only use one UDF slice for now */
slice_num = bcm_sf2_get_slice_number(layout, 0); slice_num = bcm_sf2_get_slice_number(layout, 0);
if (slice_num == UDF_NUM_SLICES) if (slice_num == UDF_NUM_SLICES) {
return -EINVAL; ret = -EINVAL;
goto out_err_flow_rule;
}
num_udf = bcm_sf2_get_num_udf_slices(layout->udfs[slice_num].slices); num_udf = bcm_sf2_get_num_udf_slices(layout->udfs[slice_num].slices);
@ -398,7 +411,7 @@ static int bcm_sf2_cfp_ipv4_rule_set(struct bcm_sf2_priv *priv, int port,
* Reserved [1] * Reserved [1]
* UDF_Valid[8] [0] * UDF_Valid[8] [0]
*/ */
core_writel(priv, v4_spec->tos << IPTOS_SHIFT | core_writel(priv, ip.key->tos << IPTOS_SHIFT |
ip_proto << IPPROTO_SHIFT | ip_frag << IP_FRAG_SHIFT | ip_proto << IPPROTO_SHIFT | ip_frag << IP_FRAG_SHIFT |
udf_upper_bits(num_udf), udf_upper_bits(num_udf),
CORE_CFP_DATA_PORT(6)); CORE_CFP_DATA_PORT(6));
@ -417,8 +430,8 @@ static int bcm_sf2_cfp_ipv4_rule_set(struct bcm_sf2_priv *priv, int port,
core_writel(priv, udf_lower_bits(num_udf) << 24, CORE_CFP_MASK_PORT(5)); core_writel(priv, udf_lower_bits(num_udf) << 24, CORE_CFP_MASK_PORT(5));
/* Program the match and the mask */ /* Program the match and the mask */
bcm_sf2_cfp_slice_ipv4(priv, v4_spec, slice_num, false); bcm_sf2_cfp_slice_ipv4(priv, ipv4.key, ports.key, slice_num, false);
bcm_sf2_cfp_slice_ipv4(priv, v4_m_spec, SLICE_NUM_MASK, true); bcm_sf2_cfp_slice_ipv4(priv, ipv4.mask, ports.mask, SLICE_NUM_MASK, true);
/* Insert into TCAM now */ /* Insert into TCAM now */
bcm_sf2_cfp_rule_addr_set(priv, rule_index); bcm_sf2_cfp_rule_addr_set(priv, rule_index);
@ -426,14 +439,14 @@ static int bcm_sf2_cfp_ipv4_rule_set(struct bcm_sf2_priv *priv, int port,
ret = bcm_sf2_cfp_op(priv, OP_SEL_WRITE | TCAM_SEL); ret = bcm_sf2_cfp_op(priv, OP_SEL_WRITE | TCAM_SEL);
if (ret) { if (ret) {
pr_err("TCAM entry at addr %d failed\n", rule_index); pr_err("TCAM entry at addr %d failed\n", rule_index);
return ret; goto out_err_flow_rule;
} }
/* Insert into Action and policer RAMs now */ /* Insert into Action and policer RAMs now */
ret = bcm_sf2_cfp_act_pol_set(priv, rule_index, port_num, ret = bcm_sf2_cfp_act_pol_set(priv, rule_index, port_num,
queue_num, true); queue_num, true);
if (ret) if (ret)
return ret; goto out_err_flow_rule;
/* Turn on CFP for this rule now */ /* Turn on CFP for this rule now */
reg = core_readl(priv, CORE_CFP_CTL_REG); reg = core_readl(priv, CORE_CFP_CTL_REG);
@ -446,6 +459,10 @@ static int bcm_sf2_cfp_ipv4_rule_set(struct bcm_sf2_priv *priv, int port,
fs->location = rule_index; fs->location = rule_index;
return 0; return 0;
out_err_flow_rule:
ethtool_rx_flow_rule_destroy(flow);
return ret;
} }
static void bcm_sf2_cfp_slice_ipv6(struct bcm_sf2_priv *priv, static void bcm_sf2_cfp_slice_ipv6(struct bcm_sf2_priv *priv,
@ -582,8 +599,12 @@ static int bcm_sf2_cfp_ipv6_rule_set(struct bcm_sf2_priv *priv, int port,
struct ethtool_rx_flow_spec *fs) struct ethtool_rx_flow_spec *fs)
{ {
struct ethtool_tcpip6_spec *v6_spec, *v6_m_spec; struct ethtool_tcpip6_spec *v6_spec, *v6_m_spec;
struct ethtool_rx_flow_spec_input input = {};
unsigned int slice_num, rule_index[2]; unsigned int slice_num, rule_index[2];
const struct cfp_udf_layout *layout; const struct cfp_udf_layout *layout;
struct ethtool_rx_flow_rule *flow;
struct flow_match_ipv6_addrs ipv6;
struct flow_match_ports ports;
u8 ip_proto, ip_frag; u8 ip_proto, ip_frag;
int ret = 0; int ret = 0;
u8 num_udf; u8 num_udf;
@ -645,6 +666,15 @@ static int bcm_sf2_cfp_ipv6_rule_set(struct bcm_sf2_priv *priv, int port,
goto out_err; goto out_err;
} }
input.fs = fs;
flow = ethtool_rx_flow_rule_create(&input);
if (IS_ERR(flow)) {
ret = PTR_ERR(flow);
goto out_err;
}
flow_rule_match_ipv6_addrs(flow->rule, &ipv6);
flow_rule_match_ports(flow->rule, &ports);
/* Apply the UDF layout for this filter */ /* Apply the UDF layout for this filter */
bcm_sf2_cfp_udf_set(priv, layout, slice_num); bcm_sf2_cfp_udf_set(priv, layout, slice_num);
@ -688,10 +718,10 @@ static int bcm_sf2_cfp_ipv6_rule_set(struct bcm_sf2_priv *priv, int port,
core_writel(priv, udf_lower_bits(num_udf) << 24, CORE_CFP_MASK_PORT(5)); core_writel(priv, udf_lower_bits(num_udf) << 24, CORE_CFP_MASK_PORT(5));
/* Slice the IPv6 source address and port */ /* Slice the IPv6 source address and port */
bcm_sf2_cfp_slice_ipv6(priv, v6_spec->ip6src, v6_spec->psrc, bcm_sf2_cfp_slice_ipv6(priv, ipv6.key->src.in6_u.u6_addr32,
slice_num, false); ports.key->src, slice_num, false);
bcm_sf2_cfp_slice_ipv6(priv, v6_m_spec->ip6src, v6_m_spec->psrc, bcm_sf2_cfp_slice_ipv6(priv, ipv6.mask->src.in6_u.u6_addr32,
SLICE_NUM_MASK, true); ports.mask->src, SLICE_NUM_MASK, true);
/* Insert into TCAM now because we need to insert a second rule */ /* Insert into TCAM now because we need to insert a second rule */
bcm_sf2_cfp_rule_addr_set(priv, rule_index[0]); bcm_sf2_cfp_rule_addr_set(priv, rule_index[0]);
@ -699,20 +729,20 @@ static int bcm_sf2_cfp_ipv6_rule_set(struct bcm_sf2_priv *priv, int port,
ret = bcm_sf2_cfp_op(priv, OP_SEL_WRITE | TCAM_SEL); ret = bcm_sf2_cfp_op(priv, OP_SEL_WRITE | TCAM_SEL);
if (ret) { if (ret) {
pr_err("TCAM entry at addr %d failed\n", rule_index[0]); pr_err("TCAM entry at addr %d failed\n", rule_index[0]);
goto out_err; goto out_err_flow_rule;
} }
/* Insert into Action and policer RAMs now */ /* Insert into Action and policer RAMs now */
ret = bcm_sf2_cfp_act_pol_set(priv, rule_index[0], port_num, ret = bcm_sf2_cfp_act_pol_set(priv, rule_index[0], port_num,
queue_num, false); queue_num, false);
if (ret) if (ret)
goto out_err; goto out_err_flow_rule;
/* Now deal with the second slice to chain this rule */ /* Now deal with the second slice to chain this rule */
slice_num = bcm_sf2_get_slice_number(layout, slice_num + 1); slice_num = bcm_sf2_get_slice_number(layout, slice_num + 1);
if (slice_num == UDF_NUM_SLICES) { if (slice_num == UDF_NUM_SLICES) {
ret = -EINVAL; ret = -EINVAL;
goto out_err; goto out_err_flow_rule;
} }
num_udf = bcm_sf2_get_num_udf_slices(layout->udfs[slice_num].slices); num_udf = bcm_sf2_get_num_udf_slices(layout->udfs[slice_num].slices);
@ -748,10 +778,10 @@ static int bcm_sf2_cfp_ipv6_rule_set(struct bcm_sf2_priv *priv, int port,
/* Mask all */ /* Mask all */
core_writel(priv, 0, CORE_CFP_MASK_PORT(5)); core_writel(priv, 0, CORE_CFP_MASK_PORT(5));
bcm_sf2_cfp_slice_ipv6(priv, v6_spec->ip6dst, v6_spec->pdst, slice_num, bcm_sf2_cfp_slice_ipv6(priv, ipv6.key->dst.in6_u.u6_addr32,
false); ports.key->dst, slice_num, false);
bcm_sf2_cfp_slice_ipv6(priv, v6_m_spec->ip6dst, v6_m_spec->pdst, bcm_sf2_cfp_slice_ipv6(priv, ipv6.mask->dst.in6_u.u6_addr32,
SLICE_NUM_MASK, true); ports.key->dst, SLICE_NUM_MASK, true);
/* Insert into TCAM now */ /* Insert into TCAM now */
bcm_sf2_cfp_rule_addr_set(priv, rule_index[1]); bcm_sf2_cfp_rule_addr_set(priv, rule_index[1]);
@ -759,7 +789,7 @@ static int bcm_sf2_cfp_ipv6_rule_set(struct bcm_sf2_priv *priv, int port,
ret = bcm_sf2_cfp_op(priv, OP_SEL_WRITE | TCAM_SEL); ret = bcm_sf2_cfp_op(priv, OP_SEL_WRITE | TCAM_SEL);
if (ret) { if (ret) {
pr_err("TCAM entry at addr %d failed\n", rule_index[1]); pr_err("TCAM entry at addr %d failed\n", rule_index[1]);
goto out_err; goto out_err_flow_rule;
} }
/* Insert into Action and policer RAMs now, set chain ID to /* Insert into Action and policer RAMs now, set chain ID to
@ -768,7 +798,7 @@ static int bcm_sf2_cfp_ipv6_rule_set(struct bcm_sf2_priv *priv, int port,
ret = bcm_sf2_cfp_act_pol_set(priv, rule_index[1], port_num, ret = bcm_sf2_cfp_act_pol_set(priv, rule_index[1], port_num,
queue_num, true); queue_num, true);
if (ret) if (ret)
goto out_err; goto out_err_flow_rule;
/* Turn on CFP for this rule now */ /* Turn on CFP for this rule now */
reg = core_readl(priv, CORE_CFP_CTL_REG); reg = core_readl(priv, CORE_CFP_CTL_REG);
@ -784,6 +814,8 @@ static int bcm_sf2_cfp_ipv6_rule_set(struct bcm_sf2_priv *priv, int port,
return ret; return ret;
out_err_flow_rule:
ethtool_rx_flow_rule_destroy(flow);
out_err: out_err:
clear_bit(rule_index[1], priv->cfp.used); clear_bit(rule_index[1], priv->cfp.used);
return ret; return ret;