mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-11 12:28:41 +08:00
d11e631194
These helpers aim to help drivers, with checking for the presence of unsupported control flags. For drivers supporting at least one control flag: flow_rule_is_supp_control_flags() For drivers using flow_rule_match_control(), but not using flags: flow_rule_has_control_flags() For drivers not using flow_rule_match_control(): flow_rule_match_has_control_flags() While primarily aimed at FLOW_DISSECTOR_KEY_CONTROL and flow_rule_match_control(), then the first two can also be used with FLOW_DISSECTOR_KEY_ENC_CONTROL and flow_rule_match_enc_control(). These helpers mirrors the existing check done in sfc: drivers/net/ethernet/sfc/tc.c +276 Only compile-tested. Signed-off-by: Asbjørn Sloth Tønnesen <ast@fiberby.net> Reviewed-by: Louis Peens <louis.peens@corigine.com> Signed-off-by: David S. Miller <davem@davemloft.net>
710 lines
19 KiB
C
710 lines
19 KiB
C
#ifndef _NET_FLOW_OFFLOAD_H
|
|
#define _NET_FLOW_OFFLOAD_H
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/list.h>
|
|
#include <linux/netlink.h>
|
|
#include <net/flow_dissector.h>
|
|
|
|
struct flow_match {
|
|
struct flow_dissector *dissector;
|
|
void *mask;
|
|
void *key;
|
|
};
|
|
|
|
struct flow_match_meta {
|
|
struct flow_dissector_key_meta *key, *mask;
|
|
};
|
|
|
|
struct flow_match_basic {
|
|
struct flow_dissector_key_basic *key, *mask;
|
|
};
|
|
|
|
struct flow_match_control {
|
|
struct flow_dissector_key_control *key, *mask;
|
|
};
|
|
|
|
struct flow_match_eth_addrs {
|
|
struct flow_dissector_key_eth_addrs *key, *mask;
|
|
};
|
|
|
|
struct flow_match_vlan {
|
|
struct flow_dissector_key_vlan *key, *mask;
|
|
};
|
|
|
|
struct flow_match_arp {
|
|
struct flow_dissector_key_arp *key, *mask;
|
|
};
|
|
|
|
struct flow_match_ipv4_addrs {
|
|
struct flow_dissector_key_ipv4_addrs *key, *mask;
|
|
};
|
|
|
|
struct flow_match_ipv6_addrs {
|
|
struct flow_dissector_key_ipv6_addrs *key, *mask;
|
|
};
|
|
|
|
struct flow_match_ip {
|
|
struct flow_dissector_key_ip *key, *mask;
|
|
};
|
|
|
|
struct flow_match_ports {
|
|
struct flow_dissector_key_ports *key, *mask;
|
|
};
|
|
|
|
struct flow_match_ports_range {
|
|
struct flow_dissector_key_ports_range *key, *mask;
|
|
};
|
|
|
|
struct flow_match_icmp {
|
|
struct flow_dissector_key_icmp *key, *mask;
|
|
};
|
|
|
|
struct flow_match_tcp {
|
|
struct flow_dissector_key_tcp *key, *mask;
|
|
};
|
|
|
|
struct flow_match_ipsec {
|
|
struct flow_dissector_key_ipsec *key, *mask;
|
|
};
|
|
|
|
struct flow_match_mpls {
|
|
struct flow_dissector_key_mpls *key, *mask;
|
|
};
|
|
|
|
struct flow_match_enc_keyid {
|
|
struct flow_dissector_key_keyid *key, *mask;
|
|
};
|
|
|
|
struct flow_match_enc_opts {
|
|
struct flow_dissector_key_enc_opts *key, *mask;
|
|
};
|
|
|
|
struct flow_match_ct {
|
|
struct flow_dissector_key_ct *key, *mask;
|
|
};
|
|
|
|
struct flow_match_pppoe {
|
|
struct flow_dissector_key_pppoe *key, *mask;
|
|
};
|
|
|
|
struct flow_match_l2tpv3 {
|
|
struct flow_dissector_key_l2tpv3 *key, *mask;
|
|
};
|
|
|
|
struct flow_rule;
|
|
|
|
void flow_rule_match_meta(const struct flow_rule *rule,
|
|
struct flow_match_meta *out);
|
|
void flow_rule_match_basic(const struct flow_rule *rule,
|
|
struct flow_match_basic *out);
|
|
void flow_rule_match_control(const struct flow_rule *rule,
|
|
struct flow_match_control *out);
|
|
void flow_rule_match_eth_addrs(const struct flow_rule *rule,
|
|
struct flow_match_eth_addrs *out);
|
|
void flow_rule_match_vlan(const struct flow_rule *rule,
|
|
struct flow_match_vlan *out);
|
|
void flow_rule_match_cvlan(const struct flow_rule *rule,
|
|
struct flow_match_vlan *out);
|
|
void flow_rule_match_arp(const struct flow_rule *rule,
|
|
struct flow_match_arp *out);
|
|
void flow_rule_match_ipv4_addrs(const struct flow_rule *rule,
|
|
struct flow_match_ipv4_addrs *out);
|
|
void flow_rule_match_ipv6_addrs(const struct flow_rule *rule,
|
|
struct flow_match_ipv6_addrs *out);
|
|
void flow_rule_match_ip(const struct flow_rule *rule,
|
|
struct flow_match_ip *out);
|
|
void flow_rule_match_ports(const struct flow_rule *rule,
|
|
struct flow_match_ports *out);
|
|
void flow_rule_match_ports_range(const struct flow_rule *rule,
|
|
struct flow_match_ports_range *out);
|
|
void flow_rule_match_tcp(const struct flow_rule *rule,
|
|
struct flow_match_tcp *out);
|
|
void flow_rule_match_ipsec(const struct flow_rule *rule,
|
|
struct flow_match_ipsec *out);
|
|
void flow_rule_match_icmp(const struct flow_rule *rule,
|
|
struct flow_match_icmp *out);
|
|
void flow_rule_match_mpls(const struct flow_rule *rule,
|
|
struct flow_match_mpls *out);
|
|
void flow_rule_match_enc_control(const struct flow_rule *rule,
|
|
struct flow_match_control *out);
|
|
void flow_rule_match_enc_ipv4_addrs(const struct flow_rule *rule,
|
|
struct flow_match_ipv4_addrs *out);
|
|
void flow_rule_match_enc_ipv6_addrs(const struct flow_rule *rule,
|
|
struct flow_match_ipv6_addrs *out);
|
|
void flow_rule_match_enc_ip(const struct flow_rule *rule,
|
|
struct flow_match_ip *out);
|
|
void flow_rule_match_enc_ports(const struct flow_rule *rule,
|
|
struct flow_match_ports *out);
|
|
void flow_rule_match_enc_keyid(const struct flow_rule *rule,
|
|
struct flow_match_enc_keyid *out);
|
|
void flow_rule_match_enc_opts(const struct flow_rule *rule,
|
|
struct flow_match_enc_opts *out);
|
|
void flow_rule_match_ct(const struct flow_rule *rule,
|
|
struct flow_match_ct *out);
|
|
void flow_rule_match_pppoe(const struct flow_rule *rule,
|
|
struct flow_match_pppoe *out);
|
|
void flow_rule_match_l2tpv3(const struct flow_rule *rule,
|
|
struct flow_match_l2tpv3 *out);
|
|
|
|
enum flow_action_id {
|
|
FLOW_ACTION_ACCEPT = 0,
|
|
FLOW_ACTION_DROP,
|
|
FLOW_ACTION_TRAP,
|
|
FLOW_ACTION_GOTO,
|
|
FLOW_ACTION_REDIRECT,
|
|
FLOW_ACTION_MIRRED,
|
|
FLOW_ACTION_REDIRECT_INGRESS,
|
|
FLOW_ACTION_MIRRED_INGRESS,
|
|
FLOW_ACTION_VLAN_PUSH,
|
|
FLOW_ACTION_VLAN_POP,
|
|
FLOW_ACTION_VLAN_MANGLE,
|
|
FLOW_ACTION_TUNNEL_ENCAP,
|
|
FLOW_ACTION_TUNNEL_DECAP,
|
|
FLOW_ACTION_MANGLE,
|
|
FLOW_ACTION_ADD,
|
|
FLOW_ACTION_CSUM,
|
|
FLOW_ACTION_MARK,
|
|
FLOW_ACTION_PTYPE,
|
|
FLOW_ACTION_PRIORITY,
|
|
FLOW_ACTION_RX_QUEUE_MAPPING,
|
|
FLOW_ACTION_WAKE,
|
|
FLOW_ACTION_QUEUE,
|
|
FLOW_ACTION_SAMPLE,
|
|
FLOW_ACTION_POLICE,
|
|
FLOW_ACTION_CT,
|
|
FLOW_ACTION_CT_METADATA,
|
|
FLOW_ACTION_MPLS_PUSH,
|
|
FLOW_ACTION_MPLS_POP,
|
|
FLOW_ACTION_MPLS_MANGLE,
|
|
FLOW_ACTION_GATE,
|
|
FLOW_ACTION_PPPOE_PUSH,
|
|
FLOW_ACTION_JUMP,
|
|
FLOW_ACTION_PIPE,
|
|
FLOW_ACTION_VLAN_PUSH_ETH,
|
|
FLOW_ACTION_VLAN_POP_ETH,
|
|
FLOW_ACTION_CONTINUE,
|
|
NUM_FLOW_ACTIONS,
|
|
};
|
|
|
|
/* This is mirroring enum pedit_header_type definition for easy mapping between
|
|
* tc pedit action. Legacy TCA_PEDIT_KEY_EX_HDR_TYPE_NETWORK is mapped to
|
|
* FLOW_ACT_MANGLE_UNSPEC, which is supported by no driver.
|
|
*/
|
|
enum flow_action_mangle_base {
|
|
FLOW_ACT_MANGLE_UNSPEC = 0,
|
|
FLOW_ACT_MANGLE_HDR_TYPE_ETH,
|
|
FLOW_ACT_MANGLE_HDR_TYPE_IP4,
|
|
FLOW_ACT_MANGLE_HDR_TYPE_IP6,
|
|
FLOW_ACT_MANGLE_HDR_TYPE_TCP,
|
|
FLOW_ACT_MANGLE_HDR_TYPE_UDP,
|
|
};
|
|
|
|
enum flow_action_hw_stats_bit {
|
|
FLOW_ACTION_HW_STATS_IMMEDIATE_BIT,
|
|
FLOW_ACTION_HW_STATS_DELAYED_BIT,
|
|
FLOW_ACTION_HW_STATS_DISABLED_BIT,
|
|
|
|
FLOW_ACTION_HW_STATS_NUM_BITS
|
|
};
|
|
|
|
enum flow_action_hw_stats {
|
|
FLOW_ACTION_HW_STATS_IMMEDIATE =
|
|
BIT(FLOW_ACTION_HW_STATS_IMMEDIATE_BIT),
|
|
FLOW_ACTION_HW_STATS_DELAYED = BIT(FLOW_ACTION_HW_STATS_DELAYED_BIT),
|
|
FLOW_ACTION_HW_STATS_ANY = FLOW_ACTION_HW_STATS_IMMEDIATE |
|
|
FLOW_ACTION_HW_STATS_DELAYED,
|
|
FLOW_ACTION_HW_STATS_DISABLED =
|
|
BIT(FLOW_ACTION_HW_STATS_DISABLED_BIT),
|
|
FLOW_ACTION_HW_STATS_DONT_CARE = BIT(FLOW_ACTION_HW_STATS_NUM_BITS) - 1,
|
|
};
|
|
|
|
typedef void (*action_destr)(void *priv);
|
|
|
|
struct flow_action_cookie {
|
|
u32 cookie_len;
|
|
u8 cookie[];
|
|
};
|
|
|
|
struct flow_action_cookie *flow_action_cookie_create(void *data,
|
|
unsigned int len,
|
|
gfp_t gfp);
|
|
void flow_action_cookie_destroy(struct flow_action_cookie *cookie);
|
|
|
|
struct flow_action_entry {
|
|
enum flow_action_id id;
|
|
u32 hw_index;
|
|
unsigned long cookie;
|
|
u64 miss_cookie;
|
|
enum flow_action_hw_stats hw_stats;
|
|
action_destr destructor;
|
|
void *destructor_priv;
|
|
union {
|
|
u32 chain_index; /* FLOW_ACTION_GOTO */
|
|
struct net_device *dev; /* FLOW_ACTION_REDIRECT */
|
|
struct { /* FLOW_ACTION_VLAN */
|
|
u16 vid;
|
|
__be16 proto;
|
|
u8 prio;
|
|
} vlan;
|
|
struct { /* FLOW_ACTION_VLAN_PUSH_ETH */
|
|
unsigned char dst[ETH_ALEN];
|
|
unsigned char src[ETH_ALEN];
|
|
} vlan_push_eth;
|
|
struct { /* FLOW_ACTION_MANGLE */
|
|
/* FLOW_ACTION_ADD */
|
|
enum flow_action_mangle_base htype;
|
|
u32 offset;
|
|
u32 mask;
|
|
u32 val;
|
|
} mangle;
|
|
struct ip_tunnel_info *tunnel; /* FLOW_ACTION_TUNNEL_ENCAP */
|
|
u32 csum_flags; /* FLOW_ACTION_CSUM */
|
|
u32 mark; /* FLOW_ACTION_MARK */
|
|
u16 ptype; /* FLOW_ACTION_PTYPE */
|
|
u16 rx_queue; /* FLOW_ACTION_RX_QUEUE_MAPPING */
|
|
u32 priority; /* FLOW_ACTION_PRIORITY */
|
|
struct { /* FLOW_ACTION_QUEUE */
|
|
u32 ctx;
|
|
u32 index;
|
|
u8 vf;
|
|
} queue;
|
|
struct { /* FLOW_ACTION_SAMPLE */
|
|
struct psample_group *psample_group;
|
|
u32 rate;
|
|
u32 trunc_size;
|
|
bool truncate;
|
|
} sample;
|
|
struct { /* FLOW_ACTION_POLICE */
|
|
u32 burst;
|
|
u64 rate_bytes_ps;
|
|
u64 peakrate_bytes_ps;
|
|
u32 avrate;
|
|
u16 overhead;
|
|
u64 burst_pkt;
|
|
u64 rate_pkt_ps;
|
|
u32 mtu;
|
|
struct {
|
|
enum flow_action_id act_id;
|
|
u32 extval;
|
|
} exceed, notexceed;
|
|
} police;
|
|
struct { /* FLOW_ACTION_CT */
|
|
int action;
|
|
u16 zone;
|
|
struct nf_flowtable *flow_table;
|
|
} ct;
|
|
struct {
|
|
unsigned long cookie;
|
|
u32 mark;
|
|
u32 labels[4];
|
|
bool orig_dir;
|
|
} ct_metadata;
|
|
struct { /* FLOW_ACTION_MPLS_PUSH */
|
|
u32 label;
|
|
__be16 proto;
|
|
u8 tc;
|
|
u8 bos;
|
|
u8 ttl;
|
|
} mpls_push;
|
|
struct { /* FLOW_ACTION_MPLS_POP */
|
|
__be16 proto;
|
|
} mpls_pop;
|
|
struct { /* FLOW_ACTION_MPLS_MANGLE */
|
|
u32 label;
|
|
u8 tc;
|
|
u8 bos;
|
|
u8 ttl;
|
|
} mpls_mangle;
|
|
struct {
|
|
s32 prio;
|
|
u64 basetime;
|
|
u64 cycletime;
|
|
u64 cycletimeext;
|
|
u32 num_entries;
|
|
struct action_gate_entry *entries;
|
|
} gate;
|
|
struct { /* FLOW_ACTION_PPPOE_PUSH */
|
|
u16 sid;
|
|
} pppoe;
|
|
};
|
|
struct flow_action_cookie *user_cookie; /* user defined action cookie */
|
|
};
|
|
|
|
struct flow_action {
|
|
unsigned int num_entries;
|
|
struct flow_action_entry entries[] __counted_by(num_entries);
|
|
};
|
|
|
|
static inline bool flow_action_has_entries(const struct flow_action *action)
|
|
{
|
|
return action->num_entries;
|
|
}
|
|
|
|
/**
|
|
* flow_offload_has_one_action() - check if exactly one action is present
|
|
* @action: tc filter flow offload action
|
|
*
|
|
* Return: true if exactly one action is present.
|
|
*/
|
|
static inline bool flow_offload_has_one_action(const struct flow_action *action)
|
|
{
|
|
return action->num_entries == 1;
|
|
}
|
|
|
|
static inline bool flow_action_is_last_entry(const struct flow_action *action,
|
|
const struct flow_action_entry *entry)
|
|
{
|
|
return entry == &action->entries[action->num_entries - 1];
|
|
}
|
|
|
|
#define flow_action_for_each(__i, __act, __actions) \
|
|
for (__i = 0, __act = &(__actions)->entries[0]; \
|
|
__i < (__actions)->num_entries; \
|
|
__act = &(__actions)->entries[++__i])
|
|
|
|
static inline bool
|
|
flow_action_mixed_hw_stats_check(const struct flow_action *action,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
const struct flow_action_entry *action_entry;
|
|
u8 last_hw_stats;
|
|
int i;
|
|
|
|
if (flow_offload_has_one_action(action))
|
|
return true;
|
|
|
|
flow_action_for_each(i, action_entry, action) {
|
|
if (i && action_entry->hw_stats != last_hw_stats) {
|
|
NL_SET_ERR_MSG_MOD(extack, "Mixing HW stats types for actions is not supported");
|
|
return false;
|
|
}
|
|
last_hw_stats = action_entry->hw_stats;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static inline const struct flow_action_entry *
|
|
flow_action_first_entry_get(const struct flow_action *action)
|
|
{
|
|
WARN_ON(!flow_action_has_entries(action));
|
|
return &action->entries[0];
|
|
}
|
|
|
|
static inline bool
|
|
__flow_action_hw_stats_check(const struct flow_action *action,
|
|
struct netlink_ext_ack *extack,
|
|
bool check_allow_bit,
|
|
enum flow_action_hw_stats_bit allow_bit)
|
|
{
|
|
const struct flow_action_entry *action_entry;
|
|
|
|
if (!flow_action_has_entries(action))
|
|
return true;
|
|
if (!flow_action_mixed_hw_stats_check(action, extack))
|
|
return false;
|
|
|
|
action_entry = flow_action_first_entry_get(action);
|
|
|
|
/* Zero is not a legal value for hw_stats, catch anyone passing it */
|
|
WARN_ON_ONCE(!action_entry->hw_stats);
|
|
|
|
if (!check_allow_bit &&
|
|
~action_entry->hw_stats & FLOW_ACTION_HW_STATS_ANY) {
|
|
NL_SET_ERR_MSG_MOD(extack, "Driver supports only default HW stats type \"any\"");
|
|
return false;
|
|
} else if (check_allow_bit &&
|
|
!(action_entry->hw_stats & BIT(allow_bit))) {
|
|
NL_SET_ERR_MSG_MOD(extack, "Driver does not support selected HW stats type");
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static inline bool
|
|
flow_action_hw_stats_check(const struct flow_action *action,
|
|
struct netlink_ext_ack *extack,
|
|
enum flow_action_hw_stats_bit allow_bit)
|
|
{
|
|
return __flow_action_hw_stats_check(action, extack, true, allow_bit);
|
|
}
|
|
|
|
static inline bool
|
|
flow_action_basic_hw_stats_check(const struct flow_action *action,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
return __flow_action_hw_stats_check(action, extack, false, 0);
|
|
}
|
|
|
|
struct flow_rule {
|
|
struct flow_match match;
|
|
struct flow_action action;
|
|
};
|
|
|
|
struct flow_rule *flow_rule_alloc(unsigned int num_actions);
|
|
|
|
static inline bool flow_rule_match_key(const struct flow_rule *rule,
|
|
enum flow_dissector_key_id key)
|
|
{
|
|
return dissector_uses_key(rule->match.dissector, key);
|
|
}
|
|
|
|
/**
|
|
* flow_rule_is_supp_control_flags() - check for supported control flags
|
|
* @supp_flags: control flags supported by driver
|
|
* @ctrl_flags: control flags present in rule
|
|
* @extack: The netlink extended ACK for reporting errors.
|
|
*
|
|
* Return: true if only supported control flags are set, false otherwise.
|
|
*/
|
|
static inline bool flow_rule_is_supp_control_flags(const u32 supp_flags,
|
|
const u32 ctrl_flags,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
if (likely((ctrl_flags & ~supp_flags) == 0))
|
|
return true;
|
|
|
|
NL_SET_ERR_MSG_FMT_MOD(extack,
|
|
"Unsupported match on control.flags %#x",
|
|
ctrl_flags);
|
|
|
|
return false;
|
|
}
|
|
|
|
/**
|
|
* flow_rule_has_control_flags() - check for presence of any control flags
|
|
* @ctrl_flags: control flags present in rule
|
|
* @extack: The netlink extended ACK for reporting errors.
|
|
*
|
|
* Return: true if control flags are set, false otherwise.
|
|
*/
|
|
static inline bool flow_rule_has_control_flags(const u32 ctrl_flags,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
return !flow_rule_is_supp_control_flags(0, ctrl_flags, extack);
|
|
}
|
|
|
|
/**
|
|
* flow_rule_match_has_control_flags() - match and check for any control flags
|
|
* @rule: The flow_rule under evaluation.
|
|
* @extack: The netlink extended ACK for reporting errors.
|
|
*
|
|
* Return: true if control flags are set, false otherwise.
|
|
*/
|
|
static inline bool flow_rule_match_has_control_flags(struct flow_rule *rule,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
struct flow_match_control match;
|
|
|
|
if (!flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_CONTROL))
|
|
return false;
|
|
|
|
flow_rule_match_control(rule, &match);
|
|
|
|
return flow_rule_has_control_flags(match.mask->flags, extack);
|
|
}
|
|
|
|
struct flow_stats {
|
|
u64 pkts;
|
|
u64 bytes;
|
|
u64 drops;
|
|
u64 lastused;
|
|
enum flow_action_hw_stats used_hw_stats;
|
|
bool used_hw_stats_valid;
|
|
};
|
|
|
|
static inline void flow_stats_update(struct flow_stats *flow_stats,
|
|
u64 bytes, u64 pkts,
|
|
u64 drops, u64 lastused,
|
|
enum flow_action_hw_stats used_hw_stats)
|
|
{
|
|
flow_stats->pkts += pkts;
|
|
flow_stats->bytes += bytes;
|
|
flow_stats->drops += drops;
|
|
flow_stats->lastused = max_t(u64, flow_stats->lastused, lastused);
|
|
|
|
/* The driver should pass value with a maximum of one bit set.
|
|
* Passing FLOW_ACTION_HW_STATS_ANY is invalid.
|
|
*/
|
|
WARN_ON(used_hw_stats == FLOW_ACTION_HW_STATS_ANY);
|
|
flow_stats->used_hw_stats |= used_hw_stats;
|
|
flow_stats->used_hw_stats_valid = true;
|
|
}
|
|
|
|
enum flow_block_command {
|
|
FLOW_BLOCK_BIND,
|
|
FLOW_BLOCK_UNBIND,
|
|
};
|
|
|
|
enum flow_block_binder_type {
|
|
FLOW_BLOCK_BINDER_TYPE_UNSPEC,
|
|
FLOW_BLOCK_BINDER_TYPE_CLSACT_INGRESS,
|
|
FLOW_BLOCK_BINDER_TYPE_CLSACT_EGRESS,
|
|
FLOW_BLOCK_BINDER_TYPE_RED_EARLY_DROP,
|
|
FLOW_BLOCK_BINDER_TYPE_RED_MARK,
|
|
};
|
|
|
|
struct flow_block {
|
|
struct list_head cb_list;
|
|
};
|
|
|
|
struct netlink_ext_ack;
|
|
|
|
struct flow_block_offload {
|
|
enum flow_block_command command;
|
|
enum flow_block_binder_type binder_type;
|
|
bool block_shared;
|
|
bool unlocked_driver_cb;
|
|
struct net *net;
|
|
struct flow_block *block;
|
|
struct list_head cb_list;
|
|
struct list_head *driver_block_list;
|
|
struct netlink_ext_ack *extack;
|
|
struct Qdisc *sch;
|
|
struct list_head *cb_list_head;
|
|
};
|
|
|
|
enum tc_setup_type;
|
|
typedef int flow_setup_cb_t(enum tc_setup_type type, void *type_data,
|
|
void *cb_priv);
|
|
|
|
struct flow_block_cb;
|
|
|
|
struct flow_block_indr {
|
|
struct list_head list;
|
|
struct net_device *dev;
|
|
struct Qdisc *sch;
|
|
enum flow_block_binder_type binder_type;
|
|
void *data;
|
|
void *cb_priv;
|
|
void (*cleanup)(struct flow_block_cb *block_cb);
|
|
};
|
|
|
|
struct flow_block_cb {
|
|
struct list_head driver_list;
|
|
struct list_head list;
|
|
flow_setup_cb_t *cb;
|
|
void *cb_ident;
|
|
void *cb_priv;
|
|
void (*release)(void *cb_priv);
|
|
struct flow_block_indr indr;
|
|
unsigned int refcnt;
|
|
};
|
|
|
|
struct flow_block_cb *flow_block_cb_alloc(flow_setup_cb_t *cb,
|
|
void *cb_ident, void *cb_priv,
|
|
void (*release)(void *cb_priv));
|
|
struct flow_block_cb *flow_indr_block_cb_alloc(flow_setup_cb_t *cb,
|
|
void *cb_ident, void *cb_priv,
|
|
void (*release)(void *cb_priv),
|
|
struct flow_block_offload *bo,
|
|
struct net_device *dev,
|
|
struct Qdisc *sch, void *data,
|
|
void *indr_cb_priv,
|
|
void (*cleanup)(struct flow_block_cb *block_cb));
|
|
void flow_block_cb_free(struct flow_block_cb *block_cb);
|
|
|
|
struct flow_block_cb *flow_block_cb_lookup(struct flow_block *block,
|
|
flow_setup_cb_t *cb, void *cb_ident);
|
|
|
|
void *flow_block_cb_priv(struct flow_block_cb *block_cb);
|
|
void flow_block_cb_incref(struct flow_block_cb *block_cb);
|
|
unsigned int flow_block_cb_decref(struct flow_block_cb *block_cb);
|
|
|
|
static inline void flow_block_cb_add(struct flow_block_cb *block_cb,
|
|
struct flow_block_offload *offload)
|
|
{
|
|
list_add_tail(&block_cb->list, &offload->cb_list);
|
|
}
|
|
|
|
static inline void flow_block_cb_remove(struct flow_block_cb *block_cb,
|
|
struct flow_block_offload *offload)
|
|
{
|
|
list_move(&block_cb->list, &offload->cb_list);
|
|
}
|
|
|
|
static inline void flow_indr_block_cb_remove(struct flow_block_cb *block_cb,
|
|
struct flow_block_offload *offload)
|
|
{
|
|
list_del(&block_cb->indr.list);
|
|
list_move(&block_cb->list, &offload->cb_list);
|
|
}
|
|
|
|
bool flow_block_cb_is_busy(flow_setup_cb_t *cb, void *cb_ident,
|
|
struct list_head *driver_block_list);
|
|
|
|
int flow_block_cb_setup_simple(struct flow_block_offload *f,
|
|
struct list_head *driver_list,
|
|
flow_setup_cb_t *cb,
|
|
void *cb_ident, void *cb_priv, bool ingress_only);
|
|
|
|
enum flow_cls_command {
|
|
FLOW_CLS_REPLACE,
|
|
FLOW_CLS_DESTROY,
|
|
FLOW_CLS_STATS,
|
|
FLOW_CLS_TMPLT_CREATE,
|
|
FLOW_CLS_TMPLT_DESTROY,
|
|
};
|
|
|
|
struct flow_cls_common_offload {
|
|
u32 chain_index;
|
|
__be16 protocol;
|
|
u32 prio;
|
|
struct netlink_ext_ack *extack;
|
|
};
|
|
|
|
struct flow_cls_offload {
|
|
struct flow_cls_common_offload common;
|
|
enum flow_cls_command command;
|
|
bool use_act_stats;
|
|
unsigned long cookie;
|
|
struct flow_rule *rule;
|
|
struct flow_stats stats;
|
|
u32 classid;
|
|
};
|
|
|
|
enum offload_act_command {
|
|
FLOW_ACT_REPLACE,
|
|
FLOW_ACT_DESTROY,
|
|
FLOW_ACT_STATS,
|
|
};
|
|
|
|
struct flow_offload_action {
|
|
struct netlink_ext_ack *extack; /* NULL in FLOW_ACT_STATS process*/
|
|
enum offload_act_command command;
|
|
enum flow_action_id id;
|
|
u32 index;
|
|
unsigned long cookie;
|
|
struct flow_stats stats;
|
|
struct flow_action action;
|
|
};
|
|
|
|
struct flow_offload_action *offload_action_alloc(unsigned int num_actions);
|
|
|
|
static inline struct flow_rule *
|
|
flow_cls_offload_flow_rule(struct flow_cls_offload *flow_cmd)
|
|
{
|
|
return flow_cmd->rule;
|
|
}
|
|
|
|
static inline void flow_block_init(struct flow_block *flow_block)
|
|
{
|
|
INIT_LIST_HEAD(&flow_block->cb_list);
|
|
}
|
|
|
|
typedef int flow_indr_block_bind_cb_t(struct net_device *dev, struct Qdisc *sch, void *cb_priv,
|
|
enum tc_setup_type type, void *type_data,
|
|
void *data,
|
|
void (*cleanup)(struct flow_block_cb *block_cb));
|
|
|
|
int flow_indr_dev_register(flow_indr_block_bind_cb_t *cb, void *cb_priv);
|
|
void flow_indr_dev_unregister(flow_indr_block_bind_cb_t *cb, void *cb_priv,
|
|
void (*release)(void *cb_priv));
|
|
int flow_indr_dev_setup_offload(struct net_device *dev, struct Qdisc *sch,
|
|
enum tc_setup_type type, void *data,
|
|
struct flow_block_offload *bo,
|
|
void (*cleanup)(struct flow_block_cb *block_cb));
|
|
bool flow_indr_dev_exists(void);
|
|
|
|
#endif /* _NET_FLOW_OFFLOAD_H */
|