mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-27 22:24:11 +08:00
net/sched: act_ct: offload UDP NEW connections
Modify the offload algorithm of UDP connections to the following: - Offload NEW connection as unidirectional. - When connection state changes to ESTABLISHED also update the hardware flow. However, in order to prevent act_ct from spamming offload add wq for every packet coming in reply direction in this state verify whether connection has already been updated to ESTABLISHED in the drivers. If that it the case, then skip flow_table and let conntrack handle such packets which will also allow conntrack to potentially promote the connection to ASSURED. - When connection state changes to ASSURED set the flow_table flow NF_FLOW_HW_BIDIRECTIONAL flag which will cause refresh mechanism to offload the reply direction. All other protocols have their offload algorithm preserved and are always offloaded as bidirectional. Note that this change tries to minimize the load on flow_table add workqueue. First, it tracks the last ctinfo that was offloaded by using new flow 'NF_FLOW_HW_ESTABLISHED' flag and doesn't schedule the refresh for reply direction packets when the offloads have already been updated with current ctinfo. Second, when 'add' task executes on workqueue it always update the offload with current flow state (by checking 'bidirectional' flow flag and obtaining actual ctinfo/cookie through meta action instead of caching any of these from the moment of scheduling the 'add' work) preventing the need from scheduling more updates if state changed concurrently while the 'add' work was pending on workqueue. Signed-off-by: Vlad Buslov <vladbu@nvidia.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
d5774cb6c5
commit
6a9bad0069
@ -369,7 +369,7 @@ static void tcf_ct_flow_tc_ifidx(struct flow_offload *entry,
|
||||
|
||||
static void tcf_ct_flow_table_add(struct tcf_ct_flow_table *ct_ft,
|
||||
struct nf_conn *ct,
|
||||
bool tcp)
|
||||
bool tcp, bool bidirectional)
|
||||
{
|
||||
struct nf_conn_act_ct_ext *act_ct_ext;
|
||||
struct flow_offload *entry;
|
||||
@ -388,6 +388,8 @@ static void tcf_ct_flow_table_add(struct tcf_ct_flow_table *ct_ft,
|
||||
ct->proto.tcp.seen[0].flags |= IP_CT_TCP_FLAG_BE_LIBERAL;
|
||||
ct->proto.tcp.seen[1].flags |= IP_CT_TCP_FLAG_BE_LIBERAL;
|
||||
}
|
||||
if (bidirectional)
|
||||
__set_bit(NF_FLOW_HW_BIDIRECTIONAL, &entry->flags);
|
||||
|
||||
act_ct_ext = nf_conn_act_ct_ext_find(ct);
|
||||
if (act_ct_ext) {
|
||||
@ -411,26 +413,34 @@ static void tcf_ct_flow_table_process_conn(struct tcf_ct_flow_table *ct_ft,
|
||||
struct nf_conn *ct,
|
||||
enum ip_conntrack_info ctinfo)
|
||||
{
|
||||
bool tcp = false;
|
||||
|
||||
if ((ctinfo != IP_CT_ESTABLISHED && ctinfo != IP_CT_ESTABLISHED_REPLY) ||
|
||||
!test_bit(IPS_ASSURED_BIT, &ct->status))
|
||||
return;
|
||||
bool tcp = false, bidirectional = true;
|
||||
|
||||
switch (nf_ct_protonum(ct)) {
|
||||
case IPPROTO_TCP:
|
||||
tcp = true;
|
||||
if (ct->proto.tcp.state != TCP_CONNTRACK_ESTABLISHED)
|
||||
if ((ctinfo != IP_CT_ESTABLISHED &&
|
||||
ctinfo != IP_CT_ESTABLISHED_REPLY) ||
|
||||
!test_bit(IPS_ASSURED_BIT, &ct->status) ||
|
||||
ct->proto.tcp.state != TCP_CONNTRACK_ESTABLISHED)
|
||||
return;
|
||||
|
||||
tcp = true;
|
||||
break;
|
||||
case IPPROTO_UDP:
|
||||
if (!nf_ct_is_confirmed(ct))
|
||||
return;
|
||||
if (!test_bit(IPS_ASSURED_BIT, &ct->status))
|
||||
bidirectional = false;
|
||||
break;
|
||||
#ifdef CONFIG_NF_CT_PROTO_GRE
|
||||
case IPPROTO_GRE: {
|
||||
struct nf_conntrack_tuple *tuple;
|
||||
|
||||
if (ct->status & IPS_NAT_MASK)
|
||||
if ((ctinfo != IP_CT_ESTABLISHED &&
|
||||
ctinfo != IP_CT_ESTABLISHED_REPLY) ||
|
||||
!test_bit(IPS_ASSURED_BIT, &ct->status) ||
|
||||
ct->status & IPS_NAT_MASK)
|
||||
return;
|
||||
|
||||
tuple = &ct->tuplehash[IP_CT_DIR_ORIGINAL].tuple;
|
||||
/* No support for GRE v1 */
|
||||
if (tuple->src.u.gre.key || tuple->dst.u.gre.key)
|
||||
@ -446,7 +456,7 @@ static void tcf_ct_flow_table_process_conn(struct tcf_ct_flow_table *ct_ft,
|
||||
ct->status & IPS_SEQ_ADJUST)
|
||||
return;
|
||||
|
||||
tcf_ct_flow_table_add(ct_ft, ct, tcp);
|
||||
tcf_ct_flow_table_add(ct_ft, ct, tcp, bidirectional);
|
||||
}
|
||||
|
||||
static bool
|
||||
@ -625,13 +635,30 @@ static bool tcf_ct_flow_table_lookup(struct tcf_ct_params *p,
|
||||
flow = container_of(tuplehash, struct flow_offload, tuplehash[dir]);
|
||||
ct = flow->ct;
|
||||
|
||||
if (dir == FLOW_OFFLOAD_DIR_REPLY &&
|
||||
!test_bit(NF_FLOW_HW_BIDIRECTIONAL, &flow->flags)) {
|
||||
/* Only offload reply direction after connection became
|
||||
* assured.
|
||||
*/
|
||||
if (test_bit(IPS_ASSURED_BIT, &ct->status))
|
||||
set_bit(NF_FLOW_HW_BIDIRECTIONAL, &flow->flags);
|
||||
else if (test_bit(NF_FLOW_HW_ESTABLISHED, &flow->flags))
|
||||
/* If flow_table flow has already been updated to the
|
||||
* established state, then don't refresh.
|
||||
*/
|
||||
return false;
|
||||
}
|
||||
|
||||
if (tcph && (unlikely(tcph->fin || tcph->rst))) {
|
||||
flow_offload_teardown(flow);
|
||||
return false;
|
||||
}
|
||||
|
||||
ctinfo = dir == FLOW_OFFLOAD_DIR_ORIGINAL ? IP_CT_ESTABLISHED :
|
||||
IP_CT_ESTABLISHED_REPLY;
|
||||
if (dir == FLOW_OFFLOAD_DIR_ORIGINAL)
|
||||
ctinfo = test_bit(IPS_SEEN_REPLY_BIT, &ct->status) ?
|
||||
IP_CT_ESTABLISHED : IP_CT_NEW;
|
||||
else
|
||||
ctinfo = IP_CT_ESTABLISHED_REPLY;
|
||||
|
||||
flow_offload_refresh(nf_ft, flow);
|
||||
nf_conntrack_get(&ct->ct_general);
|
||||
|
Loading…
Reference in New Issue
Block a user