mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-25 05:04:09 +08:00
netfilter: nft_set_pipapo: remove scratch_aligned pointer
use ->scratch for both avx2 and the generic implementation.
After previous change the scratch->map member is always aligned properly
for AVX2, so we can just use scratch->map in AVX2 too.
The alignoff delta is stored in the scratchpad so we can reconstruct
the correct address to free the area again.
Fixes: 7400b06396
("nft_set_pipapo: Introduce AVX2-based lookup implementation")
Reviewed-by: Stefano Brivio <sbrivio@redhat.com>
Signed-off-by: Florian Westphal <fw@strlen.de>
Signed-off-by: Pablo Neira Ayuso <pablo@netfilter.org>
This commit is contained in:
parent
47b1c03c3c
commit
5a8cdf6fd8
@ -1125,6 +1125,7 @@ static void pipapo_free_scratch(const struct nft_pipapo_match *m, unsigned int c
|
||||
return;
|
||||
|
||||
mem = s;
|
||||
mem -= s->align_off;
|
||||
kfree(mem);
|
||||
}
|
||||
|
||||
@ -1144,6 +1145,7 @@ static int pipapo_realloc_scratch(struct nft_pipapo_match *clone,
|
||||
struct nft_pipapo_scratch *scratch;
|
||||
#ifdef NFT_PIPAPO_ALIGN
|
||||
void *scratch_aligned;
|
||||
u32 align_off;
|
||||
#endif
|
||||
scratch = kzalloc_node(struct_size(scratch, map,
|
||||
bsize_max * 2) +
|
||||
@ -1162,8 +1164,6 @@ static int pipapo_realloc_scratch(struct nft_pipapo_match *clone,
|
||||
|
||||
pipapo_free_scratch(clone, i);
|
||||
|
||||
*per_cpu_ptr(clone->scratch, i) = scratch;
|
||||
|
||||
#ifdef NFT_PIPAPO_ALIGN
|
||||
/* Align &scratch->map (not the struct itself): the extra
|
||||
* %NFT_PIPAPO_ALIGN_HEADROOM bytes passed to kzalloc_node()
|
||||
@ -1175,8 +1175,12 @@ static int pipapo_realloc_scratch(struct nft_pipapo_match *clone,
|
||||
|
||||
scratch_aligned = NFT_PIPAPO_LT_ALIGN(&scratch->map);
|
||||
scratch_aligned -= offsetof(struct nft_pipapo_scratch, map);
|
||||
*per_cpu_ptr(clone->scratch_aligned, i) = scratch_aligned;
|
||||
align_off = scratch_aligned - (void *)scratch;
|
||||
|
||||
scratch = scratch_aligned;
|
||||
scratch->align_off = align_off;
|
||||
#endif
|
||||
*per_cpu_ptr(clone->scratch, i) = scratch;
|
||||
}
|
||||
|
||||
return 0;
|
||||
@ -1331,11 +1335,6 @@ static struct nft_pipapo_match *pipapo_clone(struct nft_pipapo_match *old)
|
||||
if (!new->scratch)
|
||||
goto out_scratch;
|
||||
|
||||
#ifdef NFT_PIPAPO_ALIGN
|
||||
new->scratch_aligned = alloc_percpu(*new->scratch_aligned);
|
||||
if (!new->scratch_aligned)
|
||||
goto out_scratch;
|
||||
#endif
|
||||
for_each_possible_cpu(i)
|
||||
*per_cpu_ptr(new->scratch, i) = NULL;
|
||||
|
||||
@ -1388,9 +1387,6 @@ out_lt:
|
||||
out_scratch_realloc:
|
||||
for_each_possible_cpu(i)
|
||||
pipapo_free_scratch(new, i);
|
||||
#ifdef NFT_PIPAPO_ALIGN
|
||||
free_percpu(new->scratch_aligned);
|
||||
#endif
|
||||
out_scratch:
|
||||
free_percpu(new->scratch);
|
||||
kfree(new);
|
||||
@ -1673,11 +1669,7 @@ static void pipapo_free_match(struct nft_pipapo_match *m)
|
||||
for_each_possible_cpu(i)
|
||||
pipapo_free_scratch(m, i);
|
||||
|
||||
#ifdef NFT_PIPAPO_ALIGN
|
||||
free_percpu(m->scratch_aligned);
|
||||
#endif
|
||||
free_percpu(m->scratch);
|
||||
|
||||
pipapo_free_fields(m);
|
||||
|
||||
kfree(m);
|
||||
@ -2171,16 +2163,6 @@ static int nft_pipapo_init(const struct nft_set *set,
|
||||
for_each_possible_cpu(i)
|
||||
*per_cpu_ptr(m->scratch, i) = NULL;
|
||||
|
||||
#ifdef NFT_PIPAPO_ALIGN
|
||||
m->scratch_aligned = alloc_percpu(struct nft_pipapo_scratch *);
|
||||
if (!m->scratch_aligned) {
|
||||
err = -ENOMEM;
|
||||
goto out_free;
|
||||
}
|
||||
for_each_possible_cpu(i)
|
||||
*per_cpu_ptr(m->scratch_aligned, i) = NULL;
|
||||
#endif
|
||||
|
||||
rcu_head_init(&m->rcu);
|
||||
|
||||
nft_pipapo_for_each_field(f, i, m) {
|
||||
@ -2211,9 +2193,6 @@ static int nft_pipapo_init(const struct nft_set *set,
|
||||
return 0;
|
||||
|
||||
out_free:
|
||||
#ifdef NFT_PIPAPO_ALIGN
|
||||
free_percpu(m->scratch_aligned);
|
||||
#endif
|
||||
free_percpu(m->scratch);
|
||||
out_scratch:
|
||||
kfree(m);
|
||||
@ -2267,9 +2246,6 @@ static void nft_pipapo_destroy(const struct nft_ctx *ctx,
|
||||
|
||||
nft_set_pipapo_match_destroy(ctx, set, m);
|
||||
|
||||
#ifdef NFT_PIPAPO_ALIGN
|
||||
free_percpu(m->scratch_aligned);
|
||||
#endif
|
||||
for_each_possible_cpu(cpu)
|
||||
pipapo_free_scratch(m, cpu);
|
||||
free_percpu(m->scratch);
|
||||
@ -2284,9 +2260,6 @@ static void nft_pipapo_destroy(const struct nft_ctx *ctx,
|
||||
if (priv->dirty)
|
||||
nft_set_pipapo_match_destroy(ctx, set, m);
|
||||
|
||||
#ifdef NFT_PIPAPO_ALIGN
|
||||
free_percpu(priv->clone->scratch_aligned);
|
||||
#endif
|
||||
for_each_possible_cpu(cpu)
|
||||
pipapo_free_scratch(priv->clone, cpu);
|
||||
free_percpu(priv->clone->scratch);
|
||||
|
@ -133,10 +133,12 @@ struct nft_pipapo_field {
|
||||
/**
|
||||
* struct nft_pipapo_scratch - percpu data used for lookup and matching
|
||||
* @map_index: Current working bitmap index, toggled between field matches
|
||||
* @align_off: Offset to get the originally allocated address
|
||||
* @map: store partial matching results during lookup
|
||||
*/
|
||||
struct nft_pipapo_scratch {
|
||||
u8 map_index;
|
||||
u32 align_off;
|
||||
unsigned long map[];
|
||||
};
|
||||
|
||||
@ -144,16 +146,12 @@ struct nft_pipapo_scratch {
|
||||
* struct nft_pipapo_match - Data used for lookup and matching
|
||||
* @field_count Amount of fields in set
|
||||
* @scratch: Preallocated per-CPU maps for partial matching results
|
||||
* @scratch_aligned: Version of @scratch aligned to NFT_PIPAPO_ALIGN bytes
|
||||
* @bsize_max: Maximum lookup table bucket size of all fields, in longs
|
||||
* @rcu Matching data is swapped on commits
|
||||
* @f: Fields, with lookup and mapping tables
|
||||
*/
|
||||
struct nft_pipapo_match {
|
||||
int field_count;
|
||||
#ifdef NFT_PIPAPO_ALIGN
|
||||
struct nft_pipapo_scratch * __percpu *scratch_aligned;
|
||||
#endif
|
||||
struct nft_pipapo_scratch * __percpu *scratch;
|
||||
size_t bsize_max;
|
||||
struct rcu_head rcu;
|
||||
|
@ -1139,7 +1139,7 @@ bool nft_pipapo_avx2_lookup(const struct net *net, const struct nft_set *set,
|
||||
*/
|
||||
kernel_fpu_begin_mask(0);
|
||||
|
||||
scratch = *raw_cpu_ptr(m->scratch_aligned);
|
||||
scratch = *raw_cpu_ptr(m->scratch);
|
||||
if (unlikely(!scratch)) {
|
||||
kernel_fpu_end();
|
||||
return false;
|
||||
|
Loading…
Reference in New Issue
Block a user