netfilter pr 2024-21-02
-----BEGIN PGP SIGNATURE----- iQJBBAABCAArFiEEgKkgxbID4Gn1hq6fcJGo2a1f9gAFAmXV2OYNHGZ3QHN0cmxl bi5kZQAKCRBwkajZrV/2AHCRD/9sHoOd4QCVVgcDr3SjpaVWikM0Zdkge65At/uY bFENWgcDsSfsH7kAQm+nwzseT+QtTk9OOv9wqWzdEYROD7sqjVK2Zv/CUs24odGj 7Wj35OLYLgUIEMlHF/G9kOuWqW61URXwXcHvoFWkew1WweAVDqi648osLWUP9qkL IFJ5729/1upq9XJc+pMxIy2Oe2zhMc4XNHsy1OCOg4fUQtDM81jgoJz0137ohCIh PW4aaSno8ZeRuFe1RKfya5+suv3WgMui/fOBmpnnhjWVxHRJvYZ926wsy/jC7xRJ E7/TdmymbzijRBEHh+IxQYZkE55XXc0E1Lj1ic653AzUWJ3tQRfD+HWg+GYj/WCu sWy1e7eRJIjYVbeB5m6ao3g47Zq1XIRXo7E2Rvt3E2beM6t9aMIMuuajBHAOEV2O pCfG4zBlEYw1SuuuoqzcXTVLKDf6WZjx1xtUAJCTks8JFTjPEwPwOQhGCv1cc/BC qox7MejeDH/L+ZreeTYnWlQr1GGokNgrmpdDx0G8GBBRUDPoP8D4GTxvNEz44XOO SfL2yl5v82GBBmsFHzC2J8BGN8KC4JyzDGupU+bcdMWCs8tSvMK0KVeankRvpdBl x4VLmdoNo6zvtOYlPOxdphhsd6xA0dFiLMgSr9f5WsIgepaC+Umxp59IfCEH/bfl 1Kcg9g== =GYgG -----END PGP SIGNATURE----- Merge tag 'nf-next-24-02-21' of https://git.kernel.org/pub/scm/linux/kernel/git/netfilter/nf-next Florian Westphal says: ==================== netfilter updates for net-next 1. Prefer KMEM_CACHE() macro to create kmem caches, from Kunwu Chan. Patches 2 and 3 consolidate nf_log NULL checks and introduces extra boundary checks on family and type to make it clear that no out of bounds access will happen. No in-tree user currently passes such values, but thats not clear from looking at the function. From Pablo Neira Ayuso. Patch 4, also from Pablo, gets rid of unneeded conditional in nft_osf init function. Patch 5, from myself, fixes erroneous Kconfig dependencies that came in an earlier net-next pull request. This should get rid of the xtables related build failure reports. Patches 6 to 10 are an update to nftables' concatenated-ranges set type to speed up element insertions. This series also compacts a few data structures and cleans up a few oddities such as reliance on ZERO_SIZE_PTR when asking to allocate a set with no elements. From myself. Patches 11 moves the nf_reinject function from the netfilter core (vmlinux) into the nfnetlink_queue backend, the only location where this is called from. Also from myself. Patch 12, from Kees Cook, switches xtables' compat layer to use unsafe_memcpy because xt_entry_target cannot easily get converted to a real flexible array (its UAPI and used inside other structs). * tag 'nf-next-24-02-21' of https://git.kernel.org/pub/scm/linux/kernel/git/netfilter/nf-next: netfilter: x_tables: Use unsafe_memcpy() for 0-sized destination netfilter: move nf_reinject into nfnetlink_queue modules netfilter: nft_set_pipapo: use GFP_KERNEL for insertions netfilter: nft_set_pipapo: speed up bulk element insertions netfilter: nft_set_pipapo: shrink data structures netfilter: nft_set_pipapo: do not rely on ZERO_SIZE_PTR netfilter: nft_set_pipapo: constify lookup fn args where possible netfilter: xtables: fix up kconfig dependencies netfilter: nft_osf: simplify init path netfilter: nf_log: validate nf_logger_find_get() netfilter: nf_log: consolidate check for NULL logger in lookup function netfilter: expect: Simplify the allocation of slab caches in nf_conntrack_expect_init ==================== Link: https://lore.kernel.org/r/20240221112637.5396-1-fw@strlen.de Signed-off-by: Jakub Kicinski <kuba@kernel.org>
This commit is contained in:
commit
4679f4f123
|
@ -370,7 +370,6 @@ __sum16 nf_checksum_partial(struct sk_buff *skb, unsigned int hook,
|
|||
u_int8_t protocol, unsigned short family);
|
||||
int nf_route(struct net *net, struct dst_entry **dst, struct flowi *fl,
|
||||
bool strict, unsigned short family);
|
||||
int nf_reroute(struct sk_buff *skb, struct nf_queue_entry *entry);
|
||||
|
||||
#include <net/flow.h>
|
||||
|
||||
|
|
|
@ -35,7 +35,6 @@ struct nf_queue_handler {
|
|||
|
||||
void nf_register_queue_handler(const struct nf_queue_handler *qh);
|
||||
void nf_unregister_queue_handler(void);
|
||||
void nf_reinject(struct nf_queue_entry *entry, unsigned int verdict);
|
||||
|
||||
bool nf_queue_entry_get_refs(struct nf_queue_entry *entry);
|
||||
void nf_queue_entry_free(struct nf_queue_entry *entry);
|
||||
|
|
|
@ -217,7 +217,7 @@ config IP_NF_NAT
|
|||
default m if NETFILTER_ADVANCED=n
|
||||
select NF_NAT
|
||||
select NETFILTER_XT_NAT
|
||||
select IP6_NF_IPTABLES_LEGACY
|
||||
select IP_NF_IPTABLES_LEGACY
|
||||
help
|
||||
This enables the `nat' table in iptables. This allows masquerading,
|
||||
port forwarding and other forms of full Network Address Port
|
||||
|
@ -329,6 +329,7 @@ config NFT_COMPAT_ARP
|
|||
config IP_NF_ARPFILTER
|
||||
tristate "arptables-legacy packet filtering support"
|
||||
select IP_NF_ARPTABLES
|
||||
depends on NETFILTER_XTABLES
|
||||
help
|
||||
ARP packet filtering defines a table `filter', which has a series of
|
||||
rules for simple ARP packet filtering at local input and
|
||||
|
|
|
@ -722,9 +722,7 @@ int nf_conntrack_expect_init(void)
|
|||
nf_ct_expect_hsize = 1;
|
||||
}
|
||||
nf_ct_expect_max = nf_ct_expect_hsize * 4;
|
||||
nf_ct_expect_cachep = kmem_cache_create("nf_conntrack_expect",
|
||||
sizeof(struct nf_conntrack_expect),
|
||||
0, 0, NULL);
|
||||
nf_ct_expect_cachep = KMEM_CACHE(nf_conntrack_expect, 0);
|
||||
if (!nf_ct_expect_cachep)
|
||||
return -ENOMEM;
|
||||
|
||||
|
|
|
@ -31,10 +31,10 @@ static struct nf_logger *__find_logger(int pf, const char *str_logger)
|
|||
int i;
|
||||
|
||||
for (i = 0; i < NF_LOG_TYPE_MAX; i++) {
|
||||
if (loggers[pf][i] == NULL)
|
||||
log = nft_log_dereference(loggers[pf][i]);
|
||||
if (!log)
|
||||
continue;
|
||||
|
||||
log = nft_log_dereference(loggers[pf][i]);
|
||||
if (!strncasecmp(str_logger, log->name, strlen(log->name)))
|
||||
return log;
|
||||
}
|
||||
|
@ -156,6 +156,11 @@ int nf_logger_find_get(int pf, enum nf_log_type type)
|
|||
struct nf_logger *logger;
|
||||
int ret = -ENOENT;
|
||||
|
||||
if (pf >= ARRAY_SIZE(loggers))
|
||||
return -EINVAL;
|
||||
if (type >= NF_LOG_TYPE_MAX)
|
||||
return -EINVAL;
|
||||
|
||||
if (pf == NFPROTO_INET) {
|
||||
ret = nf_logger_find_get(NFPROTO_IPV4, type);
|
||||
if (ret < 0)
|
||||
|
|
|
@ -248,109 +248,3 @@ int nf_queue(struct sk_buff *skb, struct nf_hook_state *state,
|
|||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(nf_queue);
|
||||
|
||||
static unsigned int nf_iterate(struct sk_buff *skb,
|
||||
struct nf_hook_state *state,
|
||||
const struct nf_hook_entries *hooks,
|
||||
unsigned int *index)
|
||||
{
|
||||
const struct nf_hook_entry *hook;
|
||||
unsigned int verdict, i = *index;
|
||||
|
||||
while (i < hooks->num_hook_entries) {
|
||||
hook = &hooks->hooks[i];
|
||||
repeat:
|
||||
verdict = nf_hook_entry_hookfn(hook, skb, state);
|
||||
if (verdict != NF_ACCEPT) {
|
||||
*index = i;
|
||||
if (verdict != NF_REPEAT)
|
||||
return verdict;
|
||||
goto repeat;
|
||||
}
|
||||
i++;
|
||||
}
|
||||
|
||||
*index = i;
|
||||
return NF_ACCEPT;
|
||||
}
|
||||
|
||||
static struct nf_hook_entries *nf_hook_entries_head(const struct net *net, u8 pf, u8 hooknum)
|
||||
{
|
||||
switch (pf) {
|
||||
#ifdef CONFIG_NETFILTER_FAMILY_BRIDGE
|
||||
case NFPROTO_BRIDGE:
|
||||
return rcu_dereference(net->nf.hooks_bridge[hooknum]);
|
||||
#endif
|
||||
case NFPROTO_IPV4:
|
||||
return rcu_dereference(net->nf.hooks_ipv4[hooknum]);
|
||||
case NFPROTO_IPV6:
|
||||
return rcu_dereference(net->nf.hooks_ipv6[hooknum]);
|
||||
default:
|
||||
WARN_ON_ONCE(1);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
return NULL;
|
||||
}
|
||||
|
||||
/* Caller must hold rcu read-side lock */
|
||||
void nf_reinject(struct nf_queue_entry *entry, unsigned int verdict)
|
||||
{
|
||||
const struct nf_hook_entry *hook_entry;
|
||||
const struct nf_hook_entries *hooks;
|
||||
struct sk_buff *skb = entry->skb;
|
||||
const struct net *net;
|
||||
unsigned int i;
|
||||
int err;
|
||||
u8 pf;
|
||||
|
||||
net = entry->state.net;
|
||||
pf = entry->state.pf;
|
||||
|
||||
hooks = nf_hook_entries_head(net, pf, entry->state.hook);
|
||||
|
||||
i = entry->hook_index;
|
||||
if (WARN_ON_ONCE(!hooks || i >= hooks->num_hook_entries)) {
|
||||
kfree_skb(skb);
|
||||
nf_queue_entry_free(entry);
|
||||
return;
|
||||
}
|
||||
|
||||
hook_entry = &hooks->hooks[i];
|
||||
|
||||
/* Continue traversal iff userspace said ok... */
|
||||
if (verdict == NF_REPEAT)
|
||||
verdict = nf_hook_entry_hookfn(hook_entry, skb, &entry->state);
|
||||
|
||||
if (verdict == NF_ACCEPT) {
|
||||
if (nf_reroute(skb, entry) < 0)
|
||||
verdict = NF_DROP;
|
||||
}
|
||||
|
||||
if (verdict == NF_ACCEPT) {
|
||||
next_hook:
|
||||
++i;
|
||||
verdict = nf_iterate(skb, &entry->state, hooks, &i);
|
||||
}
|
||||
|
||||
switch (verdict & NF_VERDICT_MASK) {
|
||||
case NF_ACCEPT:
|
||||
case NF_STOP:
|
||||
local_bh_disable();
|
||||
entry->state.okfn(entry->state.net, entry->state.sk, skb);
|
||||
local_bh_enable();
|
||||
break;
|
||||
case NF_QUEUE:
|
||||
err = nf_queue(skb, &entry->state, i, verdict);
|
||||
if (err == 1)
|
||||
goto next_hook;
|
||||
break;
|
||||
case NF_STOLEN:
|
||||
break;
|
||||
default:
|
||||
kfree_skb(skb);
|
||||
}
|
||||
|
||||
nf_queue_entry_free(entry);
|
||||
}
|
||||
EXPORT_SYMBOL(nf_reinject);
|
||||
|
|
|
@ -225,6 +225,148 @@ find_dequeue_entry(struct nfqnl_instance *queue, unsigned int id)
|
|||
return entry;
|
||||
}
|
||||
|
||||
static unsigned int nf_iterate(struct sk_buff *skb,
|
||||
struct nf_hook_state *state,
|
||||
const struct nf_hook_entries *hooks,
|
||||
unsigned int *index)
|
||||
{
|
||||
const struct nf_hook_entry *hook;
|
||||
unsigned int verdict, i = *index;
|
||||
|
||||
while (i < hooks->num_hook_entries) {
|
||||
hook = &hooks->hooks[i];
|
||||
repeat:
|
||||
verdict = nf_hook_entry_hookfn(hook, skb, state);
|
||||
if (verdict != NF_ACCEPT) {
|
||||
*index = i;
|
||||
if (verdict != NF_REPEAT)
|
||||
return verdict;
|
||||
goto repeat;
|
||||
}
|
||||
i++;
|
||||
}
|
||||
|
||||
*index = i;
|
||||
return NF_ACCEPT;
|
||||
}
|
||||
|
||||
static struct nf_hook_entries *nf_hook_entries_head(const struct net *net, u8 pf, u8 hooknum)
|
||||
{
|
||||
switch (pf) {
|
||||
#ifdef CONFIG_NETFILTER_FAMILY_BRIDGE
|
||||
case NFPROTO_BRIDGE:
|
||||
return rcu_dereference(net->nf.hooks_bridge[hooknum]);
|
||||
#endif
|
||||
case NFPROTO_IPV4:
|
||||
return rcu_dereference(net->nf.hooks_ipv4[hooknum]);
|
||||
case NFPROTO_IPV6:
|
||||
return rcu_dereference(net->nf.hooks_ipv6[hooknum]);
|
||||
default:
|
||||
WARN_ON_ONCE(1);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
return NULL;
|
||||
}
|
||||
|
||||
static int nf_ip_reroute(struct sk_buff *skb, const struct nf_queue_entry *entry)
|
||||
{
|
||||
#ifdef CONFIG_INET
|
||||
const struct ip_rt_info *rt_info = nf_queue_entry_reroute(entry);
|
||||
|
||||
if (entry->state.hook == NF_INET_LOCAL_OUT) {
|
||||
const struct iphdr *iph = ip_hdr(skb);
|
||||
|
||||
if (!(iph->tos == rt_info->tos &&
|
||||
skb->mark == rt_info->mark &&
|
||||
iph->daddr == rt_info->daddr &&
|
||||
iph->saddr == rt_info->saddr))
|
||||
return ip_route_me_harder(entry->state.net, entry->state.sk,
|
||||
skb, RTN_UNSPEC);
|
||||
}
|
||||
#endif
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int nf_reroute(struct sk_buff *skb, struct nf_queue_entry *entry)
|
||||
{
|
||||
const struct nf_ipv6_ops *v6ops;
|
||||
int ret = 0;
|
||||
|
||||
switch (entry->state.pf) {
|
||||
case AF_INET:
|
||||
ret = nf_ip_reroute(skb, entry);
|
||||
break;
|
||||
case AF_INET6:
|
||||
v6ops = rcu_dereference(nf_ipv6_ops);
|
||||
if (v6ops)
|
||||
ret = v6ops->reroute(skb, entry);
|
||||
break;
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
/* caller must hold rcu read-side lock */
|
||||
static void nf_reinject(struct nf_queue_entry *entry, unsigned int verdict)
|
||||
{
|
||||
const struct nf_hook_entry *hook_entry;
|
||||
const struct nf_hook_entries *hooks;
|
||||
struct sk_buff *skb = entry->skb;
|
||||
const struct net *net;
|
||||
unsigned int i;
|
||||
int err;
|
||||
u8 pf;
|
||||
|
||||
net = entry->state.net;
|
||||
pf = entry->state.pf;
|
||||
|
||||
hooks = nf_hook_entries_head(net, pf, entry->state.hook);
|
||||
|
||||
i = entry->hook_index;
|
||||
if (WARN_ON_ONCE(!hooks || i >= hooks->num_hook_entries)) {
|
||||
kfree_skb_reason(skb, SKB_DROP_REASON_NETFILTER_DROP);
|
||||
nf_queue_entry_free(entry);
|
||||
return;
|
||||
}
|
||||
|
||||
hook_entry = &hooks->hooks[i];
|
||||
|
||||
/* Continue traversal iff userspace said ok... */
|
||||
if (verdict == NF_REPEAT)
|
||||
verdict = nf_hook_entry_hookfn(hook_entry, skb, &entry->state);
|
||||
|
||||
if (verdict == NF_ACCEPT) {
|
||||
if (nf_reroute(skb, entry) < 0)
|
||||
verdict = NF_DROP;
|
||||
}
|
||||
|
||||
if (verdict == NF_ACCEPT) {
|
||||
next_hook:
|
||||
++i;
|
||||
verdict = nf_iterate(skb, &entry->state, hooks, &i);
|
||||
}
|
||||
|
||||
switch (verdict & NF_VERDICT_MASK) {
|
||||
case NF_ACCEPT:
|
||||
case NF_STOP:
|
||||
local_bh_disable();
|
||||
entry->state.okfn(entry->state.net, entry->state.sk, skb);
|
||||
local_bh_enable();
|
||||
break;
|
||||
case NF_QUEUE:
|
||||
err = nf_queue(skb, &entry->state, i, verdict);
|
||||
if (err == 1)
|
||||
goto next_hook;
|
||||
break;
|
||||
case NF_STOLEN:
|
||||
break;
|
||||
default:
|
||||
kfree_skb(skb);
|
||||
}
|
||||
|
||||
nf_queue_entry_free(entry);
|
||||
}
|
||||
|
||||
static void nfqnl_reinject(struct nf_queue_entry *entry, unsigned int verdict)
|
||||
{
|
||||
const struct nf_ct_hook *ct_hook;
|
||||
|
|
|
@ -63,7 +63,6 @@ static int nft_osf_init(const struct nft_ctx *ctx,
|
|||
{
|
||||
struct nft_osf *priv = nft_expr_priv(expr);
|
||||
u32 flags;
|
||||
int err;
|
||||
u8 ttl;
|
||||
|
||||
if (!tb[NFTA_OSF_DREG])
|
||||
|
@ -83,13 +82,9 @@ static int nft_osf_init(const struct nft_ctx *ctx,
|
|||
priv->flags = flags;
|
||||
}
|
||||
|
||||
err = nft_parse_register_store(ctx, tb[NFTA_OSF_DREG], &priv->dreg,
|
||||
NULL, NFT_DATA_VALUE,
|
||||
NFT_OSF_MAXGENRELEN);
|
||||
if (err < 0)
|
||||
return err;
|
||||
|
||||
return 0;
|
||||
return nft_parse_register_store(ctx, tb[NFTA_OSF_DREG], &priv->dreg,
|
||||
NULL, NFT_DATA_VALUE,
|
||||
NFT_OSF_MAXGENRELEN);
|
||||
}
|
||||
|
||||
static int nft_osf_dump(struct sk_buff *skb,
|
||||
|
|
|
@ -359,11 +359,13 @@
|
|||
*
|
||||
* Return: -1 on no match, bit position on 'match_only', 0 otherwise.
|
||||
*/
|
||||
int pipapo_refill(unsigned long *map, int len, int rules, unsigned long *dst,
|
||||
union nft_pipapo_map_bucket *mt, bool match_only)
|
||||
int pipapo_refill(unsigned long *map, unsigned int len, unsigned int rules,
|
||||
unsigned long *dst,
|
||||
const union nft_pipapo_map_bucket *mt, bool match_only)
|
||||
{
|
||||
unsigned long bitset;
|
||||
int k, ret = -1;
|
||||
unsigned int k;
|
||||
int ret = -1;
|
||||
|
||||
for (k = 0; k < len; k++) {
|
||||
bitset = map[k];
|
||||
|
@ -412,9 +414,9 @@ bool nft_pipapo_lookup(const struct net *net, const struct nft_set *set,
|
|||
struct nft_pipapo_scratch *scratch;
|
||||
unsigned long *res_map, *fill_map;
|
||||
u8 genmask = nft_genmask_cur(net);
|
||||
const struct nft_pipapo_match *m;
|
||||
const struct nft_pipapo_field *f;
|
||||
const u8 *rp = (const u8 *)key;
|
||||
struct nft_pipapo_match *m;
|
||||
struct nft_pipapo_field *f;
|
||||
bool map_index;
|
||||
int i;
|
||||
|
||||
|
@ -505,6 +507,7 @@ out:
|
|||
* @data: Key data to be matched against existing elements
|
||||
* @genmask: If set, check that element is active in given genmask
|
||||
* @tstamp: timestamp to check for expired elements
|
||||
* @gfp: the type of memory to allocate (see kmalloc).
|
||||
*
|
||||
* This is essentially the same as the lookup function, except that it matches
|
||||
* key data against the uncommitted copy and doesn't use preallocated maps for
|
||||
|
@ -515,22 +518,26 @@ out:
|
|||
static struct nft_pipapo_elem *pipapo_get(const struct net *net,
|
||||
const struct nft_set *set,
|
||||
const u8 *data, u8 genmask,
|
||||
u64 tstamp)
|
||||
u64 tstamp, gfp_t gfp)
|
||||
{
|
||||
struct nft_pipapo_elem *ret = ERR_PTR(-ENOENT);
|
||||
struct nft_pipapo *priv = nft_set_priv(set);
|
||||
struct nft_pipapo_match *m = priv->clone;
|
||||
unsigned long *res_map, *fill_map = NULL;
|
||||
struct nft_pipapo_field *f;
|
||||
const struct nft_pipapo_match *m;
|
||||
const struct nft_pipapo_field *f;
|
||||
int i;
|
||||
|
||||
res_map = kmalloc_array(m->bsize_max, sizeof(*res_map), GFP_ATOMIC);
|
||||
m = priv->clone;
|
||||
if (m->bsize_max == 0)
|
||||
return ret;
|
||||
|
||||
res_map = kmalloc_array(m->bsize_max, sizeof(*res_map), gfp);
|
||||
if (!res_map) {
|
||||
ret = ERR_PTR(-ENOMEM);
|
||||
goto out;
|
||||
}
|
||||
|
||||
fill_map = kcalloc(m->bsize_max, sizeof(*res_map), GFP_ATOMIC);
|
||||
fill_map = kcalloc(m->bsize_max, sizeof(*res_map), gfp);
|
||||
if (!fill_map) {
|
||||
ret = ERR_PTR(-ENOMEM);
|
||||
goto out;
|
||||
|
@ -608,13 +615,73 @@ nft_pipapo_get(const struct net *net, const struct nft_set *set,
|
|||
struct nft_pipapo_elem *e;
|
||||
|
||||
e = pipapo_get(net, set, (const u8 *)elem->key.val.data,
|
||||
nft_genmask_cur(net), get_jiffies_64());
|
||||
nft_genmask_cur(net), get_jiffies_64(),
|
||||
GFP_ATOMIC);
|
||||
if (IS_ERR(e))
|
||||
return ERR_CAST(e);
|
||||
|
||||
return &e->priv;
|
||||
}
|
||||
|
||||
/**
|
||||
* pipapo_realloc_mt() - Reallocate mapping table if needed upon resize
|
||||
* @f: Field containing mapping table
|
||||
* @old_rules: Amount of existing mapped rules
|
||||
* @rules: Amount of new rules to map
|
||||
*
|
||||
* Return: 0 on success, negative error code on failure.
|
||||
*/
|
||||
static int pipapo_realloc_mt(struct nft_pipapo_field *f,
|
||||
unsigned int old_rules, unsigned int rules)
|
||||
{
|
||||
union nft_pipapo_map_bucket *new_mt = NULL, *old_mt = f->mt;
|
||||
const unsigned int extra = PAGE_SIZE / sizeof(*new_mt);
|
||||
unsigned int rules_alloc = rules;
|
||||
|
||||
might_sleep();
|
||||
|
||||
if (unlikely(rules == 0))
|
||||
goto out_free;
|
||||
|
||||
/* growing and enough space left, no action needed */
|
||||
if (rules > old_rules && f->rules_alloc > rules)
|
||||
return 0;
|
||||
|
||||
/* downsize and extra slack has not grown too large */
|
||||
if (rules < old_rules) {
|
||||
unsigned int remove = f->rules_alloc - rules;
|
||||
|
||||
if (remove < (2u * extra))
|
||||
return 0;
|
||||
}
|
||||
|
||||
/* If set needs more than one page of memory for rules then
|
||||
* allocate another extra page to avoid frequent reallocation.
|
||||
*/
|
||||
if (rules > extra &&
|
||||
check_add_overflow(rules, extra, &rules_alloc))
|
||||
return -EOVERFLOW;
|
||||
|
||||
new_mt = kvmalloc_array(rules_alloc, sizeof(*new_mt), GFP_KERNEL);
|
||||
if (!new_mt)
|
||||
return -ENOMEM;
|
||||
|
||||
if (old_mt)
|
||||
memcpy(new_mt, old_mt, min(old_rules, rules) * sizeof(*new_mt));
|
||||
|
||||
if (rules > old_rules) {
|
||||
memset(new_mt + old_rules, 0,
|
||||
(rules - old_rules) * sizeof(*new_mt));
|
||||
}
|
||||
out_free:
|
||||
f->rules_alloc = rules_alloc;
|
||||
f->mt = new_mt;
|
||||
|
||||
kvfree(old_mt);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* pipapo_resize() - Resize lookup or mapping table, or both
|
||||
* @f: Field containing lookup and mapping tables
|
||||
|
@ -627,12 +694,15 @@ nft_pipapo_get(const struct net *net, const struct nft_set *set,
|
|||
*
|
||||
* Return: 0 on success, -ENOMEM on allocation failure.
|
||||
*/
|
||||
static int pipapo_resize(struct nft_pipapo_field *f, int old_rules, int rules)
|
||||
static int pipapo_resize(struct nft_pipapo_field *f,
|
||||
unsigned int old_rules, unsigned int rules)
|
||||
{
|
||||
long *new_lt = NULL, *new_p, *old_lt = f->lt, *old_p;
|
||||
union nft_pipapo_map_bucket *new_mt, *old_mt = f->mt;
|
||||
size_t new_bucket_size, copy;
|
||||
int group, bucket;
|
||||
unsigned int new_bucket_size, copy;
|
||||
int group, bucket, err;
|
||||
|
||||
if (rules >= NFT_PIPAPO_RULE0_MAX)
|
||||
return -ENOSPC;
|
||||
|
||||
new_bucket_size = DIV_ROUND_UP(rules, BITS_PER_LONG);
|
||||
#ifdef NFT_PIPAPO_ALIGN
|
||||
|
@ -672,27 +742,18 @@ static int pipapo_resize(struct nft_pipapo_field *f, int old_rules, int rules)
|
|||
}
|
||||
|
||||
mt:
|
||||
new_mt = kvmalloc(rules * sizeof(*new_mt), GFP_KERNEL);
|
||||
if (!new_mt) {
|
||||
err = pipapo_realloc_mt(f, old_rules, rules);
|
||||
if (err) {
|
||||
kvfree(new_lt);
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
memcpy(new_mt, f->mt, min(old_rules, rules) * sizeof(*new_mt));
|
||||
if (rules > old_rules) {
|
||||
memset(new_mt + old_rules, 0,
|
||||
(rules - old_rules) * sizeof(*new_mt));
|
||||
return err;
|
||||
}
|
||||
|
||||
if (new_lt) {
|
||||
f->bsize = new_bucket_size;
|
||||
NFT_PIPAPO_LT_ASSIGN(f, new_lt);
|
||||
f->lt = new_lt;
|
||||
kvfree(old_lt);
|
||||
}
|
||||
|
||||
f->mt = new_mt;
|
||||
kvfree(old_mt);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
@ -843,8 +904,8 @@ static void pipapo_lt_8b_to_4b(int old_groups, int bsize,
|
|||
*/
|
||||
static void pipapo_lt_bits_adjust(struct nft_pipapo_field *f)
|
||||
{
|
||||
unsigned int groups, bb;
|
||||
unsigned long *new_lt;
|
||||
int groups, bb;
|
||||
size_t lt_size;
|
||||
|
||||
lt_size = f->groups * NFT_PIPAPO_BUCKETS(f->bb) * f->bsize *
|
||||
|
@ -894,7 +955,7 @@ static void pipapo_lt_bits_adjust(struct nft_pipapo_field *f)
|
|||
f->groups = groups;
|
||||
f->bb = bb;
|
||||
kvfree(f->lt);
|
||||
NFT_PIPAPO_LT_ASSIGN(f, new_lt);
|
||||
f->lt = new_lt;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -911,7 +972,7 @@ static void pipapo_lt_bits_adjust(struct nft_pipapo_field *f)
|
|||
static int pipapo_insert(struct nft_pipapo_field *f, const uint8_t *k,
|
||||
int mask_bits)
|
||||
{
|
||||
int rule = f->rules, group, ret, bit_offset = 0;
|
||||
unsigned int rule = f->rules, group, ret, bit_offset = 0;
|
||||
|
||||
ret = pipapo_resize(f, f->rules, f->rules + 1);
|
||||
if (ret)
|
||||
|
@ -1216,7 +1277,7 @@ static int nft_pipapo_insert(const struct net *net, const struct nft_set *set,
|
|||
else
|
||||
end = start;
|
||||
|
||||
dup = pipapo_get(net, set, start, genmask, tstamp);
|
||||
dup = pipapo_get(net, set, start, genmask, tstamp, GFP_KERNEL);
|
||||
if (!IS_ERR(dup)) {
|
||||
/* Check if we already have the same exact entry */
|
||||
const struct nft_data *dup_key, *dup_end;
|
||||
|
@ -1238,7 +1299,8 @@ static int nft_pipapo_insert(const struct net *net, const struct nft_set *set,
|
|||
|
||||
if (PTR_ERR(dup) == -ENOENT) {
|
||||
/* Look for partially overlapping entries */
|
||||
dup = pipapo_get(net, set, end, nft_genmask_next(net), tstamp);
|
||||
dup = pipapo_get(net, set, end, nft_genmask_next(net), tstamp,
|
||||
GFP_KERNEL);
|
||||
}
|
||||
|
||||
if (PTR_ERR(dup) != -ENOENT) {
|
||||
|
@ -1251,8 +1313,14 @@ static int nft_pipapo_insert(const struct net *net, const struct nft_set *set,
|
|||
/* Validate */
|
||||
start_p = start;
|
||||
end_p = end;
|
||||
|
||||
/* some helpers return -1, or 0 >= for valid rule pos,
|
||||
* so we cannot support more than INT_MAX rules at this time.
|
||||
*/
|
||||
BUILD_BUG_ON(NFT_PIPAPO_RULE0_MAX > INT_MAX);
|
||||
|
||||
nft_pipapo_for_each_field(f, i, m) {
|
||||
if (f->rules >= (unsigned long)NFT_PIPAPO_RULE0_MAX)
|
||||
if (f->rules >= NFT_PIPAPO_RULE0_MAX)
|
||||
return -ENOSPC;
|
||||
|
||||
if (memcmp(start_p, end_p,
|
||||
|
@ -1358,18 +1426,25 @@ static struct nft_pipapo_match *pipapo_clone(struct nft_pipapo_match *old)
|
|||
if (!new_lt)
|
||||
goto out_lt;
|
||||
|
||||
NFT_PIPAPO_LT_ASSIGN(dst, new_lt);
|
||||
dst->lt = new_lt;
|
||||
|
||||
memcpy(NFT_PIPAPO_LT_ALIGN(new_lt),
|
||||
NFT_PIPAPO_LT_ALIGN(src->lt),
|
||||
src->bsize * sizeof(*dst->lt) *
|
||||
src->groups * NFT_PIPAPO_BUCKETS(src->bb));
|
||||
|
||||
dst->mt = kvmalloc(src->rules * sizeof(*src->mt), GFP_KERNEL);
|
||||
if (!dst->mt)
|
||||
goto out_mt;
|
||||
if (src->rules > 0) {
|
||||
dst->mt = kvmalloc_array(src->rules_alloc,
|
||||
sizeof(*src->mt), GFP_KERNEL);
|
||||
if (!dst->mt)
|
||||
goto out_mt;
|
||||
|
||||
memcpy(dst->mt, src->mt, src->rules * sizeof(*src->mt));
|
||||
} else {
|
||||
dst->mt = NULL;
|
||||
dst->rules_alloc = 0;
|
||||
}
|
||||
|
||||
memcpy(dst->mt, src->mt, src->rules * sizeof(*src->mt));
|
||||
src++;
|
||||
dst++;
|
||||
}
|
||||
|
@ -1423,10 +1498,10 @@ out_scratch:
|
|||
*
|
||||
* Return: Number of rules that originated from the same entry as @first.
|
||||
*/
|
||||
static int pipapo_rules_same_key(struct nft_pipapo_field *f, int first)
|
||||
static unsigned int pipapo_rules_same_key(struct nft_pipapo_field *f, unsigned int first)
|
||||
{
|
||||
struct nft_pipapo_elem *e = NULL; /* Keep gcc happy */
|
||||
int r;
|
||||
unsigned int r;
|
||||
|
||||
for (r = first; r < f->rules; r++) {
|
||||
if (r != first && e != f->mt[r].e)
|
||||
|
@ -1479,8 +1554,9 @@ static int pipapo_rules_same_key(struct nft_pipapo_field *f, int first)
|
|||
* 0 1 2
|
||||
* element pointers: 0x42 0x42 0x44
|
||||
*/
|
||||
static void pipapo_unmap(union nft_pipapo_map_bucket *mt, int rules,
|
||||
int start, int n, int to_offset, bool is_last)
|
||||
static void pipapo_unmap(union nft_pipapo_map_bucket *mt, unsigned int rules,
|
||||
unsigned int start, unsigned int n,
|
||||
unsigned int to_offset, bool is_last)
|
||||
{
|
||||
int i;
|
||||
|
||||
|
@ -1586,8 +1662,8 @@ static void pipapo_gc(struct nft_set *set, struct nft_pipapo_match *m)
|
|||
{
|
||||
struct nft_pipapo *priv = nft_set_priv(set);
|
||||
struct net *net = read_pnet(&set->net);
|
||||
unsigned int rules_f0, first_rule = 0;
|
||||
u64 tstamp = nft_net_tstamp(net);
|
||||
int rules_f0, first_rule = 0;
|
||||
struct nft_pipapo_elem *e;
|
||||
struct nft_trans_gc *gc;
|
||||
|
||||
|
@ -1597,8 +1673,8 @@ static void pipapo_gc(struct nft_set *set, struct nft_pipapo_match *m)
|
|||
|
||||
while ((rules_f0 = pipapo_rules_same_key(m->f, first_rule))) {
|
||||
union nft_pipapo_map_bucket rulemap[NFT_PIPAPO_MAX_FIELDS];
|
||||
struct nft_pipapo_field *f;
|
||||
int i, start, rules_fx;
|
||||
const struct nft_pipapo_field *f;
|
||||
unsigned int i, start, rules_fx;
|
||||
|
||||
start = first_rule;
|
||||
rules_fx = rules_f0;
|
||||
|
@ -1792,7 +1868,8 @@ static void *pipapo_deactivate(const struct net *net, const struct nft_set *set,
|
|||
{
|
||||
struct nft_pipapo_elem *e;
|
||||
|
||||
e = pipapo_get(net, set, data, nft_genmask_next(net), nft_net_tstamp(net));
|
||||
e = pipapo_get(net, set, data, nft_genmask_next(net),
|
||||
nft_net_tstamp(net), GFP_KERNEL);
|
||||
if (IS_ERR(e))
|
||||
return NULL;
|
||||
|
||||
|
@ -1976,7 +2053,7 @@ static void nft_pipapo_remove(const struct net *net, const struct nft_set *set,
|
|||
{
|
||||
struct nft_pipapo *priv = nft_set_priv(set);
|
||||
struct nft_pipapo_match *m = priv->clone;
|
||||
int rules_f0, first_rule = 0;
|
||||
unsigned int rules_f0, first_rule = 0;
|
||||
struct nft_pipapo_elem *e;
|
||||
const u8 *data;
|
||||
|
||||
|
@ -2039,9 +2116,9 @@ static void nft_pipapo_walk(const struct nft_ctx *ctx, struct nft_set *set,
|
|||
{
|
||||
struct nft_pipapo *priv = nft_set_priv(set);
|
||||
struct net *net = read_pnet(&set->net);
|
||||
struct nft_pipapo_match *m;
|
||||
struct nft_pipapo_field *f;
|
||||
int i, r;
|
||||
const struct nft_pipapo_match *m;
|
||||
const struct nft_pipapo_field *f;
|
||||
unsigned int i, r;
|
||||
|
||||
rcu_read_lock();
|
||||
if (iter->genmask == nft_genmask_cur(net))
|
||||
|
@ -2145,6 +2222,9 @@ static int nft_pipapo_init(const struct nft_set *set,
|
|||
|
||||
field_count = desc->field_count ? : 1;
|
||||
|
||||
BUILD_BUG_ON(NFT_PIPAPO_MAX_FIELDS > 255);
|
||||
BUILD_BUG_ON(NFT_PIPAPO_MAX_FIELDS != NFT_REG32_COUNT);
|
||||
|
||||
if (field_count > NFT_PIPAPO_MAX_FIELDS)
|
||||
return -EINVAL;
|
||||
|
||||
|
@ -2166,7 +2246,11 @@ static int nft_pipapo_init(const struct nft_set *set,
|
|||
rcu_head_init(&m->rcu);
|
||||
|
||||
nft_pipapo_for_each_field(f, i, m) {
|
||||
int len = desc->field_len[i] ? : set->klen;
|
||||
unsigned int len = desc->field_len[i] ? : set->klen;
|
||||
|
||||
/* f->groups is u8 */
|
||||
BUILD_BUG_ON((NFT_PIPAPO_MAX_BYTES *
|
||||
BITS_PER_BYTE / NFT_PIPAPO_GROUP_BITS_LARGE_SET) >= 256);
|
||||
|
||||
f->bb = NFT_PIPAPO_GROUP_BITS_INIT;
|
||||
f->groups = len * NFT_PIPAPO_GROUPS_PER_BYTE(f);
|
||||
|
@ -2175,7 +2259,8 @@ static int nft_pipapo_init(const struct nft_set *set,
|
|||
|
||||
f->bsize = 0;
|
||||
f->rules = 0;
|
||||
NFT_PIPAPO_LT_ASSIGN(f, NULL);
|
||||
f->rules_alloc = 0;
|
||||
f->lt = NULL;
|
||||
f->mt = NULL;
|
||||
}
|
||||
|
||||
|
@ -2211,7 +2296,7 @@ static void nft_set_pipapo_match_destroy(const struct nft_ctx *ctx,
|
|||
struct nft_pipapo_match *m)
|
||||
{
|
||||
struct nft_pipapo_field *f;
|
||||
int i, r;
|
||||
unsigned int i, r;
|
||||
|
||||
for (i = 0, f = m->f; i < m->field_count - 1; i++, f++)
|
||||
;
|
||||
|
|
|
@ -70,15 +70,9 @@
|
|||
#define NFT_PIPAPO_ALIGN_HEADROOM \
|
||||
(NFT_PIPAPO_ALIGN - ARCH_KMALLOC_MINALIGN)
|
||||
#define NFT_PIPAPO_LT_ALIGN(lt) (PTR_ALIGN((lt), NFT_PIPAPO_ALIGN))
|
||||
#define NFT_PIPAPO_LT_ASSIGN(field, x) \
|
||||
do { \
|
||||
(field)->lt_aligned = NFT_PIPAPO_LT_ALIGN(x); \
|
||||
(field)->lt = (x); \
|
||||
} while (0)
|
||||
#else
|
||||
#define NFT_PIPAPO_ALIGN_HEADROOM 0
|
||||
#define NFT_PIPAPO_LT_ALIGN(lt) (lt)
|
||||
#define NFT_PIPAPO_LT_ASSIGN(field, x) ((field)->lt = (x))
|
||||
#endif /* NFT_PIPAPO_ALIGN */
|
||||
|
||||
#define nft_pipapo_for_each_field(field, index, match) \
|
||||
|
@ -110,22 +104,20 @@ union nft_pipapo_map_bucket {
|
|||
|
||||
/**
|
||||
* struct nft_pipapo_field - Lookup, mapping tables and related data for a field
|
||||
* @groups: Amount of bit groups
|
||||
* @rules: Number of inserted rules
|
||||
* @bsize: Size of each bucket in lookup table, in longs
|
||||
* @rules_alloc: Number of allocated rules, always >= rules
|
||||
* @groups: Amount of bit groups
|
||||
* @bb: Number of bits grouped together in lookup table buckets
|
||||
* @lt: Lookup table: 'groups' rows of buckets
|
||||
* @lt_aligned: Version of @lt aligned to NFT_PIPAPO_ALIGN bytes
|
||||
* @mt: Mapping table: one bucket per rule
|
||||
*/
|
||||
struct nft_pipapo_field {
|
||||
int groups;
|
||||
unsigned long rules;
|
||||
size_t bsize;
|
||||
int bb;
|
||||
#ifdef NFT_PIPAPO_ALIGN
|
||||
unsigned long *lt_aligned;
|
||||
#endif
|
||||
unsigned int rules;
|
||||
unsigned int bsize;
|
||||
unsigned int rules_alloc;
|
||||
u8 groups;
|
||||
u8 bb;
|
||||
unsigned long *lt;
|
||||
union nft_pipapo_map_bucket *mt;
|
||||
};
|
||||
|
@ -145,15 +137,15 @@ struct nft_pipapo_scratch {
|
|||
/**
|
||||
* struct nft_pipapo_match - Data used for lookup and matching
|
||||
* @field_count: Amount of fields in set
|
||||
* @scratch: Preallocated per-CPU maps for partial matching results
|
||||
* @bsize_max: Maximum lookup table bucket size of all fields, in longs
|
||||
* @scratch: Preallocated per-CPU maps for partial matching results
|
||||
* @rcu: Matching data is swapped on commits
|
||||
* @f: Fields, with lookup and mapping tables
|
||||
*/
|
||||
struct nft_pipapo_match {
|
||||
int field_count;
|
||||
u8 field_count;
|
||||
unsigned int bsize_max;
|
||||
struct nft_pipapo_scratch * __percpu *scratch;
|
||||
size_t bsize_max;
|
||||
struct rcu_head rcu;
|
||||
struct nft_pipapo_field f[] __counted_by(field_count);
|
||||
};
|
||||
|
@ -186,8 +178,9 @@ struct nft_pipapo_elem {
|
|||
struct nft_set_ext ext;
|
||||
};
|
||||
|
||||
int pipapo_refill(unsigned long *map, int len, int rules, unsigned long *dst,
|
||||
union nft_pipapo_map_bucket *mt, bool match_only);
|
||||
int pipapo_refill(unsigned long *map, unsigned int len, unsigned int rules,
|
||||
unsigned long *dst,
|
||||
const union nft_pipapo_map_bucket *mt, bool match_only);
|
||||
|
||||
/**
|
||||
* pipapo_and_field_buckets_4bit() - Intersect 4-bit buckets
|
||||
|
@ -195,7 +188,7 @@ int pipapo_refill(unsigned long *map, int len, int rules, unsigned long *dst,
|
|||
* @dst: Area to store result
|
||||
* @data: Input data selecting table buckets
|
||||
*/
|
||||
static inline void pipapo_and_field_buckets_4bit(struct nft_pipapo_field *f,
|
||||
static inline void pipapo_and_field_buckets_4bit(const struct nft_pipapo_field *f,
|
||||
unsigned long *dst,
|
||||
const u8 *data)
|
||||
{
|
||||
|
@ -223,7 +216,7 @@ static inline void pipapo_and_field_buckets_4bit(struct nft_pipapo_field *f,
|
|||
* @dst: Area to store result
|
||||
* @data: Input data selecting table buckets
|
||||
*/
|
||||
static inline void pipapo_and_field_buckets_8bit(struct nft_pipapo_field *f,
|
||||
static inline void pipapo_and_field_buckets_8bit(const struct nft_pipapo_field *f,
|
||||
unsigned long *dst,
|
||||
const u8 *data)
|
||||
{
|
||||
|
|
|
@ -212,8 +212,9 @@ static int nft_pipapo_avx2_refill(int offset, unsigned long *map,
|
|||
* word index to be checked next (i.e. first filled word).
|
||||
*/
|
||||
static int nft_pipapo_avx2_lookup_4b_2(unsigned long *map, unsigned long *fill,
|
||||
struct nft_pipapo_field *f, int offset,
|
||||
const u8 *pkt, bool first, bool last)
|
||||
const struct nft_pipapo_field *f,
|
||||
int offset, const u8 *pkt,
|
||||
bool first, bool last)
|
||||
{
|
||||
int i, ret = -1, m256_size = f->bsize / NFT_PIPAPO_LONGS_PER_M256, b;
|
||||
u8 pg[2] = { pkt[0] >> 4, pkt[0] & 0xf };
|
||||
|
@ -274,8 +275,9 @@ nothing:
|
|||
* word index to be checked next (i.e. first filled word).
|
||||
*/
|
||||
static int nft_pipapo_avx2_lookup_4b_4(unsigned long *map, unsigned long *fill,
|
||||
struct nft_pipapo_field *f, int offset,
|
||||
const u8 *pkt, bool first, bool last)
|
||||
const struct nft_pipapo_field *f,
|
||||
int offset, const u8 *pkt,
|
||||
bool first, bool last)
|
||||
{
|
||||
int i, ret = -1, m256_size = f->bsize / NFT_PIPAPO_LONGS_PER_M256, b;
|
||||
u8 pg[4] = { pkt[0] >> 4, pkt[0] & 0xf, pkt[1] >> 4, pkt[1] & 0xf };
|
||||
|
@ -350,8 +352,9 @@ nothing:
|
|||
* word index to be checked next (i.e. first filled word).
|
||||
*/
|
||||
static int nft_pipapo_avx2_lookup_4b_8(unsigned long *map, unsigned long *fill,
|
||||
struct nft_pipapo_field *f, int offset,
|
||||
const u8 *pkt, bool first, bool last)
|
||||
const struct nft_pipapo_field *f,
|
||||
int offset, const u8 *pkt,
|
||||
bool first, bool last)
|
||||
{
|
||||
u8 pg[8] = { pkt[0] >> 4, pkt[0] & 0xf, pkt[1] >> 4, pkt[1] & 0xf,
|
||||
pkt[2] >> 4, pkt[2] & 0xf, pkt[3] >> 4, pkt[3] & 0xf,
|
||||
|
@ -445,8 +448,9 @@ nothing:
|
|||
* word index to be checked next (i.e. first filled word).
|
||||
*/
|
||||
static int nft_pipapo_avx2_lookup_4b_12(unsigned long *map, unsigned long *fill,
|
||||
struct nft_pipapo_field *f, int offset,
|
||||
const u8 *pkt, bool first, bool last)
|
||||
const struct nft_pipapo_field *f,
|
||||
int offset, const u8 *pkt,
|
||||
bool first, bool last)
|
||||
{
|
||||
u8 pg[12] = { pkt[0] >> 4, pkt[0] & 0xf, pkt[1] >> 4, pkt[1] & 0xf,
|
||||
pkt[2] >> 4, pkt[2] & 0xf, pkt[3] >> 4, pkt[3] & 0xf,
|
||||
|
@ -534,8 +538,9 @@ nothing:
|
|||
* word index to be checked next (i.e. first filled word).
|
||||
*/
|
||||
static int nft_pipapo_avx2_lookup_4b_32(unsigned long *map, unsigned long *fill,
|
||||
struct nft_pipapo_field *f, int offset,
|
||||
const u8 *pkt, bool first, bool last)
|
||||
const struct nft_pipapo_field *f,
|
||||
int offset, const u8 *pkt,
|
||||
bool first, bool last)
|
||||
{
|
||||
u8 pg[32] = { pkt[0] >> 4, pkt[0] & 0xf, pkt[1] >> 4, pkt[1] & 0xf,
|
||||
pkt[2] >> 4, pkt[2] & 0xf, pkt[3] >> 4, pkt[3] & 0xf,
|
||||
|
@ -669,8 +674,9 @@ nothing:
|
|||
* word index to be checked next (i.e. first filled word).
|
||||
*/
|
||||
static int nft_pipapo_avx2_lookup_8b_1(unsigned long *map, unsigned long *fill,
|
||||
struct nft_pipapo_field *f, int offset,
|
||||
const u8 *pkt, bool first, bool last)
|
||||
const struct nft_pipapo_field *f,
|
||||
int offset, const u8 *pkt,
|
||||
bool first, bool last)
|
||||
{
|
||||
int i, ret = -1, m256_size = f->bsize / NFT_PIPAPO_LONGS_PER_M256, b;
|
||||
unsigned long *lt = f->lt, bsize = f->bsize;
|
||||
|
@ -726,8 +732,9 @@ nothing:
|
|||
* word index to be checked next (i.e. first filled word).
|
||||
*/
|
||||
static int nft_pipapo_avx2_lookup_8b_2(unsigned long *map, unsigned long *fill,
|
||||
struct nft_pipapo_field *f, int offset,
|
||||
const u8 *pkt, bool first, bool last)
|
||||
const struct nft_pipapo_field *f,
|
||||
int offset, const u8 *pkt,
|
||||
bool first, bool last)
|
||||
{
|
||||
int i, ret = -1, m256_size = f->bsize / NFT_PIPAPO_LONGS_PER_M256, b;
|
||||
unsigned long *lt = f->lt, bsize = f->bsize;
|
||||
|
@ -790,8 +797,9 @@ nothing:
|
|||
* word index to be checked next (i.e. first filled word).
|
||||
*/
|
||||
static int nft_pipapo_avx2_lookup_8b_4(unsigned long *map, unsigned long *fill,
|
||||
struct nft_pipapo_field *f, int offset,
|
||||
const u8 *pkt, bool first, bool last)
|
||||
const struct nft_pipapo_field *f,
|
||||
int offset, const u8 *pkt,
|
||||
bool first, bool last)
|
||||
{
|
||||
int i, ret = -1, m256_size = f->bsize / NFT_PIPAPO_LONGS_PER_M256, b;
|
||||
unsigned long *lt = f->lt, bsize = f->bsize;
|
||||
|
@ -865,8 +873,9 @@ nothing:
|
|||
* word index to be checked next (i.e. first filled word).
|
||||
*/
|
||||
static int nft_pipapo_avx2_lookup_8b_6(unsigned long *map, unsigned long *fill,
|
||||
struct nft_pipapo_field *f, int offset,
|
||||
const u8 *pkt, bool first, bool last)
|
||||
const struct nft_pipapo_field *f,
|
||||
int offset, const u8 *pkt,
|
||||
bool first, bool last)
|
||||
{
|
||||
int i, ret = -1, m256_size = f->bsize / NFT_PIPAPO_LONGS_PER_M256, b;
|
||||
unsigned long *lt = f->lt, bsize = f->bsize;
|
||||
|
@ -950,8 +959,9 @@ nothing:
|
|||
* word index to be checked next (i.e. first filled word).
|
||||
*/
|
||||
static int nft_pipapo_avx2_lookup_8b_16(unsigned long *map, unsigned long *fill,
|
||||
struct nft_pipapo_field *f, int offset,
|
||||
const u8 *pkt, bool first, bool last)
|
||||
const struct nft_pipapo_field *f,
|
||||
int offset, const u8 *pkt,
|
||||
bool first, bool last)
|
||||
{
|
||||
int i, ret = -1, m256_size = f->bsize / NFT_PIPAPO_LONGS_PER_M256, b;
|
||||
unsigned long *lt = f->lt, bsize = f->bsize;
|
||||
|
@ -1042,8 +1052,9 @@ nothing:
|
|||
* word index to be checked next (i.e. first filled word).
|
||||
*/
|
||||
static int nft_pipapo_avx2_lookup_slow(unsigned long *map, unsigned long *fill,
|
||||
struct nft_pipapo_field *f, int offset,
|
||||
const u8 *pkt, bool first, bool last)
|
||||
const struct nft_pipapo_field *f,
|
||||
int offset, const u8 *pkt,
|
||||
bool first, bool last)
|
||||
{
|
||||
unsigned long bsize = f->bsize;
|
||||
int i, ret = -1, b;
|
||||
|
@ -1119,9 +1130,9 @@ bool nft_pipapo_avx2_lookup(const struct net *net, const struct nft_set *set,
|
|||
struct nft_pipapo *priv = nft_set_priv(set);
|
||||
struct nft_pipapo_scratch *scratch;
|
||||
u8 genmask = nft_genmask_cur(net);
|
||||
const struct nft_pipapo_match *m;
|
||||
const struct nft_pipapo_field *f;
|
||||
const u8 *rp = (const u8 *)key;
|
||||
struct nft_pipapo_match *m;
|
||||
struct nft_pipapo_field *f;
|
||||
unsigned long *res, *fill;
|
||||
bool map_index;
|
||||
int i, ret = 0;
|
||||
|
|
|
@ -179,43 +179,6 @@ int nf_route(struct net *net, struct dst_entry **dst, struct flowi *fl,
|
|||
}
|
||||
EXPORT_SYMBOL_GPL(nf_route);
|
||||
|
||||
static int nf_ip_reroute(struct sk_buff *skb, const struct nf_queue_entry *entry)
|
||||
{
|
||||
#ifdef CONFIG_INET
|
||||
const struct ip_rt_info *rt_info = nf_queue_entry_reroute(entry);
|
||||
|
||||
if (entry->state.hook == NF_INET_LOCAL_OUT) {
|
||||
const struct iphdr *iph = ip_hdr(skb);
|
||||
|
||||
if (!(iph->tos == rt_info->tos &&
|
||||
skb->mark == rt_info->mark &&
|
||||
iph->daddr == rt_info->daddr &&
|
||||
iph->saddr == rt_info->saddr))
|
||||
return ip_route_me_harder(entry->state.net, entry->state.sk,
|
||||
skb, RTN_UNSPEC);
|
||||
}
|
||||
#endif
|
||||
return 0;
|
||||
}
|
||||
|
||||
int nf_reroute(struct sk_buff *skb, struct nf_queue_entry *entry)
|
||||
{
|
||||
const struct nf_ipv6_ops *v6ops;
|
||||
int ret = 0;
|
||||
|
||||
switch (entry->state.pf) {
|
||||
case AF_INET:
|
||||
ret = nf_ip_reroute(skb, entry);
|
||||
break;
|
||||
case AF_INET6:
|
||||
v6ops = rcu_dereference(nf_ipv6_ops);
|
||||
if (v6ops)
|
||||
ret = v6ops->reroute(skb, entry);
|
||||
break;
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
/* Only get and check the lengths, not do any hop-by-hop stuff. */
|
||||
int nf_ip6_check_hbh_len(struct sk_buff *skb, u32 *plen)
|
||||
{
|
||||
|
|
|
@ -1142,7 +1142,8 @@ void xt_compat_target_from_user(struct xt_entry_target *t, void **dstptr,
|
|||
if (target->compat_from_user)
|
||||
target->compat_from_user(t->data, ct->data);
|
||||
else
|
||||
memcpy(t->data, ct->data, tsize - sizeof(*ct));
|
||||
unsafe_memcpy(t->data, ct->data, tsize - sizeof(*ct),
|
||||
/* UAPI 0-sized destination */);
|
||||
|
||||
tsize += off;
|
||||
t->u.user.target_size = tsize;
|
||||
|
|
Loading…
Reference in New Issue