mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-09-15 23:25:07 +00:00
954d82979b
Replace the existing /* fall through */ comments and its variants with the new pseudo-keyword macro fallthrough[1]. Also, remove unnecessary fall-through markings when it is the case. [1] https://www.kernel.org/doc/html/latest/process/deprecated.html?highlight=fallthrough#implicit-switch-case-fall-through Signed-off-by: Gustavo A. R. Silva <gustavoars@kernel.org> Signed-off-by: Pablo Neira Ayuso <pablo@netfilter.org>
705 lines
18 KiB
C
705 lines
18 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright (c) 2008-2009 Patrick McHardy <kaber@trash.net>
|
|
* Copyright (c) 2016 Pablo Neira Ayuso <pablo@netfilter.org>
|
|
*
|
|
* Development of this code funded by Astaro AG (http://www.astaro.com/)
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/if_vlan.h>
|
|
#include <linux/init.h>
|
|
#include <linux/module.h>
|
|
#include <linux/netlink.h>
|
|
#include <linux/netfilter.h>
|
|
#include <linux/netfilter/nf_tables.h>
|
|
#include <net/netfilter/nf_tables_core.h>
|
|
#include <net/netfilter/nf_tables.h>
|
|
#include <net/netfilter/nf_tables_offload.h>
|
|
/* For layer 4 checksum field offset. */
|
|
#include <linux/tcp.h>
|
|
#include <linux/udp.h>
|
|
#include <linux/icmpv6.h>
|
|
#include <linux/ip.h>
|
|
#include <linux/ipv6.h>
|
|
|
|
static bool nft_payload_rebuild_vlan_hdr(const struct sk_buff *skb, int mac_off,
|
|
struct vlan_ethhdr *veth)
|
|
{
|
|
if (skb_copy_bits(skb, mac_off, veth, ETH_HLEN))
|
|
return false;
|
|
|
|
veth->h_vlan_proto = skb->vlan_proto;
|
|
veth->h_vlan_TCI = htons(skb_vlan_tag_get(skb));
|
|
veth->h_vlan_encapsulated_proto = skb->protocol;
|
|
|
|
return true;
|
|
}
|
|
|
|
/* add vlan header into the user buffer for if tag was removed by offloads */
|
|
static bool
|
|
nft_payload_copy_vlan(u32 *d, const struct sk_buff *skb, u8 offset, u8 len)
|
|
{
|
|
int mac_off = skb_mac_header(skb) - skb->data;
|
|
u8 *vlanh, *dst_u8 = (u8 *) d;
|
|
struct vlan_ethhdr veth;
|
|
u8 vlan_hlen = 0;
|
|
|
|
if ((skb->protocol == htons(ETH_P_8021AD) ||
|
|
skb->protocol == htons(ETH_P_8021Q)) &&
|
|
offset >= VLAN_ETH_HLEN && offset < VLAN_ETH_HLEN + VLAN_HLEN)
|
|
vlan_hlen += VLAN_HLEN;
|
|
|
|
vlanh = (u8 *) &veth;
|
|
if (offset < VLAN_ETH_HLEN + vlan_hlen) {
|
|
u8 ethlen = len;
|
|
|
|
if (vlan_hlen &&
|
|
skb_copy_bits(skb, mac_off, &veth, VLAN_ETH_HLEN) < 0)
|
|
return false;
|
|
else if (!nft_payload_rebuild_vlan_hdr(skb, mac_off, &veth))
|
|
return false;
|
|
|
|
if (offset + len > VLAN_ETH_HLEN + vlan_hlen)
|
|
ethlen -= offset + len - VLAN_ETH_HLEN + vlan_hlen;
|
|
|
|
memcpy(dst_u8, vlanh + offset - vlan_hlen, ethlen);
|
|
|
|
len -= ethlen;
|
|
if (len == 0)
|
|
return true;
|
|
|
|
dst_u8 += ethlen;
|
|
offset = ETH_HLEN + vlan_hlen;
|
|
} else {
|
|
offset -= VLAN_HLEN + vlan_hlen;
|
|
}
|
|
|
|
return skb_copy_bits(skb, offset + mac_off, dst_u8, len) == 0;
|
|
}
|
|
|
|
void nft_payload_eval(const struct nft_expr *expr,
|
|
struct nft_regs *regs,
|
|
const struct nft_pktinfo *pkt)
|
|
{
|
|
const struct nft_payload *priv = nft_expr_priv(expr);
|
|
const struct sk_buff *skb = pkt->skb;
|
|
u32 *dest = ®s->data[priv->dreg];
|
|
int offset;
|
|
|
|
dest[priv->len / NFT_REG32_SIZE] = 0;
|
|
switch (priv->base) {
|
|
case NFT_PAYLOAD_LL_HEADER:
|
|
if (!skb_mac_header_was_set(skb))
|
|
goto err;
|
|
|
|
if (skb_vlan_tag_present(skb)) {
|
|
if (!nft_payload_copy_vlan(dest, skb,
|
|
priv->offset, priv->len))
|
|
goto err;
|
|
return;
|
|
}
|
|
offset = skb_mac_header(skb) - skb->data;
|
|
break;
|
|
case NFT_PAYLOAD_NETWORK_HEADER:
|
|
offset = skb_network_offset(skb);
|
|
break;
|
|
case NFT_PAYLOAD_TRANSPORT_HEADER:
|
|
if (!pkt->tprot_set)
|
|
goto err;
|
|
offset = pkt->xt.thoff;
|
|
break;
|
|
default:
|
|
BUG();
|
|
}
|
|
offset += priv->offset;
|
|
|
|
if (skb_copy_bits(skb, offset, dest, priv->len) < 0)
|
|
goto err;
|
|
return;
|
|
err:
|
|
regs->verdict.code = NFT_BREAK;
|
|
}
|
|
|
|
static const struct nla_policy nft_payload_policy[NFTA_PAYLOAD_MAX + 1] = {
|
|
[NFTA_PAYLOAD_SREG] = { .type = NLA_U32 },
|
|
[NFTA_PAYLOAD_DREG] = { .type = NLA_U32 },
|
|
[NFTA_PAYLOAD_BASE] = { .type = NLA_U32 },
|
|
[NFTA_PAYLOAD_OFFSET] = { .type = NLA_U32 },
|
|
[NFTA_PAYLOAD_LEN] = { .type = NLA_U32 },
|
|
[NFTA_PAYLOAD_CSUM_TYPE] = { .type = NLA_U32 },
|
|
[NFTA_PAYLOAD_CSUM_OFFSET] = { .type = NLA_U32 },
|
|
[NFTA_PAYLOAD_CSUM_FLAGS] = { .type = NLA_U32 },
|
|
};
|
|
|
|
static int nft_payload_init(const struct nft_ctx *ctx,
|
|
const struct nft_expr *expr,
|
|
const struct nlattr * const tb[])
|
|
{
|
|
struct nft_payload *priv = nft_expr_priv(expr);
|
|
|
|
priv->base = ntohl(nla_get_be32(tb[NFTA_PAYLOAD_BASE]));
|
|
priv->offset = ntohl(nla_get_be32(tb[NFTA_PAYLOAD_OFFSET]));
|
|
priv->len = ntohl(nla_get_be32(tb[NFTA_PAYLOAD_LEN]));
|
|
priv->dreg = nft_parse_register(tb[NFTA_PAYLOAD_DREG]);
|
|
|
|
return nft_validate_register_store(ctx, priv->dreg, NULL,
|
|
NFT_DATA_VALUE, priv->len);
|
|
}
|
|
|
|
static int nft_payload_dump(struct sk_buff *skb, const struct nft_expr *expr)
|
|
{
|
|
const struct nft_payload *priv = nft_expr_priv(expr);
|
|
|
|
if (nft_dump_register(skb, NFTA_PAYLOAD_DREG, priv->dreg) ||
|
|
nla_put_be32(skb, NFTA_PAYLOAD_BASE, htonl(priv->base)) ||
|
|
nla_put_be32(skb, NFTA_PAYLOAD_OFFSET, htonl(priv->offset)) ||
|
|
nla_put_be32(skb, NFTA_PAYLOAD_LEN, htonl(priv->len)))
|
|
goto nla_put_failure;
|
|
return 0;
|
|
|
|
nla_put_failure:
|
|
return -1;
|
|
}
|
|
|
|
static int nft_payload_offload_ll(struct nft_offload_ctx *ctx,
|
|
struct nft_flow_rule *flow,
|
|
const struct nft_payload *priv)
|
|
{
|
|
struct nft_offload_reg *reg = &ctx->regs[priv->dreg];
|
|
|
|
switch (priv->offset) {
|
|
case offsetof(struct ethhdr, h_source):
|
|
if (priv->len != ETH_ALEN)
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_ETH_ADDRS, eth_addrs,
|
|
src, ETH_ALEN, reg);
|
|
break;
|
|
case offsetof(struct ethhdr, h_dest):
|
|
if (priv->len != ETH_ALEN)
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_ETH_ADDRS, eth_addrs,
|
|
dst, ETH_ALEN, reg);
|
|
break;
|
|
case offsetof(struct ethhdr, h_proto):
|
|
if (priv->len != sizeof(__be16))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_BASIC, basic,
|
|
n_proto, sizeof(__be16), reg);
|
|
nft_offload_set_dependency(ctx, NFT_OFFLOAD_DEP_NETWORK);
|
|
break;
|
|
case offsetof(struct vlan_ethhdr, h_vlan_TCI):
|
|
if (priv->len != sizeof(__be16))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_VLAN, vlan,
|
|
vlan_tci, sizeof(__be16), reg);
|
|
break;
|
|
case offsetof(struct vlan_ethhdr, h_vlan_encapsulated_proto):
|
|
if (priv->len != sizeof(__be16))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_VLAN, vlan,
|
|
vlan_tpid, sizeof(__be16), reg);
|
|
nft_offload_set_dependency(ctx, NFT_OFFLOAD_DEP_NETWORK);
|
|
break;
|
|
case offsetof(struct vlan_ethhdr, h_vlan_TCI) + sizeof(struct vlan_hdr):
|
|
if (priv->len != sizeof(__be16))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_CVLAN, vlan,
|
|
vlan_tci, sizeof(__be16), reg);
|
|
break;
|
|
case offsetof(struct vlan_ethhdr, h_vlan_encapsulated_proto) +
|
|
sizeof(struct vlan_hdr):
|
|
if (priv->len != sizeof(__be16))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_CVLAN, vlan,
|
|
vlan_tpid, sizeof(__be16), reg);
|
|
break;
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nft_payload_offload_ip(struct nft_offload_ctx *ctx,
|
|
struct nft_flow_rule *flow,
|
|
const struct nft_payload *priv)
|
|
{
|
|
struct nft_offload_reg *reg = &ctx->regs[priv->dreg];
|
|
|
|
switch (priv->offset) {
|
|
case offsetof(struct iphdr, saddr):
|
|
if (priv->len != sizeof(struct in_addr))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_IPV4_ADDRS, ipv4, src,
|
|
sizeof(struct in_addr), reg);
|
|
break;
|
|
case offsetof(struct iphdr, daddr):
|
|
if (priv->len != sizeof(struct in_addr))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_IPV4_ADDRS, ipv4, dst,
|
|
sizeof(struct in_addr), reg);
|
|
break;
|
|
case offsetof(struct iphdr, protocol):
|
|
if (priv->len != sizeof(__u8))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_BASIC, basic, ip_proto,
|
|
sizeof(__u8), reg);
|
|
nft_offload_set_dependency(ctx, NFT_OFFLOAD_DEP_TRANSPORT);
|
|
break;
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nft_payload_offload_ip6(struct nft_offload_ctx *ctx,
|
|
struct nft_flow_rule *flow,
|
|
const struct nft_payload *priv)
|
|
{
|
|
struct nft_offload_reg *reg = &ctx->regs[priv->dreg];
|
|
|
|
switch (priv->offset) {
|
|
case offsetof(struct ipv6hdr, saddr):
|
|
if (priv->len != sizeof(struct in6_addr))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_IPV6_ADDRS, ipv6, src,
|
|
sizeof(struct in6_addr), reg);
|
|
break;
|
|
case offsetof(struct ipv6hdr, daddr):
|
|
if (priv->len != sizeof(struct in6_addr))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_IPV6_ADDRS, ipv6, dst,
|
|
sizeof(struct in6_addr), reg);
|
|
break;
|
|
case offsetof(struct ipv6hdr, nexthdr):
|
|
if (priv->len != sizeof(__u8))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_BASIC, basic, ip_proto,
|
|
sizeof(__u8), reg);
|
|
nft_offload_set_dependency(ctx, NFT_OFFLOAD_DEP_TRANSPORT);
|
|
break;
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nft_payload_offload_nh(struct nft_offload_ctx *ctx,
|
|
struct nft_flow_rule *flow,
|
|
const struct nft_payload *priv)
|
|
{
|
|
int err;
|
|
|
|
switch (ctx->dep.l3num) {
|
|
case htons(ETH_P_IP):
|
|
err = nft_payload_offload_ip(ctx, flow, priv);
|
|
break;
|
|
case htons(ETH_P_IPV6):
|
|
err = nft_payload_offload_ip6(ctx, flow, priv);
|
|
break;
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nft_payload_offload_tcp(struct nft_offload_ctx *ctx,
|
|
struct nft_flow_rule *flow,
|
|
const struct nft_payload *priv)
|
|
{
|
|
struct nft_offload_reg *reg = &ctx->regs[priv->dreg];
|
|
|
|
switch (priv->offset) {
|
|
case offsetof(struct tcphdr, source):
|
|
if (priv->len != sizeof(__be16))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_PORTS, tp, src,
|
|
sizeof(__be16), reg);
|
|
break;
|
|
case offsetof(struct tcphdr, dest):
|
|
if (priv->len != sizeof(__be16))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_PORTS, tp, dst,
|
|
sizeof(__be16), reg);
|
|
break;
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nft_payload_offload_udp(struct nft_offload_ctx *ctx,
|
|
struct nft_flow_rule *flow,
|
|
const struct nft_payload *priv)
|
|
{
|
|
struct nft_offload_reg *reg = &ctx->regs[priv->dreg];
|
|
|
|
switch (priv->offset) {
|
|
case offsetof(struct udphdr, source):
|
|
if (priv->len != sizeof(__be16))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_PORTS, tp, src,
|
|
sizeof(__be16), reg);
|
|
break;
|
|
case offsetof(struct udphdr, dest):
|
|
if (priv->len != sizeof(__be16))
|
|
return -EOPNOTSUPP;
|
|
|
|
NFT_OFFLOAD_MATCH(FLOW_DISSECTOR_KEY_PORTS, tp, dst,
|
|
sizeof(__be16), reg);
|
|
break;
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nft_payload_offload_th(struct nft_offload_ctx *ctx,
|
|
struct nft_flow_rule *flow,
|
|
const struct nft_payload *priv)
|
|
{
|
|
int err;
|
|
|
|
switch (ctx->dep.protonum) {
|
|
case IPPROTO_TCP:
|
|
err = nft_payload_offload_tcp(ctx, flow, priv);
|
|
break;
|
|
case IPPROTO_UDP:
|
|
err = nft_payload_offload_udp(ctx, flow, priv);
|
|
break;
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static int nft_payload_offload(struct nft_offload_ctx *ctx,
|
|
struct nft_flow_rule *flow,
|
|
const struct nft_expr *expr)
|
|
{
|
|
const struct nft_payload *priv = nft_expr_priv(expr);
|
|
int err;
|
|
|
|
switch (priv->base) {
|
|
case NFT_PAYLOAD_LL_HEADER:
|
|
err = nft_payload_offload_ll(ctx, flow, priv);
|
|
break;
|
|
case NFT_PAYLOAD_NETWORK_HEADER:
|
|
err = nft_payload_offload_nh(ctx, flow, priv);
|
|
break;
|
|
case NFT_PAYLOAD_TRANSPORT_HEADER:
|
|
err = nft_payload_offload_th(ctx, flow, priv);
|
|
break;
|
|
default:
|
|
err = -EOPNOTSUPP;
|
|
break;
|
|
}
|
|
return err;
|
|
}
|
|
|
|
static const struct nft_expr_ops nft_payload_ops = {
|
|
.type = &nft_payload_type,
|
|
.size = NFT_EXPR_SIZE(sizeof(struct nft_payload)),
|
|
.eval = nft_payload_eval,
|
|
.init = nft_payload_init,
|
|
.dump = nft_payload_dump,
|
|
.offload = nft_payload_offload,
|
|
};
|
|
|
|
const struct nft_expr_ops nft_payload_fast_ops = {
|
|
.type = &nft_payload_type,
|
|
.size = NFT_EXPR_SIZE(sizeof(struct nft_payload)),
|
|
.eval = nft_payload_eval,
|
|
.init = nft_payload_init,
|
|
.dump = nft_payload_dump,
|
|
.offload = nft_payload_offload,
|
|
};
|
|
|
|
static inline void nft_csum_replace(__sum16 *sum, __wsum fsum, __wsum tsum)
|
|
{
|
|
*sum = csum_fold(csum_add(csum_sub(~csum_unfold(*sum), fsum), tsum));
|
|
if (*sum == 0)
|
|
*sum = CSUM_MANGLED_0;
|
|
}
|
|
|
|
static bool nft_payload_udp_checksum(struct sk_buff *skb, unsigned int thoff)
|
|
{
|
|
struct udphdr *uh, _uh;
|
|
|
|
uh = skb_header_pointer(skb, thoff, sizeof(_uh), &_uh);
|
|
if (!uh)
|
|
return false;
|
|
|
|
return (__force bool)uh->check;
|
|
}
|
|
|
|
static int nft_payload_l4csum_offset(const struct nft_pktinfo *pkt,
|
|
struct sk_buff *skb,
|
|
unsigned int *l4csum_offset)
|
|
{
|
|
switch (pkt->tprot) {
|
|
case IPPROTO_TCP:
|
|
*l4csum_offset = offsetof(struct tcphdr, check);
|
|
break;
|
|
case IPPROTO_UDP:
|
|
if (!nft_payload_udp_checksum(skb, pkt->xt.thoff))
|
|
return -1;
|
|
fallthrough;
|
|
case IPPROTO_UDPLITE:
|
|
*l4csum_offset = offsetof(struct udphdr, check);
|
|
break;
|
|
case IPPROTO_ICMPV6:
|
|
*l4csum_offset = offsetof(struct icmp6hdr, icmp6_cksum);
|
|
break;
|
|
default:
|
|
return -1;
|
|
}
|
|
|
|
*l4csum_offset += pkt->xt.thoff;
|
|
return 0;
|
|
}
|
|
|
|
static int nft_payload_l4csum_update(const struct nft_pktinfo *pkt,
|
|
struct sk_buff *skb,
|
|
__wsum fsum, __wsum tsum)
|
|
{
|
|
int l4csum_offset;
|
|
__sum16 sum;
|
|
|
|
/* If we cannot determine layer 4 checksum offset or this packet doesn't
|
|
* require layer 4 checksum recalculation, skip this packet.
|
|
*/
|
|
if (nft_payload_l4csum_offset(pkt, skb, &l4csum_offset) < 0)
|
|
return 0;
|
|
|
|
if (skb_copy_bits(skb, l4csum_offset, &sum, sizeof(sum)) < 0)
|
|
return -1;
|
|
|
|
/* Checksum mangling for an arbitrary amount of bytes, based on
|
|
* inet_proto_csum_replace*() functions.
|
|
*/
|
|
if (skb->ip_summed != CHECKSUM_PARTIAL) {
|
|
nft_csum_replace(&sum, fsum, tsum);
|
|
if (skb->ip_summed == CHECKSUM_COMPLETE) {
|
|
skb->csum = ~csum_add(csum_sub(~(skb->csum), fsum),
|
|
tsum);
|
|
}
|
|
} else {
|
|
sum = ~csum_fold(csum_add(csum_sub(csum_unfold(sum), fsum),
|
|
tsum));
|
|
}
|
|
|
|
if (skb_ensure_writable(skb, l4csum_offset + sizeof(sum)) ||
|
|
skb_store_bits(skb, l4csum_offset, &sum, sizeof(sum)) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int nft_payload_csum_inet(struct sk_buff *skb, const u32 *src,
|
|
__wsum fsum, __wsum tsum, int csum_offset)
|
|
{
|
|
__sum16 sum;
|
|
|
|
if (skb_copy_bits(skb, csum_offset, &sum, sizeof(sum)) < 0)
|
|
return -1;
|
|
|
|
nft_csum_replace(&sum, fsum, tsum);
|
|
if (skb_ensure_writable(skb, csum_offset + sizeof(sum)) ||
|
|
skb_store_bits(skb, csum_offset, &sum, sizeof(sum)) < 0)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void nft_payload_set_eval(const struct nft_expr *expr,
|
|
struct nft_regs *regs,
|
|
const struct nft_pktinfo *pkt)
|
|
{
|
|
const struct nft_payload_set *priv = nft_expr_priv(expr);
|
|
struct sk_buff *skb = pkt->skb;
|
|
const u32 *src = ®s->data[priv->sreg];
|
|
int offset, csum_offset;
|
|
__wsum fsum, tsum;
|
|
|
|
switch (priv->base) {
|
|
case NFT_PAYLOAD_LL_HEADER:
|
|
if (!skb_mac_header_was_set(skb))
|
|
goto err;
|
|
offset = skb_mac_header(skb) - skb->data;
|
|
break;
|
|
case NFT_PAYLOAD_NETWORK_HEADER:
|
|
offset = skb_network_offset(skb);
|
|
break;
|
|
case NFT_PAYLOAD_TRANSPORT_HEADER:
|
|
if (!pkt->tprot_set)
|
|
goto err;
|
|
offset = pkt->xt.thoff;
|
|
break;
|
|
default:
|
|
BUG();
|
|
}
|
|
|
|
csum_offset = offset + priv->csum_offset;
|
|
offset += priv->offset;
|
|
|
|
if ((priv->csum_type == NFT_PAYLOAD_CSUM_INET || priv->csum_flags) &&
|
|
(priv->base != NFT_PAYLOAD_TRANSPORT_HEADER ||
|
|
skb->ip_summed != CHECKSUM_PARTIAL)) {
|
|
fsum = skb_checksum(skb, offset, priv->len, 0);
|
|
tsum = csum_partial(src, priv->len, 0);
|
|
|
|
if (priv->csum_type == NFT_PAYLOAD_CSUM_INET &&
|
|
nft_payload_csum_inet(skb, src, fsum, tsum, csum_offset))
|
|
goto err;
|
|
|
|
if (priv->csum_flags &&
|
|
nft_payload_l4csum_update(pkt, skb, fsum, tsum) < 0)
|
|
goto err;
|
|
}
|
|
|
|
if (skb_ensure_writable(skb, max(offset + priv->len, 0)) ||
|
|
skb_store_bits(skb, offset, src, priv->len) < 0)
|
|
goto err;
|
|
|
|
return;
|
|
err:
|
|
regs->verdict.code = NFT_BREAK;
|
|
}
|
|
|
|
static int nft_payload_set_init(const struct nft_ctx *ctx,
|
|
const struct nft_expr *expr,
|
|
const struct nlattr * const tb[])
|
|
{
|
|
struct nft_payload_set *priv = nft_expr_priv(expr);
|
|
|
|
priv->base = ntohl(nla_get_be32(tb[NFTA_PAYLOAD_BASE]));
|
|
priv->offset = ntohl(nla_get_be32(tb[NFTA_PAYLOAD_OFFSET]));
|
|
priv->len = ntohl(nla_get_be32(tb[NFTA_PAYLOAD_LEN]));
|
|
priv->sreg = nft_parse_register(tb[NFTA_PAYLOAD_SREG]);
|
|
|
|
if (tb[NFTA_PAYLOAD_CSUM_TYPE])
|
|
priv->csum_type =
|
|
ntohl(nla_get_be32(tb[NFTA_PAYLOAD_CSUM_TYPE]));
|
|
if (tb[NFTA_PAYLOAD_CSUM_OFFSET])
|
|
priv->csum_offset =
|
|
ntohl(nla_get_be32(tb[NFTA_PAYLOAD_CSUM_OFFSET]));
|
|
if (tb[NFTA_PAYLOAD_CSUM_FLAGS]) {
|
|
u32 flags;
|
|
|
|
flags = ntohl(nla_get_be32(tb[NFTA_PAYLOAD_CSUM_FLAGS]));
|
|
if (flags & ~NFT_PAYLOAD_L4CSUM_PSEUDOHDR)
|
|
return -EINVAL;
|
|
|
|
priv->csum_flags = flags;
|
|
}
|
|
|
|
switch (priv->csum_type) {
|
|
case NFT_PAYLOAD_CSUM_NONE:
|
|
case NFT_PAYLOAD_CSUM_INET:
|
|
break;
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
return nft_validate_register_load(priv->sreg, priv->len);
|
|
}
|
|
|
|
static int nft_payload_set_dump(struct sk_buff *skb, const struct nft_expr *expr)
|
|
{
|
|
const struct nft_payload_set *priv = nft_expr_priv(expr);
|
|
|
|
if (nft_dump_register(skb, NFTA_PAYLOAD_SREG, priv->sreg) ||
|
|
nla_put_be32(skb, NFTA_PAYLOAD_BASE, htonl(priv->base)) ||
|
|
nla_put_be32(skb, NFTA_PAYLOAD_OFFSET, htonl(priv->offset)) ||
|
|
nla_put_be32(skb, NFTA_PAYLOAD_LEN, htonl(priv->len)) ||
|
|
nla_put_be32(skb, NFTA_PAYLOAD_CSUM_TYPE, htonl(priv->csum_type)) ||
|
|
nla_put_be32(skb, NFTA_PAYLOAD_CSUM_OFFSET,
|
|
htonl(priv->csum_offset)) ||
|
|
nla_put_be32(skb, NFTA_PAYLOAD_CSUM_FLAGS, htonl(priv->csum_flags)))
|
|
goto nla_put_failure;
|
|
return 0;
|
|
|
|
nla_put_failure:
|
|
return -1;
|
|
}
|
|
|
|
static const struct nft_expr_ops nft_payload_set_ops = {
|
|
.type = &nft_payload_type,
|
|
.size = NFT_EXPR_SIZE(sizeof(struct nft_payload_set)),
|
|
.eval = nft_payload_set_eval,
|
|
.init = nft_payload_set_init,
|
|
.dump = nft_payload_set_dump,
|
|
};
|
|
|
|
static const struct nft_expr_ops *
|
|
nft_payload_select_ops(const struct nft_ctx *ctx,
|
|
const struct nlattr * const tb[])
|
|
{
|
|
enum nft_payload_bases base;
|
|
unsigned int offset, len;
|
|
|
|
if (tb[NFTA_PAYLOAD_BASE] == NULL ||
|
|
tb[NFTA_PAYLOAD_OFFSET] == NULL ||
|
|
tb[NFTA_PAYLOAD_LEN] == NULL)
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
base = ntohl(nla_get_be32(tb[NFTA_PAYLOAD_BASE]));
|
|
switch (base) {
|
|
case NFT_PAYLOAD_LL_HEADER:
|
|
case NFT_PAYLOAD_NETWORK_HEADER:
|
|
case NFT_PAYLOAD_TRANSPORT_HEADER:
|
|
break;
|
|
default:
|
|
return ERR_PTR(-EOPNOTSUPP);
|
|
}
|
|
|
|
if (tb[NFTA_PAYLOAD_SREG] != NULL) {
|
|
if (tb[NFTA_PAYLOAD_DREG] != NULL)
|
|
return ERR_PTR(-EINVAL);
|
|
return &nft_payload_set_ops;
|
|
}
|
|
|
|
if (tb[NFTA_PAYLOAD_DREG] == NULL)
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
offset = ntohl(nla_get_be32(tb[NFTA_PAYLOAD_OFFSET]));
|
|
len = ntohl(nla_get_be32(tb[NFTA_PAYLOAD_LEN]));
|
|
|
|
if (len <= 4 && is_power_of_2(len) && IS_ALIGNED(offset, len) &&
|
|
base != NFT_PAYLOAD_LL_HEADER)
|
|
return &nft_payload_fast_ops;
|
|
else
|
|
return &nft_payload_ops;
|
|
}
|
|
|
|
struct nft_expr_type nft_payload_type __read_mostly = {
|
|
.name = "payload",
|
|
.select_ops = nft_payload_select_ops,
|
|
.policy = nft_payload_policy,
|
|
.maxattr = NFTA_PAYLOAD_MAX,
|
|
.owner = THIS_MODULE,
|
|
};
|