2019-05-23 09:14:41 +00:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-or-later
|
2008-01-11 14:57:09 +00:00
|
|
|
/* SCTP kernel implementation
|
2005-04-16 22:20:36 +00:00
|
|
|
* (C) Copyright IBM Corp. 2001, 2004
|
|
|
|
* Copyright (c) 1999-2000 Cisco, Inc.
|
|
|
|
* Copyright (c) 1999-2001 Motorola, Inc.
|
|
|
|
*
|
2008-01-11 14:57:09 +00:00
|
|
|
* This file is part of the SCTP kernel implementation
|
2005-04-16 22:20:36 +00:00
|
|
|
*
|
|
|
|
* These functions handle output processing.
|
|
|
|
*
|
|
|
|
* Please send any bug reports or fixes you make to the
|
|
|
|
* email address(es):
|
2013-07-23 12:51:47 +00:00
|
|
|
* lksctp developers <linux-sctp@vger.kernel.org>
|
2005-04-16 22:20:36 +00:00
|
|
|
*
|
|
|
|
* Written or modified by:
|
|
|
|
* La Monte H.P. Yarroll <piggy@acm.org>
|
|
|
|
* Karl Knutson <karl@athena.chicago.il.us>
|
|
|
|
* Jon Grimm <jgrimm@austin.ibm.com>
|
|
|
|
* Sridhar Samudrala <sri@us.ibm.com>
|
|
|
|
*/
|
|
|
|
|
2010-08-24 13:21:08 +00:00
|
|
|
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
#include <linux/types.h>
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/wait.h>
|
|
|
|
#include <linux/time.h>
|
|
|
|
#include <linux/ip.h>
|
|
|
|
#include <linux/ipv6.h>
|
|
|
|
#include <linux/init.h>
|
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 08:04:11 +00:00
|
|
|
#include <linux/slab.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
#include <net/inet_ecn.h>
|
2009-03-21 20:41:09 +00:00
|
|
|
#include <net/ip.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
#include <net/icmp.h>
|
2008-07-17 03:20:11 +00:00
|
|
|
#include <net/net_namespace.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
#include <linux/socket.h> /* for sa_family_t */
|
|
|
|
#include <net/sock.h>
|
|
|
|
|
|
|
|
#include <net/sctp/sctp.h>
|
|
|
|
#include <net/sctp/sm.h>
|
2007-12-16 22:06:41 +00:00
|
|
|
#include <net/sctp/checksum.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* Forward declarations for private helpers. */
|
2017-08-05 11:59:57 +00:00
|
|
|
static enum sctp_xmit __sctp_packet_append_chunk(struct sctp_packet *packet,
|
|
|
|
struct sctp_chunk *chunk);
|
|
|
|
static enum sctp_xmit sctp_packet_can_append_data(struct sctp_packet *packet,
|
|
|
|
struct sctp_chunk *chunk);
|
2009-08-07 14:43:07 +00:00
|
|
|
static void sctp_packet_append_data(struct sctp_packet *packet,
|
2017-08-05 11:59:57 +00:00
|
|
|
struct sctp_chunk *chunk);
|
|
|
|
static enum sctp_xmit sctp_packet_will_fit(struct sctp_packet *packet,
|
|
|
|
struct sctp_chunk *chunk,
|
|
|
|
u16 chunk_len);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-09-04 06:34:06 +00:00
|
|
|
static void sctp_packet_reset(struct sctp_packet *packet)
|
|
|
|
{
|
2018-02-24 16:18:51 +00:00
|
|
|
/* sctp_packet_transmit() relies on this to reset size to the
|
|
|
|
* current overhead after sending packets.
|
|
|
|
*/
|
2009-09-04 06:34:06 +00:00
|
|
|
packet->size = packet->overhead;
|
2018-02-24 16:18:51 +00:00
|
|
|
|
2009-09-04 06:34:06 +00:00
|
|
|
packet->has_cookie_echo = 0;
|
|
|
|
packet->has_sack = 0;
|
|
|
|
packet->has_data = 0;
|
|
|
|
packet->has_auth = 0;
|
|
|
|
packet->ipfragok = 0;
|
|
|
|
packet->auth = NULL;
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/* Config a packet.
|
|
|
|
* This appears to be a followup set of initializations.
|
|
|
|
*/
|
2016-12-28 11:26:34 +00:00
|
|
|
void sctp_packet_config(struct sctp_packet *packet, __u32 vtag,
|
|
|
|
int ecn_capable)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2016-06-02 18:05:43 +00:00
|
|
|
struct sctp_transport *tp = packet->transport;
|
|
|
|
struct sctp_association *asoc = tp->asoc;
|
2018-04-26 19:58:54 +00:00
|
|
|
struct sctp_sock *sp = NULL;
|
2017-04-01 09:15:59 +00:00
|
|
|
struct sock *sk;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
net: sctp: rework debugging framework to use pr_debug and friends
We should get rid of all own SCTP debug printk macros and use the ones
that the kernel offers anyway instead. This makes the code more readable
and conform to the kernel code, and offers all the features of dynamic
debbuging that pr_debug() et al has, such as only turning on/off portions
of debug messages at runtime through debugfs. The runtime cost of having
CONFIG_DYNAMIC_DEBUG enabled, but none of the debug statements printing,
is negligible [1]. If kernel debugging is completly turned off, then these
statements will also compile into "empty" functions.
While we're at it, we also need to change the Kconfig option as it /now/
only refers to the ifdef'ed code portions in outqueue.c that enable further
debugging/tracing of SCTP transaction fields. Also, since SCTP_ASSERT code
was enabled with this Kconfig option and has now been removed, we
transform those code parts into WARNs resp. where appropriate BUG_ONs so
that those bugs can be more easily detected as probably not many people
have SCTP debugging permanently turned on.
To turn on all SCTP debugging, the following steps are needed:
# mount -t debugfs none /sys/kernel/debug
# echo -n 'module sctp +p' > /sys/kernel/debug/dynamic_debug/control
This can be done more fine-grained on a per file, per line basis and others
as described in [2].
[1] https://www.kernel.org/doc/ols/2009/ols2009-pages-39-46.pdf
[2] Documentation/dynamic-debug-howto.txt
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2013-06-28 17:49:40 +00:00
|
|
|
pr_debug("%s: packet:%p vtag:0x%x\n", __func__, packet, vtag);
|
2005-04-16 22:20:36 +00:00
|
|
|
packet->vtag = vtag;
|
|
|
|
|
2017-04-01 09:15:59 +00:00
|
|
|
/* do the following jobs only once for a flush schedule */
|
|
|
|
if (!sctp_packet_empty(packet))
|
|
|
|
return;
|
2016-06-02 18:05:43 +00:00
|
|
|
|
2018-02-24 16:18:51 +00:00
|
|
|
/* set packet max_size with pathmtu, then calculate overhead */
|
2017-04-01 09:15:59 +00:00
|
|
|
packet->max_size = tp->pathmtu;
|
2018-04-26 19:58:54 +00:00
|
|
|
|
2018-02-24 16:18:51 +00:00
|
|
|
if (asoc) {
|
2018-04-26 19:58:54 +00:00
|
|
|
sk = asoc->base.sk;
|
|
|
|
sp = sctp_sk(sk);
|
|
|
|
}
|
|
|
|
packet->overhead = sctp_mtu_payload(sp, 0, 0);
|
|
|
|
packet->size = packet->overhead;
|
2018-02-24 16:18:51 +00:00
|
|
|
|
2018-04-26 19:58:54 +00:00
|
|
|
if (!asoc)
|
2017-04-01 09:15:59 +00:00
|
|
|
return;
|
2016-06-02 18:05:43 +00:00
|
|
|
|
2017-04-01 09:15:59 +00:00
|
|
|
/* update dst or transport pathmtu if in need */
|
|
|
|
if (!sctp_transport_dst_check(tp)) {
|
2018-04-26 19:58:54 +00:00
|
|
|
sctp_transport_route(tp, NULL, sp);
|
2017-04-01 09:15:59 +00:00
|
|
|
if (asoc->param_flags & SPP_PMTUD_ENABLE)
|
2017-04-04 05:39:55 +00:00
|
|
|
sctp_assoc_sync_pmtu(asoc);
|
2021-06-22 18:04:56 +00:00
|
|
|
} else if (!sctp_transport_pl_enabled(tp) &&
|
2021-07-21 18:45:54 +00:00
|
|
|
asoc->param_flags & SPP_PMTUD_ENABLE) {
|
|
|
|
if (!sctp_transport_pmtu_check(tp))
|
2018-11-18 08:14:47 +00:00
|
|
|
sctp_assoc_sync_pmtu(asoc);
|
2016-06-02 18:05:43 +00:00
|
|
|
}
|
|
|
|
|
2018-10-15 11:58:29 +00:00
|
|
|
if (asoc->pmtu_pending) {
|
|
|
|
if (asoc->param_flags & SPP_PMTUD_ENABLE)
|
|
|
|
sctp_assoc_sync_pmtu(asoc);
|
|
|
|
asoc->pmtu_pending = 0;
|
|
|
|
}
|
|
|
|
|
2017-04-01 09:15:59 +00:00
|
|
|
/* If there a is a prepend chunk stick it on the list before
|
|
|
|
* any other chunks get appended.
|
|
|
|
*/
|
|
|
|
if (ecn_capable) {
|
|
|
|
struct sctp_chunk *chunk = sctp_get_ecne_prepend(asoc);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
if (chunk)
|
|
|
|
sctp_packet_append_chunk(packet, chunk);
|
|
|
|
}
|
2017-04-01 09:15:59 +00:00
|
|
|
|
|
|
|
if (!tp->dst)
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* set packet max_size with gso_max_size if gso is enabled*/
|
|
|
|
rcu_read_lock();
|
|
|
|
if (__sk_dst_get(sk) != tp->dst) {
|
|
|
|
dst_hold(tp->dst);
|
|
|
|
sk_setup_caps(sk, tp->dst);
|
|
|
|
}
|
2022-05-13 18:33:57 +00:00
|
|
|
packet->max_size = sk_can_gso(sk) ? min(READ_ONCE(tp->dst->dev->gso_max_size),
|
|
|
|
GSO_LEGACY_MAX_SIZE)
|
2017-04-01 09:15:59 +00:00
|
|
|
: asoc->pathmtu;
|
|
|
|
rcu_read_unlock();
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Initialize the packet structure. */
|
2016-12-28 11:26:34 +00:00
|
|
|
void sctp_packet_init(struct sctp_packet *packet,
|
|
|
|
struct sctp_transport *transport,
|
|
|
|
__u16 sport, __u16 dport)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
net: sctp: rework debugging framework to use pr_debug and friends
We should get rid of all own SCTP debug printk macros and use the ones
that the kernel offers anyway instead. This makes the code more readable
and conform to the kernel code, and offers all the features of dynamic
debbuging that pr_debug() et al has, such as only turning on/off portions
of debug messages at runtime through debugfs. The runtime cost of having
CONFIG_DYNAMIC_DEBUG enabled, but none of the debug statements printing,
is negligible [1]. If kernel debugging is completly turned off, then these
statements will also compile into "empty" functions.
While we're at it, we also need to change the Kconfig option as it /now/
only refers to the ifdef'ed code portions in outqueue.c that enable further
debugging/tracing of SCTP transaction fields. Also, since SCTP_ASSERT code
was enabled with this Kconfig option and has now been removed, we
transform those code parts into WARNs resp. where appropriate BUG_ONs so
that those bugs can be more easily detected as probably not many people
have SCTP debugging permanently turned on.
To turn on all SCTP debugging, the following steps are needed:
# mount -t debugfs none /sys/kernel/debug
# echo -n 'module sctp +p' > /sys/kernel/debug/dynamic_debug/control
This can be done more fine-grained on a per file, per line basis and others
as described in [2].
[1] https://www.kernel.org/doc/ols/2009/ols2009-pages-39-46.pdf
[2] Documentation/dynamic-debug-howto.txt
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2013-06-28 17:49:40 +00:00
|
|
|
pr_debug("%s: packet:%p transport:%p\n", __func__, packet, transport);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
packet->transport = transport;
|
|
|
|
packet->source_port = sport;
|
|
|
|
packet->destination_port = dport;
|
2005-07-09 04:47:49 +00:00
|
|
|
INIT_LIST_HEAD(&packet->chunk_list);
|
2018-02-24 16:18:51 +00:00
|
|
|
/* The overhead will be calculated by sctp_packet_config() */
|
|
|
|
packet->overhead = 0;
|
2009-09-04 06:34:06 +00:00
|
|
|
sctp_packet_reset(packet);
|
2005-04-16 22:20:36 +00:00
|
|
|
packet->vtag = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Free a packet. */
|
|
|
|
void sctp_packet_free(struct sctp_packet *packet)
|
|
|
|
{
|
2005-07-09 04:47:49 +00:00
|
|
|
struct sctp_chunk *chunk, *tmp;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
net: sctp: rework debugging framework to use pr_debug and friends
We should get rid of all own SCTP debug printk macros and use the ones
that the kernel offers anyway instead. This makes the code more readable
and conform to the kernel code, and offers all the features of dynamic
debbuging that pr_debug() et al has, such as only turning on/off portions
of debug messages at runtime through debugfs. The runtime cost of having
CONFIG_DYNAMIC_DEBUG enabled, but none of the debug statements printing,
is negligible [1]. If kernel debugging is completly turned off, then these
statements will also compile into "empty" functions.
While we're at it, we also need to change the Kconfig option as it /now/
only refers to the ifdef'ed code portions in outqueue.c that enable further
debugging/tracing of SCTP transaction fields. Also, since SCTP_ASSERT code
was enabled with this Kconfig option and has now been removed, we
transform those code parts into WARNs resp. where appropriate BUG_ONs so
that those bugs can be more easily detected as probably not many people
have SCTP debugging permanently turned on.
To turn on all SCTP debugging, the following steps are needed:
# mount -t debugfs none /sys/kernel/debug
# echo -n 'module sctp +p' > /sys/kernel/debug/dynamic_debug/control
This can be done more fine-grained on a per file, per line basis and others
as described in [2].
[1] https://www.kernel.org/doc/ols/2009/ols2009-pages-39-46.pdf
[2] Documentation/dynamic-debug-howto.txt
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2013-06-28 17:49:40 +00:00
|
|
|
pr_debug("%s: packet:%p\n", __func__, packet);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2005-07-09 04:47:49 +00:00
|
|
|
list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) {
|
|
|
|
list_del_init(&chunk->list);
|
2005-04-16 22:20:36 +00:00
|
|
|
sctp_chunk_free(chunk);
|
2005-07-09 04:47:49 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* This routine tries to append the chunk to the offered packet. If adding
|
|
|
|
* the chunk causes the packet to exceed the path MTU and COOKIE_ECHO chunk
|
|
|
|
* is not present in the packet, it transmits the input packet.
|
|
|
|
* Data can be bundled with a packet containing a COOKIE_ECHO chunk as long
|
|
|
|
* as it can fit in the packet, but any more data that does not fit in this
|
|
|
|
* packet can be sent only after receiving the COOKIE_ACK.
|
|
|
|
*/
|
2017-08-05 11:59:57 +00:00
|
|
|
enum sctp_xmit sctp_packet_transmit_chunk(struct sctp_packet *packet,
|
|
|
|
struct sctp_chunk *chunk,
|
|
|
|
int one_packet, gfp_t gfp)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2017-08-05 11:59:57 +00:00
|
|
|
enum sctp_xmit retval;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2017-02-27 22:30:02 +00:00
|
|
|
pr_debug("%s: packet:%p size:%zu chunk:%p size:%d\n", __func__,
|
2016-06-02 18:05:44 +00:00
|
|
|
packet, packet->size, chunk, chunk->skb ? chunk->skb->len : -1);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
switch ((retval = (sctp_packet_append_chunk(packet, chunk)))) {
|
|
|
|
case SCTP_XMIT_PMTU_FULL:
|
|
|
|
if (!packet->has_cookie_echo) {
|
sctp: save transmit error to sk_err in sctp_outq_flush
Every time when sctp calls sctp_outq_flush, it sends out the chunks of
control queue, retransmit queue and data queue. Even if some trunks are
failed to transmit, it still has to flush all the transports, as it's
the only chance to clean that transmit_list.
So the latest transmit error here should be returned back. This transmit
error is an internal error of sctp stack.
I checked all the places where it uses the transmit error (the return
value of sctp_outq_flush), most of them are actually just save it to
sk_err.
Except for sctp_assoc/endpoint_bh_rcv, they will drop the chunk if
it's failed to send a REPLY, which is actually incorrect, as we can't
be sure the error that sctp_outq_flush returns is from sending that
REPLY.
So it's meaningless for sctp_outq_flush to return error back.
This patch is to save transmit error to sk_err in sctp_outq_flush, the
new error can update the old value. Eventually, sctp_wait_for_* would
check for it.
Signed-off-by: Xin Long <lucien.xin@gmail.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-09-13 18:04:21 +00:00
|
|
|
int error = 0;
|
|
|
|
|
2016-03-10 21:33:07 +00:00
|
|
|
error = sctp_packet_transmit(packet, gfp);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (error < 0)
|
|
|
|
chunk->skb->sk->sk_err = -error;
|
|
|
|
|
|
|
|
/* If we have an empty packet, then we can NOT ever
|
|
|
|
* return PMTU_FULL.
|
|
|
|
*/
|
2008-06-19 23:08:18 +00:00
|
|
|
if (!one_packet)
|
|
|
|
retval = sctp_packet_append_chunk(packet,
|
|
|
|
chunk);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
case SCTP_XMIT_RWND_FULL:
|
|
|
|
case SCTP_XMIT_OK:
|
2014-07-22 08:59:14 +00:00
|
|
|
case SCTP_XMIT_DELAY:
|
2005-04-16 22:20:36 +00:00
|
|
|
break;
|
2007-04-21 00:09:22 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
2021-06-22 18:04:52 +00:00
|
|
|
/* Try to bundle a pad chunk into a packet with a heartbeat chunk for PLPMTUTD probe */
|
|
|
|
static enum sctp_xmit sctp_packet_bundle_pad(struct sctp_packet *pkt, struct sctp_chunk *chunk)
|
|
|
|
{
|
|
|
|
struct sctp_transport *t = pkt->transport;
|
|
|
|
struct sctp_chunk *pad;
|
|
|
|
int overhead = 0;
|
|
|
|
|
|
|
|
if (!chunk->pmtu_probe)
|
|
|
|
return SCTP_XMIT_OK;
|
|
|
|
|
|
|
|
/* calculate the Padding Data size for the pad chunk */
|
|
|
|
overhead += sizeof(struct sctphdr) + sizeof(struct sctp_chunkhdr);
|
|
|
|
overhead += sizeof(struct sctp_sender_hb_info) + sizeof(struct sctp_pad_chunk);
|
|
|
|
pad = sctp_make_pad(t->asoc, t->pl.probe_size - overhead);
|
|
|
|
if (!pad)
|
|
|
|
return SCTP_XMIT_DELAY;
|
|
|
|
|
|
|
|
list_add_tail(&pad->list, &pkt->chunk_list);
|
|
|
|
pkt->size += SCTP_PAD4(ntohs(pad->chunk_hdr->length));
|
|
|
|
chunk->transport = t;
|
|
|
|
|
|
|
|
return SCTP_XMIT_OK;
|
|
|
|
}
|
|
|
|
|
2007-09-17 02:32:45 +00:00
|
|
|
/* Try to bundle an auth chunk into the packet. */
|
2017-08-05 11:59:57 +00:00
|
|
|
static enum sctp_xmit sctp_packet_bundle_auth(struct sctp_packet *pkt,
|
|
|
|
struct sctp_chunk *chunk)
|
2007-09-17 02:32:45 +00:00
|
|
|
{
|
|
|
|
struct sctp_association *asoc = pkt->transport->asoc;
|
2017-08-05 11:59:57 +00:00
|
|
|
enum sctp_xmit retval = SCTP_XMIT_OK;
|
2007-09-17 02:32:45 +00:00
|
|
|
struct sctp_chunk *auth;
|
|
|
|
|
|
|
|
/* if we don't have an association, we can't do authentication */
|
|
|
|
if (!asoc)
|
|
|
|
return retval;
|
|
|
|
|
|
|
|
/* See if this is an auth chunk we are bundling or if
|
|
|
|
* auth is already bundled.
|
|
|
|
*/
|
2009-09-04 22:21:00 +00:00
|
|
|
if (chunk->chunk_hdr->type == SCTP_CID_AUTH || pkt->has_auth)
|
2007-09-17 02:32:45 +00:00
|
|
|
return retval;
|
|
|
|
|
|
|
|
/* if the peer did not request this chunk to be authenticated,
|
|
|
|
* don't do it
|
|
|
|
*/
|
|
|
|
if (!chunk->auth)
|
|
|
|
return retval;
|
|
|
|
|
2018-03-14 11:05:30 +00:00
|
|
|
auth = sctp_make_auth(asoc, chunk->shkey->key_id);
|
2007-09-17 02:32:45 +00:00
|
|
|
if (!auth)
|
|
|
|
return retval;
|
|
|
|
|
2018-03-14 11:05:30 +00:00
|
|
|
auth->shkey = chunk->shkey;
|
|
|
|
sctp_auth_shkey_hold(auth->shkey);
|
|
|
|
|
2012-07-02 09:59:24 +00:00
|
|
|
retval = __sctp_packet_append_chunk(pkt, auth);
|
|
|
|
|
|
|
|
if (retval != SCTP_XMIT_OK)
|
|
|
|
sctp_chunk_free(auth);
|
2007-09-17 02:32:45 +00:00
|
|
|
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/* Try to bundle a SACK with the packet. */
|
2017-08-05 11:59:57 +00:00
|
|
|
static enum sctp_xmit sctp_packet_bundle_sack(struct sctp_packet *pkt,
|
|
|
|
struct sctp_chunk *chunk)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2017-08-05 11:59:57 +00:00
|
|
|
enum sctp_xmit retval = SCTP_XMIT_OK;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* If sending DATA and haven't aleady bundled a SACK, try to
|
|
|
|
* bundle one in to the packet.
|
|
|
|
*/
|
|
|
|
if (sctp_chunk_is_data(chunk) && !pkt->has_sack &&
|
|
|
|
!pkt->has_cookie_echo) {
|
|
|
|
struct sctp_association *asoc;
|
sctp: Fix piggybacked ACKs
This patch corrects the conditions under which a SACK will be piggybacked
on a DATA packet. The previous condition was incorrect due to a
misinterpretation of RFC 4960 and/or RFC 2960. Specifically, the
following paragraph from section 6.2 had not been implemented correctly:
Before an endpoint transmits a DATA chunk, if any received DATA
chunks have not been acknowledged (e.g., due to delayed ack), the
sender should create a SACK and bundle it with the outbound DATA
chunk, as long as the size of the final SCTP packet does not exceed
the current MTU. See Section 6.2.
When about to send a DATA chunk, the code now checks to see if the SACK
timer is running. If it is, we know we have a SACK to send to the
peer, so we append the SACK (assuming available space in the packet)
and turn off the timer. For a simple request-response scenario, this
will result in the SACK being bundled with the response, meaning the
the SACK is received quickly by the client, and also meaning that no
separate SACK packet needs to be sent by the server to acknowledge the
request. Prior to this patch, a separate SACK packet would have been
sent by the server SCTP only after its delayed-ACK timer had expired
(usually 200ms). This is wasteful of bandwidth, and can also have a
major negative impact on performance due the interaction of delayed ACKs
with the Nagle algorithm.
Signed-off-by: Doug Graham <dgraham@nortel.com>
Signed-off-by: Vlad Yasevich <vladislav.yasevich@hp.com>
2009-07-29 16:05:57 +00:00
|
|
|
struct timer_list *timer;
|
2005-04-16 22:20:36 +00:00
|
|
|
asoc = pkt->transport->asoc;
|
sctp: Fix piggybacked ACKs
This patch corrects the conditions under which a SACK will be piggybacked
on a DATA packet. The previous condition was incorrect due to a
misinterpretation of RFC 4960 and/or RFC 2960. Specifically, the
following paragraph from section 6.2 had not been implemented correctly:
Before an endpoint transmits a DATA chunk, if any received DATA
chunks have not been acknowledged (e.g., due to delayed ack), the
sender should create a SACK and bundle it with the outbound DATA
chunk, as long as the size of the final SCTP packet does not exceed
the current MTU. See Section 6.2.
When about to send a DATA chunk, the code now checks to see if the SACK
timer is running. If it is, we know we have a SACK to send to the
peer, so we append the SACK (assuming available space in the packet)
and turn off the timer. For a simple request-response scenario, this
will result in the SACK being bundled with the response, meaning the
the SACK is received quickly by the client, and also meaning that no
separate SACK packet needs to be sent by the server to acknowledge the
request. Prior to this patch, a separate SACK packet would have been
sent by the server SCTP only after its delayed-ACK timer had expired
(usually 200ms). This is wasteful of bandwidth, and can also have a
major negative impact on performance due the interaction of delayed ACKs
with the Nagle algorithm.
Signed-off-by: Doug Graham <dgraham@nortel.com>
Signed-off-by: Vlad Yasevich <vladislav.yasevich@hp.com>
2009-07-29 16:05:57 +00:00
|
|
|
timer = &asoc->timers[SCTP_EVENT_TIMEOUT_SACK];
|
2005-04-16 22:20:36 +00:00
|
|
|
|
sctp: Fix piggybacked ACKs
This patch corrects the conditions under which a SACK will be piggybacked
on a DATA packet. The previous condition was incorrect due to a
misinterpretation of RFC 4960 and/or RFC 2960. Specifically, the
following paragraph from section 6.2 had not been implemented correctly:
Before an endpoint transmits a DATA chunk, if any received DATA
chunks have not been acknowledged (e.g., due to delayed ack), the
sender should create a SACK and bundle it with the outbound DATA
chunk, as long as the size of the final SCTP packet does not exceed
the current MTU. See Section 6.2.
When about to send a DATA chunk, the code now checks to see if the SACK
timer is running. If it is, we know we have a SACK to send to the
peer, so we append the SACK (assuming available space in the packet)
and turn off the timer. For a simple request-response scenario, this
will result in the SACK being bundled with the response, meaning the
the SACK is received quickly by the client, and also meaning that no
separate SACK packet needs to be sent by the server to acknowledge the
request. Prior to this patch, a separate SACK packet would have been
sent by the server SCTP only after its delayed-ACK timer had expired
(usually 200ms). This is wasteful of bandwidth, and can also have a
major negative impact on performance due the interaction of delayed ACKs
with the Nagle algorithm.
Signed-off-by: Doug Graham <dgraham@nortel.com>
Signed-off-by: Vlad Yasevich <vladislav.yasevich@hp.com>
2009-07-29 16:05:57 +00:00
|
|
|
/* If the SACK timer is running, we have a pending SACK */
|
|
|
|
if (timer_pending(timer)) {
|
2005-04-16 22:20:36 +00:00
|
|
|
struct sctp_chunk *sack;
|
2012-06-30 03:04:26 +00:00
|
|
|
|
|
|
|
if (pkt->transport->sack_generation !=
|
|
|
|
pkt->transport->asoc->peer.sack_generation)
|
|
|
|
return retval;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
asoc->a_rwnd = asoc->rwnd;
|
|
|
|
sack = sctp_make_sack(asoc);
|
|
|
|
if (sack) {
|
2012-07-02 09:59:24 +00:00
|
|
|
retval = __sctp_packet_append_chunk(pkt, sack);
|
|
|
|
if (retval != SCTP_XMIT_OK) {
|
|
|
|
sctp_chunk_free(sack);
|
|
|
|
goto out;
|
|
|
|
}
|
2019-12-09 05:45:18 +00:00
|
|
|
SCTP_INC_STATS(asoc->base.net,
|
2019-07-03 10:20:20 +00:00
|
|
|
SCTP_MIB_OUTCTRLCHUNKS);
|
|
|
|
asoc->stats.octrlchunks++;
|
2005-04-16 22:20:36 +00:00
|
|
|
asoc->peer.sack_needed = 0;
|
sctp: Fix piggybacked ACKs
This patch corrects the conditions under which a SACK will be piggybacked
on a DATA packet. The previous condition was incorrect due to a
misinterpretation of RFC 4960 and/or RFC 2960. Specifically, the
following paragraph from section 6.2 had not been implemented correctly:
Before an endpoint transmits a DATA chunk, if any received DATA
chunks have not been acknowledged (e.g., due to delayed ack), the
sender should create a SACK and bundle it with the outbound DATA
chunk, as long as the size of the final SCTP packet does not exceed
the current MTU. See Section 6.2.
When about to send a DATA chunk, the code now checks to see if the SACK
timer is running. If it is, we know we have a SACK to send to the
peer, so we append the SACK (assuming available space in the packet)
and turn off the timer. For a simple request-response scenario, this
will result in the SACK being bundled with the response, meaning the
the SACK is received quickly by the client, and also meaning that no
separate SACK packet needs to be sent by the server to acknowledge the
request. Prior to this patch, a separate SACK packet would have been
sent by the server SCTP only after its delayed-ACK timer had expired
(usually 200ms). This is wasteful of bandwidth, and can also have a
major negative impact on performance due the interaction of delayed ACKs
with the Nagle algorithm.
Signed-off-by: Doug Graham <dgraham@nortel.com>
Signed-off-by: Vlad Yasevich <vladislav.yasevich@hp.com>
2009-07-29 16:05:57 +00:00
|
|
|
if (del_timer(timer))
|
2005-04-16 22:20:36 +00:00
|
|
|
sctp_association_put(asoc);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2012-07-02 09:59:24 +00:00
|
|
|
out:
|
2005-04-16 22:20:36 +00:00
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
2012-07-02 09:59:24 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/* Append a chunk to the offered packet reporting back any inability to do
|
|
|
|
* so.
|
|
|
|
*/
|
2017-08-05 11:59:57 +00:00
|
|
|
static enum sctp_xmit __sctp_packet_append_chunk(struct sctp_packet *packet,
|
|
|
|
struct sctp_chunk *chunk)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2016-09-21 11:45:55 +00:00
|
|
|
__u16 chunk_len = SCTP_PAD4(ntohs(chunk->chunk_hdr->length));
|
2017-08-05 11:59:57 +00:00
|
|
|
enum sctp_xmit retval = SCTP_XMIT_OK;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-08-07 14:43:07 +00:00
|
|
|
/* Check to see if this chunk will fit into the packet */
|
|
|
|
retval = sctp_packet_will_fit(packet, chunk, chunk_len);
|
|
|
|
if (retval != SCTP_XMIT_OK)
|
|
|
|
goto finish;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2009-08-07 14:43:07 +00:00
|
|
|
/* We believe that this chunk is OK to add to the packet */
|
2007-09-17 02:32:45 +00:00
|
|
|
switch (chunk->chunk_hdr->type) {
|
2013-12-23 04:16:52 +00:00
|
|
|
case SCTP_CID_DATA:
|
2017-12-08 13:04:02 +00:00
|
|
|
case SCTP_CID_I_DATA:
|
2009-08-07 14:43:07 +00:00
|
|
|
/* Account for the data being in the packet */
|
|
|
|
sctp_packet_append_data(packet, chunk);
|
2005-04-16 22:20:36 +00:00
|
|
|
/* Disallow SACK bundling after DATA. */
|
|
|
|
packet->has_sack = 1;
|
2007-09-17 02:32:45 +00:00
|
|
|
/* Disallow AUTH bundling after DATA */
|
|
|
|
packet->has_auth = 1;
|
|
|
|
/* Let it be knows that packet has DATA in it */
|
|
|
|
packet->has_data = 1;
|
2009-01-22 22:53:01 +00:00
|
|
|
/* timestamp the chunk for rtx purposes */
|
|
|
|
chunk->sent_at = jiffies;
|
2016-07-09 11:47:43 +00:00
|
|
|
/* Mainly used for prsctp RTX policy */
|
|
|
|
chunk->sent_count++;
|
2007-09-17 02:32:45 +00:00
|
|
|
break;
|
2013-12-23 04:16:52 +00:00
|
|
|
case SCTP_CID_COOKIE_ECHO:
|
2005-04-16 22:20:36 +00:00
|
|
|
packet->has_cookie_echo = 1;
|
2007-09-17 02:32:45 +00:00
|
|
|
break;
|
|
|
|
|
2013-12-23 04:16:52 +00:00
|
|
|
case SCTP_CID_SACK:
|
2005-04-16 22:20:36 +00:00
|
|
|
packet->has_sack = 1;
|
2012-12-01 04:49:42 +00:00
|
|
|
if (chunk->asoc)
|
|
|
|
chunk->asoc->stats.osacks++;
|
2007-09-17 02:32:45 +00:00
|
|
|
break;
|
|
|
|
|
2013-12-23 04:16:52 +00:00
|
|
|
case SCTP_CID_AUTH:
|
2007-09-17 02:32:45 +00:00
|
|
|
packet->has_auth = 1;
|
|
|
|
packet->auth = chunk;
|
|
|
|
break;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* It is OK to send this chunk. */
|
2005-07-09 04:47:49 +00:00
|
|
|
list_add_tail(&chunk->list, &packet->chunk_list);
|
2005-04-16 22:20:36 +00:00
|
|
|
packet->size += chunk_len;
|
|
|
|
chunk->transport = packet->transport;
|
|
|
|
finish:
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
2012-07-02 09:59:24 +00:00
|
|
|
/* Append a chunk to the offered packet reporting back any inability to do
|
|
|
|
* so.
|
|
|
|
*/
|
2017-08-05 11:59:57 +00:00
|
|
|
enum sctp_xmit sctp_packet_append_chunk(struct sctp_packet *packet,
|
|
|
|
struct sctp_chunk *chunk)
|
2012-07-02 09:59:24 +00:00
|
|
|
{
|
2017-08-05 11:59:57 +00:00
|
|
|
enum sctp_xmit retval = SCTP_XMIT_OK;
|
2012-07-02 09:59:24 +00:00
|
|
|
|
net: sctp: rework debugging framework to use pr_debug and friends
We should get rid of all own SCTP debug printk macros and use the ones
that the kernel offers anyway instead. This makes the code more readable
and conform to the kernel code, and offers all the features of dynamic
debbuging that pr_debug() et al has, such as only turning on/off portions
of debug messages at runtime through debugfs. The runtime cost of having
CONFIG_DYNAMIC_DEBUG enabled, but none of the debug statements printing,
is negligible [1]. If kernel debugging is completly turned off, then these
statements will also compile into "empty" functions.
While we're at it, we also need to change the Kconfig option as it /now/
only refers to the ifdef'ed code portions in outqueue.c that enable further
debugging/tracing of SCTP transaction fields. Also, since SCTP_ASSERT code
was enabled with this Kconfig option and has now been removed, we
transform those code parts into WARNs resp. where appropriate BUG_ONs so
that those bugs can be more easily detected as probably not many people
have SCTP debugging permanently turned on.
To turn on all SCTP debugging, the following steps are needed:
# mount -t debugfs none /sys/kernel/debug
# echo -n 'module sctp +p' > /sys/kernel/debug/dynamic_debug/control
This can be done more fine-grained on a per file, per line basis and others
as described in [2].
[1] https://www.kernel.org/doc/ols/2009/ols2009-pages-39-46.pdf
[2] Documentation/dynamic-debug-howto.txt
Signed-off-by: Daniel Borkmann <dborkman@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2013-06-28 17:49:40 +00:00
|
|
|
pr_debug("%s: packet:%p chunk:%p\n", __func__, packet, chunk);
|
2012-07-02 09:59:24 +00:00
|
|
|
|
|
|
|
/* Data chunks are special. Before seeing what else we can
|
|
|
|
* bundle into this packet, check to see if we are allowed to
|
|
|
|
* send this DATA.
|
|
|
|
*/
|
|
|
|
if (sctp_chunk_is_data(chunk)) {
|
|
|
|
retval = sctp_packet_can_append_data(packet, chunk);
|
|
|
|
if (retval != SCTP_XMIT_OK)
|
|
|
|
goto finish;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Try to bundle AUTH chunk */
|
|
|
|
retval = sctp_packet_bundle_auth(packet, chunk);
|
|
|
|
if (retval != SCTP_XMIT_OK)
|
|
|
|
goto finish;
|
|
|
|
|
|
|
|
/* Try to bundle SACK chunk */
|
|
|
|
retval = sctp_packet_bundle_sack(packet, chunk);
|
|
|
|
if (retval != SCTP_XMIT_OK)
|
|
|
|
goto finish;
|
|
|
|
|
|
|
|
retval = __sctp_packet_append_chunk(packet, chunk);
|
2021-06-22 18:04:52 +00:00
|
|
|
if (retval != SCTP_XMIT_OK)
|
|
|
|
goto finish;
|
|
|
|
|
|
|
|
retval = sctp_packet_bundle_pad(packet, chunk);
|
2012-07-02 09:59:24 +00:00
|
|
|
|
|
|
|
finish:
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
2018-06-13 23:37:02 +00:00
|
|
|
static void sctp_packet_gso_append(struct sk_buff *head, struct sk_buff *skb)
|
|
|
|
{
|
|
|
|
if (SCTP_OUTPUT_CB(head)->last == head)
|
|
|
|
skb_shinfo(head)->frag_list = skb;
|
|
|
|
else
|
|
|
|
SCTP_OUTPUT_CB(head)->last->next = skb;
|
|
|
|
SCTP_OUTPUT_CB(head)->last = skb;
|
|
|
|
|
|
|
|
head->truesize += skb->truesize;
|
|
|
|
head->data_len += skb->len;
|
|
|
|
head->len += skb->len;
|
2018-11-26 06:52:44 +00:00
|
|
|
refcount_add(skb->truesize, &head->sk->sk_wmem_alloc);
|
2018-06-13 23:37:02 +00:00
|
|
|
|
|
|
|
__skb_header_release(skb);
|
|
|
|
}
|
|
|
|
|
2016-10-31 16:49:41 +00:00
|
|
|
static int sctp_packet_pack(struct sctp_packet *packet,
|
|
|
|
struct sk_buff *head, int gso, gfp_t gfp)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
struct sctp_transport *tp = packet->transport;
|
2016-10-31 16:49:41 +00:00
|
|
|
struct sctp_auth_chunk *auth = NULL;
|
2005-07-09 04:47:49 +00:00
|
|
|
struct sctp_chunk *chunk, *tmp;
|
2016-10-31 16:49:41 +00:00
|
|
|
int pkt_count = 0, pkt_size;
|
|
|
|
struct sock *sk = head->sk;
|
|
|
|
struct sk_buff *nskb;
|
2016-10-23 17:01:09 +00:00
|
|
|
int auth_len = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2016-06-02 18:05:43 +00:00
|
|
|
if (gso) {
|
|
|
|
skb_shinfo(head)->gso_type = sk->sk_gso_type;
|
2018-06-13 23:37:02 +00:00
|
|
|
SCTP_OUTPUT_CB(head)->last = head;
|
2016-10-31 16:49:41 +00:00
|
|
|
} else {
|
|
|
|
nskb = head;
|
|
|
|
pkt_size = packet->size;
|
|
|
|
goto merge;
|
2016-06-02 18:05:43 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2016-06-02 18:05:43 +00:00
|
|
|
do {
|
2016-10-31 16:49:41 +00:00
|
|
|
/* calculate the pkt_size and alloc nskb */
|
|
|
|
pkt_size = packet->overhead;
|
|
|
|
list_for_each_entry_safe(chunk, tmp, &packet->chunk_list,
|
|
|
|
list) {
|
|
|
|
int padded = SCTP_PAD4(chunk->skb->len);
|
2013-11-25 03:26:57 +00:00
|
|
|
|
2016-10-31 16:49:41 +00:00
|
|
|
if (chunk == packet->auth)
|
|
|
|
auth_len = padded;
|
|
|
|
else if (auth_len + padded + packet->overhead >
|
|
|
|
tp->pathmtu)
|
|
|
|
return 0;
|
|
|
|
else if (pkt_size + padded > tp->pathmtu)
|
|
|
|
break;
|
|
|
|
pkt_size += padded;
|
2016-06-02 18:05:43 +00:00
|
|
|
}
|
2016-10-31 16:49:41 +00:00
|
|
|
nskb = alloc_skb(pkt_size + MAX_HEADER, gfp);
|
|
|
|
if (!nskb)
|
|
|
|
return 0;
|
|
|
|
skb_reserve(nskb, packet->overhead + MAX_HEADER);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2016-10-31 16:49:41 +00:00
|
|
|
merge:
|
|
|
|
/* merge chunks into nskb and append nskb into head list */
|
2016-06-02 18:05:43 +00:00
|
|
|
pkt_size -= packet->overhead;
|
|
|
|
list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) {
|
2016-10-31 16:49:41 +00:00
|
|
|
int padding;
|
|
|
|
|
2016-06-02 18:05:43 +00:00
|
|
|
list_del_init(&chunk->list);
|
|
|
|
if (sctp_chunk_is_data(chunk)) {
|
2016-10-08 03:36:05 +00:00
|
|
|
if (!sctp_chunk_retransmitted(chunk) &&
|
|
|
|
!tp->rto_pending) {
|
2016-06-02 18:05:43 +00:00
|
|
|
chunk->rtt_in_progress = 1;
|
|
|
|
tp->rto_pending = 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-09-21 11:45:55 +00:00
|
|
|
padding = SCTP_PAD4(chunk->skb->len) - chunk->skb->len;
|
2016-06-02 18:05:43 +00:00
|
|
|
if (padding)
|
networking: convert many more places to skb_put_zero()
There were many places that my previous spatch didn't find,
as pointed out by yuan linyu in various patches.
The following spatch found many more and also removes the
now unnecessary casts:
@@
identifier p, p2;
expression len;
expression skb;
type t, t2;
@@
(
-p = skb_put(skb, len);
+p = skb_put_zero(skb, len);
|
-p = (t)skb_put(skb, len);
+p = skb_put_zero(skb, len);
)
... when != p
(
p2 = (t2)p;
-memset(p2, 0, len);
|
-memset(p, 0, len);
)
@@
type t, t2;
identifier p, p2;
expression skb;
@@
t *p;
...
(
-p = skb_put(skb, sizeof(t));
+p = skb_put_zero(skb, sizeof(t));
|
-p = (t *)skb_put(skb, sizeof(t));
+p = skb_put_zero(skb, sizeof(t));
)
... when != p
(
p2 = (t2)p;
-memset(p2, 0, sizeof(*p));
|
-memset(p, 0, sizeof(*p));
)
@@
expression skb, len;
@@
-memset(skb_put(skb, len), 0, len);
+skb_put_zero(skb, len);
Apply it to the tree (with one manual fixup to keep the
comment in vxlan.c, which spatch removed.)
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-06-16 12:29:19 +00:00
|
|
|
skb_put_zero(chunk->skb, padding);
|
2016-06-02 18:05:43 +00:00
|
|
|
|
|
|
|
if (chunk == packet->auth)
|
2016-10-31 16:49:41 +00:00
|
|
|
auth = (struct sctp_auth_chunk *)
|
|
|
|
skb_tail_pointer(nskb);
|
2016-06-02 18:05:43 +00:00
|
|
|
|
networking: introduce and use skb_put_data()
A common pattern with skb_put() is to just want to memcpy()
some data into the new space, introduce skb_put_data() for
this.
An spatch similar to the one for skb_put_zero() converts many
of the places using it:
@@
identifier p, p2;
expression len, skb, data;
type t, t2;
@@
(
-p = skb_put(skb, len);
+p = skb_put_data(skb, data, len);
|
-p = (t)skb_put(skb, len);
+p = skb_put_data(skb, data, len);
)
(
p2 = (t2)p;
-memcpy(p2, data, len);
|
-memcpy(p, data, len);
)
@@
type t, t2;
identifier p, p2;
expression skb, data;
@@
t *p;
...
(
-p = skb_put(skb, sizeof(t));
+p = skb_put_data(skb, data, sizeof(t));
|
-p = (t *)skb_put(skb, sizeof(t));
+p = skb_put_data(skb, data, sizeof(t));
)
(
p2 = (t2)p;
-memcpy(p2, data, sizeof(*p));
|
-memcpy(p, data, sizeof(*p));
)
@@
expression skb, len, data;
@@
-memcpy(skb_put(skb, len), data, len);
+skb_put_data(skb, data, len);
(again, manually post-processed to retain some comments)
Reviewed-by: Stephen Hemminger <stephen@networkplumber.org>
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-06-16 12:29:20 +00:00
|
|
|
skb_put_data(nskb, chunk->skb->data, chunk->skb->len);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2016-06-02 18:05:43 +00:00
|
|
|
pr_debug("*** Chunk:%p[%s] %s 0x%x, length:%d, chunk->skb->len:%d, rtt_in_progress:%d\n",
|
|
|
|
chunk,
|
|
|
|
sctp_cname(SCTP_ST_CHUNK(chunk->chunk_hdr->type)),
|
|
|
|
chunk->has_tsn ? "TSN" : "No TSN",
|
|
|
|
chunk->has_tsn ? ntohl(chunk->subh.data_hdr->tsn) : 0,
|
|
|
|
ntohs(chunk->chunk_hdr->length), chunk->skb->len,
|
|
|
|
chunk->rtt_in_progress);
|
|
|
|
|
2016-09-21 11:45:55 +00:00
|
|
|
pkt_size -= SCTP_PAD4(chunk->skb->len);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2016-07-07 12:39:29 +00:00
|
|
|
if (!sctp_chunk_is_data(chunk) && chunk != packet->auth)
|
2016-06-02 18:05:43 +00:00
|
|
|
sctp_chunk_free(chunk);
|
|
|
|
|
|
|
|
if (!pkt_size)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2016-10-31 16:49:41 +00:00
|
|
|
if (auth) {
|
2018-03-14 11:05:30 +00:00
|
|
|
sctp_auth_calculate_hmac(tp->asoc, nskb, auth,
|
|
|
|
packet->auth->shkey, gfp);
|
2016-10-31 16:49:41 +00:00
|
|
|
/* free auth if no more chunks, or add it back */
|
|
|
|
if (list_empty(&packet->chunk_list))
|
|
|
|
sctp_chunk_free(packet->auth);
|
|
|
|
else
|
2016-07-30 05:58:35 +00:00
|
|
|
list_add(&packet->auth->list,
|
|
|
|
&packet->chunk_list);
|
2016-07-07 12:39:29 +00:00
|
|
|
}
|
|
|
|
|
2018-06-13 23:37:02 +00:00
|
|
|
if (gso)
|
|
|
|
sctp_packet_gso_append(head, nskb);
|
2016-10-31 16:49:41 +00:00
|
|
|
|
|
|
|
pkt_count++;
|
2016-06-02 18:05:43 +00:00
|
|
|
} while (!list_empty(&packet->chunk_list));
|
2007-09-17 02:32:45 +00:00
|
|
|
|
2016-10-31 16:49:41 +00:00
|
|
|
if (gso) {
|
|
|
|
memset(head->cb, 0, max(sizeof(struct inet_skb_parm),
|
|
|
|
sizeof(struct inet6_skb_parm)));
|
|
|
|
skb_shinfo(head)->gso_segs = pkt_count;
|
|
|
|
skb_shinfo(head)->gso_size = GSO_BY_FRAGS;
|
|
|
|
goto chksum;
|
2009-04-27 22:35:52 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2016-10-31 16:49:41 +00:00
|
|
|
if (sctp_checksum_disable)
|
|
|
|
return 1;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
sctp: support for sending packet over udp4 sock
This patch does what the rfc6951#section-5.3 says for ipv4:
"Within the UDP header, the source port MUST be the local UDP
encapsulation port number of the SCTP stack, and the destination port
MUST be the remote UDP encapsulation port number maintained for the
association and the destination address to which the packet is sent
(see Section 5.1).
Because the SCTP packet is the UDP payload, the length of the UDP
packet MUST be the length of the SCTP packet plus the size of the UDP
header.
The SCTP checksum MUST be computed for IPv4 and IPv6, and the UDP
checksum SHOULD be computed for IPv4 and IPv6."
Some places need to be adjusted in sctp_packet_transmit():
1. For non-gso packets, when transport's encap_port is set, sctp
checksum has to be done in sctp_packet_pack(), as the outer
udp will use ip_summed = CHECKSUM_PARTIAL to do the offload
setting for checksum.
2. Delay calling dst_clone() and skb_dst_set() for non-udp packets
until sctp_v4_xmit(), as for udp packets, skb_dst_set() is not
needed before calling udp_tunnel_xmit_skb().
then in sctp_v4_xmit():
1. Go to udp_tunnel_xmit_skb() only when transport->encap_port and
net->sctp.udp_port both are set, as these are one for dst port
and another for src port.
2. For gso packet, SKB_GSO_UDP_TUNNEL_CSUM is set for gso_type, and
with this udp checksum can be done in __skb_udp_tunnel_segment()
for each segments after the sctp gso.
3. inner_mac_header and inner_transport_header are set, as these
will be needed in __skb_udp_tunnel_segment() to find the right
headers.
4. df and ttl are calculated, as these are the required params by
udp_tunnel_xmit_skb().
5. nocheck param has to be false, as "the UDP checksum SHOULD be
computed for IPv4 and IPv6", says in rfc6951#section-5.3.
v1->v2:
- Use sp->udp_port instead in sctp_v4_xmit(), which is more safe.
Signed-off-by: Xin Long <lucien.xin@gmail.com>
Acked-by: Marcelo Ricardo Leitner <marcelo.leitner@gmail.com>
Signed-off-by: Jakub Kicinski <kuba@kernel.org>
2020-10-29 07:05:06 +00:00
|
|
|
if (!(tp->dst->dev->features & NETIF_F_SCTP_CRC) ||
|
|
|
|
dst_xfrm(tp->dst) || packet->ipfragok || tp->encap_port) {
|
2016-10-31 16:49:41 +00:00
|
|
|
struct sctphdr *sh =
|
|
|
|
(struct sctphdr *)skb_transport_header(head);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2016-10-31 16:49:41 +00:00
|
|
|
sh->checksum = sctp_compute_cksum(head, 0);
|
|
|
|
} else {
|
|
|
|
chksum:
|
|
|
|
head->ip_summed = CHECKSUM_PARTIAL;
|
2017-05-18 13:44:40 +00:00
|
|
|
head->csum_not_inet = 1;
|
2016-10-31 16:49:41 +00:00
|
|
|
head->csum_start = skb_transport_header(head) - head->head;
|
|
|
|
head->csum_offset = offsetof(struct sctphdr, checksum);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2016-10-31 16:49:41 +00:00
|
|
|
return pkt_count;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* All packets are sent to the network through this function from
|
|
|
|
* sctp_outq_tail().
|
|
|
|
*
|
|
|
|
* The return value is always 0 for now.
|
|
|
|
*/
|
|
|
|
int sctp_packet_transmit(struct sctp_packet *packet, gfp_t gfp)
|
|
|
|
{
|
|
|
|
struct sctp_transport *tp = packet->transport;
|
|
|
|
struct sctp_association *asoc = tp->asoc;
|
|
|
|
struct sctp_chunk *chunk, *tmp;
|
|
|
|
int pkt_count, gso = 0;
|
|
|
|
struct sk_buff *head;
|
|
|
|
struct sctphdr *sh;
|
|
|
|
struct sock *sk;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2016-10-31 16:49:41 +00:00
|
|
|
pr_debug("%s: packet:%p\n", __func__, packet);
|
|
|
|
if (list_empty(&packet->chunk_list))
|
|
|
|
return 0;
|
|
|
|
chunk = list_entry(packet->chunk_list.next, struct sctp_chunk, list);
|
|
|
|
sk = chunk->skb->sk;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2021-06-22 18:04:52 +00:00
|
|
|
if (packet->size > tp->pathmtu && !packet->ipfragok && !chunk->pmtu_probe) {
|
2021-10-28 09:36:01 +00:00
|
|
|
if (tp->pl.state == SCTP_PL_ERROR) { /* do IP fragmentation if in Error state */
|
|
|
|
packet->ipfragok = 1;
|
|
|
|
} else {
|
|
|
|
if (!sk_can_gso(sk)) { /* check gso */
|
|
|
|
pr_err_once("Trying to GSO but underlying device doesn't support it.");
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
gso = 1;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2016-10-31 16:49:41 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* alloc head skb */
|
|
|
|
head = alloc_skb((gso ? packet->overhead : packet->size) +
|
|
|
|
MAX_HEADER, gfp);
|
|
|
|
if (!head)
|
|
|
|
goto out;
|
|
|
|
skb_reserve(head, packet->overhead + MAX_HEADER);
|
2018-11-18 07:07:38 +00:00
|
|
|
skb_set_owner_w(head, sk);
|
2016-10-31 16:49:41 +00:00
|
|
|
|
|
|
|
/* set sctp header */
|
networking: make skb_push & __skb_push return void pointers
It seems like a historic accident that these return unsigned char *,
and in many places that means casts are required, more often than not.
Make these functions return void * and remove all the casts across
the tree, adding a (u8 *) cast only where the unsigned char pointer
was used directly, all done with the following spatch:
@@
expression SKB, LEN;
typedef u8;
identifier fn = { skb_push, __skb_push, skb_push_rcsum };
@@
- *(fn(SKB, LEN))
+ *(u8 *)fn(SKB, LEN)
@@
expression E, SKB, LEN;
identifier fn = { skb_push, __skb_push, skb_push_rcsum };
type T;
@@
- E = ((T *)(fn(SKB, LEN)))
+ E = fn(SKB, LEN)
@@
expression SKB, LEN;
identifier fn = { skb_push, __skb_push, skb_push_rcsum };
@@
- fn(SKB, LEN)[0]
+ *(u8 *)fn(SKB, LEN)
Note that the last part there converts from push(...)[0] to the
more idiomatic *(u8 *)push(...).
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-06-16 12:29:23 +00:00
|
|
|
sh = skb_push(head, sizeof(struct sctphdr));
|
2016-10-31 16:49:41 +00:00
|
|
|
skb_reset_transport_header(head);
|
|
|
|
sh->source = htons(packet->source_port);
|
|
|
|
sh->dest = htons(packet->destination_port);
|
|
|
|
sh->vtag = htonl(packet->vtag);
|
|
|
|
sh->checksum = 0;
|
|
|
|
|
2017-04-01 09:15:59 +00:00
|
|
|
/* drop packet if no dst */
|
sctp: support for sending packet over udp4 sock
This patch does what the rfc6951#section-5.3 says for ipv4:
"Within the UDP header, the source port MUST be the local UDP
encapsulation port number of the SCTP stack, and the destination port
MUST be the remote UDP encapsulation port number maintained for the
association and the destination address to which the packet is sent
(see Section 5.1).
Because the SCTP packet is the UDP payload, the length of the UDP
packet MUST be the length of the SCTP packet plus the size of the UDP
header.
The SCTP checksum MUST be computed for IPv4 and IPv6, and the UDP
checksum SHOULD be computed for IPv4 and IPv6."
Some places need to be adjusted in sctp_packet_transmit():
1. For non-gso packets, when transport's encap_port is set, sctp
checksum has to be done in sctp_packet_pack(), as the outer
udp will use ip_summed = CHECKSUM_PARTIAL to do the offload
setting for checksum.
2. Delay calling dst_clone() and skb_dst_set() for non-udp packets
until sctp_v4_xmit(), as for udp packets, skb_dst_set() is not
needed before calling udp_tunnel_xmit_skb().
then in sctp_v4_xmit():
1. Go to udp_tunnel_xmit_skb() only when transport->encap_port and
net->sctp.udp_port both are set, as these are one for dst port
and another for src port.
2. For gso packet, SKB_GSO_UDP_TUNNEL_CSUM is set for gso_type, and
with this udp checksum can be done in __skb_udp_tunnel_segment()
for each segments after the sctp gso.
3. inner_mac_header and inner_transport_header are set, as these
will be needed in __skb_udp_tunnel_segment() to find the right
headers.
4. df and ttl are calculated, as these are the required params by
udp_tunnel_xmit_skb().
5. nocheck param has to be false, as "the UDP checksum SHOULD be
computed for IPv4 and IPv6", says in rfc6951#section-5.3.
v1->v2:
- Use sp->udp_port instead in sctp_v4_xmit(), which is more safe.
Signed-off-by: Xin Long <lucien.xin@gmail.com>
Acked-by: Marcelo Ricardo Leitner <marcelo.leitner@gmail.com>
Signed-off-by: Jakub Kicinski <kuba@kernel.org>
2020-10-29 07:05:06 +00:00
|
|
|
if (!tp->dst) {
|
2016-10-31 16:49:41 +00:00
|
|
|
IP_INC_STATS(sock_net(sk), IPSTATS_MIB_OUTNOROUTES);
|
|
|
|
kfree_skb(head);
|
|
|
|
goto out;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2016-10-31 16:49:41 +00:00
|
|
|
/* pack up chunks */
|
|
|
|
pkt_count = sctp_packet_pack(packet, head, gso, gfp);
|
|
|
|
if (!pkt_count) {
|
|
|
|
kfree_skb(head);
|
|
|
|
goto out;
|
|
|
|
}
|
2016-06-02 18:05:43 +00:00
|
|
|
pr_debug("***sctp_transmit_packet*** skb->len:%d\n", head->len);
|
|
|
|
|
2016-10-31 16:49:41 +00:00
|
|
|
/* start autoclose timer */
|
|
|
|
if (packet->has_data && sctp_state(asoc, ESTABLISHED) &&
|
|
|
|
asoc->timeouts[SCTP_EVENT_TIMEOUT_AUTOCLOSE]) {
|
|
|
|
struct timer_list *timer =
|
|
|
|
&asoc->timers[SCTP_EVENT_TIMEOUT_AUTOCLOSE];
|
|
|
|
unsigned long timeout =
|
|
|
|
asoc->timeouts[SCTP_EVENT_TIMEOUT_AUTOCLOSE];
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2016-10-31 16:49:41 +00:00
|
|
|
if (!mod_timer(timer, jiffies + timeout))
|
|
|
|
sctp_association_hold(asoc);
|
|
|
|
}
|
2016-06-02 18:05:43 +00:00
|
|
|
|
2016-10-31 16:49:41 +00:00
|
|
|
/* sctp xmit */
|
|
|
|
tp->af_specific->ecn_capable(sk);
|
|
|
|
if (asoc) {
|
|
|
|
asoc->stats.opackets += pkt_count;
|
|
|
|
if (asoc->peer.last_sent_to != tp)
|
|
|
|
asoc->peer.last_sent_to = tp;
|
2016-06-02 18:05:43 +00:00
|
|
|
}
|
|
|
|
head->ignore_df = packet->ipfragok;
|
2017-03-18 11:12:22 +00:00
|
|
|
if (tp->dst_pending_confirm)
|
2017-02-06 21:14:13 +00:00
|
|
|
skb_set_dst_pending_confirm(head, 1);
|
|
|
|
/* neighbour should be confirmed on successful transmission or
|
|
|
|
* positive error
|
|
|
|
*/
|
2017-03-18 11:12:22 +00:00
|
|
|
if (tp->af_specific->sctp_xmit(head, tp) >= 0 &&
|
|
|
|
tp->dst_pending_confirm)
|
2017-02-06 21:14:13 +00:00
|
|
|
tp->dst_pending_confirm = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2016-10-31 16:49:41 +00:00
|
|
|
out:
|
2005-07-09 04:47:49 +00:00
|
|
|
list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) {
|
|
|
|
list_del_init(&chunk->list);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (!sctp_chunk_is_data(chunk))
|
2007-02-09 14:25:18 +00:00
|
|
|
sctp_chunk_free(chunk);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2016-09-13 18:04:23 +00:00
|
|
|
sctp_packet_reset(packet);
|
2016-10-31 16:49:41 +00:00
|
|
|
return 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/********************************************************************
|
|
|
|
* 2nd Level Abstractions
|
|
|
|
********************************************************************/
|
|
|
|
|
2009-08-07 14:43:07 +00:00
|
|
|
/* This private function check to see if a chunk can be added */
|
2017-08-05 11:59:57 +00:00
|
|
|
static enum sctp_xmit sctp_packet_can_append_data(struct sctp_packet *packet,
|
|
|
|
struct sctp_chunk *chunk)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2009-08-07 14:43:07 +00:00
|
|
|
size_t datasize, rwnd, inflight, flight_size;
|
2005-04-16 22:20:36 +00:00
|
|
|
struct sctp_transport *transport = packet->transport;
|
|
|
|
struct sctp_association *asoc = transport->asoc;
|
|
|
|
struct sctp_outq *q = &asoc->outqueue;
|
|
|
|
|
|
|
|
/* RFC 2960 6.1 Transmission of DATA Chunks
|
|
|
|
*
|
|
|
|
* A) At any given time, the data sender MUST NOT transmit new data to
|
|
|
|
* any destination transport address if its peer's rwnd indicates
|
|
|
|
* that the peer has no buffer space (i.e. rwnd is 0, see Section
|
|
|
|
* 6.2.1). However, regardless of the value of rwnd (including if it
|
|
|
|
* is 0), the data sender can always have one DATA chunk in flight to
|
|
|
|
* the receiver if allowed by cwnd (see rule B below). This rule
|
|
|
|
* allows the sender to probe for a change in rwnd that the sender
|
|
|
|
* missed due to the SACK having been lost in transit from the data
|
|
|
|
* receiver to the data sender.
|
|
|
|
*/
|
|
|
|
|
|
|
|
rwnd = asoc->peer.rwnd;
|
2009-08-07 14:43:07 +00:00
|
|
|
inflight = q->outstanding_bytes;
|
|
|
|
flight_size = transport->flight_size;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
datasize = sctp_data_size(chunk);
|
|
|
|
|
2014-07-22 08:59:08 +00:00
|
|
|
if (datasize > rwnd && inflight > 0)
|
|
|
|
/* We have (at least) one data chunk in flight,
|
|
|
|
* so we can't fall back to rule 6.1 B).
|
|
|
|
*/
|
|
|
|
return SCTP_XMIT_RWND_FULL;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* RFC 2960 6.1 Transmission of DATA Chunks
|
|
|
|
*
|
|
|
|
* B) At any given time, the sender MUST NOT transmit new data
|
|
|
|
* to a given transport address if it has cwnd or more bytes
|
|
|
|
* of data outstanding to that transport address.
|
|
|
|
*/
|
|
|
|
/* RFC 7.2.4 & the Implementers Guide 2.8.
|
|
|
|
*
|
|
|
|
* 3) ...
|
|
|
|
* When a Fast Retransmit is being performed the sender SHOULD
|
|
|
|
* ignore the value of cwnd and SHOULD NOT delay retransmission.
|
|
|
|
*/
|
2014-07-22 08:59:08 +00:00
|
|
|
if (chunk->fast_retransmit != SCTP_NEED_FRTX &&
|
|
|
|
flight_size >= transport->cwnd)
|
|
|
|
return SCTP_XMIT_RWND_FULL;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* Nagle's algorithm to solve small-packet problem:
|
|
|
|
* Inhibit the sending of new chunks when new outgoing data arrives
|
|
|
|
* if any previously transmitted data on the connection remains
|
|
|
|
* unacknowledged.
|
|
|
|
*/
|
|
|
|
|
2017-02-18 17:52:46 +00:00
|
|
|
if ((sctp_sk(asoc->base.sk)->nodelay || inflight == 0) &&
|
2017-03-26 16:21:15 +00:00
|
|
|
!asoc->force_delay)
|
2017-02-18 17:52:46 +00:00
|
|
|
/* Nothing unacked */
|
2014-07-22 08:59:08 +00:00
|
|
|
return SCTP_XMIT_OK;
|
|
|
|
|
|
|
|
if (!sctp_packet_empty(packet))
|
|
|
|
/* Append to packet */
|
|
|
|
return SCTP_XMIT_OK;
|
|
|
|
|
|
|
|
if (!sctp_state(asoc, ESTABLISHED))
|
|
|
|
return SCTP_XMIT_OK;
|
|
|
|
|
|
|
|
/* Check whether this chunk and all the rest of pending data will fit
|
|
|
|
* or delay in hopes of bundling a full sized packet.
|
|
|
|
*/
|
2017-06-30 03:52:20 +00:00
|
|
|
if (chunk->skb->len + q->out_qlen > transport->pathmtu -
|
2017-12-08 13:04:02 +00:00
|
|
|
packet->overhead - sctp_datachk_len(&chunk->asoc->stream) - 4)
|
2014-07-22 08:59:08 +00:00
|
|
|
/* Enough data queued to fill a packet */
|
|
|
|
return SCTP_XMIT_OK;
|
|
|
|
|
|
|
|
/* Don't delay large message writes that may have been fragmented */
|
|
|
|
if (!chunk->msg->can_delay)
|
|
|
|
return SCTP_XMIT_OK;
|
|
|
|
|
|
|
|
/* Defer until all data acked or packet full */
|
2014-07-22 08:59:14 +00:00
|
|
|
return SCTP_XMIT_DELAY;
|
2009-08-07 14:43:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* This private function does management things when adding DATA chunk */
|
|
|
|
static void sctp_packet_append_data(struct sctp_packet *packet,
|
|
|
|
struct sctp_chunk *chunk)
|
|
|
|
{
|
|
|
|
struct sctp_transport *transport = packet->transport;
|
|
|
|
size_t datasize = sctp_data_size(chunk);
|
|
|
|
struct sctp_association *asoc = transport->asoc;
|
|
|
|
u32 rwnd = asoc->peer.rwnd;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/* Keep track of how many bytes are in flight over this transport. */
|
|
|
|
transport->flight_size += datasize;
|
|
|
|
|
|
|
|
/* Keep track of how many bytes are in flight to the receiver. */
|
|
|
|
asoc->outqueue.outstanding_bytes += datasize;
|
|
|
|
|
2011-12-19 04:11:40 +00:00
|
|
|
/* Update our view of the receiver's rwnd. */
|
2005-04-16 22:20:36 +00:00
|
|
|
if (datasize < rwnd)
|
|
|
|
rwnd -= datasize;
|
|
|
|
else
|
|
|
|
rwnd = 0;
|
|
|
|
|
|
|
|
asoc->peer.rwnd = rwnd;
|
sctp: Fix mis-ordering of user space data when multihoming in use
Recently had a bug reported to me, in which the user was sending
packets with a payload containing a sequence number. The packets
were getting delivered in order according the chunk TSN values, but
the sequence values in the payload were arriving out of order. At
first I thought it must be an application error, but we eventually
found it to be a problem on the transmit side in the sctp stack.
The conditions for the error are that multihoming must be in use,
and it helps if each transport has a different pmtu. The problem
occurs in sctp_outq_flush. Basically we dequeue packets from the
data queue, and attempt to append them to the orrered packet for a
given transport. After we append a data chunk we add the trasport
to the end of a list of transports to have their packets sent at
the end of sctp_outq_flush. The problem occurs when a data chunks
fills up a offered packet on a transport. The function that does
the appending (sctp_packet_transmit_chunk), will try to call
sctp_packet_transmit on the full packet, and then append the chunk
to a new packet. This call to sctp_packet_transmit, sends that
packet ahead of the others that may be queued in the transport_list
in sctp_outq_flush. The result is that frames that were sent in one
order from the user space sending application get re-ordered prior
to tsn assignment in sctp_packet_transmit, resulting in mis-sequencing
of data payloads, even though tsn ordering is correct.
The fix is to change where we assign a tsn. By doing this earlier,
we are then free to place chunks in packets, whatever way we
see fit and the protocol will make sure to do all the appropriate
re-ordering on receive as is needed.
Signed-off-by: Neil Horman <nhorman@tuxdriver.com>
Reported-by: William Reich <reich@ulticom.com>
Signed-off-by: Vlad Yasevich <vladislav.yasevich@hp.com>
2009-11-23 20:54:00 +00:00
|
|
|
sctp_chunk_assign_tsn(chunk);
|
2017-12-08 13:04:02 +00:00
|
|
|
asoc->stream.si->assign_number(chunk);
|
2009-08-07 14:43:07 +00:00
|
|
|
}
|
|
|
|
|
2017-08-05 11:59:57 +00:00
|
|
|
static enum sctp_xmit sctp_packet_will_fit(struct sctp_packet *packet,
|
|
|
|
struct sctp_chunk *chunk,
|
|
|
|
u16 chunk_len)
|
2009-08-07 14:43:07 +00:00
|
|
|
{
|
2017-08-05 11:59:57 +00:00
|
|
|
enum sctp_xmit retval = SCTP_XMIT_OK;
|
2016-09-08 09:54:11 +00:00
|
|
|
size_t psize, pmtu, maxsize;
|
2009-08-07 14:43:07 +00:00
|
|
|
|
2018-03-14 11:05:30 +00:00
|
|
|
/* Don't bundle in this packet if this chunk's auth key doesn't
|
|
|
|
* match other chunks already enqueued on this packet. Also,
|
|
|
|
* don't bundle the chunk with auth key if other chunks in this
|
|
|
|
* packet don't have auth key.
|
|
|
|
*/
|
|
|
|
if ((packet->auth && chunk->shkey != packet->auth->shkey) ||
|
|
|
|
(!packet->auth && chunk->shkey &&
|
|
|
|
chunk->chunk_hdr->type != SCTP_CID_AUTH))
|
|
|
|
return SCTP_XMIT_PMTU_FULL;
|
|
|
|
|
2009-08-07 14:43:07 +00:00
|
|
|
psize = packet->size;
|
2016-06-02 18:05:43 +00:00
|
|
|
if (packet->transport->asoc)
|
|
|
|
pmtu = packet->transport->asoc->pathmtu;
|
|
|
|
else
|
|
|
|
pmtu = packet->transport->pathmtu;
|
2009-08-07 14:43:07 +00:00
|
|
|
|
|
|
|
/* Decide if we need to fragment or resubmit later. */
|
2016-06-02 18:05:43 +00:00
|
|
|
if (psize + chunk_len > pmtu) {
|
|
|
|
/* It's OK to fragment at IP level if any one of the following
|
2009-08-07 14:43:07 +00:00
|
|
|
* is true:
|
2016-06-02 18:05:43 +00:00
|
|
|
* 1. The packet is empty (meaning this chunk is greater
|
|
|
|
* the MTU)
|
|
|
|
* 2. The packet doesn't have any data in it yet and data
|
|
|
|
* requires authentication.
|
2009-08-07 14:43:07 +00:00
|
|
|
*/
|
2016-06-02 18:05:43 +00:00
|
|
|
if (sctp_packet_empty(packet) ||
|
2009-08-07 14:43:07 +00:00
|
|
|
(!packet->has_data && chunk->auth)) {
|
|
|
|
/* We no longer do re-fragmentation.
|
|
|
|
* Just fragment at the IP layer, if we
|
|
|
|
* actually hit this condition
|
|
|
|
*/
|
|
|
|
packet->ipfragok = 1;
|
2016-06-02 18:05:43 +00:00
|
|
|
goto out;
|
2009-08-07 14:43:07 +00:00
|
|
|
}
|
2016-06-02 18:05:43 +00:00
|
|
|
|
2016-09-08 09:54:11 +00:00
|
|
|
/* Similarly, if this chunk was built before a PMTU
|
|
|
|
* reduction, we have to fragment it at IP level now. So
|
|
|
|
* if the packet already contains something, we need to
|
|
|
|
* flush.
|
|
|
|
*/
|
|
|
|
maxsize = pmtu - packet->overhead;
|
|
|
|
if (packet->auth)
|
2016-09-21 11:45:55 +00:00
|
|
|
maxsize -= SCTP_PAD4(packet->auth->skb->len);
|
2016-09-08 09:54:11 +00:00
|
|
|
if (chunk_len > maxsize)
|
|
|
|
retval = SCTP_XMIT_PMTU_FULL;
|
|
|
|
|
2016-06-02 18:05:43 +00:00
|
|
|
/* It is also okay to fragment if the chunk we are
|
|
|
|
* adding is a control chunk, but only if current packet
|
|
|
|
* is not a GSO one otherwise it causes fragmentation of
|
|
|
|
* a large frame. So in this case we allow the
|
|
|
|
* fragmentation by forcing it to be in a new packet.
|
|
|
|
*/
|
|
|
|
if (!sctp_chunk_is_data(chunk) && packet->has_data)
|
|
|
|
retval = SCTP_XMIT_PMTU_FULL;
|
|
|
|
|
|
|
|
if (psize + chunk_len > packet->max_size)
|
|
|
|
/* Hit GSO/PMTU limit, gotta flush */
|
|
|
|
retval = SCTP_XMIT_PMTU_FULL;
|
|
|
|
|
|
|
|
if (!packet->transport->burst_limited &&
|
|
|
|
psize + chunk_len > (packet->transport->cwnd >> 1))
|
|
|
|
/* Do not allow a single GSO packet to use more
|
|
|
|
* than half of cwnd.
|
|
|
|
*/
|
|
|
|
retval = SCTP_XMIT_PMTU_FULL;
|
|
|
|
|
|
|
|
if (packet->transport->burst_limited &&
|
|
|
|
psize + chunk_len > (packet->transport->burst_limited >> 1))
|
|
|
|
/* Do not allow a single GSO packet to use more
|
|
|
|
* than half of original cwnd.
|
|
|
|
*/
|
|
|
|
retval = SCTP_XMIT_PMTU_FULL;
|
|
|
|
/* Otherwise it will fit in the GSO packet */
|
2009-08-07 14:43:07 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2016-06-02 18:05:43 +00:00
|
|
|
out:
|
2005-04-16 22:20:36 +00:00
|
|
|
return retval;
|
|
|
|
}
|