1 /* SCTP kernel implementation
2 * (C) Copyright IBM Corp. 2001, 2004
3 * Copyright (c) 1999-2000 Cisco, Inc.
4 * Copyright (c) 1999-2001 Motorola, Inc.
5 *
6 * This file is part of the SCTP kernel implementation
7 *
8 * These functions handle output processing.
9 *
10 * This SCTP implementation is free software;
11 * you can redistribute it and/or modify it under the terms of
12 * the GNU General Public License as published by
13 * the Free Software Foundation; either version 2, or (at your option)
14 * any later version.
15 *
16 * This SCTP implementation is distributed in the hope that it
17 * will be useful, but WITHOUT ANY WARRANTY; without even the implied
18 * ************************
19 * warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
20 * See the GNU General Public License for more details.
21 *
22 * You should have received a copy of the GNU General Public License
23 * along with GNU CC; see the file COPYING. If not, write to
24 * the Free Software Foundation, 59 Temple Place - Suite 330,
25 * Boston, MA 02111-1307, USA.
26 *
27 * Please send any bug reports or fixes you make to the
28 * email address(es):
29 * lksctp developers <lksctp-developers@lists.sourceforge.net>
30 *
31 * Or submit a bug report through the following website:
32 * http://www.sf.net/projects/lksctp
33 *
34 * Written or modified by:
35 * La Monte H.P. Yarroll <piggy@acm.org>
36 * Karl Knutson <karl@athena.chicago.il.us>
37 * Jon Grimm <jgrimm@austin.ibm.com>
38 * Sridhar Samudrala <sri@us.ibm.com>
39 *
40 * Any bugs reported given to us we will try to fix... any fixes shared will
41 * be incorporated into the next SCTP release.
42 */
43
44 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
45
46 #include <linux/types.h>
47 #include <linux/kernel.h>
48 #include <linux/wait.h>
49 #include <linux/time.h>
50 #include <linux/ip.h>
51 #include <linux/ipv6.h>
52 #include <linux/init.h>
53 #include <linux/slab.h>
54 #include <net/inet_ecn.h>
55 #include <net/ip.h>
56 #include <net/icmp.h>
57 #include <net/net_namespace.h>
58
59 #include <linux/socket.h> /* for sa_family_t */
60 #include <net/sock.h>
61
62 #include <net/sctp/sctp.h>
63 #include <net/sctp/sm.h>
64 #include <net/sctp/checksum.h>
65
66 /* Forward declarations for private helpers. */
67 static sctp_xmit_t sctp_packet_can_append_data(struct sctp_packet *packet,
68 struct sctp_chunk *chunk);
69 static void sctp_packet_append_data(struct sctp_packet *packet,
70 struct sctp_chunk *chunk);
71 static sctp_xmit_t sctp_packet_will_fit(struct sctp_packet *packet,
72 struct sctp_chunk *chunk,
73 u16 chunk_len);
74
sctp_packet_reset(struct sctp_packet * packet)75 static void sctp_packet_reset(struct sctp_packet *packet)
76 {
77 packet->size = packet->overhead;
78 packet->has_cookie_echo = 0;
79 packet->has_sack = 0;
80 packet->has_data = 0;
81 packet->has_auth = 0;
82 packet->ipfragok = 0;
83 packet->auth = NULL;
84 }
85
86 /* Config a packet.
87 * This appears to be a followup set of initializations.
88 */
sctp_packet_config(struct sctp_packet * packet,__u32 vtag,int ecn_capable)89 struct sctp_packet *sctp_packet_config(struct sctp_packet *packet,
90 __u32 vtag, int ecn_capable)
91 {
92 struct sctp_chunk *chunk = NULL;
93
94 SCTP_DEBUG_PRINTK("%s: packet:%p vtag:0x%x\n", __func__,
95 packet, vtag);
96
97 packet->vtag = vtag;
98
99 if (ecn_capable && sctp_packet_empty(packet)) {
100 chunk = sctp_get_ecne_prepend(packet->transport->asoc);
101
102 /* If there a is a prepend chunk stick it on the list before
103 * any other chunks get appended.
104 */
105 if (chunk)
106 sctp_packet_append_chunk(packet, chunk);
107 }
108
109 return packet;
110 }
111
112 /* Initialize the packet structure. */
sctp_packet_init(struct sctp_packet * packet,struct sctp_transport * transport,__u16 sport,__u16 dport)113 struct sctp_packet *sctp_packet_init(struct sctp_packet *packet,
114 struct sctp_transport *transport,
115 __u16 sport, __u16 dport)
116 {
117 struct sctp_association *asoc = transport->asoc;
118 size_t overhead;
119
120 SCTP_DEBUG_PRINTK("%s: packet:%p transport:%p\n", __func__,
121 packet, transport);
122
123 packet->transport = transport;
124 packet->source_port = sport;
125 packet->destination_port = dport;
126 INIT_LIST_HEAD(&packet->chunk_list);
127 if (asoc) {
128 struct sctp_sock *sp = sctp_sk(asoc->base.sk);
129 overhead = sp->pf->af->net_header_len;
130 } else {
131 overhead = sizeof(struct ipv6hdr);
132 }
133 overhead += sizeof(struct sctphdr);
134 packet->overhead = overhead;
135 sctp_packet_reset(packet);
136 packet->vtag = 0;
137 packet->malloced = 0;
138 return packet;
139 }
140
141 /* Free a packet. */
sctp_packet_free(struct sctp_packet * packet)142 void sctp_packet_free(struct sctp_packet *packet)
143 {
144 struct sctp_chunk *chunk, *tmp;
145
146 SCTP_DEBUG_PRINTK("%s: packet:%p\n", __func__, packet);
147
148 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) {
149 list_del_init(&chunk->list);
150 sctp_chunk_free(chunk);
151 }
152
153 if (packet->malloced)
154 kfree(packet);
155 }
156
157 /* This routine tries to append the chunk to the offered packet. If adding
158 * the chunk causes the packet to exceed the path MTU and COOKIE_ECHO chunk
159 * is not present in the packet, it transmits the input packet.
160 * Data can be bundled with a packet containing a COOKIE_ECHO chunk as long
161 * as it can fit in the packet, but any more data that does not fit in this
162 * packet can be sent only after receiving the COOKIE_ACK.
163 */
sctp_packet_transmit_chunk(struct sctp_packet * packet,struct sctp_chunk * chunk,int one_packet)164 sctp_xmit_t sctp_packet_transmit_chunk(struct sctp_packet *packet,
165 struct sctp_chunk *chunk,
166 int one_packet)
167 {
168 sctp_xmit_t retval;
169 int error = 0;
170
171 SCTP_DEBUG_PRINTK("%s: packet:%p chunk:%p\n", __func__,
172 packet, chunk);
173
174 switch ((retval = (sctp_packet_append_chunk(packet, chunk)))) {
175 case SCTP_XMIT_PMTU_FULL:
176 if (!packet->has_cookie_echo) {
177 error = sctp_packet_transmit(packet);
178 if (error < 0)
179 chunk->skb->sk->sk_err = -error;
180
181 /* If we have an empty packet, then we can NOT ever
182 * return PMTU_FULL.
183 */
184 if (!one_packet)
185 retval = sctp_packet_append_chunk(packet,
186 chunk);
187 }
188 break;
189
190 case SCTP_XMIT_RWND_FULL:
191 case SCTP_XMIT_OK:
192 case SCTP_XMIT_NAGLE_DELAY:
193 break;
194 }
195
196 return retval;
197 }
198
199 /* Try to bundle an auth chunk into the packet. */
sctp_packet_bundle_auth(struct sctp_packet * pkt,struct sctp_chunk * chunk)200 static sctp_xmit_t sctp_packet_bundle_auth(struct sctp_packet *pkt,
201 struct sctp_chunk *chunk)
202 {
203 struct sctp_association *asoc = pkt->transport->asoc;
204 struct sctp_chunk *auth;
205 sctp_xmit_t retval = SCTP_XMIT_OK;
206
207 /* if we don't have an association, we can't do authentication */
208 if (!asoc)
209 return retval;
210
211 /* See if this is an auth chunk we are bundling or if
212 * auth is already bundled.
213 */
214 if (chunk->chunk_hdr->type == SCTP_CID_AUTH || pkt->has_auth)
215 return retval;
216
217 /* if the peer did not request this chunk to be authenticated,
218 * don't do it
219 */
220 if (!chunk->auth)
221 return retval;
222
223 auth = sctp_make_auth(asoc);
224 if (!auth)
225 return retval;
226
227 retval = sctp_packet_append_chunk(pkt, auth);
228
229 return retval;
230 }
231
232 /* Try to bundle a SACK with the packet. */
sctp_packet_bundle_sack(struct sctp_packet * pkt,struct sctp_chunk * chunk)233 static sctp_xmit_t sctp_packet_bundle_sack(struct sctp_packet *pkt,
234 struct sctp_chunk *chunk)
235 {
236 sctp_xmit_t retval = SCTP_XMIT_OK;
237
238 /* If sending DATA and haven't aleady bundled a SACK, try to
239 * bundle one in to the packet.
240 */
241 if (sctp_chunk_is_data(chunk) && !pkt->has_sack &&
242 !pkt->has_cookie_echo) {
243 struct sctp_association *asoc;
244 struct timer_list *timer;
245 asoc = pkt->transport->asoc;
246 timer = &asoc->timers[SCTP_EVENT_TIMEOUT_SACK];
247
248 /* If the SACK timer is running, we have a pending SACK */
249 if (timer_pending(timer)) {
250 struct sctp_chunk *sack;
251 asoc->a_rwnd = asoc->rwnd;
252 sack = sctp_make_sack(asoc);
253 if (sack) {
254 retval = sctp_packet_append_chunk(pkt, sack);
255 asoc->peer.sack_needed = 0;
256 if (del_timer(timer))
257 sctp_association_put(asoc);
258 }
259 }
260 }
261 return retval;
262 }
263
264 /* Append a chunk to the offered packet reporting back any inability to do
265 * so.
266 */
sctp_packet_append_chunk(struct sctp_packet * packet,struct sctp_chunk * chunk)267 sctp_xmit_t sctp_packet_append_chunk(struct sctp_packet *packet,
268 struct sctp_chunk *chunk)
269 {
270 sctp_xmit_t retval = SCTP_XMIT_OK;
271 __u16 chunk_len = WORD_ROUND(ntohs(chunk->chunk_hdr->length));
272
273 SCTP_DEBUG_PRINTK("%s: packet:%p chunk:%p\n", __func__, packet,
274 chunk);
275
276 /* Data chunks are special. Before seeing what else we can
277 * bundle into this packet, check to see if we are allowed to
278 * send this DATA.
279 */
280 if (sctp_chunk_is_data(chunk)) {
281 retval = sctp_packet_can_append_data(packet, chunk);
282 if (retval != SCTP_XMIT_OK)
283 goto finish;
284 }
285
286 /* Try to bundle AUTH chunk */
287 retval = sctp_packet_bundle_auth(packet, chunk);
288 if (retval != SCTP_XMIT_OK)
289 goto finish;
290
291 /* Try to bundle SACK chunk */
292 retval = sctp_packet_bundle_sack(packet, chunk);
293 if (retval != SCTP_XMIT_OK)
294 goto finish;
295
296 /* Check to see if this chunk will fit into the packet */
297 retval = sctp_packet_will_fit(packet, chunk, chunk_len);
298 if (retval != SCTP_XMIT_OK)
299 goto finish;
300
301 /* We believe that this chunk is OK to add to the packet */
302 switch (chunk->chunk_hdr->type) {
303 case SCTP_CID_DATA:
304 /* Account for the data being in the packet */
305 sctp_packet_append_data(packet, chunk);
306 /* Disallow SACK bundling after DATA. */
307 packet->has_sack = 1;
308 /* Disallow AUTH bundling after DATA */
309 packet->has_auth = 1;
310 /* Let it be knows that packet has DATA in it */
311 packet->has_data = 1;
312 /* timestamp the chunk for rtx purposes */
313 chunk->sent_at = jiffies;
314 break;
315 case SCTP_CID_COOKIE_ECHO:
316 packet->has_cookie_echo = 1;
317 break;
318
319 case SCTP_CID_SACK:
320 packet->has_sack = 1;
321 break;
322
323 case SCTP_CID_AUTH:
324 packet->has_auth = 1;
325 packet->auth = chunk;
326 break;
327 }
328
329 /* It is OK to send this chunk. */
330 list_add_tail(&chunk->list, &packet->chunk_list);
331 packet->size += chunk_len;
332 chunk->transport = packet->transport;
333 finish:
334 return retval;
335 }
336
sctp_packet_release_owner(struct sk_buff * skb)337 static void sctp_packet_release_owner(struct sk_buff *skb)
338 {
339 sk_free(skb->sk);
340 }
341
sctp_packet_set_owner_w(struct sk_buff * skb,struct sock * sk)342 static void sctp_packet_set_owner_w(struct sk_buff *skb, struct sock *sk)
343 {
344 skb_orphan(skb);
345 skb->sk = sk;
346 skb->destructor = sctp_packet_release_owner;
347
348 /*
349 * The data chunks have already been accounted for in sctp_sendmsg(),
350 * therefore only reserve a single byte to keep socket around until
351 * the packet has been transmitted.
352 */
353 atomic_inc(&sk->sk_wmem_alloc);
354 }
355
356 /* All packets are sent to the network through this function from
357 * sctp_outq_tail().
358 *
359 * The return value is a normal kernel error return value.
360 */
sctp_packet_transmit(struct sctp_packet * packet)361 int sctp_packet_transmit(struct sctp_packet *packet)
362 {
363 struct sctp_transport *tp = packet->transport;
364 struct sctp_association *asoc = tp->asoc;
365 struct sctphdr *sh;
366 struct sk_buff *nskb;
367 struct sctp_chunk *chunk, *tmp;
368 struct sock *sk;
369 int err = 0;
370 int padding; /* How much padding do we need? */
371 __u8 has_data = 0;
372 struct dst_entry *dst = tp->dst;
373 unsigned char *auth = NULL; /* pointer to auth in skb data */
374 __u32 cksum_buf_len = sizeof(struct sctphdr);
375
376 SCTP_DEBUG_PRINTK("%s: packet:%p\n", __func__, packet);
377
378 /* Do NOT generate a chunkless packet. */
379 if (list_empty(&packet->chunk_list))
380 return err;
381
382 /* Set up convenience variables... */
383 chunk = list_entry(packet->chunk_list.next, struct sctp_chunk, list);
384 sk = chunk->skb->sk;
385
386 /* Allocate the new skb. */
387 nskb = alloc_skb(packet->size + LL_MAX_HEADER, GFP_ATOMIC);
388 if (!nskb)
389 goto nomem;
390
391 /* Make sure the outbound skb has enough header room reserved. */
392 skb_reserve(nskb, packet->overhead + LL_MAX_HEADER);
393
394 /* Set the owning socket so that we know where to get the
395 * destination IP address.
396 */
397 sctp_packet_set_owner_w(nskb, sk);
398
399 if (!sctp_transport_dst_check(tp)) {
400 sctp_transport_route(tp, NULL, sctp_sk(sk));
401 if (asoc && (asoc->param_flags & SPP_PMTUD_ENABLE)) {
402 sctp_assoc_sync_pmtu(asoc);
403 }
404 }
405 dst = dst_clone(tp->dst);
406 skb_dst_set(nskb, dst);
407 if (!dst)
408 goto no_route;
409
410 /* Build the SCTP header. */
411 sh = (struct sctphdr *)skb_push(nskb, sizeof(struct sctphdr));
412 skb_reset_transport_header(nskb);
413 sh->source = htons(packet->source_port);
414 sh->dest = htons(packet->destination_port);
415
416 /* From 6.8 Adler-32 Checksum Calculation:
417 * After the packet is constructed (containing the SCTP common
418 * header and one or more control or DATA chunks), the
419 * transmitter shall:
420 *
421 * 1) Fill in the proper Verification Tag in the SCTP common
422 * header and initialize the checksum field to 0's.
423 */
424 sh->vtag = htonl(packet->vtag);
425 sh->checksum = 0;
426
427 /**
428 * 6.10 Bundling
429 *
430 * An endpoint bundles chunks by simply including multiple
431 * chunks in one outbound SCTP packet. ...
432 */
433
434 /**
435 * 3.2 Chunk Field Descriptions
436 *
437 * The total length of a chunk (including Type, Length and
438 * Value fields) MUST be a multiple of 4 bytes. If the length
439 * of the chunk is not a multiple of 4 bytes, the sender MUST
440 * pad the chunk with all zero bytes and this padding is not
441 * included in the chunk length field. The sender should
442 * never pad with more than 3 bytes.
443 *
444 * [This whole comment explains WORD_ROUND() below.]
445 */
446 SCTP_DEBUG_PRINTK("***sctp_transmit_packet***\n");
447 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) {
448 list_del_init(&chunk->list);
449 if (sctp_chunk_is_data(chunk)) {
450 /* 6.3.1 C4) When data is in flight and when allowed
451 * by rule C5, a new RTT measurement MUST be made each
452 * round trip. Furthermore, new RTT measurements
453 * SHOULD be made no more than once per round-trip
454 * for a given destination transport address.
455 */
456
457 if (!tp->rto_pending) {
458 chunk->rtt_in_progress = 1;
459 tp->rto_pending = 1;
460 }
461 has_data = 1;
462 }
463
464 padding = WORD_ROUND(chunk->skb->len) - chunk->skb->len;
465 if (padding)
466 memset(skb_put(chunk->skb, padding), 0, padding);
467
468 /* if this is the auth chunk that we are adding,
469 * store pointer where it will be added and put
470 * the auth into the packet.
471 */
472 if (chunk == packet->auth)
473 auth = skb_tail_pointer(nskb);
474
475 cksum_buf_len += chunk->skb->len;
476 memcpy(skb_put(nskb, chunk->skb->len),
477 chunk->skb->data, chunk->skb->len);
478
479 SCTP_DEBUG_PRINTK("%s %p[%s] %s 0x%x, %s %d, %s %d, %s %d\n",
480 "*** Chunk", chunk,
481 sctp_cname(SCTP_ST_CHUNK(
482 chunk->chunk_hdr->type)),
483 chunk->has_tsn ? "TSN" : "No TSN",
484 chunk->has_tsn ?
485 ntohl(chunk->subh.data_hdr->tsn) : 0,
486 "length", ntohs(chunk->chunk_hdr->length),
487 "chunk->skb->len", chunk->skb->len,
488 "rtt_in_progress", chunk->rtt_in_progress);
489
490 /*
491 * If this is a control chunk, this is our last
492 * reference. Free data chunks after they've been
493 * acknowledged or have failed.
494 */
495 if (!sctp_chunk_is_data(chunk))
496 sctp_chunk_free(chunk);
497 }
498
499 /* SCTP-AUTH, Section 6.2
500 * The sender MUST calculate the MAC as described in RFC2104 [2]
501 * using the hash function H as described by the MAC Identifier and
502 * the shared association key K based on the endpoint pair shared key
503 * described by the shared key identifier. The 'data' used for the
504 * computation of the AUTH-chunk is given by the AUTH chunk with its
505 * HMAC field set to zero (as shown in Figure 6) followed by all
506 * chunks that are placed after the AUTH chunk in the SCTP packet.
507 */
508 if (auth)
509 sctp_auth_calculate_hmac(asoc, nskb,
510 (struct sctp_auth_chunk *)auth,
511 GFP_ATOMIC);
512
513 /* 2) Calculate the Adler-32 checksum of the whole packet,
514 * including the SCTP common header and all the
515 * chunks.
516 *
517 * Note: Adler-32 is no longer applicable, as has been replaced
518 * by CRC32-C as described in <draft-ietf-tsvwg-sctpcsum-02.txt>.
519 */
520 if (!sctp_checksum_disable) {
521 if (!(dst->dev->features & NETIF_F_SCTP_CSUM) ||
522 (dst_xfrm(dst) != NULL) || packet->ipfragok) {
523 __u32 crc32 = sctp_start_cksum((__u8 *)sh, cksum_buf_len);
524
525 /* 3) Put the resultant value into the checksum field in the
526 * common header, and leave the rest of the bits unchanged.
527 */
528 sh->checksum = sctp_end_cksum(crc32);
529 } else {
530 /* no need to seed pseudo checksum for SCTP */
531 nskb->ip_summed = CHECKSUM_PARTIAL;
532 nskb->csum_start = (skb_transport_header(nskb) -
533 nskb->head);
534 nskb->csum_offset = offsetof(struct sctphdr, checksum);
535 }
536 }
537
538 /* IP layer ECN support
539 * From RFC 2481
540 * "The ECN-Capable Transport (ECT) bit would be set by the
541 * data sender to indicate that the end-points of the
542 * transport protocol are ECN-capable."
543 *
544 * Now setting the ECT bit all the time, as it should not cause
545 * any problems protocol-wise even if our peer ignores it.
546 *
547 * Note: The works for IPv6 layer checks this bit too later
548 * in transmission. See IP6_ECN_flow_xmit().
549 */
550 (*tp->af_specific->ecn_capable)(nskb->sk);
551
552 /* Set up the IP options. */
553 /* BUG: not implemented
554 * For v4 this all lives somewhere in sk->sk_opt...
555 */
556
557 /* Dump that on IP! */
558 if (asoc && asoc->peer.last_sent_to != tp) {
559 /* Considering the multiple CPU scenario, this is a
560 * "correcter" place for last_sent_to. --xguo
561 */
562 asoc->peer.last_sent_to = tp;
563 }
564
565 if (has_data) {
566 struct timer_list *timer;
567 unsigned long timeout;
568
569 /* Restart the AUTOCLOSE timer when sending data. */
570 if (sctp_state(asoc, ESTABLISHED) && asoc->autoclose) {
571 timer = &asoc->timers[SCTP_EVENT_TIMEOUT_AUTOCLOSE];
572 timeout = asoc->timeouts[SCTP_EVENT_TIMEOUT_AUTOCLOSE];
573
574 if (!mod_timer(timer, jiffies + timeout))
575 sctp_association_hold(asoc);
576 }
577 }
578
579 SCTP_DEBUG_PRINTK("***sctp_transmit_packet*** skb len %d\n",
580 nskb->len);
581
582 nskb->local_df = packet->ipfragok;
583 (*tp->af_specific->sctp_xmit)(nskb, tp);
584
585 out:
586 sctp_packet_reset(packet);
587 return err;
588 no_route:
589 kfree_skb(nskb);
590 IP_INC_STATS_BH(&init_net, IPSTATS_MIB_OUTNOROUTES);
591
592 /* FIXME: Returning the 'err' will effect all the associations
593 * associated with a socket, although only one of the paths of the
594 * association is unreachable.
595 * The real failure of a transport or association can be passed on
596 * to the user via notifications. So setting this error may not be
597 * required.
598 */
599 /* err = -EHOSTUNREACH; */
600 err:
601 /* Control chunks are unreliable so just drop them. DATA chunks
602 * will get resent or dropped later.
603 */
604
605 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) {
606 list_del_init(&chunk->list);
607 if (!sctp_chunk_is_data(chunk))
608 sctp_chunk_free(chunk);
609 }
610 goto out;
611 nomem:
612 err = -ENOMEM;
613 goto err;
614 }
615
616 /********************************************************************
617 * 2nd Level Abstractions
618 ********************************************************************/
619
620 /* This private function check to see if a chunk can be added */
sctp_packet_can_append_data(struct sctp_packet * packet,struct sctp_chunk * chunk)621 static sctp_xmit_t sctp_packet_can_append_data(struct sctp_packet *packet,
622 struct sctp_chunk *chunk)
623 {
624 sctp_xmit_t retval = SCTP_XMIT_OK;
625 size_t datasize, rwnd, inflight, flight_size;
626 struct sctp_transport *transport = packet->transport;
627 struct sctp_association *asoc = transport->asoc;
628 struct sctp_outq *q = &asoc->outqueue;
629
630 /* RFC 2960 6.1 Transmission of DATA Chunks
631 *
632 * A) At any given time, the data sender MUST NOT transmit new data to
633 * any destination transport address if its peer's rwnd indicates
634 * that the peer has no buffer space (i.e. rwnd is 0, see Section
635 * 6.2.1). However, regardless of the value of rwnd (including if it
636 * is 0), the data sender can always have one DATA chunk in flight to
637 * the receiver if allowed by cwnd (see rule B below). This rule
638 * allows the sender to probe for a change in rwnd that the sender
639 * missed due to the SACK having been lost in transit from the data
640 * receiver to the data sender.
641 */
642
643 rwnd = asoc->peer.rwnd;
644 inflight = q->outstanding_bytes;
645 flight_size = transport->flight_size;
646
647 datasize = sctp_data_size(chunk);
648
649 if (datasize > rwnd) {
650 if (inflight > 0) {
651 /* We have (at least) one data chunk in flight,
652 * so we can't fall back to rule 6.1 B).
653 */
654 retval = SCTP_XMIT_RWND_FULL;
655 goto finish;
656 }
657 }
658
659 /* RFC 2960 6.1 Transmission of DATA Chunks
660 *
661 * B) At any given time, the sender MUST NOT transmit new data
662 * to a given transport address if it has cwnd or more bytes
663 * of data outstanding to that transport address.
664 */
665 /* RFC 7.2.4 & the Implementers Guide 2.8.
666 *
667 * 3) ...
668 * When a Fast Retransmit is being performed the sender SHOULD
669 * ignore the value of cwnd and SHOULD NOT delay retransmission.
670 */
671 if (chunk->fast_retransmit != SCTP_NEED_FRTX)
672 if (flight_size >= transport->cwnd) {
673 retval = SCTP_XMIT_RWND_FULL;
674 goto finish;
675 }
676
677 /* Nagle's algorithm to solve small-packet problem:
678 * Inhibit the sending of new chunks when new outgoing data arrives
679 * if any previously transmitted data on the connection remains
680 * unacknowledged.
681 */
682 if (!sctp_sk(asoc->base.sk)->nodelay && sctp_packet_empty(packet) &&
683 inflight && sctp_state(asoc, ESTABLISHED)) {
684 unsigned max = transport->pathmtu - packet->overhead;
685 unsigned len = chunk->skb->len + q->out_qlen;
686
687 /* Check whether this chunk and all the rest of pending
688 * data will fit or delay in hopes of bundling a full
689 * sized packet.
690 * Don't delay large message writes that may have been
691 * fragmeneted into small peices.
692 */
693 if ((len < max) && chunk->msg->can_delay) {
694 retval = SCTP_XMIT_NAGLE_DELAY;
695 goto finish;
696 }
697 }
698
699 finish:
700 return retval;
701 }
702
703 /* This private function does management things when adding DATA chunk */
sctp_packet_append_data(struct sctp_packet * packet,struct sctp_chunk * chunk)704 static void sctp_packet_append_data(struct sctp_packet *packet,
705 struct sctp_chunk *chunk)
706 {
707 struct sctp_transport *transport = packet->transport;
708 size_t datasize = sctp_data_size(chunk);
709 struct sctp_association *asoc = transport->asoc;
710 u32 rwnd = asoc->peer.rwnd;
711
712 /* Keep track of how many bytes are in flight over this transport. */
713 transport->flight_size += datasize;
714
715 /* Keep track of how many bytes are in flight to the receiver. */
716 asoc->outqueue.outstanding_bytes += datasize;
717
718 /* Update our view of the receiver's rwnd. */
719 if (datasize < rwnd)
720 rwnd -= datasize;
721 else
722 rwnd = 0;
723
724 asoc->peer.rwnd = rwnd;
725 /* Has been accepted for transmission. */
726 if (!asoc->peer.prsctp_capable)
727 chunk->msg->can_abandon = 0;
728 sctp_chunk_assign_tsn(chunk);
729 sctp_chunk_assign_ssn(chunk);
730 }
731
sctp_packet_will_fit(struct sctp_packet * packet,struct sctp_chunk * chunk,u16 chunk_len)732 static sctp_xmit_t sctp_packet_will_fit(struct sctp_packet *packet,
733 struct sctp_chunk *chunk,
734 u16 chunk_len)
735 {
736 size_t psize;
737 size_t pmtu;
738 int too_big;
739 sctp_xmit_t retval = SCTP_XMIT_OK;
740
741 psize = packet->size;
742 pmtu = ((packet->transport->asoc) ?
743 (packet->transport->asoc->pathmtu) :
744 (packet->transport->pathmtu));
745
746 too_big = (psize + chunk_len > pmtu);
747
748 /* Decide if we need to fragment or resubmit later. */
749 if (too_big) {
750 /* It's OK to fragmet at IP level if any one of the following
751 * is true:
752 * 1. The packet is empty (meaning this chunk is greater
753 * the MTU)
754 * 2. The chunk we are adding is a control chunk
755 * 3. The packet doesn't have any data in it yet and data
756 * requires authentication.
757 */
758 if (sctp_packet_empty(packet) || !sctp_chunk_is_data(chunk) ||
759 (!packet->has_data && chunk->auth)) {
760 /* We no longer do re-fragmentation.
761 * Just fragment at the IP layer, if we
762 * actually hit this condition
763 */
764 packet->ipfragok = 1;
765 } else {
766 retval = SCTP_XMIT_PMTU_FULL;
767 }
768 }
769
770 return retval;
771 }
772