1 /* $KAME: sctp_pcb.c,v 1.37 2004/08/17 06:28:02 t-momose Exp $ */
4 * Copyright (c) 2001, 2002, 2003, 2004 Cisco Systems, Inc.
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
10 * 1. Redistributions of source code must retain the above copyright
11 * notice, this list of conditions and the following disclaimer.
12 * 2. Redistributions in binary form must reproduce the above copyright
13 * notice, this list of conditions and the following disclaimer in the
14 * documentation and/or other materials provided with the distribution.
15 * 3. All advertising materials mentioning features or use of this software
16 * must display the following acknowledgement:
17 * This product includes software developed by Cisco Systems, Inc.
18 * 4. Neither the name of the project nor the names of its contributors
19 * may be used to endorse or promote products derived from this software
20 * without specific prior written permission.
22 * THIS SOFTWARE IS PROVIDED BY CISCO SYSTEMS AND CONTRIBUTORS ``AS IS'' AND
23 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
24 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
25 * ARE DISCLAIMED. IN NO EVENT SHALL CISCO SYSTEMS OR CONTRIBUTORS BE LIABLE
26 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
27 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
28 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
29 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
30 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
31 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
34 #if !(defined(__OpenBSD__) || defined(__APPLE__))
35 #include "opt_ipsec.h"
37 #if defined(__FreeBSD__) || defined(__DragonFly__)
38 #include "opt_compat.h"
39 #include "opt_inet6.h"
42 #if defined(__NetBSD__)
47 #elif !defined(__OpenBSD__)
51 #include <sys/param.h>
52 #include <sys/systm.h>
53 #include <sys/malloc.h>
55 #include <sys/domain.h>
56 #include <sys/protosw.h>
57 #include <sys/socket.h>
58 #include <sys/socketvar.h>
61 #include <sys/kernel.h>
62 #include <sys/sysctl.h>
63 #include <sys/thread2.h>
64 #if defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__)
65 #include <sys/random.h>
67 #if defined(__NetBSD__)
70 #if defined(__OpenBSD__)
71 #include <dev/rndvar.h>
74 #if defined(__APPLE__)
75 #include <netinet/sctp_callout.h>
76 #elif defined(__OpenBSD__)
77 #include <sys/timeout.h>
79 #include <sys/callout.h>
82 #if (defined(__FreeBSD__) && __FreeBSD_version >= 500000)
83 #include <sys/limits.h>
85 #include <machine/limits.h>
87 #include <machine/cpu.h>
90 #include <net/if_types.h>
91 #include <net/route.h>
92 #include <netinet/in.h>
93 #include <netinet/in_systm.h>
94 #include <netinet/ip.h>
95 #include <netinet/in_pcb.h>
96 #include <netinet/in_var.h>
97 #include <netinet/ip_var.h>
100 #include <netinet/ip6.h>
101 #include <netinet6/ip6_var.h>
102 #include <netinet6/scope6_var.h>
103 #if defined(__FreeBSD__) || defined(__NetBSD__) || defined(__DragonFly__)
104 #include <netinet6/in6_pcb.h>
105 #elif defined(__OpenBSD__)
106 #include <netinet/in_pcb.h>
112 #include <netinet6/ipsec.h>
113 #include <netproto/key/key.h>
119 #include <netinet/sctp_var.h>
120 #include <netinet/sctp_pcb.h>
121 #include <netinet/sctputil.h>
122 #include <netinet/sctp.h>
123 #include <netinet/sctp_header.h>
124 #include <netinet/sctp_asconf.h>
125 #include <netinet/sctp_output.h>
126 #include <netinet/sctp_timer.h>
128 #ifndef SCTP_PCBHASHSIZE
129 /* default number of association hash buckets in each endpoint */
130 #define SCTP_PCBHASHSIZE 256
134 u_int32_t sctp_debug_on = 0;
135 #endif /* SCTP_DEBUG */
137 u_int32_t sctp_pegs[SCTP_NUMBER_OF_PEGS];
139 int sctp_pcbtblsize = SCTP_PCBHASHSIZE;
141 struct sctp_epinfo sctppcbinfo;
143 /* FIX: we don't handle multiple link local scopes */
144 /* "scopeless" replacement IN6_ARE_ADDR_EQUAL */
146 SCTP6_ARE_ADDR_EQUAL(struct in6_addr *a, struct in6_addr *b)
148 struct in6_addr tmp_a, tmp_b;
149 /* use a copy of a and b */
152 in6_clearscope(&tmp_a);
153 in6_clearscope(&tmp_b);
154 return (IN6_ARE_ADDR_EQUAL(&tmp_a, &tmp_b));
158 extern int ipport_firstauto;
159 extern int ipport_lastauto;
160 extern int ipport_hifirstauto;
161 extern int ipport_hilastauto;
164 #if defined(__FreeBSD__) && __FreeBSD_version > 500000
167 void sctp_validate_no_locks(void);
170 SCTP_INP_RLOCK(struct sctp_inpcb *inp)
172 struct sctp_tcb *stcb;
173 LIST_FOREACH(stcb, &inp->sctp_asoc_list, sctp_tcblist) {
174 if (mtx_owned(&(stcb)->tcb_mtx))
175 panic("I own TCB lock?");
177 if (mtx_owned(&(inp)->inp_mtx))
178 panic("INP Recursive Lock-R");
179 mtx_lock(&(inp)->inp_mtx);
183 SCTP_INP_WLOCK(struct sctp_inpcb *inp)
189 SCTP_INP_INFO_RLOCK(void)
191 struct sctp_inpcb *inp;
192 struct sctp_tcb *stcb;
193 LIST_FOREACH(inp, &sctppcbinfo.listhead, sctp_list) {
194 if (mtx_owned(&(inp)->inp_mtx))
195 panic("info-lock and own inp lock?");
196 LIST_FOREACH(stcb, &inp->sctp_asoc_list, sctp_tcblist) {
197 if (mtx_owned(&(stcb)->tcb_mtx))
198 panic("Info lock and own a tcb lock?");
201 if (mtx_owned(&sctppcbinfo.ipi_ep_mtx))
202 panic("INP INFO Recursive Lock-R");
203 mtx_lock(&sctppcbinfo.ipi_ep_mtx);
207 SCTP_INP_INFO_WLOCK(void)
209 SCTP_INP_INFO_RLOCK();
213 void sctp_validate_no_locks(void)
215 struct sctp_inpcb *inp;
216 struct sctp_tcb *stcb;
218 if (mtx_owned(&sctppcbinfo.ipi_ep_mtx))
219 panic("INP INFO lock is owned?");
221 LIST_FOREACH(inp, &sctppcbinfo.listhead, sctp_list) {
222 if (mtx_owned(&(inp)->inp_mtx))
223 panic("You own an INP lock?");
224 LIST_FOREACH(stcb, &inp->sctp_asoc_list, sctp_tcblist) {
225 if (mtx_owned(&(stcb)->tcb_mtx))
226 panic("You own a TCB lock?");
235 sctp_fill_pcbinfo(struct sctp_pcbinfo *spcb)
237 /* We really don't need
238 * to lock this, but I will
239 * just because it does not hurt.
241 SCTP_INP_INFO_RLOCK();
242 spcb->ep_count = sctppcbinfo.ipi_count_ep;
243 spcb->asoc_count = sctppcbinfo.ipi_count_asoc;
244 spcb->laddr_count = sctppcbinfo.ipi_count_laddr;
245 spcb->raddr_count = sctppcbinfo.ipi_count_raddr;
246 spcb->chk_count = sctppcbinfo.ipi_count_chunk;
247 spcb->sockq_count = sctppcbinfo.ipi_count_sockq;
248 spcb->mbuf_track = sctppcbinfo.mbuf_track;
249 SCTP_INP_INFO_RUNLOCK();
254 * Notes on locks for FreeBSD 5 and up. All association
255 * lookups that have a definte ep, the INP structure is
256 * assumed to be locked for reading. If we need to go
257 * find the INP (ususally when a **inp is passed) then
258 * we must lock the INFO structure first and if needed
259 * lock the INP too. Note that if we lock it we must
265 * Given a endpoint, look and find in its association list any association
266 * with the "to" address given. This can be a "from" address, too, for
267 * inbound packets. For outbound packets it is a true "to" address.
269 static struct sctp_tcb *
270 sctp_tcb_special_locate(struct sctp_inpcb **inp_p, struct sockaddr *from,
271 struct sockaddr *to, struct sctp_nets **netp)
273 /**** ASSUMSES THE CALLER holds the INP_INFO_RLOCK */
276 * Note for this module care must be taken when observing what to is
277 * for. In most of the rest of the code the TO field represents my
278 * peer and the FROM field represents my address. For this module it
279 * is reversed of that.
282 * If we support the TCP model, then we must now dig through to
283 * see if we can find our endpoint in the list of tcp ep's.
285 uint16_t lport, rport;
286 struct sctppcbhead *ephead;
287 struct sctp_inpcb *inp;
288 struct sctp_laddr *laddr;
289 struct sctp_tcb *stcb;
290 struct sctp_nets *net;
292 if ((to == NULL) || (from == NULL)) {
296 if (to->sa_family == AF_INET && from->sa_family == AF_INET) {
297 lport = ((struct sockaddr_in *)to)->sin_port;
298 rport = ((struct sockaddr_in *)from)->sin_port;
299 } else if (to->sa_family == AF_INET6 && from->sa_family == AF_INET6) {
300 lport = ((struct sockaddr_in6 *)to)->sin6_port;
301 rport = ((struct sockaddr_in6 *)from)->sin6_port;
305 ephead = &sctppcbinfo.sctp_tcpephash[SCTP_PCBHASH_ALLADDR(
306 (lport + rport), sctppcbinfo.hashtcpmark)];
308 * Ok now for each of the guys in this bucket we must look
310 * - Does the remote port match.
311 * - Does there single association's addresses match this
313 * If so we update p_ep to point to this ep and return the
316 LIST_FOREACH(inp, ephead, sctp_hash) {
317 if (lport != inp->sctp_lport) {
321 /* check to see if the ep has one of the addresses */
322 if ((inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) == 0) {
323 /* We are NOT bound all, so look further */
326 LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
327 if (laddr->ifa == NULL) {
329 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
330 kprintf("An ounce of prevention is worth a pound of cure\n");
335 if (laddr->ifa->ifa_addr == NULL) {
337 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
338 kprintf("ifa with a NULL address\n");
343 if (laddr->ifa->ifa_addr->sa_family ==
345 /* see if it matches */
346 struct sockaddr_in *intf_addr, *sin;
347 intf_addr = (struct sockaddr_in *)
348 laddr->ifa->ifa_addr;
349 sin = (struct sockaddr_in *)to;
350 if (from->sa_family == AF_INET) {
351 if (sin->sin_addr.s_addr ==
352 intf_addr->sin_addr.s_addr) {
354 SCTP_INP_RUNLOCK(inp);
358 struct sockaddr_in6 *intf_addr6;
359 struct sockaddr_in6 *sin6;
360 sin6 = (struct sockaddr_in6 *)
362 intf_addr6 = (struct sockaddr_in6 *)
363 laddr->ifa->ifa_addr;
365 if (SCTP6_ARE_ADDR_EQUAL(&sin6->sin6_addr,
366 &intf_addr6->sin6_addr)) {
368 SCTP_INP_RUNLOCK(inp);
375 /* This endpoint does not have this address */
376 SCTP_INP_RUNLOCK(inp);
381 * Ok if we hit here the ep has the address, does it hold the
385 stcb = LIST_FIRST(&inp->sctp_asoc_list);
387 SCTP_INP_RUNLOCK(inp);
391 if (stcb->rport != rport) {
392 /* remote port does not match. */
393 SCTP_TCB_UNLOCK(stcb);
394 SCTP_INP_RUNLOCK(inp);
397 /* Does this TCB have a matching address? */
398 TAILQ_FOREACH(net, &stcb->asoc.nets, sctp_next) {
399 if (net->ro._l_addr.sa.sa_family != from->sa_family) {
400 /* not the same family, can't be a match */
403 if (from->sa_family == AF_INET) {
404 struct sockaddr_in *sin, *rsin;
405 sin = (struct sockaddr_in *)&net->ro._l_addr;
406 rsin = (struct sockaddr_in *)from;
407 if (sin->sin_addr.s_addr ==
408 rsin->sin_addr.s_addr) {
413 /* Update the endpoint pointer */
415 SCTP_INP_RUNLOCK(inp);
419 struct sockaddr_in6 *sin6, *rsin6;
420 sin6 = (struct sockaddr_in6 *)&net->ro._l_addr;
421 rsin6 = (struct sockaddr_in6 *)from;
422 if (SCTP6_ARE_ADDR_EQUAL(&sin6->sin6_addr,
423 &rsin6->sin6_addr)) {
428 /* Update the endpoint pointer */
430 SCTP_INP_RUNLOCK(inp);
435 SCTP_TCB_UNLOCK(stcb);
437 SCTP_INP_RUNLOCK(inp);
443 sctp_findassociation_ep_asconf(struct mbuf *m, int iphlen, int offset,
444 struct sctphdr *sh, struct sctp_inpcb **inp_p, struct sctp_nets **netp)
446 struct sctp_tcb *stcb;
447 struct sockaddr_in *sin;
448 struct sockaddr_in6 *sin6;
449 struct sockaddr_storage local_store, remote_store;
451 struct sctp_paramhdr parm_buf, *phdr;
454 memset(&local_store, 0, sizeof(local_store));
455 memset(&remote_store, 0, sizeof(remote_store));
457 /* First get the destination address setup too. */
458 iph = mtod(m, struct ip *);
459 if (iph->ip_v == IPVERSION) {
461 sin = (struct sockaddr_in *)&local_store;
462 sin->sin_family = AF_INET;
463 sin->sin_len = sizeof(*sin);
464 sin->sin_port = sh->dest_port;
465 sin->sin_addr.s_addr = iph->ip_dst.s_addr ;
466 } else if (iph->ip_v == (IPV6_VERSION >> 4)) {
469 ip6 = mtod(m, struct ip6_hdr *);
470 sin6 = (struct sockaddr_in6 *)&local_store;
471 sin6->sin6_family = AF_INET6;
472 sin6->sin6_len = sizeof(*sin6);
473 sin6->sin6_port = sh->dest_port;
474 sin6->sin6_addr = ip6->ip6_dst;
479 phdr = sctp_get_next_param(m, offset + sizeof(struct sctp_asconf_chunk),
480 &parm_buf, sizeof(struct sctp_paramhdr));
483 if (sctp_debug_on & SCTP_DEBUG_INPUT3) {
484 kprintf("sctp_process_control: failed to get asconf lookup addr\n");
486 #endif /* SCTP_DEBUG */
489 ptype = (int)((u_int)ntohs(phdr->param_type));
490 /* get the correlation address */
491 if (ptype == SCTP_IPV6_ADDRESS) {
492 /* ipv6 address param */
493 struct sctp_ipv6addr_param *p6, p6_buf;
494 if (ntohs(phdr->param_length) != sizeof(struct sctp_ipv6addr_param)) {
498 p6 = (struct sctp_ipv6addr_param *)sctp_get_next_param(m,
499 offset + sizeof(struct sctp_asconf_chunk),
500 &p6_buf.ph, sizeof(*p6));
503 if (sctp_debug_on & SCTP_DEBUG_INPUT3) {
504 kprintf("sctp_process_control: failed to get asconf v6 lookup addr\n");
506 #endif /* SCTP_DEBUG */
509 sin6 = (struct sockaddr_in6 *)&remote_store;
510 sin6->sin6_family = AF_INET6;
511 sin6->sin6_len = sizeof(*sin6);
512 sin6->sin6_port = sh->src_port;
513 memcpy(&sin6->sin6_addr, &p6->addr, sizeof(struct in6_addr));
514 } else if (ptype == SCTP_IPV4_ADDRESS) {
515 /* ipv4 address param */
516 struct sctp_ipv4addr_param *p4, p4_buf;
517 if (ntohs(phdr->param_length) != sizeof(struct sctp_ipv4addr_param)) {
521 p4 = (struct sctp_ipv4addr_param *)sctp_get_next_param(m,
522 offset + sizeof(struct sctp_asconf_chunk),
523 &p4_buf.ph, sizeof(*p4));
526 if (sctp_debug_on & SCTP_DEBUG_INPUT3) {
527 kprintf("sctp_process_control: failed to get asconf v4 lookup addr\n");
529 #endif /* SCTP_DEBUG */
532 sin = (struct sockaddr_in *)&remote_store;
533 sin->sin_family = AF_INET;
534 sin->sin_len = sizeof(*sin);
535 sin->sin_port = sh->src_port;
536 memcpy(&sin->sin_addr, &p4->addr, sizeof(struct in_addr));
538 /* invalid address param type */
542 stcb = sctp_findassociation_ep_addr(inp_p,
543 (struct sockaddr *)&remote_store, netp,
544 (struct sockaddr *)&local_store, NULL);
549 sctp_findassociation_ep_addr(struct sctp_inpcb **inp_p, struct sockaddr *remote,
550 struct sctp_nets **netp, struct sockaddr *local, struct sctp_tcb *locked_tcb)
552 struct sctpasochead *head;
553 struct sctp_inpcb *inp;
554 struct sctp_tcb *stcb;
555 struct sctp_nets *net;
559 if (remote->sa_family == AF_INET) {
560 rport = (((struct sockaddr_in *)remote)->sin_port);
561 } else if (remote->sa_family == AF_INET6) {
562 rport = (((struct sockaddr_in6 *)remote)->sin6_port);
567 /* UN-lock so we can do proper locking here
568 * this occurs when called from load_addresses_from_init.
570 SCTP_TCB_UNLOCK(locked_tcb);
572 SCTP_INP_INFO_RLOCK();
573 if (inp->sctp_flags & SCTP_PCB_FLAGS_TCPTYPE) {
575 * Now either this guy is our listner or it's the connector.
576 * If it is the one that issued the connect, then it's only
577 * chance is to be the first TCB in the list. If it is the
578 * acceptor, then do the special_lookup to hash and find the
581 if (inp->sctp_flags & SCTP_PCB_FLAGS_ACCEPTING) {
582 /* to is peer addr, from is my addr */
583 stcb = sctp_tcb_special_locate(inp_p, remote, local,
585 if ((stcb != NULL) && (locked_tcb == NULL)){
586 /* we have a locked tcb, lower refcount */
588 SCTP_INP_DECR_REF(inp);
589 SCTP_INP_WUNLOCK(inp);
591 if (locked_tcb != NULL) {
592 SCTP_INP_RLOCK(locked_tcb->sctp_ep);
593 SCTP_TCB_LOCK(locked_tcb);
594 SCTP_INP_RUNLOCK(locked_tcb->sctp_ep);
596 SCTP_TCB_UNLOCK(stcb);
598 SCTP_INP_INFO_RUNLOCK();
602 stcb = LIST_FIRST(&inp->sctp_asoc_list);
607 if (stcb->rport != rport) {
608 /* remote port does not match. */
609 SCTP_TCB_UNLOCK(stcb);
612 /* now look at the list of remote addresses */
613 TAILQ_FOREACH(net, &stcb->asoc.nets, sctp_next) {
614 if (net->ro._l_addr.sa.sa_family !=
616 /* not the same family */
619 if (remote->sa_family == AF_INET) {
620 struct sockaddr_in *sin, *rsin;
621 sin = (struct sockaddr_in *)
623 rsin = (struct sockaddr_in *)remote;
624 if (sin->sin_addr.s_addr ==
625 rsin->sin_addr.s_addr) {
630 if (locked_tcb == NULL) {
631 SCTP_INP_DECR_REF(inp);
633 SCTP_INP_WUNLOCK(inp);
634 SCTP_INP_INFO_RUNLOCK();
637 } else if (remote->sa_family == AF_INET6) {
638 struct sockaddr_in6 *sin6, *rsin6;
639 sin6 = (struct sockaddr_in6 *)&net->ro._l_addr;
640 rsin6 = (struct sockaddr_in6 *)remote;
641 if (SCTP6_ARE_ADDR_EQUAL(&sin6->sin6_addr,
642 &rsin6->sin6_addr)) {
647 if (locked_tcb == NULL) {
648 SCTP_INP_DECR_REF(inp);
650 SCTP_INP_WUNLOCK(inp);
651 SCTP_INP_INFO_RUNLOCK();
656 SCTP_TCB_UNLOCK(stcb);
660 head = &inp->sctp_tcbhash[SCTP_PCBHASH_ALLADDR(rport,
661 inp->sctp_hashmark)];
665 LIST_FOREACH(stcb, head, sctp_tcbhash) {
666 if (stcb->rport != rport) {
667 /* remote port does not match */
670 /* now look at the list of remote addresses */
672 TAILQ_FOREACH(net, &stcb->asoc.nets, sctp_next) {
673 if (net->ro._l_addr.sa.sa_family !=
675 /* not the same family */
678 if (remote->sa_family == AF_INET) {
679 struct sockaddr_in *sin, *rsin;
680 sin = (struct sockaddr_in *)
682 rsin = (struct sockaddr_in *)remote;
683 if (sin->sin_addr.s_addr ==
684 rsin->sin_addr.s_addr) {
689 if (locked_tcb == NULL) {
690 SCTP_INP_DECR_REF(inp);
692 SCTP_INP_WUNLOCK(inp);
693 SCTP_INP_INFO_RUNLOCK();
696 } else if (remote->sa_family == AF_INET6) {
697 struct sockaddr_in6 *sin6, *rsin6;
698 sin6 = (struct sockaddr_in6 *)
700 rsin6 = (struct sockaddr_in6 *)remote;
701 if (SCTP6_ARE_ADDR_EQUAL(&sin6->sin6_addr,
702 &rsin6->sin6_addr)) {
707 if (locked_tcb == NULL) {
708 SCTP_INP_DECR_REF(inp);
710 SCTP_INP_WUNLOCK(inp);
711 SCTP_INP_INFO_RUNLOCK();
716 SCTP_TCB_UNLOCK(stcb);
720 /* clean up for returning null */
722 if (locked_tcb->sctp_ep != inp) {
723 SCTP_INP_RLOCK(locked_tcb->sctp_ep);
724 SCTP_TCB_LOCK(locked_tcb);
725 SCTP_INP_RUNLOCK(locked_tcb->sctp_ep);
727 SCTP_TCB_LOCK(locked_tcb);
729 SCTP_INP_WUNLOCK(inp);
730 SCTP_INP_INFO_RUNLOCK();
736 * Find an association for a specific endpoint using the association id
737 * given out in the COMM_UP notification
740 sctp_findassociation_ep_asocid(struct sctp_inpcb *inp, caddr_t asoc_id)
743 * Use my the assoc_id to find a endpoint
745 struct sctpasochead *head;
746 struct sctp_tcb *stcb;
749 if (asoc_id == 0 || inp == NULL) {
752 SCTP_INP_INFO_RLOCK();
753 vtag = (u_int32_t)(uintptr_t)asoc_id;
754 head = &sctppcbinfo.sctp_asochash[SCTP_PCBHASH_ASOC(vtag,
755 sctppcbinfo.hashasocmark)];
758 SCTP_INP_INFO_RUNLOCK();
761 LIST_FOREACH(stcb, head, sctp_asocs) {
762 SCTP_INP_RLOCK(stcb->sctp_ep);
764 SCTP_INP_RUNLOCK(stcb->sctp_ep);
765 if (stcb->asoc.my_vtag == vtag) {
767 if (inp != stcb->sctp_ep) {
768 /* some other guy has the
769 * same vtag active (vtag collision).
771 sctp_pegs[SCTP_VTAG_BOGUS]++;
772 SCTP_TCB_UNLOCK(stcb);
775 sctp_pegs[SCTP_VTAG_EXPR]++;
776 SCTP_INP_INFO_RUNLOCK();
779 SCTP_TCB_UNLOCK(stcb);
781 SCTP_INP_INFO_RUNLOCK();
785 static struct sctp_inpcb *
786 sctp_endpoint_probe(struct sockaddr *nam, struct sctppcbhead *head,
789 struct sctp_inpcb *inp;
790 struct sockaddr_in *sin;
791 struct sockaddr_in6 *sin6;
792 struct sctp_laddr *laddr;
794 /* Endpoing probe expects
795 * that the INP_INFO is locked.
797 if (nam->sa_family == AF_INET) {
798 sin = (struct sockaddr_in *)nam;
800 } else if (nam->sa_family == AF_INET6) {
801 sin6 = (struct sockaddr_in6 *)nam;
804 /* unsupported family */
810 LIST_FOREACH(inp, head, sctp_hash) {
813 if ((inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) &&
814 (inp->sctp_lport == lport)) {
816 if ((nam->sa_family == AF_INET) &&
817 (inp->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) &&
818 #if defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__)
819 (((struct inpcb *)inp)->inp_flags & IN6P_IPV6_V6ONLY)
821 #if defined(__OpenBSD__)
822 (0) /* For open bsd we do dual bind only */
824 (((struct in6pcb *)inp)->in6p_flags & IN6P_IPV6_V6ONLY)
828 /* IPv4 on a IPv6 socket with ONLY IPv6 set */
829 SCTP_INP_RUNLOCK(inp);
832 /* A V6 address and the endpoint is NOT bound V6 */
833 if (nam->sa_family == AF_INET6 &&
834 (inp->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) == 0) {
835 SCTP_INP_RUNLOCK(inp);
838 SCTP_INP_RUNLOCK(inp);
841 SCTP_INP_RUNLOCK(inp);
844 if ((nam->sa_family == AF_INET) &&
845 (sin->sin_addr.s_addr == INADDR_ANY)) {
846 /* Can't hunt for one that has no address specified */
848 } else if ((nam->sa_family == AF_INET6) &&
849 (IN6_IS_ADDR_UNSPECIFIED(&sin6->sin6_addr))) {
850 /* Can't hunt for one that has no address specified */
854 * ok, not bound to all so see if we can find a EP bound to this
858 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
859 kprintf("Ok, there is NO bound-all available for port:%x\n", ntohs(lport));
862 LIST_FOREACH(inp, head, sctp_hash) {
864 if ((inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL)) {
865 SCTP_INP_RUNLOCK(inp);
869 * Ok this could be a likely candidate, look at all of
872 if (inp->sctp_lport != lport) {
873 SCTP_INP_RUNLOCK(inp);
877 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
878 kprintf("Ok, found matching local port\n");
881 LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
882 if (laddr->ifa == NULL) {
884 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
885 kprintf("An ounce of prevention is worth a pound of cure\n");
891 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
892 kprintf("Ok laddr->ifa:%p is possible, ",
896 if (laddr->ifa->ifa_addr == NULL) {
898 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
899 kprintf("Huh IFA as an ifa_addr=NULL, ");
905 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
906 kprintf("Ok laddr->ifa:%p is possible, ",
907 laddr->ifa->ifa_addr);
908 sctp_print_address(laddr->ifa->ifa_addr);
909 kprintf("looking for ");
910 sctp_print_address(nam);
913 if (laddr->ifa->ifa_addr->sa_family == nam->sa_family) {
914 /* possible, see if it matches */
915 struct sockaddr_in *intf_addr;
916 intf_addr = (struct sockaddr_in *)
917 laddr->ifa->ifa_addr;
918 if (nam->sa_family == AF_INET) {
919 if (sin->sin_addr.s_addr ==
920 intf_addr->sin_addr.s_addr) {
922 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
923 kprintf("YES, return ep:%p\n", inp);
926 SCTP_INP_RUNLOCK(inp);
929 } else if (nam->sa_family == AF_INET6) {
930 struct sockaddr_in6 *intf_addr6;
931 intf_addr6 = (struct sockaddr_in6 *)
932 laddr->ifa->ifa_addr;
933 if (SCTP6_ARE_ADDR_EQUAL(&sin6->sin6_addr,
934 &intf_addr6->sin6_addr)) {
936 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
937 kprintf("YES, return ep:%p\n", inp);
940 SCTP_INP_RUNLOCK(inp);
945 SCTP_INP_RUNLOCK(inp);
949 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
950 kprintf("NO, Falls out to NULL\n");
958 sctp_pcb_findep(struct sockaddr *nam, int find_tcp_pool, int have_lock)
961 * First we check the hash table to see if someone has this port
962 * bound with just the port.
964 struct sctp_inpcb *inp;
965 struct sctppcbhead *head;
968 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
969 kprintf("Looking for endpoint %d :",
970 ntohs(((struct sockaddr_in *)nam)->sin_port));
971 sctp_print_address(nam);
974 if (nam->sa_family == AF_INET) {
975 lport = ((struct sockaddr_in *)nam)->sin_port;
976 } else if (nam->sa_family == AF_INET6) {
977 lport = ((struct sockaddr_in6 *)nam)->sin6_port;
979 /* unsupported family */
983 * I could cheat here and just cast to one of the types but we will
984 * do it right. It also provides the check against an Unsupported
987 /* Find the head of the ALLADDR chain */
989 SCTP_INP_INFO_RLOCK();
990 head = &sctppcbinfo.sctp_ephash[SCTP_PCBHASH_ALLADDR(lport,
991 sctppcbinfo.hashmark)];
993 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
994 kprintf("Main hash to lookup at head:%p\n", head);
997 inp = sctp_endpoint_probe(nam, head, lport);
1000 * If the TCP model exists it could be that the main listening
1001 * endpoint is gone but there exists a connected socket for this
1002 * guy yet. If so we can return the first one that we find. This
1003 * may NOT be the correct one but the sctp_findassociation_ep_addr
1004 * has further code to look at all TCP models.
1006 if (inp == NULL && find_tcp_pool) {
1009 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1010 kprintf("EP was NULL and TCP model is supported\n");
1013 for (i = 0; i < sctppcbinfo.hashtblsize; i++) {
1015 * This is real gross, but we do NOT have a remote
1016 * port at this point depending on who is calling. We
1017 * must therefore look for ANY one that matches our
1020 head = &sctppcbinfo.sctp_tcpephash[i];
1021 if (LIST_FIRST(head)) {
1022 inp = sctp_endpoint_probe(nam, head, lport);
1031 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1032 kprintf("EP to return is %p\n", inp);
1035 if (have_lock == 0) {
1037 SCTP_INP_WLOCK(inp);
1038 SCTP_INP_INCR_REF(inp);
1039 SCTP_INP_WUNLOCK(inp);
1041 SCTP_INP_INFO_RUNLOCK();
1044 SCTP_INP_WLOCK(inp);
1045 SCTP_INP_INCR_REF(inp);
1046 SCTP_INP_WUNLOCK(inp);
1053 * Find an association for an endpoint with the pointer to whom you want
1054 * to send to and the endpoint pointer. The address can be IPv4 or IPv6.
1055 * We may need to change the *to to some other struct like a mbuf...
1058 sctp_findassociation_addr_sa(struct sockaddr *to, struct sockaddr *from,
1059 struct sctp_inpcb **inp_p, struct sctp_nets **netp, int find_tcp_pool)
1061 struct sctp_inpcb *inp;
1062 struct sctp_tcb *retval;
1064 SCTP_INP_INFO_RLOCK();
1065 if (find_tcp_pool) {
1066 if (inp_p != NULL) {
1067 retval = sctp_tcb_special_locate(inp_p, from, to, netp);
1069 retval = sctp_tcb_special_locate(&inp, from, to, netp);
1071 if (retval != NULL) {
1072 SCTP_INP_INFO_RUNLOCK();
1076 inp = sctp_pcb_findep(to, 0, 1);
1077 if (inp_p != NULL) {
1080 SCTP_INP_INFO_RUNLOCK();
1087 * ok, we have an endpoint, now lets find the assoc for it (if any)
1088 * we now place the source address or from in the to of the find
1089 * endpoint call. Since in reality this chain is used from the
1090 * inbound packet side.
1092 if (inp_p != NULL) {
1093 return (sctp_findassociation_ep_addr(inp_p, from, netp, to, NULL));
1095 return (sctp_findassociation_ep_addr(&inp, from, netp, to, NULL));
1101 * This routine will grub through the mbuf that is a INIT or INIT-ACK and
1102 * find all addresses that the sender has specified in any address list.
1103 * Each address will be used to lookup the TCB and see if one exits.
1105 static struct sctp_tcb *
1106 sctp_findassociation_special_addr(struct mbuf *m, int iphlen, int offset,
1107 struct sctphdr *sh, struct sctp_inpcb **inp_p, struct sctp_nets **netp,
1108 struct sockaddr *dest)
1110 struct sockaddr_in sin4;
1111 struct sockaddr_in6 sin6;
1112 struct sctp_paramhdr *phdr, parm_buf;
1113 struct sctp_tcb *retval;
1114 u_int32_t ptype, plen;
1116 memset(&sin4, 0, sizeof(sin4));
1117 memset(&sin6, 0, sizeof(sin6));
1118 sin4.sin_len = sizeof(sin4);
1119 sin4.sin_family = AF_INET;
1120 sin4.sin_port = sh->src_port;
1121 sin6.sin6_len = sizeof(sin6);
1122 sin6.sin6_family = AF_INET6;
1123 sin6.sin6_port = sh->src_port;
1126 offset += sizeof(struct sctp_init_chunk);
1128 phdr = sctp_get_next_param(m, offset, &parm_buf, sizeof(parm_buf));
1129 while (phdr != NULL) {
1130 /* now we must see if we want the parameter */
1131 ptype = ntohs(phdr->param_type);
1132 plen = ntohs(phdr->param_length);
1135 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1136 kprintf("sctp_findassociation_special_addr: Impossible length in parameter\n");
1138 #endif /* SCTP_DEBUG */
1141 if (ptype == SCTP_IPV4_ADDRESS &&
1142 plen == sizeof(struct sctp_ipv4addr_param)) {
1143 /* Get the rest of the address */
1144 struct sctp_ipv4addr_param ip4_parm, *p4;
1146 phdr = sctp_get_next_param(m, offset,
1147 (struct sctp_paramhdr *)&ip4_parm, plen);
1151 p4 = (struct sctp_ipv4addr_param *)phdr;
1152 memcpy(&sin4.sin_addr, &p4->addr, sizeof(p4->addr));
1154 retval = sctp_findassociation_ep_addr(inp_p,
1155 (struct sockaddr *)&sin4, netp, dest, NULL);
1156 if (retval != NULL) {
1159 } else if (ptype == SCTP_IPV6_ADDRESS &&
1160 plen == sizeof(struct sctp_ipv6addr_param)) {
1161 /* Get the rest of the address */
1162 struct sctp_ipv6addr_param ip6_parm, *p6;
1164 phdr = sctp_get_next_param(m, offset,
1165 (struct sctp_paramhdr *)&ip6_parm, plen);
1169 p6 = (struct sctp_ipv6addr_param *)phdr;
1170 memcpy(&sin6.sin6_addr, &p6->addr, sizeof(p6->addr));
1172 retval = sctp_findassociation_ep_addr(inp_p,
1173 (struct sockaddr *)&sin6, netp, dest, NULL);
1174 if (retval != NULL) {
1178 offset += SCTP_SIZE32(plen);
1179 phdr = sctp_get_next_param(m, offset, &parm_buf,
1185 static struct sctp_tcb *
1186 sctp_findassoc_by_vtag(struct sockaddr *from, uint32_t vtag,
1187 struct sctp_inpcb **inp_p, struct sctp_nets **netp, uint16_t rport,
1191 * Use my vtag to hash. If we find it we then verify the source addr
1192 * is in the assoc. If all goes well we save a bit on rec of a packet.
1194 struct sctpasochead *head;
1195 struct sctp_nets *net;
1196 struct sctp_tcb *stcb;
1198 SCTP_INP_INFO_RLOCK();
1199 head = &sctppcbinfo.sctp_asochash[SCTP_PCBHASH_ASOC(vtag,
1200 sctppcbinfo.hashasocmark)];
1203 SCTP_INP_INFO_RUNLOCK();
1206 LIST_FOREACH(stcb, head, sctp_asocs) {
1207 SCTP_INP_RLOCK(stcb->sctp_ep);
1208 SCTP_TCB_LOCK(stcb);
1209 SCTP_INP_RUNLOCK(stcb->sctp_ep);
1210 if (stcb->asoc.my_vtag == vtag) {
1212 if (stcb->rport != rport) {
1214 * we could remove this if vtags are unique
1215 * across the system.
1217 SCTP_TCB_UNLOCK(stcb);
1220 if (stcb->sctp_ep->sctp_lport != lport) {
1222 * we could remove this if vtags are unique
1223 * across the system.
1225 SCTP_TCB_UNLOCK(stcb);
1228 net = sctp_findnet(stcb, from);
1232 sctp_pegs[SCTP_VTAG_EXPR]++;
1233 *inp_p = stcb->sctp_ep;
1234 SCTP_INP_INFO_RUNLOCK();
1237 /* not him, this should only
1238 * happen in rare cases so
1241 sctp_pegs[SCTP_VTAG_BOGUS]++;
1244 SCTP_TCB_UNLOCK(stcb);
1246 SCTP_INP_INFO_RUNLOCK();
1251 * Find an association with the pointer to the inbound IP packet. This
1252 * can be a IPv4 or IPv6 packet.
1255 sctp_findassociation_addr(struct mbuf *m, int iphlen, int offset,
1256 struct sctphdr *sh, struct sctp_chunkhdr *ch,
1257 struct sctp_inpcb **inp_p, struct sctp_nets **netp)
1261 struct sctp_tcb *retval;
1262 struct sockaddr_storage to_store, from_store;
1263 struct sockaddr *to = (struct sockaddr *)&to_store;
1264 struct sockaddr *from = (struct sockaddr *)&from_store;
1265 struct sctp_inpcb *inp;
1268 iph = mtod(m, struct ip *);
1269 if (iph->ip_v == IPVERSION) {
1271 struct sockaddr_in *to4, *from4;
1273 to4 = (struct sockaddr_in *)&to_store;
1274 from4 = (struct sockaddr_in *)&from_store;
1275 bzero(to4, sizeof(*to4));
1276 bzero(from4, sizeof(*from4));
1277 from4->sin_family = to4->sin_family = AF_INET;
1278 from4->sin_len = to4->sin_len = sizeof(struct sockaddr_in);
1279 from4->sin_addr.s_addr = iph->ip_src.s_addr;
1280 to4->sin_addr.s_addr = iph->ip_dst.s_addr ;
1281 from4->sin_port = sh->src_port;
1282 to4->sin_port = sh->dest_port;
1283 } else if (iph->ip_v == (IPV6_VERSION >> 4)) {
1285 struct ip6_hdr *ip6;
1286 struct sockaddr_in6 *to6, *from6;
1288 ip6 = mtod(m, struct ip6_hdr *);
1289 to6 = (struct sockaddr_in6 *)&to_store;
1290 from6 = (struct sockaddr_in6 *)&from_store;
1291 bzero(to6, sizeof(*to6));
1292 bzero(from6, sizeof(*from6));
1293 from6->sin6_family = to6->sin6_family = AF_INET6;
1294 from6->sin6_len = to6->sin6_len = sizeof(struct sockaddr_in6);
1295 to6->sin6_addr = ip6->ip6_dst;
1296 from6->sin6_addr = ip6->ip6_src;
1297 from6->sin6_port = sh->src_port;
1298 to6->sin6_port = sh->dest_port;
1299 /* Get the scopes in properly to the sin6 addr's */
1300 in6_recoverscope(to6, &to6->sin6_addr, NULL);
1301 #if defined(SCTP_BASE_FREEBSD) || defined(__APPLE__) || defined(__DragonFly__)
1302 in6_embedscope(&to6->sin6_addr, to6, NULL, NULL);
1304 in6_embedscope(&to6->sin6_addr, to6);
1307 in6_recoverscope(from6, &from6->sin6_addr, NULL);
1308 #if defined(SCTP_BASE_FREEBSD) || defined(__APPLE__) || defined(__DragonFly__)
1309 in6_embedscope(&from6->sin6_addr, from6, NULL, NULL);
1311 in6_embedscope(&from6->sin6_addr, from6);
1314 /* Currently not supported. */
1318 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1319 kprintf("Looking for port %d address :",
1320 ntohs(((struct sockaddr_in *)to)->sin_port));
1321 sctp_print_address(to);
1322 kprintf("From for port %d address :",
1323 ntohs(((struct sockaddr_in *)from)->sin_port));
1324 sctp_print_address(from);
1329 /* we only go down this path if vtag is non-zero */
1330 retval = sctp_findassoc_by_vtag(from, ntohl(sh->v_tag),
1331 inp_p, netp, sh->src_port, sh->dest_port);
1337 if ((ch->chunk_type != SCTP_INITIATION) &&
1338 (ch->chunk_type != SCTP_INITIATION_ACK) &&
1339 (ch->chunk_type != SCTP_COOKIE_ACK) &&
1340 (ch->chunk_type != SCTP_COOKIE_ECHO)) {
1341 /* Other chunk types go to the tcp pool. */
1345 retval = sctp_findassociation_addr_sa(to, from, inp_p, netp,
1349 retval = sctp_findassociation_addr_sa(to, from, &inp, netp,
1353 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1354 kprintf("retval:%p inp:%p\n", retval, inp);
1357 if (retval == NULL && inp) {
1358 /* Found a EP but not this address */
1360 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1361 kprintf("Found endpoint %p but no asoc - ep state:%x\n",
1362 inp, inp->sctp_flags);
1365 if ((ch->chunk_type == SCTP_INITIATION) ||
1366 (ch->chunk_type == SCTP_INITIATION_ACK)) {
1368 * special hook, we do NOT return linp or an
1369 * association that is linked to an existing
1370 * association that is under the TCP pool (i.e. no
1371 * listener exists). The endpoint finding routine
1372 * will always find a listner before examining the
1375 if (inp->sctp_flags & SCTP_PCB_FLAGS_IN_TCPPOOL) {
1377 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1378 kprintf("Gak, its in the TCP pool... return NULL");
1387 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1388 kprintf("Now doing SPECIAL find\n");
1391 retval = sctp_findassociation_special_addr(m, iphlen,
1392 offset, sh, inp_p, netp, to);
1396 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1397 kprintf("retval is %p\n", retval);
1403 extern int sctp_max_burst_default;
1405 extern unsigned int sctp_delayed_sack_time_default;
1406 extern unsigned int sctp_heartbeat_interval_default;
1407 extern unsigned int sctp_pmtu_raise_time_default;
1408 extern unsigned int sctp_shutdown_guard_time_default;
1409 extern unsigned int sctp_secret_lifetime_default;
1411 extern unsigned int sctp_rto_max_default;
1412 extern unsigned int sctp_rto_min_default;
1413 extern unsigned int sctp_rto_initial_default;
1414 extern unsigned int sctp_init_rto_max_default;
1415 extern unsigned int sctp_valid_cookie_life_default;
1416 extern unsigned int sctp_init_rtx_max_default;
1417 extern unsigned int sctp_assoc_rtx_max_default;
1418 extern unsigned int sctp_path_rtx_max_default;
1419 extern unsigned int sctp_nr_outgoing_streams_default;
1422 * allocate a sctp_inpcb and setup a temporary binding to a port/all
1423 * addresses. This way if we don't get a bind we by default pick a ephemeral
1424 * port with all addresses bound.
1427 sctp_inpcb_alloc(struct socket *so)
1430 * we get called when a new endpoint starts up. We need to allocate
1431 * the sctp_inpcb structure from the zone and init it. Mark it as
1432 * unbound and find a port that we can use as an ephemeral with
1433 * INADDR_ANY. If the user binds later no problem we can then add
1434 * in the specific addresses. And setup the default parameters for
1438 struct sctp_inpcb *inp, *n_inp;
1440 struct timeval time;
1446 * This code audits the entire INP list to see if
1447 * any ep's that are in the GONE state are now
1448 * all free. This should not happen really since when
1449 * the last association if freed we should end up deleting
1450 * the inpcb. This code including the locks should
1451 * be taken out ... since the last set of fixes I
1452 * have not seen the "Found a GONE on list" has not
1453 * came out. But i am paranoid and we will leave this
1454 * in at the cost of efficency on allocation of PCB's.
1455 * Probably we should move this to the invariant
1458 /* #ifdef INVARIANTS*/
1459 SCTP_INP_INFO_RLOCK();
1460 inp = LIST_FIRST(&sctppcbinfo.listhead);
1462 n_inp = LIST_NEXT(inp, sctp_list);
1463 if (inp->sctp_flags & SCTP_PCB_FLAGS_SOCKET_GONE) {
1464 if (LIST_FIRST(&inp->sctp_asoc_list) == NULL) {
1465 /* finish the job now */
1466 kprintf("Found a GONE on list\n");
1467 SCTP_INP_INFO_RUNLOCK();
1468 sctp_inpcb_free(inp, 1);
1469 SCTP_INP_INFO_RLOCK();
1474 SCTP_INP_INFO_RUNLOCK();
1475 /* #endif INVARIANTS*/
1477 SCTP_INP_INFO_WLOCK();
1478 inp = (struct sctp_inpcb *)SCTP_ZONE_GET(sctppcbinfo.ipi_zone_ep);
1480 kprintf("Out of SCTP-INPCB structures - no resources\n");
1481 SCTP_INP_INFO_WUNLOCK();
1486 bzero(inp, sizeof(*inp));
1488 /* bump generations */
1489 inp->ip_inp.inp.inp_socket = so;
1491 /* setup socket pointers */
1492 inp->sctp_socket = so;
1494 /* setup inpcb socket too */
1495 inp->ip_inp.inp.inp_socket = so;
1496 inp->sctp_frag_point = SCTP_DEFAULT_MAXSEGMENT;
1498 #if !(defined(__OpenBSD__) || defined(__APPLE__))
1500 struct inpcbpolicy *pcb_sp = NULL;
1501 error = ipsec_init_policy(so, &pcb_sp);
1502 /* Arrange to share the policy */
1503 inp->ip_inp.inp.inp_sp = pcb_sp;
1504 ((struct in6pcb *)(&inp->ip_inp.inp))->in6p_sp = pcb_sp;
1507 /* not sure what to do for openbsd here */
1511 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_ep, inp);
1512 SCTP_INP_INFO_WUNLOCK();
1516 sctppcbinfo.ipi_count_ep++;
1517 #if defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__)
1518 inp->ip_inp.inp.inp_gencnt = ++sctppcbinfo.ipi_gencnt_ep;
1519 inp->ip_inp.inp.inp_ip_ttl = ip_defttl;
1521 inp->inp_ip_ttl = ip_defttl;
1522 inp->inp_ip_tos = 0;
1525 so->so_pcb = (caddr_t)inp;
1527 if ((so->so_type == SOCK_DGRAM) ||
1528 (so->so_type == SOCK_SEQPACKET)) {
1529 /* UDP style socket */
1530 inp->sctp_flags = (SCTP_PCB_FLAGS_UDPTYPE |
1531 SCTP_PCB_FLAGS_UNBOUND);
1532 inp->sctp_flags |= (SCTP_PCB_FLAGS_RECVDATAIOEVNT);
1533 } else if (so->so_type == SOCK_STREAM) {
1534 /* TCP style socket */
1535 inp->sctp_flags = (SCTP_PCB_FLAGS_TCPTYPE |
1536 SCTP_PCB_FLAGS_UNBOUND);
1537 inp->sctp_flags |= (SCTP_PCB_FLAGS_RECVDATAIOEVNT);
1540 * unsupported socket type (RAW, etc)- in case we missed
1543 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_ep, inp);
1544 SCTP_INP_INFO_WUNLOCK();
1545 return (EOPNOTSUPP);
1547 inp->sctp_tcbhash = hashinit(sctp_pcbtblsize,
1552 #if defined(__NetBSD__) || defined(__OpenBSD__)
1555 &inp->sctp_hashmark);
1556 if (inp->sctp_tcbhash == NULL) {
1557 kprintf("Out of SCTP-INPCB->hashinit - no resources\n");
1558 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_ep, inp);
1559 SCTP_INP_INFO_WUNLOCK();
1563 SCTP_INP_LOCK_INIT(inp);
1564 SCTP_ASOC_CREATE_LOCK_INIT(inp);
1565 /* lock the new ep */
1566 SCTP_INP_WLOCK(inp);
1568 /* add it to the info area */
1569 LIST_INSERT_HEAD(&sctppcbinfo.listhead, inp, sctp_list);
1570 SCTP_INP_INFO_WUNLOCK();
1572 LIST_INIT(&inp->sctp_addr_list);
1573 LIST_INIT(&inp->sctp_asoc_list);
1574 TAILQ_INIT(&inp->sctp_queue_list);
1575 /* Init the timer structure for signature change */
1576 #if defined (__FreeBSD__) && __FreeBSD_version >= 500000
1577 callout_init(&inp->sctp_ep.signature_change.timer, 0);
1579 callout_init(&inp->sctp_ep.signature_change.timer);
1581 inp->sctp_ep.signature_change.type = SCTP_TIMER_TYPE_NEWCOOKIE;
1583 /* now init the actual endpoint default data */
1586 /* setup the base timeout information */
1587 m->sctp_timeoutticks[SCTP_TIMER_SEND] = SEC_TO_TICKS(SCTP_SEND_SEC); /* needed ? */
1588 m->sctp_timeoutticks[SCTP_TIMER_INIT] = SEC_TO_TICKS(SCTP_INIT_SEC); /* needed ? */
1589 m->sctp_timeoutticks[SCTP_TIMER_RECV] = MSEC_TO_TICKS(sctp_delayed_sack_time_default);
1590 m->sctp_timeoutticks[SCTP_TIMER_HEARTBEAT] = sctp_heartbeat_interval_default; /* this is in MSEC */
1591 m->sctp_timeoutticks[SCTP_TIMER_PMTU] = SEC_TO_TICKS(sctp_pmtu_raise_time_default);
1592 m->sctp_timeoutticks[SCTP_TIMER_MAXSHUTDOWN] = SEC_TO_TICKS(sctp_shutdown_guard_time_default);
1593 m->sctp_timeoutticks[SCTP_TIMER_SIGNATURE] = SEC_TO_TICKS(sctp_secret_lifetime_default);
1594 /* all max/min max are in ms */
1595 m->sctp_maxrto = sctp_rto_max_default;
1596 m->sctp_minrto = sctp_rto_min_default;
1597 m->initial_rto = sctp_rto_initial_default;
1598 m->initial_init_rto_max = sctp_init_rto_max_default;
1600 m->max_open_streams_intome = MAX_SCTP_STREAMS;
1602 m->max_init_times = sctp_init_rtx_max_default;
1603 m->max_send_times = sctp_assoc_rtx_max_default;
1604 m->def_net_failure = sctp_path_rtx_max_default;
1605 m->sctp_sws_sender = SCTP_SWS_SENDER_DEF;
1606 m->sctp_sws_receiver = SCTP_SWS_RECEIVER_DEF;
1607 m->max_burst = sctp_max_burst_default;
1608 /* number of streams to pre-open on a association */
1609 m->pre_open_stream_count = sctp_nr_outgoing_streams_default;
1611 /* Add adaption cookie */
1612 m->adaption_layer_indicator = 0x504C5253;
1614 /* seed random number generator */
1615 m->random_counter = 1;
1616 m->store_at = SCTP_SIGNATURE_SIZE;
1617 #if (defined(__FreeBSD__) && (__FreeBSD_version < 500000)) || defined(__DragonFly__)
1618 read_random_unlimited(m->random_numbers, sizeof(m->random_numbers));
1619 #elif defined(__APPLE__) || (defined(__FreeBSD__) && (__FreeBSD_version > 500000))
1620 read_random(m->random_numbers, sizeof(m->random_numbers));
1621 #elif defined(__OpenBSD__)
1622 get_random_bytes(m->random_numbers, sizeof(m->random_numbers));
1623 #elif defined(__NetBSD__) && NRND > 0
1624 rnd_extract_data(m->random_numbers, sizeof(m->random_numbers),
1628 u_int32_t *ranm, *ranp;
1629 ranp = (u_int32_t *)&m->random_numbers;
1630 ranm = ranp + (SCTP_SIGNATURE_ALOC_SIZE/sizeof(u_int32_t));
1631 if ((u_long)ranp % 4) {
1632 /* not a even boundary? */
1633 ranp = (u_int32_t *)SCTP_SIZE32((u_long)ranp);
1635 while (ranp < ranm) {
1641 sctp_fill_random_store(m);
1643 /* Minimum cookie size */
1644 m->size_of_a_cookie = (sizeof(struct sctp_init_msg) * 2) +
1645 sizeof(struct sctp_state_cookie);
1646 m->size_of_a_cookie += SCTP_SIGNATURE_SIZE;
1648 /* Setup the initial secret */
1649 SCTP_GETTIME_TIMEVAL(&time);
1650 m->time_of_secret_change = time.tv_sec;
1652 for (i = 0; i < SCTP_NUMBER_OF_SECRETS; i++) {
1653 m->secret_key[0][i] = sctp_select_initial_TSN(m);
1655 sctp_timer_start(SCTP_TIMER_TYPE_NEWCOOKIE, inp, NULL, NULL);
1657 /* How long is a cookie good for ? */
1658 m->def_cookie_life = sctp_valid_cookie_life_default;
1659 SCTP_INP_WUNLOCK(inp);
1665 sctp_move_pcb_and_assoc(struct sctp_inpcb *old_inp, struct sctp_inpcb *new_inp,
1666 struct sctp_tcb *stcb)
1668 uint16_t lport, rport;
1669 struct sctppcbhead *head;
1670 struct sctp_laddr *laddr, *oladdr;
1672 SCTP_TCB_UNLOCK(stcb);
1673 SCTP_INP_INFO_WLOCK();
1674 SCTP_INP_WLOCK(old_inp);
1675 SCTP_INP_WLOCK(new_inp);
1676 SCTP_TCB_LOCK(stcb);
1678 new_inp->sctp_ep.time_of_secret_change =
1679 old_inp->sctp_ep.time_of_secret_change;
1680 memcpy(new_inp->sctp_ep.secret_key, old_inp->sctp_ep.secret_key,
1681 sizeof(old_inp->sctp_ep.secret_key));
1682 new_inp->sctp_ep.current_secret_number =
1683 old_inp->sctp_ep.current_secret_number;
1684 new_inp->sctp_ep.last_secret_number =
1685 old_inp->sctp_ep.last_secret_number;
1686 new_inp->sctp_ep.size_of_a_cookie = old_inp->sctp_ep.size_of_a_cookie;
1688 /* Copy the port across */
1689 lport = new_inp->sctp_lport = old_inp->sctp_lport;
1690 rport = stcb->rport;
1691 /* Pull the tcb from the old association */
1692 LIST_REMOVE(stcb, sctp_tcbhash);
1693 LIST_REMOVE(stcb, sctp_tcblist);
1695 /* Now insert the new_inp into the TCP connected hash */
1696 head = &sctppcbinfo.sctp_tcpephash[SCTP_PCBHASH_ALLADDR((lport + rport),
1697 sctppcbinfo.hashtcpmark)];
1699 LIST_INSERT_HEAD(head, new_inp, sctp_hash);
1701 /* Now move the tcb into the endpoint list */
1702 LIST_INSERT_HEAD(&new_inp->sctp_asoc_list, stcb, sctp_tcblist);
1704 * Question, do we even need to worry about the ep-hash since
1705 * we only have one connection? Probably not :> so lets
1706 * get rid of it and not suck up any kernel memory in that.
1708 SCTP_INP_INFO_WUNLOCK();
1709 stcb->sctp_socket = new_inp->sctp_socket;
1710 stcb->sctp_ep = new_inp;
1711 if (new_inp->sctp_tcbhash != NULL) {
1712 hashdestroy(new_inp->sctp_tcbhash, M_PCB,
1713 new_inp->sctp_hashmark);
1714 new_inp->sctp_tcbhash = NULL;
1716 if ((new_inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) == 0) {
1717 /* Subset bound, so copy in the laddr list from the old_inp */
1718 LIST_FOREACH(oladdr, &old_inp->sctp_addr_list, sctp_nxt_addr) {
1719 laddr = (struct sctp_laddr *)SCTP_ZONE_GET(
1720 sctppcbinfo.ipi_zone_laddr);
1721 if (laddr == NULL) {
1723 * Gak, what can we do? This assoc is really
1724 * HOSED. We probably should send an abort
1728 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1729 kprintf("Association hosed in TCP model, out of laddr memory\n");
1731 #endif /* SCTP_DEBUG */
1734 sctppcbinfo.ipi_count_laddr++;
1735 sctppcbinfo.ipi_gencnt_laddr++;
1736 bzero(laddr, sizeof(*laddr));
1737 laddr->ifa = oladdr->ifa;
1738 LIST_INSERT_HEAD(&new_inp->sctp_addr_list, laddr,
1740 new_inp->laddr_count++;
1743 SCTP_INP_WUNLOCK(new_inp);
1744 SCTP_INP_WUNLOCK(old_inp);
1748 sctp_isport_inuse(struct sctp_inpcb *inp, uint16_t lport)
1750 struct sctppcbhead *head;
1751 struct sctp_inpcb *t_inp;
1753 head = &sctppcbinfo.sctp_ephash[SCTP_PCBHASH_ALLADDR(lport,
1754 sctppcbinfo.hashmark)];
1755 LIST_FOREACH(t_inp, head, sctp_hash) {
1756 if (t_inp->sctp_lport != lport) {
1759 /* This one is in use. */
1760 /* check the v6/v4 binding issue */
1761 if ((t_inp->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) &&
1762 #if defined(__FreeBSD__)
1763 (((struct inpcb *)t_inp)->inp_flags & IN6P_IPV6_V6ONLY)
1765 #if defined(__OpenBSD__)
1766 (0) /* For open bsd we do dual bind only */
1768 (((struct in6pcb *)t_inp)->in6p_flags & IN6P_IPV6_V6ONLY)
1772 if (inp->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) {
1773 /* collision in V6 space */
1776 /* inp is BOUND_V4 no conflict */
1779 } else if (t_inp->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) {
1780 /* t_inp is bound v4 and v6, conflict always */
1783 /* t_inp is bound only V4 */
1784 if ((inp->sctp_flags & SCTP_PCB_FLAGS_BOUND_V6) &&
1785 #if defined(__FreeBSD__)
1786 (((struct inpcb *)inp)->inp_flags & IN6P_IPV6_V6ONLY)
1788 #if defined(__OpenBSD__)
1789 (0) /* For open bsd we do dual bind only */
1791 (((struct in6pcb *)inp)->in6p_flags & IN6P_IPV6_V6ONLY)
1798 /* else fall through to conflict */
1805 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
1807 * Don't know why, but without this there is an unknown reference when
1808 * compiling NetBSD... hmm
1810 extern void in6_sin6_2_sin (struct sockaddr_in *, struct sockaddr_in6 *sin6);
1815 #if (defined(__FreeBSD__) && __FreeBSD_version >= 500000) || defined(__DragonFly__)
1816 sctp_inpcb_bind(struct socket *so, struct sockaddr *addr, struct thread *p)
1818 sctp_inpcb_bind(struct socket *so, struct sockaddr *addr, struct proc *p)
1821 /* bind a ep to a socket address */
1822 struct sctppcbhead *head;
1823 struct sctp_inpcb *inp, *inp_tmp;
1824 struct inpcb *ip_inp;
1832 inp = (struct sctp_inpcb *)so->so_pcb;
1833 ip_inp = (struct inpcb *)so->so_pcb;
1835 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
1837 kprintf("Bind called port:%d\n",
1838 ntohs(((struct sockaddr_in *)addr)->sin_port));
1840 sctp_print_address(addr);
1843 #endif /* SCTP_DEBUG */
1844 if ((inp->sctp_flags & SCTP_PCB_FLAGS_UNBOUND) == 0) {
1845 /* already did a bind, subsequent binds NOT allowed ! */
1850 if (addr->sa_family == AF_INET) {
1851 struct sockaddr_in *sin;
1853 /* IPV6_V6ONLY socket? */
1855 #if defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__)
1856 (ip_inp->inp_flags & IN6P_IPV6_V6ONLY)
1858 #if defined(__OpenBSD__)
1859 (0) /* For openbsd we do dual bind only */
1861 (((struct in6pcb *)inp)->in6p_flags & IN6P_IPV6_V6ONLY)
1868 if (addr->sa_len != sizeof(*sin))
1871 sin = (struct sockaddr_in *)addr;
1872 lport = sin->sin_port;
1874 if (sin->sin_addr.s_addr != INADDR_ANY) {
1877 } else if (addr->sa_family == AF_INET6) {
1878 /* Only for pure IPv6 Address. (No IPv4 Mapped!) */
1879 struct sockaddr_in6 *sin6;
1881 sin6 = (struct sockaddr_in6 *)addr;
1883 if (addr->sa_len != sizeof(*sin6))
1886 lport = sin6->sin6_port;
1887 if (!IN6_IS_ADDR_UNSPECIFIED(&sin6->sin6_addr)) {
1889 /* KAME hack: embed scopeid */
1890 #if defined(SCTP_BASE_FREEBSD) || defined(__APPLE__) || defined(__DragonFly__)
1891 if (in6_embedscope(&sin6->sin6_addr, sin6,
1894 #elif defined(__FreeBSD__)
1895 error = scope6_check_id(sin6, ip6_use_defzone);
1899 if (in6_embedscope(&sin6->sin6_addr, sin6) != 0) {
1904 #ifndef SCOPEDROUTING
1905 /* this must be cleared for ifa_ifwithaddr() */
1906 sin6->sin6_scope_id = 0;
1907 #endif /* SCOPEDROUTING */
1909 return (EAFNOSUPPORT);
1912 SCTP_INP_INFO_WLOCK();
1913 SCTP_INP_WLOCK(inp);
1914 /* increase our count due to the unlock we do */
1915 SCTP_INP_INCR_REF(inp);
1918 * Did the caller specify a port? if so we must see if a
1919 * ep already has this one bound.
1921 /* got to be root to get at low ports */
1922 if (ntohs(lport) < IPPORT_RESERVED) {
1925 #if __FreeBSD_version >= 500000
1926 suser_cred(p->td_ucred, 0)
1930 #elif defined(__NetBSD__) || defined(__APPLE__)
1931 suser(p->p_ucred, &p->p_acflag)
1932 #elif defined(__DragonFly__)
1933 priv_check(p, PRIV_ROOT)
1938 SCTP_INP_DECR_REF(inp);
1939 SCTP_INP_WUNLOCK(inp);
1940 SCTP_INP_INFO_WUNLOCK();
1945 SCTP_INP_DECR_REF(inp);
1946 SCTP_INP_WUNLOCK(inp);
1947 SCTP_INP_INFO_WUNLOCK();
1950 SCTP_INP_WUNLOCK(inp);
1951 inp_tmp = sctp_pcb_findep(addr, 0, 1);
1952 if (inp_tmp != NULL) {
1953 /* lock guy returned and lower count
1954 * note that we are not bound so inp_tmp
1955 * should NEVER be inp. And it is this
1956 * inp (inp_tmp) that gets the reference
1957 * bump, so we must lower it.
1959 SCTP_INP_WLOCK(inp_tmp);
1960 SCTP_INP_DECR_REF(inp_tmp);
1961 SCTP_INP_WUNLOCK(inp_tmp);
1964 SCTP_INP_INFO_WUNLOCK();
1965 return (EADDRNOTAVAIL);
1967 SCTP_INP_WLOCK(inp);
1969 /* verify that no lport is not used by a singleton */
1970 if (sctp_isport_inuse(inp, lport)) {
1971 /* Sorry someone already has this one bound */
1972 SCTP_INP_DECR_REF(inp);
1973 SCTP_INP_WUNLOCK(inp);
1974 SCTP_INP_INFO_WUNLOCK();
1975 return (EADDRNOTAVAIL);
1980 * get any port but lets make sure no one has any address
1981 * with this port bound
1985 * setup the inp to the top (I could use the union but this
1988 uint32_t port_guess;
1989 uint16_t port_attempt;
1993 port_guess = sctp_select_initial_TSN(&inp->sctp_ep);
1994 port_attempt = (port_guess & 0x0000ffff);
1995 if (port_attempt == 0) {
1998 if (port_attempt < IPPORT_RESERVED) {
1999 port_attempt += IPPORT_RESERVED;
2002 if (sctp_isport_inuse(inp, htons(port_attempt)) == 0) {
2003 /* got a port we can use */
2007 /* try upper half */
2009 port_attempt = ((port_guess >> 16) & 0x0000ffff);
2010 if (port_attempt == 0) {
2013 if (port_attempt < IPPORT_RESERVED) {
2014 port_attempt += IPPORT_RESERVED;
2016 if (sctp_isport_inuse(inp, htons(port_attempt)) == 0) {
2017 /* got a port we can use */
2021 /* try two half's added together */
2023 port_attempt = (((port_guess >> 16) & 0x0000ffff) + (port_guess & 0x0000ffff));
2024 if (port_attempt == 0) {
2025 /* get a new random number */
2028 if (port_attempt < IPPORT_RESERVED) {
2029 port_attempt += IPPORT_RESERVED;
2031 if (sctp_isport_inuse(inp, htons(port_attempt)) == 0) {
2032 /* got a port we can use */
2037 /* we don't get out of the loop until we have a port */
2038 lport = htons(port_attempt);
2040 SCTP_INP_DECR_REF(inp);
2041 if (inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
2042 /* this really should not happen. The guy
2043 * did a non-blocking bind and then did a close
2046 SCTP_INP_WUNLOCK(inp);
2047 SCTP_INP_INFO_WUNLOCK();
2050 /* ok we look clear to give out this port, so lets setup the binding */
2052 /* binding to all addresses, so just set in the proper flags */
2053 inp->sctp_flags |= (SCTP_PCB_FLAGS_BOUNDALL |
2054 SCTP_PCB_FLAGS_DO_ASCONF);
2055 /* set the automatic addr changes from kernel flag */
2056 if (sctp_auto_asconf == 0) {
2057 inp->sctp_flags &= ~SCTP_PCB_FLAGS_AUTO_ASCONF;
2059 inp->sctp_flags |= SCTP_PCB_FLAGS_AUTO_ASCONF;
2063 * bind specific, make sure flags is off and add a new address
2064 * structure to the sctp_addr_list inside the ep structure.
2066 * We will need to allocate one and insert it at the head.
2067 * The socketopt call can just insert new addresses in there
2068 * as well. It will also have to do the embed scope kame hack
2069 * too (before adding).
2072 struct sockaddr_storage store_sa;
2074 memset(&store_sa, 0, sizeof(store_sa));
2075 if (addr->sa_family == AF_INET) {
2076 struct sockaddr_in *sin;
2078 sin = (struct sockaddr_in *)&store_sa;
2079 memcpy(sin, addr, sizeof(struct sockaddr_in));
2081 } else if (addr->sa_family == AF_INET6) {
2082 struct sockaddr_in6 *sin6;
2084 sin6 = (struct sockaddr_in6 *)&store_sa;
2085 memcpy(sin6, addr, sizeof(struct sockaddr_in6));
2086 sin6->sin6_port = 0;
2089 * first find the interface with the bound address
2090 * need to zero out the port to find the address! yuck!
2091 * can't do this earlier since need port for sctp_pcb_findep()
2093 ifa = sctp_find_ifa_by_addr((struct sockaddr *)&store_sa);
2095 /* Can't find an interface with that address */
2096 SCTP_INP_WUNLOCK(inp);
2097 SCTP_INP_INFO_WUNLOCK();
2098 return (EADDRNOTAVAIL);
2100 if (addr->sa_family == AF_INET6) {
2101 struct in6_ifaddr *ifa6;
2102 ifa6 = (struct in6_ifaddr *)ifa;
2104 * allow binding of deprecated addresses as per
2105 * RFC 2462 and ipng discussion
2107 if (ifa6->ia6_flags & (IN6_IFF_DETACHED |
2109 IN6_IFF_NOTREADY)) {
2110 /* Can't bind a non-existent addr. */
2111 SCTP_INP_WUNLOCK(inp);
2112 SCTP_INP_INFO_WUNLOCK();
2116 /* we're not bound all */
2117 inp->sctp_flags &= ~SCTP_PCB_FLAGS_BOUNDALL;
2118 #if 0 /* use sysctl now */
2119 /* don't allow automatic addr changes from kernel */
2120 inp->sctp_flags &= ~SCTP_PCB_FLAGS_AUTO_ASCONF;
2122 /* set the automatic addr changes from kernel flag */
2123 if (sctp_auto_asconf == 0) {
2124 inp->sctp_flags &= ~SCTP_PCB_FLAGS_AUTO_ASCONF;
2126 inp->sctp_flags |= SCTP_PCB_FLAGS_AUTO_ASCONF;
2128 /* allow bindx() to send ASCONF's for binding changes */
2129 inp->sctp_flags |= SCTP_PCB_FLAGS_DO_ASCONF;
2130 /* add this address to the endpoint list */
2131 error = sctp_insert_laddr(&inp->sctp_addr_list, ifa);
2133 SCTP_INP_WUNLOCK(inp);
2134 SCTP_INP_INFO_WUNLOCK();
2139 /* find the bucket */
2140 head = &sctppcbinfo.sctp_ephash[SCTP_PCBHASH_ALLADDR(lport,
2141 sctppcbinfo.hashmark)];
2142 /* put it in the bucket */
2143 LIST_INSERT_HEAD(head, inp, sctp_hash);
2145 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
2146 kprintf("Main hash to bind at head:%p, bound port:%d\n", head, ntohs(lport));
2149 /* set in the port */
2150 inp->sctp_lport = lport;
2152 /* turn off just the unbound flag */
2153 inp->sctp_flags &= ~SCTP_PCB_FLAGS_UNBOUND;
2154 SCTP_INP_WUNLOCK(inp);
2155 SCTP_INP_INFO_WUNLOCK();
2161 sctp_iterator_inp_being_freed(struct sctp_inpcb *inp, struct sctp_inpcb *inp_next)
2163 struct sctp_iterator *it;
2164 /* We enter with the only the ITERATOR_LOCK in place and
2165 * A write lock on the inp_info stuff.
2168 /* Go through all iterators, we must do this since
2169 * it is possible that some iterator does NOT have
2170 * the lock, but is waiting for it. And the one that
2171 * had the lock has either moved in the last iteration
2172 * or we just cleared it above. We need to find all
2173 * of those guys. The list of iterators should never
2174 * be very big though.
2176 LIST_FOREACH(it, &sctppcbinfo.iteratorhead, sctp_nxt_itr) {
2177 if (it == inp->inp_starting_point_for_iterator)
2178 /* skip this guy, he's special */
2180 if (it->inp == inp) {
2181 /* This is tricky and we DON'T lock the iterator.
2182 * Reason is he's running but waiting for me since
2183 * inp->inp_starting_point_for_iterator has the lock
2184 * on me (the guy above we skipped). This tells us
2185 * its is not running but waiting for inp->inp_starting_point_for_iterator
2186 * to be released by the guy that does have our INP in a lock.
2188 if (it->iterator_flags & SCTP_ITERATOR_DO_SINGLE_INP) {
2192 /* set him up to do the next guy not me */
2198 it = inp->inp_starting_point_for_iterator;
2200 if (it->iterator_flags & SCTP_ITERATOR_DO_SINGLE_INP) {
2209 /* release sctp_inpcb unbind the port */
2211 sctp_inpcb_free(struct sctp_inpcb *inp, int immediate)
2214 * Here we free a endpoint. We must find it (if it is in the Hash
2215 * table) and remove it from there. Then we must also find it in
2216 * the overall list and remove it from there. After all removals are
2217 * complete then any timer has to be stopped. Then start the actual
2219 * a) Any local lists.
2220 * b) Any associations.
2221 * c) The hash of all associations.
2222 * d) finally the ep itself.
2224 struct sctp_inpcb *inp_save;
2225 struct sctp_tcb *asoc, *nasoc;
2226 struct sctp_laddr *laddr, *nladdr;
2227 struct inpcb *ip_pcb;
2229 struct sctp_socket_q_list *sq;
2230 #if !defined(__FreeBSD__) || __FreeBSD_version < 500000
2235 SCTP_ASOC_CREATE_LOCK(inp);
2236 SCTP_INP_WLOCK(inp);
2238 if (inp->sctp_flags & SCTP_PCB_FLAGS_SOCKET_ALLGONE) {
2239 /* been here before */
2240 kprintf("Endpoint was all gone (dup free)?\n");
2241 SCTP_INP_WUNLOCK(inp);
2242 SCTP_ASOC_CREATE_UNLOCK(inp);
2245 sctp_timer_stop(SCTP_TIMER_TYPE_NEWCOOKIE, inp, NULL, NULL);
2248 sctp_m_freem(inp->control);
2249 inp->control = NULL;
2252 sctp_m_freem(inp->pkt);
2255 so = inp->sctp_socket;
2256 ip_pcb = &inp->ip_inp.inp; /* we could just cast the main
2257 * pointer here but I will
2258 * be nice :> (i.e. ip_pcb = ep;)
2261 if (immediate == 0) {
2264 for ((asoc = LIST_FIRST(&inp->sctp_asoc_list)); asoc != NULL;
2266 nasoc = LIST_NEXT(asoc, sctp_tcblist);
2267 if ((SCTP_GET_STATE(&asoc->asoc) == SCTP_STATE_COOKIE_WAIT) ||
2268 (SCTP_GET_STATE(&asoc->asoc) == SCTP_STATE_COOKIE_ECHOED)) {
2269 /* Just abandon things in the front states */
2270 SCTP_TCB_LOCK(asoc);
2271 SCTP_INP_WUNLOCK(inp);
2272 sctp_free_assoc(inp, asoc);
2273 SCTP_INP_WLOCK(inp);
2276 asoc->asoc.state |= SCTP_STATE_CLOSED_SOCKET;
2278 if ((asoc->asoc.size_on_delivery_queue > 0) ||
2279 (asoc->asoc.size_on_reasm_queue > 0) ||
2280 (asoc->asoc.size_on_all_streams > 0) ||
2281 (so && (so->so_rcv.ssb_cc > 0))
2283 /* Left with Data unread */
2284 struct mbuf *op_err;
2285 MGET(op_err, MB_DONTWAIT, MT_DATA);
2287 /* Fill in the user initiated abort */
2288 struct sctp_paramhdr *ph;
2290 sizeof(struct sctp_paramhdr);
2292 struct sctp_paramhdr *);
2293 ph->param_type = htons(
2294 SCTP_CAUSE_USER_INITIATED_ABT);
2295 ph->param_length = htons(op_err->m_len);
2297 SCTP_TCB_LOCK(asoc);
2298 sctp_send_abort_tcb(asoc, op_err);
2300 SCTP_INP_WUNLOCK(inp);
2301 sctp_free_assoc(inp, asoc);
2302 SCTP_INP_WLOCK(inp);
2304 } else if (TAILQ_EMPTY(&asoc->asoc.send_queue) &&
2305 TAILQ_EMPTY(&asoc->asoc.sent_queue)) {
2306 if ((SCTP_GET_STATE(&asoc->asoc) != SCTP_STATE_SHUTDOWN_SENT) &&
2307 (SCTP_GET_STATE(&asoc->asoc) != SCTP_STATE_SHUTDOWN_ACK_SENT)) {
2308 /* there is nothing queued to send, so I send shutdown */
2309 SCTP_TCB_LOCK(asoc);
2310 sctp_send_shutdown(asoc, asoc->asoc.primary_destination);
2311 asoc->asoc.state = SCTP_STATE_SHUTDOWN_SENT;
2312 sctp_timer_start(SCTP_TIMER_TYPE_SHUTDOWN, asoc->sctp_ep, asoc,
2313 asoc->asoc.primary_destination);
2314 sctp_timer_start(SCTP_TIMER_TYPE_SHUTDOWNGUARD, asoc->sctp_ep, asoc,
2315 asoc->asoc.primary_destination);
2316 sctp_chunk_output(inp, asoc, 1);
2317 SCTP_TCB_UNLOCK(asoc);
2320 /* mark into shutdown pending */
2321 asoc->asoc.state |= SCTP_STATE_SHUTDOWN_PENDING;
2325 /* now is there some left in our SHUTDOWN state? */
2327 inp->sctp_flags |= SCTP_PCB_FLAGS_SOCKET_GONE;
2328 SCTP_INP_WUNLOCK(inp);
2329 SCTP_ASOC_CREATE_UNLOCK(inp);
2333 #if defined(__FreeBSD__) && __FreeBSD_version >= 503000
2334 if (inp->refcount) {
2335 sctp_timer_start(SCTP_TIMER_TYPE_INPKILL, inp, NULL, NULL);
2336 SCTP_INP_WUNLOCK(inp);
2337 SCTP_ASOC_CREATE_UNLOCK(inp);
2341 inp->sctp_flags |= SCTP_PCB_FLAGS_SOCKET_ALLGONE;
2342 #if !defined(__FreeBSD__) || __FreeBSD_version < 500000
2343 rt = ip_pcb->inp_route.ro_rt;
2346 callout_stop(&inp->sctp_ep.signature_change.timer);
2349 /* First take care of socket level things */
2352 /* XXX IPsec cleanup here */
2353 if (ip_pcb->inp_tdb_in)
2354 TAILQ_REMOVE(&ip_pcb->inp_tdb_in->tdb_inp_in,
2355 ip_pcb, inp_tdb_in_next);
2356 if (ip_pcb->inp_tdb_out)
2357 TAILQ_REMOVE(&ip_pcb->inp_tdb_out->tdb_inp_out, ip_pcb,
2359 if (ip_pcb->inp_ipsec_localid)
2360 ipsp_reffree(ip_pcb->inp_ipsec_localid);
2361 if (ip_pcb->inp_ipsec_remoteid)
2362 ipsp_reffree(ip_pcb->inp_ipsec_remoteid);
2363 if (ip_pcb->inp_ipsec_localcred)
2364 ipsp_reffree(ip_pcb->inp_ipsec_localcred);
2365 if (ip_pcb->inp_ipsec_remotecred)
2366 ipsp_reffree(ip_pcb->inp_ipsec_remotecred);
2367 if (ip_pcb->inp_ipsec_localauth)
2368 ipsp_reffree(ip_pcb->inp_ipsec_localauth);
2369 if (ip_pcb->inp_ipsec_remoteauth)
2370 ipsp_reffree(ip_pcb->inp_ipsec_remoteauth);
2372 ipsec4_delete_pcbpolicy(ip_pcb);
2375 #if defined(__FreeBSD__) && __FreeBSD_version > 500000
2383 if (ip_pcb->inp_options) {
2384 m_free(ip_pcb->inp_options);
2385 ip_pcb->inp_options = 0;
2387 #if !defined(__FreeBSD__) || __FreeBSD_version < 500000
2390 ip_pcb->inp_route.ro_rt = 0;
2393 if (ip_pcb->inp_moptions) {
2394 ip_freemoptions(ip_pcb->inp_moptions);
2395 ip_pcb->inp_moptions = 0;
2397 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
2400 ip_pcb->inp_vflag = 0;
2403 /* Now the sctp_pcb things */
2406 * free each asoc if it is not already closed/free. we can't use
2407 * the macro here since le_next will get freed as part of the
2408 * sctp_free_assoc() call.
2411 for ((asoc = LIST_FIRST(&inp->sctp_asoc_list)); asoc != NULL;
2413 nasoc = LIST_NEXT(asoc, sctp_tcblist);
2414 SCTP_TCB_LOCK(asoc);
2415 if (SCTP_GET_STATE(&asoc->asoc) != SCTP_STATE_COOKIE_WAIT) {
2416 struct mbuf *op_err;
2417 MGET(op_err, MB_DONTWAIT, MT_DATA);
2419 /* Fill in the user initiated abort */
2420 struct sctp_paramhdr *ph;
2421 op_err->m_len = sizeof(struct sctp_paramhdr);
2422 ph = mtod(op_err, struct sctp_paramhdr *);
2423 ph->param_type = htons(
2424 SCTP_CAUSE_USER_INITIATED_ABT);
2425 ph->param_length = htons(op_err->m_len);
2427 sctp_send_abort_tcb(asoc, op_err);
2431 * sctp_free_assoc() will call sctp_inpcb_free(),
2432 * if SCTP_PCB_FLAGS_SOCKET_GONE set.
2433 * So, we clear it before sctp_free_assoc() making sure
2434 * no double sctp_inpcb_free().
2436 inp->sctp_flags &= ~SCTP_PCB_FLAGS_SOCKET_GONE;
2437 SCTP_INP_WUNLOCK(inp);
2438 sctp_free_assoc(inp, asoc);
2439 SCTP_INP_WLOCK(inp);
2441 while ((sq = TAILQ_FIRST(&inp->sctp_queue_list)) != NULL) {
2442 TAILQ_REMOVE(&inp->sctp_queue_list, sq, next_sq);
2443 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_sockq, sq);
2444 sctppcbinfo.ipi_count_sockq--;
2445 sctppcbinfo.ipi_gencnt_sockq++;
2447 inp->sctp_socket = 0;
2448 /* Now first we remove ourselves from the overall list of all EP's */
2450 /* Unlock inp first, need correct order */
2451 SCTP_INP_WUNLOCK(inp);
2452 /* now iterator lock */
2453 SCTP_ITERATOR_LOCK();
2455 SCTP_INP_INFO_WLOCK();
2456 /* now reget the inp lock */
2457 SCTP_INP_WLOCK(inp);
2459 inp_save = LIST_NEXT(inp, sctp_list);
2460 LIST_REMOVE(inp, sctp_list);
2462 * Now the question comes as to if this EP was ever bound at all.
2463 * If it was, then we must pull it out of the EP hash list.
2465 if ((inp->sctp_flags & SCTP_PCB_FLAGS_UNBOUND) !=
2466 SCTP_PCB_FLAGS_UNBOUND) {
2468 * ok, this guy has been bound. It's port is somewhere
2469 * in the sctppcbinfo hash table. Remove it!
2471 LIST_REMOVE(inp, sctp_hash);
2473 /* fix any iterators only after out of the list */
2474 sctp_iterator_inp_being_freed(inp, inp_save);
2475 SCTP_ITERATOR_UNLOCK();
2477 * if we have an address list the following will free the list of
2478 * ifaddr's that are set into this ep. Again macro limitations here,
2479 * since the LIST_FOREACH could be a bad idea.
2481 for ((laddr = LIST_FIRST(&inp->sctp_addr_list)); laddr != NULL;
2483 nladdr = LIST_NEXT(laddr, sctp_nxt_addr);
2484 LIST_REMOVE(laddr, sctp_nxt_addr);
2485 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_laddr, laddr);
2486 sctppcbinfo.ipi_gencnt_laddr++;
2487 sctppcbinfo.ipi_count_laddr--;
2489 /* Now lets see about freeing the EP hash table. */
2490 if (inp->sctp_tcbhash != NULL) {
2491 hashdestroy(inp->sctp_tcbhash, M_PCB, inp->sctp_hashmark);
2492 inp->sctp_tcbhash = 0;
2494 SCTP_INP_WUNLOCK(inp);
2495 SCTP_ASOC_CREATE_UNLOCK(inp);
2496 SCTP_INP_LOCK_DESTROY(inp);
2497 SCTP_ASOC_CREATE_LOCK_DESTROY(inp);
2499 /* Now we must put the ep memory back into the zone pool */
2500 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_ep, inp);
2501 sctppcbinfo.ipi_count_ep--;
2503 SCTP_INP_INFO_WUNLOCK();
2508 sctp_findnet(struct sctp_tcb *stcb, struct sockaddr *addr)
2510 struct sctp_nets *net;
2512 struct sockaddr_in *sin;
2514 /* why do we need to check the port for a nets list on an assoc? */
2515 /* use the peer's/remote port for lookup if unspecified */
2516 sin = (struct sockaddr_in *)addr;
2517 if (stcb->rport != sin->sin_port) {
2518 /* we cheat and just a sin for this test */
2522 /* locate the address */
2523 TAILQ_FOREACH(net, &stcb->asoc.nets, sctp_next) {
2524 if (sctp_cmpaddr(addr, (struct sockaddr *)&net->ro._l_addr))
2532 * add's a remote endpoint address, done with the INIT/INIT-ACK
2533 * as well as when a ASCONF arrives that adds it. It will also
2534 * initialize all the cwnd stats of stuff.
2537 sctp_is_address_on_local_host(struct sockaddr *addr)
2541 TAILQ_FOREACH(ifn, &ifnet, if_list) {
2542 struct ifaddr_container *ifac;
2544 TAILQ_FOREACH(ifac, &ifn->if_addrheads[mycpuid], ifa_link) {
2545 struct ifaddr *ifa = ifac->ifa;
2547 if (addr->sa_family == ifa->ifa_addr->sa_family) {
2549 if (addr->sa_family == AF_INET) {
2550 struct sockaddr_in *sin, *sin_c;
2551 sin = (struct sockaddr_in *)addr;
2552 sin_c = (struct sockaddr_in *)
2554 if (sin->sin_addr.s_addr ==
2555 sin_c->sin_addr.s_addr) {
2556 /* we are on the same machine */
2559 } else if (addr->sa_family == AF_INET6) {
2560 struct sockaddr_in6 *sin6, *sin_c6;
2561 sin6 = (struct sockaddr_in6 *)addr;
2562 sin_c6 = (struct sockaddr_in6 *)
2564 if (SCTP6_ARE_ADDR_EQUAL(&sin6->sin6_addr,
2565 &sin_c6->sin6_addr)) {
2566 /* we are on the same machine */
2577 sctp_add_remote_addr(struct sctp_tcb *stcb, struct sockaddr *newaddr,
2578 int set_scope, int from)
2581 * The following is redundant to the same lines in the
2582 * sctp_aloc_assoc() but is needed since other's call the add
2585 struct sctp_nets *net, *netfirst;
2589 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
2590 kprintf("Adding an address (from:%d) to the peer: ", from);
2591 sctp_print_address(newaddr);
2594 netfirst = sctp_findnet(stcb, newaddr);
2597 * Lie and return ok, we don't want to make the association
2598 * go away for this behavior. It will happen in the TCP model
2599 * in a connected socket. It does not reach the hash table
2600 * until after the association is built so it can't be found.
2601 * Mark as reachable, since the initial creation will have
2602 * been cleared and the NOT_IN_ASSOC flag will have been
2603 * added... and we don't want to end up removing it back out.
2605 if (netfirst->dest_state & SCTP_ADDR_UNCONFIRMED) {
2606 netfirst->dest_state = (SCTP_ADDR_REACHABLE|
2607 SCTP_ADDR_UNCONFIRMED);
2609 netfirst->dest_state = SCTP_ADDR_REACHABLE;
2615 if (newaddr->sa_family == AF_INET) {
2616 struct sockaddr_in *sin;
2617 sin = (struct sockaddr_in *)newaddr;
2618 if (sin->sin_addr.s_addr == 0) {
2619 /* Invalid address */
2622 /* zero out the bzero area */
2623 memset(&sin->sin_zero, 0, sizeof(sin->sin_zero));
2625 /* assure len is set */
2626 sin->sin_len = sizeof(struct sockaddr_in);
2628 #ifdef SCTP_DONT_DO_PRIVADDR_SCOPE
2629 stcb->ipv4_local_scope = 1;
2631 if (IN4_ISPRIVATE_ADDRESS(&sin->sin_addr)) {
2632 stcb->asoc.ipv4_local_scope = 1;
2634 #endif /* SCTP_DONT_DO_PRIVADDR_SCOPE */
2636 if (sctp_is_address_on_local_host(newaddr)) {
2637 stcb->asoc.loopback_scope = 1;
2638 stcb->asoc.ipv4_local_scope = 1;
2639 stcb->asoc.local_scope = 1;
2640 stcb->asoc.site_scope = 1;
2645 if (sctp_is_address_on_local_host(newaddr)) {
2646 stcb->asoc.loopback_scope = 1;
2647 stcb->asoc.ipv4_local_scope = 1;
2648 stcb->asoc.local_scope = 1;
2649 stcb->asoc.site_scope = 1;
2652 /* Validate the address is in scope */
2653 if ((IN4_ISPRIVATE_ADDRESS(&sin->sin_addr)) &&
2654 (stcb->asoc.ipv4_local_scope == 0)) {
2658 } else if (newaddr->sa_family == AF_INET6) {
2659 struct sockaddr_in6 *sin6;
2660 sin6 = (struct sockaddr_in6 *)newaddr;
2661 if (IN6_IS_ADDR_UNSPECIFIED(&sin6->sin6_addr)) {
2662 /* Invalid address */
2665 /* assure len is set */
2666 sin6->sin6_len = sizeof(struct sockaddr_in6);
2668 if (sctp_is_address_on_local_host(newaddr)) {
2669 stcb->asoc.loopback_scope = 1;
2670 stcb->asoc.local_scope = 1;
2671 stcb->asoc.ipv4_local_scope = 1;
2672 stcb->asoc.site_scope = 1;
2673 } else if (IN6_IS_ADDR_LINKLOCAL(&sin6->sin6_addr)) {
2675 * If the new destination is a LINK_LOCAL
2676 * we must have common site scope. Don't set
2677 * the local scope since we may not share all
2678 * links, only loopback can do this.
2679 * Links on the local network would also
2680 * be on our private network for v4 too.
2682 stcb->asoc.ipv4_local_scope = 1;
2683 stcb->asoc.site_scope = 1;
2684 } else if (IN6_IS_ADDR_SITELOCAL(&sin6->sin6_addr)) {
2686 * If the new destination is SITE_LOCAL
2687 * then we must have site scope in common.
2689 stcb->asoc.site_scope = 1;
2694 if (sctp_is_address_on_local_host(newaddr)) {
2695 stcb->asoc.loopback_scope = 1;
2696 stcb->asoc.ipv4_local_scope = 1;
2697 stcb->asoc.local_scope = 1;
2698 stcb->asoc.site_scope = 1;
2701 /* Validate the address is in scope */
2702 if (IN6_IS_ADDR_LOOPBACK(&sin6->sin6_addr) &&
2703 (stcb->asoc.loopback_scope == 0)) {
2705 } else if (IN6_IS_ADDR_LINKLOCAL(&sin6->sin6_addr) &&
2706 (stcb->asoc.local_scope == 0)) {
2708 } else if (IN6_IS_ADDR_SITELOCAL(&sin6->sin6_addr) &&
2709 (stcb->asoc.site_scope == 0)) {
2714 /* not supported family type */
2717 net = (struct sctp_nets *)SCTP_ZONE_GET(sctppcbinfo.ipi_zone_net);
2721 sctppcbinfo.ipi_count_raddr++;
2722 sctppcbinfo.ipi_gencnt_raddr++;
2723 bzero(net, sizeof(*net));
2724 memcpy(&net->ro._l_addr, newaddr, newaddr->sa_len);
2725 if (newaddr->sa_family == AF_INET) {
2726 ((struct sockaddr_in *)&net->ro._l_addr)->sin_port = stcb->rport;
2727 } else if (newaddr->sa_family == AF_INET6) {
2728 ((struct sockaddr_in6 *)&net->ro._l_addr)->sin6_port = stcb->rport;
2730 net->addr_is_local = sctp_is_address_on_local_host(newaddr);
2731 net->failure_threshold = stcb->asoc.def_net_failure;
2732 if (addr_inscope == 0) {
2734 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
2735 kprintf("Adding an address which is OUT OF SCOPE\n");
2737 #endif /* SCTP_DEBUG */
2738 net->dest_state = (SCTP_ADDR_REACHABLE |
2739 SCTP_ADDR_OUT_OF_SCOPE);
2742 /* 8 is passed by connect_x */
2743 net->dest_state = SCTP_ADDR_REACHABLE;
2745 net->dest_state = SCTP_ADDR_REACHABLE |
2746 SCTP_ADDR_UNCONFIRMED;
2748 net->RTO = stcb->asoc.initial_rto;
2749 stcb->asoc.numnets++;
2752 /* Init the timer structure */
2753 #if defined(__FreeBSD__) && __FreeBSD_version >= 500000
2754 callout_init(&net->rxt_timer.timer, 0);
2755 callout_init(&net->pmtu_timer.timer, 0);
2757 callout_init(&net->rxt_timer.timer);
2758 callout_init(&net->pmtu_timer.timer);
2761 /* Now generate a route for this guy */
2762 /* KAME hack: embed scopeid */
2763 if (newaddr->sa_family == AF_INET6) {
2764 struct sockaddr_in6 *sin6;
2765 sin6 = (struct sockaddr_in6 *)&net->ro._l_addr;
2766 #if defined(SCTP_BASE_FREEBSD) || defined(__APPLE__) || defined(__DragonFly__)
2767 in6_embedscope(&sin6->sin6_addr, sin6,
2768 &stcb->sctp_ep->ip_inp.inp, NULL);
2770 in6_embedscope(&sin6->sin6_addr, sin6);
2772 #ifndef SCOPEDROUTING
2773 sin6->sin6_scope_id = 0;
2776 #if defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__)
2777 rtalloc_ign((struct route *)&net->ro, 0UL);
2779 rtalloc((struct route *)&net->ro);
2781 if (newaddr->sa_family == AF_INET6) {
2782 struct sockaddr_in6 *sin6;
2783 sin6 = (struct sockaddr_in6 *)&net->ro._l_addr;
2784 in6_recoverscope(sin6, &sin6->sin6_addr, NULL);
2786 if ((net->ro.ro_rt) &&
2787 (net->ro.ro_rt->rt_ifp)) {
2788 net->mtu = net->ro.ro_rt->rt_ifp->if_mtu;
2790 stcb->asoc.smallest_mtu = net->mtu;
2792 /* start things off to match mtu of interface please. */
2793 net->ro.ro_rt->rt_rmx.rmx_mtu = net->ro.ro_rt->rt_ifp->if_mtu;
2795 net->mtu = stcb->asoc.smallest_mtu;
2797 if (stcb->asoc.smallest_mtu > net->mtu) {
2798 stcb->asoc.smallest_mtu = net->mtu;
2800 /* We take the max of the burst limit times a MTU or the INITIAL_CWND.
2801 * We then limit this to 4 MTU's of sending.
2803 net->cwnd = min((net->mtu * 4), max((stcb->asoc.max_burst * net->mtu), SCTP_INITIAL_CWND));
2805 /* we always get at LEAST 2 MTU's */
2806 if (net->cwnd < (2 * net->mtu)) {
2807 net->cwnd = 2 * net->mtu;
2810 net->ssthresh = stcb->asoc.peers_rwnd;
2812 net->src_addr_selected = 0;
2813 netfirst = TAILQ_FIRST(&stcb->asoc.nets);
2814 if (net->ro.ro_rt == NULL) {
2815 /* Since we have no route put it at the back */
2816 TAILQ_INSERT_TAIL(&stcb->asoc.nets, net, sctp_next);
2817 } else if (netfirst == NULL) {
2818 /* We are the first one in the pool. */
2819 TAILQ_INSERT_HEAD(&stcb->asoc.nets, net, sctp_next);
2820 } else if (netfirst->ro.ro_rt == NULL) {
2822 * First one has NO route. Place this one ahead of the
2825 TAILQ_INSERT_HEAD(&stcb->asoc.nets, net, sctp_next);
2826 } else if (net->ro.ro_rt->rt_ifp != netfirst->ro.ro_rt->rt_ifp) {
2828 * This one has a different interface than the one at the
2829 * top of the list. Place it ahead.
2831 TAILQ_INSERT_HEAD(&stcb->asoc.nets, net, sctp_next);
2834 * Ok we have the same interface as the first one. Move
2835 * forward until we find either
2836 * a) one with a NULL route... insert ahead of that
2837 * b) one with a different ifp.. insert after that.
2838 * c) end of the list.. insert at the tail.
2840 struct sctp_nets *netlook;
2842 netlook = TAILQ_NEXT(netfirst, sctp_next);
2843 if (netlook == NULL) {
2844 /* End of the list */
2845 TAILQ_INSERT_TAIL(&stcb->asoc.nets, net,
2848 } else if (netlook->ro.ro_rt == NULL) {
2849 /* next one has NO route */
2850 TAILQ_INSERT_BEFORE(netfirst, net, sctp_next);
2852 } else if (netlook->ro.ro_rt->rt_ifp !=
2853 net->ro.ro_rt->rt_ifp) {
2854 TAILQ_INSERT_AFTER(&stcb->asoc.nets, netlook,
2860 } while (netlook != NULL);
2862 /* got to have a primary set */
2863 if (stcb->asoc.primary_destination == 0) {
2864 stcb->asoc.primary_destination = net;
2865 } else if ((stcb->asoc.primary_destination->ro.ro_rt == NULL) &&
2867 /* No route to current primary adopt new primary */
2868 stcb->asoc.primary_destination = net;
2870 sctp_timer_start(SCTP_TIMER_TYPE_PATHMTURAISE, stcb->sctp_ep, stcb,
2878 * allocate an association and add it to the endpoint. The caller must
2879 * be careful to add all additional addresses once they are know right
2880 * away or else the assoc will be may experience a blackout scenario.
2883 sctp_aloc_assoc(struct sctp_inpcb *inp, struct sockaddr *firstaddr,
2884 int for_a_init, int *error, uint32_t override_tag)
2886 struct sctp_tcb *stcb;
2887 struct sctp_association *asoc;
2888 struct sctpasochead *head;
2893 * Assumption made here:
2894 * Caller has done a sctp_findassociation_ep_addr(ep, addr's);
2895 * to make sure the address does not exist already.
2897 if (sctppcbinfo.ipi_count_asoc >= SCTP_MAX_NUM_OF_ASOC) {
2898 /* Hit max assoc, sorry no more */
2902 SCTP_INP_RLOCK(inp);
2903 if (inp->sctp_flags & SCTP_PCB_FLAGS_IN_TCPPOOL) {
2905 * If its in the TCP pool, its NOT allowed to create an
2906 * association. The parent listener needs to call
2907 * sctp_aloc_assoc.. or the one-2-many socket. If a
2908 * peeled off, or connected one does this.. its an error.
2910 SCTP_INP_RUNLOCK(inp);
2916 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
2917 kprintf("Allocate an association for peer:");
2919 sctp_print_address(firstaddr);
2922 kprintf("Port:%d\n",
2923 ntohs(((struct sockaddr_in *)firstaddr)->sin_port));
2925 #endif /* SCTP_DEBUG */
2926 if (firstaddr->sa_family == AF_INET) {
2927 struct sockaddr_in *sin;
2928 sin = (struct sockaddr_in *)firstaddr;
2929 if ((sin->sin_port == 0) || (sin->sin_addr.s_addr == 0)) {
2930 /* Invalid address */
2932 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
2933 kprintf("peer address invalid\n");
2936 SCTP_INP_RUNLOCK(inp);
2940 rport = sin->sin_port;
2941 } else if (firstaddr->sa_family == AF_INET6) {
2942 struct sockaddr_in6 *sin6;
2943 sin6 = (struct sockaddr_in6 *)firstaddr;
2944 if ((sin6->sin6_port == 0) ||
2945 (IN6_IS_ADDR_UNSPECIFIED(&sin6->sin6_addr))) {
2946 /* Invalid address */
2948 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
2949 kprintf("peer address invalid\n");
2952 SCTP_INP_RUNLOCK(inp);
2956 rport = sin6->sin6_port;
2958 /* not supported family type */
2960 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
2961 kprintf("BAD family %d\n", firstaddr->sa_family);
2964 SCTP_INP_RUNLOCK(inp);
2968 SCTP_INP_RUNLOCK(inp);
2969 if (inp->sctp_flags & SCTP_PCB_FLAGS_UNBOUND) {
2971 * If you have not performed a bind, then we need to do
2972 * the ephemerial bind for you.
2975 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
2976 kprintf("Doing implicit BIND\n");
2980 if ((err = sctp_inpcb_bind(inp->sctp_socket,
2982 #if (defined(__FreeBSD__) && __FreeBSD_version >= 500000) || defined(__DragonFly__)
2988 /* bind error, probably perm */
2990 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
2991 kprintf("BIND FAILS ret:%d\n", err);
2999 stcb = (struct sctp_tcb *)SCTP_ZONE_GET(sctppcbinfo.ipi_zone_asoc);
3001 /* out of memory? */
3003 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
3004 kprintf("aloc_assoc: no assoc mem left, stcb=NULL\n");
3010 sctppcbinfo.ipi_count_asoc++;
3011 sctppcbinfo.ipi_gencnt_asoc++;
3013 bzero(stcb, sizeof(*stcb));
3015 SCTP_TCB_LOCK_INIT(stcb);
3016 /* setup back pointer's */
3017 stcb->sctp_ep = inp;
3018 stcb->sctp_socket = inp->sctp_socket;
3019 if ((err = sctp_init_asoc(inp, asoc, for_a_init, override_tag))) {
3021 SCTP_TCB_LOCK_DESTROY (stcb);
3022 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_asoc, stcb);
3023 sctppcbinfo.ipi_count_asoc--;
3025 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
3026 kprintf("aloc_assoc: couldn't init asoc, out of mem?!\n");
3033 stcb->rport = rport;
3034 SCTP_INP_INFO_WLOCK();
3035 SCTP_INP_WLOCK(inp);
3036 if (inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
3037 /* inpcb freed while alloc going on */
3038 SCTP_TCB_LOCK_DESTROY (stcb);
3039 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_asoc, stcb);
3040 SCTP_INP_WUNLOCK(inp);
3041 SCTP_INP_INFO_WUNLOCK();
3042 sctppcbinfo.ipi_count_asoc--;
3044 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
3045 kprintf("aloc_assoc: couldn't init asoc, out of mem?!\n");
3051 SCTP_TCB_LOCK(stcb);
3053 /* now that my_vtag is set, add it to the hash */
3054 head = &sctppcbinfo.sctp_asochash[SCTP_PCBHASH_ASOC(stcb->asoc.my_vtag,
3055 sctppcbinfo.hashasocmark)];
3056 /* put it in the bucket in the vtag hash of assoc's for the system */
3057 LIST_INSERT_HEAD(head, stcb, sctp_asocs);
3058 SCTP_INP_INFO_WUNLOCK();
3061 if ((err = sctp_add_remote_addr(stcb, firstaddr, 1, 1))) {
3062 /* failure.. memory error? */
3064 kfree(asoc->strmout, M_PCB);
3065 if (asoc->mapping_array)
3066 kfree(asoc->mapping_array, M_PCB);
3068 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_asoc, stcb);
3069 sctppcbinfo.ipi_count_asoc--;
3071 if (sctp_debug_on & SCTP_DEBUG_PCB3) {
3072 kprintf("aloc_assoc: couldn't add remote addr!\n");
3075 SCTP_TCB_LOCK_DESTROY (stcb);
3079 /* Init all the timers */
3080 #if defined(__FreeBSD__) && __FreeBSD_version >= 500000
3081 callout_init(&asoc->hb_timer.timer, 0);
3082 callout_init(&asoc->dack_timer.timer, 0);
3083 callout_init(&asoc->asconf_timer.timer, 0);
3084 callout_init(&asoc->shut_guard_timer.timer, 0);
3085 callout_init(&asoc->autoclose_timer.timer, 0);
3086 callout_init(&asoc->delayed_event_timer.timer, 0);
3088 callout_init(&asoc->hb_timer.timer);
3089 callout_init(&asoc->dack_timer.timer);
3090 callout_init(&asoc->asconf_timer.timer);
3091 callout_init(&asoc->shut_guard_timer.timer);
3092 callout_init(&asoc->autoclose_timer.timer);
3093 callout_init(&asoc->delayed_event_timer.timer);
3095 LIST_INSERT_HEAD(&inp->sctp_asoc_list, stcb, sctp_tcblist);
3096 /* now file the port under the hash as well */
3097 if (inp->sctp_tcbhash != NULL) {
3098 head = &inp->sctp_tcbhash[SCTP_PCBHASH_ALLADDR(stcb->rport,
3099 inp->sctp_hashmark)];
3100 LIST_INSERT_HEAD(head, stcb, sctp_tcbhash);
3102 SCTP_INP_WUNLOCK(inp);
3104 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
3105 kprintf("Association %p now allocated\n", stcb);
3112 sctp_free_remote_addr(struct sctp_nets *net)
3117 if (net->ref_count <= 0) {
3118 /* stop timer if running */
3119 callout_stop(&net->rxt_timer.timer);
3120 callout_stop(&net->pmtu_timer.timer);
3121 net->dest_state = SCTP_ADDR_NOT_REACHABLE;
3122 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_net, net);
3123 sctppcbinfo.ipi_count_raddr--;
3128 * remove a remote endpoint address from an association, it
3129 * will fail if the address does not exist.
3132 sctp_del_remote_addr(struct sctp_tcb *stcb, struct sockaddr *remaddr)
3135 * Here we need to remove a remote address. This is quite simple, we
3136 * first find it in the list of address for the association
3137 * (tasoc->asoc.nets) and then if it is there, we do a LIST_REMOVE on
3139 * Note we do not allow it to be removed if there are no other
3142 struct sctp_association *asoc;
3143 struct sctp_nets *net, *net_tmp;
3145 if (asoc->numnets < 2) {
3146 /* Must have at LEAST two remote addresses */
3149 /* locate the address */
3150 for (net = TAILQ_FIRST(&asoc->nets); net != NULL; net = net_tmp) {
3151 net_tmp = TAILQ_NEXT(net, sctp_next);
3152 if (net->ro._l_addr.sa.sa_family != remaddr->sa_family) {
3155 if (sctp_cmpaddr((struct sockaddr *)&net->ro._l_addr,
3157 /* we found the guy */
3159 TAILQ_REMOVE(&asoc->nets, net, sctp_next);
3160 sctp_free_remote_addr(net);
3161 if (net == asoc->primary_destination) {
3163 struct sctp_nets *lnet;
3164 lnet = TAILQ_FIRST(&asoc->nets);
3165 /* Try to find a confirmed primary */
3166 asoc->primary_destination =
3167 sctp_find_alternate_net(stcb, lnet);
3169 if (net == asoc->last_data_chunk_from) {
3171 asoc->last_data_chunk_from =
3172 TAILQ_FIRST(&asoc->nets);
3174 if (net == asoc->last_control_chunk_from) {
3176 asoc->last_control_chunk_from =
3177 TAILQ_FIRST(&asoc->nets);
3179 if (net == asoc->asconf_last_sent_to) {
3181 asoc->asconf_last_sent_to =
3182 TAILQ_FIRST(&asoc->nets);
3193 sctp_add_vtag_to_timewait(struct sctp_inpcb *inp, u_int32_t tag)
3195 struct sctpvtaghead *chain;
3196 struct sctp_tagblock *twait_block;
3199 SCTP_GETTIME_TIMEVAL(&now);
3200 chain = &sctppcbinfo.vtag_timewait[(tag % SCTP_STACK_VTAG_HASH_SIZE)];
3202 if (!LIST_EMPTY(chain)) {
3203 /* Block(s) present, lets find space, and expire on the fly */
3204 LIST_FOREACH(twait_block, chain, sctp_nxt_tagblock) {
3205 for (i = 0; i < SCTP_NUMBER_IN_VTAG_BLOCK; i++) {
3206 if ((twait_block->vtag_block[i].v_tag == 0) &&
3208 twait_block->vtag_block[0].tv_sec_at_expire =
3209 now.tv_sec + SCTP_TIME_WAIT;
3210 twait_block->vtag_block[0].v_tag = tag;
3212 } else if ((twait_block->vtag_block[i].v_tag) &&
3213 ((long)twait_block->vtag_block[i].tv_sec_at_expire >
3215 /* Audit expires this guy */
3216 twait_block->vtag_block[i].tv_sec_at_expire = 0;
3217 twait_block->vtag_block[i].v_tag = 0;
3219 /* Reuse it for my new tag */
3220 twait_block->vtag_block[0].tv_sec_at_expire = now.tv_sec + SCTP_TIME_WAIT;
3221 twait_block->vtag_block[0].v_tag = tag;
3228 * We only do up to the block where we can
3229 * place our tag for audits
3235 /* Need to add a new block to chain */
3237 twait_block = kmalloc(sizeof(struct sctp_tagblock), M_PCB,
3239 if (twait_block == NULL) {
3242 memset(twait_block, 0, sizeof(struct sctp_timewait));
3243 LIST_INSERT_HEAD(chain, twait_block, sctp_nxt_tagblock);
3244 twait_block->vtag_block[0].tv_sec_at_expire = now.tv_sec +
3246 twait_block->vtag_block[0].v_tag = tag;
3252 sctp_iterator_asoc_being_freed(struct sctp_inpcb *inp, struct sctp_tcb *stcb)
3254 struct sctp_iterator *it;
3258 /* Unlock the tcb lock we do this so
3259 * we avoid a dead lock scenario where
3260 * the iterator is waiting on the TCB lock
3261 * and the TCB lock is waiting on the iterator
3264 SCTP_ITERATOR_LOCK();
3265 SCTP_INP_INFO_WLOCK();
3266 SCTP_INP_WLOCK(inp);
3267 SCTP_TCB_LOCK(stcb);
3269 it = stcb->asoc.stcb_starting_point_for_iterator;
3273 if (it->inp != stcb->sctp_ep) {
3274 /* hm, focused on the wrong one? */
3277 if (it->stcb != stcb) {
3280 it->stcb = LIST_NEXT(stcb, sctp_tcblist);
3281 if (it->stcb == NULL) {
3282 /* done with all asoc's in this assoc */
3283 if (it->iterator_flags & SCTP_ITERATOR_DO_SINGLE_INP) {
3287 it->inp = LIST_NEXT(inp, sctp_list);
3293 * Free the association after un-hashing the remote port.
3296 sctp_free_assoc(struct sctp_inpcb *inp, struct sctp_tcb *stcb)
3298 struct sctp_association *asoc;
3299 struct sctp_nets *net, *prev;
3300 struct sctp_laddr *laddr;
3301 struct sctp_tmit_chunk *chk;
3302 struct sctp_asconf_addr *aparam;
3303 struct sctp_socket_q_list *sq;
3305 /* first, lets purge the entry from the hash table. */
3306 if (stcb->asoc.state == 0) {
3307 kprintf("Freeing already free association:%p - huh??\n",
3313 /* now clean up any other timers */
3314 callout_stop(&asoc->hb_timer.timer);
3315 callout_stop(&asoc->dack_timer.timer);
3316 callout_stop(&asoc->asconf_timer.timer);
3317 callout_stop(&asoc->shut_guard_timer.timer);
3318 callout_stop(&asoc->autoclose_timer.timer);
3319 callout_stop(&asoc->delayed_event_timer.timer);
3320 TAILQ_FOREACH(net, &asoc->nets, sctp_next) {
3321 callout_stop(&net->rxt_timer.timer);
3322 callout_stop(&net->pmtu_timer.timer);
3325 /* Iterator asoc being freed we send an
3326 * unlocked TCB. It returns with INP_INFO
3327 * and INP write locked and the TCB locked
3328 * too and of course the iterator lock
3329 * in place as well..
3331 SCTP_TCB_UNLOCK(stcb);
3332 sctp_iterator_asoc_being_freed(inp, stcb);
3334 /* Null all of my entry's on the socket q */
3335 TAILQ_FOREACH(sq, &inp->sctp_queue_list, next_sq) {
3336 if (sq->tcb == stcb) {
3341 if (inp->sctp_tcb_at_block == (void *)stcb) {
3342 inp->error_on_block = ECONNRESET;
3345 if (inp->sctp_tcbhash) {
3346 LIST_REMOVE(stcb, sctp_tcbhash);
3348 /* Now lets remove it from the list of ALL associations in the EP */
3349 LIST_REMOVE(stcb, sctp_tcblist);
3350 SCTP_INP_WUNLOCK(inp);
3351 SCTP_ITERATOR_UNLOCK();
3354 /* pull from vtag hash */
3355 LIST_REMOVE(stcb, sctp_asocs);
3358 * Now before we can free the assoc, we must remove all of the
3359 * networks and any other allocated space.. i.e. add removes here
3360 * before the SCTP_ZONE_FREE() of the tasoc entry.
3363 sctp_add_vtag_to_timewait(inp, asoc->my_vtag);
3364 SCTP_INP_INFO_WUNLOCK();
3366 while (!TAILQ_EMPTY(&asoc->nets)) {
3367 net = TAILQ_FIRST(&asoc->nets);
3368 /* pull from list */
3369 if ((sctppcbinfo.ipi_count_raddr == 0) || (prev == net)) {
3373 TAILQ_REMOVE(&asoc->nets, net, sctp_next);
3376 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_net, net);
3377 sctppcbinfo.ipi_count_raddr--;
3380 * The chunk lists and such SHOULD be empty but we check them
3383 /* anything on the wheel needs to be removed */
3384 while (!TAILQ_EMPTY(&asoc->out_wheel)) {
3385 struct sctp_stream_out *outs;
3386 outs = TAILQ_FIRST(&asoc->out_wheel);
3387 TAILQ_REMOVE(&asoc->out_wheel, outs, next_spoke);
3388 /* now clean up any chunks here */
3389 chk = TAILQ_FIRST(&outs->outqueue);
3391 TAILQ_REMOVE(&outs->outqueue, chk, sctp_next);
3393 sctp_m_freem(chk->data);
3398 /* Free the chunk */
3399 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
3400 sctppcbinfo.ipi_count_chunk--;
3401 sctppcbinfo.ipi_gencnt_chunk++;
3402 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3403 panic("Chunk count is negative");
3405 chk = TAILQ_FIRST(&outs->outqueue);
3407 outs = TAILQ_FIRST(&asoc->out_wheel);
3410 if (asoc->pending_reply) {
3411 kfree(asoc->pending_reply, M_PCB);
3412 asoc->pending_reply = NULL;
3414 chk = TAILQ_FIRST(&asoc->pending_reply_queue);
3416 TAILQ_REMOVE(&asoc->pending_reply_queue, chk, sctp_next);
3418 sctp_m_freem(chk->data);
3423 /* Free the chunk */
3424 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
3425 sctppcbinfo.ipi_count_chunk--;
3426 sctppcbinfo.ipi_gencnt_chunk++;
3427 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3428 panic("Chunk count is negative");
3430 chk = TAILQ_FIRST(&asoc->pending_reply_queue);
3432 /* pending send queue SHOULD be empty */
3433 if (!TAILQ_EMPTY(&asoc->send_queue)) {
3434 chk = TAILQ_FIRST(&asoc->send_queue);
3436 TAILQ_REMOVE(&asoc->send_queue, chk, sctp_next);
3438 sctp_m_freem(chk->data);
3441 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
3442 sctppcbinfo.ipi_count_chunk--;
3443 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3444 panic("Chunk count is negative");
3446 sctppcbinfo.ipi_gencnt_chunk++;
3447 chk = TAILQ_FIRST(&asoc->send_queue);
3450 /* sent queue SHOULD be empty */
3451 if (!TAILQ_EMPTY(&asoc->sent_queue)) {
3452 chk = TAILQ_FIRST(&asoc->sent_queue);
3454 TAILQ_REMOVE(&asoc->sent_queue, chk, sctp_next);
3456 sctp_m_freem(chk->data);
3459 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
3460 sctppcbinfo.ipi_count_chunk--;
3461 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3462 panic("Chunk count is negative");
3464 sctppcbinfo.ipi_gencnt_chunk++;
3465 chk = TAILQ_FIRST(&asoc->sent_queue);
3468 /* control queue MAY not be empty */
3469 if (!TAILQ_EMPTY(&asoc->control_send_queue)) {
3470 chk = TAILQ_FIRST(&asoc->control_send_queue);
3472 TAILQ_REMOVE(&asoc->control_send_queue, chk, sctp_next);
3474 sctp_m_freem(chk->data);
3477 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
3478 sctppcbinfo.ipi_count_chunk--;
3479 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3480 panic("Chunk count is negative");
3482 sctppcbinfo.ipi_gencnt_chunk++;
3483 chk = TAILQ_FIRST(&asoc->control_send_queue);
3486 if (!TAILQ_EMPTY(&asoc->reasmqueue)) {
3487 chk = TAILQ_FIRST(&asoc->reasmqueue);
3489 TAILQ_REMOVE(&asoc->reasmqueue, chk, sctp_next);
3491 sctp_m_freem(chk->data);
3494 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
3495 sctppcbinfo.ipi_count_chunk--;
3496 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3497 panic("Chunk count is negative");
3499 sctppcbinfo.ipi_gencnt_chunk++;
3500 chk = TAILQ_FIRST(&asoc->reasmqueue);
3503 if (!TAILQ_EMPTY(&asoc->delivery_queue)) {
3504 chk = TAILQ_FIRST(&asoc->delivery_queue);
3506 TAILQ_REMOVE(&asoc->delivery_queue, chk, sctp_next);
3508 sctp_m_freem(chk->data);
3511 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
3512 sctppcbinfo.ipi_count_chunk--;
3513 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3514 panic("Chunk count is negative");
3516 sctppcbinfo.ipi_gencnt_chunk++;
3517 chk = TAILQ_FIRST(&asoc->delivery_queue);
3520 if (asoc->mapping_array) {
3521 kfree(asoc->mapping_array, M_PCB);
3522 asoc->mapping_array = NULL;
3525 /* the stream outs */
3526 if (asoc->strmout) {
3527 kfree(asoc->strmout, M_PCB);
3528 asoc->strmout = NULL;
3530 asoc->streamoutcnt = 0;
3533 for (i = 0; i < asoc->streamincnt; i++) {
3534 if (!TAILQ_EMPTY(&asoc->strmin[i].inqueue)) {
3535 /* We have somethings on the streamin queue */
3536 chk = TAILQ_FIRST(&asoc->strmin[i].inqueue);
3538 TAILQ_REMOVE(&asoc->strmin[i].inqueue,
3541 sctp_m_freem(chk->data);
3544 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk,
3546 sctppcbinfo.ipi_count_chunk--;
3547 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
3548 panic("Chunk count is negative");
3550 sctppcbinfo.ipi_gencnt_chunk++;
3551 chk = TAILQ_FIRST(&asoc->strmin[i].inqueue);
3555 kfree(asoc->strmin, M_PCB);
3556 asoc->strmin = NULL;
3558 asoc->streamincnt = 0;
3559 /* local addresses, if any */
3560 while (!LIST_EMPTY(&asoc->sctp_local_addr_list)) {
3561 laddr = LIST_FIRST(&asoc->sctp_local_addr_list);
3562 LIST_REMOVE(laddr, sctp_nxt_addr);
3563 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_laddr, laddr);
3564 sctppcbinfo.ipi_count_laddr--;
3566 /* pending asconf (address) parameters */
3567 while (!TAILQ_EMPTY(&asoc->asconf_queue)) {
3568 aparam = TAILQ_FIRST(&asoc->asconf_queue);
3569 TAILQ_REMOVE(&asoc->asconf_queue, aparam, next);
3570 kfree(aparam, M_PCB);
3572 if (asoc->last_asconf_ack_sent != NULL) {
3573 sctp_m_freem(asoc->last_asconf_ack_sent);
3574 asoc->last_asconf_ack_sent = NULL;
3576 /* Insert new items here :> */
3578 /* Get rid of LOCK */
3579 SCTP_TCB_LOCK_DESTROY(stcb);
3581 /* now clean up the tasoc itself */
3582 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_asoc, stcb);
3583 sctppcbinfo.ipi_count_asoc--;
3584 if ((inp->sctp_socket->so_snd.ssb_cc) ||
3585 (inp->sctp_socket->so_snd.ssb_mbcnt)) {
3586 /* This will happen when a abort is done */
3587 inp->sctp_socket->so_snd.ssb_cc = 0;
3588 inp->sctp_socket->so_snd.ssb_mbcnt = 0;
3590 if (inp->sctp_flags & SCTP_PCB_FLAGS_TCPTYPE) {
3591 if ((inp->sctp_flags & SCTP_PCB_FLAGS_IN_TCPPOOL) == 0) {
3592 if (inp->sctp_flags & SCTP_PCB_FLAGS_CONNECTED) {
3594 * For the base fd, that is NOT in TCP pool we
3595 * turn off the connected flag. This allows
3596 * non-listening endpoints to connect/shutdown/
3599 inp->sctp_flags &= ~SCTP_PCB_FLAGS_CONNECTED;
3600 soisdisconnected(inp->sctp_socket);
3603 * For those that are in the TCP pool we just leave
3604 * so it cannot be used. When they close the fd we
3609 if (inp->sctp_flags & SCTP_PCB_FLAGS_SOCKET_GONE) {
3610 sctp_inpcb_free(inp, 0);
3616 * determine if a destination is "reachable" based upon the addresses
3617 * bound to the current endpoint (e.g. only v4 or v6 currently bound)
3620 * FIX: if we allow assoc-level bindx(), then this needs to be fixed
3621 * to use assoc level v4/v6 flags, as the assoc *may* not have the
3622 * same address types bound as its endpoint
3625 sctp_destination_is_reachable(struct sctp_tcb *stcb, struct sockaddr *destaddr)
3627 struct sctp_inpcb *inp;
3630 /* No locks here, the TCB, in all cases is already
3631 * locked and an assoc is up. There is either a
3632 * INP lock by the caller applied (in asconf case when
3633 * deleting an address) or NOT in the HB case, however
3634 * if HB then the INP increment is up and the INP
3635 * will not be removed (on top of the fact that
3636 * we have a TCB lock). So we only want to
3637 * read the sctp_flags, which is either bound-all
3638 * or not.. no protection needed since once an
3639 * assoc is up you can't be changing your binding.
3641 inp = stcb->sctp_ep;
3642 if (inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) {
3643 /* if bound all, destination is not restricted */
3644 /* RRS: Question during lock work: Is this
3645 * correct? If you are bound-all you still
3646 * might need to obey the V4--V6 flags???
3647 * IMO this bound-all stuff needs to be removed!
3651 /* NOTE: all "scope" checks are done when local addresses are added */
3652 if (destaddr->sa_family == AF_INET6) {
3653 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
3654 answer = inp->inp_vflag & INP_IPV6;
3656 answer = inp->ip_inp.inp.inp_vflag & INP_IPV6;
3658 } else if (destaddr->sa_family == AF_INET) {
3659 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
3660 answer = inp->inp_vflag & INP_IPV4;
3662 answer = inp->ip_inp.inp.inp_vflag & INP_IPV4;
3665 /* invalid family, so it's unreachable */
3672 * update the inp_vflags on an endpoint
3675 sctp_update_ep_vflag(struct sctp_inpcb *inp) {
3676 struct sctp_laddr *laddr;
3678 /* first clear the flag */
3679 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
3682 inp->ip_inp.inp.inp_vflag = 0;
3684 /* set the flag based on addresses on the ep list */
3685 LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
3686 if (laddr->ifa == NULL) {
3688 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
3689 kprintf("An ounce of prevention is worth a pound of cure\n");
3691 #endif /* SCTP_DEBUG */
3694 if (laddr->ifa->ifa_addr) {
3697 if (laddr->ifa->ifa_addr->sa_family == AF_INET6) {
3698 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
3699 inp->inp_vflag |= INP_IPV6;
3701 inp->ip_inp.inp.inp_vflag |= INP_IPV6;
3703 } else if (laddr->ifa->ifa_addr->sa_family == AF_INET) {
3704 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
3705 inp->inp_vflag |= INP_IPV4;
3707 inp->ip_inp.inp.inp_vflag |= INP_IPV4;
3714 * Add the address to the endpoint local address list
3715 * There is nothing to be done if we are bound to all addresses
3718 sctp_add_local_addr_ep(struct sctp_inpcb *inp, struct ifaddr *ifa)
3720 struct sctp_laddr *laddr;
3724 if (inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) {
3725 /* You are already bound to all. You have it already */
3728 if (ifa->ifa_addr->sa_family == AF_INET6) {
3729 struct in6_ifaddr *ifa6;
3730 ifa6 = (struct in6_ifaddr *)ifa;
3731 if (ifa6->ia6_flags & (IN6_IFF_DETACHED |
3732 IN6_IFF_DEPRECATED | IN6_IFF_ANYCAST | IN6_IFF_NOTREADY))
3733 /* Can't bind a non-existent addr. */
3736 /* first, is it already present? */
3737 LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
3738 if (laddr->ifa == ifa) {
3744 if (((inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) == 0) && (fnd == 0)) {
3745 /* Not bound to all */
3746 error = sctp_insert_laddr(&inp->sctp_addr_list, ifa);
3750 /* update inp_vflag flags */
3751 if (ifa->ifa_addr->sa_family == AF_INET6) {
3752 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
3753 inp->inp_vflag |= INP_IPV6;
3755 inp->ip_inp.inp.inp_vflag |= INP_IPV6;
3757 } else if (ifa->ifa_addr->sa_family == AF_INET) {
3758 #if !(defined(__FreeBSD__) || defined(__APPLE__) || defined(__DragonFly__))
3759 inp->inp_vflag |= INP_IPV4;
3761 inp->ip_inp.inp.inp_vflag |= INP_IPV4;
3770 * select a new (hopefully reachable) destination net
3771 * (should only be used when we deleted an ep addr that is the
3772 * only usable source address to reach the destination net)
3775 sctp_select_primary_destination(struct sctp_tcb *stcb)
3777 struct sctp_nets *net;
3779 TAILQ_FOREACH(net, &stcb->asoc.nets, sctp_next) {
3780 /* for now, we'll just pick the first reachable one we find */
3781 if (net->dest_state & SCTP_ADDR_UNCONFIRMED)
3783 if (sctp_destination_is_reachable(stcb,
3784 (struct sockaddr *)&net->ro._l_addr)) {
3785 /* found a reachable destination */
3786 stcb->asoc.primary_destination = net;
3789 /* I can't there from here! ...we're gonna die shortly... */
3794 * Delete the address from the endpoint local address list
3795 * There is nothing to be done if we are bound to all addresses
3798 sctp_del_local_addr_ep(struct sctp_inpcb *inp, struct ifaddr *ifa)
3800 struct sctp_laddr *laddr;
3803 if (inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) {
3804 /* You are already bound to all. You have it already */
3808 LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
3809 if (laddr->ifa == ifa) {
3814 if (fnd && (inp->laddr_count < 2)) {
3815 /* can't delete unless there are at LEAST 2 addresses */
3818 if (((inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) == 0) && (fnd)) {
3820 * clean up any use of this address
3821 * go through our associations and clear any
3822 * last_used_address that match this one
3823 * for each assoc, see if a new primary_destination is needed
3825 struct sctp_tcb *stcb;
3827 /* clean up "next_addr_touse" */
3828 if (inp->next_addr_touse == laddr)
3829 /* delete this address */
3830 inp->next_addr_touse = NULL;
3832 /* clean up "last_used_address" */
3833 LIST_FOREACH(stcb, &inp->sctp_asoc_list, sctp_tcblist) {
3834 if (stcb->asoc.last_used_address == laddr)
3835 /* delete this address */
3836 stcb->asoc.last_used_address = NULL;
3837 } /* for each tcb */
3839 /* remove it from the ep list */
3840 sctp_remove_laddr(laddr);
3842 /* update inp_vflag flags */
3843 sctp_update_ep_vflag(inp);
3844 /* select a new primary destination if needed */
3845 LIST_FOREACH(stcb, &inp->sctp_asoc_list, sctp_tcblist) {
3846 /* presume caller (sctp_asconf.c) already owns INP lock */
3847 SCTP_TCB_LOCK(stcb);
3848 if (sctp_destination_is_reachable(stcb,
3849 (struct sockaddr *)&stcb->asoc.primary_destination->ro._l_addr) == 0) {
3850 sctp_select_primary_destination(stcb);
3852 SCTP_TCB_UNLOCK(stcb);
3853 } /* for each tcb */
3859 * Add the addr to the TCB local address list
3860 * For the BOUNDALL or dynamic case, this is a "pending" address list
3861 * (eg. addresses waiting for an ASCONF-ACK response)
3862 * For the subset binding, static case, this is a "valid" address list
3865 sctp_add_local_addr_assoc(struct sctp_tcb *stcb, struct ifaddr *ifa)
3867 struct sctp_laddr *laddr;
3870 /* Assumes TCP is locked.. and possiblye
3871 * the INP. May need to confirm/fix that if
3872 * we need it and is not the case.
3874 if (ifa->ifa_addr->sa_family == AF_INET6) {
3875 struct in6_ifaddr *ifa6;
3876 ifa6 = (struct in6_ifaddr *)ifa;
3877 if (ifa6->ia6_flags & (IN6_IFF_DETACHED |
3878 /* IN6_IFF_DEPRECATED | */
3881 /* Can't bind a non-existent addr. */
3884 /* does the address already exist? */
3885 LIST_FOREACH(laddr, &stcb->asoc.sctp_local_addr_list, sctp_nxt_addr) {
3886 if (laddr->ifa == ifa) {
3891 /* add to the list */
3892 error = sctp_insert_laddr(&stcb->asoc.sctp_local_addr_list, ifa);
3899 * insert an laddr entry with the given ifa for the desired list
3902 sctp_insert_laddr(struct sctpladdr *list, struct ifaddr *ifa) {
3903 struct sctp_laddr *laddr;
3905 laddr = (struct sctp_laddr *)SCTP_ZONE_GET(sctppcbinfo.ipi_zone_laddr);
3906 if (laddr == NULL) {
3907 /* out of memory? */
3910 sctppcbinfo.ipi_count_laddr++;
3911 sctppcbinfo.ipi_gencnt_laddr++;
3912 bzero(laddr, sizeof(*laddr));
3915 LIST_INSERT_HEAD(list, laddr, sctp_nxt_addr);
3921 * Remove an laddr entry from the local address list (on an assoc)
3924 sctp_remove_laddr(struct sctp_laddr *laddr)
3926 /* remove from the list */
3927 LIST_REMOVE(laddr, sctp_nxt_addr);
3928 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_laddr, laddr);
3929 sctppcbinfo.ipi_count_laddr--;
3930 sctppcbinfo.ipi_gencnt_laddr++;
3934 * Remove an address from the TCB local address list
3937 sctp_del_local_addr_assoc(struct sctp_tcb *stcb, struct ifaddr *ifa)
3939 struct sctp_inpcb *inp;
3940 struct sctp_laddr *laddr;
3942 /* This is called by asconf work. It is assumed that
3943 * a) The TCB is locked
3945 * b) The INP is locked.
3946 * This is true in as much as I can trace through
3947 * the entry asconf code where I did these locks.
3948 * Again, the ASCONF code is a bit different in
3949 * that it does lock the INP during its work often
3950 * times. This must be since we don't want other
3951 * proc's looking up things while what they are
3952 * looking up is changing :-D
3955 inp = stcb->sctp_ep;
3956 /* if subset bound and don't allow ASCONF's, can't delete last */
3957 if (((inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) == 0) &&
3958 ((inp->sctp_flags & SCTP_PCB_FLAGS_DO_ASCONF) == 0)) {
3959 if (stcb->asoc.numnets < 2) {
3960 /* can't delete last address */
3965 LIST_FOREACH(laddr, &stcb->asoc.sctp_local_addr_list, sctp_nxt_addr) {
3966 /* remove the address if it exists */
3967 if (laddr->ifa == NULL)
3969 if (laddr->ifa == ifa) {
3970 sctp_remove_laddr(laddr);
3975 /* address not found! */
3980 * Remove an address from the TCB local address list
3981 * lookup using a sockaddr addr
3984 sctp_del_local_addr_assoc_sa(struct sctp_tcb *stcb, struct sockaddr *sa)
3986 struct sctp_inpcb *inp;
3987 struct sctp_laddr *laddr;
3988 struct sockaddr *l_sa;
3991 * This function I find does not seem to have a caller.
3992 * As such we NEED TO DELETE this code. If we do
3993 * find a caller, the caller MUST have locked the TCB
3994 * at the least and probably the INP as well.
3996 inp = stcb->sctp_ep;
3997 /* if subset bound and don't allow ASCONF's, can't delete last */
3998 if (((inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) == 0) &&
3999 ((inp->sctp_flags & SCTP_PCB_FLAGS_DO_ASCONF) == 0)) {
4000 if (stcb->asoc.numnets < 2) {
4001 /* can't delete last address */
4006 LIST_FOREACH(laddr, &stcb->asoc.sctp_local_addr_list, sctp_nxt_addr) {
4007 /* make sure the address exists */
4008 if (laddr->ifa == NULL)
4010 if (laddr->ifa->ifa_addr == NULL)
4013 l_sa = laddr->ifa->ifa_addr;
4014 if (l_sa->sa_family == AF_INET6) {
4016 struct sockaddr_in6 *sin1, *sin2;
4017 sin1 = (struct sockaddr_in6 *)l_sa;
4018 sin2 = (struct sockaddr_in6 *)sa;
4019 if (memcmp(&sin1->sin6_addr, &sin2->sin6_addr,
4020 sizeof(struct in6_addr)) == 0) {
4022 sctp_remove_laddr(laddr);
4025 } else if (l_sa->sa_family == AF_INET) {
4027 struct sockaddr_in *sin1, *sin2;
4028 sin1 = (struct sockaddr_in *)l_sa;
4029 sin2 = (struct sockaddr_in *)sa;
4030 if (sin1->sin_addr.s_addr == sin2->sin_addr.s_addr) {
4032 sctp_remove_laddr(laddr);
4036 /* invalid family */
4040 /* address not found! */
4044 static char sctp_pcb_initialized = 0;
4046 #if defined(__FreeBSD__) || defined(__APPLE__)
4048 /* not used by DragonFly SCTP_ZONE_INIT macro */
4049 static int sctp_max_number_of_assoc = SCTP_MAX_NUM_OF_ASOC;
4050 static int sctp_scale_up_for_address = SCTP_SCALE_FOR_ADDR;
4052 #endif /* FreeBSD || APPLE || DragonFly */
4054 #ifndef SCTP_TCBHASHSIZE
4055 #define SCTP_TCBHASHSIZE 1024
4058 #ifndef SCTP_CHUNKQUEUE_SCALE
4059 #define SCTP_CHUNKQUEUE_SCALE 10
4066 * SCTP initialization for the PCB structures
4067 * should be called by the sctp_init() funciton.
4070 int hashtblsize = SCTP_TCBHASHSIZE;
4072 #if defined(__FreeBSD__) || defined(__APPLE__)
4073 /* not used by DragonFly SCTP_ZONE_INIT macro */
4074 int sctp_chunkscale = SCTP_CHUNKQUEUE_SCALE;
4077 if (sctp_pcb_initialized != 0) {
4078 /* error I was called twice */
4081 sctp_pcb_initialized = 1;
4083 /* Init all peg counts */
4084 for (i = 0; i < SCTP_NUMBER_OF_PEGS; i++) {
4088 /* init the empty list of (All) Endpoints */
4089 LIST_INIT(&sctppcbinfo.listhead);
4091 /* init the iterator head */
4092 LIST_INIT(&sctppcbinfo.iteratorhead);
4094 /* init the hash table of endpoints */
4095 #if defined(__FreeBSD__)
4096 #if defined(__FreeBSD_cc_version) && __FreeBSD_cc_version >= 440000
4097 TUNABLE_INT_FETCH("net.inet.sctp.tcbhashsize", &hashtblsize);
4098 TUNABLE_INT_FETCH("net.inet.sctp.pcbhashsize", &sctp_pcbtblsize);
4099 TUNABLE_INT_FETCH("net.inet.sctp.chunkscale", &sctp_chunkscale);
4101 TUNABLE_INT_FETCH("net.inet.sctp.tcbhashsize", SCTP_TCBHASHSIZE,
4103 TUNABLE_INT_FETCH("net.inet.sctp.pcbhashsize", SCTP_PCBHASHSIZE,
4105 TUNABLE_INT_FETCH("net.inet.sctp.chunkscale", SCTP_CHUNKQUEUE_SCALE,
4110 sctppcbinfo.sctp_asochash = hashinit((hashtblsize * 31),
4115 #if defined(__NetBSD__) || defined(__OpenBSD__)
4118 &sctppcbinfo.hashasocmark);
4120 sctppcbinfo.sctp_ephash = hashinit(hashtblsize,
4125 #if defined(__NetBSD__) || defined(__OpenBSD__)
4128 &sctppcbinfo.hashmark);
4130 sctppcbinfo.sctp_tcpephash = hashinit(hashtblsize,
4135 #if defined(__NetBSD__) || defined(__OpenBSD__)
4138 &sctppcbinfo.hashtcpmark);
4140 sctppcbinfo.hashtblsize = hashtblsize;
4142 /* init the zones */
4144 * FIX ME: Should check for NULL returns, but if it does fail we
4145 * are doomed to panic anyways... add later maybe.
4147 SCTP_ZONE_INIT(sctppcbinfo.ipi_zone_ep, "sctp_ep",
4148 sizeof(struct sctp_inpcb), maxsockets);
4150 SCTP_ZONE_INIT(sctppcbinfo.ipi_zone_asoc, "sctp_asoc",
4151 sizeof(struct sctp_tcb), sctp_max_number_of_assoc);
4153 SCTP_ZONE_INIT(sctppcbinfo.ipi_zone_laddr, "sctp_laddr",
4154 sizeof(struct sctp_laddr),
4155 (sctp_max_number_of_assoc * sctp_scale_up_for_address));
4157 SCTP_ZONE_INIT(sctppcbinfo.ipi_zone_net, "sctp_raddr",
4158 sizeof(struct sctp_nets),
4159 (sctp_max_number_of_assoc * sctp_scale_up_for_address));
4161 SCTP_ZONE_INIT(sctppcbinfo.ipi_zone_chunk, "sctp_chunk",
4162 sizeof(struct sctp_tmit_chunk),
4163 (sctp_max_number_of_assoc * sctp_scale_up_for_address *
4166 SCTP_ZONE_INIT(sctppcbinfo.ipi_zone_sockq, "sctp_sockq",
4167 sizeof(struct sctp_socket_q_list),
4168 (sctp_max_number_of_assoc * sctp_scale_up_for_address *
4171 /* Master Lock INIT for info structure */
4172 SCTP_INP_INFO_LOCK_INIT();
4173 SCTP_ITERATOR_LOCK_INIT();
4174 /* not sure if we need all the counts */
4175 sctppcbinfo.ipi_count_ep = 0;
4176 sctppcbinfo.ipi_gencnt_ep = 0;
4177 /* assoc/tcb zone info */
4178 sctppcbinfo.ipi_count_asoc = 0;
4179 sctppcbinfo.ipi_gencnt_asoc = 0;
4180 /* local addrlist zone info */
4181 sctppcbinfo.ipi_count_laddr = 0;
4182 sctppcbinfo.ipi_gencnt_laddr = 0;
4183 /* remote addrlist zone info */
4184 sctppcbinfo.ipi_count_raddr = 0;
4185 sctppcbinfo.ipi_gencnt_raddr = 0;
4187 sctppcbinfo.ipi_count_chunk = 0;
4188 sctppcbinfo.ipi_gencnt_chunk = 0;
4190 /* socket queue zone info */
4191 sctppcbinfo.ipi_count_sockq = 0;
4192 sctppcbinfo.ipi_gencnt_sockq = 0;
4195 sctppcbinfo.mbuf_track = 0;
4197 #if defined(__FreeBSD__) || defined(__OpenBSD__) || defined(__APPLE__) \
4198 || defined(__DragonFly__)
4199 sctppcbinfo.lastlow = ipport_firstauto;
4201 sctppcbinfo.lastlow = anonportmin;
4203 /* Init the TIMEWAIT list */
4204 for (i = 0; i < SCTP_STACK_VTAG_HASH_SIZE; i++) {
4205 LIST_INIT(&sctppcbinfo.vtag_timewait[i]);
4208 #if defined(_SCTP_NEEDS_CALLOUT_) && !defined(__APPLE__)
4209 TAILQ_INIT(&sctppcbinfo.callqueue);
4215 sctp_load_addresses_from_init(struct sctp_tcb *stcb, struct mbuf *m,
4216 int iphlen, int offset, int limit, struct sctphdr *sh,
4217 struct sockaddr *altsa)
4220 * grub through the INIT pulling addresses and
4221 * loading them to the nets structure in the asoc.
4222 * The from address in the mbuf should also be loaded
4223 * (if it is not already). This routine can be called
4224 * with either INIT or INIT-ACK's as long as the
4225 * m points to the IP packet and the offset points
4226 * to the beginning of the parameters.
4228 struct sctp_inpcb *inp, *l_inp;
4229 struct sctp_nets *net, *net_tmp;
4231 struct sctp_paramhdr *phdr, parm_buf;
4232 struct sctp_tcb *stcb_tmp;
4233 u_int16_t ptype, plen;
4234 struct sockaddr *sa;
4235 struct sockaddr_storage dest_store;
4236 struct sockaddr *local_sa = (struct sockaddr *)&dest_store;
4237 struct sockaddr_in sin;
4238 struct sockaddr_in6 sin6;
4240 /* First get the destination address setup too. */
4241 memset(&sin, 0, sizeof(sin));
4242 memset(&sin6, 0, sizeof(sin6));
4244 sin.sin_family = AF_INET;
4245 sin.sin_len = sizeof(sin);
4246 sin.sin_port = stcb->rport;
4248 sin6.sin6_family = AF_INET6;
4249 sin6.sin6_len = sizeof(struct sockaddr_in6);
4250 sin6.sin6_port = stcb->rport;
4251 if (altsa == NULL) {
4252 iph = mtod(m, struct ip *);
4253 if (iph->ip_v == IPVERSION) {
4255 struct sockaddr_in *sin_2;
4256 sin_2 = (struct sockaddr_in *)(local_sa);
4257 memset(sin_2, 0, sizeof(sin));
4258 sin_2->sin_family = AF_INET;
4259 sin_2->sin_len = sizeof(sin);
4260 sin_2->sin_port = sh->dest_port;
4261 sin_2->sin_addr.s_addr = iph->ip_dst.s_addr ;
4262 sin.sin_addr = iph->ip_src;
4263 sa = (struct sockaddr *)&sin;
4264 } else if (iph->ip_v == (IPV6_VERSION >> 4)) {
4266 struct ip6_hdr *ip6;
4267 struct sockaddr_in6 *sin6_2;
4269 ip6 = mtod(m, struct ip6_hdr *);
4270 sin6_2 = (struct sockaddr_in6 *)(local_sa);
4271 memset(sin6_2, 0, sizeof(sin6));
4272 sin6_2->sin6_family = AF_INET6;
4273 sin6_2->sin6_len = sizeof(struct sockaddr_in6);
4274 sin6_2->sin6_port = sh->dest_port;
4275 sin6.sin6_addr = ip6->ip6_src;
4276 sa = (struct sockaddr *)&sin6;
4282 * For cookies we use the src address NOT from the packet
4283 * but from the original INIT
4287 /* Turn off ECN until we get through all params */
4288 stcb->asoc.ecn_allowed = 0;
4290 TAILQ_FOREACH(net, &stcb->asoc.nets, sctp_next) {
4291 /* mark all addresses that we have currently on the list */
4292 net->dest_state |= SCTP_ADDR_NOT_IN_ASSOC;
4294 /* does the source address already exist? if so skip it */
4295 l_inp = inp = stcb->sctp_ep;
4296 stcb_tmp = sctp_findassociation_ep_addr(&inp, sa, &net_tmp, local_sa, stcb);
4297 if ((stcb_tmp == NULL && inp == stcb->sctp_ep) || inp == NULL) {
4298 /* we must add the source address */
4299 /* no scope set here since we have a tcb already. */
4300 if ((sa->sa_family == AF_INET) &&
4301 (stcb->asoc.ipv4_addr_legal)) {
4302 if (sctp_add_remote_addr(stcb, sa, 0, 2)) {
4305 } else if ((sa->sa_family == AF_INET6) &&
4306 (stcb->asoc.ipv6_addr_legal)) {
4307 if (sctp_add_remote_addr(stcb, sa, 0, 3)) {
4312 if (net_tmp != NULL && stcb_tmp == stcb) {
4313 net_tmp->dest_state &= ~SCTP_ADDR_NOT_IN_ASSOC;
4314 } else if (stcb_tmp != stcb) {
4315 /* It belongs to another association? */
4319 /* since a unlock occured we must check the
4320 * TCB's state and the pcb's gone flags.
4322 if (l_inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
4323 /* the user freed the ep */
4326 if (stcb->asoc.state == 0) {
4327 /* the assoc was freed? */
4331 /* now we must go through each of the params. */
4332 phdr = sctp_get_next_param(m, offset, &parm_buf, sizeof(parm_buf));
4334 ptype = ntohs(phdr->param_type);
4335 plen = ntohs(phdr->param_length);
4336 /*kprintf("ptype => %d, plen => %d\n", ptype, plen);*/
4337 if (offset + plen > limit) {
4343 if ((ptype == SCTP_IPV4_ADDRESS) &&
4344 (stcb->asoc.ipv4_addr_legal)) {
4345 struct sctp_ipv4addr_param *p4, p4_buf;
4346 /* ok get the v4 address and check/add */
4347 phdr = sctp_get_next_param(m, offset,
4348 (struct sctp_paramhdr *)&p4_buf, sizeof(p4_buf));
4349 if (plen != sizeof(struct sctp_ipv4addr_param) ||
4353 p4 = (struct sctp_ipv4addr_param *)phdr;
4354 sin.sin_addr.s_addr = p4->addr;
4355 sa = (struct sockaddr *)&sin;
4356 inp = stcb->sctp_ep;
4357 stcb_tmp = sctp_findassociation_ep_addr(&inp, sa, &net,
4360 if ((stcb_tmp== NULL && inp == stcb->sctp_ep) ||
4362 /* we must add the source address */
4363 /* no scope set since we have a tcb already */
4365 /* we must validate the state again here */
4366 if (l_inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
4367 /* the user freed the ep */
4370 if (stcb->asoc.state == 0) {
4371 /* the assoc was freed? */
4374 if (sctp_add_remote_addr(stcb, sa, 0, 4)) {
4377 } else if (stcb_tmp == stcb) {
4378 if (l_inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
4379 /* the user freed the ep */
4382 if (stcb->asoc.state == 0) {
4383 /* the assoc was freed? */
4389 ~SCTP_ADDR_NOT_IN_ASSOC;
4392 /* strange, address is in another assoc?
4393 * straighten out locks.
4395 SCTP_TCB_UNLOCK(stcb_tmp);
4396 SCTP_INP_RLOCK(inp);
4397 if (l_inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
4398 /* the user freed the ep */
4399 SCTP_INP_RUNLOCK(l_inp);
4402 if (stcb->asoc.state == 0) {
4403 /* the assoc was freed? */
4404 SCTP_INP_RUNLOCK(l_inp);
4407 SCTP_TCB_LOCK(stcb);
4408 SCTP_INP_RUNLOCK(stcb->sctp_ep);
4411 } else if ((ptype == SCTP_IPV6_ADDRESS) &&
4412 (stcb->asoc.ipv6_addr_legal)) {
4413 /* ok get the v6 address and check/add */
4414 struct sctp_ipv6addr_param *p6, p6_buf;
4415 phdr = sctp_get_next_param(m, offset,
4416 (struct sctp_paramhdr *)&p6_buf, sizeof(p6_buf));
4417 if (plen != sizeof(struct sctp_ipv6addr_param) ||
4421 p6 = (struct sctp_ipv6addr_param *)phdr;
4422 memcpy((caddr_t)&sin6.sin6_addr, p6->addr,
4424 sa = (struct sockaddr *)&sin6;
4425 inp = stcb->sctp_ep;
4426 stcb_tmp= sctp_findassociation_ep_addr(&inp, sa, &net,
4428 if (stcb_tmp == NULL && (inp == stcb->sctp_ep ||
4430 /* we must validate the state again here */
4431 if (l_inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
4432 /* the user freed the ep */
4435 if (stcb->asoc.state == 0) {
4436 /* the assoc was freed? */
4439 /* we must add the address, no scope set */
4440 if (sctp_add_remote_addr(stcb, sa, 0, 5)) {
4443 } else if (stcb_tmp == stcb) {
4444 /* we must validate the state again here */
4445 if (l_inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
4446 /* the user freed the ep */
4449 if (stcb->asoc.state == 0) {
4450 /* the assoc was freed? */
4456 ~SCTP_ADDR_NOT_IN_ASSOC;
4459 /* strange, address is in another assoc?
4460 * straighten out locks.
4462 SCTP_TCB_UNLOCK(stcb_tmp);
4463 SCTP_INP_RLOCK(l_inp);
4464 /* we must validate the state again here */
4465 if (l_inp->sctp_flags & (SCTP_PCB_FLAGS_SOCKET_GONE|SCTP_PCB_FLAGS_SOCKET_ALLGONE)) {
4466 /* the user freed the ep */
4467 SCTP_INP_RUNLOCK(l_inp);
4470 if (stcb->asoc.state == 0) {
4471 /* the assoc was freed? */
4472 SCTP_INP_RUNLOCK(l_inp);
4475 SCTP_TCB_LOCK(stcb);
4476 SCTP_INP_RUNLOCK(l_inp);
4479 } else if (ptype == SCTP_ECN_CAPABLE) {
4480 stcb->asoc.ecn_allowed = 1;
4481 } else if (ptype == SCTP_ULP_ADAPTION) {
4482 if (stcb->asoc.state != SCTP_STATE_OPEN) {
4483 struct sctp_adaption_layer_indication ai, *aip;
4485 phdr = sctp_get_next_param(m, offset,
4486 (struct sctp_paramhdr *)&ai, sizeof(ai));
4487 aip = (struct sctp_adaption_layer_indication *)phdr;
4488 sctp_ulp_notify(SCTP_NOTIFY_ADAPTION_INDICATION,
4489 stcb, ntohl(aip->indication), NULL);
4491 } else if (ptype == SCTP_SET_PRIM_ADDR) {
4492 struct sctp_asconf_addr_param lstore, *fee;
4493 struct sctp_asconf_addrv4_param *fii;
4495 struct sockaddr *lsa = NULL;
4497 stcb->asoc.peer_supports_asconf = 1;
4498 stcb->asoc.peer_supports_asconf_setprim = 1;
4499 if (plen > sizeof(lstore)) {
4502 phdr = sctp_get_next_param(m, offset,
4503 (struct sctp_paramhdr *)&lstore, plen);
4508 fee = (struct sctp_asconf_addr_param *)phdr;
4509 lptype = ntohs(fee->addrp.ph.param_type);
4510 if (lptype == SCTP_IPV4_ADDRESS) {
4512 sizeof(struct sctp_asconf_addrv4_param)) {
4513 kprintf("Sizeof setprim in init/init ack not %d but %d - ignored\n",
4514 (int)sizeof(struct sctp_asconf_addrv4_param),
4517 fii = (struct sctp_asconf_addrv4_param *)fee;
4518 sin.sin_addr.s_addr = fii->addrp.addr;
4519 lsa = (struct sockaddr *)&sin;
4521 } else if (lptype == SCTP_IPV6_ADDRESS) {
4523 sizeof(struct sctp_asconf_addr_param)) {
4524 kprintf("Sizeof setprim (v6) in init/init ack not %d but %d - ignored\n",
4525 (int)sizeof(struct sctp_asconf_addr_param),
4528 memcpy(sin6.sin6_addr.s6_addr,
4530 sizeof(fee->addrp.addr));
4531 lsa = (struct sockaddr *)&sin6;
4535 sctp_set_primary_addr(stcb, sa, NULL);
4538 } else if (ptype == SCTP_PRSCTP_SUPPORTED) {
4539 /* Peer supports pr-sctp */
4540 stcb->asoc.peer_supports_prsctp = 1;
4541 } else if (ptype == SCTP_SUPPORTED_CHUNK_EXT) {
4542 /* A supported extension chunk */
4543 struct sctp_supported_chunk_types_param *pr_supported;
4544 uint8_t local_store[128];
4547 phdr = sctp_get_next_param(m, offset,
4548 (struct sctp_paramhdr *)&local_store, plen);
4552 stcb->asoc.peer_supports_asconf = 0;
4553 stcb->asoc.peer_supports_asconf_setprim = 0;
4554 stcb->asoc.peer_supports_prsctp = 0;
4555 stcb->asoc.peer_supports_pktdrop = 0;
4556 stcb->asoc.peer_supports_strreset = 0;
4557 pr_supported = (struct sctp_supported_chunk_types_param *)phdr;
4558 num_ent = plen - sizeof(struct sctp_paramhdr);
4559 for (i=0; i<num_ent; i++) {
4560 switch (pr_supported->chunk_types[i]) {
4562 stcb->asoc.peer_supports_asconf = 1;
4563 stcb->asoc.peer_supports_asconf_setprim = 1;
4565 case SCTP_ASCONF_ACK:
4566 stcb->asoc.peer_supports_asconf = 1;
4567 stcb->asoc.peer_supports_asconf_setprim = 1;
4569 case SCTP_FORWARD_CUM_TSN:
4570 stcb->asoc.peer_supports_prsctp = 1;
4572 case SCTP_PACKET_DROPPED:
4573 stcb->asoc.peer_supports_pktdrop = 1;
4575 case SCTP_STREAM_RESET:
4576 stcb->asoc.peer_supports_strreset = 1;
4579 /* one I have not learned yet */
4584 } else if (ptype == SCTP_ECN_NONCE_SUPPORTED) {
4585 /* Peer supports ECN-nonce */
4586 stcb->asoc.peer_supports_ecn_nonce = 1;
4587 stcb->asoc.ecn_nonce_allowed = 1;
4588 } else if ((ptype == SCTP_HEARTBEAT_INFO) ||
4589 (ptype == SCTP_STATE_COOKIE) ||
4590 (ptype == SCTP_UNRECOG_PARAM) ||
4591 (ptype == SCTP_COOKIE_PRESERVE) ||
4592 (ptype == SCTP_SUPPORTED_ADDRTYPE) ||
4593 (ptype == SCTP_ADD_IP_ADDRESS) ||
4594 (ptype == SCTP_DEL_IP_ADDRESS) ||
4595 (ptype == SCTP_ERROR_CAUSE_IND) ||
4596 (ptype == SCTP_SUCCESS_REPORT)) {
4599 if ((ptype & 0x8000) == 0x0000) {
4600 /* must stop processing the rest of
4601 * the param's. Any report bits were
4602 * handled with the call to sctp_arethere_unrecognized_parameters()
4603 * when the INIT or INIT-ACK was first seen.
4608 offset += SCTP_SIZE32(plen);
4609 if (offset >= limit) {
4612 phdr = sctp_get_next_param(m, offset, &parm_buf,
4615 /* Now check to see if we need to purge any addresses */
4616 for (net = TAILQ_FIRST(&stcb->asoc.nets); net != NULL; net = net_tmp) {
4617 net_tmp = TAILQ_NEXT(net, sctp_next);
4618 if ((net->dest_state & SCTP_ADDR_NOT_IN_ASSOC) ==
4619 SCTP_ADDR_NOT_IN_ASSOC) {
4620 /* This address has been removed from the asoc */
4621 /* remove and free it */
4622 stcb->asoc.numnets--;
4623 TAILQ_REMOVE(&stcb->asoc.nets, net, sctp_next);
4624 sctp_free_remote_addr(net);
4625 if (net == stcb->asoc.primary_destination) {
4626 stcb->asoc.primary_destination = NULL;
4627 sctp_select_primary_destination(stcb);
4635 sctp_set_primary_addr(struct sctp_tcb *stcb, struct sockaddr *sa,
4636 struct sctp_nets *net)
4638 /* make sure the requested primary address exists in the assoc */
4639 if (net == NULL && sa)
4640 net = sctp_findnet(stcb, sa);
4643 /* didn't find the requested primary address! */
4646 /* set the primary address */
4647 if (net->dest_state & SCTP_ADDR_UNCONFIRMED) {
4648 /* Must be confirmed */
4651 stcb->asoc.primary_destination = net;
4652 net->dest_state &= ~SCTP_ADDR_WAS_PRIMARY;
4659 sctp_is_vtag_good(struct sctp_inpcb *inp, u_int32_t tag, struct timeval *now)
4662 * This function serves two purposes. It will see if a TAG can be
4663 * re-used and return 1 for yes it is ok and 0 for don't use that
4665 * A secondary function it will do is purge out old tags that can
4668 struct sctpasochead *head;
4669 struct sctpvtaghead *chain;
4670 struct sctp_tagblock *twait_block;
4671 struct sctp_tcb *stcb;
4674 SCTP_INP_INFO_WLOCK();
4675 chain = &sctppcbinfo.vtag_timewait[(tag % SCTP_STACK_VTAG_HASH_SIZE)];
4676 /* First is the vtag in use ? */
4678 head = &sctppcbinfo.sctp_asochash[SCTP_PCBHASH_ASOC(tag,
4679 sctppcbinfo.hashasocmark)];
4681 SCTP_INP_INFO_WUNLOCK();
4684 LIST_FOREACH(stcb, head, sctp_asocs) {
4685 if (stcb->asoc.my_vtag == tag) {
4686 /* We should remove this if and
4687 * return 0 always if we want vtags
4688 * unique across all endpoints. For
4689 * now within a endpoint is ok.
4691 if (inp == stcb->sctp_ep) {
4692 /* bad tag, in use */
4693 SCTP_INP_INFO_WUNLOCK();
4698 if (!LIST_EMPTY(chain)) {
4700 * Block(s) are present, lets see if we have this tag in
4703 LIST_FOREACH(twait_block, chain, sctp_nxt_tagblock) {
4704 for (i = 0; i < SCTP_NUMBER_IN_VTAG_BLOCK; i++) {
4705 if (twait_block->vtag_block[i].v_tag == 0) {
4708 } else if ((long)twait_block->vtag_block[i].tv_sec_at_expire >
4710 /* Audit expires this guy */
4711 twait_block->vtag_block[i].tv_sec_at_expire = 0;
4712 twait_block->vtag_block[i].v_tag = 0;
4713 } else if (twait_block->vtag_block[i].v_tag ==
4715 /* Bad tag, sorry :< */
4716 SCTP_INP_INFO_WUNLOCK();
4722 /* Not found, ok to use the tag */
4723 SCTP_INP_INFO_WUNLOCK();
4729 * Delete the address from the endpoint local address list
4730 * Lookup using a sockaddr address (ie. not an ifaddr)
4733 sctp_del_local_addr_ep_sa(struct sctp_inpcb *inp, struct sockaddr *sa)
4735 struct sctp_laddr *laddr;
4736 struct sockaddr *l_sa;
4738 /* Here is another function I cannot find a
4739 * caller for. As such we SHOULD delete it
4740 * if we have no users. If we find a user that
4741 * user MUST have the INP locked.
4745 if (inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) {
4746 /* You are already bound to all. You have it already */
4750 LIST_FOREACH(laddr, &inp->sctp_addr_list, sctp_nxt_addr) {
4751 /* make sure the address exists */
4752 if (laddr->ifa == NULL)
4754 if (laddr->ifa->ifa_addr == NULL)
4757 l_sa = laddr->ifa->ifa_addr;
4758 if (l_sa->sa_family == AF_INET6) {
4760 struct sockaddr_in6 *sin1, *sin2;
4761 sin1 = (struct sockaddr_in6 *)l_sa;
4762 sin2 = (struct sockaddr_in6 *)sa;
4763 if (memcmp(&sin1->sin6_addr, &sin2->sin6_addr,
4764 sizeof(struct in6_addr)) == 0) {
4769 } else if (l_sa->sa_family == AF_INET) {
4771 struct sockaddr_in *sin1, *sin2;
4772 sin1 = (struct sockaddr_in *)l_sa;
4773 sin2 = (struct sockaddr_in *)sa;
4774 if (sin1->sin_addr.s_addr == sin2->sin_addr.s_addr) {
4780 /* invalid family */
4785 if (found && inp->laddr_count < 2) {
4786 /* can't delete unless there are at LEAST 2 addresses */
4790 if (found && (inp->sctp_flags & SCTP_PCB_FLAGS_BOUNDALL) == 0) {
4792 * remove it from the ep list, this should NOT be
4793 * done until its really gone from the interface list and
4794 * we won't be receiving more of these. Probably right
4795 * away. If we do allow a removal of an address from
4796 * an association (sub-set bind) than this should NOT
4797 * be called until the all ASCONF come back from this
4800 sctp_remove_laddr(laddr);
4808 sctp_drain_mbufs(struct sctp_inpcb *inp, struct sctp_tcb *stcb)
4811 * We must hunt this association for MBUF's past the cumack
4812 * (i.e. out of order data that we can renege on).
4814 struct sctp_association *asoc;
4815 struct sctp_tmit_chunk *chk, *nchk;
4816 u_int32_t cumulative_tsn_p1, tsn;
4817 int cnt, strmat, gap;
4818 /* We look for anything larger than the cum-ack + 1 */
4821 cumulative_tsn_p1 = asoc->cumulative_tsn + 1;
4823 /* First look in the re-assembly queue */
4824 chk = TAILQ_FIRST(&asoc->reasmqueue);
4826 /* Get the next one */
4827 nchk = TAILQ_NEXT(chk, sctp_next);
4828 if (compare_with_wrap(chk->rec.data.TSN_seq,
4829 cumulative_tsn_p1, MAX_TSN)) {
4830 /* Yep it is above cum-ack */
4832 tsn = chk->rec.data.TSN_seq;
4833 if (tsn >= asoc->mapping_array_base_tsn) {
4834 gap = tsn - asoc->mapping_array_base_tsn;
4836 gap = (MAX_TSN - asoc->mapping_array_base_tsn) +
4839 asoc->size_on_reasm_queue -= chk->send_size;
4840 asoc->cnt_on_reasm_queue--;
4841 SCTP_UNSET_TSN_PRESENT(asoc->mapping_array, gap);
4842 TAILQ_REMOVE(&asoc->reasmqueue, chk, sctp_next);
4844 sctp_m_freem(chk->data);
4847 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
4848 sctppcbinfo.ipi_count_chunk--;
4849 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
4850 panic("Chunk count is negative");
4852 sctppcbinfo.ipi_gencnt_chunk++;
4856 /* Ok that was fun, now we will drain all the inbound streams? */
4857 for (strmat = 0; strmat < asoc->streamincnt; strmat++) {
4858 chk = TAILQ_FIRST(&asoc->strmin[strmat].inqueue);
4860 nchk = TAILQ_NEXT(chk, sctp_next);
4861 if (compare_with_wrap(chk->rec.data.TSN_seq,
4862 cumulative_tsn_p1, MAX_TSN)) {
4863 /* Yep it is above cum-ack */
4865 tsn = chk->rec.data.TSN_seq;
4866 if (tsn >= asoc->mapping_array_base_tsn) {
4868 asoc->mapping_array_base_tsn;
4871 asoc->mapping_array_base_tsn) +
4874 asoc->size_on_all_streams -= chk->send_size;
4875 asoc->cnt_on_all_streams--;
4877 SCTP_UNSET_TSN_PRESENT(asoc->mapping_array,
4879 TAILQ_REMOVE(&asoc->strmin[strmat].inqueue,
4882 sctp_m_freem(chk->data);
4885 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_chunk, chk);
4886 sctppcbinfo.ipi_count_chunk--;
4887 if ((int)sctppcbinfo.ipi_count_chunk < 0) {
4888 panic("Chunk count is negative");
4890 sctppcbinfo.ipi_gencnt_chunk++;
4896 * Question, should we go through the delivery queue?
4897 * The only reason things are on here is the app not reading OR a
4898 * p-d-api up. An attacker COULD send enough in to initiate the
4899 * PD-API and then send a bunch of stuff to other streams... these
4900 * would wind up on the delivery queue.. and then we would not get
4901 * to them. But in order to do this I then have to back-track and
4902 * un-deliver sequence numbers in streams.. el-yucko. I think for
4903 * now we will NOT look at the delivery queue and leave it to be
4904 * something to consider later. An alternative would be to abort
4905 * the P-D-API with a notification and then deliver the data....
4906 * Or another method might be to keep track of how many times the
4907 * situation occurs and if we see a possible attack underway just
4908 * abort the association.
4911 if (sctp_debug_on & SCTP_DEBUG_PCB1) {
4913 kprintf("Freed %d chunks from reneg harvest\n", cnt);
4916 #endif /* SCTP_DEBUG */
4919 * Another issue, in un-setting the TSN's in the mapping array we
4920 * DID NOT adjust the higest_tsn marker. This will cause one of
4921 * two things to occur. It may cause us to do extra work in checking
4922 * for our mapping array movement. More importantly it may cause us
4923 * to SACK every datagram. This may not be a bad thing though since
4924 * we will recover once we get our cum-ack above and all this stuff
4925 * we dumped recovered.
4933 * We must walk the PCB lists for ALL associations here. The system
4934 * is LOW on MBUF's and needs help. This is where reneging will
4935 * occur. We really hope this does NOT happen!
4937 struct sctp_inpcb *inp;
4938 struct sctp_tcb *stcb;
4940 SCTP_INP_INFO_RLOCK();
4941 LIST_FOREACH(inp, &sctppcbinfo.listhead, sctp_list) {
4942 /* For each endpoint */
4943 SCTP_INP_RLOCK(inp);
4944 LIST_FOREACH(stcb, &inp->sctp_asoc_list, sctp_tcblist) {
4945 /* For each association */
4946 SCTP_TCB_LOCK(stcb);
4947 sctp_drain_mbufs(inp, stcb);
4948 SCTP_TCB_UNLOCK(stcb);
4950 SCTP_INP_RUNLOCK(inp);
4952 SCTP_INP_INFO_RUNLOCK();
4956 sctp_add_to_socket_q(struct sctp_inpcb *inp, struct sctp_tcb *stcb)
4958 struct sctp_socket_q_list *sq;
4960 /* write lock on INP assumed */
4961 if ((inp == NULL) || (stcb == NULL)) {
4965 sq = (struct sctp_socket_q_list *)SCTP_ZONE_GET(
4966 sctppcbinfo.ipi_zone_sockq);
4968 /* out of sq structs */
4971 sctppcbinfo.ipi_count_sockq++;
4972 sctppcbinfo.ipi_gencnt_sockq++;
4974 stcb->asoc.cnt_msg_on_sb++;
4976 TAILQ_INSERT_TAIL(&inp->sctp_queue_list, sq, next_sq);
4982 sctp_remove_from_socket_q(struct sctp_inpcb *inp)
4984 struct sctp_tcb *stcb = NULL;
4985 struct sctp_socket_q_list *sq;
4987 /* W-Lock on INP assumed held */
4988 sq = TAILQ_FIRST(&inp->sctp_queue_list);
4993 TAILQ_REMOVE(&inp->sctp_queue_list, sq, next_sq);
4994 SCTP_ZONE_FREE(sctppcbinfo.ipi_zone_sockq, sq);
4995 sctppcbinfo.ipi_count_sockq--;
4996 sctppcbinfo.ipi_gencnt_sockq++;
4998 stcb->asoc.cnt_msg_on_sb--;
5004 sctp_initiate_iterator(asoc_func af, uint32_t pcb_state, uint32_t asoc_state,
5005 void *argp, uint32_t argi, end_func ef,
5006 struct sctp_inpcb *s_inp)
5008 struct sctp_iterator *it=NULL;
5013 it = kmalloc(sizeof(struct sctp_iterator), M_PCB, M_WAITOK);
5014 memset(it, 0, sizeof(*it));
5015 it->function_toapply = af;
5016 it->function_atend = ef;
5019 it->pcb_flags = pcb_state;
5020 it->asoc_state = asoc_state;
5023 it->iterator_flags = SCTP_ITERATOR_DO_SINGLE_INP;
5025 SCTP_INP_INFO_RLOCK();
5026 it->inp = LIST_FIRST(&sctppcbinfo.listhead);
5027 SCTP_INP_INFO_RUNLOCK();
5028 it->iterator_flags = SCTP_ITERATOR_DO_ALL_INP;
5031 /* Init the timer */
5032 #if defined(__FreeBSD__) && __FreeBSD_version >= 500000
5033 callout_init(&it->tmr.timer, 0);
5035 callout_init(&it->tmr.timer);
5037 /* add to the list of all iterators */
5038 SCTP_INP_INFO_WLOCK();
5039 LIST_INSERT_HEAD(&sctppcbinfo.iteratorhead, it, sctp_nxt_itr);
5040 SCTP_INP_INFO_WUNLOCK();
5041 sctp_iterator_timer(it);
5047 * Callout/Timer routines for OS that doesn't have them
5049 #ifdef _SCTP_NEEDS_CALLOUT_
5055 callout_init(struct callout *c)
5057 bzero(c, sizeof(*c));
5061 callout_reset(struct callout *c, int to_ticks, void (*ftn)(void *), void *arg)
5063 if (c->c_flags & CALLOUT_PENDING)
5067 * We could spl down here and back up at the TAILQ_INSERT_TAIL,
5068 * but there's no point since doing this setup doesn't take much
5075 c->c_flags = (CALLOUT_ACTIVE | CALLOUT_PENDING);
5078 c->c_time = to_ticks; /* just store the requested timeout */
5079 timeout(ftn, arg, to_ticks);
5081 c->c_time = ticks + to_ticks;
5082 TAILQ_INSERT_TAIL(&sctppcbinfo.callqueue, c, tqe);
5087 callout_stop(struct callout *c)
5090 * Don't attempt to delete a callout that's not on the queue.
5092 if (!(c->c_flags & CALLOUT_PENDING)) {
5093 c->c_flags &= ~CALLOUT_ACTIVE;
5096 c->c_flags &= ~(CALLOUT_ACTIVE | CALLOUT_PENDING| CALLOUT_FIRED);
5098 /* thread_call_cancel(c->c_call); */
5099 untimeout(c->c_func, c->c_arg);
5101 TAILQ_REMOVE(&sctppcbinfo.callqueue, c, tqe);
5107 #if !defined(__APPLE__)
5111 struct callout *c, *n;
5112 struct calloutlist locallist;
5115 /* run through and subtract and mark all callouts */
5116 c = TAILQ_FIRST(&sctppcbinfo.callqueue);
5118 n = TAILQ_NEXT(c, tqe);
5119 if (c->c_time <= ticks) {
5120 c->c_flags |= CALLOUT_FIRED;
5122 TAILQ_REMOVE(&sctppcbinfo.callqueue, c, tqe);
5124 TAILQ_INIT(&locallist);
5127 /* move off of main list */
5128 TAILQ_INSERT_TAIL(&locallist, c, tqe);
5132 /* Now all the ones on the locallist must be called */
5134 c = TAILQ_FIRST(&locallist);
5137 TAILQ_REMOVE(&locallist, c, tqe);
5138 /* now validate that it did not get canceled */
5139 if (c->c_flags & CALLOUT_FIRED) {
5140 c->c_flags &= ~CALLOUT_PENDING;
5141 (*c->c_func)(c->c_arg);
5143 c = TAILQ_FIRST(&locallist);
5148 #endif /* _SCTP_NEEDS_CALLOUT_ */