/* * Copyright (c) 2008-2016, 2022-2023 Apple Inc. All rights reserved. * * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * * This file contains Original Code and/or Modifications of Original Code * as defined in and that are subject to the Apple Public Source License * Version 2.0 (the 'License'). You may not use this file except in * compliance with the License. The rights granted to you under the License * may not be used to create, or enable the creation or redistribution of, * unlawful or unlicensed copies of an Apple operating system, or to * circumvent, violate, or enable the circumvention or violation of, any * terms of an Apple operating system software license agreement. * * Please obtain a copy of the License at * http://www.opensource.apple.com/apsl/ and read it before using this file. * * The Original Code and all software distributed under the License are * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES, * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT. * Please see the License for the specific language governing rights and * limitations under the License. * * @APPLE_OSREFERENCE_LICENSE_HEADER_END@ */ /* $FreeBSD: src/sys/netinet6/esp_input.c,v 1.1.2.3 2001/07/03 11:01:50 ume Exp $ */ /* $KAME: esp_input.c,v 1.55 2001/03/23 08:08:47 itojun Exp $ */ /* * Copyright (C) 1995, 1996, 1997, and 1998 WIDE Project. * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. Neither the name of the project nor the names of its contributors * may be used to endorse or promote products derived from this software * without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE PROJECT AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE PROJECT OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. */ #define _IP_VHL /* * RFC1827/2406 Encapsulated Security Payload. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #if SKYWALK #include #endif // SKYWALK #include #define DBG_LAYER_BEG NETDBG_CODE(DBG_NETIPSEC, 1) #define DBG_LAYER_END NETDBG_CODE(DBG_NETIPSEC, 3) #define DBG_FNC_ESPIN NETDBG_CODE(DBG_NETIPSEC, (6 << 8)) #define DBG_FNC_DECRYPT NETDBG_CODE(DBG_NETIPSEC, (7 << 8)) #define IPLEN_FLIPPED #define ESPMAXLEN \ (sizeof(struct esp) < sizeof(struct newesp) \ ? sizeof(struct newesp) : sizeof(struct esp)) static struct ip * esp4_input_strip_udp_encap(struct mbuf *m, int iphlen) { // strip the udp header that's encapsulating ESP struct ip *ip; u_int8_t stripsiz = (u_int8_t)sizeof(struct udphdr); ip = mtod(m, __typeof__(ip)); ovbcopy((caddr_t)ip, (caddr_t)(((u_char *)ip) + stripsiz), iphlen); m->m_data += stripsiz; m->m_len -= stripsiz; m->m_pkthdr.len -= stripsiz; ip = mtod(m, __typeof__(ip)); ip->ip_len = ip->ip_len - stripsiz; ip->ip_p = IPPROTO_ESP; return ip; } static struct ip6_hdr * esp6_input_strip_udp_encap(struct mbuf *m, int ip6hlen) { // strip the udp header that's encapsulating ESP struct ip6_hdr *ip6; u_int8_t stripsiz = (u_int8_t)sizeof(struct udphdr); ip6 = mtod(m, __typeof__(ip6)); ovbcopy((caddr_t)ip6, (caddr_t)(((u_char *)ip6) + stripsiz), ip6hlen); m->m_data += stripsiz; m->m_len -= stripsiz; m->m_pkthdr.len -= stripsiz; ip6 = mtod(m, __typeof__(ip6)); ip6->ip6_plen = htons(ntohs(ip6->ip6_plen) - stripsiz); ip6->ip6_nxt = IPPROTO_ESP; return ip6; } static void esp_input_log(struct mbuf *m, struct secasvar *sav, u_int32_t spi, u_int32_t seq) { if (net_mpklog_enabled && (sav->sah->ipsec_if->if_xflags & IFXF_MPK_LOG) == IFXF_MPK_LOG) { struct tcphdr th = {}; u_int32_t proto_len = 0; u_int8_t iphlen = 0; u_int8_t proto = 0; struct ip *inner_ip = mtod(m, struct ip *); if (IP_VHL_V(inner_ip->ip_vhl) == 4) { iphlen = (u_int8_t)(IP_VHL_HL(inner_ip->ip_vhl) << 2); proto = inner_ip->ip_p; } else if (IP_VHL_V(inner_ip->ip_vhl) == 6) { struct ip6_hdr *inner_ip6 = mtod(m, struct ip6_hdr *); iphlen = sizeof(struct ip6_hdr); proto = inner_ip6->ip6_nxt; } if (proto == IPPROTO_TCP) { if ((int)(iphlen + sizeof(th)) <= m->m_pkthdr.len) { m_copydata(m, iphlen, sizeof(th), (u_int8_t *)&th); } proto_len = m->m_pkthdr.len - iphlen - (th.th_off << 2); MPKL_ESP_INPUT_TCP(esp_mpkl_log_object, ntohl(spi), seq, ntohs(th.th_sport), ntohs(th.th_dport), ntohl(th.th_seq), proto_len); } } } void esp4_input(struct mbuf *m, int off) { (void)esp4_input_extended(m, off, NULL); } struct mbuf * esp4_input_extended(struct mbuf *m, int off, ifnet_t interface) { struct ip *ip; struct ip6_hdr *ip6; struct esp *esp; struct esptail esptail; u_int32_t spi; u_int32_t seq; u_int32_t replay_index = 0; struct secasvar *sav = NULL; size_t taillen; u_int16_t nxt; const struct esp_algorithm *algo; int ivlen; size_t esplen; u_int8_t hlen; sa_family_t ifamily; struct mbuf *out_m = NULL; KERNEL_DEBUG(DBG_FNC_ESPIN | DBG_FUNC_START, 0, 0, 0, 0, 0); /* sanity check for alignment. */ if (off % 4 != 0 || m->m_pkthdr.len % 4 != 0) { ipseclog((LOG_ERR, "IPv4 ESP input: packet alignment problem " "(off=%d, pktlen=%d)\n", off, m->m_pkthdr.len)); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } if (m->m_len < off + ESPMAXLEN) { m = m_pullup(m, off + ESPMAXLEN); if (!m) { ipseclog((LOG_DEBUG, "IPv4 ESP input: can't pullup in esp4_input\n")); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } } m->m_pkthdr.csum_flags &= ~CSUM_RX_FLAGS; /* Expect 32-bit aligned data pointer on strict-align platforms */ MBUF_STRICT_DATA_ALIGNMENT_CHECK_32(m); ip = mtod(m, struct ip *); // expect udp-encap and esp packets only if (ip->ip_p != IPPROTO_ESP && !(ip->ip_p == IPPROTO_UDP && off >= sizeof(struct udphdr))) { ipseclog((LOG_DEBUG, "IPv4 ESP input: invalid protocol type\n")); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } esp = (struct esp *)(void *)(((u_int8_t *)ip) + off); #ifdef _IP_VHL hlen = (u_int8_t)(IP_VHL_HL(ip->ip_vhl) << 2); #else hlen = ip->ip_hl << 2; #endif /* find the sassoc. */ spi = esp->esp_spi; if ((sav = key_allocsa_extended(AF_INET, (caddr_t)&ip->ip_src, (caddr_t)&ip->ip_dst, IFSCOPE_NONE, IPPROTO_ESP, spi, interface)) == 0) { ipseclog((LOG_WARNING, "IPv4 ESP input: no key association found for spi %u (0x%08x)\n", (u_int32_t)ntohl(spi), (u_int32_t)ntohl(spi))); IPSEC_STAT_INCREMENT(ipsecstat.in_nosa); goto bad; } KEYDEBUG(KEYDEBUG_IPSEC_STAMP, printf("DP esp4_input called to allocate SA:0x%llx\n", (uint64_t)VM_KERNEL_ADDRPERM(sav))); if (sav->state != SADB_SASTATE_MATURE && sav->state != SADB_SASTATE_DYING) { ipseclog((LOG_DEBUG, "IPv4 ESP input: non-mature/dying SA found for spi %u (0x%08x)\n", (u_int32_t)ntohl(spi), (u_int32_t)ntohl(spi))); IPSEC_STAT_INCREMENT(ipsecstat.in_badspi); goto bad; } algo = esp_algorithm_lookup(sav->alg_enc); if (!algo) { ipseclog((LOG_DEBUG, "IPv4 ESP input: " "unsupported encryption algorithm for spi %u (0x%08x)\n", (u_int32_t)ntohl(spi), (u_int32_t)ntohl(spi))); IPSEC_STAT_INCREMENT(ipsecstat.in_badspi); goto bad; } /* check if we have proper ivlen information */ ivlen = sav->ivlen; if (ivlen < 0) { ipseclog((LOG_ERR, "inproper ivlen in IPv4 ESP input: %s %s\n", ipsec4_logpacketstr(ip, spi), ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } seq = ntohl(((struct newesp *)esp)->esp_seq); if ((sav->flags2 & SADB_X_EXT_SA2_SEQ_PER_TRAFFIC_CLASS) == SADB_X_EXT_SA2_SEQ_PER_TRAFFIC_CLASS) { replay_index = seq >> PER_TC_REPLAY_WINDOW_SN_SHIFT; } if (!((sav->flags & SADB_X_EXT_OLD) == 0 && sav->replay[replay_index] != NULL && ((sav->alg_auth && sav->key_auth) || algo->finalizedecrypt))) { goto noreplaycheck; } if ((sav->alg_auth == SADB_X_AALG_NULL || sav->alg_auth == SADB_AALG_NONE) && !algo->finalizedecrypt) { goto noreplaycheck; } /* * check for sequence number. */ _CASSERT(MBUF_TC_MAX <= UINT8_MAX); if (ipsec_chkreplay(seq, sav, (u_int8_t)replay_index)) { ; /*okey*/ } else { IPSEC_STAT_INCREMENT(ipsecstat.in_espreplay); ipseclog((LOG_WARNING, "replay packet in IPv4 ESP input: seq(%u) idx(%u) %s %s\n", seq, (u_int8_t)replay_index, ipsec4_logpacketstr(ip, spi), ipsec_logsastr(sav))); goto bad; } /* Save ICV from packet for verification later */ size_t siz = 0; unsigned char saved_icv[AH_MAXSUMSIZE] __attribute__((aligned(4))); if (algo->finalizedecrypt) { siz = algo->icvlen; VERIFY(siz <= USHRT_MAX); if (m->m_pkthdr.len < off + ESPMAXLEN + siz) { ipseclog((LOG_DEBUG, "invalid ESP packet length %u, missing ICV\n", m->m_pkthdr.len)); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } m_copydata(m, m->m_pkthdr.len - (u_short)siz, (u_short)siz, (caddr_t) saved_icv); } else { /* check ICV immediately */ u_char sum0[AH_MAXSUMSIZE] __attribute__((aligned(4))); u_char sum[AH_MAXSUMSIZE] __attribute__((aligned(4))); const struct ah_algorithm *sumalgo; sumalgo = ah_algorithm_lookup(sav->alg_auth); if (!sumalgo) { goto noreplaycheck; } siz = (((*sumalgo->sumsiz)(sav) + 3) & ~(4 - 1)); if (m->m_pkthdr.len < off + ESPMAXLEN + siz) { ipseclog((LOG_DEBUG, "invalid ESP packet length %u, missing ICV\n", m->m_pkthdr.len)); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } if (AH_MAXSUMSIZE < siz) { ipseclog((LOG_DEBUG, "internal error: AH_MAXSUMSIZE must be larger than %u\n", (u_int32_t)siz)); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } m_copydata(m, m->m_pkthdr.len - (int)siz, (int)siz, (caddr_t) &sum0[0]); if (esp_auth(m, off, m->m_pkthdr.len - off - siz, sav, sum)) { ipseclog((LOG_WARNING, "auth fail in IPv4 ESP input: %s %s\n", ipsec4_logpacketstr(ip, spi), ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsecstat.in_espauthfail); goto bad; } if (cc_cmp_safe(siz, sum0, sum)) { ipseclog((LOG_WARNING, "cc_cmp fail in IPv4 ESP input: %s %s\n", ipsec4_logpacketstr(ip, spi), ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsecstat.in_espauthfail); goto bad; } m->m_flags |= M_AUTHIPDGM; IPSEC_STAT_INCREMENT(ipsecstat.in_espauthsucc); /* * update replay window. */ if ((sav->flags & SADB_X_EXT_OLD) == 0 && sav->replay[replay_index] != NULL) { if (ipsec_updatereplay(seq, sav, (u_int8_t)replay_index)) { IPSEC_STAT_INCREMENT(ipsecstat.in_espreplay); goto bad; } } } /* strip off the authentication data */ m_adj(m, (int)-siz); ip = mtod(m, struct ip *); #ifdef IPLEN_FLIPPED ip->ip_len = ip->ip_len - (u_short)siz; #else ip->ip_len = htons(ntohs(ip->ip_len) - siz); #endif noreplaycheck: /* process main esp header. */ if (sav->flags & SADB_X_EXT_OLD) { /* RFC 1827 */ esplen = sizeof(struct esp); } else { /* RFC 2406 */ if (sav->flags & SADB_X_EXT_DERIV) { esplen = sizeof(struct esp); } else { esplen = sizeof(struct newesp); } } if (m->m_pkthdr.len < off + esplen + ivlen + sizeof(esptail)) { ipseclog((LOG_WARNING, "IPv4 ESP input: packet too short\n")); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } if (m->m_len < off + esplen + ivlen) { m = m_pullup(m, (int)(off + esplen + ivlen)); if (!m) { ipseclog((LOG_DEBUG, "IPv4 ESP input: can't pullup in esp4_input\n")); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } ip = mtod(m, struct ip *); } /* * pre-compute and cache intermediate key */ if (esp_schedule(algo, sav) != 0) { IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } /* * decrypt the packet. */ if (!algo->decrypt) { panic("internal error: no decrypt function"); } KERNEL_DEBUG(DBG_FNC_DECRYPT | DBG_FUNC_START, 0, 0, 0, 0, 0); if ((*algo->decrypt)(m, off, sav, algo, ivlen)) { /* m is already freed */ m = NULL; ipseclog((LOG_ERR, "decrypt fail in IPv4 ESP input: %s\n", ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); KERNEL_DEBUG(DBG_FNC_DECRYPT | DBG_FUNC_END, 1, 0, 0, 0, 0); goto bad; } KERNEL_DEBUG(DBG_FNC_DECRYPT | DBG_FUNC_END, 2, 0, 0, 0, 0); IPSEC_STAT_INCREMENT(ipsecstat.in_esphist[sav->alg_enc]); m->m_flags |= M_DECRYPTED; if (algo->finalizedecrypt) { if ((*algo->finalizedecrypt)(sav, saved_icv, algo->icvlen)) { ipseclog((LOG_ERR, "esp4 packet decryption ICV failure: %s\n", ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsecstat.in_espauthfail); KERNEL_DEBUG(DBG_FNC_DECRYPT | DBG_FUNC_END, 1, 0, 0, 0, 0); goto bad; } else { m->m_flags |= M_AUTHIPDGM; IPSEC_STAT_INCREMENT(ipsecstat.in_espauthsucc); /* * update replay window. */ if ((sav->flags & SADB_X_EXT_OLD) == 0 && sav->replay[replay_index] != NULL) { if (ipsec_updatereplay(seq, sav, (u_int8_t)replay_index)) { IPSEC_STAT_INCREMENT(ipsecstat.in_espreplay); goto bad; } } } } /* * find the trailer of the ESP. */ m_copydata(m, m->m_pkthdr.len - sizeof(esptail), sizeof(esptail), (caddr_t)&esptail); nxt = esptail.esp_nxt; taillen = esptail.esp_padlen + sizeof(esptail); if (m->m_pkthdr.len < taillen || m->m_pkthdr.len - taillen < hlen) { /*?*/ ipseclog((LOG_WARNING, "bad pad length in IPv4 ESP input: %s %s\n", ipsec4_logpacketstr(ip, spi), ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } /* strip off the trailing pad area. */ m_adj(m, (int)-taillen); ip = mtod(m, struct ip *); #ifdef IPLEN_FLIPPED ip->ip_len = ip->ip_len - (u_short)taillen; #else ip->ip_len = htons(ntohs(ip->ip_len) - taillen); #endif if (ip->ip_p == IPPROTO_UDP) { // offset includes the outer ip and udp header lengths. if (m->m_len < off) { m = m_pullup(m, off); if (!m) { ipseclog((LOG_DEBUG, "IPv4 ESP input: invalid udp encapsulated ESP packet length \n")); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } ip = mtod(m, struct ip *); } // check the UDP encap header to detect changes in the source port, and then strip the header off -= sizeof(struct udphdr); // off no longer includes the udphdr's size // if peer is behind nat and this is the latest esp packet if ((sav->flags & SADB_X_EXT_NATT_DETECTED_PEER) != 0 && (sav->flags & SADB_X_EXT_OLD) == 0 && seq && sav->replay[replay_index] && seq >= sav->replay[replay_index]->lastseq) { struct udphdr *encap_uh = (__typeof__(encap_uh))(void *)((caddr_t)ip + off); if (encap_uh->uh_sport && ntohs(encap_uh->uh_sport) != sav->remote_ike_port) { sav->remote_ike_port = ntohs(encap_uh->uh_sport); } } ip = esp4_input_strip_udp_encap(m, off); esp = (struct esp *)(void *)(((u_int8_t *)ip) + off); } /* was it transmitted over the IPsec tunnel SA? */ if (ipsec4_tunnel_validate(m, (int)(off + esplen + ivlen), nxt, sav, &ifamily)) { ifaddr_t ifa; struct sockaddr_storage addr; /* * strip off all the headers that precedes ESP header. * IP4 xx ESP IP4' payload -> IP4' payload * * XXX more sanity checks * XXX relationship with gif? */ u_int8_t tos, otos; u_int8_t inner_ip_proto = 0; int sum; tos = ip->ip_tos; m_adj(m, (int)(off + esplen + ivlen)); if (ifamily == AF_INET) { struct sockaddr_in *ipaddr; if (m->m_len < sizeof(*ip)) { m = m_pullup(m, sizeof(*ip)); if (!m) { IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } } ip = mtod(m, struct ip *); /* ECN consideration. */ otos = ip->ip_tos; if (ip_ecn_egress(ip4_ipsec_ecn, &tos, &ip->ip_tos) == 0) { IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } if (otos != ip->ip_tos) { sum = ~ntohs(ip->ip_sum) & 0xffff; sum += (~otos & 0xffff) + ip->ip_tos; sum = (sum >> 16) + (sum & 0xffff); sum += (sum >> 16); /* add carry */ ip->ip_sum = htons(~sum & 0xffff); } if (!key_checktunnelsanity(sav, AF_INET, (caddr_t)&ip->ip_src, (caddr_t)&ip->ip_dst)) { ipseclog((LOG_ERR, "ipsec tunnel address mismatch " "in ESP input: %s %s\n", ipsec4_logpacketstr(ip, spi), ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } inner_ip_proto = ip->ip_p; bzero(&addr, sizeof(addr)); ipaddr = (__typeof__(ipaddr)) & addr; ipaddr->sin_family = AF_INET; ipaddr->sin_len = sizeof(*ipaddr); ipaddr->sin_addr = ip->ip_dst; } else if (ifamily == AF_INET6) { struct sockaddr_in6 *ip6addr; /* * m_pullup is prohibited in KAME IPv6 input processing * but there's no other way! */ if (m->m_len < sizeof(*ip6)) { m = m_pullup(m, sizeof(*ip6)); if (!m) { IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } } /* * Expect 32-bit aligned data pointer on strict-align * platforms. */ MBUF_STRICT_DATA_ALIGNMENT_CHECK_32(m); ip6 = mtod(m, struct ip6_hdr *); /* ECN consideration. */ if (ip64_ecn_egress(ip4_ipsec_ecn, &tos, &ip6->ip6_flow) == 0) { IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } if (!key_checktunnelsanity(sav, AF_INET6, (caddr_t)&ip6->ip6_src, (caddr_t)&ip6->ip6_dst)) { ipseclog((LOG_ERR, "ipsec tunnel address mismatch " "in ESP input: %s %s\n", ipsec6_logpacketstr(ip6, spi), ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } inner_ip_proto = ip6->ip6_nxt; bzero(&addr, sizeof(addr)); ip6addr = (__typeof__(ip6addr)) & addr; ip6addr->sin6_family = AF_INET6; ip6addr->sin6_len = sizeof(*ip6addr); ip6addr->sin6_addr = ip6->ip6_dst; } else { ipseclog((LOG_ERR, "ipsec tunnel unsupported address family " "in ESP input\n")); goto bad; } key_sa_recordxfer(sav, m->m_pkthdr.len); if (ipsec_incr_history_count(m, IPPROTO_ESP, spi) != 0 || ipsec_incr_history_count(m, IPPROTO_IPV4, 0) != 0) { IPSEC_STAT_INCREMENT(ipsecstat.in_nomem); goto bad; } // update the receiving interface address based on the inner address ifa = ifa_ifwithaddr((struct sockaddr *)&addr); if (ifa) { m->m_pkthdr.rcvif = ifa->ifa_ifp; ifa_remref(ifa); } /* Clear the csum flags, they can't be valid for the inner headers */ m->m_pkthdr.csum_flags = 0; // Input via IPsec interface lck_mtx_lock(sadb_mutex); ifnet_t ipsec_if = sav->sah->ipsec_if; if (ipsec_if != NULL) { // If an interface is found, add a reference count before dropping the lock ifnet_reference(ipsec_if); } lck_mtx_unlock(sadb_mutex); if ((m->m_pkthdr.pkt_flags & PKTF_WAKE_PKT) == PKTF_WAKE_PKT) { if (m->m_pkthdr.rcvif != NULL) { if_ports_used_match_mbuf(m->m_pkthdr.rcvif, ifamily, m); } else { ipseclog((LOG_ERR, "no input interface for ipsec wake packet\n")); } } if (ipsec_if != NULL) { esp_input_log(m, sav, spi, seq); // Return mbuf if (interface != NULL && interface == ipsec_if) { out_m = m; ifnet_release(ipsec_if); goto done; } errno_t inject_error = ipsec_inject_inbound_packet(ipsec_if, m); ifnet_release(ipsec_if); if (inject_error == 0) { m = NULL; goto done; } else { goto bad; } } if (proto_input(ifamily == AF_INET ? PF_INET : PF_INET6, m) != 0) { goto bad; } nxt = IPPROTO_DONE; KERNEL_DEBUG(DBG_FNC_ESPIN | DBG_FUNC_END, 2, 0, 0, 0, 0); } else { /* * strip off ESP header and IV. * even in m_pulldown case, we need to strip off ESP so that * we can always compute checksum for AH correctly. */ size_t stripsiz; stripsiz = esplen + ivlen; ip = mtod(m, struct ip *); ovbcopy((caddr_t)ip, (caddr_t)(((u_char *)ip) + stripsiz), off); m->m_data += stripsiz; m->m_len -= stripsiz; m->m_pkthdr.len -= stripsiz; ip = mtod(m, struct ip *); #ifdef IPLEN_FLIPPED ip->ip_len = ip->ip_len - (u_short)stripsiz; #else ip->ip_len = htons(ntohs(ip->ip_len) - stripsiz); #endif ip->ip_p = (u_int8_t)nxt; key_sa_recordxfer(sav, m->m_pkthdr.len); if (ipsec_incr_history_count(m, IPPROTO_ESP, spi) != 0) { IPSEC_STAT_INCREMENT(ipsecstat.in_nomem); goto bad; } /* * Set the csum valid flag, if we authenticated the * packet, the payload shouldn't be corrupt unless * it was corrupted before being signed on the other * side. */ if (nxt == IPPROTO_TCP || nxt == IPPROTO_UDP) { m->m_pkthdr.csum_flags = CSUM_DATA_VALID | CSUM_PSEUDO_HDR; m->m_pkthdr.csum_data = 0xFFFF; _CASSERT(offsetof(struct pkthdr, csum_data) == offsetof(struct pkthdr, csum_rx_val)); } if (nxt != IPPROTO_DONE) { if ((ip_protox[nxt]->pr_flags & PR_LASTHDR) != 0 && ipsec4_in_reject(m, NULL)) { IPSEC_STAT_INCREMENT(ipsecstat.in_polvio); goto bad; } KERNEL_DEBUG(DBG_FNC_ESPIN | DBG_FUNC_END, 3, 0, 0, 0, 0); /* translate encapsulated UDP port ? */ if ((sav->flags & SADB_X_EXT_NATT_MULTIPLEUSERS) != 0) { struct udphdr *udp; if (nxt != IPPROTO_UDP) { /* not UPD packet - drop it */ IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } if (m->m_len < off + sizeof(struct udphdr)) { m = m_pullup(m, off + sizeof(struct udphdr)); if (!m) { ipseclog((LOG_DEBUG, "IPv4 ESP input: can't pullup UDP header in esp4_input\n")); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } ip = mtod(m, struct ip *); } udp = (struct udphdr *)(void *)(((u_int8_t *)ip) + off); lck_mtx_lock(sadb_mutex); if (sav->natt_encapsulated_src_port == 0) { sav->natt_encapsulated_src_port = udp->uh_sport; } else if (sav->natt_encapsulated_src_port != udp->uh_sport) { /* something wrong */ IPSEC_STAT_INCREMENT(ipsecstat.in_inval); lck_mtx_unlock(sadb_mutex); goto bad; } lck_mtx_unlock(sadb_mutex); udp->uh_sport = htons(sav->remote_ike_port); udp->uh_sum = 0; } DTRACE_IP6(receive, struct mbuf *, m, struct inpcb *, NULL, struct ip *, ip, struct ifnet *, m->m_pkthdr.rcvif, struct ip *, ip, struct ip6_hdr *, NULL); // Input via IPsec interface legacy path lck_mtx_lock(sadb_mutex); ifnet_t ipsec_if = sav->sah->ipsec_if; if (ipsec_if != NULL) { // If an interface is found, add a reference count before dropping the lock ifnet_reference(ipsec_if); } lck_mtx_unlock(sadb_mutex); if (ipsec_if != NULL) { int mlen; if ((mlen = m_length2(m, NULL)) < hlen) { ipseclog((LOG_DEBUG, "IPv4 ESP input: decrypted packet too short %d < %u\n", mlen, hlen)); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); ifnet_release(ipsec_if); goto bad; } ip->ip_len = htons(ip->ip_len + hlen); ip->ip_off = htons(ip->ip_off); ip->ip_sum = 0; ip->ip_sum = ip_cksum_hdr_in(m, hlen); esp_input_log(m, sav, spi, seq); if ((m->m_pkthdr.pkt_flags & PKTF_WAKE_PKT) == PKTF_WAKE_PKT) { if_ports_used_match_mbuf(ipsec_if, PF_INET, m); } // Return mbuf if (interface != NULL && interface == ipsec_if) { out_m = m; ifnet_release(ipsec_if); goto done; } errno_t inject_error = ipsec_inject_inbound_packet(ipsec_if, m); ifnet_release(ipsec_if); if (inject_error == 0) { m = NULL; goto done; } else { goto bad; } } if ((m->m_pkthdr.pkt_flags & PKTF_WAKE_PKT) == PKTF_WAKE_PKT) { if_ports_used_match_mbuf(m->m_pkthdr.rcvif, PF_INET, m); if (m->m_pkthdr.rcvif == NULL) { ipseclog((LOG_ERR, "no input interface for ipsec wake packet\n")); } } ip_proto_dispatch_in(m, off, (u_int8_t)nxt, 0); } else { m_freem(m); } m = NULL; } done: if (sav) { KEYDEBUG(KEYDEBUG_IPSEC_STAMP, printf("DP esp4_input call free SA:0x%llx\n", (uint64_t)VM_KERNEL_ADDRPERM(sav))); key_freesav(sav, KEY_SADB_UNLOCKED); } IPSEC_STAT_INCREMENT(ipsecstat.in_success); return out_m; bad: if (sav) { KEYDEBUG(KEYDEBUG_IPSEC_STAMP, printf("DP esp4_input call free SA:0x%llx\n", (uint64_t)VM_KERNEL_ADDRPERM(sav))); key_freesav(sav, KEY_SADB_UNLOCKED); } if (m) { m_freem(m); } KERNEL_DEBUG(DBG_FNC_ESPIN | DBG_FUNC_END, 4, 0, 0, 0, 0); return out_m; } int esp6_input(struct mbuf **mp, int *offp, int proto) { return esp6_input_extended(mp, offp, proto, NULL); } int esp6_input_extended(struct mbuf **mp, int *offp, int proto, ifnet_t interface) { struct mbuf *m = *mp; int off = *offp; struct ip *ip; struct ip6_hdr *ip6; struct esp *esp; struct esptail esptail; u_int32_t spi; u_int32_t seq; u_int32_t replay_index = 0; struct secasvar *sav = NULL; u_int16_t nxt; const struct esp_algorithm *algo; int ivlen; size_t esplen; u_int16_t taillen; sa_family_t ifamily; /* sanity check for alignment. */ if (off % 4 != 0 || m->m_pkthdr.len % 4 != 0) { ipseclog((LOG_ERR, "IPv6 ESP input: packet alignment problem " "(off=%d, pktlen=%d)\n", off, m->m_pkthdr.len)); IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } #ifndef PULLDOWN_TEST IP6_EXTHDR_CHECK(m, off, ESPMAXLEN, {return IPPROTO_DONE;}); esp = (struct esp *)(void *)(mtod(m, caddr_t) + off); #else IP6_EXTHDR_GET(esp, struct esp *, m, off, ESPMAXLEN); if (esp == NULL) { IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); return IPPROTO_DONE; } #endif m->m_pkthdr.csum_flags &= ~CSUM_RX_FLAGS; /* Expect 32-bit data aligned pointer on strict-align platforms */ MBUF_STRICT_DATA_ALIGNMENT_CHECK_32(m); ip6 = mtod(m, struct ip6_hdr *); if (__improbable(ntohs(ip6->ip6_plen) == 0)) { ipseclog((LOG_ERR, "IPv6 ESP input: " "ESP with IPv6 jumbogram is not supported.\n")); IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } if (__improbable(proto != IPPROTO_ESP && !(proto == IPPROTO_UDP && off >= (sizeof(struct udphdr) + sizeof(struct ip6_hdr))))) { ipseclog((LOG_DEBUG, "IPv6 ESP input: invalid protocol type\n")); IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } /* find the sassoc. */ spi = esp->esp_spi; if ((sav = key_allocsa_extended(AF_INET6, (caddr_t)&ip6->ip6_src, (caddr_t)&ip6->ip6_dst, interface != NULL ? interface->if_index : IFSCOPE_UNKNOWN, IPPROTO_ESP, spi, interface)) == 0) { ipseclog((LOG_WARNING, "IPv6 ESP input: no key association found for spi %u (0x%08x) seq %u" " src %04x:%04x:%04x:%04x:%04x:%04x:%04x:%04x" " dst %04x:%04x:%04x:%04x:%04x:%04x:%04x:%04x if %s\n", (u_int32_t)ntohl(spi), (u_int32_t)ntohl(spi), ntohl(((struct newesp *)esp)->esp_seq), ntohs(ip6->ip6_src.__u6_addr.__u6_addr16[0]), ntohs(ip6->ip6_src.__u6_addr.__u6_addr16[1]), ntohs(ip6->ip6_src.__u6_addr.__u6_addr16[2]), ntohs(ip6->ip6_src.__u6_addr.__u6_addr16[3]), ntohs(ip6->ip6_src.__u6_addr.__u6_addr16[4]), ntohs(ip6->ip6_src.__u6_addr.__u6_addr16[5]), ntohs(ip6->ip6_src.__u6_addr.__u6_addr16[6]), ntohs(ip6->ip6_src.__u6_addr.__u6_addr16[7]), ntohs(ip6->ip6_dst.__u6_addr.__u6_addr16[0]), ntohs(ip6->ip6_dst.__u6_addr.__u6_addr16[1]), ntohs(ip6->ip6_dst.__u6_addr.__u6_addr16[2]), ntohs(ip6->ip6_dst.__u6_addr.__u6_addr16[3]), ntohs(ip6->ip6_dst.__u6_addr.__u6_addr16[4]), ntohs(ip6->ip6_dst.__u6_addr.__u6_addr16[5]), ntohs(ip6->ip6_dst.__u6_addr.__u6_addr16[6]), ntohs(ip6->ip6_dst.__u6_addr.__u6_addr16[7]), ((interface != NULL) ? if_name(interface) : "NONE"))); IPSEC_STAT_INCREMENT(ipsec6stat.in_nosa); goto bad; } KEYDEBUG(KEYDEBUG_IPSEC_STAMP, printf("DP esp6_input called to allocate SA:0x%llx\n", (uint64_t)VM_KERNEL_ADDRPERM(sav))); if (sav->state != SADB_SASTATE_MATURE && sav->state != SADB_SASTATE_DYING) { ipseclog((LOG_DEBUG, "IPv6 ESP input: non-mature/dying SA found for spi %u (0x%08x)\n", (u_int32_t)ntohl(spi), (u_int32_t)ntohl(spi))); IPSEC_STAT_INCREMENT(ipsec6stat.in_badspi); goto bad; } algo = esp_algorithm_lookup(sav->alg_enc); if (!algo) { ipseclog((LOG_DEBUG, "IPv6 ESP input: " "unsupported encryption algorithm for spi %u (0x%08x)\n", (u_int32_t)ntohl(spi), (u_int32_t)ntohl(spi))); IPSEC_STAT_INCREMENT(ipsec6stat.in_badspi); goto bad; } /* check if we have proper ivlen information */ ivlen = sav->ivlen; if (__improbable(ivlen < 0)) { ipseclog((LOG_ERR, "improper ivlen in IPv6 ESP input: %s %s\n", ipsec6_logpacketstr(ip6, spi), ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsec6stat.in_badspi); goto bad; } seq = ntohl(((struct newesp *)esp)->esp_seq); if ((sav->flags2 & SADB_X_EXT_SA2_SEQ_PER_TRAFFIC_CLASS) == SADB_X_EXT_SA2_SEQ_PER_TRAFFIC_CLASS) { replay_index = seq >> PER_TC_REPLAY_WINDOW_SN_SHIFT; } if (!((sav->flags & SADB_X_EXT_OLD) == 0 && sav->replay[replay_index] != NULL && ((sav->alg_auth && sav->key_auth) || algo->finalizedecrypt))) { goto noreplaycheck; } if ((sav->alg_auth == SADB_X_AALG_NULL || sav->alg_auth == SADB_AALG_NONE) && !algo->finalizedecrypt) { goto noreplaycheck; } /* * check for sequence number. */ if (ipsec_chkreplay(seq, sav, (u_int8_t)replay_index)) { ; /*okey*/ } else { IPSEC_STAT_INCREMENT(ipsec6stat.in_espreplay); ipseclog((LOG_WARNING, "replay packet in IPv6 ESP input: seq(%u) idx(%u) %s %s\n", seq, (u_int8_t)replay_index, ipsec6_logpacketstr(ip6, spi), ipsec_logsastr(sav))); goto bad; } /* Save ICV from packet for verification later */ size_t siz = 0; unsigned char saved_icv[AH_MAXSUMSIZE] __attribute__((aligned(4))); if (algo->finalizedecrypt) { siz = algo->icvlen; VERIFY(siz <= UINT16_MAX); if (m->m_pkthdr.len < off + ESPMAXLEN + siz) { ipseclog((LOG_DEBUG, "invalid ESP packet length %u, missing ICV\n", m->m_pkthdr.len)); IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } m_copydata(m, m->m_pkthdr.len - (int)siz, (int)siz, (caddr_t) saved_icv); } else { /* check ICV immediately */ u_char sum0[AH_MAXSUMSIZE] __attribute__((aligned(4))); u_char sum[AH_MAXSUMSIZE] __attribute__((aligned(4))); const struct ah_algorithm *sumalgo; sumalgo = ah_algorithm_lookup(sav->alg_auth); if (!sumalgo) { goto noreplaycheck; } siz = (((*sumalgo->sumsiz)(sav) + 3) & ~(4 - 1)); if (m->m_pkthdr.len < off + ESPMAXLEN + siz) { ipseclog((LOG_DEBUG, "invalid ESP packet length %u, missing ICV\n", m->m_pkthdr.len)); IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } if (__improbable(AH_MAXSUMSIZE < siz)) { ipseclog((LOG_DEBUG, "internal error: AH_MAXSUMSIZE must be larger than %u\n", (u_int32_t)siz)); IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } m_copydata(m, m->m_pkthdr.len - (int)siz, (int)siz, (caddr_t) &sum0[0]); if (esp_auth(m, off, m->m_pkthdr.len - off - siz, sav, sum)) { ipseclog((LOG_WARNING, "auth fail in IPv6 ESP input: %s %s\n", ipsec6_logpacketstr(ip6, spi), ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsec6stat.in_espauthfail); goto bad; } if (cc_cmp_safe(siz, sum0, sum)) { ipseclog((LOG_WARNING, "auth fail in IPv6 ESP input: %s %s\n", ipsec6_logpacketstr(ip6, spi), ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsec6stat.in_espauthfail); goto bad; } m->m_flags |= M_AUTHIPDGM; IPSEC_STAT_INCREMENT(ipsec6stat.in_espauthsucc); /* * update replay window. */ if ((sav->flags & SADB_X_EXT_OLD) == 0 && sav->replay[replay_index] != NULL) { if (ipsec_updatereplay(seq, sav, (u_int8_t)replay_index)) { IPSEC_STAT_INCREMENT(ipsec6stat.in_espreplay); goto bad; } } } /* strip off the authentication data */ m_adj(m, (int)-siz); ip6 = mtod(m, struct ip6_hdr *); ip6->ip6_plen = htons(ntohs(ip6->ip6_plen) - (u_int16_t)siz); noreplaycheck: /* process main esp header. */ if (sav->flags & SADB_X_EXT_OLD) { /* RFC 1827 */ esplen = sizeof(struct esp); } else { /* RFC 2406 */ if (sav->flags & SADB_X_EXT_DERIV) { esplen = sizeof(struct esp); } else { esplen = sizeof(struct newesp); } } if (m->m_pkthdr.len < off + esplen + ivlen + sizeof(esptail)) { ipseclog((LOG_WARNING, "IPv6 ESP input: packet too short\n")); IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } #ifndef PULLDOWN_TEST IP6_EXTHDR_CHECK(m, off, (int)(esplen + ivlen), goto bad); /*XXX*/ #else IP6_EXTHDR_GET(esp, struct esp *, m, off, esplen + ivlen); if (esp == NULL) { IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); m = NULL; goto bad; } #endif ip6 = mtod(m, struct ip6_hdr *); /*set it again just in case*/ /* * pre-compute and cache intermediate key */ if (esp_schedule(algo, sav) != 0) { IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } /* * decrypt the packet. */ if (!algo->decrypt) { panic("internal error: no decrypt function"); } KERNEL_DEBUG(DBG_FNC_DECRYPT | DBG_FUNC_START, 0, 0, 0, 0, 0); if ((*algo->decrypt)(m, off, sav, algo, ivlen)) { /* m is already freed */ m = NULL; ipseclog((LOG_ERR, "decrypt fail in IPv6 ESP input: %s\n", ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); KERNEL_DEBUG(DBG_FNC_DECRYPT | DBG_FUNC_END, 1, 0, 0, 0, 0); goto bad; } KERNEL_DEBUG(DBG_FNC_DECRYPT | DBG_FUNC_END, 2, 0, 0, 0, 0); IPSEC_STAT_INCREMENT(ipsec6stat.in_esphist[sav->alg_enc]); m->m_flags |= M_DECRYPTED; if (algo->finalizedecrypt) { if ((*algo->finalizedecrypt)(sav, saved_icv, algo->icvlen)) { ipseclog((LOG_ERR, "esp6 packet decryption ICV failure: %s\n", ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsec6stat.in_espauthfail); KERNEL_DEBUG(DBG_FNC_DECRYPT | DBG_FUNC_END, 1, 0, 0, 0, 0); goto bad; } else { m->m_flags |= M_AUTHIPDGM; IPSEC_STAT_INCREMENT(ipsec6stat.in_espauthsucc); /* * update replay window. */ if ((sav->flags & SADB_X_EXT_OLD) == 0 && sav->replay[replay_index] != NULL) { if (ipsec_updatereplay(seq, sav, (u_int8_t)replay_index)) { IPSEC_STAT_INCREMENT(ipsec6stat.in_espreplay); goto bad; } } } } /* * find the trailer of the ESP. */ m_copydata(m, m->m_pkthdr.len - sizeof(esptail), sizeof(esptail), (caddr_t)&esptail); nxt = esptail.esp_nxt; taillen = esptail.esp_padlen + sizeof(esptail); if (m->m_pkthdr.len < taillen || m->m_pkthdr.len - taillen < sizeof(struct ip6_hdr)) { /*?*/ ipseclog((LOG_WARNING, "bad pad length in IPv6 ESP input: %s %s\n", ipsec6_logpacketstr(ip6, spi), ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } /* strip off the trailing pad area. */ m_adj(m, -taillen); ip6 = mtod(m, struct ip6_hdr *); ip6->ip6_plen = htons(ntohs(ip6->ip6_plen) - taillen); if (proto == IPPROTO_UDP) { // offset includes the outer ip and udp header lengths. if (m->m_len < off) { m = m_pullup(m, off); if (!m) { ipseclog((LOG_DEBUG, "IPv6 ESP input: invalid udp encapsulated ESP packet length\n")); IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } ip6 = mtod(m, struct ip6_hdr *); } // check the UDP encap header to detect changes in the source port, and then strip the header off -= sizeof(struct udphdr); // off no longer includes the udphdr's size // if peer is behind nat and this is the latest esp packet if ((sav->flags & SADB_X_EXT_NATT_DETECTED_PEER) != 0 && (sav->flags & SADB_X_EXT_OLD) == 0 && seq && sav->replay[replay_index] && seq >= sav->replay[replay_index]->lastseq) { struct udphdr *encap_uh = (__typeof__(encap_uh))(void *)((caddr_t)ip6 + off); if (encap_uh->uh_sport && ntohs(encap_uh->uh_sport) != sav->remote_ike_port) { sav->remote_ike_port = ntohs(encap_uh->uh_sport); } } ip6 = esp6_input_strip_udp_encap(m, off); esp = (struct esp *)(void *)(((u_int8_t *)ip6) + off); } /* was it transmitted over the IPsec tunnel SA? */ if (ipsec6_tunnel_validate(m, (int)(off + esplen + ivlen), nxt, sav, &ifamily)) { ifaddr_t ifa; struct sockaddr_storage addr; u_int8_t inner_ip_proto = 0; /* * strip off all the headers that precedes ESP header. * IP6 xx ESP IP6' payload -> IP6' payload * * XXX more sanity checks * XXX relationship with gif? */ u_int32_t flowinfo; /*net endian*/ flowinfo = ip6->ip6_flow; m_adj(m, (int)(off + esplen + ivlen)); if (ifamily == AF_INET6) { struct sockaddr_in6 *ip6addr; if (m->m_len < sizeof(*ip6)) { #ifndef PULLDOWN_TEST /* * m_pullup is prohibited in KAME IPv6 input processing * but there's no other way! */ #else /* okay to pullup in m_pulldown style */ #endif m = m_pullup(m, sizeof(*ip6)); if (!m) { IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } } ip6 = mtod(m, struct ip6_hdr *); /* ECN consideration. */ if (ip6_ecn_egress(ip6_ipsec_ecn, &flowinfo, &ip6->ip6_flow) == 0) { IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } if (!key_checktunnelsanity(sav, AF_INET6, (caddr_t)&ip6->ip6_src, (caddr_t)&ip6->ip6_dst)) { ipseclog((LOG_ERR, "ipsec tunnel address mismatch " "in IPv6 ESP input: %s %s\n", ipsec6_logpacketstr(ip6, spi), ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } inner_ip_proto = ip6->ip6_nxt; bzero(&addr, sizeof(addr)); ip6addr = (__typeof__(ip6addr)) & addr; ip6addr->sin6_family = AF_INET6; ip6addr->sin6_len = sizeof(*ip6addr); ip6addr->sin6_addr = ip6->ip6_dst; } else if (ifamily == AF_INET) { struct sockaddr_in *ipaddr; if (m->m_len < sizeof(*ip)) { m = m_pullup(m, sizeof(*ip)); if (!m) { IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } } u_int8_t otos; int sum; ip = mtod(m, struct ip *); otos = ip->ip_tos; /* ECN consideration. */ if (ip46_ecn_egress(ip6_ipsec_ecn, &flowinfo, &ip->ip_tos) == 0) { IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } if (otos != ip->ip_tos) { sum = ~ntohs(ip->ip_sum) & 0xffff; sum += (~otos & 0xffff) + ip->ip_tos; sum = (sum >> 16) + (sum & 0xffff); sum += (sum >> 16); /* add carry */ ip->ip_sum = htons(~sum & 0xffff); } if (!key_checktunnelsanity(sav, AF_INET, (caddr_t)&ip->ip_src, (caddr_t)&ip->ip_dst)) { ipseclog((LOG_ERR, "ipsec tunnel address mismatch " "in ESP input: %s %s\n", ipsec4_logpacketstr(ip, spi), ipsec_logsastr(sav))); IPSEC_STAT_INCREMENT(ipsecstat.in_inval); goto bad; } inner_ip_proto = ip->ip_p; bzero(&addr, sizeof(addr)); ipaddr = (__typeof__(ipaddr)) & addr; ipaddr->sin_family = AF_INET; ipaddr->sin_len = sizeof(*ipaddr); ipaddr->sin_addr = ip->ip_dst; } key_sa_recordxfer(sav, m->m_pkthdr.len); if (ipsec_incr_history_count(m, IPPROTO_ESP, spi) != 0 || ipsec_incr_history_count(m, IPPROTO_IPV6, 0) != 0) { IPSEC_STAT_INCREMENT(ipsec6stat.in_nomem); goto bad; } // update the receiving interface address based on the inner address ifa = ifa_ifwithaddr((struct sockaddr *)&addr); if (ifa) { m->m_pkthdr.rcvif = ifa->ifa_ifp; ifa_remref(ifa); } // Input via IPsec interface lck_mtx_lock(sadb_mutex); ifnet_t ipsec_if = sav->sah->ipsec_if; if (ipsec_if != NULL) { // If an interface is found, add a reference count before dropping the lock ifnet_reference(ipsec_if); } lck_mtx_unlock(sadb_mutex); if ((m->m_pkthdr.pkt_flags & PKTF_WAKE_PKT) == PKTF_WAKE_PKT) { if_ports_used_match_mbuf(m->m_pkthdr.rcvif, ifamily, m); if (m->m_pkthdr.rcvif == NULL) { ipseclog((LOG_ERR, "no input interface for ipsec wake packet\n")); } } if (ipsec_if != NULL) { esp_input_log(m, sav, spi, seq); // Return mbuf if (interface != NULL && interface == ipsec_if) { ifnet_release(ipsec_if); goto done; } errno_t inject_error = ipsec_inject_inbound_packet(ipsec_if, m); ifnet_release(ipsec_if); if (inject_error == 0) { m = NULL; nxt = IPPROTO_DONE; goto done; } else { goto bad; } } if (proto_input(ifamily == AF_INET ? PF_INET : PF_INET6, m) != 0) { goto bad; } nxt = IPPROTO_DONE; } else { /* * strip off ESP header and IV. * even in m_pulldown case, we need to strip off ESP so that * we can always compute checksum for AH correctly. */ u_int16_t stripsiz; char *prvnxtp; /* * Set the next header field of the previous header correctly. */ prvnxtp = ip6_get_prevhdr(m, off); /* XXX */ *prvnxtp = (u_int8_t)nxt; VERIFY(esplen + ivlen <= UINT16_MAX); stripsiz = (u_int16_t)(esplen + ivlen); ip6 = mtod(m, struct ip6_hdr *); if (m->m_len >= stripsiz + off) { ovbcopy((caddr_t)ip6, ((caddr_t)ip6) + stripsiz, off); m->m_data += stripsiz; m->m_len -= stripsiz; m->m_pkthdr.len -= stripsiz; } else { /* * this comes with no copy if the boundary is on * cluster */ struct mbuf *n; n = m_split(m, off, M_DONTWAIT); if (n == NULL) { /* m is retained by m_split */ goto bad; } m_adj(n, stripsiz); /* m_cat does not update m_pkthdr.len */ m->m_pkthdr.len += n->m_pkthdr.len; m_cat(m, n); } #ifndef PULLDOWN_TEST /* * KAME requires that the packet to be contiguous on the * mbuf. We need to make that sure. * this kind of code should be avoided. * XXX other conditions to avoid running this part? */ if (m->m_len != m->m_pkthdr.len) { struct mbuf *n = NULL; int maxlen; MGETHDR(n, M_DONTWAIT, MT_HEADER); /* MAC-OK */ maxlen = MHLEN; if (n) { M_COPY_PKTHDR(n, m); } if (n && m->m_pkthdr.len > maxlen) { MCLGET(n, M_DONTWAIT); maxlen = MCLBYTES; if ((n->m_flags & M_EXT) == 0) { m_free(n); n = NULL; } } if (!n) { printf("esp6_input: mbuf allocation failed\n"); goto bad; } if (m->m_pkthdr.len <= maxlen) { m_copydata(m, 0, m->m_pkthdr.len, mtod(n, caddr_t)); n->m_len = m->m_pkthdr.len; n->m_pkthdr.len = m->m_pkthdr.len; n->m_next = NULL; m_freem(m); } else { m_copydata(m, 0, maxlen, mtod(n, caddr_t)); n->m_len = maxlen; n->m_pkthdr.len = m->m_pkthdr.len; n->m_next = m; m_adj(m, maxlen); m->m_flags &= ~M_PKTHDR; } m = n; } #endif ip6 = mtod(m, struct ip6_hdr *); ip6->ip6_plen = htons(ntohs(ip6->ip6_plen) - stripsiz); key_sa_recordxfer(sav, m->m_pkthdr.len); if (ipsec_incr_history_count(m, IPPROTO_ESP, spi) != 0) { IPSEC_STAT_INCREMENT(ipsec6stat.in_nomem); goto bad; } /* * Set the csum valid flag, if we authenticated the * packet, the payload shouldn't be corrupt unless * it was corrupted before being signed on the other * side. */ if (nxt == IPPROTO_TCP || nxt == IPPROTO_UDP) { m->m_pkthdr.csum_flags = CSUM_DATA_VALID | CSUM_PSEUDO_HDR; m->m_pkthdr.csum_data = 0xFFFF; _CASSERT(offsetof(struct pkthdr, csum_data) == offsetof(struct pkthdr, csum_rx_val)); } // Input via IPsec interface lck_mtx_lock(sadb_mutex); ifnet_t ipsec_if = sav->sah->ipsec_if; if (ipsec_if != NULL) { // If an interface is found, add a reference count before dropping the lock ifnet_reference(ipsec_if); } lck_mtx_unlock(sadb_mutex); if (ipsec_if != NULL) { esp_input_log(m, sav, spi, seq); if ((m->m_pkthdr.pkt_flags & PKTF_WAKE_PKT) == PKTF_WAKE_PKT) { if_ports_used_match_mbuf(ipsec_if, PF_INET6, m); } // Return mbuf if (interface != NULL && interface == ipsec_if) { ifnet_release(ipsec_if); goto done; } errno_t inject_error = ipsec_inject_inbound_packet(ipsec_if, m); ifnet_release(ipsec_if); if (inject_error == 0) { m = NULL; nxt = IPPROTO_DONE; goto done; } else { goto bad; } } else { if ((m->m_pkthdr.pkt_flags & PKTF_WAKE_PKT) == PKTF_WAKE_PKT) { if_ports_used_match_mbuf(m->m_pkthdr.rcvif, PF_INET, m); if (m->m_pkthdr.rcvif == NULL) { ipseclog((LOG_ERR, "no input interface for ipsec wake packet\n")); } } } } done: *offp = off; *mp = m; if (sav) { KEYDEBUG(KEYDEBUG_IPSEC_STAMP, printf("DP esp6_input call free SA:0x%llx\n", (uint64_t)VM_KERNEL_ADDRPERM(sav))); key_freesav(sav, KEY_SADB_UNLOCKED); } IPSEC_STAT_INCREMENT(ipsec6stat.in_success); return nxt; bad: if (sav) { KEYDEBUG(KEYDEBUG_IPSEC_STAMP, printf("DP esp6_input call free SA:0x%llx\n", (uint64_t)VM_KERNEL_ADDRPERM(sav))); key_freesav(sav, KEY_SADB_UNLOCKED); } if (m) { m_freem(m); } if (interface != NULL) { *mp = NULL; } return IPPROTO_DONE; } void esp6_ctlinput(int cmd, struct sockaddr *sa, void *d, __unused struct ifnet *ifp) { const struct newesp *espp; struct newesp esp; struct ip6ctlparam *ip6cp = NULL, ip6cp1; struct secasvar *sav; struct ip6_hdr *ip6; struct mbuf *m; int off = 0; struct sockaddr_in6 *sa6_src, *sa6_dst; if (sa->sa_family != AF_INET6 || sa->sa_len != sizeof(struct sockaddr_in6)) { return; } if ((unsigned)cmd >= PRC_NCMDS) { return; } /* if the parameter is from icmp6, decode it. */ if (d != NULL) { ip6cp = (struct ip6ctlparam *)d; m = ip6cp->ip6c_m; ip6 = ip6cp->ip6c_ip6; off = ip6cp->ip6c_off; } else { m = NULL; ip6 = NULL; } if (ip6) { /* * Notify the error to all possible sockets via pfctlinput2. * Since the upper layer information (such as protocol type, * source and destination ports) is embedded in the encrypted * data and might have been cut, we can't directly call * an upper layer ctlinput function. However, the pcbnotify * function will consider source and destination addresses * as well as the flow info value, and may be able to find * some PCB that should be notified. * Although pfctlinput2 will call esp6_ctlinput(), there is * no possibility of an infinite loop of function calls, * because we don't pass the inner IPv6 header. */ bzero(&ip6cp1, sizeof(ip6cp1)); ip6cp1.ip6c_src = ip6cp->ip6c_src; pfctlinput2(cmd, sa, (void *)&ip6cp1); /* * Then go to special cases that need ESP header information. * XXX: We assume that when ip6 is non NULL, * M and OFF are valid. */ /* check if we can safely examine src and dst ports */ if (m->m_pkthdr.len < off + sizeof(esp)) { return; } if (m->m_len < off + sizeof(esp)) { /* * this should be rare case, * so we compromise on this copy... */ m_copydata(m, off, sizeof(esp), (caddr_t)&esp); espp = &esp; } else { espp = (struct newesp*)(void *)(mtod(m, caddr_t) + off); } if (cmd == PRC_MSGSIZE) { int valid = 0; /* * Check to see if we have a valid SA corresponding to * the address in the ICMP message payload. */ sa6_src = ip6cp->ip6c_src; sa6_dst = (struct sockaddr_in6 *)(void *)sa; sav = key_allocsa(AF_INET6, (caddr_t)&sa6_src->sin6_addr, (caddr_t)&sa6_dst->sin6_addr, sa6_dst->sin6_scope_id, IPPROTO_ESP, espp->esp_spi); if (sav) { if (sav->state == SADB_SASTATE_MATURE || sav->state == SADB_SASTATE_DYING) { valid++; } key_freesav(sav, KEY_SADB_UNLOCKED); } /* XXX Further validation? */ /* * Depending on the value of "valid" and routing table * size (mtudisc_{hi,lo}wat), we will: * - recalcurate the new MTU and create the * corresponding routing entry, or * - ignore the MTU change notification. */ icmp6_mtudisc_update((struct ip6ctlparam *)d, valid); } } else { /* we normally notify any pcb here */ } } int esp_kpipe_input(ifnet_t interface, kern_packet_t sph, kern_packet_t dph) { struct newesp *esp = NULL; struct esptail *esptail = NULL; struct secasvar *sav = NULL; struct ipsecstat *stat = NULL; const struct esp_algorithm *e_algo = NULL; const struct ah_algorithm *a_algo = NULL; caddr_t src_ip = NULL, dst_ip = NULL; uint8_t *sbaddr = NULL, *dbaddr = NULL; uint8_t *src_payload = NULL, *dst_payload = NULL; uint8_t *iv = NULL; size_t iphlen = 0; size_t auth_size = 0; size_t esphlen = 0; u_int32_t replay_index = 0; int af = 0, ivlen = 0; int err = 0; uint32_t slim = 0, slen = 0; uint32_t dlim = 0; uint8_t dscp = 0, nxt_proto = 0; KERNEL_DEBUG(DBG_FNC_ESPIN | DBG_FUNC_START, 0, 0, 0, 0, 0); MD_BUFLET_ADDR(SK_PTR_ADDR_KPKT(sph), sbaddr); kern_buflet_t sbuf = __packet_get_next_buflet(sph, NULL); VERIFY(sbuf != NULL); slen = __buflet_get_data_length(sbuf); slim = __buflet_get_data_limit(sbuf); slim -= __buflet_get_data_offset(sbuf); MD_BUFLET_ADDR(SK_PTR_ADDR_KPKT(dph), dbaddr); kern_buflet_t dbuf = __packet_get_next_buflet(dph, NULL); VERIFY(dbuf != NULL); dlim = __buflet_get_data_limit(dbuf); dlim -= __buflet_get_data_offset(dbuf); struct ip *ip_hdr = (struct ip *)(void *)sbaddr; ASSERT(IP_HDR_ALIGNED_P(ip_hdr)); u_int ip_vers = IP_VHL_V(ip_hdr->ip_vhl); switch (ip_vers) { case IPVERSION: { #ifdef _IP_VHL iphlen = IP_VHL_HL(ip_hdr->ip_vhl) << 2; #else /* _IP_VHL */ iphlen = ip_hdr->ip_hl << 2; #endif /* _IP_VHL */ nxt_proto = ip_hdr->ip_p; dscp = ip_hdr->ip_tos >> IPTOS_DSCP_SHIFT; src_ip = (caddr_t)&ip_hdr->ip_src; dst_ip = (caddr_t)&ip_hdr->ip_dst; stat = &ipsecstat; af = AF_INET; break; } case 6: { struct ip6_hdr *ip6 = (struct ip6_hdr *)sbaddr; iphlen = sizeof(struct ip6_hdr); nxt_proto = ip6->ip6_nxt; dscp = (ntohl(ip6->ip6_flow) & IP6FLOW_DSCP_MASK) >> IP6FLOW_DSCP_SHIFT; src_ip = (caddr_t)&ip6->ip6_src; dst_ip = (caddr_t)&ip6->ip6_dst; stat = &ipsec6stat; af = AF_INET6; if (__improbable(ip6->ip6_plen == 0)) { esp_packet_log_err("esp kpipe input, jumbogram not supported"); IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } break; } default: { esp_log_info("esp kpipe input, ipversion %u, SPI=%x", ip_vers, ntohl(sav->spi)); err = EINVAL; goto bad; } } if (__improbable(dlim < slen)) { esp_packet_log_err("esp kpipe input, output buffer is short(%u), " "compared to input buffer(%u) SPI=%x\n", dlim, slen, ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_inval); err = EINVAL; goto bad; } if (__improbable(nxt_proto != IPPROTO_ESP)) { esp_packet_log_err("esp kpipe input, invalid nxt proto %u", nxt_proto); IPSEC_STAT_INCREMENT(stat->in_inval); err = EINVAL; goto bad; } if (__improbable(slen < (iphlen + sizeof(struct newesp)))) { esp_packet_log_err("esp kpipe input, slen too short %u", slen); IPSEC_STAT_INCREMENT(stat->in_inval); err = EINVAL; goto bad; } esp = (struct newesp *)(void *)(sbaddr + iphlen); sav = key_allocsa_extended(af, src_ip, dst_ip, interface != NULL ? interface->if_index: IFSCOPE_UNKNOWN, IPPROTO_ESP, esp->esp_spi, interface); if (__improbable(sav == NULL)) { if (ipsec_debug) { char src_buf[MAX_IPv6_STR_LEN] = {}; char dst_buf[MAX_IPv6_STR_LEN] = {}; inet_ntop(af, src_ip, src_buf, sizeof(src_buf)); inet_ntop(af, dst_ip, dst_buf, sizeof(src_buf)); esp_packet_log_err("esp kpipe input, no SA found for SPI=%x, " "packet %s<->%s", ntohl(esp->esp_spi), src_buf, dst_buf); } IPSEC_STAT_INCREMENT(stat->in_nosa); err = ENOENT; goto bad; } if (__improbable(sav->sah == NULL)) { esp_log_err("esp kpipe input, sah is NULL\n"); IPSEC_STAT_INCREMENT(stat->in_nosa); err = ENOENT; goto bad; } if (__improbable(sav->sah->saidx.mode != IPSEC_MODE_TRANSPORT)) { esp_log_err("ipsec tunnel mode not supported " "in kpipe mode, SPI=%x\n", ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_nosa); err = EINVAL; goto bad; } if (__improbable((sav->flags & (SADB_X_EXT_OLD | SADB_X_EXT_DERIV | SADB_X_EXT_NATT | SADB_X_EXT_NATT_MULTIPLEUSERS | SADB_X_EXT_CYCSEQ | SADB_X_EXT_PMASK)) != 0)) { esp_log_err("sadb flag %x not supported in " "kpipe mode, SPI=%x\n", sav->flags, ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_nosa); err = EINVAL; goto bad; } if (__improbable(sav->state != SADB_SASTATE_MATURE && sav->state != SADB_SASTATE_DYING)) { esp_log_info("esp kpipe input, invalid SA state %u, SPI=%x", sav->state, ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_inval); err = EINVAL; goto bad; } if ((sav->flags2 & SADB_X_EXT_SA2_SEQ_PER_TRAFFIC_CLASS) == SADB_X_EXT_SA2_SEQ_PER_TRAFFIC_CLASS) { replay_index = ntohl(esp->esp_seq) >> PER_TC_REPLAY_WINDOW_SN_SHIFT; } if (__improbable(sav->replay[replay_index] == NULL)) { esp_log_err("esp kpipe input, missing replay window, SPI=%x\n", ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_inval); err = EINVAL; goto bad; } /* * check for sequence number */ if (__improbable(!ipsec_chkreplay(ntohl(esp->esp_seq), sav, (uint8_t)replay_index))) { esp_packet_log_err("esp kpipe input, replay packet, " "seq(%u), idx(%u), SPI=%x\n", ntohl(esp->esp_seq), replay_index, ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_espreplay); err = EINVAL; goto bad; } e_algo = esp_algorithm_lookup(sav->alg_enc); if (__improbable(e_algo == NULL)) { esp_log_info("esp kpipe input, unsupported algorithm(%d) for, SPI=%x", sav->alg_enc, ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_inval); err = EINVAL; goto bad; } if ((sav->flags & SADB_X_EXT_IIV) == 0) { ivlen = sav->ivlen; if (__improbable(ivlen < 0)) { panic("esp kpipe input: invalid ivlen(%d) SPI=%x", ivlen, ntohl(sav->spi)); /* NOTREACHED */ __builtin_unreachable(); } iv = sbaddr + iphlen + sizeof(struct newesp); } esphlen = sizeof(struct newesp) + ivlen; if (sav->alg_auth != SADB_X_AALG_NULL && sav->alg_auth != SADB_AALG_NONE) { a_algo = ah_algorithm_lookup(sav->alg_auth); if (a_algo != NULL && sav->key_auth != NULL) { auth_size = (((*a_algo->sumsiz)(sav) + 3) & ~(4 - 1)); VERIFY(auth_size < AH_MAXSUMSIZE); if (__improbable(slen < iphlen + esphlen + auth_size)) { esp_packet_log_err("esp kpipe input, input buffer " "does not contain auth, SPI=%x\n", ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_espauthfail); err = EBADMSG; goto bad; } /* * Use destination buffer to store authentication * tag for comparison. */ uint8_t *auth_buf = dbaddr + dlim - auth_size; if (__improbable((err = esp_auth_data(sav, (uint8_t *)esp, slen - iphlen - auth_size, auth_buf, auth_size)) != 0)) { esp_packet_log_err("esp kpipe input, esp auth " "data failed, SPI=%x\n", ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_espauthfail); err = EBADMSG; goto bad; } if (__improbable(cc_cmp_safe(auth_size, auth_buf, sbaddr + slen - auth_size) != 0)) { esp_packet_log_err("esp kpipe input, auth compare " "failed, SPI=%x\n", ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_espauthfail); err = EBADMSG; goto bad; } IPSEC_STAT_INCREMENT(stat->in_espauthsucc); } } else if (e_algo->finalizedecrypt) { auth_size = e_algo->icvlen; } if (__improbable(slen <= (iphlen + esphlen + sizeof(struct esptail) + auth_size))) { esp_packet_log_err("esp kpipe input, input buffer is short(%u), " "to contain ivlen and esptail SPI=%x\n", slen, ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_inval); err = EBADMSG; goto bad; } /* * pre-compute and cache intermediate key */ if (__improbable((err = esp_schedule(e_algo, sav)) != 0)) { esp_log_info("esp schedule failed %d, SPI=%x\n", err, ntohl(sav->spi)); IPSEC_STAT_INCREMENT(ipsec6stat.in_inval); goto bad; } VERIFY(e_algo->decrypt_pkt); KERNEL_DEBUG(DBG_FNC_DECRYPT | DBG_FUNC_START, 0, 0, 0, 0, 0); src_payload = sbaddr + iphlen + esphlen; dst_payload = dbaddr + iphlen; uint16_t encrypted_payload_len = (uint16_t)(slen - iphlen - esphlen - auth_size); if (__improbable((err = (*e_algo->decrypt_pkt)(sav, src_payload, encrypted_payload_len, esp, iv, ivlen, dst_payload, encrypted_payload_len)) != 0)) { esp_packet_log_err("esp kpipe input: decryption failed, SPI=%x\n", ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_inval); KERNEL_DEBUG(DBG_FNC_DECRYPT | DBG_FUNC_END, 1, 0, 0, 0, 0); goto bad; } KERNEL_DEBUG(DBG_FNC_DECRYPT | DBG_FUNC_END, 2, 0, 0, 0, 0); IPSEC_STAT_INCREMENT(stat->in_esphist[sav->alg_enc]); if (e_algo->finalizedecrypt) { if (__improbable((err = (*e_algo->finalizedecrypt)(sav, sbaddr + slen - auth_size, e_algo->icvlen)) != 0)) { esp_packet_log_err("esp kpipe input: ICV failed, SPI=%x\n", ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_espauthfail); KERNEL_DEBUG(DBG_FNC_DECRYPT | DBG_FUNC_END, 3, 0, 0, 0, 0); goto bad; } IPSEC_STAT_INCREMENT(stat->in_espauthsucc); KERNEL_DEBUG(DBG_FNC_DECRYPT | DBG_FUNC_END, 4, 0, 0, 0, 0); } if (__improbable(ipsec_updatereplay(ntohl(esp->esp_seq), sav, (uint8_t)replay_index))) { esp_packet_log_err("esp kpipe input: update replay failed, SPI=%x\n", ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_espreplay); goto bad; } esptail = (struct esptail *)(dst_payload + encrypted_payload_len - sizeof(struct esptail)); nxt_proto = esptail->esp_nxt; size_t taillen = sizeof(struct esptail) + esptail->esp_padlen; if (__improbable(encrypted_payload_len <= taillen)) { esp_packet_log_err("esp kpipe input: encrypted payload len %u, " "is invalid, taillen %zu, SPI=%x\n", encrypted_payload_len, taillen, ntohl(sav->spi)); IPSEC_STAT_INCREMENT(stat->in_inval); goto bad; } uint16_t decrypted_payload_len = encrypted_payload_len - (uint16_t)taillen; switch (ip_vers) { case IPVERSION: { struct ip *ip = (struct ip *)(void *)dbaddr; ASSERT(IP_HDR_ALIGNED_P(ip)); ip->ip_p = nxt_proto; ip->ip_len = htons((uint16_t)(iphlen + decrypted_payload_len)); ip->ip_sum = 0; // Recalculate checksum ip->ip_sum = in_cksum_hdr_opt(ip); break; } case 6: { struct ip6_hdr *ip6 = (struct ip6_hdr *)dbaddr; ip6->ip6_plen = htons((uint16_t)decrypted_payload_len); ip6->ip6_nxt = nxt_proto; break; } } if (nxt_proto == IPPROTO_TCP || nxt_proto == IPPROTO_UDP) { __packet_set_inet_checksum(dph, PACKET_CSUM_DATA_VALID | PACKET_CSUM_PSEUDO_HDR, 0, 0xFFFF, FALSE); } __buflet_set_data_length(dbuf, (uint16_t)(iphlen + decrypted_payload_len)); key_sa_recordxfer(sav, iphlen + decrypted_payload_len); IPSEC_STAT_INCREMENT(stat->in_success); key_freesav(sav, KEY_SADB_UNLOCKED); KERNEL_DEBUG(DBG_FNC_ESPIN | DBG_FUNC_END, 1, 0, 0, 0, 0); return 0; bad: if (sav != NULL) { key_freesav(sav, KEY_SADB_UNLOCKED); sav = NULL; } KERNEL_DEBUG(DBG_FNC_ESPIN | DBG_FUNC_END, 2, err, 0, 0, 0); return err; }