summaryrefslogtreecommitdiff
path: root/usr/src/uts/common/inet/ip/ip_output.c
diff options
context:
space:
mode:
Diffstat (limited to 'usr/src/uts/common/inet/ip/ip_output.c')
-rw-r--r--usr/src/uts/common/inet/ip/ip_output.c2554
1 files changed, 2554 insertions, 0 deletions
diff --git a/usr/src/uts/common/inet/ip/ip_output.c b/usr/src/uts/common/inet/ip/ip_output.c
new file mode 100644
index 0000000000..a4940fd3e8
--- /dev/null
+++ b/usr/src/uts/common/inet/ip/ip_output.c
@@ -0,0 +1,2554 @@
+/*
+ * CDDL HEADER START
+ *
+ * The contents of this file are subject to the terms of the
+ * Common Development and Distribution License (the "License").
+ * You may not use this file except in compliance with the License.
+ *
+ * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
+ * or http://www.opensolaris.org/os/licensing.
+ * See the License for the specific language governing permissions
+ * and limitations under the License.
+ *
+ * When distributing Covered Code, include this CDDL HEADER in each
+ * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
+ * If applicable, add the following below this CDDL HEADER, with the
+ * fields enclosed by brackets "[]" replaced with your own identifying
+ * information: Portions Copyright [yyyy] [name of copyright owner]
+ *
+ * CDDL HEADER END
+ */
+
+/*
+ * Copyright 2009 Sun Microsystems, Inc. All rights reserved.
+ * Use is subject to license terms.
+ */
+/* Copyright (c) 1990 Mentat Inc. */
+
+#include <sys/types.h>
+#include <sys/stream.h>
+#include <sys/strsubr.h>
+#include <sys/dlpi.h>
+#include <sys/strsun.h>
+#include <sys/zone.h>
+#include <sys/ddi.h>
+#include <sys/sunddi.h>
+#include <sys/cmn_err.h>
+#include <sys/debug.h>
+#include <sys/atomic.h>
+
+#include <sys/systm.h>
+#include <sys/param.h>
+#include <sys/kmem.h>
+#include <sys/sdt.h>
+#include <sys/socket.h>
+#include <sys/mac.h>
+#include <net/if.h>
+#include <net/if_arp.h>
+#include <net/route.h>
+#include <sys/sockio.h>
+#include <netinet/in.h>
+#include <net/if_dl.h>
+
+#include <inet/common.h>
+#include <inet/mi.h>
+#include <inet/mib2.h>
+#include <inet/nd.h>
+#include <inet/arp.h>
+#include <inet/snmpcom.h>
+#include <inet/kstatcom.h>
+
+#include <netinet/igmp_var.h>
+#include <netinet/ip6.h>
+#include <netinet/icmp6.h>
+#include <netinet/sctp.h>
+
+#include <inet/ip.h>
+#include <inet/ip_impl.h>
+#include <inet/ip6.h>
+#include <inet/ip6_asp.h>
+#include <inet/tcp.h>
+#include <inet/ip_multi.h>
+#include <inet/ip_if.h>
+#include <inet/ip_ire.h>
+#include <inet/ip_ftable.h>
+#include <inet/ip_rts.h>
+#include <inet/optcom.h>
+#include <inet/ip_ndp.h>
+#include <inet/ip_listutils.h>
+#include <netinet/igmp.h>
+#include <netinet/ip_mroute.h>
+#include <inet/ipp_common.h>
+
+#include <net/pfkeyv2.h>
+#include <inet/sadb.h>
+#include <inet/ipsec_impl.h>
+#include <inet/ipdrop.h>
+#include <inet/ip_netinfo.h>
+
+#include <sys/pattr.h>
+#include <inet/ipclassifier.h>
+#include <inet/sctp_ip.h>
+#include <inet/sctp/sctp_impl.h>
+#include <inet/udp_impl.h>
+#include <sys/sunddi.h>
+
+#include <sys/tsol/label.h>
+#include <sys/tsol/tnet.h>
+
+#ifdef DEBUG
+extern boolean_t skip_sctp_cksum;
+#endif
+
+static int ip_verify_nce(mblk_t *, ip_xmit_attr_t *);
+static int ip_verify_dce(mblk_t *, ip_xmit_attr_t *);
+static boolean_t ip_verify_lso(ill_t *, ip_xmit_attr_t *);
+static boolean_t ip_verify_zcopy(ill_t *, ip_xmit_attr_t *);
+static void ip_output_simple_broadcast(ip_xmit_attr_t *, mblk_t *);
+
+/*
+ * There are two types of output functions for IP used for different
+ * purposes:
+ * - ip_output_simple() is when sending ICMP errors, TCP resets, etc when there
+ * is no context in the form of a conn_t. However, there is a
+ * ip_xmit_attr_t that the callers use to influence interface selection
+ * (needed for ICMP echo as well as IPv6 link-locals) and IPsec.
+ *
+ * - conn_ip_output() is used when sending packets with a conn_t and
+ * ip_set_destination has been called to cache information. In that case
+ * various socket options are recorded in the ip_xmit_attr_t and should
+ * be taken into account.
+ */
+
+/*
+ * The caller *must* have called conn_connect() or ip_attr_connect()
+ * before calling conn_ip_output(). The caller needs to redo that each time
+ * the destination IP address or port changes, as well as each time there is
+ * a change to any socket option that would modify how packets are routed out
+ * of the box (e.g., SO_DONTROUTE, IP_NEXTHOP, IP_BOUND_IF).
+ *
+ * The ULP caller has to serialize the use of a single ip_xmit_attr_t.
+ * We assert for that here.
+ */
+int
+conn_ip_output(mblk_t *mp, ip_xmit_attr_t *ixa)
+{
+ iaflags_t ixaflags = ixa->ixa_flags;
+ ire_t *ire;
+ nce_t *nce;
+ dce_t *dce;
+ ill_t *ill;
+ ip_stack_t *ipst = ixa->ixa_ipst;
+ int error;
+
+ /* We defer ipIfStatsHCOutRequests until an error or we have an ill */
+
+ ASSERT(ixa->ixa_ire != NULL);
+ /* Note there is no ixa_nce when reject and blackhole routes */
+ ASSERT(ixa->ixa_dce != NULL); /* Could be default dce */
+
+#ifdef DEBUG
+ ASSERT(ixa->ixa_curthread == NULL);
+ ixa->ixa_curthread = curthread;
+#endif
+
+ /*
+ * Even on labeled systems we can have a NULL ixa_tsl e.g.,
+ * for IGMP/MLD traffic.
+ */
+
+ ire = ixa->ixa_ire;
+
+ /*
+ * If the ULP says the (old) IRE resulted in reachability we
+ * record this before determine whether to use a new IRE.
+ * No locking for performance reasons.
+ */
+ if (ixaflags & IXAF_REACH_CONF)
+ ire->ire_badcnt = 0;
+
+ /*
+ * Has routing changed since we cached the results of the lookup?
+ *
+ * This check captures all of:
+ * - the cached ire being deleted (by means of the special
+ * IRE_GENERATION_CONDEMNED)
+ * - A potentially better ire being added (ire_generation being
+ * increased)
+ * - A deletion of the nexthop ire that was used when we did the
+ * lookup.
+ * - An addition of a potentially better nexthop ire.
+ * The last two are handled by walking and increasing the generation
+ * number on all dependant IREs in ire_flush_cache().
+ *
+ * The check also handles all cases of RTF_REJECT and RTF_BLACKHOLE
+ * since we ensure that each time we set ixa_ire to such an IRE we
+ * make sure the ixa_ire_generation does not match (by using
+ * IRE_GENERATION_VERIFY).
+ */
+ if (ire->ire_generation != ixa->ixa_ire_generation) {
+ error = ip_verify_ire(mp, ixa);
+ if (error != 0) {
+ ip_drop_output("ipIfStatsOutDiscards - verify ire",
+ mp, NULL);
+ goto drop;
+ }
+ ire = ixa->ixa_ire;
+ ASSERT(ire != NULL);
+ if (ire->ire_flags & (RTF_REJECT|RTF_BLACKHOLE)) {
+#ifdef DEBUG
+ ASSERT(ixa->ixa_curthread == curthread);
+ ixa->ixa_curthread = NULL;
+#endif
+ ire->ire_ob_pkt_count++;
+ /* ixa_dce might be condemned; use default one */
+ return ((ire->ire_sendfn)(ire, mp, mp->b_rptr, ixa,
+ &ipst->ips_dce_default->dce_ident));
+ }
+ /*
+ * If the ncec changed then ip_verify_ire already set
+ * ixa->ixa_dce_generation = DCE_GENERATION_VERIFY;
+ * so we can recheck the interface mtu.
+ */
+
+ /*
+ * Note that ire->ire_generation could already have changed.
+ * We catch that next time we send a packet.
+ */
+ }
+
+ /*
+ * No need to lock access to ixa_nce since the ip_xmit_attr usage
+ * is single threaded.
+ */
+ ASSERT(ixa->ixa_nce != NULL);
+ nce = ixa->ixa_nce;
+ if (nce->nce_is_condemned) {
+ error = ip_verify_nce(mp, ixa);
+ /*
+ * In case ZEROCOPY capability become not available, we
+ * copy the message and free the original one. We might
+ * be copying more data than needed but it doesn't hurt
+ * since such change rarely happens.
+ */
+ switch (error) {
+ case 0:
+ break;
+ case ENOTSUP: { /* ZEROCOPY */
+ mblk_t *nmp;
+
+ if ((nmp = copymsg(mp)) != NULL) {
+ freemsg(mp);
+ mp = nmp;
+
+ break;
+ }
+ /* FALLTHROUGH */
+ }
+ default:
+ ip_drop_output("ipIfStatsOutDiscards - verify nce",
+ mp, NULL);
+ goto drop;
+ }
+ ire = ixa->ixa_ire;
+ ASSERT(ire != NULL);
+ if (ire->ire_flags & (RTF_REJECT|RTF_BLACKHOLE)) {
+#ifdef DEBUG
+ ASSERT(ixa->ixa_curthread == curthread);
+ ixa->ixa_curthread = NULL;
+#endif
+ ire->ire_ob_pkt_count++;
+ /* ixa_dce might be condemned; use default one */
+ return ((ire->ire_sendfn)(ire, mp, mp->b_rptr,
+ ixa, &ipst->ips_dce_default->dce_ident));
+ }
+ ASSERT(ixa->ixa_nce != NULL);
+ nce = ixa->ixa_nce;
+
+ /*
+ * Note that some other event could already have made
+ * the new nce condemned. We catch that next time we
+ * try to send a packet.
+ */
+ }
+ /*
+ * If there is no per-destination dce_t then we have a reference to
+ * the default dce_t (which merely contains the dce_ipid).
+ * The generation check captures both the introduction of a
+ * per-destination dce_t (e.g., due to ICMP packet too big) and
+ * any change to the per-destination dce (including it becoming
+ * condemned by use of the special DCE_GENERATION_CONDEMNED).
+ */
+ dce = ixa->ixa_dce;
+
+ /*
+ * To avoid a periodic timer to increase the path MTU we
+ * look at dce_last_change_time each time we send a packet.
+ */
+ if ((dce->dce_flags & DCEF_PMTU) &&
+ (TICK_TO_SEC(lbolt64) - dce->dce_last_change_time >
+ ipst->ips_ip_pathmtu_interval)) {
+ /*
+ * Older than 20 minutes. Drop the path MTU information.
+ * Since the path MTU changes as a result of this, twiddle
+ * ixa_dce_generation to make us go through the dce
+ * verification code in conn_ip_output.
+ */
+ mutex_enter(&dce->dce_lock);
+ dce->dce_flags &= ~(DCEF_PMTU|DCEF_TOO_SMALL_PMTU);
+ dce->dce_last_change_time = TICK_TO_SEC(lbolt64);
+ mutex_exit(&dce->dce_lock);
+ dce_increment_generation(dce);
+ }
+
+ if (dce->dce_generation != ixa->ixa_dce_generation) {
+ error = ip_verify_dce(mp, ixa);
+ if (error != 0) {
+ ip_drop_output("ipIfStatsOutDiscards - verify dce",
+ mp, NULL);
+ goto drop;
+ }
+ dce = ixa->ixa_dce;
+
+ /*
+ * Note that some other event could already have made the
+ * new dce's generation number change.
+ * We catch that next time we try to send a packet.
+ */
+ }
+
+ ill = nce->nce_ill;
+
+ /*
+ * An initial ixa_fragsize was set in ip_set_destination
+ * and we update it if any routing changes above.
+ * A change to ill_mtu with ifconfig will increase all dce_generation
+ * so that we will detect that with the generation check.
+ */
+
+ /*
+ * Caller needs to make sure IXAF_VERIFY_SRC is not set if
+ * conn_unspec_src.
+ */
+ if ((ixaflags & IXAF_VERIFY_SOURCE) &&
+ ixa->ixa_src_generation != ipst->ips_src_generation) {
+ /* Check if the IP source is still assigned to the host. */
+ uint_t gen;
+
+ if (!ip_verify_src(mp, ixa, &gen)) {
+ /* Don't send a packet with a source that isn't ours */
+ error = EADDRNOTAVAIL;
+ ip_drop_output("ipIfStatsOutDiscards - invalid src",
+ mp, NULL);
+ goto drop;
+ }
+ /* The source is still valid - update the generation number */
+ ixa->ixa_src_generation = gen;
+ }
+
+ /*
+ * We don't have an IRE when we fragment, hence ire_ob_pkt_count
+ * can only count the use prior to fragmentation. However the MIB
+ * counters on the ill will be incremented in post fragmentation.
+ */
+ ire->ire_ob_pkt_count++;
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCOutRequests);
+
+ /*
+ * Based on ire_type and ire_flags call one of:
+ * ire_send_local_v* - for IRE_LOCAL and IRE_LOOPBACK
+ * ire_send_multirt_v* - if RTF_MULTIRT
+ * ire_send_noroute_v* - if RTF_REJECT or RTF_BLACHOLE
+ * ire_send_multicast_v* - for IRE_MULTICAST
+ * ire_send_broadcast_v4 - for IRE_BROADCAST
+ * ire_send_wire_v* - for the rest.
+ */
+#ifdef DEBUG
+ ASSERT(ixa->ixa_curthread == curthread);
+ ixa->ixa_curthread = NULL;
+#endif
+ return ((ire->ire_sendfn)(ire, mp, mp->b_rptr, ixa, &dce->dce_ident));
+
+drop:
+ if (ixaflags & IXAF_IS_IPV4) {
+ BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsHCOutRequests);
+ BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsOutDiscards);
+ } else {
+ BUMP_MIB(&ipst->ips_ip6_mib, ipIfStatsHCOutRequests);
+ BUMP_MIB(&ipst->ips_ip6_mib, ipIfStatsOutDiscards);
+ }
+ freemsg(mp);
+#ifdef DEBUG
+ ASSERT(ixa->ixa_curthread == curthread);
+ ixa->ixa_curthread = NULL;
+#endif
+ return (error);
+}
+
+/*
+ * Handle both IPv4 and IPv6. Sets the generation number
+ * to allow the caller to know when to call us again.
+ * Returns true if the source address in the packet is a valid source.
+ * We handle callers which try to send with a zero address (since we only
+ * get here if UNSPEC_SRC is not set).
+ */
+boolean_t
+ip_verify_src(mblk_t *mp, ip_xmit_attr_t *ixa, uint_t *generationp)
+{
+ ip_stack_t *ipst = ixa->ixa_ipst;
+
+ /*
+ * Need to grab the generation number before we check to
+ * avoid a race with a change to the set of local addresses.
+ * No lock needed since the thread which updates the set of local
+ * addresses use ipif/ill locks and exit those (hence a store memory
+ * barrier) before doing the atomic increase of ips_src_generation.
+ */
+ if (generationp != NULL)
+ *generationp = ipst->ips_src_generation;
+
+ if (ixa->ixa_flags & IXAF_IS_IPV4) {
+ ipha_t *ipha = (ipha_t *)mp->b_rptr;
+
+ if (ipha->ipha_src == INADDR_ANY)
+ return (B_FALSE);
+
+ return (ip_laddr_verify_v4(ipha->ipha_src, ixa->ixa_zoneid,
+ ipst, B_FALSE) != IPVL_BAD);
+ } else {
+ ip6_t *ip6h = (ip6_t *)mp->b_rptr;
+ uint_t scopeid;
+
+ if (IN6_IS_ADDR_UNSPECIFIED(&ip6h->ip6_src))
+ return (B_FALSE);
+
+ if (ixa->ixa_flags & IXAF_SCOPEID_SET)
+ scopeid = ixa->ixa_scopeid;
+ else
+ scopeid = 0;
+
+ return (ip_laddr_verify_v6(&ip6h->ip6_src, ixa->ixa_zoneid,
+ ipst, B_FALSE, scopeid) != IPVL_BAD);
+ }
+}
+
+/*
+ * Handle both IPv4 and IPv6. Reverify/recalculate the IRE to use.
+ */
+int
+ip_verify_ire(mblk_t *mp, ip_xmit_attr_t *ixa)
+{
+ uint_t gen;
+ ire_t *ire;
+ nce_t *nce;
+ int error;
+ boolean_t multirt = B_FALSE;
+
+ /*
+ * Redo ip_select_route.
+ * Need to grab generation number as part of the lookup to
+ * avoid race.
+ */
+ error = 0;
+ ire = ip_select_route_pkt(mp, ixa, &gen, &error, &multirt);
+ ASSERT(ire != NULL); /* IRE_NOROUTE if none found */
+ if (error != 0) {
+ ire_refrele(ire);
+ return (error);
+ }
+
+ if (ixa->ixa_ire != NULL)
+ ire_refrele_notr(ixa->ixa_ire);
+#ifdef DEBUG
+ ire_refhold_notr(ire);
+ ire_refrele(ire);
+#endif
+ ixa->ixa_ire = ire;
+ ixa->ixa_ire_generation = gen;
+ if (multirt) {
+ if (ixa->ixa_flags & IXAF_IS_IPV4)
+ ixa->ixa_postfragfn = ip_postfrag_multirt_v4;
+ else
+ ixa->ixa_postfragfn = ip_postfrag_multirt_v6;
+ ixa->ixa_flags |= IXAF_MULTIRT_MULTICAST;
+ } else {
+ ixa->ixa_postfragfn = ire->ire_postfragfn;
+ ixa->ixa_flags &= ~IXAF_MULTIRT_MULTICAST;
+ }
+
+ /*
+ * Don't look for an nce for reject or blackhole.
+ * They have ire_generation set to IRE_GENERATION_VERIFY which
+ * makes conn_ip_output avoid references to ixa_nce.
+ */
+ if (ire->ire_flags & (RTF_REJECT|RTF_BLACKHOLE)) {
+ ASSERT(ixa->ixa_ire_generation == IRE_GENERATION_VERIFY);
+ ixa->ixa_dce_generation = DCE_GENERATION_VERIFY;
+ return (0);
+ }
+
+ /* The NCE could now be different */
+ nce = ire_to_nce_pkt(ire, mp);
+ if (nce == NULL) {
+ /*
+ * Allocation failure. Make sure we redo ire/nce selection
+ * next time we send.
+ */
+ ixa->ixa_ire_generation = IRE_GENERATION_VERIFY;
+ ixa->ixa_dce_generation = DCE_GENERATION_VERIFY;
+ return (ENOBUFS);
+ }
+ if (nce == ixa->ixa_nce) {
+ /* No change */
+ nce_refrele(nce);
+ return (0);
+ }
+
+ /*
+ * Since the path MTU might change as a result of this
+ * route change, we twiddle ixa_dce_generation to
+ * make conn_ip_output go through the ip_verify_dce code.
+ */
+ ixa->ixa_dce_generation = DCE_GENERATION_VERIFY;
+
+ if (ixa->ixa_nce != NULL)
+ nce_refrele(ixa->ixa_nce);
+ ixa->ixa_nce = nce;
+ return (0);
+}
+
+/*
+ * Handle both IPv4 and IPv6. Reverify/recalculate the NCE to use.
+ */
+static int
+ip_verify_nce(mblk_t *mp, ip_xmit_attr_t *ixa)
+{
+ ire_t *ire = ixa->ixa_ire;
+ nce_t *nce;
+ int error = 0;
+ ipha_t *ipha = NULL;
+ ip6_t *ip6h = NULL;
+
+ if (ire->ire_ipversion == IPV4_VERSION)
+ ipha = (ipha_t *)mp->b_rptr;
+ else
+ ip6h = (ip6_t *)mp->b_rptr;
+
+ nce = ire_handle_condemned_nce(ixa->ixa_nce, ire, ipha, ip6h, B_TRUE);
+ if (nce == NULL) {
+ /* Try to find a better ire */
+ return (ip_verify_ire(mp, ixa));
+ }
+
+ /*
+ * The hardware offloading capabilities, for example LSO, of the
+ * interface might have changed, so do sanity verification here.
+ */
+ if (ixa->ixa_flags & IXAF_VERIFY_LSO) {
+ if (!ip_verify_lso(nce->nce_ill, ixa)) {
+ ASSERT(ixa->ixa_notify != NULL);
+ ixa->ixa_notify(ixa->ixa_notify_cookie, ixa,
+ IXAN_LSO, 0);
+ error = ENOTSUP;
+ }
+ }
+
+ /*
+ * Verify ZEROCOPY capability of underlying ill. Notify the ULP with
+ * any ZEROCOPY changes. In case ZEROCOPY capability is not available
+ * any more, return error so that conn_ip_output() can take care of
+ * the ZEROCOPY message properly. It's safe to continue send the
+ * message when ZEROCOPY newly become available.
+ */
+ if (ixa->ixa_flags & IXAF_VERIFY_ZCOPY) {
+ if (!ip_verify_zcopy(nce->nce_ill, ixa)) {
+ ASSERT(ixa->ixa_notify != NULL);
+ ixa->ixa_notify(ixa->ixa_notify_cookie, ixa,
+ IXAN_ZCOPY, 0);
+ if ((ixa->ixa_flags & IXAF_ZCOPY_CAPAB) == 0)
+ error = ENOTSUP;
+ }
+ }
+
+ /*
+ * Since the path MTU might change as a result of this
+ * change, we twiddle ixa_dce_generation to
+ * make conn_ip_output go through the ip_verify_dce code.
+ */
+ ixa->ixa_dce_generation = DCE_GENERATION_VERIFY;
+
+ nce_refrele(ixa->ixa_nce);
+ ixa->ixa_nce = nce;
+ return (error);
+}
+
+/*
+ * Handle both IPv4 and IPv6. Reverify/recalculate the DCE to use.
+ */
+static int
+ip_verify_dce(mblk_t *mp, ip_xmit_attr_t *ixa)
+{
+ dce_t *dce;
+ uint_t gen;
+ uint_t pmtu;
+
+ dce = dce_lookup_pkt(mp, ixa, &gen);
+ ASSERT(dce != NULL);
+
+ dce_refrele_notr(ixa->ixa_dce);
+#ifdef DEBUG
+ dce_refhold_notr(dce);
+ dce_refrele(dce);
+#endif
+ ixa->ixa_dce = dce;
+ ixa->ixa_dce_generation = gen;
+
+ /* Extract the (path) mtu from the dce, ncec_ill etc */
+ pmtu = ip_get_pmtu(ixa);
+
+ /*
+ * Tell ULP about PMTU changes - increase or decrease - by returning
+ * an error if IXAF_VERIFY_PMTU is set. In such case, ULP should update
+ * both ixa_pmtu and ixa_fragsize appropriately.
+ *
+ * If ULP doesn't set that flag then we need to update ixa_fragsize
+ * since routing could have changed the ill after after ixa_fragsize
+ * was set previously in the conn_ip_output path or in
+ * ip_set_destination.
+ *
+ * In case of LSO, ixa_fragsize might be greater than ixa_pmtu.
+ *
+ * In the case of a path MTU increase we send the packet after the
+ * notify to the ULP.
+ */
+ if (ixa->ixa_flags & IXAF_VERIFY_PMTU) {
+ if (ixa->ixa_pmtu != pmtu) {
+ uint_t oldmtu = ixa->ixa_pmtu;
+
+ DTRACE_PROBE2(verify_pmtu, uint32_t, pmtu,
+ uint32_t, ixa->ixa_pmtu);
+ ASSERT(ixa->ixa_notify != NULL);
+ ixa->ixa_notify(ixa->ixa_notify_cookie, ixa,
+ IXAN_PMTU, pmtu);
+ if (pmtu < oldmtu)
+ return (EMSGSIZE);
+ }
+ } else {
+ ixa->ixa_fragsize = pmtu;
+ }
+ return (0);
+}
+
+/*
+ * Verify LSO usability. Keep the return value simple to indicate whether
+ * the LSO capability has changed. Handle both IPv4 and IPv6.
+ */
+static boolean_t
+ip_verify_lso(ill_t *ill, ip_xmit_attr_t *ixa)
+{
+ ill_lso_capab_t *lsoc = &ixa->ixa_lso_capab;
+ ill_lso_capab_t *new_lsoc = ill->ill_lso_capab;
+
+ if (ixa->ixa_flags & IXAF_LSO_CAPAB) {
+ /*
+ * Not unsable any more.
+ */
+ if ((ixa->ixa_flags & IXAF_IPSEC_SECURE) ||
+ (ixa->ixa_ire->ire_type & (IRE_LOCAL | IRE_LOOPBACK)) ||
+ (ixa->ixa_ire->ire_flags & RTF_MULTIRT) ||
+ ((ixa->ixa_flags & IXAF_IS_IPV4) ?
+ !ILL_LSO_TCP_IPV4_USABLE(ill) :
+ !ILL_LSO_TCP_IPV6_USABLE(ill))) {
+ ixa->ixa_flags &= ~IXAF_LSO_CAPAB;
+
+ return (B_FALSE);
+ }
+
+ /*
+ * Capability has changed, refresh the copy in ixa.
+ */
+ if (lsoc->ill_lso_max != new_lsoc->ill_lso_max) {
+ *lsoc = *new_lsoc;
+
+ return (B_FALSE);
+ }
+ } else { /* Was not usable */
+ if (!(ixa->ixa_flags & IXAF_IPSEC_SECURE) &&
+ !(ixa->ixa_ire->ire_type & (IRE_LOCAL | IRE_LOOPBACK)) &&
+ !(ixa->ixa_ire->ire_flags & RTF_MULTIRT) &&
+ ((ixa->ixa_flags & IXAF_IS_IPV4) ?
+ ILL_LSO_TCP_IPV4_USABLE(ill) :
+ ILL_LSO_TCP_IPV6_USABLE(ill))) {
+ *lsoc = *new_lsoc;
+ ixa->ixa_flags |= IXAF_LSO_CAPAB;
+
+ return (B_FALSE);
+ }
+ }
+
+ return (B_TRUE);
+}
+
+/*
+ * Verify ZEROCOPY usability. Keep the return value simple to indicate whether
+ * the ZEROCOPY capability has changed. Handle both IPv4 and IPv6.
+ */
+static boolean_t
+ip_verify_zcopy(ill_t *ill, ip_xmit_attr_t *ixa)
+{
+ if (ixa->ixa_flags & IXAF_ZCOPY_CAPAB) {
+ /*
+ * Not unsable any more.
+ */
+ if ((ixa->ixa_flags & IXAF_IPSEC_SECURE) ||
+ (ixa->ixa_ire->ire_type & (IRE_LOCAL | IRE_LOOPBACK)) ||
+ (ixa->ixa_ire->ire_flags & RTF_MULTIRT) ||
+ !ILL_ZCOPY_USABLE(ill)) {
+ ixa->ixa_flags &= ~IXAF_ZCOPY_CAPAB;
+
+ return (B_FALSE);
+ }
+ } else { /* Was not usable */
+ if (!(ixa->ixa_flags & IXAF_IPSEC_SECURE) &&
+ !(ixa->ixa_ire->ire_type & (IRE_LOCAL | IRE_LOOPBACK)) &&
+ !(ixa->ixa_ire->ire_flags & RTF_MULTIRT) &&
+ ILL_ZCOPY_USABLE(ill)) {
+ ixa->ixa_flags |= IXAF_ZCOPY_CAPAB;
+
+ return (B_FALSE);
+ }
+ }
+
+ return (B_TRUE);
+}
+
+
+/*
+ * When there is no conn_t context, this will send a packet.
+ * The caller must *not* have called conn_connect() or ip_attr_connect()
+ * before calling ip_output_simple().
+ * Handles IPv4 and IPv6. Returns zero or an errno such as ENETUNREACH.
+ * Honors IXAF_SET_SOURCE.
+ *
+ * We acquire the ire and after calling ire_sendfn we release
+ * the hold on the ire. Ditto for the nce and dce.
+ *
+ * This assumes that the caller has set the following in ip_xmit_attr_t:
+ * ixa_tsl, ixa_zoneid, and ixa_ipst must always be set.
+ * If ixa_ifindex is non-zero it means send out that ill. (If it is
+ * an upper IPMP ill we load balance across the group; if a lower we send
+ * on that lower ill without load balancing.)
+ * IXAF_IS_IPV4 must be set correctly.
+ * If IXAF_IPSEC_SECURE is set then the ixa_ipsec_* fields must be set.
+ * If IXAF_NO_IPSEC is set we'd skip IPsec policy lookup.
+ * If neither of those two are set we do an IPsec policy lookup.
+ *
+ * We handle setting things like
+ * ixa_pktlen
+ * ixa_ip_hdr_length
+ * ixa->ixa_protocol
+ *
+ * The caller may set ixa_xmit_hint, which is used for ECMP selection and
+ * transmit ring selecting in GLD.
+ *
+ * The caller must do an ixa_cleanup() to release any IPsec references
+ * after we return.
+ */
+int
+ip_output_simple(mblk_t *mp, ip_xmit_attr_t *ixa)
+{
+ ts_label_t *effective_tsl = NULL;
+ int err;
+
+ ASSERT(ixa->ixa_ipst != NULL);
+
+ if (is_system_labeled()) {
+ ip_stack_t *ipst = ixa->ixa_ipst;
+
+ if (ixa->ixa_flags & IXAF_IS_IPV4) {
+ err = tsol_check_label_v4(ixa->ixa_tsl, ixa->ixa_zoneid,
+ &mp, CONN_MAC_DEFAULT, B_FALSE, ixa->ixa_ipst,
+ &effective_tsl);
+ } else {
+ err = tsol_check_label_v6(ixa->ixa_tsl, ixa->ixa_zoneid,
+ &mp, CONN_MAC_DEFAULT, B_FALSE, ixa->ixa_ipst,
+ &effective_tsl);
+ }
+ if (err != 0) {
+ ip2dbg(("tsol_check: label check failed (%d)\n", err));
+ BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsHCOutRequests);
+ BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsOutDiscards);
+ ip_drop_output("tsol_check_label", mp, NULL);
+ freemsg(mp);
+ return (err);
+ }
+ if (effective_tsl != NULL) {
+ /* Update the label */
+ ip_xmit_attr_replace_tsl(ixa, effective_tsl);
+ }
+ }
+
+ if (ixa->ixa_flags & IXAF_IS_IPV4)
+ return (ip_output_simple_v4(mp, ixa));
+ else
+ return (ip_output_simple_v6(mp, ixa));
+}
+
+int
+ip_output_simple_v4(mblk_t *mp, ip_xmit_attr_t *ixa)
+{
+ ipha_t *ipha;
+ ipaddr_t firsthop; /* In IP header */
+ ipaddr_t dst; /* End of source route, or ipha_dst if none */
+ ire_t *ire;
+ ipaddr_t setsrc; /* RTF_SETSRC */
+ int error;
+ ill_t *ill = NULL;
+ dce_t *dce = NULL;
+ nce_t *nce;
+ iaflags_t ixaflags = ixa->ixa_flags;
+ ip_stack_t *ipst = ixa->ixa_ipst;
+ boolean_t repeat = B_FALSE;
+ boolean_t multirt = B_FALSE;
+
+ ipha = (ipha_t *)mp->b_rptr;
+ ASSERT(IPH_HDR_VERSION(ipha) == IPV4_VERSION);
+
+ /*
+ * Even on labeled systems we can have a NULL ixa_tsl e.g.,
+ * for IGMP/MLD traffic.
+ */
+
+ /* Caller already set flags */
+ ASSERT(ixa->ixa_flags & IXAF_IS_IPV4);
+
+ ASSERT(ixa->ixa_nce == NULL);
+
+ ixa->ixa_pktlen = ntohs(ipha->ipha_length);
+ ASSERT(ixa->ixa_pktlen == msgdsize(mp));
+ ixa->ixa_ip_hdr_length = IPH_HDR_LENGTH(ipha);
+ ixa->ixa_protocol = ipha->ipha_protocol;
+
+ /*
+ * Assumes that source routed packets have already been massaged by
+ * the ULP (ip_massage_options) and as a result ipha_dst is the next
+ * hop in the source route. The final destination is used for IPsec
+ * policy and DCE lookup.
+ */
+ firsthop = ipha->ipha_dst;
+ dst = ip_get_dst(ipha);
+
+repeat_ire:
+ error = 0;
+ setsrc = INADDR_ANY;
+ ire = ip_select_route_v4(firsthop, ixa, NULL, &setsrc, &error,
+ &multirt);
+ ASSERT(ire != NULL); /* IRE_NOROUTE if none found */
+ if (error != 0) {
+ BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsHCOutRequests);
+ BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards - select route", mp, NULL);
+ freemsg(mp);
+ goto done;
+ }
+
+ if (ire->ire_flags & (RTF_BLACKHOLE|RTF_REJECT)) {
+ /* ire_ill might be NULL hence need to skip some code */
+ if (ixaflags & IXAF_SET_SOURCE)
+ ipha->ipha_src = htonl(INADDR_LOOPBACK);
+ ixa->ixa_fragsize = IP_MAXPACKET;
+ ill = NULL;
+ nce = NULL;
+ ire->ire_ob_pkt_count++;
+ BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsHCOutRequests);
+ /* No dce yet; use default one */
+ error = (ire->ire_sendfn)(ire, mp, ipha, ixa,
+ &ipst->ips_dce_default->dce_ident);
+ goto done;
+ }
+
+ /* Note that ipha_dst is only used for IRE_MULTICAST */
+ nce = ire_to_nce(ire, ipha->ipha_dst, NULL);
+ if (nce == NULL) {
+ /* Allocation failure? */
+ ip_drop_output("ire_to_nce", mp, ill);
+ freemsg(mp);
+ error = ENOBUFS;
+ goto done;
+ }
+ if (nce->nce_is_condemned) {
+ nce_t *nce1;
+
+ nce1 = ire_handle_condemned_nce(nce, ire, ipha, NULL, B_TRUE);
+ nce_refrele(nce);
+ if (nce1 == NULL) {
+ if (!repeat) {
+ /* Try finding a better IRE */
+ repeat = B_TRUE;
+ ire_refrele(ire);
+ goto repeat_ire;
+ }
+ /* Tried twice - drop packet */
+ BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsOutDiscards);
+ ip_drop_output("No nce", mp, ill);
+ freemsg(mp);
+ error = ENOBUFS;
+ goto done;
+ }
+ nce = nce1;
+ }
+
+ /*
+ * For multicast with multirt we have a flag passed back from
+ * ire_lookup_multi_ill_v4 since we don't have an IRE for each
+ * possible multicast address.
+ * We also need a flag for multicast since we can't check
+ * whether RTF_MULTIRT is set in ixa_ire for multicast.
+ */
+ if (multirt) {
+ ixa->ixa_postfragfn = ip_postfrag_multirt_v4;
+ ixa->ixa_flags |= IXAF_MULTIRT_MULTICAST;
+ } else {
+ ixa->ixa_postfragfn = ire->ire_postfragfn;
+ ixa->ixa_flags &= ~IXAF_MULTIRT_MULTICAST;
+ }
+ ASSERT(ixa->ixa_nce == NULL);
+ ixa->ixa_nce = nce;
+
+ /*
+ * Check for a dce_t with a path mtu.
+ */
+ dce = dce_lookup_v4(dst, ipst, NULL);
+ ASSERT(dce != NULL);
+
+ if (!(ixaflags & IXAF_PMTU_DISCOVERY)) {
+ ixa->ixa_fragsize = ip_get_base_mtu(nce->nce_ill, ire);
+ } else if (dce->dce_flags & DCEF_PMTU) {
+ /*
+ * To avoid a periodic timer to increase the path MTU we
+ * look at dce_last_change_time each time we send a packet.
+ */
+ if (TICK_TO_SEC(lbolt64) - dce->dce_last_change_time >
+ ipst->ips_ip_pathmtu_interval) {
+ /*
+ * Older than 20 minutes. Drop the path MTU information.
+ */
+ mutex_enter(&dce->dce_lock);
+ dce->dce_flags &= ~(DCEF_PMTU|DCEF_TOO_SMALL_PMTU);
+ dce->dce_last_change_time = TICK_TO_SEC(lbolt64);
+ mutex_exit(&dce->dce_lock);
+ dce_increment_generation(dce);
+ ixa->ixa_fragsize = ip_get_base_mtu(nce->nce_ill, ire);
+ } else {
+ uint_t fragsize;
+
+ fragsize = ip_get_base_mtu(nce->nce_ill, ire);
+ if (fragsize > dce->dce_pmtu)
+ fragsize = dce->dce_pmtu;
+ ixa->ixa_fragsize = fragsize;
+ }
+ } else {
+ ixa->ixa_fragsize = ip_get_base_mtu(nce->nce_ill, ire);
+ }
+
+ /*
+ * We use use ire_nexthop_ill (and not ncec_ill) to avoid the under ipmp
+ * interface for source address selection.
+ */
+ ill = ire_nexthop_ill(ire);
+
+ if (ixaflags & IXAF_SET_SOURCE) {
+ ipaddr_t src;
+
+ /*
+ * We use the final destination to get
+ * correct selection for source routed packets
+ */
+
+ /* If unreachable we have no ill but need some source */
+ if (ill == NULL) {
+ src = htonl(INADDR_LOOPBACK);
+ error = 0;
+ } else {
+ error = ip_select_source_v4(ill, setsrc, dst,
+ ixa->ixa_multicast_ifaddr, ixa->ixa_zoneid, ipst,
+ &src, NULL, NULL);
+ }
+ if (error != 0) {
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCOutRequests);
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards - no source",
+ mp, ill);
+ freemsg(mp);
+ goto done;
+ }
+ ipha->ipha_src = src;
+ } else if (ixaflags & IXAF_VERIFY_SOURCE) {
+ /* Check if the IP source is assigned to the host. */
+ if (!ip_verify_src(mp, ixa, NULL)) {
+ /* Don't send a packet with a source that isn't ours */
+ BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsHCOutRequests);
+ BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards - invalid source",
+ mp, ill);
+ freemsg(mp);
+ error = EADDRNOTAVAIL;
+ goto done;
+ }
+ }
+
+
+ /*
+ * Check against global IPsec policy to set the AH/ESP attributes.
+ * IPsec will set IXAF_IPSEC_* and ixa_ipsec_* as appropriate.
+ */
+ if (!(ixaflags & (IXAF_NO_IPSEC|IXAF_IPSEC_SECURE))) {
+ ASSERT(ixa->ixa_ipsec_policy == NULL);
+ mp = ip_output_attach_policy(mp, ipha, NULL, NULL, ixa);
+ if (mp == NULL) {
+ /* MIB and ip_drop_packet already done */
+ return (EHOSTUNREACH); /* IPsec policy failure */
+ }
+ }
+
+ if (ill != NULL) {
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCOutRequests);
+ } else {
+ BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsHCOutRequests);
+ }
+
+ /*
+ * We update the statistics on the most specific IRE i.e., the first
+ * one we found.
+ * We don't have an IRE when we fragment, hence ire_ob_pkt_count
+ * can only count the use prior to fragmentation. However the MIB
+ * counters on the ill will be incremented in post fragmentation.
+ */
+ ire->ire_ob_pkt_count++;
+
+ /*
+ * Based on ire_type and ire_flags call one of:
+ * ire_send_local_v4 - for IRE_LOCAL and IRE_LOOPBACK
+ * ire_send_multirt_v4 - if RTF_MULTIRT
+ * ire_send_noroute_v4 - if RTF_REJECT or RTF_BLACHOLE
+ * ire_send_multicast_v4 - for IRE_MULTICAST
+ * ire_send_broadcast_v4 - for IRE_BROADCAST
+ * ire_send_wire_v4 - for the rest.
+ */
+ error = (ire->ire_sendfn)(ire, mp, ipha, ixa, &dce->dce_ident);
+done:
+ ire_refrele(ire);
+ if (dce != NULL)
+ dce_refrele(dce);
+ if (ill != NULL)
+ ill_refrele(ill);
+ if (ixa->ixa_nce != NULL)
+ nce_refrele(ixa->ixa_nce);
+ ixa->ixa_nce = NULL;
+ return (error);
+}
+
+/*
+ * ire_sendfn() functions.
+ * These functions use the following xmit_attr:
+ * - ixa_fragsize - read to determine whether or not to fragment
+ * - IXAF_IPSEC_SECURE - to determine whether or not to invoke IPsec
+ * - ixa_ipsec_* are used inside IPsec
+ * - IXAF_SET_SOURCE - replace IP source in broadcast case.
+ * - IXAF_LOOPBACK_COPY - for multicast and broadcast
+ */
+
+
+/*
+ * ire_sendfn for IRE_LOCAL and IRE_LOOPBACK
+ *
+ * The checks for restrict_interzone_loopback are done in ire_route_recursive.
+ */
+/* ARGSUSED4 */
+int
+ire_send_local_v4(ire_t *ire, mblk_t *mp, void *iph_arg,
+ ip_xmit_attr_t *ixa, uint32_t *identp)
+{
+ ipha_t *ipha = (ipha_t *)iph_arg;
+ ip_stack_t *ipst = ixa->ixa_ipst;
+ ill_t *ill = ire->ire_ill;
+ ip_recv_attr_t iras; /* NOTE: No bzero for performance */
+ uint_t pktlen = ixa->ixa_pktlen;
+
+ /*
+ * No fragmentation, no nce, no application of IPsec,
+ * and no ipha_ident assignment.
+ *
+ * Note different order between IP provider and FW_HOOKS than in
+ * send_wire case.
+ */
+
+ /*
+ * DTrace this as ip:::send. A packet blocked by FW_HOOKS will fire the
+ * send probe, but not the receive probe.
+ */
+ DTRACE_IP7(send, mblk_t *, mp, conn_t *, NULL, void_ip_t *,
+ ipha, __dtrace_ipsr_ill_t *, ill, ipha_t *, ipha, ip6_t *, NULL,
+ int, 1);
+
+ if (HOOKS4_INTERESTED_LOOPBACK_OUT(ipst)) {
+ int error;
+
+ DTRACE_PROBE4(ip4__loopback__out__start, ill_t *, NULL,
+ ill_t *, ill, ipha_t *, ipha, mblk_t *, mp);
+ FW_HOOKS(ipst->ips_ip4_loopback_out_event,
+ ipst->ips_ipv4firewall_loopback_out,
+ NULL, ill, ipha, mp, mp, 0, ipst, error);
+ DTRACE_PROBE1(ip4__loopback__out__end, mblk_t *, mp);
+ if (mp == NULL)
+ return (error);
+
+ /*
+ * Even if the destination was changed by the filter we use the
+ * forwarding decision that was made based on the address
+ * in ip_output/ip_set_destination.
+ */
+ /* Length could be different */
+ ipha = (ipha_t *)mp->b_rptr;
+ pktlen = ntohs(ipha->ipha_length);
+ }
+
+ /*
+ * If a callback is enabled then we need to know the
+ * source and destination zoneids for the packet. We already
+ * have those handy.
+ */
+ if (ipst->ips_ip4_observe.he_interested) {
+ zoneid_t szone, dzone;
+ zoneid_t stackzoneid;
+
+ stackzoneid = netstackid_to_zoneid(
+ ipst->ips_netstack->netstack_stackid);
+
+ if (stackzoneid == GLOBAL_ZONEID) {
+ /* Shared-IP zone */
+ dzone = ire->ire_zoneid;
+ szone = ixa->ixa_zoneid;
+ } else {
+ szone = dzone = stackzoneid;
+ }
+ ipobs_hook(mp, IPOBS_HOOK_LOCAL, szone, dzone, ill, ipst);
+ }
+
+ /* Handle lo0 stats */
+ ipst->ips_loopback_packets++;
+
+ /* Map ixa to ira including IPsec policies */
+ ipsec_out_to_in(ixa, ill, &iras);
+ iras.ira_pktlen = pktlen;
+
+ if (!IS_SIMPLE_IPH(ipha)) {
+ ip_output_local_options(ipha, ipst);
+ iras.ira_flags |= IRAF_IPV4_OPTIONS;
+ }
+
+ if (HOOKS4_INTERESTED_LOOPBACK_IN(ipst)) {
+ int error;
+
+ DTRACE_PROBE4(ip4__loopback__in__start, ill_t *, ill,
+ ill_t *, NULL, ipha_t *, ipha, mblk_t *, mp);
+ FW_HOOKS(ipst->ips_ip4_loopback_in_event,
+ ipst->ips_ipv4firewall_loopback_in,
+ ill, NULL, ipha, mp, mp, 0, ipst, error);
+
+ DTRACE_PROBE1(ip4__loopback__in__end, mblk_t *, mp);
+ if (mp == NULL) {
+ ira_cleanup(&iras, B_FALSE);
+ return (error);
+ }
+ /*
+ * Even if the destination was changed by the filter we use the
+ * forwarding decision that was made based on the address
+ * in ip_output/ip_set_destination.
+ */
+ /* Length could be different */
+ ipha = (ipha_t *)mp->b_rptr;
+ pktlen = iras.ira_pktlen = ntohs(ipha->ipha_length);
+ }
+
+ DTRACE_IP7(receive, mblk_t *, mp, conn_t *, NULL, void_ip_t *,
+ ipha, __dtrace_ipsr_ill_t *, ill, ipha_t *, ipha, ip6_t *, NULL,
+ int, 1);
+
+ ire->ire_ib_pkt_count++;
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInReceives);
+ UPDATE_MIB(ill->ill_ip_mib, ipIfStatsHCInOctets, pktlen);
+
+ /* Destined to ire_zoneid - use that for fanout */
+ iras.ira_zoneid = ire->ire_zoneid;
+
+ if (is_system_labeled()) {
+ iras.ira_flags |= IRAF_SYSTEM_LABELED;
+
+ /*
+ * This updates ira_cred, ira_tsl and ira_free_flags based
+ * on the label. We don't expect this to ever fail for
+ * loopback packets, so we silently drop the packet should it
+ * fail.
+ */
+ if (!tsol_get_pkt_label(mp, IPV4_VERSION, &iras)) {
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
+ ip_drop_input("tsol_get_pkt_label", mp, ill);
+ freemsg(mp);
+ return (0);
+ }
+ ASSERT(iras.ira_tsl != NULL);
+
+ /* tsol_get_pkt_label sometimes does pullupmsg */
+ ipha = (ipha_t *)mp->b_rptr;
+ }
+
+ ip_fanout_v4(mp, ipha, &iras);
+
+ /* We moved any IPsec refs from ixa to iras */
+ ira_cleanup(&iras, B_FALSE);
+ return (0);
+}
+
+/*
+ * ire_sendfn for IRE_BROADCAST
+ * If the broadcast address is present on multiple ills and ixa_ifindex
+ * isn't set, then we generate
+ * a separate datagram (potentially with different source address) for
+ * those ills. In any case, only one copy is looped back to ip_input_v4.
+ */
+int
+ire_send_broadcast_v4(ire_t *ire, mblk_t *mp, void *iph_arg,
+ ip_xmit_attr_t *ixa, uint32_t *identp)
+{
+ ipha_t *ipha = (ipha_t *)iph_arg;
+ ip_stack_t *ipst = ixa->ixa_ipst;
+ irb_t *irb = ire->ire_bucket;
+ ire_t *ire1;
+ mblk_t *mp1;
+ ipha_t *ipha1;
+ iaflags_t ixaflags = ixa->ixa_flags;
+ nce_t *nce1, *nce_orig;
+
+ /*
+ * Unless ire_send_multirt_v4 already set a ttl, force the
+ * ttl to a smallish value.
+ */
+ if (!(ixa->ixa_flags & IXAF_NO_TTL_CHANGE)) {
+ /*
+ * To avoid broadcast storms, we usually set the TTL to 1 for
+ * broadcasts. This can
+ * be overridden stack-wide through the ip_broadcast_ttl
+ * ndd tunable, or on a per-connection basis through the
+ * IP_BROADCAST_TTL socket option.
+ *
+ * If SO_DONTROUTE/IXAF_DONTROUTE is set, then ire_send_wire_v4
+ * will force ttl to one after we've set this.
+ */
+ if (ixaflags & IXAF_BROADCAST_TTL_SET)
+ ipha->ipha_ttl = ixa->ixa_broadcast_ttl;
+ else
+ ipha->ipha_ttl = ipst->ips_ip_broadcast_ttl;
+ }
+ /*
+ * Make sure we get a loopback copy (after IPsec and frag)
+ * Skip hardware checksum so that loopback copy is checksumed.
+ */
+ ixa->ixa_flags |= IXAF_LOOPBACK_COPY | IXAF_NO_HW_CKSUM;
+
+ /* Do we need to potentially generate multiple copies? */
+ if (irb->irb_ire_cnt == 1 || ixa->ixa_ifindex != 0)
+ return (ire_send_wire_v4(ire, mp, ipha, ixa, identp));
+
+ /*
+ * Loop over all IRE_BROADCAST in the bucket (might only be one).
+ * Note that everything in the bucket has the same destination address.
+ */
+ irb_refhold(irb);
+ for (ire1 = irb->irb_ire; ire1 != NULL; ire1 = ire1->ire_next) {
+ /* We do the main IRE after the end of the loop */
+ if (ire1 == ire)
+ continue;
+
+ /*
+ * Only IREs for the same IP address should be in the same
+ * bucket.
+ * But could have IRE_HOSTs in the case of CGTP.
+ * If we find any multirt routes we bail out of the loop
+ * and just do the single packet at the end; ip_postfrag_multirt
+ * will duplicate the packet.
+ */
+ ASSERT(ire1->ire_addr == ire->ire_addr);
+ if (!(ire1->ire_type & IRE_BROADCAST))
+ continue;
+
+ if (IRE_IS_CONDEMNED(ire1))
+ continue;
+
+ if (ixa->ixa_zoneid != ALL_ZONES &&
+ ire->ire_zoneid != ire1->ire_zoneid)
+ continue;
+
+ ASSERT(ire->ire_ill != ire1->ire_ill && ire1->ire_ill != NULL);
+
+ if (ire1->ire_flags & RTF_MULTIRT)
+ break;
+
+ /*
+ * For IPMP we only send for the ipmp_ill. arp_nce_init() will
+ * ensure that this goes out on the cast_ill.
+ */
+ if (IS_UNDER_IPMP(ire1->ire_ill))
+ continue;
+
+ mp1 = copymsg(mp);
+ if (mp1 == NULL) {
+ BUMP_MIB(ire1->ire_ill->ill_ip_mib,
+ ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards",
+ mp, ire1->ire_ill);
+ continue;
+ }
+
+ ipha1 = (ipha_t *)mp1->b_rptr;
+ if (ixa->ixa_flags & IXAF_SET_SOURCE) {
+ /*
+ * Need to pick a different source address for each
+ * interface. If we have a global IPsec policy and
+ * no per-socket policy then we punt to
+ * ip_output_simple_v4 using a separate ip_xmit_attr_t.
+ */
+ if (ixaflags & IXAF_IPSEC_GLOBAL_POLICY) {
+ ip_output_simple_broadcast(ixa, mp1);
+ continue;
+ }
+ /* Pick a new source address for each interface */
+ if (ip_select_source_v4(ire1->ire_ill, INADDR_ANY,
+ ipha1->ipha_dst, INADDR_ANY, ixa->ixa_zoneid, ipst,
+ &ipha1->ipha_src, NULL, NULL) != 0) {
+ BUMP_MIB(ire1->ire_ill->ill_ip_mib,
+ ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards - select "
+ "broadcast source", mp1, ire1->ire_ill);
+ freemsg(mp1);
+ continue;
+ }
+ /*
+ * Check against global IPsec policy to set the AH/ESP
+ * attributes. IPsec will set IXAF_IPSEC_* and
+ * ixa_ipsec_* as appropriate.
+ */
+ if (!(ixaflags & (IXAF_NO_IPSEC|IXAF_IPSEC_SECURE))) {
+ ASSERT(ixa->ixa_ipsec_policy == NULL);
+ mp1 = ip_output_attach_policy(mp1, ipha, NULL,
+ NULL, ixa);
+ if (mp1 == NULL) {
+ /*
+ * MIB and ip_drop_packet already
+ * done
+ */
+ continue;
+ }
+ }
+ }
+ /* Make sure we have an NCE on this ill */
+ nce1 = arp_nce_init(ire1->ire_ill, ire1->ire_addr,
+ ire1->ire_type);
+ if (nce1 == NULL) {
+ BUMP_MIB(ire1->ire_ill->ill_ip_mib,
+ ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards - broadcast nce",
+ mp1, ire1->ire_ill);
+ freemsg(mp1);
+ continue;
+ }
+ nce_orig = ixa->ixa_nce;
+ ixa->ixa_nce = nce1;
+
+ ire_refhold(ire1);
+ /*
+ * Ignore any errors here. We just collect the errno for
+ * the main ire below
+ */
+ (void) ire_send_wire_v4(ire1, mp1, ipha1, ixa, identp);
+ ire_refrele(ire1);
+
+ ixa->ixa_nce = nce_orig;
+ nce_refrele(nce1);
+
+ ixa->ixa_flags &= ~IXAF_LOOPBACK_COPY;
+ }
+ irb_refrele(irb);
+ /* Finally, the main one */
+
+ /*
+ * For IPMP we only send broadcasts on the ipmp_ill.
+ */
+ if (IS_UNDER_IPMP(ire->ire_ill)) {
+ freemsg(mp);
+ return (0);
+ }
+
+ return (ire_send_wire_v4(ire, mp, ipha, ixa, identp));
+}
+
+/*
+ * Send a packet using a different source address and different
+ * IPsec policy.
+ */
+static void
+ip_output_simple_broadcast(ip_xmit_attr_t *ixa, mblk_t *mp)
+{
+ ip_xmit_attr_t ixas;
+
+ bzero(&ixas, sizeof (ixas));
+ ixas.ixa_flags = IXAF_BASIC_SIMPLE_V4;
+ ixas.ixa_zoneid = ixa->ixa_zoneid;
+ ixas.ixa_ifindex = 0;
+ ixas.ixa_ipst = ixa->ixa_ipst;
+ ixas.ixa_cred = ixa->ixa_cred;
+ ixas.ixa_cpid = ixa->ixa_cpid;
+ ixas.ixa_tsl = ixa->ixa_tsl;
+ ixas.ixa_multicast_ttl = IP_DEFAULT_MULTICAST_TTL;
+
+ (void) ip_output_simple(mp, &ixas);
+ ixa_cleanup(&ixas);
+}
+
+
+static void
+multirt_check_v4(ire_t *ire, ipha_t *ipha, ip_xmit_attr_t *ixa)
+{
+ ip_stack_t *ipst = ixa->ixa_ipst;
+
+ /* Limit the TTL on multirt packets */
+ if (ire->ire_type & IRE_MULTICAST) {
+ if (ipha->ipha_ttl > 1) {
+ ip2dbg(("ire_send_multirt_v4: forcing multicast "
+ "multirt TTL to 1 (was %d), dst 0x%08x\n",
+ ipha->ipha_ttl, ntohl(ire->ire_addr)));
+ ipha->ipha_ttl = 1;
+ }
+ ixa->ixa_flags |= IXAF_NO_TTL_CHANGE;
+ } else if ((ipst->ips_ip_multirt_ttl > 0) &&
+ (ipha->ipha_ttl > ipst->ips_ip_multirt_ttl)) {
+ ipha->ipha_ttl = ipst->ips_ip_multirt_ttl;
+ /*
+ * Need to ensure we don't increase the ttl should we go through
+ * ire_send_broadcast or multicast.
+ */
+ ixa->ixa_flags |= IXAF_NO_TTL_CHANGE;
+ }
+}
+
+/*
+ * ire_sendfn for IRE_MULTICAST
+ */
+int
+ire_send_multicast_v4(ire_t *ire, mblk_t *mp, void *iph_arg,
+ ip_xmit_attr_t *ixa, uint32_t *identp)
+{
+ ipha_t *ipha = (ipha_t *)iph_arg;
+ ip_stack_t *ipst = ixa->ixa_ipst;
+ ill_t *ill = ire->ire_ill;
+ iaflags_t ixaflags = ixa->ixa_flags;
+
+ /*
+ * The IRE_MULTICAST is the same whether or not multirt is in use.
+ * Hence we need special-case code.
+ */
+ if (ixaflags & IXAF_MULTIRT_MULTICAST)
+ multirt_check_v4(ire, ipha, ixa);
+
+ /*
+ * Check if anything in ip_input_v4 wants a copy of the transmitted
+ * packet (after IPsec and fragmentation)
+ *
+ * 1. Multicast routers always need a copy unless SO_DONTROUTE is set
+ * RSVP and the rsvp daemon is an example of a
+ * protocol and user level process that
+ * handles it's own routing. Hence, it uses the
+ * SO_DONTROUTE option to accomplish this.
+ * 2. If the sender has set IP_MULTICAST_LOOP, then we just
+ * check whether there are any receivers for the group on the ill
+ * (ignoring the zoneid).
+ * 3. If IP_MULTICAST_LOOP is not set, then we check if there are
+ * any members in other shared-IP zones.
+ * If such members exist, then we indicate that the sending zone
+ * shouldn't get a loopback copy to preserve the IP_MULTICAST_LOOP
+ * behavior.
+ *
+ * When we loopback we skip hardware checksum to make sure loopback
+ * copy is checksumed.
+ *
+ * Note that ire_ill is the upper in the case of IPMP.
+ */
+ ixa->ixa_flags &= ~(IXAF_LOOPBACK_COPY | IXAF_NO_HW_CKSUM);
+ if (ipst->ips_ip_g_mrouter && ill->ill_mrouter_cnt > 0 &&
+ !(ixaflags & IXAF_DONTROUTE)) {
+ ixa->ixa_flags |= IXAF_LOOPBACK_COPY | IXAF_NO_HW_CKSUM;
+ } else if (ixaflags & IXAF_MULTICAST_LOOP) {
+ /*
+ * If this zone or any other zone has members then loopback
+ * a copy.
+ */
+ if (ill_hasmembers_v4(ill, ipha->ipha_dst))
+ ixa->ixa_flags |= IXAF_LOOPBACK_COPY | IXAF_NO_HW_CKSUM;
+ } else if (ipst->ips_netstack->netstack_numzones > 1) {
+ /*
+ * This zone should not have a copy. But there are some other
+ * zones which might have members.
+ */
+ if (ill_hasmembers_otherzones_v4(ill, ipha->ipha_dst,
+ ixa->ixa_zoneid)) {
+ ixa->ixa_flags |= IXAF_NO_LOOP_ZONEID_SET;
+ ixa->ixa_no_loop_zoneid = ixa->ixa_zoneid;
+ ixa->ixa_flags |= IXAF_LOOPBACK_COPY | IXAF_NO_HW_CKSUM;
+ }
+ }
+
+ /*
+ * Unless ire_send_multirt_v4 or icmp_output_hdrincl already set a ttl,
+ * force the ttl to the IP_MULTICAST_TTL value
+ */
+ if (!(ixaflags & IXAF_NO_TTL_CHANGE)) {
+ ipha->ipha_ttl = ixa->ixa_multicast_ttl;
+ }
+
+ return (ire_send_wire_v4(ire, mp, ipha, ixa, identp));
+}
+
+/*
+ * ire_sendfn for IREs with RTF_MULTIRT
+ */
+int
+ire_send_multirt_v4(ire_t *ire, mblk_t *mp, void *iph_arg,
+ ip_xmit_attr_t *ixa, uint32_t *identp)
+{
+ ipha_t *ipha = (ipha_t *)iph_arg;
+
+ multirt_check_v4(ire, ipha, ixa);
+
+ if (ire->ire_type & IRE_MULTICAST)
+ return (ire_send_multicast_v4(ire, mp, ipha, ixa, identp));
+ else if (ire->ire_type & IRE_BROADCAST)
+ return (ire_send_broadcast_v4(ire, mp, ipha, ixa, identp));
+ else
+ return (ire_send_wire_v4(ire, mp, ipha, ixa, identp));
+}
+
+/*
+ * ire_sendfn for IREs with RTF_REJECT/RTF_BLACKHOLE, including IRE_NOROUTE
+ */
+int
+ire_send_noroute_v4(ire_t *ire, mblk_t *mp, void *iph_arg,
+ ip_xmit_attr_t *ixa, uint32_t *identp)
+{
+ ip_stack_t *ipst = ixa->ixa_ipst;
+ ipha_t *ipha = (ipha_t *)iph_arg;
+ ill_t *ill;
+ ip_recv_attr_t iras;
+ boolean_t dummy;
+
+ /* We assign an IP ident for nice errors */
+ ipha->ipha_ident = atomic_add_32_nv(identp, 1);
+
+ BUMP_MIB(&ipst->ips_ip_mib, ipIfStatsOutNoRoutes);
+
+ if (ire->ire_type & IRE_NOROUTE) {
+ /* A lack of a route as opposed to RTF_REJECT|BLACKHOLE */
+ ip_rts_change(RTM_MISS, ipha->ipha_dst, 0, 0, 0, 0, 0, 0,
+ RTA_DST, ipst);
+ }
+
+ if (ire->ire_flags & RTF_BLACKHOLE) {
+ ip_drop_output("ipIfStatsOutNoRoutes RTF_BLACKHOLE", mp, NULL);
+ freemsg(mp);
+ /* No error even for local senders - silent blackhole */
+ return (0);
+ }
+ ip_drop_output("ipIfStatsOutNoRoutes RTF_REJECT", mp, NULL);
+
+ /*
+ * We need an ill_t for the ip_recv_attr_t even though this packet
+ * was never received and icmp_unreachable doesn't currently use
+ * ira_ill.
+ */
+ ill = ill_lookup_on_name("lo0", B_FALSE,
+ !(ixa->ixa_flags & IRAF_IS_IPV4), &dummy, ipst);
+ if (ill == NULL) {
+ freemsg(mp);
+ return (EHOSTUNREACH);
+ }
+
+ bzero(&iras, sizeof (iras));
+ /* Map ixa to ira including IPsec policies */
+ ipsec_out_to_in(ixa, ill, &iras);
+
+ if (ip_source_routed(ipha, ipst)) {
+ icmp_unreachable(mp, ICMP_SOURCE_ROUTE_FAILED, &iras);
+ } else {
+ icmp_unreachable(mp, ICMP_HOST_UNREACHABLE, &iras);
+ }
+ /* We moved any IPsec refs from ixa to iras */
+ ira_cleanup(&iras, B_FALSE);
+ ill_refrele(ill);
+ return (EHOSTUNREACH);
+}
+
+/*
+ * Calculate a checksum ignoring any hardware capabilities
+ *
+ * Returns B_FALSE if the packet was too short for the checksum. Caller
+ * should free and do stats.
+ */
+static boolean_t
+ip_output_sw_cksum_v4(mblk_t *mp, ipha_t *ipha, ip_xmit_attr_t *ixa)
+{
+ ip_stack_t *ipst = ixa->ixa_ipst;
+ uint_t pktlen = ixa->ixa_pktlen;
+ uint16_t *cksump;
+ uint32_t cksum;
+ uint8_t protocol = ixa->ixa_protocol;
+ uint16_t ip_hdr_length = ixa->ixa_ip_hdr_length;
+ ipaddr_t dst = ipha->ipha_dst;
+ ipaddr_t src = ipha->ipha_src;
+
+ /* Just in case it contained garbage */
+ DB_CKSUMFLAGS(mp) &= ~HCK_FLAGS;
+
+ /*
+ * Calculate ULP checksum
+ */
+ if (protocol == IPPROTO_TCP) {
+ cksump = IPH_TCPH_CHECKSUMP(ipha, ip_hdr_length);
+ cksum = IP_TCP_CSUM_COMP;
+ } else if (protocol == IPPROTO_UDP) {
+ cksump = IPH_UDPH_CHECKSUMP(ipha, ip_hdr_length);
+ cksum = IP_UDP_CSUM_COMP;
+ } else if (protocol == IPPROTO_SCTP) {
+ sctp_hdr_t *sctph;
+
+ ASSERT(MBLKL(mp) >= (ip_hdr_length + sizeof (*sctph)));
+ sctph = (sctp_hdr_t *)(mp->b_rptr + ip_hdr_length);
+ /*
+ * Zero out the checksum field to ensure proper
+ * checksum calculation.
+ */
+ sctph->sh_chksum = 0;
+#ifdef DEBUG
+ if (!skip_sctp_cksum)
+#endif
+ sctph->sh_chksum = sctp_cksum(mp, ip_hdr_length);
+ goto ip_hdr_cksum;
+ } else {
+ goto ip_hdr_cksum;
+ }
+
+ /* ULP puts the checksum field is in the first mblk */
+ ASSERT(((uchar_t *)cksump) + sizeof (uint16_t) <= mp->b_wptr);
+
+ /*
+ * We accumulate the pseudo header checksum in cksum.
+ * This is pretty hairy code, so watch close. One
+ * thing to keep in mind is that UDP and TCP have
+ * stored their respective datagram lengths in their
+ * checksum fields. This lines things up real nice.
+ */
+ cksum += (dst >> 16) + (dst & 0xFFFF) + (src >> 16) + (src & 0xFFFF);
+
+ cksum = IP_CSUM(mp, ip_hdr_length, cksum);
+ /*
+ * For UDP/IPv4 a zero means that the packets wasn't checksummed.
+ * Change to 0xffff
+ */
+ if (protocol == IPPROTO_UDP && cksum == 0)
+ *cksump = ~cksum;
+ else
+ *cksump = cksum;
+
+ IP_STAT(ipst, ip_out_sw_cksum);
+ IP_STAT_UPDATE(ipst, ip_out_sw_cksum_bytes, pktlen);
+
+ip_hdr_cksum:
+ /* Calculate IPv4 header checksum */
+ ipha->ipha_hdr_checksum = 0;
+ ipha->ipha_hdr_checksum = ip_csum_hdr(ipha);
+ return (B_TRUE);
+}
+
+/*
+ * Calculate the ULP checksum - try to use hardware.
+ * In the case of MULTIRT, broadcast or multicast the
+ * IXAF_NO_HW_CKSUM is set in which case we use software.
+ *
+ * If the hardware supports IP header checksum offload; then clear the
+ * contents of IP header checksum field as expected by NIC.
+ * Do this only if we offloaded either full or partial sum.
+ *
+ * Returns B_FALSE if the packet was too short for the checksum. Caller
+ * should free and do stats.
+ */
+static boolean_t
+ip_output_cksum_v4(iaflags_t ixaflags, mblk_t *mp, ipha_t *ipha,
+ ip_xmit_attr_t *ixa, ill_t *ill)
+{
+ uint_t pktlen = ixa->ixa_pktlen;
+ uint16_t *cksump;
+ uint16_t hck_flags;
+ uint32_t cksum;
+ uint8_t protocol = ixa->ixa_protocol;
+ uint16_t ip_hdr_length = ixa->ixa_ip_hdr_length;
+
+ if ((ixaflags & IXAF_NO_HW_CKSUM) || !ILL_HCKSUM_CAPABLE(ill) ||
+ !dohwcksum) {
+ return (ip_output_sw_cksum_v4(mp, ipha, ixa));
+ }
+
+ /*
+ * Calculate ULP checksum. Note that we don't use cksump and cksum
+ * if the ill has FULL support.
+ */
+ if (protocol == IPPROTO_TCP) {
+ cksump = IPH_TCPH_CHECKSUMP(ipha, ip_hdr_length);
+ cksum = IP_TCP_CSUM_COMP; /* Pseudo-header cksum */
+ } else if (protocol == IPPROTO_UDP) {
+ cksump = IPH_UDPH_CHECKSUMP(ipha, ip_hdr_length);
+ cksum = IP_UDP_CSUM_COMP; /* Pseudo-header cksum */
+ } else if (protocol == IPPROTO_SCTP) {
+ sctp_hdr_t *sctph;
+
+ ASSERT(MBLKL(mp) >= (ip_hdr_length + sizeof (*sctph)));
+ sctph = (sctp_hdr_t *)(mp->b_rptr + ip_hdr_length);
+ /*
+ * Zero out the checksum field to ensure proper
+ * checksum calculation.
+ */
+ sctph->sh_chksum = 0;
+#ifdef DEBUG
+ if (!skip_sctp_cksum)
+#endif
+ sctph->sh_chksum = sctp_cksum(mp, ip_hdr_length);
+ goto ip_hdr_cksum;
+ } else {
+ ip_hdr_cksum:
+ /* Calculate IPv4 header checksum */
+ ipha->ipha_hdr_checksum = 0;
+ ipha->ipha_hdr_checksum = ip_csum_hdr(ipha);
+ return (B_TRUE);
+ }
+
+ /* ULP puts the checksum field is in the first mblk */
+ ASSERT(((uchar_t *)cksump) + sizeof (uint16_t) <= mp->b_wptr);
+
+ /*
+ * Underlying interface supports hardware checksum offload for
+ * the payload; leave the payload checksum for the hardware to
+ * calculate. N.B: We only need to set up checksum info on the
+ * first mblk.
+ */
+ hck_flags = ill->ill_hcksum_capab->ill_hcksum_txflags;
+
+ DB_CKSUMFLAGS(mp) &= ~HCK_FLAGS;
+ if (hck_flags & HCKSUM_INET_FULL_V4) {
+ /*
+ * Hardware calculates pseudo-header, header and the
+ * payload checksums, so clear the checksum field in
+ * the protocol header.
+ */
+ *cksump = 0;
+ DB_CKSUMFLAGS(mp) |= HCK_FULLCKSUM;
+
+ ipha->ipha_hdr_checksum = 0;
+ if (hck_flags & HCKSUM_IPHDRCKSUM) {
+ DB_CKSUMFLAGS(mp) |= HCK_IPV4_HDRCKSUM;
+ } else {
+ ipha->ipha_hdr_checksum = ip_csum_hdr(ipha);
+ }
+ return (B_TRUE);
+ }
+ if ((hck_flags) & HCKSUM_INET_PARTIAL) {
+ ipaddr_t dst = ipha->ipha_dst;
+ ipaddr_t src = ipha->ipha_src;
+ /*
+ * Partial checksum offload has been enabled. Fill
+ * the checksum field in the protocol header with the
+ * pseudo-header checksum value.
+ *
+ * We accumulate the pseudo header checksum in cksum.
+ * This is pretty hairy code, so watch close. One
+ * thing to keep in mind is that UDP and TCP have
+ * stored their respective datagram lengths in their
+ * checksum fields. This lines things up real nice.
+ */
+ cksum += (dst >> 16) + (dst & 0xFFFF) +
+ (src >> 16) + (src & 0xFFFF);
+ cksum += *(cksump);
+ cksum = (cksum & 0xFFFF) + (cksum >> 16);
+ *(cksump) = (cksum & 0xFFFF) + (cksum >> 16);
+
+ /*
+ * Offsets are relative to beginning of IP header.
+ */
+ DB_CKSUMSTART(mp) = ip_hdr_length;
+ DB_CKSUMSTUFF(mp) = (uint8_t *)cksump - (uint8_t *)ipha;
+ DB_CKSUMEND(mp) = pktlen;
+ DB_CKSUMFLAGS(mp) |= HCK_PARTIALCKSUM;
+
+ ipha->ipha_hdr_checksum = 0;
+ if (hck_flags & HCKSUM_IPHDRCKSUM) {
+ DB_CKSUMFLAGS(mp) |= HCK_IPV4_HDRCKSUM;
+ } else {
+ ipha->ipha_hdr_checksum = ip_csum_hdr(ipha);
+ }
+ return (B_TRUE);
+ }
+ /* Hardware capabilities include neither full nor partial IPv4 */
+ return (ip_output_sw_cksum_v4(mp, ipha, ixa));
+}
+
+/*
+ * ire_sendfn for offlink and onlink destinations.
+ * Also called from the multicast, broadcast, multirt send functions.
+ *
+ * Assumes that the caller has a hold on the ire.
+ *
+ * This function doesn't care if the IRE just became condemned since that
+ * can happen at any time.
+ */
+/* ARGSUSED */
+int
+ire_send_wire_v4(ire_t *ire, mblk_t *mp, void *iph_arg,
+ ip_xmit_attr_t *ixa, uint32_t *identp)
+{
+ ip_stack_t *ipst = ixa->ixa_ipst;
+ ipha_t *ipha = (ipha_t *)iph_arg;
+ iaflags_t ixaflags = ixa->ixa_flags;
+ ill_t *ill;
+
+ ASSERT(ixa->ixa_nce != NULL);
+ ill = ixa->ixa_nce->nce_ill;
+
+ if (ixaflags & IXAF_DONTROUTE)
+ ipha->ipha_ttl = 1;
+
+ /*
+ * Assign an ident value for this packet. There could be other
+ * threads targeting the same destination, so we have to arrange
+ * for a atomic increment. Note that we use a 32-bit atomic add
+ * because it has better performance than its 16-bit sibling.
+ *
+ * Normally ixa_extra_ident is 0, but in the case of LSO it will
+ * be the number of TCP segments that the driver/hardware will
+ * extraly construct.
+ *
+ * If running in cluster mode and if the source address
+ * belongs to a replicated service then vector through
+ * cl_inet_ipident vector to allocate ip identifier
+ * NOTE: This is a contract private interface with the
+ * clustering group.
+ */
+ if (cl_inet_ipident != NULL) {
+ ipaddr_t src = ipha->ipha_src;
+ ipaddr_t dst = ipha->ipha_dst;
+ netstackid_t stack_id = ipst->ips_netstack->netstack_stackid;
+
+ ASSERT(cl_inet_isclusterwide != NULL);
+ if ((*cl_inet_isclusterwide)(stack_id, IPPROTO_IP,
+ AF_INET, (uint8_t *)(uintptr_t)src, NULL)) {
+ /*
+ * Note: not correct with LSO since we can't allocate
+ * ixa_extra_ident+1 consecutive values.
+ */
+ ipha->ipha_ident = (*cl_inet_ipident)(stack_id,
+ IPPROTO_IP, AF_INET, (uint8_t *)(uintptr_t)src,
+ (uint8_t *)(uintptr_t)dst, NULL);
+ } else {
+ ipha->ipha_ident = atomic_add_32_nv(identp,
+ ixa->ixa_extra_ident + 1);
+ }
+ } else {
+ ipha->ipha_ident = atomic_add_32_nv(identp,
+ ixa->ixa_extra_ident + 1);
+ }
+#ifndef _BIG_ENDIAN
+ ipha->ipha_ident = htons(ipha->ipha_ident);
+#endif
+
+ /*
+ * This might set b_band, thus the IPsec and fragmentation
+ * code in IP ensures that b_band is updated in the first mblk.
+ */
+ if (IPP_ENABLED(IPP_LOCAL_OUT, ipst)) {
+ /* ip_process translates an IS_UNDER_IPMP */
+ mp = ip_process(IPP_LOCAL_OUT, mp, ill, ill);
+ if (mp == NULL) {
+ /* ip_drop_packet and MIB done */
+ return (0); /* Might just be delayed */
+ }
+ }
+
+ /*
+ * Verify any IPv4 options.
+ *
+ * The presense of IP options also forces the network stack to
+ * calculate the checksum in software. This is because:
+ *
+ * Wrap around: certain partial-checksum NICs (eri, ce) limit
+ * the size of "start offset" width to 6-bit. This effectively
+ * sets the largest value of the offset to 64-bytes, starting
+ * from the MAC header. When the cumulative MAC and IP headers
+ * exceed such limit, the offset will wrap around. This causes
+ * the checksum to be calculated at the wrong place.
+ *
+ * IPv4 source routing: none of the full-checksum capable NICs
+ * is capable of correctly handling the IPv4 source-routing
+ * option for purposes of calculating the pseudo-header; the
+ * actual destination is different from the destination in the
+ * header which is that of the next-hop. (This case may not be
+ * true for NICs which can parse IPv6 extension headers, but
+ * we choose to simplify the implementation by not offloading
+ * checksum when they are present.)
+ */
+ if (!IS_SIMPLE_IPH(ipha)) {
+ ixaflags = ixa->ixa_flags |= IXAF_NO_HW_CKSUM;
+ /* An IS_UNDER_IPMP ill is ok here */
+ if (ip_output_options(mp, ipha, ixa, ill)) {
+ /* Packet has been consumed and ICMP error sent */
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsOutDiscards);
+ return (EINVAL);
+ }
+ }
+
+ /*
+ * To handle IPsec/iptun's labeling needs we need to tag packets
+ * while we still have ixa_tsl
+ */
+ if (is_system_labeled() && ixa->ixa_tsl != NULL &&
+ (ill->ill_mactype == DL_6TO4 || ill->ill_mactype == DL_IPV4 ||
+ ill->ill_mactype == DL_IPV6)) {
+ cred_t *newcr;
+
+ newcr = copycred_from_tslabel(ixa->ixa_cred, ixa->ixa_tsl,
+ KM_NOSLEEP);
+ if (newcr == NULL) {
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards - newcr",
+ mp, ill);
+ freemsg(mp);
+ return (ENOBUFS);
+ }
+ mblk_setcred(mp, newcr, NOPID);
+ crfree(newcr); /* mblk_setcred did its own crhold */
+ }
+
+ if (ixa->ixa_pktlen > ixa->ixa_fragsize ||
+ (ixaflags & IXAF_IPSEC_SECURE)) {
+ uint32_t pktlen;
+
+ pktlen = ixa->ixa_pktlen;
+ if (ixaflags & IXAF_IPSEC_SECURE)
+ pktlen += ipsec_out_extra_length(ixa);
+
+ if (pktlen > IP_MAXPACKET)
+ return (EMSGSIZE);
+
+ if (ixaflags & IXAF_SET_ULP_CKSUM) {
+ /*
+ * Compute ULP checksum and IP header checksum
+ * using software
+ */
+ if (!ip_output_sw_cksum_v4(mp, ipha, ixa)) {
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards", mp, ill);
+ freemsg(mp);
+ return (EINVAL);
+ }
+ } else {
+ /* Calculate IPv4 header checksum */
+ ipha->ipha_hdr_checksum = 0;
+ ipha->ipha_hdr_checksum = ip_csum_hdr(ipha);
+ }
+
+ /*
+ * If this packet would generate a icmp_frag_needed
+ * message, we need to handle it before we do the IPsec
+ * processing. Otherwise, we need to strip the IPsec
+ * headers before we send up the message to the ULPs
+ * which becomes messy and difficult.
+ *
+ * We check using IXAF_DONTFRAG. The DF bit in the header
+ * is not inspected - it will be copied to any generated
+ * fragments.
+ */
+ if ((pktlen > ixa->ixa_fragsize) &&
+ (ixaflags & IXAF_DONTFRAG)) {
+ /* Generate ICMP and return error */
+ ip_recv_attr_t iras;
+
+ DTRACE_PROBE4(ip4__fragsize__fail, uint_t, pktlen,
+ uint_t, ixa->ixa_fragsize, uint_t, ixa->ixa_pktlen,
+ uint_t, ixa->ixa_pmtu);
+
+ bzero(&iras, sizeof (iras));
+ /* Map ixa to ira including IPsec policies */
+ ipsec_out_to_in(ixa, ill, &iras);
+
+ ip_drop_output("ICMP_FRAG_NEEDED", mp, ill);
+ icmp_frag_needed(mp, ixa->ixa_fragsize, &iras);
+ /* We moved any IPsec refs from ixa to iras */
+ ira_cleanup(&iras, B_FALSE);
+ return (EMSGSIZE);
+ }
+ DTRACE_PROBE4(ip4__fragsize__ok, uint_t, pktlen,
+ uint_t, ixa->ixa_fragsize, uint_t, ixa->ixa_pktlen,
+ uint_t, ixa->ixa_pmtu);
+
+ if (ixaflags & IXAF_IPSEC_SECURE) {
+ /*
+ * Pass in sufficient information so that
+ * IPsec can determine whether to fragment, and
+ * which function to call after fragmentation.
+ */
+ return (ipsec_out_process(mp, ixa));
+ }
+ return (ip_fragment_v4(mp, ixa->ixa_nce, ixaflags,
+ ixa->ixa_pktlen, ixa->ixa_fragsize, ixa->ixa_xmit_hint,
+ ixa->ixa_zoneid, ixa->ixa_no_loop_zoneid,
+ ixa->ixa_postfragfn, &ixa->ixa_cookie));
+ }
+ if (ixaflags & IXAF_SET_ULP_CKSUM) {
+ /* Compute ULP checksum and IP header checksum */
+ /* An IS_UNDER_IPMP ill is ok here */
+ if (!ip_output_cksum_v4(ixaflags, mp, ipha, ixa, ill)) {
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards", mp, ill);
+ freemsg(mp);
+ return (EINVAL);
+ }
+ } else {
+ /* Calculate IPv4 header checksum */
+ ipha->ipha_hdr_checksum = 0;
+ ipha->ipha_hdr_checksum = ip_csum_hdr(ipha);
+ }
+ return ((ixa->ixa_postfragfn)(mp, ixa->ixa_nce, ixaflags,
+ ixa->ixa_pktlen, ixa->ixa_xmit_hint, ixa->ixa_zoneid,
+ ixa->ixa_no_loop_zoneid, &ixa->ixa_cookie));
+}
+
+/*
+ * Send mp into ip_input
+ * Common for IPv4 and IPv6
+ */
+void
+ip_postfrag_loopback(mblk_t *mp, nce_t *nce, iaflags_t ixaflags,
+ uint_t pkt_len, zoneid_t nolzid)
+{
+ rtc_t rtc;
+ ill_t *ill = nce->nce_ill;
+ ip_recv_attr_t iras; /* NOTE: No bzero for performance */
+ ncec_t *ncec;
+
+ ncec = nce->nce_common;
+ iras.ira_flags = IRAF_VERIFY_IP_CKSUM | IRAF_VERIFY_ULP_CKSUM |
+ IRAF_LOOPBACK | IRAF_L2SRC_LOOPBACK;
+ if (ncec->ncec_flags & NCE_F_BCAST)
+ iras.ira_flags |= IRAF_L2DST_BROADCAST;
+ else if (ncec->ncec_flags & NCE_F_MCAST)
+ iras.ira_flags |= IRAF_L2DST_MULTICAST;
+
+ iras.ira_free_flags = 0;
+ iras.ira_cred = NULL;
+ iras.ira_cpid = NOPID;
+ iras.ira_tsl = NULL;
+ iras.ira_zoneid = ALL_ZONES;
+ iras.ira_pktlen = pkt_len;
+ UPDATE_MIB(ill->ill_ip_mib, ipIfStatsHCInOctets, iras.ira_pktlen);
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsHCInReceives);
+
+ if (ixaflags & IXAF_IS_IPV4)
+ iras.ira_flags |= IRAF_IS_IPV4;
+
+ iras.ira_ill = iras.ira_rill = ill;
+ iras.ira_ruifindex = ill->ill_phyint->phyint_ifindex;
+ iras.ira_rifindex = iras.ira_ruifindex;
+ iras.ira_mhip = NULL;
+
+ iras.ira_flags |= ixaflags & IAF_MASK;
+ iras.ira_no_loop_zoneid = nolzid;
+
+ /* Broadcast and multicast doesn't care about the squeue */
+ iras.ira_sqp = NULL;
+
+ rtc.rtc_ire = NULL;
+ if (ixaflags & IXAF_IS_IPV4) {
+ ipha_t *ipha = (ipha_t *)mp->b_rptr;
+
+ rtc.rtc_ipaddr = INADDR_ANY;
+
+ (*ill->ill_inputfn)(mp, ipha, &ipha->ipha_dst, &iras, &rtc);
+ if (rtc.rtc_ire != NULL) {
+ ASSERT(rtc.rtc_ipaddr != INADDR_ANY);
+ ire_refrele(rtc.rtc_ire);
+ }
+ } else {
+ ip6_t *ip6h = (ip6_t *)mp->b_rptr;
+
+ rtc.rtc_ip6addr = ipv6_all_zeros;
+
+ (*ill->ill_inputfn)(mp, ip6h, &ip6h->ip6_dst, &iras, &rtc);
+ if (rtc.rtc_ire != NULL) {
+ ASSERT(!IN6_IS_ADDR_UNSPECIFIED(&rtc.rtc_ip6addr));
+ ire_refrele(rtc.rtc_ire);
+ }
+ }
+ /* Any references to clean up? No hold on ira */
+ if (iras.ira_flags & (IRAF_IPSEC_SECURE|IRAF_SYSTEM_LABELED))
+ ira_cleanup(&iras, B_FALSE);
+}
+
+/*
+ * Post fragmentation function for IRE_MULTICAST and IRE_BROADCAST which
+ * looks at the IXAF_LOOPBACK_COPY flag.
+ * Common for IPv4 and IPv6.
+ *
+ * If the loopback copy fails (due to no memory) but we send the packet out
+ * on the wire we return no failure. Only in the case we supress the wire
+ * sending do we take the loopback failure into account.
+ *
+ * Note that we do not perform DTRACE_IP7 and FW_HOOKS for the looped back copy.
+ * Those operations are performed on this packet in ip_xmit() and it would
+ * be odd to do it twice for the same packet.
+ */
+int
+ip_postfrag_loopcheck(mblk_t *mp, nce_t *nce, iaflags_t ixaflags,
+ uint_t pkt_len, uint32_t xmit_hint, zoneid_t szone, zoneid_t nolzid,
+ uintptr_t *ixacookie)
+{
+ ill_t *ill = nce->nce_ill;
+ int error = 0;
+
+ /*
+ * Check for IXAF_LOOPBACK_COPY - send a copy to ip as if the driver
+ * had looped it back
+ */
+ if (ixaflags & IXAF_LOOPBACK_COPY) {
+ mblk_t *mp1;
+
+ mp1 = copymsg(mp);
+ if (mp1 == NULL) {
+ /* Failed to deliver the loopback copy. */
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards", mp, ill);
+ error = ENOBUFS;
+ } else {
+ ip_postfrag_loopback(mp1, nce, ixaflags, pkt_len,
+ nolzid);
+ }
+ }
+
+ /*
+ * If TTL = 0 then only do the loopback to this host i.e. we are
+ * done. We are also done if this was the
+ * loopback interface since it is sufficient
+ * to loopback one copy of a multicast packet.
+ */
+ if (ixaflags & IXAF_IS_IPV4) {
+ ipha_t *ipha = (ipha_t *)mp->b_rptr;
+
+ if (ipha->ipha_ttl == 0) {
+ ip_drop_output("multicast ipha_ttl not sent to wire",
+ mp, ill);
+ freemsg(mp);
+ return (error);
+ }
+ } else {
+ ip6_t *ip6h = (ip6_t *)mp->b_rptr;
+
+ if (ip6h->ip6_hops == 0) {
+ ip_drop_output("multicast ipha_ttl not sent to wire",
+ mp, ill);
+ freemsg(mp);
+ return (error);
+ }
+ }
+ if (nce->nce_ill->ill_wq == NULL) {
+ /* Loopback interface */
+ ip_drop_output("multicast on lo0 not sent to wire", mp, ill);
+ freemsg(mp);
+ return (error);
+ }
+
+ return (ip_xmit(mp, nce, ixaflags, pkt_len, xmit_hint, szone, 0,
+ ixacookie));
+}
+
+/*
+ * Post fragmentation function for RTF_MULTIRT routes.
+ * Since IRE_BROADCASTs can have RTF_MULTIRT, this function
+ * checks IXAF_LOOPBACK_COPY.
+ *
+ * If no packet is sent due to failures then we return an errno, but if at
+ * least one succeeded we return zero.
+ */
+int
+ip_postfrag_multirt_v4(mblk_t *mp, nce_t *nce, iaflags_t ixaflags,
+ uint_t pkt_len, uint32_t xmit_hint, zoneid_t szone, zoneid_t nolzid,
+ uintptr_t *ixacookie)
+{
+ irb_t *irb;
+ ipha_t *ipha = (ipha_t *)mp->b_rptr;
+ ire_t *ire;
+ ire_t *ire1;
+ mblk_t *mp1;
+ nce_t *nce1;
+ ill_t *ill = nce->nce_ill;
+ ill_t *ill1;
+ ip_stack_t *ipst = ill->ill_ipst;
+ int error = 0;
+ int num_sent = 0;
+ int err;
+ uint_t ire_type;
+ ipaddr_t nexthop;
+
+ ASSERT(ixaflags & IXAF_IS_IPV4);
+
+ /* Check for IXAF_LOOPBACK_COPY */
+ if (ixaflags & IXAF_LOOPBACK_COPY) {
+ mblk_t *mp1;
+
+ mp1 = copymsg(mp);
+ if (mp1 == NULL) {
+ /* Failed to deliver the loopback copy. */
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards", mp, ill);
+ error = ENOBUFS;
+ } else {
+ ip_postfrag_loopback(mp1, nce, ixaflags, pkt_len,
+ nolzid);
+ }
+ }
+
+ /*
+ * Loop over RTF_MULTIRT for ipha_dst in the same bucket. Send
+ * a copy to each one.
+ * Use the nce (nexthop) and ipha_dst to find the ire.
+ *
+ * MULTIRT is not designed to work with shared-IP zones thus we don't
+ * need to pass a zoneid or a label to the IRE lookup.
+ */
+ if (V4_PART_OF_V6(nce->nce_addr) == ipha->ipha_dst) {
+ /* Broadcast and multicast case */
+ ire = ire_ftable_lookup_v4(ipha->ipha_dst, 0, 0, 0,
+ NULL, ALL_ZONES, NULL, MATCH_IRE_DSTONLY, 0, ipst, NULL);
+ } else {
+ ipaddr_t v4addr = V4_PART_OF_V6(nce->nce_addr);
+
+ /* Unicast case */
+ ire = ire_ftable_lookup_v4(ipha->ipha_dst, 0, v4addr, 0,
+ NULL, ALL_ZONES, NULL, MATCH_IRE_GW, 0, ipst, NULL);
+ }
+
+ if (ire == NULL ||
+ (ire->ire_flags & (RTF_REJECT|RTF_BLACKHOLE)) ||
+ !(ire->ire_flags & RTF_MULTIRT)) {
+ /* Drop */
+ ip_drop_output("ip_postfrag_multirt didn't find route",
+ mp, nce->nce_ill);
+ if (ire != NULL)
+ ire_refrele(ire);
+ return (ENETUNREACH);
+ }
+
+ irb = ire->ire_bucket;
+ irb_refhold(irb);
+ for (ire1 = irb->irb_ire; ire1 != NULL; ire1 = ire1->ire_next) {
+ /*
+ * For broadcast we can have a mixture of IRE_BROADCAST and
+ * IRE_HOST due to the manually added IRE_HOSTs that are used
+ * to trigger the creation of the special CGTP broadcast routes.
+ * Thus we have to skip if ire_type doesn't match the original.
+ */
+ if (IRE_IS_CONDEMNED(ire1) ||
+ !(ire1->ire_flags & RTF_MULTIRT) ||
+ ire1->ire_type != ire->ire_type)
+ continue;
+
+ /* Do the ire argument one after the loop */
+ if (ire1 == ire)
+ continue;
+
+ ill1 = ire_nexthop_ill(ire1);
+ if (ill1 == NULL) {
+ /*
+ * This ire might not have been picked by
+ * ire_route_recursive, in which case ire_dep might
+ * not have been setup yet.
+ * We kick ire_route_recursive to try to resolve
+ * starting at ire1.
+ */
+ ire_t *ire2;
+
+ ire2 = ire_route_recursive_impl_v4(ire1,
+ ire1->ire_addr, ire1->ire_type, ire1->ire_ill,
+ ire1->ire_zoneid, NULL, MATCH_IRE_DSTONLY,
+ B_TRUE, 0, ipst, NULL, NULL, NULL);
+ if (ire2 != NULL)
+ ire_refrele(ire2);
+ ill1 = ire_nexthop_ill(ire1);
+ }
+
+ if (ill1 == NULL) {
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards - no ill",
+ mp, ill);
+ error = ENETUNREACH;
+ continue;
+ }
+
+ /* Pick the addr and type to use for arp_nce_init */
+ if (nce->nce_common->ncec_flags & NCE_F_BCAST) {
+ ire_type = IRE_BROADCAST;
+ nexthop = ire1->ire_gateway_addr;
+ } else if (nce->nce_common->ncec_flags & NCE_F_MCAST) {
+ ire_type = IRE_MULTICAST;
+ nexthop = ipha->ipha_dst;
+ } else {
+ ire_type = ire1->ire_type; /* Doesn't matter */
+ nexthop = ire1->ire_gateway_addr;
+ }
+
+ /* If IPMP meta or under, then we just drop */
+ if (ill1->ill_grp != NULL) {
+ BUMP_MIB(ill1->ill_ip_mib, ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards - IPMP",
+ mp, ill1);
+ ill_refrele(ill1);
+ error = ENETUNREACH;
+ continue;
+ }
+
+ nce1 = arp_nce_init(ill1, nexthop, ire_type);
+ if (nce1 == NULL) {
+ BUMP_MIB(ill1->ill_ip_mib, ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards - no nce",
+ mp, ill1);
+ ill_refrele(ill1);
+ error = ENETUNREACH;
+ continue;
+ }
+ mp1 = copymsg(mp);
+ if (mp1 == NULL) {
+ BUMP_MIB(ill1->ill_ip_mib, ipIfStatsOutDiscards);
+ ip_drop_output("ipIfStatsOutDiscards", mp, ill1);
+ nce_refrele(nce1);
+ ill_refrele(ill1);
+ error = ENOBUFS;
+ continue;
+ }
+ /* Preserve HW checksum for this copy */
+ DB_CKSUMSTART(mp1) = DB_CKSUMSTART(mp);
+ DB_CKSUMSTUFF(mp1) = DB_CKSUMSTUFF(mp);
+ DB_CKSUMEND(mp1) = DB_CKSUMEND(mp);
+ DB_CKSUMFLAGS(mp1) = DB_CKSUMFLAGS(mp);
+ DB_LSOMSS(mp1) = DB_LSOMSS(mp);
+
+ ire1->ire_ob_pkt_count++;
+ err = ip_xmit(mp1, nce1, ixaflags, pkt_len, xmit_hint, szone,
+ 0, ixacookie);
+ if (err == 0)
+ num_sent++;
+ else
+ error = err;
+ nce_refrele(nce1);
+ ill_refrele(ill1);
+ }
+ irb_refrele(irb);
+ ire_refrele(ire);
+ /* Finally, the main one */
+ err = ip_xmit(mp, nce, ixaflags, pkt_len, xmit_hint, szone, 0,
+ ixacookie);
+ if (err == 0)
+ num_sent++;
+ else
+ error = err;
+ if (num_sent > 0)
+ return (0);
+ else
+ return (error);
+}
+
+/*
+ * Verify local connectivity. This check is called by ULP fusion code.
+ * The generation number on an IRE_LOCAL or IRE_LOOPBACK only changes if
+ * the interface is brought down and back up. So we simply fail the local
+ * process. The caller, TCP Fusion, should unfuse the connection.
+ */
+boolean_t
+ip_output_verify_local(ip_xmit_attr_t *ixa)
+{
+ ire_t *ire = ixa->ixa_ire;
+
+ if (!(ire->ire_type & (IRE_LOCAL | IRE_LOOPBACK)))
+ return (B_FALSE);
+
+ return (ixa->ixa_ire->ire_generation == ixa->ixa_ire_generation);
+}
+
+/*
+ * Local process for ULP loopback, TCP Fusion. Handle both IPv4 and IPv6.
+ *
+ * The caller must call ip_output_verify_local() first. This function handles
+ * IPobs, FW_HOOKS, and/or IPsec cases sequentially.
+ */
+mblk_t *
+ip_output_process_local(mblk_t *mp, ip_xmit_attr_t *ixa, boolean_t hooks_out,
+ boolean_t hooks_in, conn_t *peer_connp)
+{
+ ill_t *ill = ixa->ixa_ire->ire_ill;
+ ipha_t *ipha = NULL;
+ ip6_t *ip6h = NULL;
+ ip_stack_t *ipst = ixa->ixa_ipst;
+ iaflags_t ixaflags = ixa->ixa_flags;
+ ip_recv_attr_t iras;
+ int error;
+
+ ASSERT(mp != NULL);
+
+ if (ixaflags & IXAF_IS_IPV4) {
+ ipha = (ipha_t *)mp->b_rptr;
+
+ /*
+ * If a callback is enabled then we need to know the
+ * source and destination zoneids for the packet. We already
+ * have those handy.
+ */
+ if (ipst->ips_ip4_observe.he_interested) {
+ zoneid_t szone, dzone;
+ zoneid_t stackzoneid;
+
+ stackzoneid = netstackid_to_zoneid(
+ ipst->ips_netstack->netstack_stackid);
+
+ if (stackzoneid == GLOBAL_ZONEID) {
+ /* Shared-IP zone */
+ dzone = ixa->ixa_ire->ire_zoneid;
+ szone = ixa->ixa_zoneid;
+ } else {
+ szone = dzone = stackzoneid;
+ }
+ ipobs_hook(mp, IPOBS_HOOK_LOCAL, szone, dzone, ill,
+ ipst);
+ }
+ DTRACE_IP7(send, mblk_t *, mp, conn_t *, NULL, void_ip_t *,
+ ipha, __dtrace_ipsr_ill_t *, ill, ipha_t *, ipha, ip6_t *,
+ NULL, int, 1);
+
+ /* FW_HOOKS: LOOPBACK_OUT */
+ if (hooks_out) {
+ DTRACE_PROBE4(ip4__loopback__out__start, ill_t *, NULL,
+ ill_t *, ill, ipha_t *, ipha, mblk_t *, mp);
+ FW_HOOKS(ipst->ips_ip4_loopback_out_event,
+ ipst->ips_ipv4firewall_loopback_out,
+ NULL, ill, ipha, mp, mp, 0, ipst, error);
+ DTRACE_PROBE1(ip4__loopback__out__end, mblk_t *, mp);
+ }
+ if (mp == NULL)
+ return (NULL);
+
+ /* FW_HOOKS: LOOPBACK_IN */
+ if (hooks_in) {
+ DTRACE_PROBE4(ip4__loopback__in__start, ill_t *, ill,
+ ill_t *, NULL, ipha_t *, ipha, mblk_t *, mp);
+ FW_HOOKS(ipst->ips_ip4_loopback_in_event,
+ ipst->ips_ipv4firewall_loopback_in,
+ ill, NULL, ipha, mp, mp, 0, ipst, error);
+ DTRACE_PROBE1(ip4__loopback__in__end, mblk_t *, mp);
+ }
+ if (mp == NULL)
+ return (NULL);
+
+ DTRACE_IP7(receive, mblk_t *, mp, conn_t *, NULL, void_ip_t *,
+ ipha, __dtrace_ipsr_ill_t *, ill, ipha_t *, ipha, ip6_t *,
+ NULL, int, 1);
+
+ /* Inbound IPsec polocies */
+ if (peer_connp != NULL) {
+ /* Map ixa to ira including IPsec policies. */
+ ipsec_out_to_in(ixa, ill, &iras);
+ mp = ipsec_check_inbound_policy(mp, peer_connp, ipha,
+ NULL, &iras);
+ }
+ } else {
+ ip6h = (ip6_t *)mp->b_rptr;
+
+ /*
+ * If a callback is enabled then we need to know the
+ * source and destination zoneids for the packet. We already
+ * have those handy.
+ */
+ if (ipst->ips_ip6_observe.he_interested) {
+ zoneid_t szone, dzone;
+ zoneid_t stackzoneid;
+
+ stackzoneid = netstackid_to_zoneid(
+ ipst->ips_netstack->netstack_stackid);
+
+ if (stackzoneid == GLOBAL_ZONEID) {
+ /* Shared-IP zone */
+ dzone = ixa->ixa_ire->ire_zoneid;
+ szone = ixa->ixa_zoneid;
+ } else {
+ szone = dzone = stackzoneid;
+ }
+ ipobs_hook(mp, IPOBS_HOOK_LOCAL, szone, dzone, ill,
+ ipst);
+ }
+ DTRACE_IP7(send, mblk_t *, mp, conn_t *, NULL, void_ip_t *,
+ ip6h, __dtrace_ipsr_ill_t *, ill, ipha_t *, NULL, ip6_t *,
+ ip6h, int, 1);
+
+ /* FW_HOOKS: LOOPBACK_OUT */
+ if (hooks_out) {
+ DTRACE_PROBE4(ip6__loopback__out__start, ill_t *, NULL,
+ ill_t *, ill, ip6_t *, ip6h, mblk_t *, mp);
+ FW_HOOKS6(ipst->ips_ip6_loopback_out_event,
+ ipst->ips_ipv6firewall_loopback_out,
+ NULL, ill, ip6h, mp, mp, 0, ipst, error);
+ DTRACE_PROBE1(ip6__loopback__out__end, mblk_t *, mp);
+ }
+ if (mp == NULL)
+ return (NULL);
+
+ /* FW_HOOKS: LOOPBACK_IN */
+ if (hooks_in) {
+ DTRACE_PROBE4(ip6__loopback__in__start, ill_t *, ill,
+ ill_t *, NULL, ip6_t *, ip6h, mblk_t *, mp);
+ FW_HOOKS6(ipst->ips_ip6_loopback_in_event,
+ ipst->ips_ipv6firewall_loopback_in,
+ ill, NULL, ip6h, mp, mp, 0, ipst, error);
+ DTRACE_PROBE1(ip6__loopback__in__end, mblk_t *, mp);
+ }
+ if (mp == NULL)
+ return (NULL);
+
+ DTRACE_IP7(receive, mblk_t *, mp, conn_t *, NULL, void_ip_t *,
+ ip6h, __dtrace_ipsr_ill_t *, ill, ipha_t *, NULL, ip6_t *,
+ ip6h, int, 1);
+
+ /* Inbound IPsec polocies */
+ if (peer_connp != NULL) {
+ /* Map ixa to ira including IPsec policies. */
+ ipsec_out_to_in(ixa, ill, &iras);
+ mp = ipsec_check_inbound_policy(mp, peer_connp, NULL,
+ ip6h, &iras);
+ }
+ }
+
+ if (mp == NULL) {
+ BUMP_MIB(ill->ill_ip_mib, ipIfStatsInDiscards);
+ ip_drop_input("ipIfStatsInDiscards", NULL, ill);
+ }
+
+ return (mp);
+}