summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--usr/src/uts/common/io/ixgbe/ixgbe_tx.c714
-rw-r--r--usr/src/uts/common/mapfiles/ddi.mapfile3
2 files changed, 296 insertions, 421 deletions
diff --git a/usr/src/uts/common/io/ixgbe/ixgbe_tx.c b/usr/src/uts/common/io/ixgbe/ixgbe_tx.c
index b6926fc3b7..eb91f4af09 100644
--- a/usr/src/uts/common/io/ixgbe/ixgbe_tx.c
+++ b/usr/src/uts/common/io/ixgbe/ixgbe_tx.c
@@ -27,19 +27,21 @@
* Copyright (c) 2008, 2010, Oracle and/or its affiliates. All rights reserved.
* Copyright 2012 Nexenta Systems, Inc. All rights reserved.
* Copyright 2016 OmniTI Computer Consulting, Inc. All rights reserved.
- * Copyright 2017 Joyent, Inc.
+ * Copyright 2020 Joyent, Inc.
*/
#include "ixgbe_sw.h"
-static int ixgbe_tx_copy(ixgbe_tx_ring_t *, tx_control_block_t *, mblk_t *,
- uint32_t, boolean_t);
-static int ixgbe_tx_bind(ixgbe_tx_ring_t *, tx_control_block_t *, mblk_t *,
- uint32_t);
+static int ixgbe_tx_copy(ixgbe_tx_ring_t *, tx_control_block_t **,
+ link_list_t *, const void *, size_t);
+static int ixgbe_tx_bind(ixgbe_tx_ring_t *, tx_control_block_t **,
+ link_list_t *, uint8_t *, size_t);
+static uint_t ixgbe_tcb_done(tx_control_block_t *);
static int ixgbe_tx_fill_ring(ixgbe_tx_ring_t *, link_list_t *,
ixgbe_tx_context_t *, size_t);
static void ixgbe_save_desc(tx_control_block_t *, uint64_t, size_t);
-static tx_control_block_t *ixgbe_get_free_list(ixgbe_tx_ring_t *);
+static tx_control_block_t *ixgbe_get_free_list(ixgbe_tx_ring_t *,
+ link_list_t *);
static int ixgbe_get_context(mblk_t *, ixgbe_tx_context_t *);
static boolean_t ixgbe_check_context(ixgbe_tx_ring_t *,
@@ -65,42 +67,71 @@ static void ixgbe_fill_context(struct ixgbe_adv_tx_context_desc *,
* they will be processed by using bcopy; otherwise, they will
* be processed by using DMA binding.
*
- * To process the mblk, a tx control block is got from the
- * free list. One tx control block contains one tx buffer, which
- * is used to copy mblk fragments' data; and one tx DMA handle,
- * which is used to bind a mblk fragment with DMA resource.
+ * To process the mblk, for each fragment, we pass a pointer to the location
+ * of the current transmit control block (tcb) (initialized to NULL) to either
+ * ixgbe_tx_copy() or ixgbe_tx_bind() (based on the size of the mblk fragment).
+ * ixgbe_tx_copy() and ixgbe_tx_bind() will either continue to use the current
+ * if possible, or close out the current tcb, allocate a new tcb, and update
+ * the passed location (tx_control_block_t **) to reflect the new current tcb.
*
- * Several small mblk fragments can be copied into one tx control
- * block's buffer, and then the buffer will be transmitted with
- * one tx descriptor.
+ * Since bound mblk fragments require their own tcb, the close, allocate new,
+ * and update steps occur on every call to ixgbe_tx_bind(), but since
+ * consecutive small mblk fragments can be combined into a single tcb, the
+ * close, allocate new, and update steps may not occur on every call to
+ * ixgbe_tx_copy(). If the current tcb is already being used to copy data and
+ * we call ixgbe_tx_copy(), if there is enough room in the current tcb for
+ * the current mblk fragment, we append the data from the mblk fragment. If
+ * we call ixgbe_tx_copy() and the current tcb isn't being used to copy (i.e.
+ * the previous iteration of the loop called ixgbe_tx_bind()), or doesn't
+ * have enough space for the mblk fragment, we close out the current tcb,
+ * grab a new tcb from the free list, and update the current tcb to the
+ * newly obtained tcb.
*
- * A large fragment only binds with one tx control block's DMA
- * handle, and it can span several tx descriptors for transmitting.
+ * When LSO (large segment offload) is enabled, we first copy the packet
+ * headers (ethernet, IP, and TCP/UDP) into their own descriptor before
+ * processing the remainder of the packet. The remaining bytes of the packet
+ * are then copied or mapped based on the fragment size as described above.
+ *
+ * Through the entire processing of a packet, we keep track of the number of
+ * DMA descriptors being used (either bound or pre-bound buffers used for
+ * copying) by this packet. Each tcb requires at least one DMA descriptor, but
+ * may require more than one. When a tcb is closed by ixgbe_tx_bind() or
+ * ixgbe_tx_copy(), it does so by calling ixgbe_tcb_done() which returns the
+ * number of DMA descriptors that are closed (ready for the HW). Since the
+ * hardware limits the number of descriptors that can be used to transmit a
+ * single packet, if the total number DMA descriptors required to transmit
+ * this packet exceeds this limit, we perform a msgpullup() and try again.
+ * Since our DMA attributes limit the number of DMA cookies allowed to
+ * map a single span of memory to a value (MAX_COOKIE) less than the
+ * maximum number of descriptors allowed for a packet (IXGBE_TX_DESC_LIMIT),
+ * as long as sufficient tcbs are available, we should always be able to
+ * process a packet that's contained in a single mblk_t (no additional
+ * fragments).
+ *
+ * Once all of the tcbs have been setup, ixgbe_tx_fill_ring() is called to
+ * setup the tx ring to transmit the tcbs and then tell the HW to start
+ * transmitting. When transmission is complete, an interrupt is triggered
+ * which calls the appropriate recycle routine to place the tcbs that were
+ * used in transmission back in the free list. We also may also try to
+ * recycle any available tcbs when the size of the tcb free list gets low
+ * or if the watchdog timer triggers.
*
- * So to transmit a packet (mblk), several tx control blocks can
- * be used. After the processing, those tx control blocks will
- * be put to the work list.
*/
mblk_t *
-ixgbe_ring_tx(void *arg, mblk_t *mp)
+ixgbe_ring_tx(void *arg, mblk_t *orig_mp)
{
ixgbe_tx_ring_t *tx_ring = (ixgbe_tx_ring_t *)arg;
ixgbe_t *ixgbe = tx_ring->ixgbe;
- tx_type_t current_flag, next_flag;
- uint32_t current_len, next_len;
+ mblk_t *mp = orig_mp;
+ mblk_t *pull_mp = NULL;
+ tx_control_block_t *tcb;
+ size_t mbsize, offset, len;
uint32_t desc_total;
- size_t mbsize;
+ uint32_t copy_thresh;
int desc_num;
- boolean_t copy_done, eop;
- mblk_t *current_mp, *next_mp, *nmp, *pull_mp = NULL;
- tx_control_block_t *tcb;
- ixgbe_tx_context_t tx_context, *ctx;
+ ixgbe_tx_context_t tx_context, *ctx = NULL;
link_list_t pending_list;
- uint32_t len, hdr_frag_len, hdr_len;
- uint32_t copy_thresh;
- mblk_t *hdr_new_mp = NULL;
- mblk_t *hdr_pre_mp = NULL;
- mblk_t *hdr_nmp = NULL;
+ boolean_t limit_retry = B_FALSE;
ASSERT(mp->b_next == NULL);
@@ -115,11 +146,7 @@ ixgbe_ring_tx(void *arg, mblk_t *mp)
copy_thresh = ixgbe->tx_copy_thresh;
- /* Get the mblk size */
- mbsize = 0;
- for (nmp = mp; nmp != NULL; nmp = nmp->b_cont) {
- mbsize += MBLKL(nmp);
- }
+ mbsize = msgsize(mp);
if (ixgbe->tx_hcksum_enable) {
/*
@@ -145,11 +172,16 @@ ixgbe_ring_tx(void *arg, mblk_t *mp)
IXGBE_DEBUGLOG_0(ixgbe, "ixgbe_tx: packet oversize");
return (NULL);
}
- } else {
- ctx = NULL;
}
/*
+ * If we use too many descriptors (see comments below), we may do
+ * pull_mp = msgpullup(orig_mp, -1), and jump back to here. As such,
+ * any time we error return past here, we should check and free
+ * pull_mp if != NULL.
+ */
+retry:
+ /*
* Check and recycle tx descriptors.
* The recycle threshold here should be selected carefully
*/
@@ -165,7 +197,9 @@ ixgbe_ring_tx(void *arg, mblk_t *mp)
if (tx_ring->tbd_free < ixgbe->tx_overload_thresh) {
tx_ring->reschedule = B_TRUE;
tx_ring->stat_overload++;
- return (mp);
+ if (pull_mp != NULL)
+ freemsg(pull_mp);
+ return (orig_mp);
}
/*
@@ -175,313 +209,149 @@ ixgbe_ring_tx(void *arg, mblk_t *mp)
* It is used to reduce the lock contention of the tx_lock.
*/
LINK_LIST_INIT(&pending_list);
+
+ tcb = NULL;
desc_num = 0;
desc_total = 0;
+ offset = 0;
/*
- * The software should guarantee LSO packet header(MAC+IP+TCP)
- * to be within one descriptor. Here we reallocate and refill the
- * the header if it's physical memory non-contiguous.
+ * For LSO, we always copy the packet header (Ethernet + IP + TCP/UDP)
+ * into a single descriptor separate from the remaining data.
*/
if ((ctx != NULL) && ctx->lso_flag) {
- /* find the last fragment of the header */
- len = MBLKL(mp);
- ASSERT(len > 0);
- hdr_nmp = mp;
+ size_t hdr_len;
+
hdr_len = ctx->ip_hdr_len + ctx->mac_hdr_len + ctx->l4_hdr_len;
- while (len < hdr_len) {
- hdr_pre_mp = hdr_nmp;
- hdr_nmp = hdr_nmp->b_cont;
- len += MBLKL(hdr_nmp);
- }
- /*
- * If the header and the payload are in different mblks,
- * we simply force the header to be copied into pre-allocated
- * page-aligned buffer.
- */
- if (len == hdr_len)
- goto adjust_threshold;
- hdr_frag_len = hdr_len - (len - MBLKL(hdr_nmp));
/*
- * There are two cases we need to reallocate a mblk for the
- * last header fragment:
- * 1. the header is in multiple mblks and the last fragment
- * share the same mblk with the payload
- * 2. the header is in a single mblk shared with the payload
- * and the header is physical memory non-contiguous
+ * copy the first hdr_len bytes of mp (i.e. the Ethernet, IP,
+ * and TCP/UDP headers) into tcb.
*/
- if ((hdr_nmp != mp) ||
- (P2NPHASE((uintptr_t)hdr_nmp->b_rptr, ixgbe->sys_page_size)
- < hdr_len)) {
- tx_ring->stat_lso_header_fail++;
+ for (len = hdr_len; mp != NULL && len > 0; mp = mp->b_cont) {
+ size_t mlen = MBLKL(mp);
+ size_t amt = MIN(mlen, len);
+ int ret;
+
+ ret = ixgbe_tx_copy(tx_ring, &tcb, &pending_list,
+ mp->b_rptr, amt);
+ /*
+ * Since we're trying to copy all of the headers into
+ * a single buffer in a single tcb, if ixgbe_tx_copy()
+ * returns anything but 0, it means either no tcbs
+ * are available (< 0), or while copying, we spilled
+ * over and couldn't fit all the headers into a
+ * single tcb.
+ */
+ if (ret != 0) {
+ if (ret > 0)
+ tx_ring->stat_lso_header_fail++;
+ goto tx_failure;
+ }
+
+ len -= amt;
+
/*
- * reallocate the mblk for the last header fragment,
- * expect to bcopy into pre-allocated page-aligned
- * buffer
+ * If we copy less than the full amount of this
+ * mblk_t, we have some amount to copy below.
*/
- hdr_new_mp = allocb(hdr_frag_len, 0);
- if (!hdr_new_mp)
- return (mp);
- bcopy(hdr_nmp->b_rptr, hdr_new_mp->b_rptr,
- hdr_frag_len);
- /* link the new header fragment with the other parts */
- hdr_new_mp->b_wptr = hdr_new_mp->b_rptr + hdr_frag_len;
- hdr_new_mp->b_cont = hdr_nmp;
- if (hdr_pre_mp)
- hdr_pre_mp->b_cont = hdr_new_mp;
- else
- mp = hdr_new_mp;
- hdr_nmp->b_rptr += hdr_frag_len;
+ if (amt < mlen) {
+ offset = amt;
+ break;
+ }
}
-adjust_threshold:
+
+ ASSERT0(len);
+
/*
- * adjust the bcopy threshhold to guarantee
- * the header to use bcopy way
+ * Finish off the header tcb, and start anew for the
+ * rest of the packet.
*/
- if (copy_thresh < hdr_len)
- copy_thresh = hdr_len;
+ desc_total += ixgbe_tcb_done(tcb);
+ tcb = NULL;
}
- current_mp = mp;
- current_len = MBLKL(current_mp);
- /*
- * Decide which method to use for the first fragment
- */
- current_flag = (current_len <= copy_thresh) ?
- USE_COPY : USE_DMA;
/*
- * If the mblk includes several contiguous small fragments,
- * they may be copied into one buffer. This flag is used to
- * indicate whether there are pending fragments that need to
- * be copied to the current tx buffer.
+ * Process each remaining segment in the packet -- either binding
+ * the dblk_t or copying the contents of the dblk_t to an already
+ * bound buffer. When we copy, we will accumulate consecutive small
+ * (less than copy_thresh bytes) segments into a single tcb buffer
+ * until no more can fit (or we encounter a segment larger than
+ * copy_thresh and bind the dblk_t).
*
- * If this flag is B_TRUE, it indicates that a new tx control
- * block is needed to process the next fragment using either
- * copy or DMA binding.
+ * Both ixgbe_tx_bind() and ixgbe_tx_copy() will allocate new
+ * transmit control blocks (tcb)s as needed (and append them onto
+ * 'pending_list'). Both functions also replace 'tcb' with the new
+ * tcb when they allocate a new tcb.
*
- * Otherwise, it indicates that the next fragment will be
- * copied to the current tx buffer that is maintained by the
- * current tx control block. No new tx control block is needed.
+ * We stop trying to process the packet once the number of descriptors
+ * used equals IXGBE_TX_DESC_LIMIT. Even if we're copying into the
+ * IXGBE_TX_DESC_LIMIT-th descriptor, we won't have room to add a
+ * context descriptor (since we're already at the limit), so there's
+ * no point in continuing. We'll pull up the mblk_t (see below)
+ * and try again.
*/
- copy_done = B_TRUE;
- while (current_mp) {
- next_mp = current_mp->b_cont;
- eop = (next_mp == NULL); /* Last fragment of the packet? */
- next_len = eop ? 0: MBLKL(next_mp);
-
- /*
- * When the current fragment is an empty fragment, if
- * the next fragment will still be copied to the current
- * tx buffer, we cannot skip this fragment here. Because
- * the copy processing is pending for completion. We have
- * to process this empty fragment in the tx_copy routine.
- *
- * If the copy processing is completed or a DMA binding
- * processing is just completed, we can just skip this
- * empty fragment.
- */
- if ((current_len == 0) && (copy_done)) {
- current_mp = next_mp;
- current_len = next_len;
- current_flag = (current_len <= copy_thresh) ?
- USE_COPY : USE_DMA;
- continue;
- }
-
- if (copy_done) {
- /*
- * Get a new tx control block from the free list
- */
- tcb = ixgbe_get_free_list(tx_ring);
-
- if (tcb == NULL) {
- tx_ring->stat_fail_no_tcb++;
- goto tx_failure;
- }
-
- /*
- * Push the tx control block to the pending list
- * to avoid using lock too early
- */
- LIST_PUSH_TAIL(&pending_list, &tcb->link);
- }
-
- if (current_flag == USE_COPY) {
- /*
- * Check whether to use bcopy or DMA binding to process
- * the next fragment, and if using bcopy, whether we
- * need to continue copying the next fragment into the
- * current tx buffer.
- */
- ASSERT((tcb->tx_buf.len + current_len) <=
- tcb->tx_buf.size);
+ while (mp != NULL && desc_total < IXGBE_TX_DESC_LIMIT) {
+ uint8_t *rptr = mp->b_rptr + offset;
+ int ret;
- if (eop) {
- /*
- * This is the last fragment of the packet, so
- * the copy processing will be completed with
- * this fragment.
- */
- next_flag = USE_NONE;
- copy_done = B_TRUE;
- } else if ((tcb->tx_buf.len + current_len + next_len) >
- tcb->tx_buf.size) {
- /*
- * If the next fragment is too large to be
- * copied to the current tx buffer, we need
- * to complete the current copy processing.
- */
- next_flag = (next_len > copy_thresh) ?
- USE_DMA: USE_COPY;
- copy_done = B_TRUE;
- } else if (next_len > copy_thresh) {
- /*
- * The next fragment needs to be processed with
- * DMA binding. So the copy prcessing will be
- * completed with the current fragment.
- */
- next_flag = USE_DMA;
- copy_done = B_TRUE;
- } else {
- /*
- * Continue to copy the next fragment to the
- * current tx buffer.
- */
- next_flag = USE_COPY;
- copy_done = B_FALSE;
- }
+ len = MBLKL(mp) - offset;
+ offset = 0;
- desc_num = ixgbe_tx_copy(tx_ring, tcb, current_mp,
- current_len, copy_done);
+ if (len > copy_thresh) {
+ ret = ixgbe_tx_bind(tx_ring, &tcb, &pending_list, rptr,
+ len);
} else {
- /*
- * Check whether to use bcopy or DMA binding to process
- * the next fragment.
- */
- next_flag = (next_len > copy_thresh) ?
- USE_DMA: USE_COPY;
- ASSERT(copy_done == B_TRUE);
-
- desc_num = ixgbe_tx_bind(tx_ring, tcb, current_mp,
- current_len);
+ ret = ixgbe_tx_copy(tx_ring, &tcb, &pending_list, rptr,
+ len);
}
- if (desc_num > 0)
- desc_total += desc_num;
- else if (desc_num < 0)
+ if (ret < 0)
goto tx_failure;
- current_mp = next_mp;
- current_len = next_len;
- current_flag = next_flag;
+ desc_total += ret;
+ mp = mp->b_cont;
}
- /*
- * Attach the mblk to the last tx control block
- */
- ASSERT(tcb);
- ASSERT(tcb->mp == NULL);
- tcb->mp = mp;
+ /* Finish off the last tcb */
+ desc_total += ixgbe_tcb_done(tcb);
/*
* 82598/82599 chipset has a limitation that no more than 32 tx
- * descriptors can be transmited out at one time.
+ * descriptors can be transmited out at one time. As noted above,
+ * we need to include space for a context descriptor in case its
+ * necessary, so we do this even if desc_total == IXGBE_TX_DESC_LIMIT
+ * as well as when it exceeds the limit.
*
- * Here is a workaround for it: pull up the mblk then send it
- * out with bind way. By doing so, no more than MAX_COOKIE (18)
- * descriptors is needed.
+ * If we exceed this limit, we take the hit, do a msgpullup(), and
+ * then try again. Our DMA attributes guarantee we should never use
+ * more than MAX_COOKIE (18) descriptors to map a single mblk_t, so we
+ * should only need to retry once.
*/
- if (desc_total + 1 > IXGBE_TX_DESC_LIMIT) {
- tx_ring->stat_break_tbd_limit++;
+ if (desc_total >= IXGBE_TX_DESC_LIMIT) {
+ /* We shouldn't hit this path twice */
+ VERIFY0(limit_retry);
- /*
- * Discard the mblk and free the used resources
- */
- tcb = (tx_control_block_t *)LIST_GET_HEAD(&pending_list);
- while (tcb) {
- tcb->mp = NULL;
- ixgbe_free_tcb(tcb);
- tcb = (tx_control_block_t *)
- LIST_GET_NEXT(&pending_list, &tcb->link);
- }
+ tx_ring->stat_break_tbd_limit++;
- /*
- * Return the tx control blocks in the pending list to
- * the free list.
- */
+ /* Release all the tcbs we used previously */
ixgbe_put_free_list(tx_ring, &pending_list);
-
- /*
- * pull up the mblk and send it out with bind way
- */
- if ((pull_mp = msgpullup(mp, -1)) == NULL) {
- tx_ring->reschedule = B_TRUE;
-
- /*
- * If new mblk has been allocted for the last header
- * fragment of a LSO packet, we should restore the
- * modified mp.
- */
- if (hdr_new_mp) {
- hdr_new_mp->b_cont = NULL;
- freeb(hdr_new_mp);
- hdr_nmp->b_rptr -= hdr_frag_len;
- if (hdr_pre_mp)
- hdr_pre_mp->b_cont = hdr_nmp;
- else
- mp = hdr_nmp;
- }
- return (mp);
- }
-
- LINK_LIST_INIT(&pending_list);
desc_total = 0;
+ offset = 0;
- /*
- * if the packet is a LSO packet, we simply
- * transmit the header in one descriptor using the copy way
- */
- if ((ctx != NULL) && ctx->lso_flag) {
- hdr_len = ctx->ip_hdr_len + ctx->mac_hdr_len +
- ctx->l4_hdr_len;
-
- tcb = ixgbe_get_free_list(tx_ring);
- if (tcb == NULL) {
- tx_ring->stat_fail_no_tcb++;
- goto tx_failure;
- }
- desc_num = ixgbe_tx_copy(tx_ring, tcb, pull_mp,
- hdr_len, B_TRUE);
- LIST_PUSH_TAIL(&pending_list, &tcb->link);
- desc_total += desc_num;
-
- pull_mp->b_rptr += hdr_len;
- }
-
- tcb = ixgbe_get_free_list(tx_ring);
- if (tcb == NULL) {
- tx_ring->stat_fail_no_tcb++;
- goto tx_failure;
- }
- if ((ctx != NULL) && ctx->lso_flag) {
- desc_num = ixgbe_tx_bind(tx_ring, tcb, pull_mp,
- mbsize - hdr_len);
- } else {
- desc_num = ixgbe_tx_bind(tx_ring, tcb, pull_mp,
- mbsize);
- }
- if (desc_num < 0) {
- goto tx_failure;
+ pull_mp = msgpullup(orig_mp, -1);
+ if (pull_mp == NULL) {
+ tx_ring->reschedule = B_TRUE;
+ return (orig_mp);
}
- LIST_PUSH_TAIL(&pending_list, &tcb->link);
- desc_total += desc_num;
- tcb->mp = pull_mp;
+ mp = pull_mp;
+ limit_retry = B_TRUE;
+ goto retry;
}
/*
- * Before fill the tx descriptor ring with the data, we need to
+ * Before filling the tx descriptor ring with the data, we need to
* ensure there are adequate free descriptors for transmit
* (including one context descriptor).
* Do not use up all the tx descriptors.
@@ -506,22 +376,31 @@ adjust_threshold:
goto tx_failure;
}
+ /*
+ * Attach the mblk_t we've setup to the last control block.
+ * This is only done once we know there are enough free descriptors
+ * to transmit so that the cleanup in tx_failure doesn't try to
+ * call freemsg() on mp (since we will want to return it).
+ */
+ tcb->mp = (pull_mp != NULL) ? pull_mp : orig_mp;
+
desc_num = ixgbe_tx_fill_ring(tx_ring, &pending_list, ctx,
mbsize);
ASSERT((desc_num == desc_total) || (desc_num == (desc_total + 1)));
tx_ring->stat_obytes += mbsize;
- tx_ring->stat_opackets ++;
+ tx_ring->stat_opackets++;
mutex_exit(&tx_ring->tx_lock);
/*
- * now that the transmission succeeds, need to free the original
- * mp if we used the pulling up mblk for transmission.
+ * Now that tx is done, if we pulled up the original message, we
+ * can free the original message since it is no longer being
+ * used.
*/
- if (pull_mp) {
- freemsg(mp);
+ if (pull_mp != NULL) {
+ freemsg(orig_mp);
}
return (NULL);
@@ -535,31 +414,10 @@ tx_failure:
}
/*
- * If new mblk has been allocted for the last header
- * fragment of a LSO packet, we should restore the
- * modified mp.
- */
- if (hdr_new_mp) {
- hdr_new_mp->b_cont = NULL;
- freeb(hdr_new_mp);
- hdr_nmp->b_rptr -= hdr_frag_len;
- if (hdr_pre_mp)
- hdr_pre_mp->b_cont = hdr_nmp;
- else
- mp = hdr_nmp;
- }
- /*
- * Discard the mblk and free the used resources
+ * tcb->mp should not be set until we know we can transmit (see above),
+ * so it should always be NULL if we get here.
*/
- tcb = (tx_control_block_t *)LIST_GET_HEAD(&pending_list);
- while (tcb) {
- tcb->mp = NULL;
-
- ixgbe_free_tcb(tcb);
-
- tcb = (tx_control_block_t *)
- LIST_GET_NEXT(&pending_list, &tcb->link);
- }
+ VERIFY3P(tcb->mp, ==, NULL);
/*
* Return the tx control blocks in the pending list to the free list.
@@ -569,22 +427,44 @@ tx_failure:
/* Transmit failed, do not drop the mblk, rechedule the transmit */
tx_ring->reschedule = B_TRUE;
- return (mp);
+ return (orig_mp);
}
/*
* ixgbe_tx_copy
*
- * Copy the mblk fragment to the pre-allocated tx buffer
+ * Copy the mblk fragment to the pre-allocated tx buffer. Return -1 on error,
+ * otherwise return the number of descriptors we've completed in this call.
*/
static int
-ixgbe_tx_copy(ixgbe_tx_ring_t *tx_ring, tx_control_block_t *tcb, mblk_t *mp,
- uint32_t len, boolean_t copy_done)
+ixgbe_tx_copy(ixgbe_tx_ring_t *tx_ring, tx_control_block_t **tcbp,
+ link_list_t *pending_list, const void *buf, size_t len)
{
+ tx_control_block_t *tcb = *tcbp;
dma_buffer_t *tx_buf;
- uint32_t desc_num;
- _NOTE(ARGUNUSED(tx_ring));
+ uint32_t desc_num = 0;
+
+ /*
+ * We need a new tcb -- either the current one (tcb) is NULL because
+ * we just started, tcb is being used for DMA, or tcb isn't large enough
+ * to hold the contents we need to copy.
+ */
+ if (tcb == NULL || tcb->tx_type == USE_DMA ||
+ tcb->tx_buf.len + len > tcb->tx_buf.size) {
+ tx_control_block_t *newtcb;
+
+ newtcb = ixgbe_get_free_list(tx_ring, pending_list);
+ if (newtcb == NULL)
+ return (-1);
+ newtcb->tx_type = USE_COPY;
+
+ if (tcb != NULL)
+ desc_num += ixgbe_tcb_done(tcb);
+ *tcbp = tcb = newtcb;
+ }
+
+ ASSERT3S(tcb->tx_type, ==, USE_COPY);
tx_buf = &tcb->tx_buf;
/*
@@ -598,62 +478,43 @@ ixgbe_tx_copy(ixgbe_tx_ring_t *tx_ring, tx_control_block_t *tcb, mblk_t *mp,
* fragment.
*/
if (len > 0) {
- bcopy(mp->b_rptr, tx_buf->address + tx_buf->len, len);
+ bcopy(buf, tx_buf->address + tx_buf->len, len);
tx_buf->len += len;
tcb->frag_num++;
}
- desc_num = 0;
-
- /*
- * If it is the last fragment copied to the current tx buffer,
- * in other words, if there's no remaining fragment or the remaining
- * fragment requires a new tx control block to process, we need to
- * complete the current copy processing by syncing up the current
- * DMA buffer and saving the descriptor data.
- */
- if (copy_done) {
- /*
- * Sync the DMA buffer of the packet data
- */
- DMA_SYNC(tx_buf, DDI_DMA_SYNC_FORDEV);
-
- tcb->tx_type = USE_COPY;
-
- /*
- * Save the address and length to the private data structure
- * of the tx control block, which will be used to fill the
- * tx descriptor ring after all the fragments are processed.
- */
- ixgbe_save_desc(tcb, tx_buf->dma_address, tx_buf->len);
- desc_num++;
- }
-
return (desc_num);
}
/*
* ixgbe_tx_bind
*
- * Bind the mblk fragment with DMA
+ * Bind the mblk fragment with DMA. Returns -1 on error, otherwise it
+ * returns the number of descriptors completed in this call. This count
+ * can include descriptors that weren't filled in by the current call to
+ * ixgbe_tx_bind() but were being used (but not yet completed) in previous
+ * calls to ixgbe_tx_bind() or ixgbe_tx_copy().
*/
static int
-ixgbe_tx_bind(ixgbe_tx_ring_t *tx_ring, tx_control_block_t *tcb, mblk_t *mp,
- uint32_t len)
+ixgbe_tx_bind(ixgbe_tx_ring_t *tx_ring, tx_control_block_t **tcbp,
+ link_list_t *pending_list, uint8_t *buf, size_t len)
{
- int status, i;
- ddi_dma_cookie_t dma_cookie;
- uint_t ncookies;
- int desc_num;
+ tx_control_block_t *tcb = NULL;
+ uint_t desc_num = 0;
+ int status;
+
+ tcb = ixgbe_get_free_list(tx_ring, pending_list);
+ if (tcb == NULL)
+ return (-1);
/*
* Use DMA binding to process the mblk fragment
*/
status = ddi_dma_addr_bind_handle(tcb->tx_dma_handle, NULL,
- (caddr_t)mp->b_rptr, len,
+ (caddr_t)buf, len,
DDI_DMA_WRITE | DDI_DMA_STREAMING, DDI_DMA_DONTWAIT,
- 0, &dma_cookie, &ncookies);
+ 0, NULL, NULL);
if (status != DDI_DMA_MAPPED) {
tx_ring->stat_fail_dma_bind++;
@@ -662,25 +523,51 @@ ixgbe_tx_bind(ixgbe_tx_ring_t *tx_ring, tx_control_block_t *tcb, mblk_t *mp,
tcb->frag_num++;
tcb->tx_type = USE_DMA;
+
/*
- * Each fragment can span several cookies. One cookie will have
- * one tx descriptor to transmit.
+ * If there was an old tcb, we're about to replace it. Finish
+ * setting up the old tcb so we can replace it with the new one.
*/
- desc_num = 0;
- for (i = ncookies; i > 0; i--) {
- /*
- * Save the address and length to the private data structure
- * of the tx control block, which will be used to fill the
- * tx descriptor ring after all the fragments are processed.
- */
- ixgbe_save_desc(tcb,
- dma_cookie.dmac_laddress,
- dma_cookie.dmac_size);
+ if (*tcbp != NULL)
+ desc_num += ixgbe_tcb_done(*tcbp);
- desc_num++;
+ *tcbp = tcb;
+ return (desc_num);
+}
+
+/*
+ * Once we're done populating a tcb (either by binding or copying into
+ * a buffer in the tcb), get it ready for tx and return the number of
+ * descriptors used.
+ */
+static uint_t
+ixgbe_tcb_done(tx_control_block_t *tcb)
+{
+ uint_t desc_num = 0;
+
+ if (tcb->tx_type == USE_DMA) {
+ const ddi_dma_cookie_t *c;
- if (i > 1)
- ddi_dma_nextcookie(tcb->tx_dma_handle, &dma_cookie);
+ for (c = ddi_dma_cookie_iter(tcb->tx_dma_handle, NULL);
+ c != NULL;
+ c = ddi_dma_cookie_iter(tcb->tx_dma_handle, c)) {
+ /*
+ * Save the address and length to the private data
+ * structure of the tx control block, which will be
+ * used to fill the tx descriptor ring after all the
+ * fragments are processed.
+ */
+ ixgbe_save_desc(tcb, c->dmac_laddress, c->dmac_size);
+ desc_num++;
+ }
+ } else if (tcb->tx_type == USE_COPY) {
+ dma_buffer_t *tx_buf = &tcb->tx_buf;
+
+ DMA_SYNC(tx_buf, DDI_DMA_SYNC_FORDEV);
+ ixgbe_save_desc(tcb, tx_buf->dma_address, tx_buf->len);
+ desc_num++;
+ } else {
+ panic("invalid tcb type");
}
return (desc_num);
@@ -1364,21 +1251,6 @@ ixgbe_tx_recycle_legacy(ixgbe_tx_ring_t *tx_ring)
mutex_exit(&tx_ring->recycle_lock);
/*
- * Free the resources used by the tx control blocks
- * in the pending list
- */
- tcb = (tx_control_block_t *)LIST_GET_HEAD(&pending_list);
- while (tcb != NULL) {
- /*
- * Release the resources occupied by the tx control block
- */
- ixgbe_free_tcb(tcb);
-
- tcb = (tx_control_block_t *)
- LIST_GET_NEXT(&pending_list, &tcb->link);
- }
-
- /*
* Add the tx control blocks in the pending list to the free list.
*/
ixgbe_put_free_list(tx_ring, &pending_list);
@@ -1511,21 +1383,6 @@ ixgbe_tx_recycle_head_wb(ixgbe_tx_ring_t *tx_ring)
mutex_exit(&tx_ring->recycle_lock);
/*
- * Free the resources used by the tx control blocks
- * in the pending list
- */
- tcb = (tx_control_block_t *)LIST_GET_HEAD(&pending_list);
- while (tcb) {
- /*
- * Release the resources occupied by the tx control block
- */
- ixgbe_free_tcb(tcb);
-
- tcb = (tx_control_block_t *)
- LIST_GET_NEXT(&pending_list, &tcb->link);
- }
-
- /*
* Add the tx control blocks in the pending list to the free list.
*/
ixgbe_put_free_list(tx_ring, &pending_list);
@@ -1543,6 +1400,9 @@ ixgbe_tx_recycle_head_wb(ixgbe_tx_ring_t *tx_ring)
void
ixgbe_free_tcb(tx_control_block_t *tcb)
{
+ if (tcb == NULL)
+ return;
+
switch (tcb->tx_type) {
case USE_COPY:
/*
@@ -1576,14 +1436,15 @@ ixgbe_free_tcb(tx_control_block_t *tcb)
}
/*
- * ixgbe_get_free_list - Get a free tx control block from the free list
+ * ixgbe_get_free_list - Get a free tx control block from the free list.
+ * Returns the tx control block and appends it to list.
*
* The atomic operation on the number of the available tx control block
* in the free list is used to keep this routine mutual exclusive with
* the routine ixgbe_put_check_list.
*/
static tx_control_block_t *
-ixgbe_get_free_list(ixgbe_tx_ring_t *tx_ring)
+ixgbe_get_free_list(ixgbe_tx_ring_t *tx_ring, link_list_t *list)
{
tx_control_block_t *tcb;
@@ -1591,8 +1452,10 @@ ixgbe_get_free_list(ixgbe_tx_ring_t *tx_ring)
* Check and update the number of the free tx control block
* in the free list.
*/
- if (ixgbe_atomic_reserve(&tx_ring->tcb_free, 1) < 0)
+ if (ixgbe_atomic_reserve(&tx_ring->tcb_free, 1) < 0) {
+ tx_ring->stat_fail_no_tcb++;
return (NULL);
+ }
mutex_enter(&tx_ring->tcb_head_lock);
@@ -1604,6 +1467,7 @@ ixgbe_get_free_list(ixgbe_tx_ring_t *tx_ring)
mutex_exit(&tx_ring->tcb_head_lock);
+ LIST_PUSH_TAIL(list, &tcb->link);
return (tcb);
}
@@ -1623,6 +1487,16 @@ ixgbe_put_free_list(ixgbe_tx_ring_t *tx_ring, link_list_t *pending_list)
int tcb_num;
tx_control_block_t *tcb;
+ for (tcb = (tx_control_block_t *)LIST_GET_HEAD(pending_list);
+ tcb != NULL;
+ tcb = (tx_control_block_t *)LIST_GET_NEXT(pending_list, tcb)) {
+ /*
+ * Despite the name, ixgbe_free_tcb() just releases the
+ * resources in tcb, but does not free tcb itself.
+ */
+ ixgbe_free_tcb(tcb);
+ }
+
mutex_enter(&tx_ring->tcb_tail_lock);
index = tx_ring->tcb_tail;
diff --git a/usr/src/uts/common/mapfiles/ddi.mapfile b/usr/src/uts/common/mapfiles/ddi.mapfile
index abbc802cd8..89d1073019 100644
--- a/usr/src/uts/common/mapfiles/ddi.mapfile
+++ b/usr/src/uts/common/mapfiles/ddi.mapfile
@@ -10,7 +10,7 @@
#
#
-# Copyright 2019 Joyent, Inc.
+# Copyright 2020 Joyent, Inc.
# Copyright 2020 RackTop Systems, Inc.
#
@@ -77,6 +77,7 @@ SYMBOL_SCOPE {
ddi_dev_regsize { FLAGS = EXTERN };
ddi_dma_addr_bind_handle { FLAGS = EXTERN };
ddi_dma_alloc_handle { FLAGS = EXTERN };
+ ddi_dma_cookie_iter { FLAGS = EXTERN };
ddi_dma_free_handle { FLAGS = EXTERN };
ddi_dma_mem_alloc { FLAGS = EXTERN };
ddi_dma_mem_free { FLAGS = EXTERN };