summaryrefslogtreecommitdiff
path: root/usr/src/tools/ctf/cvt/ctfmerge.c
diff options
context:
space:
mode:
Diffstat (limited to 'usr/src/tools/ctf/cvt/ctfmerge.c')
-rw-r--r--usr/src/tools/ctf/cvt/ctfmerge.c1004
1 files changed, 0 insertions, 1004 deletions
diff --git a/usr/src/tools/ctf/cvt/ctfmerge.c b/usr/src/tools/ctf/cvt/ctfmerge.c
deleted file mode 100644
index d2db789e5f..0000000000
--- a/usr/src/tools/ctf/cvt/ctfmerge.c
+++ /dev/null
@@ -1,1004 +0,0 @@
-/*
- * CDDL HEADER START
- *
- * The contents of this file are subject to the terms of the
- * Common Development and Distribution License (the "License").
- * You may not use this file except in compliance with the License.
- *
- * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
- * or http://www.opensolaris.org/os/licensing.
- * See the License for the specific language governing permissions
- * and limitations under the License.
- *
- * When distributing Covered Code, include this CDDL HEADER in each
- * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
- * If applicable, add the following below this CDDL HEADER, with the
- * fields enclosed by brackets "[]" replaced with your own identifying
- * information: Portions Copyright [yyyy] [name of copyright owner]
- *
- * CDDL HEADER END
- */
-/*
- * Copyright 2008 Sun Microsystems, Inc. All rights reserved.
- * Use is subject to license terms.
- */
-
-/*
- * Given several files containing CTF data, merge and uniquify that data into
- * a single CTF section in an output file.
- *
- * Merges can proceed independently. As such, we perform the merges in parallel
- * using a worker thread model. A given glob of CTF data (either all of the CTF
- * data from a single input file, or the result of one or more merges) can only
- * be involved in a single merge at any given time, so the process decreases in
- * parallelism, especially towards the end, as more and more files are
- * consolidated, finally resulting in a single merge of two large CTF graphs.
- * Unfortunately, the last merge is also the slowest, as the two graphs being
- * merged are each the product of merges of half of the input files.
- *
- * The algorithm consists of two phases, described in detail below. The first
- * phase entails the merging of CTF data in groups of eight. The second phase
- * takes the results of Phase I, and merges them two at a time. This disparity
- * is due to an observation that the merge time increases at least quadratically
- * with the size of the CTF data being merged. As such, merges of CTF graphs
- * newly read from input files are much faster than merges of CTF graphs that
- * are themselves the results of prior merges.
- *
- * A further complication is the need to ensure the repeatability of CTF merges.
- * That is, a merge should produce the same output every time, given the same
- * input. In both phases, this consistency requirement is met by imposing an
- * ordering on the merge process, thus ensuring that a given set of input files
- * are merged in the same order every time.
- *
- * Phase I
- *
- * The main thread reads the input files one by one, transforming the CTF
- * data they contain into tdata structures. When a given file has been read
- * and parsed, it is placed on the work queue for retrieval by worker threads.
- *
- * Central to Phase I is the Work In Progress (wip) array, which is used to
- * merge batches of files in a predictable order. Files are read by the main
- * thread, and are merged into wip array elements in round-robin order. When
- * the number of files merged into a given array slot equals the batch size,
- * the merged CTF graph in that array is added to the done slot in order by
- * array slot.
- *
- * For example, consider a case where we have five input files, a batch size
- * of two, a wip array size of two, and two worker threads (T1 and T2).
- *
- * 1. The wip array elements are assigned initial batch numbers 0 and 1.
- * 2. T1 reads an input file from the input queue (wq_queue). This is the
- * first input file, so it is placed into wip[0]. The second file is
- * similarly read and placed into wip[1]. The wip array slots now contain
- * one file each (wip_nmerged == 1).
- * 3. T1 reads the third input file, which it merges into wip[0]. The
- * number of files in wip[0] is equal to the batch size.
- * 4. T2 reads the fourth input file, which it merges into wip[1]. wip[1]
- * is now full too.
- * 5. T2 attempts to place the contents of wip[1] on the done queue
- * (wq_done_queue), but it can't, since the batch ID for wip[1] is 1.
- * Batch 0 needs to be on the done queue before batch 1 can be added, so
- * T2 blocks on wip[1]'s cv.
- * 6. T1 attempts to place the contents of wip[0] on the done queue, and
- * succeeds, updating wq_lastdonebatch to 0. It clears wip[0], and sets
- * its batch ID to 2. T1 then signals wip[1]'s cv to awaken T2.
- * 7. T2 wakes up, notices that wq_lastdonebatch is 0, which means that
- * batch 1 can now be added. It adds wip[1] to the done queue, clears
- * wip[1], and sets its batch ID to 3. It signals wip[0]'s cv, and
- * restarts.
- *
- * The above process continues until all input files have been consumed. At
- * this point, a pair of barriers are used to allow a single thread to move
- * any partial batches from the wip array to the done array in batch ID order.
- * When this is complete, wq_done_queue is moved to wq_queue, and Phase II
- * begins.
- *
- * Locking Semantics (Phase I)
- *
- * The input queue (wq_queue) and the done queue (wq_done_queue) are
- * protected by separate mutexes - wq_queue_lock and wq_done_queue. wip
- * array slots are protected by their own mutexes, which must be grabbed
- * before releasing the input queue lock. The wip array lock is dropped
- * when the thread restarts the loop. If the array slot was full, the
- * array lock will be held while the slot contents are added to the done
- * queue. The done queue lock is used to protect the wip slot cv's.
- *
- * The pow number is protected by the queue lock. The master batch ID
- * and last completed batch (wq_lastdonebatch) counters are protected *in
- * Phase I* by the done queue lock.
- *
- * Phase II
- *
- * When Phase II begins, the queue consists of the merged batches from the
- * first phase. Assume we have five batches:
- *
- * Q: a b c d e
- *
- * Using the same batch ID mechanism we used in Phase I, but without the wip
- * array, worker threads remove two entries at a time from the beginning of
- * the queue. These two entries are merged, and are added back to the tail
- * of the queue, as follows:
- *
- * Q: a b c d e # start
- * Q: c d e ab # a, b removed, merged, added to end
- * Q: e ab cd # c, d removed, merged, added to end
- * Q: cd eab # e, ab removed, merged, added to end
- * Q: cdeab # cd, eab removed, merged, added to end
- *
- * When one entry remains on the queue, with no merges outstanding, Phase II
- * finishes. We pre-determine the stopping point by pre-calculating the
- * number of nodes that will appear on the list. In the example above, the
- * number (wq_ninqueue) is 9. When ninqueue is 1, we conclude Phase II by
- * signaling the main thread via wq_done_cv.
- *
- * Locking Semantics (Phase II)
- *
- * The queue (wq_queue), ninqueue, and the master batch ID and last
- * completed batch counters are protected by wq_queue_lock. The done
- * queue and corresponding lock are unused in Phase II as is the wip array.
- *
- * Uniquification
- *
- * We want the CTF data that goes into a given module to be as small as
- * possible. For example, we don't want it to contain any type data that may
- * be present in another common module. As such, after creating the master
- * tdata_t for a given module, we can, if requested by the user, uniquify it
- * against the tdata_t from another module (genunix in the case of the SunOS
- * kernel). We perform a merge between the tdata_t for this module and the
- * tdata_t from genunix. Nodes found in this module that are not present in
- * genunix are added to a third tdata_t - the uniquified tdata_t.
- *
- * Additive Merges
- *
- * In some cases, for example if we are issuing a new version of a common
- * module in a patch, we need to make sure that the CTF data already present
- * in that module does not change. Changes to this data would void the CTF
- * data in any module that uniquified against the common module. To preserve
- * the existing data, we can perform what is known as an additive merge. In
- * this case, a final uniquification is performed against the CTF data in the
- * previous version of the module. The result will be the placement of new
- * and changed data after the existing data, thus preserving the existing type
- * ID space.
- *
- * Saving the result
- *
- * When the merges are complete, the resulting tdata_t is placed into the
- * output file, replacing the .SUNW_ctf section (if any) already in that file.
- *
- * The person who changes the merging thread code in this file without updating
- * this comment will not live to see the stock hit five.
- */
-
-#include <stdio.h>
-#include <stdlib.h>
-#include <unistd.h>
-#include <pthread.h>
-#include <assert.h>
-#include <synch.h>
-#include <signal.h>
-#include <libgen.h>
-#include <string.h>
-#include <errno.h>
-#include <alloca.h>
-#include <sys/param.h>
-#include <sys/types.h>
-#include <sys/mman.h>
-#include <sys/sysconf.h>
-
-#include "ctf_headers.h"
-#include "ctftools.h"
-#include "ctfmerge.h"
-#include "traverse.h"
-#include "memory.h"
-#include "fifo.h"
-#include "barrier.h"
-
-#pragma init(bigheap)
-
-#define MERGE_PHASE1_BATCH_SIZE 8
-#define MERGE_PHASE1_MAX_SLOTS 5
-#define MERGE_INPUT_THROTTLE_LEN 10
-
-const char *progname;
-static char *outfile = NULL;
-static char *tmpname = NULL;
-static int dynsym;
-int debug_level = DEBUG_LEVEL;
-static size_t maxpgsize = 0x400000;
-
-
-void
-usage(void)
-{
- (void) fprintf(stderr,
- "Usage: %s [-fstv] -l label | -L labelenv -o outfile file ...\n"
- " %s [-fstv] -l label | -L labelenv -o outfile -d uniqfile\n"
- " %*s [-D uniqlabel] file ...\n"
- " %s [-fstv] -l label | -L labelenv -o outfile -w withfile "
- "file ...\n"
- " %s -c srcfile destfile\n"
- "\n"
- " Note: if -L labelenv is specified and labelenv is not set in\n"
- " the environment, a default value is used.\n",
- progname, progname, strlen(progname), " ",
- progname, progname);
-}
-
-static void
-bigheap(void)
-{
- size_t big, *size;
- int sizes;
- struct memcntl_mha mha;
-
- /*
- * First, get the available pagesizes.
- */
- if ((sizes = getpagesizes(NULL, 0)) == -1)
- return;
-
- if (sizes == 1 || (size = alloca(sizeof (size_t) * sizes)) == NULL)
- return;
-
- if (getpagesizes(size, sizes) == -1)
- return;
-
- while (size[sizes - 1] > maxpgsize)
- sizes--;
-
- /* set big to the largest allowed page size */
- big = size[sizes - 1];
- if (big & (big - 1)) {
- /*
- * The largest page size is not a power of two for some
- * inexplicable reason; return.
- */
- return;
- }
-
- /*
- * Now, align our break to the largest page size.
- */
- if (brk((void *)((((uintptr_t)sbrk(0) - 1) & ~(big - 1)) + big)) != 0)
- return;
-
- /*
- * set the preferred page size for the heap
- */
- mha.mha_cmd = MHA_MAPSIZE_BSSBRK;
- mha.mha_flags = 0;
- mha.mha_pagesize = big;
-
- (void) memcntl(NULL, 0, MC_HAT_ADVISE, (caddr_t)&mha, 0, 0);
-}
-
-static void
-finalize_phase_one(workqueue_t *wq)
-{
- int startslot, i;
-
- /*
- * wip slots are cleared out only when maxbatchsz td's have been merged
- * into them. We're not guaranteed that the number of files we're
- * merging is a multiple of maxbatchsz, so there will be some partial
- * groups in the wip array. Move them to the done queue in batch ID
- * order, starting with the slot containing the next batch that would
- * have been placed on the done queue, followed by the others.
- * One thread will be doing this while the others wait at the barrier
- * back in worker_thread(), so we don't need to worry about pesky things
- * like locks.
- */
-
- for (startslot = -1, i = 0; i < wq->wq_nwipslots; i++) {
- if (wq->wq_wip[i].wip_batchid == wq->wq_lastdonebatch + 1) {
- startslot = i;
- break;
- }
- }
-
- assert(startslot != -1);
-
- for (i = startslot; i < startslot + wq->wq_nwipslots; i++) {
- int slotnum = i % wq->wq_nwipslots;
- wip_t *wipslot = &wq->wq_wip[slotnum];
-
- if (wipslot->wip_td != NULL) {
- debug(2, "clearing slot %d (%d) (saving %d)\n",
- slotnum, i, wipslot->wip_nmerged);
- } else
- debug(2, "clearing slot %d (%d)\n", slotnum, i);
-
- if (wipslot->wip_td != NULL) {
- fifo_add(wq->wq_donequeue, wipslot->wip_td);
- wq->wq_wip[slotnum].wip_td = NULL;
- }
- }
-
- wq->wq_lastdonebatch = wq->wq_next_batchid++;
-
- debug(2, "phase one done: donequeue has %d items\n",
- fifo_len(wq->wq_donequeue));
-}
-
-static void
-init_phase_two(workqueue_t *wq)
-{
- int num;
-
- /*
- * We're going to continually merge the first two entries on the queue,
- * placing the result on the end, until there's nothing left to merge.
- * At that point, everything will have been merged into one. The
- * initial value of ninqueue needs to be equal to the total number of
- * entries that will show up on the queue, both at the start of the
- * phase and as generated by merges during the phase.
- */
- wq->wq_ninqueue = num = fifo_len(wq->wq_donequeue);
- while (num != 1) {
- wq->wq_ninqueue += num / 2;
- num = num / 2 + num % 2;
- }
-
- /*
- * Move the done queue to the work queue. We won't be using the done
- * queue in phase 2.
- */
- assert(fifo_len(wq->wq_queue) == 0);
- fifo_free(wq->wq_queue, NULL);
- wq->wq_queue = wq->wq_donequeue;
-}
-
-static void
-wip_save_work(workqueue_t *wq, wip_t *slot, int slotnum)
-{
- pthread_mutex_lock(&wq->wq_donequeue_lock);
-
- while (wq->wq_lastdonebatch + 1 < slot->wip_batchid)
- pthread_cond_wait(&slot->wip_cv, &wq->wq_donequeue_lock);
- assert(wq->wq_lastdonebatch + 1 == slot->wip_batchid);
-
- fifo_add(wq->wq_donequeue, slot->wip_td);
- wq->wq_lastdonebatch++;
- pthread_cond_signal(&wq->wq_wip[(slotnum + 1) %
- wq->wq_nwipslots].wip_cv);
-
- /* reset the slot for next use */
- slot->wip_td = NULL;
- slot->wip_batchid = wq->wq_next_batchid++;
-
- pthread_mutex_unlock(&wq->wq_donequeue_lock);
-}
-
-static void
-wip_add_work(wip_t *slot, tdata_t *pow)
-{
- if (slot->wip_td == NULL) {
- slot->wip_td = pow;
- slot->wip_nmerged = 1;
- } else {
- debug(2, "%d: merging %p into %p\n", pthread_self(),
- (void *)pow, (void *)slot->wip_td);
-
- merge_into_master(pow, slot->wip_td, NULL, 0);
- tdata_free(pow);
-
- slot->wip_nmerged++;
- }
-}
-
-static void
-worker_runphase1(workqueue_t *wq)
-{
- wip_t *wipslot;
- tdata_t *pow;
- int wipslotnum, pownum;
-
- for (;;) {
- pthread_mutex_lock(&wq->wq_queue_lock);
-
- while (fifo_empty(wq->wq_queue)) {
- if (wq->wq_nomorefiles == 1) {
- pthread_cond_broadcast(&wq->wq_work_avail);
- pthread_mutex_unlock(&wq->wq_queue_lock);
-
- /* on to phase 2 ... */
- return;
- }
-
- pthread_cond_wait(&wq->wq_work_avail,
- &wq->wq_queue_lock);
- }
-
- /* there's work to be done! */
- pow = fifo_remove(wq->wq_queue);
- pownum = wq->wq_nextpownum++;
- pthread_cond_broadcast(&wq->wq_work_removed);
-
- assert(pow != NULL);
-
- /* merge it into the right slot */
- wipslotnum = pownum % wq->wq_nwipslots;
- wipslot = &wq->wq_wip[wipslotnum];
-
- pthread_mutex_lock(&wipslot->wip_lock);
-
- pthread_mutex_unlock(&wq->wq_queue_lock);
-
- wip_add_work(wipslot, pow);
-
- if (wipslot->wip_nmerged == wq->wq_maxbatchsz)
- wip_save_work(wq, wipslot, wipslotnum);
-
- pthread_mutex_unlock(&wipslot->wip_lock);
- }
-}
-
-static void
-worker_runphase2(workqueue_t *wq)
-{
- tdata_t *pow1, *pow2;
- int batchid;
-
- for (;;) {
- pthread_mutex_lock(&wq->wq_queue_lock);
-
- if (wq->wq_ninqueue == 1) {
- pthread_cond_broadcast(&wq->wq_work_avail);
- pthread_mutex_unlock(&wq->wq_queue_lock);
-
- debug(2, "%d: entering p2 completion barrier\n",
- pthread_self());
- if (barrier_wait(&wq->wq_bar1)) {
- pthread_mutex_lock(&wq->wq_queue_lock);
- wq->wq_alldone = 1;
- pthread_cond_signal(&wq->wq_alldone_cv);
- pthread_mutex_unlock(&wq->wq_queue_lock);
- }
-
- return;
- }
-
- if (fifo_len(wq->wq_queue) < 2) {
- pthread_cond_wait(&wq->wq_work_avail,
- &wq->wq_queue_lock);
- pthread_mutex_unlock(&wq->wq_queue_lock);
- continue;
- }
-
- /* there's work to be done! */
- pow1 = fifo_remove(wq->wq_queue);
- pow2 = fifo_remove(wq->wq_queue);
- wq->wq_ninqueue -= 2;
-
- batchid = wq->wq_next_batchid++;
-
- pthread_mutex_unlock(&wq->wq_queue_lock);
-
- debug(2, "%d: merging %p into %p\n", pthread_self(),
- (void *)pow1, (void *)pow2);
- merge_into_master(pow1, pow2, NULL, 0);
- tdata_free(pow1);
-
- /*
- * merging is complete. place at the tail of the queue in
- * proper order.
- */
- pthread_mutex_lock(&wq->wq_queue_lock);
- while (wq->wq_lastdonebatch + 1 != batchid) {
- pthread_cond_wait(&wq->wq_done_cv,
- &wq->wq_queue_lock);
- }
-
- wq->wq_lastdonebatch = batchid;
-
- fifo_add(wq->wq_queue, pow2);
- debug(2, "%d: added %p to queue, len now %d, ninqueue %d\n",
- pthread_self(), (void *)pow2, fifo_len(wq->wq_queue),
- wq->wq_ninqueue);
- pthread_cond_broadcast(&wq->wq_done_cv);
- pthread_cond_signal(&wq->wq_work_avail);
- pthread_mutex_unlock(&wq->wq_queue_lock);
- }
-}
-
-/*
- * Main loop for worker threads.
- */
-static void *
-worker_thread(void *ptr)
-{
- workqueue_t *wq = ptr;
-
- worker_runphase1(wq);
-
- debug(2, "%d: entering first barrier\n", pthread_self());
-
- if (barrier_wait(&wq->wq_bar1)) {
-
- debug(2, "%d: doing work in first barrier\n", pthread_self());
-
- finalize_phase_one(wq);
-
- init_phase_two(wq);
-
- debug(2, "%d: ninqueue is %d, %d on queue\n", pthread_self(),
- wq->wq_ninqueue, fifo_len(wq->wq_queue));
- }
-
- debug(2, "%d: entering second barrier\n", pthread_self());
-
- (void) barrier_wait(&wq->wq_bar2);
-
- debug(2, "%d: phase 1 complete\n", pthread_self());
-
- worker_runphase2(wq);
- return (NULL);
-}
-
-/*
- * Pass a tdata_t tree, built from an input file, off to the work queue for
- * consumption by worker threads.
- */
-static int
-merge_ctf_cb(tdata_t *td, char *name, void *arg)
-{
- workqueue_t *wq = arg;
-
- debug(3, "Adding tdata %p for processing\n", (void *)td);
-
- pthread_mutex_lock(&wq->wq_queue_lock);
- while (fifo_len(wq->wq_queue) > wq->wq_ithrottle) {
- debug(2, "Throttling input (len = %d, throttle = %d)\n",
- fifo_len(wq->wq_queue), wq->wq_ithrottle);
- pthread_cond_wait(&wq->wq_work_removed, &wq->wq_queue_lock);
- }
-
- fifo_add(wq->wq_queue, td);
- debug(1, "Thread %d announcing %s\n", pthread_self(), name);
- pthread_cond_broadcast(&wq->wq_work_avail);
- pthread_mutex_unlock(&wq->wq_queue_lock);
-
- return (1);
-}
-
-/*
- * This program is intended to be invoked from a Makefile, as part of the build.
- * As such, in the event of a failure or user-initiated interrupt (^C), we need
- * to ensure that a subsequent re-make will cause ctfmerge to be executed again.
- * Unfortunately, ctfmerge will usually be invoked directly after (and as part
- * of the same Makefile rule as) a link, and will operate on the linked file
- * in place. If we merely exit upon receipt of a SIGINT, a subsequent make
- * will notice that the *linked* file is newer than the object files, and thus
- * will not reinvoke ctfmerge. The only way to ensure that a subsequent make
- * reinvokes ctfmerge, is to remove the file to which we are adding CTF
- * data (confusingly named the output file). This means that the link will need
- * to happen again, but links are generally fast, and we can't allow the merge
- * to be skipped.
- *
- * Another possibility would be to block SIGINT entirely - to always run to
- * completion. The run time of ctfmerge can, however, be measured in minutes
- * in some cases, so this is not a valid option.
- */
-static void
-handle_sig(int sig)
-{
- terminate("Caught signal %d - exiting\n", sig);
-}
-
-static void
-terminate_cleanup(void)
-{
- int dounlink = getenv("CTFMERGE_TERMINATE_NO_UNLINK") ? 0 : 1;
-
- if (tmpname != NULL && dounlink)
- unlink(tmpname);
-
- if (outfile == NULL)
- return;
-
- if (dounlink) {
- fprintf(stderr, "Removing %s\n", outfile);
- unlink(outfile);
- }
-}
-
-static void
-copy_ctf_data(char *srcfile, char *destfile)
-{
- tdata_t *srctd;
-
- if (read_ctf(&srcfile, 1, NULL, read_ctf_save_cb, &srctd, 1) == 0)
- terminate("No CTF data found in source file %s\n", srcfile);
-
- tmpname = mktmpname(destfile, ".ctf");
- write_ctf(srctd, destfile, tmpname, CTF_COMPRESS);
- if (rename(tmpname, destfile) != 0) {
- terminate("Couldn't rename temp file %s to %s", tmpname,
- destfile);
- }
- free(tmpname);
- tdata_free(srctd);
-}
-
-static void
-wq_init(workqueue_t *wq, int nfiles)
-{
- int throttle, nslots, i;
-
- if (getenv("CTFMERGE_MAX_SLOTS"))
- nslots = atoi(getenv("CTFMERGE_MAX_SLOTS"));
- else
- nslots = MERGE_PHASE1_MAX_SLOTS;
-
- if (getenv("CTFMERGE_PHASE1_BATCH_SIZE"))
- wq->wq_maxbatchsz = atoi(getenv("CTFMERGE_PHASE1_BATCH_SIZE"));
- else
- wq->wq_maxbatchsz = MERGE_PHASE1_BATCH_SIZE;
-
- nslots = MIN(nslots, (nfiles + wq->wq_maxbatchsz - 1) /
- wq->wq_maxbatchsz);
-
- wq->wq_wip = xcalloc(sizeof (wip_t) * nslots);
- wq->wq_nwipslots = nslots;
- wq->wq_nthreads = MIN(sysconf(_SC_NPROCESSORS_ONLN) * 3 / 2, nslots);
- wq->wq_thread = xmalloc(sizeof (pthread_t) * wq->wq_nthreads);
-
- if (getenv("CTFMERGE_INPUT_THROTTLE"))
- throttle = atoi(getenv("CTFMERGE_INPUT_THROTTLE"));
- else
- throttle = MERGE_INPUT_THROTTLE_LEN;
- wq->wq_ithrottle = throttle * wq->wq_nthreads;
-
- debug(1, "Using %d slots, %d threads\n", wq->wq_nwipslots,
- wq->wq_nthreads);
-
- wq->wq_next_batchid = 0;
-
- for (i = 0; i < nslots; i++) {
- pthread_mutex_init(&wq->wq_wip[i].wip_lock, NULL);
- wq->wq_wip[i].wip_batchid = wq->wq_next_batchid++;
- }
-
- pthread_mutex_init(&wq->wq_queue_lock, NULL);
- wq->wq_queue = fifo_new();
- pthread_cond_init(&wq->wq_work_avail, NULL);
- pthread_cond_init(&wq->wq_work_removed, NULL);
- wq->wq_ninqueue = nfiles;
- wq->wq_nextpownum = 0;
-
- pthread_mutex_init(&wq->wq_donequeue_lock, NULL);
- wq->wq_donequeue = fifo_new();
- wq->wq_lastdonebatch = -1;
-
- pthread_cond_init(&wq->wq_done_cv, NULL);
-
- pthread_cond_init(&wq->wq_alldone_cv, NULL);
- wq->wq_alldone = 0;
-
- barrier_init(&wq->wq_bar1, wq->wq_nthreads);
- barrier_init(&wq->wq_bar2, wq->wq_nthreads);
-
- wq->wq_nomorefiles = 0;
-}
-
-static void
-start_threads(workqueue_t *wq)
-{
- sigset_t sets;
- int i;
-
- sigemptyset(&sets);
- sigaddset(&sets, SIGINT);
- sigaddset(&sets, SIGQUIT);
- sigaddset(&sets, SIGTERM);
- pthread_sigmask(SIG_BLOCK, &sets, NULL);
-
- for (i = 0; i < wq->wq_nthreads; i++) {
- pthread_create(&wq->wq_thread[i], NULL,
- worker_thread, wq);
- }
-
- sigset(SIGINT, handle_sig);
- sigset(SIGQUIT, handle_sig);
- sigset(SIGTERM, handle_sig);
- pthread_sigmask(SIG_UNBLOCK, &sets, NULL);
-}
-
-static void
-join_threads(workqueue_t *wq)
-{
- int i;
-
- for (i = 0; i < wq->wq_nthreads; i++) {
- pthread_join(wq->wq_thread[i], NULL);
- }
-}
-
-static int
-strcompare(const void *p1, const void *p2)
-{
- char *s1 = *((char **)p1);
- char *s2 = *((char **)p2);
-
- return (strcmp(s1, s2));
-}
-
-/*
- * Core work queue structure; passed to worker threads on thread creation
- * as the main point of coordination. Allocate as a static structure; we
- * could have put this into a local variable in main, but passing a pointer
- * into your stack to another thread is fragile at best and leads to some
- * hard-to-debug failure modes.
- */
-static workqueue_t wq;
-
-int
-main(int argc, char **argv)
-{
- tdata_t *mstrtd, *savetd;
- char *uniqfile = NULL, *uniqlabel = NULL;
- char *withfile = NULL;
- char *label = NULL;
- char **ifiles, **tifiles;
- int verbose = 0, docopy = 0;
- int write_fuzzy_match = 0;
- int require_ctf = 0;
- int nifiles, nielems;
- int c, i, idx, tidx, err;
-
- progname = basename(argv[0]);
-
- ctf_altexec("CTFMERGE_ALTEXEC", argc, argv);
-
- if (getenv("CTFMERGE_DEBUG_LEVEL"))
- debug_level = atoi(getenv("CTFMERGE_DEBUG_LEVEL"));
-
- err = 0;
- while ((c = getopt(argc, argv, ":cd:D:fl:L:o:tvw:s")) != EOF) {
- switch (c) {
- case 'c':
- docopy = 1;
- break;
- case 'd':
- /* Uniquify against `uniqfile' */
- uniqfile = optarg;
- break;
- case 'D':
- /* Uniquify against label `uniqlabel' in `uniqfile' */
- uniqlabel = optarg;
- break;
- case 'f':
- write_fuzzy_match = CTF_FUZZY_MATCH;
- break;
- case 'l':
- /* Label merged types with `label' */
- label = optarg;
- break;
- case 'L':
- /* Label merged types with getenv(`label`) */
- if ((label = getenv(optarg)) == NULL)
- label = CTF_DEFAULT_LABEL;
- break;
- case 'o':
- /* Place merged types in CTF section in `outfile' */
- outfile = optarg;
- break;
- case 't':
- /* Insist *all* object files built from C have CTF */
- require_ctf = 1;
- break;
- case 'v':
- /* More debugging information */
- verbose = 1;
- break;
- case 'w':
- /* Additive merge with data from `withfile' */
- withfile = optarg;
- break;
- case 's':
- /* use the dynsym rather than the symtab */
- dynsym = CTF_USE_DYNSYM;
- break;
- default:
- usage();
- exit(2);
- }
- }
-
- /* Validate arguments */
- if (docopy) {
- if (uniqfile != NULL || uniqlabel != NULL || label != NULL ||
- outfile != NULL || withfile != NULL || dynsym != 0)
- err++;
-
- if (argc - optind != 2)
- err++;
- } else {
- if (uniqfile != NULL && withfile != NULL)
- err++;
-
- if (uniqlabel != NULL && uniqfile == NULL)
- err++;
-
- if (outfile == NULL || label == NULL)
- err++;
-
- if (argc - optind == 0)
- err++;
- }
-
- if (err) {
- usage();
- exit(2);
- }
-
- if (uniqfile && access(uniqfile, R_OK) != 0) {
- warning("Uniquification file %s couldn't be opened and "
- "will be ignored.\n", uniqfile);
- uniqfile = NULL;
- }
- if (withfile && access(withfile, R_OK) != 0) {
- warning("With file %s couldn't be opened and will be "
- "ignored.\n", withfile);
- withfile = NULL;
- }
- if (outfile && access(outfile, R_OK|W_OK) != 0)
- terminate("Cannot open output file %s for r/w", outfile);
-
- /*
- * This is ugly, but we don't want to have to have a separate tool
- * (yet) just for copying an ELF section with our specific requirements,
- * so we shoe-horn a copier into ctfmerge.
- */
- if (docopy) {
- copy_ctf_data(argv[optind], argv[optind + 1]);
-
- exit(0);
- }
-
- set_terminate_cleanup(terminate_cleanup);
-
- /* Sort the input files and strip out duplicates */
- nifiles = argc - optind;
- ifiles = xmalloc(sizeof (char *) * nifiles);
- tifiles = xmalloc(sizeof (char *) * nifiles);
-
- for (i = 0; i < nifiles; i++)
- tifiles[i] = argv[optind + i];
- qsort(tifiles, nifiles, sizeof (char *), (int (*)())strcompare);
-
- ifiles[0] = tifiles[0];
- for (idx = 0, tidx = 1; tidx < nifiles; tidx++) {
- if (strcmp(ifiles[idx], tifiles[tidx]) != 0)
- ifiles[++idx] = tifiles[tidx];
- }
- nifiles = idx + 1;
-
- /* Make sure they all exist */
- if ((nielems = count_files(ifiles, nifiles)) < 0)
- terminate("Some input files were inaccessible\n");
-
- /* Prepare for the merge */
- wq_init(&wq, nielems);
-
- start_threads(&wq);
-
- /*
- * Start the merge
- *
- * We're reading everything from each of the object files, so we
- * don't need to specify labels.
- */
- if (read_ctf(ifiles, nifiles, NULL, merge_ctf_cb,
- &wq, require_ctf) == 0) {
- /*
- * If we're verifying that C files have CTF, it's safe to
- * assume that in this case, we're building only from assembly
- * inputs.
- */
- if (require_ctf)
- exit(0);
- terminate("No ctf sections found to merge\n");
- }
-
- pthread_mutex_lock(&wq.wq_queue_lock);
- wq.wq_nomorefiles = 1;
- pthread_cond_broadcast(&wq.wq_work_avail);
- pthread_mutex_unlock(&wq.wq_queue_lock);
-
- pthread_mutex_lock(&wq.wq_queue_lock);
- while (wq.wq_alldone == 0)
- pthread_cond_wait(&wq.wq_alldone_cv, &wq.wq_queue_lock);
- pthread_mutex_unlock(&wq.wq_queue_lock);
-
- join_threads(&wq);
-
- /*
- * All requested files have been merged, with the resulting tree in
- * mstrtd. savetd is the tree that will be placed into the output file.
- *
- * Regardless of whether we're doing a normal uniquification or an
- * additive merge, we need a type tree that has been uniquified
- * against uniqfile or withfile, as appropriate.
- *
- * If we're doing a uniquification, we stuff the resulting tree into
- * outfile. Otherwise, we add the tree to the tree already in withfile.
- */
- assert(fifo_len(wq.wq_queue) == 1);
- mstrtd = fifo_remove(wq.wq_queue);
-
- if (verbose || debug_level) {
- debug(2, "Statistics for td %p\n", (void *)mstrtd);
-
- iidesc_stats(mstrtd->td_iihash);
- }
-
- if (uniqfile != NULL || withfile != NULL) {
- char *reffile, *reflabel = NULL;
- tdata_t *reftd;
-
- if (uniqfile != NULL) {
- reffile = uniqfile;
- reflabel = uniqlabel;
- } else
- reffile = withfile;
-
- if (read_ctf(&reffile, 1, reflabel, read_ctf_save_cb,
- &reftd, require_ctf) == 0) {
- terminate("No CTF data found in reference file %s\n",
- reffile);
- }
-
- savetd = tdata_new();
-
- if (CTF_TYPE_ISCHILD(reftd->td_nextid))
- terminate("No room for additional types in master\n");
-
- savetd->td_nextid = withfile ? reftd->td_nextid :
- CTF_INDEX_TO_TYPE(1, TRUE);
- merge_into_master(mstrtd, reftd, savetd, 0);
-
- tdata_label_add(savetd, label, CTF_LABEL_LASTIDX);
-
- if (withfile) {
- /*
- * savetd holds the new data to be added to the withfile
- */
- tdata_t *withtd = reftd;
-
- tdata_merge(withtd, savetd);
-
- savetd = withtd;
- } else {
- char uniqname[MAXPATHLEN];
- labelent_t *parle;
-
- parle = tdata_label_top(reftd);
-
- savetd->td_parlabel = xstrdup(parle->le_name);
-
- strncpy(uniqname, reffile, sizeof (uniqname));
- uniqname[MAXPATHLEN - 1] = '\0';
- savetd->td_parname = xstrdup(basename(uniqname));
- }
-
- } else {
- /*
- * No post processing. Write the merged tree as-is into the
- * output file.
- */
- tdata_label_free(mstrtd);
- tdata_label_add(mstrtd, label, CTF_LABEL_LASTIDX);
-
- savetd = mstrtd;
- }
-
- tmpname = mktmpname(outfile, ".ctf");
- write_ctf(savetd, outfile, tmpname,
- CTF_COMPRESS | write_fuzzy_match | dynsym);
- if (rename(tmpname, outfile) != 0)
- terminate("Couldn't rename output temp file %s", tmpname);
- free(tmpname);
-
- return (0);
-}