diff options
Diffstat (limited to 'usr/src/lib/libcpc/common/libcpc.c')
-rw-r--r-- | usr/src/lib/libcpc/common/libcpc.c | 1091 |
1 files changed, 1091 insertions, 0 deletions
diff --git a/usr/src/lib/libcpc/common/libcpc.c b/usr/src/lib/libcpc/common/libcpc.c new file mode 100644 index 0000000000..8b0f3b48fd --- /dev/null +++ b/usr/src/lib/libcpc/common/libcpc.c @@ -0,0 +1,1091 @@ +/* + * CDDL HEADER START + * + * The contents of this file are subject to the terms of the + * Common Development and Distribution License, Version 1.0 only + * (the "License"). You may not use this file except in compliance + * with the License. + * + * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE + * or http://www.opensolaris.org/os/licensing. + * See the License for the specific language governing permissions + * and limitations under the License. + * + * When distributing Covered Code, include this CDDL HEADER in each + * file and include the License file at usr/src/OPENSOLARIS.LICENSE. + * If applicable, add the following below this CDDL HEADER, with the + * fields enclosed by brackets "[]" replaced with your own identifying + * information: Portions Copyright [yyyy] [name of copyright owner] + * + * CDDL HEADER END + */ +/* + * Copyright 2004 Sun Microsystems, Inc. All rights reserved. + * Use is subject to license terms. + */ + +#pragma ident "%Z%%M% %I% %E% SMI" + +#include <libcpc.h> +#include <stdio.h> +#include <stdlib.h> +#include <errno.h> +#include <strings.h> +#include <unistd.h> +#include <stropts.h> +#include <libintl.h> +#include <signal.h> +#include <sys/syscall.h> +#include <sys/types.h> +#include <sys/processor.h> +#include <sys/procset.h> + +#include "libcpc_impl.h" + +#define MASK32 0xFFFFFFFF + +/* + * The library uses the cpc_lock field of the cpc_t struct to protect access to + * the linked lists inside the cpc_t, and only the linked lists. It is NOT used + * to protect against a user shooting his/herself in the foot (such as, for + * instance, destroying the same set at the same time from different threads.). + * + * SIGEMT needs to be blocked while holding the lock, to prevent deadlock among + * an app holding the lock and a signal handler attempting to sample or bind. + */ + +static char *cpc_get_list(int which, int arg); +static void cpc_err(cpc_t *cpc, const char *fn, int subcode, ...); +static int cpc_set_valid(cpc_t *cpc, cpc_set_t *set); +static int cpc_lock(cpc_t *cpc); +static void cpc_unlock(cpc_t *cpc, int blocked); +static int cpc_valid_event(cpc_t *cpc, uint_t pic, const char *ev); +static int cpc_valid_attr(cpc_t *cpc, char *attr); +static void cpc_invalidate_pctx(cpc_t *cpc, pctx_t *pctx); + +cpc_t * +cpc_open(int ver) +{ + cpc_t *cpc; + void (*sigsaved)(); + int error = 0; + int i; + int j; + + if (ver != CPC_VER_CURRENT) { + /* + * v1 clients must stick to the v1 interface: cpc_version() + */ + errno = EINVAL; + return (NULL); + } + + /* + * Call the syscall with invalid parameters. If we get ENOSYS this CPU + * has no CPC support. We need to block SIGSYS because the syscall code + * will send the signal if the system call fails to load. + */ + sigsaved = signal(SIGSYS, SIG_IGN); + if (syscall(SYS_cpc, -1, -1, -1, -1, -1) != -1) { + (void) signal(SIGSYS, sigsaved); + errno = EINVAL; + return (NULL); + } + error = errno; + (void) signal(SIGSYS, sigsaved); + + if (error != EINVAL) { + errno = error; + return (NULL); + } + + if ((cpc = malloc(sizeof (cpc_t))) == NULL) { + errno = ENOMEM; + return (NULL); + } + + cpc->cpc_npic = syscall(SYS_cpc, CPC_NPIC, -1, 0, 0, 0); + cpc->cpc_caps = syscall(SYS_cpc, CPC_CAPS, -1, 0, 0, 0); + + if (syscall(SYS_cpc, CPC_IMPL_NAME, -1, &cpc->cpc_cciname, 0, 0) != 0) + return (NULL); + if (syscall(SYS_cpc, CPC_CPUREF, -1, &cpc->cpc_cpuref, 0, 0) != 0) + return (NULL); + + + if ((cpc->cpc_attrlist = cpc_get_list(CPC_LIST_ATTRS, 0)) == NULL) { + free(cpc); + return (NULL); + } + + if ((cpc->cpc_evlist = malloc(cpc->cpc_npic * sizeof (char *))) == + NULL) { + free(cpc->cpc_attrlist); + free(cpc); + return (NULL); + } + + for (i = 0; i < cpc->cpc_npic; i++) { + if ((cpc->cpc_evlist[i] = cpc_get_list(CPC_LIST_EVENTS, i)) == + NULL) + break; + } + if (i != cpc->cpc_npic) { + for (j = 0; j < i; j++) + free(cpc->cpc_evlist[j]); + free(cpc->cpc_evlist); + free(cpc->cpc_attrlist); + free(cpc); + return (NULL); + } + + cpc->cpc_sets = NULL; + cpc->cpc_bufs = NULL; + cpc->cpc_errfn = NULL; + (void) mutex_init(&cpc->cpc_lock, USYNC_THREAD, NULL); + __pctx_cpc_register_callback(cpc_invalidate_pctx); + + return (cpc); +} + +/* + * Ensure state is cleaned up: + * + * - Hardware is unbound + * - Sets are all destroyed + * - Bufs are all freed + */ +int +cpc_close(cpc_t *cpc) +{ + while (cpc->cpc_sets != NULL) { + if (cpc->cpc_sets->cs_state != CS_UNBOUND) + (void) cpc_unbind(cpc, cpc->cpc_sets); + (void) cpc_set_destroy(cpc, cpc->cpc_sets); + } + + while (cpc->cpc_bufs != NULL) + (void) cpc_buf_destroy(cpc, cpc->cpc_bufs); + + free(cpc); + return (0); +} + +cpc_set_t * +cpc_set_create(cpc_t *cpc) +{ + cpc_set_t *set; + int sigblocked; + + if ((set = malloc(sizeof (*set))) == NULL) { + errno = ENOMEM; + return (NULL); + } + + set->cs_request = NULL; + set->cs_nreqs = 0; + set->cs_state = CS_UNBOUND; + set->cs_fd = -1; + set->cs_pctx = NULL; + set->cs_id = -1; + set->cs_thr = NULL; + + sigblocked = cpc_lock(cpc); + set->cs_next = cpc->cpc_sets; + cpc->cpc_sets = set; + cpc_unlock(cpc, sigblocked); + + return (set); +} + +int +cpc_set_destroy(cpc_t *cpc, cpc_set_t *set) +{ + cpc_set_t *csp, *prev; + cpc_request_t *req, *next; + int sigblocked; + + /* + * Remove this set from the cpc handle's list of sets. + */ + sigblocked = cpc_lock(cpc); + for (csp = prev = cpc->cpc_sets; csp != NULL; csp = csp->cs_next) { + if (csp == set) + break; + prev = csp; + } + if (csp == NULL) { + cpc_unlock(cpc, sigblocked); + errno = EINVAL; + return (-1); + } + if (csp == cpc->cpc_sets) + cpc->cpc_sets = csp->cs_next; + prev->cs_next = csp->cs_next; + cpc_unlock(cpc, sigblocked); + + if (csp->cs_state != CS_UNBOUND) + (void) cpc_unbind(cpc, csp); + + for (req = csp->cs_request; req != NULL; req = next) { + next = req->cr_next; + + if (req->cr_nattrs != 0) + free(req->cr_attr); + + free(req); + } + + + free(set); + + return (0); +} + +/*ARGSUSED*/ +int +cpc_set_add_request(cpc_t *cpc, cpc_set_t *set, const char *event, + uint64_t preset, uint_t flags, uint_t nattrs, const cpc_attr_t *attrs) +{ + cpc_request_t *req; + const char *fn = "cpc_set_add_request"; + int i; + int npics = cpc_npic(cpc); + + if (cpc_set_valid(cpc, set) != 0 || set->cs_state != CS_UNBOUND) { + errno = EINVAL; + return (-1); + } + + for (i = 0; i < npics; i++) + if (cpc_valid_event(cpc, i, event)) + break; + if (i == npics) { + cpc_err(cpc, fn, CPC_INVALID_EVENT); + errno = EINVAL; + return (-1); + } + + if ((req = malloc(sizeof (*req))) == NULL) { + errno = ENOMEM; + return (-1); + } + + (void) strncpy(req->cr_event, event, CPC_MAX_EVENT_LEN); + req->cr_preset = preset; + req->cr_flags = flags; + req->cr_nattrs = nattrs; + req->cr_index = set->cs_nreqs; + req->cr_attr = NULL; + + if (nattrs != 0) { + for (i = 0; i < nattrs; i++) { + /* + * Verify that each attribute name is legal and valid. + */ + if (attrs[i].ca_name[0] == '\0' || + cpc_valid_attr(cpc, attrs[i].ca_name) == 0) { + cpc_err(cpc, fn, CPC_INVALID_ATTRIBUTE); + goto inval; + } + + /* + * If the user requested a specific picnum, ensure that + * the pic can count the requested event. + */ + if (strncmp("picnum", attrs[i].ca_name, 8) == 0) { + if (attrs[i].ca_val >= npics) { + cpc_err(cpc, fn, CPC_INVALID_PICNUM); + goto inval; + } + + if (cpc_valid_event(cpc, attrs[i].ca_val, + req->cr_event) == 0) { + cpc_err(cpc, fn, CPC_PIC_NOT_CAPABLE); + goto inval; + } + } + } + + if ((req->cr_attr = malloc(nattrs * sizeof (kcpc_attr_t))) + == NULL) { + free(req); + return (-1); + } + + for (i = 0; i < nattrs; i++) { + req->cr_attr[i].ka_val = attrs[i].ca_val; + (void) strncpy(req->cr_attr[i].ka_name, + attrs[i].ca_name, CPC_MAX_ATTR_LEN); + } + } else + req->cr_attr = NULL; + + req->cr_next = set->cs_request; + set->cs_request = req; + set->cs_nreqs++; + + return (req->cr_index); + +inval: + free(req); + errno = EINVAL; + return (-1); +} + +cpc_buf_t * +cpc_buf_create(cpc_t *cpc, cpc_set_t *set) +{ + cpc_buf_t *buf; + int sigblocked; + + if (cpc_set_valid(cpc, set) != 0) { + errno = EINVAL; + return (NULL); + } + + if ((buf = malloc(sizeof (*buf))) == NULL) + return (NULL); + + buf->cb_size = set->cs_nreqs * sizeof (uint64_t); + if ((buf->cb_data = malloc(buf->cb_size)) == NULL) { + free(buf); + return (NULL); + } + + bzero(buf->cb_data, buf->cb_size); + + buf->cb_hrtime = 0; + buf->cb_tick = 0; + + sigblocked = cpc_lock(cpc); + buf->cb_next = cpc->cpc_bufs; + cpc->cpc_bufs = buf; + cpc_unlock(cpc, sigblocked); + + return (buf); +} + +int +cpc_buf_destroy(cpc_t *cpc, cpc_buf_t *buf) +{ + cpc_buf_t *cbp, *prev; + int sigblocked; + + /* + * Remove this buf from the cpc handle's list of bufs. + */ + sigblocked = cpc_lock(cpc); + for (cbp = prev = cpc->cpc_bufs; cbp != NULL; cbp = cbp->cb_next) { + if (cbp == buf) + break; + prev = cbp; + } + if (cbp == NULL) { + cpc_unlock(cpc, sigblocked); + errno = EINVAL; + return (-1); + } + if (cbp == cpc->cpc_bufs) + cpc->cpc_bufs = cbp->cb_next; + prev->cb_next = cbp->cb_next; + + cpc_unlock(cpc, sigblocked); + free(cbp->cb_data); + free(cbp); + + return (0); +} + +/*ARGSUSED*/ +int +cpc_bind_curlwp(cpc_t *cpc, cpc_set_t *set, uint_t flags) +{ + char *packed_set; + size_t packsize; + int ret; + int subcode = -1; + + /* + * We don't bother checking cpc_set_valid() here, because this is in the + * fast path of an app doing SIGEMT-based profiling as they restart the + * counters from their signal handler. + */ + if (CPC_SET_VALID_FLAGS(flags) == 0 || set->cs_nreqs <= 0) { + errno = EINVAL; + return (-1); + } + + if ((packed_set = __cpc_pack_set(set, flags, &packsize)) == NULL) { + errno = ENOMEM; + return (-1); + } + + ret = syscall(SYS_cpc, CPC_BIND, -1, packed_set, packsize, &subcode); + free(packed_set); + + if (ret != 0) { + if (subcode != -1) + cpc_err(cpc, "cpc_bind_curlwp", subcode); + return (-1); + } + + set->cs_thr = thr_self(); + set->cs_state = CS_BOUND_CURLWP; + return (ret); +} + +/*ARGSUSED*/ +int +cpc_bind_pctx(cpc_t *cpc, pctx_t *pctx, id_t id, cpc_set_t *set, uint_t flags) +{ + char *packed_set; + size_t packsize; + int ret; + int subcode = -1; + + /* + * cpc_bind_pctx() currently has no valid flags. + */ + if (flags != 0 || cpc_set_valid(cpc, set) != 0 || set->cs_nreqs <= 0) { + errno = EINVAL; + return (-1); + } + + if ((packed_set = __cpc_pack_set(set, flags, &packsize)) == NULL) { + errno = ENOMEM; + return (-1); + } + + ret = __pctx_cpc(pctx, cpc, CPC_BIND, id, packed_set, (void *)packsize, + (void *)&subcode, -1); + + free(packed_set); + + if (ret == 0) { + set->cs_pctx = pctx; + set->cs_id = id; + set->cs_state = CS_BOUND_PCTX; + } else if (subcode != -1) + cpc_err(cpc, "cpc_bind_pctx", subcode); + + return (ret); +} + +/*ARGSUSED*/ +int +cpc_bind_cpu(cpc_t *cpc, processorid_t id, cpc_set_t *set, uint_t flags) +{ + int fd; + char *packed_set; + size_t packsize; + __cpc_args_t cpc_args; + int error; + const char *fn = "cpc_bind_cpu"; + int subcode = -1; + + /* + * cpc_bind_cpu() currently has no valid flags. + */ + if (flags != 0 || cpc_set_valid(cpc, set) != 0 || set->cs_nreqs <= 0) { + errno = EINVAL; + return (-1); + } + + if (processor_bind(P_LWPID, P_MYID, id, &set->cs_obind) == -1) { + cpc_err(cpc, fn, CPC_PBIND_FAILED); + return (-1); + } + + if ((fd = open(CPUDRV_SHARED, O_RDWR)) < 0) { + error = errno; + (void) processor_bind(P_LWPID, P_MYID, set->cs_obind, NULL); + errno = error; + return (-1); + } + + /* + * To avoid leaking file descriptors, if we find an existing fd here we + * just close it. This is only a problem if a user attempts to bind the + * same set to different CPUs without first unbinding it. + */ + if (set->cs_fd != -1) + (void) close(set->cs_fd); + set->cs_fd = fd; + + if ((packed_set = __cpc_pack_set(set, flags, &packsize)) == NULL) { + (void) close(fd); + (void) processor_bind(P_LWPID, P_MYID, set->cs_obind, NULL); + errno = ENOMEM; + return (-1); + } + + cpc_args.udata1 = packed_set; + cpc_args.udata2 = (void *)packsize; + cpc_args.udata3 = (void *)&subcode; + + if (ioctl(fd, CPCIO_BIND, &cpc_args) != 0) { + error = errno; + free(packed_set); + (void) close(fd); + (void) processor_bind(P_LWPID, P_MYID, set->cs_obind, NULL); + if (subcode != -1) + cpc_err(cpc, fn, subcode); + errno = error; + return (-1); + } + + free(packed_set); + + set->cs_thr = thr_self(); + set->cs_state = CS_BOUND_CPU; + + return (0); +} + +/*ARGSUSED*/ +int +cpc_request_preset(cpc_t *cpc, int index, uint64_t preset) +{ + return (syscall(SYS_cpc, CPC_PRESET, -1, index, + (uint32_t)(preset >> 32), (uint32_t)(preset & MASK32))); +} + +/*ARGSUSED*/ +int +cpc_set_restart(cpc_t *cpc, cpc_set_t *set) +{ + return (syscall(SYS_cpc, CPC_RESTART, -1, 0, 0, 0)); +} + +/*ARGSUSED*/ +int +cpc_unbind(cpc_t *cpc, cpc_set_t *set) +{ + int ret = 0; + int error; + + if (cpc_set_valid(cpc, set) != 0) { + errno = EINVAL; + return (-1); + } + + switch (set->cs_state) { + case CS_UNBOUND: + errno = EINVAL; + return (-1); + case CS_BOUND_CURLWP: + ret = syscall(SYS_cpc, CPC_RELE, -1, 0, 0, 0); + error = errno; + break; + case CS_BOUND_CPU: + ret = ioctl(set->cs_fd, CPCIO_RELE, NULL); + error = errno; + (void) close(set->cs_fd); + set->cs_fd = -1; + (void) processor_bind(P_LWPID, P_MYID, set->cs_obind, NULL); + break; + case CS_BOUND_PCTX: + if (set->cs_pctx != NULL) { + ret = __pctx_cpc(set->cs_pctx, cpc, CPC_RELE, + set->cs_id, 0, 0, 0, 0); + error = errno; + } + break; + } + + set->cs_thr = NULL; + set->cs_id = -1; + set->cs_state = CS_UNBOUND; + if (ret != 0) + errno = error; + return (ret); +} + +/*ARGSUSED*/ +int +cpc_set_sample(cpc_t *cpc, cpc_set_t *set, cpc_buf_t *buf) +{ + __cpc_args_t args; + + /* + * The following check ensures that only the most recently bound set + * can be sampled, as binding a set invalidates all other sets in the + * cpc_t. + */ + if (set->cs_state == CS_UNBOUND || + buf->cb_size != set->cs_nreqs * sizeof (uint64_t)) { + errno = EINVAL; + return (-1); + } + + switch (set->cs_state) { + case CS_BOUND_CURLWP: + return (syscall(SYS_cpc, CPC_SAMPLE, -1, buf->cb_data, + &buf->cb_hrtime, &buf->cb_tick)); + case CS_BOUND_CPU: + args.udata1 = buf->cb_data; + args.udata2 = &buf->cb_hrtime; + args.udata3 = &buf->cb_tick; + return (ioctl(set->cs_fd, CPCIO_SAMPLE, &args)); + case CS_BOUND_PCTX: + return (__pctx_cpc(set->cs_pctx, cpc, CPC_SAMPLE, set->cs_id, + buf->cb_data, &buf->cb_hrtime, &buf->cb_tick, + buf->cb_size)); + } + + errno = EINVAL; + return (-1); +} + +/*ARGSUSED*/ +void +cpc_buf_sub(cpc_t *cpc, cpc_buf_t *ds, cpc_buf_t *a, cpc_buf_t *b) +{ + int i; + + if (a->cb_size != ds->cb_size || b->cb_size != ds->cb_size) + return; + + ds->cb_hrtime = (a->cb_hrtime > b->cb_hrtime) ? + a->cb_hrtime : b->cb_hrtime; + ds->cb_tick = a->cb_tick - b->cb_tick; + + for (i = 0; i < ds->cb_size / sizeof (uint64_t); i++) + ds->cb_data[i] = a->cb_data[i] - b->cb_data[i]; +} + +/*ARGSUSED*/ +void +cpc_buf_add(cpc_t *cpc, cpc_buf_t *ds, cpc_buf_t *a, cpc_buf_t *b) +{ + int i; + + if (a->cb_size != ds->cb_size || b->cb_size != ds->cb_size) + return; + + ds->cb_hrtime = (a->cb_hrtime > b->cb_hrtime) ? + a->cb_hrtime : b->cb_hrtime; + ds->cb_tick = a->cb_tick + b->cb_tick; + + for (i = 0; i < ds->cb_size / sizeof (uint64_t); i++) + ds->cb_data[i] = a->cb_data[i] + b->cb_data[i]; +} + +/*ARGSUSED*/ +void +cpc_buf_copy(cpc_t *cpc, cpc_buf_t *ds, cpc_buf_t *src) +{ + if (ds->cb_size != src->cb_size) + return; + + bcopy(src->cb_data, ds->cb_data, ds->cb_size); + ds->cb_hrtime = src->cb_hrtime; + ds->cb_tick = src->cb_tick; +} + +/*ARGSUSED*/ +void +cpc_buf_zero(cpc_t *cpc, cpc_buf_t *buf) +{ + bzero(buf->cb_data, buf->cb_size); + buf->cb_hrtime = 0; + buf->cb_tick = 0; +} + +/* + * Gets or sets the value of the request specified by index. + */ +/*ARGSUSED*/ +int +cpc_buf_get(cpc_t *cpc, cpc_buf_t *buf, int index, uint64_t *val) +{ + *val = buf->cb_data[index]; + + return (0); +} + +/*ARGSUSED*/ +int +cpc_buf_set(cpc_t *cpc, cpc_buf_t *buf, int index, uint64_t val) +{ + buf->cb_data[index] = val; + + return (0); +} + +/*ARGSUSED*/ +hrtime_t +cpc_buf_hrtime(cpc_t *cpc, cpc_buf_t *buf) +{ + return (buf->cb_hrtime); +} + +/*ARGSUSED*/ +uint64_t +cpc_buf_tick(cpc_t *cpc, cpc_buf_t *buf) +{ + return (buf->cb_tick); +} + +static char * +cpc_get_list(int which, int arg) +{ + int szcmd; + int size; + char *list; + + if (which == CPC_LIST_ATTRS) + szcmd = CPC_ATTRLIST_SIZE; + else + szcmd = CPC_EVLIST_SIZE; + + if (syscall(SYS_cpc, szcmd, -1, &size, arg, 0) != 0) + return (NULL); + + if ((list = malloc(size)) == NULL) + return (NULL); + + if (syscall(SYS_cpc, which, -1, list, arg, 0) != 0) { + free(list); + return (NULL); + } + + return (list); +} + +/*ARGSUSED*/ +void +cpc_walk_requests(cpc_t *cpc, cpc_set_t *set, void *arg, + void (*action)(void *arg, int index, const char *event, uint64_t preset, + uint_t flags, int nattrs, const cpc_attr_t *attrs)) +{ + cpc_request_t *rp; + cpc_attr_t *attrs = NULL; + int i; + + for (rp = set->cs_request; rp != NULL; rp = rp->cr_next) { + /* + * Need to reconstruct a temporary cpc_attr_t array for req. + */ + if (rp->cr_nattrs != 0) + if ((attrs = malloc(rp->cr_nattrs * + sizeof (cpc_attr_t))) == NULL) + return; + for (i = 0; i < rp->cr_nattrs; i++) { + attrs[i].ca_name = rp->cr_attr[i].ka_name; + attrs[i].ca_val = rp->cr_attr[i].ka_val; + } + + action(arg, rp->cr_index, rp->cr_event, rp->cr_preset, + rp->cr_flags, rp->cr_nattrs, attrs); + + if (rp->cr_nattrs != 0) + free(attrs); + } +} + +/*ARGSUSED*/ +void +cpc_walk_events_all(cpc_t *cpc, void *arg, + void (*action)(void *arg, const char *event)) +{ + char **list; + char *p, *e; + int i; + int ncounters = cpc_npic(cpc); + cpc_strhash_t *hash; + + if ((list = malloc(ncounters * sizeof (char *))) == NULL) + return; + + if ((hash = __cpc_strhash_alloc()) == NULL) { + free(list); + return; + } + + for (i = 0; i < ncounters; i++) { + if ((list[i] = strdup(cpc->cpc_evlist[i])) == NULL) + goto err; + p = list[i]; + while ((e = strchr(p, ',')) != NULL) { + *e = '\0'; + if (__cpc_strhash_add(hash, p) == -1) + goto err; + p = e + 1; + } + if (__cpc_strhash_add(hash, p) == -1) + goto err; + } + + while ((p = __cpc_strhash_next(hash)) != NULL) + action(arg, p); + +err: + __cpc_strhash_free(hash); + for (i = 0; i < ncounters; i++) + free(list[i]); + free(list); +} + +/*ARGSUSED*/ +void +cpc_walk_events_pic(cpc_t *cpc, uint_t picno, void *arg, + void (*action)(void *arg, uint_t picno, const char *event)) +{ + char *p; + char *e; + char *list; + + if (picno >= cpc->cpc_npic) { + errno = EINVAL; + return; + } + + if ((list = strdup(cpc->cpc_evlist[picno])) == NULL) + return; + + /* + * List now points to a comma-separated list of events supported by + * the designated pic. + */ + p = list; + while ((e = strchr(p, ',')) != NULL) { + *e = '\0'; + action(arg, picno, p); + p = e + 1; + } + action(arg, picno, p); + + free(list); +} + +/*ARGSUSED*/ +void +cpc_walk_attrs(cpc_t *cpc, void *arg, + void (*action)(void *arg, const char *attr)) +{ + char *p; + char *e; + char *list; + + if ((list = strdup(cpc->cpc_attrlist)) == NULL) + return; + + /* + * Platforms with no attributes will return an empty string. + */ + if (*list == '\0') + return; + + /* + * List now points to a comma-separated list of attributes supported by + * the underlying platform. + */ + p = list; + while ((e = strchr(p, ',')) != NULL) { + *e = '\0'; + action(arg, p); + p = e + 1; + } + action(arg, p); + + free(list); +} + +/*ARGSUSED*/ +int +cpc_enable(cpc_t *cpc) +{ + return (syscall(SYS_cpc, CPC_ENABLE, -1, 0, 0, 0)); +} + +/*ARGSUSED*/ +int +cpc_disable(cpc_t *cpc) +{ + return (syscall(SYS_cpc, CPC_DISABLE, -1, 0, 0, 0)); +} + +/*ARGSUSED*/ +uint_t +cpc_npic(cpc_t *cpc) +{ + return (cpc->cpc_npic); +} + +/*ARGSUSED*/ +uint_t +cpc_caps(cpc_t *cpc) +{ + return (cpc->cpc_caps); +} + +const char * +cpc_cciname(cpc_t *cpc) +{ + return (cpc->cpc_cciname); +} + +const char * +cpc_cpuref(cpc_t *cpc) +{ + return (cpc->cpc_cpuref); +} + +int +cpc_seterrhndlr(cpc_t *cpc, cpc_errhndlr_t *fn) +{ + cpc->cpc_errfn = fn; + return (0); +} + +/* + * These strings may contain printf() conversion specifiers. + */ +static const char *errstr[] = { +"", /* zero slot filler */ +"Unknown event\n", /* CPC_INVALID_EVENT */ +"Invalid counter number\n", /* CPC_INVALID_PICNUM */ +"Unknown attribute\n", /* CPC_INVALID_ATTRIBUTE */ +"Attribute out of range\n", /* CPC_ATTRIBUTE_OUT_OF_RANGE */ +"Hardware resource unavailable\n", /* CPC_RESOURCE_UNAVAIL */ +"Counter cannot count requested event\n", /* CPC_PIC_NOT_CAPABLE */ +"Invalid flags in a request\n", /* CPC_REQ_INVALID_FLAGS */ +"Requests conflict with each other\n", /* CPC_CONFLICTING_REQS */ +"Attribute requires the cpc_cpu privilege\n", /* CPC_ATTR_REQUIRES_PRIVILEGE */ +"Couldn't bind LWP to requested processor\n" /* CPC_PBIND_FAILED */ +}; + +/*VARARGS3*/ +static void +cpc_err(cpc_t *cpc, const char *fn, int subcode, ...) +{ + va_list ap; + const char *str; + int error; + + /* + * If subcode is -1, there is no specific description for this error. + */ + if (subcode == -1) + return; + + /* + * We need to preserve errno across calls to this function to prevent it + * from being clobbered while here, or in the user's error handler. + */ + error = errno; + + str = dgettext(TEXT_DOMAIN, errstr[subcode]); + + va_start(ap, subcode); + if (cpc->cpc_errfn != NULL) + cpc->cpc_errfn(fn, subcode, str, ap); + else { + /* + * If printf() conversion specifiers are added to the errstr[] + * table, this call needs to be changed to vfprintf(). + */ + (void) fprintf(stderr, "libcpc: %s: %s", fn, str); + } + va_end(ap); + + errno = error; +} + +/* + * Hook used by libpctx to alert libcpc when a pctx handle is going away. + * This is necessary to prevent libcpc from attempting a libpctx operation on a + * stale and invalid pctx_t handle. Since pctx_t's are cached by libcpc, we need + * to be notified when they go away. + */ +static void +cpc_invalidate_pctx(cpc_t *cpc, pctx_t *pctx) +{ + cpc_set_t *set; + int sigblocked; + + sigblocked = cpc_lock(cpc); + for (set = cpc->cpc_sets; set != NULL; set = set->cs_next) + if (set->cs_pctx == pctx) + set->cs_pctx = NULL; + cpc_unlock(cpc, sigblocked); +} + +/* + * Check that the set is valid; if so it will be in the cpc handle's + * list of sets. The lock protects the list of sets, but not the set + * itself. + */ +static int +cpc_set_valid(cpc_t *cpc, cpc_set_t *set) +{ + cpc_set_t *csp; + int sigblocked; + + sigblocked = cpc_lock(cpc); + for (csp = cpc->cpc_sets; csp != NULL; csp = csp->cs_next) + if (csp == set) + break; + cpc_unlock(cpc, sigblocked); + if (csp == NULL) + return (-1); + return (0); +} + +static int +cpc_lock(cpc_t *cpc) +{ + int ret = (sigset(SIGEMT, SIG_HOLD) == SIG_HOLD); + (void) mutex_lock(&cpc->cpc_lock); + return (ret); +} + +static void +cpc_unlock(cpc_t *cpc, int sigblocked) +{ + (void) mutex_unlock(&cpc->cpc_lock); + if (sigblocked == 0) + (void) sigrelse(SIGEMT); +} + +struct priv { + const char *name; + int found; +}; + +/*ARGSUSED*/ +static void +ev_walker(void *arg, uint_t picno, const char *ev) +{ + if (strcmp(((struct priv *)arg)->name, ev) == 0) + ((struct priv *)arg)->found = 1; +} + +static void +at_walker(void *arg, const char *at) +{ + if (strcmp(((struct priv *)arg)->name, at) == 0) + ((struct priv *)arg)->found = 1; +} + +static int +cpc_valid_event(cpc_t *cpc, uint_t pic, const char *ev) +{ + struct priv pr = { NULL, 0 }; + + pr.name = ev; + cpc_walk_events_pic(cpc, pic, &pr, ev_walker); + return (pr.found); +} + +static int +cpc_valid_attr(cpc_t *cpc, char *attr) +{ + struct priv pr = { NULL, 0 }; + + pr.name = attr; + cpc_walk_attrs(cpc, &pr, at_walker); + return (pr.found); +} |