summaryrefslogtreecommitdiff
path: root/usr/src/uts/intel/os/cpc_subr.c
diff options
context:
space:
mode:
authorRichard Lowe <richlowe@richlowe.net>2021-06-04 15:15:12 -0500
committerRichard Lowe <richlowe@richlowe.net>2021-08-16 12:46:39 -0500
commitf0089e391b2bc4be2755f1a1b51fb4cd9b8f3988 (patch)
treec4ac2f5e703ed459d50bcee7ddb38a993d961520 /usr/src/uts/intel/os/cpc_subr.c
parentd083fed0c91296a88878f7a468910ad5b5c888ea (diff)
downloadillumos-gate-f0089e391b2bc4be2755f1a1b51fb4cd9b8f3988.tar.gz
13941 intel code and headers should not look ia32 specific
Reviewed by: Hans Rosenfeld <rosenfeld@grumpf.hope-2000.org> Reviewed by: Toomas Soome <tsoome@me.com> Reviewed by: Patrick Mooney <pmooney@pfmooney.com> Approved by: Garret D'Amore <garrett@damore.org>
Diffstat (limited to 'usr/src/uts/intel/os/cpc_subr.c')
-rw-r--r--usr/src/uts/intel/os/cpc_subr.c274
1 files changed, 274 insertions, 0 deletions
diff --git a/usr/src/uts/intel/os/cpc_subr.c b/usr/src/uts/intel/os/cpc_subr.c
new file mode 100644
index 0000000000..71e1ebaeee
--- /dev/null
+++ b/usr/src/uts/intel/os/cpc_subr.c
@@ -0,0 +1,274 @@
+/*
+ * CDDL HEADER START
+ *
+ * The contents of this file are subject to the terms of the
+ * Common Development and Distribution License (the "License").
+ * You may not use this file except in compliance with the License.
+ *
+ * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
+ * or http://www.opensolaris.org/os/licensing.
+ * See the License for the specific language governing permissions
+ * and limitations under the License.
+ *
+ * When distributing Covered Code, include this CDDL HEADER in each
+ * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
+ * If applicable, add the following below this CDDL HEADER, with the
+ * fields enclosed by brackets "[]" replaced with your own identifying
+ * information: Portions Copyright [yyyy] [name of copyright owner]
+ *
+ * CDDL HEADER END
+ */
+/*
+ * Copyright (c) 1999, 2010, Oracle and/or its affiliates. All rights reserved.
+ * Copyright 2021 Joyent, Inc.
+ */
+
+/*
+ * x86-specific routines used by the CPU Performance counter driver.
+ */
+
+#include <sys/types.h>
+#include <sys/time.h>
+#include <sys/atomic.h>
+#include <sys/regset.h>
+#include <sys/privregs.h>
+#include <sys/x86_archext.h>
+#include <sys/cpuvar.h>
+#include <sys/machcpuvar.h>
+#include <sys/archsystm.h>
+#include <sys/cpc_pcbe.h>
+#include <sys/cpc_impl.h>
+#include <sys/x_call.h>
+#include <sys/cmn_err.h>
+#include <sys/cmt.h>
+#include <sys/spl.h>
+#include <sys/apic.h>
+
+static const uint64_t allstopped = 0;
+static kcpc_ctx_t *(*overflow_intr_handler)(caddr_t);
+
+/* Do threads share performance monitoring hardware? */
+static int strands_perfmon_shared = 0;
+
+int kcpc_hw_overflow_intr_installed; /* set by APIC code */
+extern kcpc_ctx_t *kcpc_overflow_intr(caddr_t arg, uint64_t bitmap);
+
+extern int kcpc_counts_include_idle; /* Project Private /etc/system variable */
+
+void (*kcpc_hw_enable_cpc_intr)(void); /* set by APIC code */
+
+int
+kcpc_hw_add_ovf_intr(kcpc_ctx_t *(*handler)(caddr_t))
+{
+ if (x86_type != X86_TYPE_P6)
+ return (0);
+ overflow_intr_handler = handler;
+ return (ipltospl(APIC_PCINT_IPL));
+}
+
+void
+kcpc_hw_rem_ovf_intr(void)
+{
+ overflow_intr_handler = NULL;
+}
+
+/*
+ * Hook used on P4 systems to catch online/offline events.
+ */
+/*ARGSUSED*/
+static int
+kcpc_cpu_setup(cpu_setup_t what, int cpuid, void *arg)
+{
+ pg_cmt_t *chip_pg;
+ int active_cpus_cnt;
+
+ if (what != CPU_ON)
+ return (0);
+
+ /*
+ * If any CPU-bound contexts exist, we don't need to invalidate
+ * anything, as no per-LWP contexts can coexist.
+ */
+ if (kcpc_cpuctx || dtrace_cpc_in_use)
+ return (0);
+
+ /*
+ * If this chip now has more than 1 active cpu, we must invalidate all
+ * contexts in the system.
+ */
+ chip_pg = (pg_cmt_t *)pghw_find_pg(cpu[cpuid], PGHW_CHIP);
+ if (chip_pg != NULL) {
+ active_cpus_cnt = GROUP_SIZE(&chip_pg->cmt_cpus_actv);
+ if (active_cpus_cnt > 1)
+ kcpc_invalidate_all();
+ }
+
+ return (0);
+}
+
+static kmutex_t cpu_setup_lock; /* protects setup_registered */
+static int setup_registered;
+
+
+void
+kcpc_hw_init(cpu_t *cp)
+{
+ kthread_t *t = cp->cpu_idle_thread;
+ uint32_t versionid;
+ struct cpuid_regs cpuid;
+
+ strands_perfmon_shared = 0;
+ if (is_x86_feature(x86_featureset, X86FSET_HTT)) {
+ if (cpuid_getvendor(cpu[0]) == X86_VENDOR_Intel) {
+ /*
+ * Intel processors that support Architectural
+ * Performance Monitoring Version 3 have per strand
+ * performance monitoring hardware.
+ * Hence we can allow use of performance counters on
+ * multiple strands on the same core simultaneously.
+ */
+ cpuid.cp_eax = 0x0;
+ (void) __cpuid_insn(&cpuid);
+ if (cpuid.cp_eax < 0xa) {
+ strands_perfmon_shared = 1;
+ } else {
+ cpuid.cp_eax = 0xa;
+ (void) __cpuid_insn(&cpuid);
+
+ versionid = cpuid.cp_eax & 0xFF;
+ if (versionid < 3) {
+ strands_perfmon_shared = 1;
+ }
+ }
+ } else if (cpuid_getvendor(cpu[0]) == X86_VENDOR_AMD ||
+ cpuid_getvendor(cpu[0]) == X86_VENDOR_HYGON) {
+ /*
+ * On AMD systems with HT, all of the performance
+ * monitors exist on a per-logical CPU basis.
+ */
+ strands_perfmon_shared = 0;
+ } else {
+ strands_perfmon_shared = 1;
+ }
+ }
+
+ if (strands_perfmon_shared) {
+ mutex_enter(&cpu_setup_lock);
+ if (setup_registered == 0) {
+ mutex_enter(&cpu_lock);
+ register_cpu_setup_func(kcpc_cpu_setup, NULL);
+ mutex_exit(&cpu_lock);
+ setup_registered = 1;
+ }
+ mutex_exit(&cpu_setup_lock);
+ }
+
+ mutex_init(&cp->cpu_cpc_ctxlock, "cpu_cpc_ctxlock", MUTEX_DEFAULT, 0);
+
+ if (kcpc_counts_include_idle)
+ return;
+
+ installctx(t, cp, kcpc_idle_save, kcpc_idle_restore,
+ NULL, NULL, NULL, NULL, NULL);
+}
+
+void
+kcpc_hw_fini(cpu_t *cp)
+{
+ ASSERT(cp->cpu_idle_thread == NULL);
+
+ mutex_destroy(&cp->cpu_cpc_ctxlock);
+}
+
+#define BITS(v, u, l) \
+ (((v) >> (l)) & ((1 << (1 + (u) - (l))) - 1))
+
+#define PCBE_NAMELEN 30 /* Enough Room for pcbe.manuf.model.family.stepping */
+
+/*
+ * Examine the processor and load an appropriate PCBE.
+ */
+int
+kcpc_hw_load_pcbe(void)
+{
+ return (kcpc_pcbe_tryload(cpuid_getvendorstr(CPU), cpuid_getfamily(CPU),
+ cpuid_getmodel(CPU), cpuid_getstep(CPU)));
+}
+
+/*
+ * Called by the generic framework to check if it's OK to bind a set to a CPU.
+ */
+int
+kcpc_hw_cpu_hook(processorid_t cpuid, ulong_t *kcpc_cpumap)
+{
+ cpu_t *cpu, *p;
+ pg_t *chip_pg;
+ pg_cpu_itr_t itr;
+
+ if (!strands_perfmon_shared)
+ return (0);
+
+ /*
+ * Only one logical CPU on each Pentium 4 HT CPU may be bound to at
+ * once.
+ *
+ * This loop is protected by holding cpu_lock, in order to properly
+ * access the cpu_t of the desired cpu.
+ */
+ mutex_enter(&cpu_lock);
+ if ((cpu = cpu_get(cpuid)) == NULL) {
+ mutex_exit(&cpu_lock);
+ return (-1);
+ }
+
+ chip_pg = (pg_t *)pghw_find_pg(cpu, PGHW_CHIP);
+
+ PG_CPU_ITR_INIT(chip_pg, itr);
+ while ((p = pg_cpu_next(&itr)) != NULL) {
+ if (p == cpu)
+ continue;
+ if (BT_TEST(kcpc_cpumap, p->cpu_id)) {
+ mutex_exit(&cpu_lock);
+ return (-1);
+ }
+ }
+
+ mutex_exit(&cpu_lock);
+ return (0);
+}
+
+/*
+ * Called by the generic framework to check if it's OK to bind a set to an LWP.
+ */
+int
+kcpc_hw_lwp_hook(void)
+{
+ pg_cmt_t *chip;
+ group_t *chips;
+ group_iter_t i;
+
+ if (!strands_perfmon_shared)
+ return (0);
+
+ /*
+ * Only one CPU per chip may be online.
+ */
+ mutex_enter(&cpu_lock);
+
+ chips = pghw_set_lookup(PGHW_CHIP);
+ if (chips == NULL) {
+ mutex_exit(&cpu_lock);
+ return (0);
+ }
+
+ group_iter_init(&i);
+ while ((chip = group_iterate(chips, &i)) != NULL) {
+ if (GROUP_SIZE(&chip->cmt_cpus_actv) > 1) {
+ mutex_exit(&cpu_lock);
+ return (-1);
+ }
+ }
+
+ mutex_exit(&cpu_lock);
+ return (0);
+}