summaryrefslogtreecommitdiff
path: root/usr/src/uts/sun4u/starfire/io/idn.c
diff options
context:
space:
mode:
Diffstat (limited to 'usr/src/uts/sun4u/starfire/io/idn.c')
-rw-r--r--usr/src/uts/sun4u/starfire/io/idn.c5779
1 files changed, 5779 insertions, 0 deletions
diff --git a/usr/src/uts/sun4u/starfire/io/idn.c b/usr/src/uts/sun4u/starfire/io/idn.c
new file mode 100644
index 0000000000..7d4e9edcf9
--- /dev/null
+++ b/usr/src/uts/sun4u/starfire/io/idn.c
@@ -0,0 +1,5779 @@
+/*
+ * CDDL HEADER START
+ *
+ * The contents of this file are subject to the terms of the
+ * Common Development and Distribution License, Version 1.0 only
+ * (the "License"). You may not use this file except in compliance
+ * with the License.
+ *
+ * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
+ * or http://www.opensolaris.org/os/licensing.
+ * See the License for the specific language governing permissions
+ * and limitations under the License.
+ *
+ * When distributing Covered Code, include this CDDL HEADER in each
+ * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
+ * If applicable, add the following below this CDDL HEADER, with the
+ * fields enclosed by brackets "[]" replaced with your own identifying
+ * information: Portions Copyright [yyyy] [name of copyright owner]
+ *
+ * CDDL HEADER END
+ */
+/*
+ * Copyright 2005 Sun Microsystems, Inc. All rights reserved.
+ * Use is subject to license terms.
+ */
+
+#pragma ident "%Z%%M% %I% %E% SMI"
+
+#include <sys/types.h>
+#include <sys/sysmacros.h>
+#include <sys/open.h>
+#include <sys/param.h>
+#include <sys/machparam.h>
+#include <sys/systm.h>
+#include <sys/signal.h>
+#include <sys/cred.h>
+#include <sys/user.h>
+#include <sys/proc.h>
+#include <sys/vnode.h>
+#include <sys/uio.h>
+#include <sys/buf.h>
+#include <sys/file.h>
+#include <sys/kmem.h>
+#include <sys/stat.h>
+#include <sys/stream.h>
+#include <sys/stropts.h>
+#include <sys/strsubr.h>
+#include <sys/strsun.h>
+#include <inet/common.h>
+#include <inet/mi.h>
+#include <inet/nd.h>
+#include <sys/poll.h>
+#include <sys/utsname.h>
+#include <sys/debug.h>
+#include <sys/conf.h>
+#include <sys/ddi.h>
+#include <sys/sunddi.h>
+#include <sys/errno.h>
+#include <sys/modctl.h>
+#include <sys/machsystm.h>
+#include <sys/promif.h>
+#include <sys/prom_plat.h>
+#include <sys/obpdefs.h>
+#include <vm/seg_kmem.h>
+#include <vm/seg_kp.h>
+#include <sys/kstat.h>
+#include <sys/membar.h>
+#include <sys/ivintr.h>
+#include <sys/vm_machparam.h>
+#include <sys/x_call.h>
+#include <sys/cpuvar.h>
+#include <sys/archsystm.h>
+#include <sys/dmv.h>
+
+#include <sys/idn.h>
+#include <sys/idn_xf.h>
+#include <sys/cpu_sgnblk_defs.h>
+#include <sys/cpu_sgn.h>
+
+struct idn_gkstat sg_kstat;
+
+#define MBXTBL_PART_REPORT ((caddr_t)1)
+#define MBXTBL_FULL_REPORT ((caddr_t)2)
+
+idn_domain_t idn_domain[MAX_DOMAINS];
+idn_global_t idn;
+int idn_debug;
+int idn_snoop;
+int idn_history;
+
+typedef enum {
+ IDN_GPROPS_OKAY,
+ IDN_GPROPS_UNCHECKED,
+ IDN_GPROPS_ERROR
+} idn_gprops_t;
+
+struct idn_history idnhlog;
+
+/*
+ * IDN "tunables".
+ */
+int idn_smr_size;
+int idn_nwr_size;
+int idn_lowat;
+int idn_hiwat;
+int idn_protocol_nservers;
+int idn_awolmsg_interval;
+int idn_smr_bufsize;
+int idn_slab_bufcount;
+int idn_slab_prealloc;
+int idn_slab_maxperdomain;
+int idn_slab_mintotal;
+int idn_window_max;
+int idn_window_incr;
+int idn_window_emax;
+int idn_reclaim_min;
+int idn_reclaim_max;
+int idn_mbox_per_net;
+int idn_max_nets;
+
+int idn_netsvr_spin_count;
+int idn_netsvr_wait_min;
+int idn_netsvr_wait_max;
+int idn_netsvr_wait_shift;
+
+int idn_checksum;
+
+int idn_msgwait_nego;
+int idn_msgwait_cfg;
+int idn_msgwait_con;
+int idn_msgwait_fin;
+int idn_msgwait_cmd;
+int idn_msgwait_data;
+
+int idn_retryfreq_nego;
+int idn_retryfreq_con;
+int idn_retryfreq_fin;
+
+int idn_window_emax; /* calculated */
+int idn_slab_maxperdomain; /* calculated */
+
+/*
+ * DMV interrupt support.
+ */
+int idn_pil;
+int idn_dmv_pending_max;
+idn_dmv_msg_t *idn_iv_queue[NCPU];
+int idn_intr_index[NCPU]; /* idn_handler ONLY */
+static idn_dmv_data_t *idn_dmv_data;
+
+int idn_sigbpil;
+
+idnparam_t idn_param_arr[] = {
+{ 0, 1, 0, /* 0 */ "idn_modunloadable" },
+};
+
+/*
+ * Parameters that are only accessible in a DEBUG driver.
+ */
+static char *idn_param_debug_only[] = {
+#if 0
+ "idn_checksum",
+#endif /* 0 */
+ 0
+};
+
+/*
+ * Parameters that are READ-ONLY.
+ */
+static char *idn_param_read_only[] = {
+#if 0
+ "idn_window_emax",
+ "idn_slab_maxperdomain",
+#endif /* 0 */
+ 0
+};
+
+static struct idn_global_props {
+ int p_min, p_max, p_def;
+ char *p_string;
+ int *p_var;
+} idn_global_props[] = {
+{ 0, 0, 0, "idn_debug", &idn_debug },
+{ 0, 1, 0, "idn_history", &idn_history },
+{ 0, IDN_SMR_MAXSIZE,
+ 0, "idn_smr_size", &idn_smr_size },
+{ 0, IDN_SMR_MAXSIZE,
+ 0, "idn_nwr_size", &idn_nwr_size },
+{ 1, 512*1024,
+ 1, "idn_lowat", &idn_lowat },
+{ 1*1024,
+ 1*1024*1024,
+ 256*1024,
+ "idn_hiwat", &idn_hiwat },
+{ IDN_SMR_BUFSIZE_MIN,
+ IDN_SMR_BUFSIZE_MAX,
+ IDN_SMR_BUFSIZE_DEF,
+ "idn_smr_bufsize", &idn_smr_bufsize },
+{ 4, 1024, 32, "idn_slab_bufcount", &idn_slab_bufcount },
+{ 0, 10, 0, "idn_slab_prealloc", &idn_slab_prealloc },
+{ 2, MAX_DOMAINS,
+ 8, "idn_slab_mintotal", &idn_slab_mintotal },
+{ 8, 256, 64, "idn_window_max", &idn_window_max },
+{ 0, 32, 8, "idn_window_incr", &idn_window_incr },
+{ 1, 128, 5, "idn_reclaim_min", &idn_reclaim_min },
+{ 0, 128, 0, "idn_reclaim_max", &idn_reclaim_max },
+{ 1, IDN_MAXMAX_NETS,
+ 8, "idn_max_nets", &idn_max_nets },
+{ 31, 511, 127, "idn_mbox_per_net", &idn_mbox_per_net },
+{ 0, 1, 1, "idn_checksum", &idn_checksum },
+{ 0, 10000, 500, "idn_netsvr_spin_count",
+ &idn_netsvr_spin_count },
+{ 0, 30*100, 40, "idn_netsvr_wait_min", &idn_netsvr_wait_min },
+{ 0, 60*100, 16*100, "idn_netsvr_wait_max", &idn_netsvr_wait_max },
+{ 1, 5, 1, "idn_netsvr_wait_shift",
+ &idn_netsvr_wait_shift },
+{ 1, MAX_DOMAINS,
+ IDN_PROTOCOL_NSERVERS,
+ "idn_protocol_nservers",
+ &idn_protocol_nservers },
+{ 0, 3600, IDN_AWOLMSG_INTERVAL,
+ "idn_awolmsg_interval", &idn_awolmsg_interval },
+{ 10, 300, IDN_MSGWAIT_NEGO,
+ "idn_msgwait_nego", &idn_msgwait_nego },
+{ 10, 300, IDN_MSGWAIT_CFG,
+ "idn_msgwait_cfg", &idn_msgwait_cfg },
+{ 10, 300, IDN_MSGWAIT_CON,
+ "idn_msgwait_con", &idn_msgwait_con },
+{ 10, 300, IDN_MSGWAIT_FIN,
+ "idn_msgwait_fin", &idn_msgwait_fin },
+{ 10, 300, IDN_MSGWAIT_CMD,
+ "idn_msgwait_cmd", &idn_msgwait_cmd },
+{ 10, 300, IDN_MSGWAIT_DATA,
+ "idn_msgwait_data", &idn_msgwait_data },
+{ 1, 60, IDN_RETRYFREQ_NEGO,
+ "idn_retryfreq_nego", &idn_retryfreq_nego },
+{ 1, 60, IDN_RETRYFREQ_CON,
+ "idn_retryfreq_con", &idn_retryfreq_con },
+{ 1, 60, IDN_RETRYFREQ_FIN,
+ "idn_retryfreq_fin", &idn_retryfreq_fin },
+{ 1, 9, IDN_PIL,
+ "idn_pil", &idn_pil },
+{ 1, 9, IDN_SIGBPIL,
+ "idn_sigbpil", &idn_sigbpil },
+{ 8, 512, IDN_DMV_PENDING_MAX,
+ "idn_dmv_pending_max", &idn_dmv_pending_max },
+{ 0, 0, 0, NULL, NULL }
+};
+
+struct idn *idn_i2s_table[IDN_MAXMAX_NETS << 1];
+clock_t idn_msg_waittime[IDN_NUM_MSGTYPES];
+clock_t idn_msg_retrytime[(int)IDN_NUM_RETRYTYPES];
+
+static caddr_t idn_ndlist; /* head of 'named dispatch' var list */
+
+static int idnattach(dev_info_t *, ddi_attach_cmd_t);
+static int idndetach(dev_info_t *, ddi_detach_cmd_t);
+static int idnopen(register queue_t *, dev_t *, int, int, cred_t *);
+static int idnclose(queue_t *, int, cred_t *);
+static int idnwput(queue_t *, mblk_t *);
+static int idnwsrv(queue_t *);
+static int idnrput(queue_t *, mblk_t *);
+static void idnioctl(queue_t *, mblk_t *);
+static idn_gprops_t idn_check_conf(dev_info_t *dip, processorid_t *cpuid);
+static int idn_size_check();
+static void idn_xmit_monitor_init();
+static void idn_xmit_monitor_deinit();
+static void idn_init_msg_waittime();
+static void idn_init_msg_retrytime();
+static void idn_sigb_setup(cpu_sgnblk_t *sigbp, void *arg);
+static int idn_init(dev_info_t *dip);
+static int idn_deinit();
+static void idn_sigbhandler_create();
+static void idn_sigbhandler_kill();
+static uint_t idn_sigbhandler_wakeup(caddr_t arg1, caddr_t arg2);
+static void idn_sigbhandler_thread(struct sigbintr **sbpp);
+static void idn_sigbhandler(processorid_t cpuid, cpu_sgnblk_t *sgnblkp);
+static int idn_info(idnsb_info_t *sfp);
+static int idn_init_smr();
+static void idn_deinit_smr();
+static int idn_prom_getsmr(uint_t *smrsz, uint64_t *paddrp,
+ uint64_t *sizep);
+static int idn_init_handler();
+static void idn_deinit_handler();
+static uint_t idn_handler(caddr_t unused, caddr_t unused2);
+/*
+ * ioctl services
+ */
+static int idnioc_link(idnop_t *idnop);
+static int idnioc_unlink(idnop_t *idnop);
+static int idn_rw_mem(idnop_t *idnop);
+static int idn_send_ping(idnop_t *idnop);
+
+static void idn_domains_init(struct hwconfig *local_hw);
+static void idn_domains_deinit();
+static void idn_retrytask_init();
+static void idn_retrytask_deinit();
+static void idn_gkstat_init();
+static void idn_gkstat_deinit();
+static int idn_gkstat_update();
+static void idn_timercache_init();
+static void idn_timercache_deinit();
+static void idn_dopers_init();
+static void idn_dopers_deinit();
+
+static void idn_param_cleanup();
+static int idn_param_get(queue_t *q, mblk_t *mp, caddr_t cp, cred_t *cr);
+static int idn_param_set(queue_t *q, mblk_t *mp, char *value, caddr_t cp,
+ cred_t *cr);
+static int idn_param_register(register idnparam_t *idnpa, int count);
+static int idn_slabpool_report(queue_t *wq, mblk_t *mp, caddr_t cp,
+ cred_t *cr);
+static int idn_buffer_report(queue_t *wq, mblk_t *mp, caddr_t cp,
+ cred_t *cr);
+static int idn_mboxtbl_report(queue_t *wq, mblk_t *mp, caddr_t cp,
+ cred_t *cr);
+static int idn_mainmbox_report(queue_t *wq, mblk_t *mp, caddr_t cp,
+ cred_t *cr);
+static void idn_mainmbox_domain_report(queue_t *wq, mblk_t *mp, int domid,
+ idn_mainmbox_t *mmp, char *mbxtype);
+static int idn_global_report(queue_t *wq, mblk_t *mp, caddr_t cp,
+ cred_t *cr);
+static int idn_domain_report(queue_t *wq, mblk_t *mp, caddr_t cp,
+ cred_t *cr);
+static int idn_get_net_binding(queue_t *wq, mblk_t *mp, caddr_t cp,
+ cred_t *cr);
+static int idn_set_net_binding(queue_t *wq, mblk_t *mp, char *value,
+ caddr_t cp, cred_t *cr);
+
+/*
+ * String definitions used for DEBUG and non-DEBUG.
+ */
+const char *idnm_str[] = {
+/* 0 */ "null",
+/* 1 */ "nego",
+/* 2 */ "con",
+/* 3 */ "cfg",
+/* 4 */ "fin",
+/* 5 */ "cmd",
+/* 6 */ "data",
+};
+
+const char *idnds_str[] = {
+/* 0 */ "CLOSED",
+/* 1 */ "NEGO_PEND",
+/* 2 */ "NEGO_SENT",
+/* 3 */ "NEGO_RCVD",
+/* 4 */ "CONFIG",
+/* 5 */ "CON_PEND",
+/* 6 */ "CON_SENT",
+/* 7 */ "CON_RCVD",
+/* 8 */ "CON_READY",
+/* 9 */ "CONNECTED",
+/* 10 */ "FIN_PEND",
+/* 11 */ "FIN_SENT",
+/* 12 */ "FIN_RCVD",
+/* 13 */ "DMAP"
+};
+
+const char *idnxs_str[] = {
+/* 0 */ "PEND",
+/* 1 */ "SENT",
+/* 2 */ "RCVD",
+/* 3 */ "FINAL",
+/* 4 */ "NIL"
+};
+
+const char *idngs_str[] = {
+/* 0 */ "OFFLINE",
+/* 1 */ "CONNECT",
+/* 2 */ "ONLINE",
+/* 3 */ "DISCONNECT",
+/* 4 */ "RECONFIG",
+/* 5 */ "unknown",
+/* 6 */ "unknown",
+/* 7 */ "unknown",
+/* 8 */ "unknown",
+/* 9 */ "unknown",
+/* 10 */ "IGNORE"
+};
+
+const char *idncmd_str[] = {
+/* 0 */ "unknown",
+/* 1 */ "SLABALLOC",
+/* 2 */ "SLABFREE",
+/* 3 */ "SLABREAP",
+/* 4 */ "NODENAME"
+};
+
+const char *idncon_str[] = {
+/* 0 */ "OFF",
+/* 1 */ "NORMAL",
+/* 2 */ "QUERY"
+};
+
+const char *idnfin_str[] = {
+/* 0 */ "OFF",
+/* 1 */ "NORMAL",
+/* 2 */ "FORCE_SOFT",
+/* 3 */ "FORCE_HARD",
+/* 4 */ "QUERY"
+};
+
+const char *idnfinopt_str[] = {
+/* 0 */ "NONE",
+/* 1 */ "UNLINK",
+/* 2 */ "RELINK"
+};
+
+const char *idnfinarg_str[] = {
+/* 0 */ "NONE",
+/* 1 */ "SMRBAD",
+/* 2 */ "CPUCFG",
+/* 3 */ "HWERR",
+/* 4 */ "CFGERR_FATAL",
+/* 5 */ "CFGERR_MTU",
+/* 6 */ "CFGERR_BUF",
+/* 7 */ "CFGERR_SLAB",
+/* 8 */ "CFGERR_NWR",
+/* 9 */ "CFGERR_NETS",
+/* 10 */ "CFGERR_MBOX",
+/* 11 */ "CFGERR_NMCADR",
+/* 12 */ "CFGERR_MCADR",
+/* 13 */ "CFGERR_CKSUM",
+/* 14 */ "CFGERR_SMR",
+};
+
+const char *idnsync_str[] = {
+/* 0 */ "NIL",
+/* 1 */ "CONNECT",
+/* 2 */ "DISCONNECT"
+};
+
+const char *idnreg_str[] = {
+/* 0 */ "REG",
+/* 1 */ "NEW",
+/* 2 */ "QUERY"
+};
+
+const char *idnnack_str[] = {
+/* 0 */ "unknown",
+/* 1 */ "NOCONN",
+/* 2 */ "BADCHAN",
+/* 3 */ "BADCFG",
+/* 4 */ "BADCMD",
+/* 5 */ "RETRY",
+/* 6 */ "DUP",
+/* 7 */ "EXIT",
+/* 8 */ "--reserved1",
+/* 9 */ "--reserved2",
+/* 10 */ "--reserved3"
+};
+
+const char *idnop_str[] = {
+/* 0 */ "DISCONNECTED",
+/* 1 */ "CONNECTED",
+/* 2 */ "ERROR"
+};
+
+const char *chanop_str[] = {
+/* 0 */ "OPEN",
+/* 1 */ "SOFT_CLOSE",
+/* 2 */ "HARD_CLOSE",
+/* 3 */ "OFFLINE",
+/* 4 */ "ONLINE"
+};
+
+const char *chanaction_str[] = {
+/* 0 */ "DETACH",
+/* 1 */ "STOP",
+/* 2 */ "SUSPEND",
+/* 3 */ "RESUME",
+/* 4 */ "RESTART",
+/* 5 */ "ATTACH"
+};
+
+const char *timer_str[] = {
+/* 0 */ "NIL",
+/* 1 */ "MSG"
+};
+
+static struct module_info idnrinfo = {
+ IDNIDNUM, /* mi_idnum */
+ IDNNAME, /* mi_idname */
+ IDNMINPSZ, /* mi_minpsz */
+ IDNMAXPSZ, /* mi_maxpsz */
+ 0, /* mi_hiwat - see IDN_HIWAT */
+ 0 /* mi_lowat - see IDN_LOWAT */
+};
+
+static struct module_info idnwinfo = {
+ IDNIDNUM, /* mi_idnum */
+ IDNNAME, /* mi_idname */
+ IDNMINPSZ, /* mi_minpsz */
+ IDNMAXPSZ, /* mi_maxpsz */
+ 0, /* mi_hiwat - see IDN_HIWAT */
+ 0 /* mi_lowat - see IDN_LOWAT */
+};
+
+static struct qinit idnrinit = {
+ idnrput, /* qi_putp */
+ NULL, /* qi_srvp */
+ idnopen, /* qi_qopen */
+ idnclose, /* qi_qclose */
+ NULL, /* qi_qadmin */
+ &idnrinfo, /* qi_minfo */
+ NULL, /* qi_mstat */
+ NULL, /* qi_rwp */
+ NULL, /* qi_infop */
+ STRUIOT_DONTCARE /* qi_struiot */
+};
+
+static struct qinit idnwinit = {
+ idnwput, /* qi_putp */
+ idnwsrv, /* qi_srvp */
+ NULL, /* qi_qopen */
+ NULL, /* qi_qclose */
+ NULL, /* qi_qadmin */
+ &idnwinfo, /* qi_minfo */
+ NULL, /* qi_mstat */
+ NULL, /* qi_rwp */
+ NULL, /* qi_infop */
+ STRUIOT_DONTCARE /* qi_struiot */
+};
+
+struct streamtab idninfo = {
+ &idnrinit, /* st_rdinit */
+ &idnwinit, /* st_wrinit */
+ NULL, /* st_muxrinit */
+ NULL, /* st_muxwinit */
+};
+
+/*
+ * Module linkage information (cb_ops & dev_ops) for the kernel.
+ */
+
+static struct cb_ops cb_idnops = {
+ nulldev, /* cb_open */
+ nulldev, /* cb_close */
+ nodev, /* cb_strategy */
+ nodev, /* cb_print */
+ nodev, /* cb_dump */
+ nodev, /* cb_read */
+ nodev, /* cb_write */
+ nodev, /* cb_ioctl */
+ nodev, /* cb_devmap */
+ nodev, /* cb_mmap */
+ nodev, /* cb_segmap */
+ nochpoll, /* cb_chpoll */
+ ddi_prop_op, /* cb_prop_op */
+ &idninfo, /* cb_stream */
+ D_MP, /* cb_flag */
+ CB_REV, /* cb_rev */
+ nodev, /* cb_aread */
+ nodev, /* cb_awrite */
+};
+
+static struct dev_ops idnops = {
+ DEVO_REV, /* devo_rev */
+ 0, /* devo_refcnt */
+ ddi_no_info, /* devo_getinfo */
+ nulldev, /* devo_identify */
+ nulldev, /* devo_probe */
+ idnattach, /* devo_attach */
+ idndetach, /* devo_detach */
+ nodev, /* devo_reset */
+ &cb_idnops, /* devo_cb_ops */
+ (struct bus_ops *)NULL, /* devo_bus_ops */
+ NULL /* devo_power */
+};
+
+extern cpuset_t cpu_ready_set;
+
+static struct modldrv modldrv = {
+ &mod_driverops, /* This module is a pseudo driver */
+ IDNDESC " %I%",
+ &idnops
+};
+
+static struct modlinkage modlinkage = {
+ MODREV_1,
+ &modldrv,
+ NULL
+};
+
+/*
+ * --------------------------------------------------
+ */
+int
+_init(void)
+{
+ idn.version = IDN_VERSION;
+
+ return (mod_install(&modlinkage));
+}
+
+int
+_fini(void)
+{
+ return (mod_remove(&modlinkage));
+}
+
+int
+_info(struct modinfo *modinfop)
+{
+ return (mod_info(&modlinkage, modinfop));
+}
+
+/*
+ * ----------------------------------------------
+ */
+static int
+idnattach(dev_info_t *dip, ddi_attach_cmd_t cmd)
+{
+ int instance;
+ int doinit = 0;
+ processorid_t bcpuid;
+ struct idn *sip;
+ struct idnstr *stp;
+ procname_t proc = "idnattach";
+
+
+#ifndef lint
+ ASSERT(sizeof (idnsb_t) == IDNSB_SIZE);
+ ASSERT((uint_t)&((struct idnsb *)0)->id_hwchkpt[0] == 0x40);
+#endif /* lint */
+
+ switch (cmd) {
+ case DDI_RESUME:
+ sip = ddi_get_driver_private(dip);
+ /*
+ * sip may have not yet been set if the
+ * OBP environment variable (idn-smr-size)
+ * was not set.
+ */
+ if (sip == NULL)
+ return (DDI_FAILURE);
+ /*
+ * RESUME IDN services.
+ */
+ IDN_GLOCK_SHARED();
+ if (idn.state != IDNGS_OFFLINE) {
+ cmn_err(CE_WARN,
+ "IDN: 101: not in expected OFFLINE state "
+ "for DDI_RESUME");
+ ASSERT(0);
+ }
+ IDN_GUNLOCK();
+
+ /*
+ * RESUME DLPI services.
+ */
+ sip->si_flags &= ~IDNSUSPENDED;
+
+ rw_enter(&idn.struprwlock, RW_READER);
+ for (stp = idn.strup; stp; stp = stp->ss_nextp)
+ if (stp->ss_sip == sip) {
+ doinit = 1;
+ break;
+ }
+ rw_exit(&idn.struprwlock);
+ if (doinit)
+ (void) idndl_init(sip);
+
+ return (DDI_SUCCESS);
+
+ case DDI_ATTACH:
+ break;
+
+ default:
+ return (DDI_FAILURE);
+ }
+
+ instance = ddi_get_instance(dip);
+
+ PR_DRV("%s: instance = %d\n", proc, instance);
+
+ if (idn_check_conf(dip, &bcpuid) == IDN_GPROPS_ERROR)
+ return (DDI_FAILURE);
+
+ mutex_enter(&idn.siplock);
+
+ if (ddi_create_minor_node(dip, IDNNAME, S_IFCHR, instance,
+ DDI_NT_NET, CLONE_DEV) == DDI_FAILURE) {
+ mutex_exit(&idn.siplock);
+ return (DDI_FAILURE);
+ }
+
+ if (idn.smr.ready == 0) {
+ if (idn_init_smr() == 0) {
+ idn.enabled = 1;
+#ifdef DEBUG
+ cmn_err(CE_NOTE, "!IDN: Driver enabled");
+#endif /* DEBUG */
+ } else {
+ cmn_err(CE_NOTE,
+ "!IDN: 102: driver disabled "
+ "- check OBP environment "
+ "(idn-smr-size)");
+ mutex_exit(&idn.siplock);
+ return (DDI_SUCCESS);
+ }
+ }
+
+ ASSERT(idn.smr.ready || idn.enabled);
+
+ if (idn.dip == NULL) {
+ doinit = 1;
+
+ if (idn_size_check()) {
+ idn_deinit_smr();
+ ddi_remove_minor_node(dip, NULL);
+ mutex_exit(&idn.siplock);
+ return (DDI_FAILURE);
+ }
+
+ if (idn_init(dip)) {
+ idn_deinit_smr();
+ ddi_remove_minor_node(dip, NULL);
+ mutex_exit(&idn.siplock);
+ return (DDI_FAILURE);
+ }
+ }
+
+ ASSERT(idn.dip);
+
+ /*
+ * This must occur _after_ idn_init() since
+ * it assumes idn_chanservers_init() has been
+ * called.
+ */
+ idn_chanserver_bind(ddi_get_instance(dip), bcpuid);
+
+ /*
+ * DLPI supporting stuff.
+ */
+ sip = GETSTRUCT(struct idn, 1);
+ sip->si_dip = dip;
+ ddi_set_driver_private(dip, sip);
+ sip->si_nextp = idn.sip;
+ idn.sip = sip;
+ IDN_SET_INST2SIP(instance, sip);
+ mutex_exit(&idn.siplock);
+
+ if (doinit)
+ idndl_dlpi_init(); /* initializes idninfoack */
+ /*
+ * Get our local IDN ethernet address.
+ */
+ idndl_localetheraddr(sip, &sip->si_ouraddr);
+ idndl_statinit(sip);
+
+ if (doinit) {
+ idn_gkstat_init();
+ /*
+ * Add our sigblock SSP interrupt handler.
+ */
+ mutex_enter(&idn.sigbintr.sb_mutex);
+ idn_sigbhandler_create();
+ mutex_exit(&idn.sigbintr.sb_mutex);
+
+ if (sgnblk_poll_register(idn_sigbhandler) == 0) {
+ mutex_enter(&idn.sigbintr.sb_mutex);
+ idn_sigbhandler_kill();
+ idn.sigbintr.sb_cpuid = (uchar_t)-1;
+ idn.sigbintr.sb_busy = IDNSIGB_INACTIVE;
+ mutex_exit(&idn.sigbintr.sb_mutex);
+
+ idn_gkstat_deinit();
+
+ mutex_enter(&idn.siplock);
+ (void) idn_deinit();
+ IDN_SET_INST2SIP(instance, NULL);
+ idn.sip = sip->si_nextp;
+ mutex_exit(&idn.siplock);
+
+ ddi_remove_minor_node(dip, NULL);
+
+ return (DDI_FAILURE);
+ }
+ /*
+ * We require sigblkp[cpu0] to be mapped for hardware
+ * configuration determination and also auto-linking
+ * on bootup.
+ */
+ if (sgnblk_poll_reference(idn_sigb_setup, NULL) != 0) {
+ sgnblk_poll_unregister(idn_sigbhandler);
+ mutex_enter(&idn.sigbintr.sb_mutex);
+ idn_sigbhandler_kill();
+ idn.sigbintr.sb_cpuid = (uchar_t)-1;
+ idn.sigbintr.sb_busy = IDNSIGB_INACTIVE;
+ mutex_exit(&idn.sigbintr.sb_mutex);
+
+ idn_gkstat_deinit();
+
+ mutex_enter(&idn.siplock);
+ (void) idn_deinit();
+ IDN_SET_INST2SIP(instance, NULL);
+ idn.sip = sip->si_nextp;
+ mutex_exit(&idn.siplock);
+
+ ddi_remove_minor_node(dip, NULL);
+
+ cmn_err(CE_WARN,
+ "IDN: 103: unable to reference sigblock area");
+
+ return (DDI_FAILURE);
+ }
+
+ idn_init_autolink();
+ }
+
+ ddi_report_dev(dip);
+
+ return (DDI_SUCCESS);
+}
+
+/*
+ * ----------------------------------------------
+ */
+static int
+idndetach(dev_info_t *dip, ddi_detach_cmd_t cmd)
+{
+ int err = 0;
+ int instance;
+ struct idn *sip, *hsip, *tsip;
+ procname_t proc = "idndetach";
+
+ sip = ddi_get_driver_private(dip);
+ instance = ddi_get_instance(dip);
+
+ switch (cmd) {
+ case DDI_SUSPEND:
+ if (sip == NULL)
+ return (DDI_FAILURE);
+ /*
+ * SUSPEND IDN services.
+ * - Actually don't suspend anything, we just
+ * make sure we're not connected per DR protocol.
+ * If we really wanted to suspend it should
+ * be done _after_ DLPI is suspended so that
+ * we're not competing with that traffic.
+ */
+ IDN_GLOCK_SHARED();
+
+ if (idn.state != IDNGS_OFFLINE) {
+ int d;
+
+ cmn_err(CE_WARN,
+ "IDN: 104: cannot suspend while active "
+ "(state = %s)",
+ idngs_str[idn.state]);
+
+ for (d = 0; d < MAX_DOMAINS; d++) {
+ idn_domain_t *dp;
+
+ dp = &idn_domain[d];
+ if (dp->dcpu < 0)
+ continue;
+
+ cmn_err(CE_CONT,
+ "IDN: 121: domain %d (CPU %d, name "
+ "\"%s\", state %s)\n",
+ d, dp->dcpu, dp->dname,
+ idnds_str[dp->dstate]);
+ }
+ err = 1;
+ }
+
+ IDN_GUNLOCK();
+
+ if (err)
+ return (DDI_FAILURE);
+ /*
+ * SUSPEND DLPI services.
+ */
+ sip->si_flags |= IDNSUSPENDED;
+
+ idndl_uninit(sip);
+
+ return (DDI_FAILURE);
+
+ case DDI_DETACH:
+ if (idn.enabled == 0) {
+ ddi_remove_minor_node(dip, NULL);
+ ASSERT(idn.dip == NULL);
+ return (DDI_SUCCESS);
+ }
+ if (!IDN_MODUNLOADABLE)
+ return (DDI_FAILURE);
+ break;
+
+ default:
+ return (DDI_FAILURE);
+ }
+
+ PR_DRV("%s: instance = %d\n", proc, instance);
+
+ if (sip == NULL) {
+ /*
+ * No resources allocated.
+ */
+ return (DDI_SUCCESS);
+ }
+
+ mutex_enter(&idn.siplock);
+ if (idn.sip && (idn.sip->si_nextp == NULL)) {
+ /*
+ * This is our last stream connection
+ * going away. Time to deinit and flag
+ * the SSP we're (IDN) DOWN.
+ */
+ if (idn_deinit()) {
+ /*
+ * Must still be active.
+ */
+ mutex_exit(&idn.siplock);
+ return (DDI_FAILURE);
+ }
+ idn_deinit_autolink();
+ /*
+ * Remove our sigblock SSP interrupt handler.
+ */
+ sgnblk_poll_unregister(idn_sigbhandler);
+ mutex_enter(&idn.sigbintr.sb_mutex);
+ idn_sigbhandler_kill();
+ idn.sigbintr.sb_cpuid = (uchar_t)-1;
+ idn.sigbintr.sb_busy = IDNSIGB_NOTREADY;
+ mutex_exit(&idn.sigbintr.sb_mutex);
+ /*
+ * Remove our reference to the sigblock area.
+ */
+ sgnblk_poll_unreference(idn_sigb_setup);
+ idn_gkstat_deinit();
+ }
+
+ ddi_remove_minor_node(dip, NULL);
+
+ /*
+ * Remove this instance from our linked list.
+ */
+ IDN_SET_INST2SIP(instance, NULL);
+ if ((hsip = tsip = idn.sip) == sip) {
+ idn.sip = sip->si_nextp;
+ } else {
+ for (; hsip && (sip != hsip); tsip = hsip,
+ hsip = hsip->si_nextp)
+ ;
+ if (hsip)
+ tsip->si_nextp = hsip->si_nextp;
+ }
+ mutex_exit(&idn.siplock);
+ if (sip->si_ksp)
+ kstat_delete(sip->si_ksp);
+
+ ddi_set_driver_private(dip, NULL);
+
+ FREESTRUCT(sip, struct idn, 1);
+
+ return (DDI_SUCCESS);
+}
+
+/*
+ * ----------------------------------------------
+ */
+static idn_gprops_t
+idn_check_conf(dev_info_t *dip, processorid_t *cpuid)
+{
+ static idn_gprops_t global_props = IDN_GPROPS_UNCHECKED;
+
+ if (global_props == IDN_GPROPS_UNCHECKED) {
+ int p;
+
+ global_props = IDN_GPROPS_OKAY;
+
+ for (p = 0; idn_global_props[p].p_string; p++) {
+ char *str;
+ int *var;
+ int val, v_min, v_max, v_def;
+
+ str = idn_global_props[p].p_string;
+ var = (int *)idn_global_props[p].p_var;
+ v_min = idn_global_props[p].p_min;
+ v_max = idn_global_props[p].p_max;
+ v_def = idn_global_props[p].p_def;
+ ASSERT(str && var);
+
+ val = ddi_prop_get_int(DDI_DEV_T_ANY, dip,
+ DDI_PROP_DONTPASS |
+ DDI_PROP_NOTPROM,
+ str, v_def);
+ if ((v_min != v_max) &&
+ ((val < v_min) || (val > v_max))) {
+ cmn_err(CE_WARN,
+ "IDN: 105: driver parameter "
+ "(%s) specified (%d) out of "
+ "range [%d - %d]",
+ str, val, v_min, v_max);
+ global_props = IDN_GPROPS_ERROR;
+ } else {
+ *var = val;
+ }
+ }
+ }
+
+ *cpuid = ddi_prop_get_int(DDI_DEV_T_ANY, dip,
+ DDI_PROP_DONTPASS | DDI_PROP_NOTPROM,
+ "bind_cpu", -1);
+
+ return (global_props);
+}
+
+static int
+idn_size_check()
+{
+ int i, cnt;
+ int rv = 0;
+ ulong_t mboxareasize;
+ int max_num_slabs;
+ procname_t proc = "idn_size_check";
+
+ if (IDN_NWR_SIZE == 0)
+ IDN_NWR_SIZE = IDN_SMR_SIZE;
+
+ if (IDN_NWR_SIZE > IDN_SMR_SIZE) {
+ cmn_err(CE_WARN,
+ "IDN: 106: idn_nwr_size(%d) > idn_smr_size(%d)"
+ " - Limiting to %d MB",
+ IDN_NWR_SIZE, IDN_SMR_SIZE, IDN_SMR_SIZE);
+ IDN_NWR_SIZE = IDN_SMR_SIZE;
+ }
+
+ if (MB2B(IDN_NWR_SIZE) < IDN_SLAB_SIZE) {
+ cmn_err(CE_WARN,
+ "IDN: 107: memory region(%lu) < slab size(%u)",
+ MB2B(IDN_NWR_SIZE), IDN_SLAB_SIZE);
+ rv = -1;
+ }
+
+ if (IDN_LOWAT >= IDN_HIWAT) {
+ cmn_err(CE_WARN,
+ "IDN: 108: idn_lowat(%d) >= idn_hiwat(%d)",
+ IDN_LOWAT, IDN_HIWAT);
+ rv = -1;
+ }
+
+ mboxareasize = (ulong_t)(IDN_MBOXAREA_SIZE + (IDN_SMR_BUFSIZE - 1));
+ mboxareasize &= ~((ulong_t)IDN_SMR_BUFSIZE - 1);
+#ifdef DEBUG
+ if ((ulong_t)IDN_SLAB_SIZE < mboxareasize) {
+ PR_DRV("%s: slab size(%ld) < mailbox area(%ld)",
+ proc, IDN_SLAB_SIZE, mboxareasize);
+ /* not fatal */
+ }
+#endif /* DEBUG */
+
+ if ((mboxareasize + (ulong_t)IDN_SLAB_SIZE) > MB2B(IDN_NWR_SIZE)) {
+ cmn_err(CE_WARN,
+ "IDN: 109: mailbox area(%lu) + slab size(%u) "
+ "> nwr region(%lu)",
+ mboxareasize, IDN_SLAB_SIZE,
+ MB2B(IDN_NWR_SIZE));
+ rv = -1;
+ }
+
+ max_num_slabs = (int)((MB2B(IDN_NWR_SIZE) - mboxareasize) /
+ (ulong_t)IDN_SLAB_SIZE);
+ if (max_num_slabs < IDN_SLAB_MINTOTAL) {
+ cmn_err(CE_WARN,
+ "IDN: 110: maximum number of slabs(%d) < "
+ "minimum required(%d)",
+ max_num_slabs, IDN_SLAB_MINTOTAL);
+ rv = -1;
+ } else {
+ IDN_SLAB_MAXPERDOMAIN = max_num_slabs / IDN_SLAB_MINTOTAL;
+ }
+
+#if 0
+ if ((IDN_MTU + sizeof (struct ether_header)) > IDN_DATA_SIZE) {
+ cmn_err(CE_WARN,
+ "IDN: (IDN_MTU(%d) + ether_header(%d)) "
+ "> IDN_DATA_SIZE(%lu)",
+ IDN_MTU, sizeof (struct ether_header),
+ IDN_DATA_SIZE);
+ rv = -1;
+ }
+#endif /* 0 */
+
+ if (IDN_SMR_BUFSIZE & (IDN_ALIGNSIZE - 1)) {
+ cmn_err(CE_WARN,
+ "IDN: 111: idn_smr_bufsize(%d) not on a "
+ "64 byte boundary", IDN_SMR_BUFSIZE);
+ rv = -1;
+ }
+
+ for (i = cnt = 0;
+ (cnt <= 1) && (((ulong_t)1 << i) < MB2B(IDN_NWR_SIZE));
+ i++)
+ if ((1 << i) & IDN_SMR_BUFSIZE)
+ cnt++;
+ if ((i > 0) && (!cnt || (cnt > 1))) {
+ cmn_err(CE_WARN,
+ "IDN: 112: idn_smr_bufsize(%d) not a power of 2",
+ IDN_SMR_BUFSIZE);
+ rv = -1;
+ }
+
+ if ((IDN_MBOX_PER_NET & 1) == 0) {
+ cmn_err(CE_WARN,
+ "IDN: 113: idn_mbox_per_net(%d) must be an "
+ "odd number", IDN_MBOX_PER_NET);
+ rv = -1;
+ }
+
+ if (idn.nchannels > 0)
+ IDN_WINDOW_EMAX = IDN_WINDOW_MAX +
+ ((idn.nchannels - 1) * IDN_WINDOW_INCR);
+
+ if (IDN_NETSVR_WAIT_MIN > IDN_NETSVR_WAIT_MAX) {
+ cmn_err(CE_WARN,
+ "IDN: 115: idn_netsvr_wait_min(%d) cannot be "
+ "greater than idn_netsvr_wait_max(%d)",
+ IDN_NETSVR_WAIT_MIN,
+ IDN_NETSVR_WAIT_MAX);
+ rv = -1;
+ }
+
+ return (rv);
+}
+
+static int
+idn_init_smr()
+{
+ uint64_t obp_paddr;
+ uint64_t obp_size; /* in Bytes */
+ uint_t smr_size; /* in MBytes */
+ pgcnt_t npages;
+ procname_t proc = "idn_init_smr";
+
+ if (idn.smr.ready)
+ return (0);
+
+ if (idn_prom_getsmr(&smr_size, &obp_paddr, &obp_size) < 0)
+ return (-1);
+
+ PR_PROTO("%s: smr_size = %d, obp_paddr = 0x%llx, obp_size = 0x%llx\n",
+ proc, smr_size, obp_paddr, obp_size);
+
+ if (IDN_SMR_SIZE)
+ smr_size = MIN(smr_size, IDN_SMR_SIZE);
+
+ npages = btopr(MB2B(smr_size));
+
+ idn.smr.prom_paddr = obp_paddr;
+ idn.smr.prom_size = obp_size;
+ idn.smr.vaddr = vmem_alloc(heap_arena, ptob(npages), VM_SLEEP);
+ ASSERT(((ulong_t)idn.smr.vaddr & MMU_PAGEOFFSET) == 0);
+ idn.smr.locpfn = (pfn_t)(obp_paddr >> MMU_PAGESHIFT);
+ idn.smr.rempfn = idn.smr.rempfnlim = PFN_INVALID;
+ IDN_SMR_SIZE = smr_size;
+
+ PR_PROTO("%s: smr vaddr = %p\n", proc, idn.smr.vaddr);
+
+ smr_remap(&kas, idn.smr.vaddr, idn.smr.locpfn, IDN_SMR_SIZE);
+
+ idn.localid = PADDR_TO_DOMAINID(obp_paddr);
+
+ idn.smr.ready = 1;
+
+ return (0);
+}
+
+static void
+idn_deinit_smr()
+{
+ pgcnt_t npages;
+
+ if (idn.smr.ready == 0)
+ return;
+
+ smr_remap(&kas, idn.smr.vaddr, PFN_INVALID, IDN_SMR_SIZE);
+
+ npages = btopr(MB2B(IDN_SMR_SIZE));
+
+ vmem_free(heap_arena, idn.smr.vaddr, ptob(npages));
+
+ idn.localid = IDN_NIL_DOMID;
+
+ IDN_SMR_SIZE = 0;
+
+ idn.smr.ready = 0;
+}
+
+/*ARGSUSED1*/
+static void
+idn_sigb_setup(cpu_sgnblk_t *sigbp, void *arg)
+{
+ procname_t proc = "idn_sigb_setup";
+
+ PR_PROTO("%s: Setting sigb to %p\n", proc, sigbp);
+
+ mutex_enter(&idn.idnsb_mutex);
+ if (sigbp == NULL) {
+ idn.idnsb = NULL;
+ idn.idnsb_eventp = NULL;
+ mutex_exit(&idn.idnsb_mutex);
+ return;
+ }
+ idn.idnsb_eventp = (idnsb_event_t *)sigbp->sigb_idn;
+ idn.idnsb = (idnsb_t *)&idn.idnsb_eventp->idn_reserved1;
+ mutex_exit(&idn.idnsb_mutex);
+}
+
+static int
+idn_init(dev_info_t *dip)
+{
+ struct hwconfig local_hw;
+ procname_t proc = "idn_init";
+
+
+ ASSERT(MUTEX_HELD(&idn.siplock));
+
+ if (!idn.enabled) {
+ cmn_err(CE_WARN,
+ "IDN: 117: IDN not enabled");
+ return (-1);
+ }
+
+ if (idn.dip != NULL) {
+ PR_DRV("%s: already initialized (dip = 0x%x)\n",
+ proc, (uint_t)idn.dip);
+ return (0);
+ }
+
+ /*
+ * Determine our local domain's hardware configuration.
+ */
+ if (get_hw_config(&local_hw)) {
+ cmn_err(CE_WARN,
+ "IDN: 118: hardware config not appropriate");
+ return (-1);
+ }
+
+ PR_DRV("%s: locpfn = 0x%lx\n", proc, idn.smr.locpfn);
+ PR_DRV("%s: rempfn = 0x%lx\n", proc, idn.smr.rempfn);
+ PR_DRV("%s: smrsize = %d MB\n", proc, IDN_SMR_SIZE);
+
+ rw_init(&idn.grwlock, NULL, RW_DEFAULT, NULL);
+ rw_init(&idn.struprwlock, NULL, RW_DEFAULT, NULL);
+ mutex_init(&idn.sync.sz_mutex, NULL, MUTEX_DEFAULT, NULL);
+ mutex_init(&idn.sipwenlock, NULL, MUTEX_DEFAULT, NULL);
+
+ /*
+ * Calculate proper value for idn.bframe_shift.
+ * Kind of hokey as it assume knowledge of the format
+ * of the idnparam_t structure.
+ */
+ {
+ int s;
+
+ for (s = 0; (1 << s) < IDN_SMR_BUFSIZE_MIN; s++)
+ ;
+ idn.bframe_shift = s;
+ PR_DRV("%s: idn.bframe_shift = %d, minbuf = %ld\n",
+ proc, idn.bframe_shift, IDN_SMR_BUFSIZE_MIN);
+
+ ASSERT((uint_t)IDN_OFFSET2BFRAME(MB2B(idn_smr_size)) <
+ (1 << 24));
+ }
+
+ idn_xmit_monitor_init();
+
+ /*
+ * Initialize the domain op (dopers) stuff.
+ */
+ idn_dopers_init();
+
+ /*
+ * Initialize the timer (kmem) cache used for timeout
+ * structures.
+ */
+ idn_timercache_init();
+
+ /*
+ * Initialize the slab waiting areas.
+ */
+ (void) smr_slabwaiter_init();
+
+ /*
+ * Initialize retryjob kmem cache.
+ */
+ idn_retrytask_init();
+
+ idn_init_msg_waittime();
+ idn_init_msg_retrytime();
+
+ /*
+ * Initialize idn_domains[] and local domains information
+ * include idn_global information.
+ */
+ idn_domains_init(&local_hw);
+
+ /*
+ * Start up IDN protocol servers.
+ */
+ if (idn_protocol_init(idn_protocol_nservers) <= 0) {
+ cmn_err(CE_WARN,
+ "IDN: 119: failed to initialize %d protocol servers",
+ idn_protocol_nservers);
+ idn_domains_deinit();
+ idn_retrytask_deinit();
+ smr_slabwaiter_deinit();
+ idn_timercache_deinit();
+ idn_dopers_deinit();
+ idn_xmit_monitor_deinit();
+ mutex_destroy(&idn.sipwenlock);
+ mutex_destroy(&idn.sync.sz_mutex);
+ rw_destroy(&idn.grwlock);
+ rw_destroy(&idn.struprwlock);
+ return (-1);
+ }
+
+ /*
+ * Initialize chan_servers array.
+ */
+ (void) idn_chanservers_init();
+
+ /*
+ * Need to register the IDN handler with the DMV subsystem.
+ *
+ * Need to prevent the IDN driver from being unloaded
+ * once loaded since DMV's may come in at any time.
+ * If the driver is not loaded and the idn_dmv_handler
+ * has been registered with the DMV, system will crash.
+ */
+ (void) idn_init_handler();
+
+ idn.dip = dip;
+ IDN_GLOCK_EXCL();
+ IDN_GSTATE_TRANSITION(IDNGS_OFFLINE);
+ IDN_GUNLOCK();
+
+ return (0);
+}
+
+static int
+idn_deinit()
+{
+ procname_t proc = "idn_deinit";
+
+ ASSERT(MUTEX_HELD(&idn.siplock));
+
+ IDN_GLOCK_EXCL();
+
+ if (idn.state != IDNGS_OFFLINE) {
+ int d;
+
+ cmn_err(CE_WARN,
+ "IDN: 120: cannot deinit while active "
+ "(state = %s)", idngs_str[idn.state]);
+
+ for (d = 0; d < MAX_DOMAINS; d++) {
+ idn_domain_t *dp;
+
+ dp = &idn_domain[d];
+ if (dp->dcpu < 0)
+ continue;
+
+ cmn_err(CE_CONT,
+ "IDN: 121: domain %d (CPU %d, "
+ "name \"%s\", state %s)\n",
+ d, dp->dcpu, dp->dname,
+ idnds_str[dp->dstate]);
+ }
+ IDN_GUNLOCK();
+ return (-1);
+ }
+
+ if (idn.dip == NULL) {
+ PR_DRV("%s: already deinitialized\n", proc);
+ IDN_GUNLOCK();
+ return (0);
+ }
+
+ IDN_GSTATE_TRANSITION(IDNGS_IGNORE);
+
+ IDN_GUNLOCK();
+
+ idn_xmit_monitor_deinit();
+
+ idn_deinit_handler();
+
+ idn_chanservers_deinit();
+
+ idn.nchannels = 0;
+ ASSERT(idn.chan_servers == NULL);
+
+ smr_slabpool_deinit();
+
+ idn_protocol_deinit();
+
+ idn_domains_deinit();
+
+ smr_slabwaiter_deinit();
+
+ idn_retrytask_deinit();
+
+ idn_timercache_deinit();
+
+ idn_dopers_deinit();
+
+ ASSERT(idn.localid == IDN_NIL_DOMID);
+
+ IDN_SET_MASTERID(IDN_NIL_DOMID);
+
+ idn_deinit_smr();
+
+ mutex_destroy(&idn.sipwenlock);
+ mutex_destroy(&idn.sync.sz_mutex);
+ rw_destroy(&idn.grwlock);
+ rw_destroy(&idn.struprwlock);
+
+ idn.dip = NULL;
+
+ return (0);
+}
+
+static void
+idn_xmit_monitor_init()
+{
+ mutex_init(&idn.xmit_lock, NULL, MUTEX_DEFAULT, NULL);
+ idn.xmit_tid = (timeout_id_t)NULL;
+ CHANSET_ZERO(idn.xmit_chanset_wanted);
+}
+
+static void
+idn_xmit_monitor_deinit()
+{
+ timeout_id_t tid;
+
+ mutex_enter(&idn.xmit_lock);
+ CHANSET_ZERO(idn.xmit_chanset_wanted);
+ if ((tid = idn.xmit_tid) != (timeout_id_t)NULL) {
+ idn.xmit_tid = (timeout_id_t)NULL;
+ mutex_exit(&idn.xmit_lock);
+ (void) untimeout(tid);
+ } else {
+ mutex_exit(&idn.xmit_lock);
+ }
+ mutex_destroy(&idn.xmit_lock);
+}
+
+static void
+idn_init_msg_waittime()
+{
+ idn_msg_waittime[IDNP_NULL] = -1;
+ idn_msg_waittime[IDNP_NEGO] = idn_msgwait_nego * hz;
+ idn_msg_waittime[IDNP_CFG] = idn_msgwait_cfg * hz;
+ idn_msg_waittime[IDNP_CON] = idn_msgwait_con * hz;
+ idn_msg_waittime[IDNP_FIN] = idn_msgwait_fin * hz;
+ idn_msg_waittime[IDNP_CMD] = idn_msgwait_cmd * hz;
+ idn_msg_waittime[IDNP_DATA] = idn_msgwait_data * hz;
+}
+
+static void
+idn_init_msg_retrytime()
+{
+ idn_msg_retrytime[(int)IDNRETRY_NIL] = -1;
+ idn_msg_retrytime[(int)IDNRETRY_NEGO] = idn_retryfreq_nego * hz;
+ idn_msg_retrytime[(int)IDNRETRY_CON] = idn_retryfreq_con * hz;
+ idn_msg_retrytime[(int)IDNRETRY_CONQ] = idn_retryfreq_con * hz;
+ idn_msg_retrytime[(int)IDNRETRY_FIN] = idn_retryfreq_fin * hz;
+ idn_msg_retrytime[(int)IDNRETRY_FINQ] = idn_retryfreq_fin * hz;
+}
+
+/*
+ * ----------------------------------------------
+ */
+/*ARGSUSED*/
+static int
+idnopen(register queue_t *rq, dev_t *devp, int flag, int sflag, cred_t *crp)
+{
+ register int err = 0;
+ int minordev;
+ struct idnstr *stp, **pstp;
+ procname_t proc = "idnopen";
+
+ ASSERT(sflag != MODOPEN);
+
+ IDN_GLOCK_EXCL();
+
+ rw_enter(&idn.struprwlock, RW_WRITER);
+ mutex_enter(&idn.sipwenlock);
+ pstp = &idn.strup;
+
+ if (idn.enabled == 0) {
+ PR_DRV("%s: Driver disabled (check OBP:idn-smr-size)\n",
+ proc);
+ mutex_exit(&idn.sipwenlock);
+ rw_exit(&idn.struprwlock);
+ IDN_GUNLOCK();
+ return (EACCES);
+ }
+
+ if (!idn_ndlist &&
+ idn_param_register(idn_param_arr, A_CNT(idn_param_arr))) {
+ PR_DRV("%s: failed to register ndd parameters\n", proc);
+ mutex_exit(&idn.sipwenlock);
+ rw_exit(&idn.struprwlock);
+ IDN_GUNLOCK();
+ return (ENOMEM);
+ }
+ IDN_GUNLOCK();
+
+ if (sflag == CLONEOPEN) {
+ minordev = 0;
+ for (stp = *pstp; stp; pstp = &stp->ss_nextp, stp = *pstp) {
+ if (minordev < stp->ss_minor)
+ break;
+ minordev++;
+ }
+ *devp = makedevice(getmajor(*devp), minordev);
+ } else {
+ minordev = getminor(*devp);
+ }
+ if (rq->q_ptr)
+ goto done;
+
+ stp = GETSTRUCT(struct idnstr, 1);
+ stp->ss_rq = rq;
+ stp->ss_minor = minordev;
+ rw_init(&stp->ss_rwlock, NULL, RW_DEFAULT, NULL);
+ /*
+ * DLPI stuff
+ */
+ stp->ss_sip = NULL;
+ stp->ss_state = DL_UNATTACHED;
+ stp->ss_sap = 0;
+ stp->ss_flags = 0;
+ stp->ss_mccount = 0;
+ stp->ss_mctab = NULL;
+
+ /*
+ * Link new entry into list of actives.
+ */
+ stp->ss_nextp = *pstp;
+ *pstp = stp;
+
+ WR(rq)->q_ptr = rq->q_ptr = (void *)stp;
+ /*
+ * Disable automatic enabling of our write service
+ * procedure. We control this explicitly.
+ */
+ noenable(WR(rq));
+
+ /*
+ * Set our STREAMs queue maximum packet size that
+ * we'll accept and our high/low water marks.
+ */
+ (void) strqset(WR(rq), QMAXPSZ, 0, IDN_DATA_SIZE);
+ (void) strqset(WR(rq), QLOWAT, 0, IDN_LOWAT);
+ (void) strqset(WR(rq), QHIWAT, 0, IDN_HIWAT);
+ (void) strqset(rq, QMAXPSZ, 0, IDN_DATA_SIZE);
+ (void) strqset(rq, QLOWAT, 0, IDN_LOWAT);
+ (void) strqset(rq, QHIWAT, 0, IDN_HIWAT);
+
+done:
+ mutex_exit(&idn.sipwenlock);
+ rw_exit(&idn.struprwlock);
+
+ (void) qassociate(rq, -1);
+ qprocson(rq);
+
+ return (err);
+}
+
+/*
+ * ----------------------------------------------
+ */
+/*ARGSUSED1*/
+static int
+idnclose(queue_t *rq, int flag, cred_t *crp)
+{
+ struct idnstr *stp, **pstp;
+
+ ASSERT(rq->q_ptr);
+
+ qprocsoff(rq);
+ /*
+ * Guaranteed to be single threaded with respect
+ * to this stream at this point.
+ */
+
+ stp = (struct idnstr *)rq->q_ptr;
+
+ if (stp->ss_sip)
+ idndl_dodetach(stp);
+
+ rw_enter(&idn.struprwlock, RW_WRITER);
+ mutex_enter(&idn.sipwenlock);
+ pstp = &idn.strup;
+ for (stp = *pstp; stp; pstp = &stp->ss_nextp, stp = *pstp)
+ if (stp == (struct idnstr *)rq->q_ptr)
+ break;
+ ASSERT(stp);
+ ASSERT(stp->ss_rq == rq);
+ *pstp = stp->ss_nextp;
+
+ rw_destroy(&stp->ss_rwlock);
+ FREESTRUCT(stp, struct idnstr, 1);
+
+ WR(rq)->q_ptr = rq->q_ptr = NULL;
+ mutex_exit(&idn.sipwenlock);
+ rw_exit(&idn.struprwlock);
+
+ idn_param_cleanup();
+ (void) qassociate(rq, -1);
+
+ return (0);
+}
+
+/*
+ * ----------------------------------------------
+ */
+static int
+idnwput(register queue_t *wq, register mblk_t *mp)
+{
+ register struct idnstr *stp;
+ struct idn *sip;
+ procname_t proc = "idnwput";
+
+ stp = (struct idnstr *)wq->q_ptr;
+ sip = stp->ss_sip;
+
+ switch (DB_TYPE(mp)) {
+ case M_IOCTL:
+ idnioctl(wq, mp);
+ break;
+
+ case M_DATA:
+ if (((stp->ss_flags & (IDNSFAST|IDNSRAW)) == 0) ||
+ (stp->ss_state != DL_IDLE) ||
+ (sip == NULL)) {
+ PR_DLPI("%s: fl=0x%lx, st=0x%lx, ret(EPROTO)\n",
+ proc, stp->ss_flags, stp->ss_state);
+ merror(wq, mp, EPROTO);
+
+ } else if (wq->q_first) {
+ if (putq(wq, mp) == 0)
+ freemsg(mp);
+ /*
+ * We're only holding the reader lock,
+ * but that's okay since this field
+ * is just a soft-flag.
+ */
+ sip->si_wantw = 1;
+ qenable(wq);
+
+ } else if (sip->si_flags & IDNPROMISC) {
+ if (putq(wq, mp) == 0) {
+ PR_DLPI("%s: putq failed\n", proc);
+ freemsg(mp);
+ } else {
+ PR_DLPI("%s: putq succeeded\n", proc);
+ }
+ qenable(wq);
+
+ } else {
+ PR_DLPI("%s: idndl_start(sip=0x%x)\n",
+ proc, (uint_t)sip);
+ rw_enter(&stp->ss_rwlock, RW_READER);
+ (void) idndl_start(wq, mp, sip);
+ rw_exit(&stp->ss_rwlock);
+ }
+ break;
+
+ case M_PROTO:
+ case M_PCPROTO:
+ /*
+ * Break the association between the current thread
+ * and the thread that calls idndl_proto() to resolve
+ * the problem of idn_chan_server() threads which
+ * loop back around to call idndl_proto and try to
+ * recursively acquire internal locks.
+ */
+ if (putq(wq, mp) == 0)
+ freemsg(mp);
+ qenable(wq);
+ break;
+
+ case M_FLUSH:
+ PR_STR("%s: M_FLUSH request (flush = %d)\n",
+ proc, (int)*mp->b_rptr);
+ if (*mp->b_rptr & FLUSHW) {
+ flushq(wq, FLUSHALL);
+ *mp->b_rptr &= ~FLUSHW;
+ }
+ if (*mp->b_rptr & FLUSHR)
+ qreply(wq, mp);
+ else
+ freemsg(mp);
+ break;
+
+ default:
+ PR_STR("%s: unexpected DB_TYPE 0x%x\n",
+ proc, DB_TYPE(mp));
+ freemsg(mp);
+ break;
+ }
+
+ return (0);
+}
+
+/*
+ * ----------------------------------------------
+ */
+static int
+idnwsrv(queue_t *wq)
+{
+ mblk_t *mp;
+ int err = 0;
+ struct idnstr *stp;
+ struct idn *sip;
+ procname_t proc = "idnwsrv";
+
+ stp = (struct idnstr *)wq->q_ptr;
+ sip = stp->ss_sip;
+
+ while (mp = getq(wq)) {
+ switch (DB_TYPE(mp)) {
+ case M_DATA:
+ if (sip) {
+ PR_DLPI("%s: idndl_start(sip=0x%x)\n",
+ proc, (uint_t)sip);
+ rw_enter(&stp->ss_rwlock, RW_READER);
+ err = idndl_start(wq, mp, sip);
+ rw_exit(&stp->ss_rwlock);
+ if (err)
+ goto done;
+ } else {
+ PR_DLPI("%s: NO sip to start msg\n", proc);
+ freemsg(mp);
+ }
+ break;
+
+ case M_PROTO:
+ case M_PCPROTO:
+ idndl_proto(wq, mp);
+ break;
+
+ default:
+ ASSERT(0);
+ PR_STR("%s: unexpected db_type (%d)\n",
+ proc, DB_TYPE(mp));
+ freemsg(mp);
+ break;
+ }
+ }
+done:
+ return (0);
+}
+
+/*
+ * ----------------------------------------------
+ */
+static int
+idnrput(register queue_t *rq, register mblk_t *mp)
+{
+ register int err = 0;
+ procname_t proc = "idnrput";
+
+ switch (DB_TYPE(mp)) {
+ case M_DATA:
+ /*
+ * Should not reach here with data packets
+ * if running DLPI.
+ */
+ cmn_err(CE_WARN,
+ "IDN: 123: unexpected M_DATA packets for "
+ "q_stream 0x%x", (uint_t)rq->q_stream);
+ freemsg(mp);
+ err = ENXIO;
+ break;
+
+ case M_FLUSH:
+ PR_STR("%s: M_FLUSH request (flush = %d)\n",
+ proc, (int)*mp->b_rptr);
+ if (*mp->b_rptr & FLUSHR)
+ flushq(rq, FLUSHALL);
+ (void) putnext(rq, mp);
+ break;
+
+ case M_ERROR:
+ PR_STR("%s: M_ERROR (error = %d) coming through\n",
+ proc, (int)*mp->b_rptr);
+ (void) putnext(rq, mp);
+ break;
+ default:
+ PR_STR("%s: unexpected DB_TYPE 0x%x\n",
+ proc, DB_TYPE(mp));
+ freemsg(mp);
+ err = ENXIO;
+ break;
+ }
+
+ return (err);
+}
+
+/*
+ * ----------------------------------------------
+ * Not allowed to enqueue messages! Only M_DATA messages
+ * can be enqueued on the write stream.
+ * ----------------------------------------------
+ */
+static void
+idnioctl(register queue_t *wq, register mblk_t *mp)
+{
+ register struct iocblk *iocp;
+ register int cmd;
+ idnop_t *idnop = NULL;
+ int error = 0;
+ int argsize;
+ procname_t proc = "idnioctl";
+
+ iocp = (struct iocblk *)mp->b_rptr;
+ cmd = iocp->ioc_cmd;
+
+ /*
+ * Intercept DLPI ioctl's.
+ */
+ if (VALID_DLPIOP(cmd)) {
+ PR_STR("%s: DLPI ioctl(%d)\n", proc, cmd);
+ error = idnioc_dlpi(wq, mp, &argsize);
+ goto done;
+ }
+
+ /*
+ * Validate expected arguments.
+ */
+ if (!VALID_IDNIOCTL(cmd)) {
+ PR_STR("%s: invalid cmd (0x%x)\n", proc, cmd);
+ error = EINVAL;
+ goto done;
+
+ } else if (!VALID_NDOP(cmd)) {
+ error = miocpullup(mp, sizeof (idnop_t));
+ if (error != 0) {
+ PR_STR("%s: idnioc(cmd = 0x%x) miocpullup "
+ "failed (%d)\n", proc, cmd, error);
+ goto done;
+ }
+ }
+
+ argsize = mp->b_cont->b_wptr - mp->b_cont->b_rptr;
+ idnop = (idnop_t *)mp->b_cont->b_rptr;
+
+ switch (cmd) {
+ case IDNIOC_LINK:
+ error = idnioc_link(idnop);
+ break;
+
+ case IDNIOC_UNLINK:
+ error = idnioc_unlink(idnop);
+ break;
+
+ case IDNIOC_MEM_RW:
+ error = idn_rw_mem(idnop);
+ break;
+
+ case IDNIOC_PING:
+ error = idn_send_ping(idnop);
+ break;
+
+ case ND_SET:
+ IDN_GLOCK_EXCL();
+ if (!nd_getset(wq, idn_ndlist, mp)) {
+ IDN_GUNLOCK();
+ error = ENOENT;
+ break;
+ }
+ IDN_GUNLOCK();
+ qreply(wq, mp);
+ return;
+
+ case ND_GET:
+ IDN_GLOCK_SHARED();
+ if (!nd_getset(wq, idn_ndlist, mp)) {
+ IDN_GUNLOCK();
+ error = ENOENT;
+ break;
+ }
+ IDN_GUNLOCK();
+ qreply(wq, mp);
+ return;
+
+ default:
+ PR_STR("%s: invalid cmd 0x%x\n", proc, cmd);
+ error = EINVAL;
+ break;
+ }
+
+done:
+ if (error == 0)
+ miocack(wq, mp, argsize, 0);
+ else
+ miocnak(wq, mp, 0, error);
+}
+
+/*
+ * This thread actually services the SSI_LINK/UNLINK calls
+ * asynchronously that come via BBSRAM. This is necessary
+ * since we can't process them from within the context of
+ * the interrupt handler in which idn_sigbhandler() is
+ * called.
+ */
+static void
+idn_sigbhandler_thread(struct sigbintr **sbpp)
+{
+ int d, pri, rv;
+ struct sigbintr *sbp;
+ sigbmbox_t *mbp;
+ idn_fin_t fintype;
+ idnsb_data_t *sdp;
+ idnsb_info_t *sfp;
+ idnsb_error_t *sep;
+ idn_domain_t *dp;
+ procname_t proc = "idn_sigbhandler_thread";
+
+
+ sbp = *sbpp;
+
+ PR_PROTO("%s: KICKED OFF (sigbintr pointer = 0x%x)\n",
+ proc, (uint_t)sbp);
+
+ ASSERT(sbp == &idn.sigbintr);
+
+ mutex_enter(&idn.sigbintr.sb_mutex);
+
+ while (sbp->sb_busy != IDNSIGB_DIE) {
+ cpu_sgnblk_t *sigbp;
+
+ while ((sbp->sb_busy != IDNSIGB_ACTIVE) &&
+ (sbp->sb_busy != IDNSIGB_DIE)) {
+ cv_wait(&sbp->sb_cv, &idn.sigbintr.sb_mutex);
+ PR_PROTO("%s: AWAKENED (busy = %d)\n",
+ proc, (int)sbp->sb_busy);
+ }
+ if (sbp->sb_busy == IDNSIGB_DIE) {
+ PR_PROTO("%s: DIE REQUESTED\n", proc);
+ break;
+ }
+
+ if ((sigbp = cpu_sgnblkp[sbp->sb_cpuid]) == NULL) {
+ cmn_err(CE_WARN,
+ "IDN: 124: sigblk for CPU ID %d "
+ "is NULL", sbp->sb_cpuid);
+ sbp->sb_busy = IDNSIGB_INACTIVE;
+ continue;
+ }
+
+ mbp = &sigbp->sigb_host_mbox;
+
+ if (mbp->flag != SIGB_MBOX_BUSY) {
+ PR_PROTO("%s: sigblk mbox flag (%d) != BUSY (%d)\n",
+ proc, mbp->flag, SIGB_MBOX_BUSY);
+ sbp->sb_busy = IDNSIGB_INACTIVE;
+ continue;
+ }
+ /*
+ * The sb_busy bit is set and the mailbox flag
+ * indicates BUSY also, so we effectively have things locked.
+ * So, we can drop the critical sb_mutex which we want to
+ * do since it pushes us to PIL 14 while we hold it and we
+ * don't want to run at PIL 14 across IDN code.
+ */
+ mutex_exit(&idn.sigbintr.sb_mutex);
+
+ sdp = (idnsb_data_t *)mbp->data;
+ sep = (idnsb_error_t *)&sdp->ssb_error;
+ INIT_IDNKERR(sep);
+
+ if (mbp->len != sizeof (idnsb_data_t)) {
+ PR_PROTO("%s: sigblk mbox length (%d) != "
+ "expected (%d)\n", proc, mbp->len,
+ sizeof (idnsb_data_t));
+ SET_IDNKERR_ERRNO(sep, EINVAL);
+ SET_IDNKERR_IDNERR(sep, IDNKERR_DATA_LEN);
+ SET_IDNKERR_PARAM0(sep, sizeof (idnsb_data_t));
+
+ goto sberr;
+
+ }
+ if (idn.enabled == 0) {
+#ifdef DEBUG
+ cmn_err(CE_NOTE,
+ "IDN: 102: driver disabled "
+ "- check OBP environment "
+ "(idn-smr-size)");
+#else /* DEBUG */
+ cmn_err(CE_NOTE,
+ "!IDN: 102: driver disabled "
+ "- check OBP environment "
+ "(idn-smr-size)");
+#endif /* DEBUG */
+ SET_IDNKERR_ERRNO(sep, EACCES);
+ SET_IDNKERR_IDNERR(sep, IDNKERR_DRV_DISABLED);
+
+ goto sberr;
+
+ }
+
+ switch (mbp->cmd) {
+
+ case SSI_LINK:
+ {
+ idnsb_link_t slp;
+
+ bcopy(&sdp->ssb_link, &slp, sizeof (slp));
+
+ if (slp.master_pri < 0) {
+ pri = IDNVOTE_MINPRI;
+ } else if (slp.master_pri > 0) {
+ /*
+ * If I'm already in a IDN network,
+ * then my vote priority is set to
+ * the max, otherwise it's one-less.
+ */
+ pri = IDNVOTE_MAXPRI;
+ IDN_GLOCK_SHARED();
+ if (idn.ndomains <= 1)
+ pri--;
+ IDN_GUNLOCK();
+ } else {
+ pri = IDNVOTE_DEFPRI;
+ }
+
+ PR_PROTO("%s: SSI_LINK(cpuid = %d, domid = %d, "
+ "pri = %d (req = %d), t/o = %d)\n",
+ proc, slp.cpuid, slp.domid, pri,
+ slp.master_pri, slp.timeout);
+
+ rv = idn_link(slp.domid, slp.cpuid, pri,
+ slp.timeout, sep);
+ SET_IDNKERR_ERRNO(sep, rv);
+ (void) idn_info(&sdp->ssb_info);
+ break;
+ }
+
+ case SSI_UNLINK:
+ {
+ idnsb_unlink_t sup;
+ idn_domain_t *xdp;
+ domainset_t domset;
+
+ bcopy(&sdp->ssb_unlink, &sup, sizeof (sup));
+
+ PR_PROTO("%s: SSI_UNLINK(c = %d, d = %d, bs = 0x%x, "
+ "f = %d, is = 0x%x, t/o = %d)\n",
+ proc, sup.cpuid, sup.domid, sup.boardset,
+ sup.force, sup.idnset, sup.timeout);
+
+ domset = idn.domset.ds_trans_on |
+ idn.domset.ds_connected |
+ idn.domset.ds_trans_off |
+ idn.domset.ds_awol |
+ idn.domset.ds_relink;
+
+ if (VALID_DOMAINID(sup.domid)) {
+ dp = &idn_domain[sup.domid];
+ } else if (VALID_CPUID(sup.cpuid)) {
+ for (d = 0; d < MAX_DOMAINS; d++) {
+ xdp = &idn_domain[d];
+
+ if ((xdp->dcpu == IDN_NIL_DCPU) &&
+ !DOMAIN_IN_SET(domset, d))
+ continue;
+
+ if (CPU_IN_SET(xdp->dcpuset,
+ sup.cpuid))
+ break;
+ }
+ dp = (d == MAX_DOMAINS) ? NULL : xdp;
+ }
+ if ((dp == NULL) && sup.boardset) {
+ for (d = 0; d < MAX_DOMAINS; d++) {
+ xdp = &idn_domain[d];
+
+ if ((xdp->dcpu == IDN_NIL_DCPU) &&
+ !DOMAIN_IN_SET(domset, d))
+ continue;
+
+ if (xdp->dhw.dh_boardset &
+ sup.boardset)
+ break;
+ }
+ dp = (d == MAX_DOMAINS) ? NULL : xdp;
+ }
+ if (dp == NULL) {
+ SET_IDNKERR_ERRNO(sep, EINVAL);
+ SET_IDNKERR_IDNERR(sep, IDNKERR_INVALID_DOMAIN);
+ SET_IDNKERR_PARAM0(sep, sup.domid);
+ SET_IDNKERR_PARAM1(sep, sup.cpuid);
+ (void) idn_info(&sdp->ssb_info);
+ goto sberr;
+ } else {
+ sup.domid = dp->domid;
+ }
+
+ switch (sup.force) {
+ case SSIFORCE_OFF:
+ fintype = IDNFIN_NORMAL;
+ break;
+
+ case SSIFORCE_SOFT:
+ fintype = IDNFIN_FORCE_SOFT;
+ break;
+
+ case SSIFORCE_HARD:
+ fintype = IDNFIN_FORCE_HARD;
+ break;
+ default:
+ SET_IDNKERR_ERRNO(sep, EINVAL);
+ SET_IDNKERR_IDNERR(sep, IDNKERR_INVALID_FORCE);
+ SET_IDNKERR_PARAM0(sep, sup.force);
+ (void) idn_info(&sdp->ssb_info);
+ goto sberr;
+ }
+
+ rv = idn_unlink(sup.domid, sup.idnset, fintype,
+ IDNFIN_OPT_UNLINK, sup.timeout, sep);
+ SET_IDNKERR_ERRNO(sep, rv);
+ (void) idn_info(&sdp->ssb_info);
+ break;
+ }
+
+ case SSI_INFO:
+ sfp = &sdp->ssb_info;
+
+ PR_PROTO("%s: SSI_INFO\n", proc);
+
+ rv = idn_info(sfp);
+ SET_IDNKERR_ERRNO(sep, rv);
+ if (rv != 0) {
+ SET_IDNKERR_IDNERR(sep, IDNKERR_INFO_FAILED);
+ }
+ break;
+
+ default:
+ ASSERT(0);
+ SET_IDNKERR_ERRNO(sep, EINVAL);
+ SET_IDNKERR_IDNERR(sep, IDNKERR_INVALID_CMD);
+ SET_IDNKERR_PARAM0(sep, mbp->cmd);
+ break;
+ }
+
+sberr:
+
+ if (GET_IDNKERR_ERRNO(sep) != 0) {
+ cmn_err(CE_WARN,
+#ifdef DEBUG
+ "IDN: 125: op (%s) failed, returning "
+ "(%d/0x%x [%d, %d, %d])",
+#else /* DEBUG */
+ "!IDN: 125: op (%s) failed, returning "
+ "(%d/0x%x [%d, %d, %d])",
+#endif /* DEBUG */
+ (mbp->cmd == SSI_LINK) ? "LINK" :
+ (mbp->cmd == SSI_UNLINK) ? "UNLINK" :
+ (mbp->cmd == SSI_INFO) ?
+ "INFO" : "UNKNOWN",
+ GET_IDNKERR_ERRNO(sep),
+ GET_IDNKERR_IDNERR(sep),
+ GET_IDNKERR_PARAM0(sep),
+ GET_IDNKERR_PARAM1(sep),
+ GET_IDNKERR_PARAM2(sep));
+ }
+
+ PR_PROTO("%s: returning errno = %d, idnerr = %d, "
+ "params = [%d, %d, %d]\n",
+ proc, GET_IDNKERR_ERRNO(sep), GET_IDNKERR_IDNERR(sep),
+ GET_IDNKERR_PARAM0(sep), GET_IDNKERR_PARAM1(sep),
+ GET_IDNKERR_PARAM2(sep));
+
+ mutex_enter(&idn.sigbintr.sb_mutex);
+ ASSERT((sbp->sb_busy == IDNSIGB_ACTIVE) ||
+ (sbp->sb_busy == IDNSIGB_DIE));
+ mbp->cmd |= SSI_ACK;
+ if (sbp->sb_busy == IDNSIGB_ACTIVE)
+ sbp->sb_busy = IDNSIGB_INACTIVE;
+ /*
+ * Set flag which kicks off response to SSP.
+ */
+ membar_stst_ldst();
+ mbp->flag = HOST_TO_CBS;
+ }
+
+ /*
+ * Wake up the dude that killed us!
+ */
+ idn.sigb_threadp = NULL;
+ cv_signal(&sbp->sb_cv);
+ mutex_exit(&idn.sigbintr.sb_mutex);
+ thread_exit();
+}
+
+/*
+ * Create the thread that will service sigb interrupts.
+ */
+static void
+idn_sigbhandler_create()
+{
+ struct sigbintr *sbp;
+
+ if (idn.sigb_threadp) {
+ cmn_err(CE_WARN,
+ "IDN: 126: sigbhandler thread already "
+ "exists (0x%x)", (uint_t)idn.sigb_threadp);
+ return;
+ }
+ cv_init(&idn.sigbintr.sb_cv, NULL, CV_DEFAULT, NULL);
+ sbp = &idn.sigbintr;
+ sbp->sb_busy = IDNSIGB_INACTIVE;
+ idn.sigb_threadp = thread_create(NULL, 0,
+ idn_sigbhandler_thread, &sbp, sizeof (sbp), &p0,
+ TS_RUN, minclsyspri);
+ sbp->sb_inum = add_softintr((uint_t)idn_sigbpil,
+ idn_sigbhandler_wakeup, 0);
+}
+
+static void
+idn_sigbhandler_kill()
+{
+ if (idn.sigb_threadp) {
+ struct sigbintr *sbp;
+
+ sbp = &idn.sigbintr;
+ if (sbp->sb_inum != 0)
+ rem_softintr(sbp->sb_inum);
+ sbp->sb_inum = 0;
+ sbp->sb_busy = IDNSIGB_DIE;
+ cv_signal(&sbp->sb_cv);
+ while (idn.sigb_threadp != NULL)
+ cv_wait(&sbp->sb_cv, &idn.sigbintr.sb_mutex);
+ sbp->sb_busy = IDNSIGB_INACTIVE;
+ cv_destroy(&sbp->sb_cv);
+ }
+}
+
+/*ARGSUSED0*/
+static uint_t
+idn_sigbhandler_wakeup(caddr_t arg1, caddr_t arg2)
+{
+ mutex_enter(&idn.sigbintr.sb_mutex);
+ if (idn.sigbintr.sb_busy == IDNSIGB_STARTED) {
+ idn.sigbintr.sb_busy = IDNSIGB_ACTIVE;
+ cv_signal(&idn.sigbintr.sb_cv);
+ }
+ mutex_exit(&idn.sigbintr.sb_mutex);
+
+ return (DDI_INTR_CLAIMED);
+}
+
+static void
+idn_sigbhandler(processorid_t cpuid, cpu_sgnblk_t *sgnblkp)
+{
+ struct sigbintr *sbp = &idn.sigbintr;
+ sigbmbox_t *mbp;
+ idnsb_data_t *sdp;
+ idnsb_error_t *sep;
+ uint32_t cmd;
+ int sigb_lock = 0;
+
+ ASSERT(sgnblkp);
+
+ mbp = &sgnblkp->sigb_host_mbox;
+ sdp = (idnsb_data_t *)mbp->data;
+ sep = &sdp->ssb_error;
+ cmd = mbp->cmd;
+
+ if ((mbp->flag != CBS_TO_HOST) || !VALID_IDNSIGBCMD(cmd)) {
+ /*
+ * Not a valid IDN command. Just bail out.
+ */
+ return;
+ }
+
+ mbp->flag = SIGB_MBOX_BUSY;
+ SET_IDNKERR_ERRNO(sep, 0);
+
+ if (cmd & SSI_ACK) {
+ /*
+ * Hmmm...weird, the ACK bit is set.
+ */
+ SET_IDNKERR_ERRNO(sep, EPROTO);
+ SET_IDNKERR_IDNERR(sep, IDNKERR_INVALID_CMD);
+ SET_IDNKERR_PARAM0(sep, cmd);
+ goto sigb_done;
+ }
+
+ if (!mutex_tryenter(&idn.sigbintr.sb_mutex)) {
+ /*
+ * Couldn't get the lock. Driver is either
+ * not quite all the way up or is shutting down
+ * for some reason. Caller should spin again.
+ */
+ cmd |= SSI_ACK;
+ SET_IDNKERR_ERRNO(sep, EBUSY);
+ SET_IDNKERR_IDNERR(sep, IDNKERR_SIGBINTR_LOCKED);
+ goto sigb_done;
+ }
+ sigb_lock = 1;
+
+ if ((idn.sigb_threadp == NULL) ||
+ (sbp->sb_busy == IDNSIGB_NOTREADY)) {
+ cmd |= SSI_ACK;
+ SET_IDNKERR_ERRNO(sep, EAGAIN);
+ SET_IDNKERR_IDNERR(sep, IDNKERR_SIGBINTR_NOTRDY);
+ goto sigb_done;
+ }
+
+ if (sbp->sb_busy != IDNSIGB_INACTIVE) {
+ cmd |= SSI_ACK;
+ SET_IDNKERR_ERRNO(sep, EBUSY);
+ SET_IDNKERR_IDNERR(sep, IDNKERR_SIGBINTR_BUSY);
+ goto sigb_done;
+ }
+
+ sbp->sb_cpuid = (uchar_t)cpuid & 0xff;
+ membar_stst_ldst();
+ sbp->sb_busy = IDNSIGB_STARTED;
+ /*
+ * The sb_busy bit is set and the mailbox flag
+ * indicates BUSY also, so we effectively have things locked.
+ * So, we can drop the critical sb_mutex which we want to
+ * do since it pushes us to PIL 14 while we hold it and we
+ * don't want to run at PIL 14 across IDN code.
+ *
+ * Send interrupt to cause idn_sigbhandler_thread to wakeup.
+ * We cannot do wakeup (cv_signal) directly from here since
+ * we're executing from a high-level (14) interrupt.
+ */
+ setsoftint(sbp->sb_inum);
+
+sigb_done:
+
+ if (GET_IDNKERR_ERRNO(sep) != 0) {
+ mbp->len = sizeof (idnsb_data_t);
+ mbp->cmd = cmd;
+ membar_stst_ldst();
+ mbp->flag = HOST_TO_CBS;
+ }
+
+ if (sigb_lock)
+ mutex_exit(&idn.sigbintr.sb_mutex);
+}
+
+static int
+idn_info(idnsb_info_t *sfp)
+{
+ int count, d;
+ idn_domain_t *dp;
+ idnsb_info_t sinfo;
+ int local_id, master_id;
+ procname_t proc = "idn_info";
+
+ bzero(&sinfo, sizeof (sinfo));
+ sinfo.master_index = (uchar_t)-1;
+ sinfo.master_cpuid = (uchar_t)-1;
+ sinfo.local_index = (uchar_t)-1;
+ sinfo.local_cpuid = (uchar_t)-1;
+
+ IDN_GLOCK_SHARED();
+
+ sinfo.idn_state = (uchar_t)idn.state;
+
+ switch (idn.state) {
+ case IDNGS_OFFLINE:
+ sinfo.idn_active = SSISTATE_INACTIVE;
+ PR_PROTO("%s: idn_state (%s) = INACTIVE\n",
+ proc, idngs_str[idn.state]);
+ break;
+
+ case IDNGS_IGNORE:
+ PR_PROTO("%s: IGNORING IDN_INFO call...\n", proc);
+ IDN_GUNLOCK();
+ return (EIO);
+
+ default:
+ sinfo.idn_active = SSISTATE_ACTIVE;
+ PR_PROTO("%s: idn_state (%s) = ACTIVE\n",
+ proc, idngs_str[idn.state]);
+ break;
+ }
+ master_id = IDN_GET_MASTERID();
+ local_id = idn.localid;
+
+ /*
+ * Need to drop idn.grwlock before acquiring domain locks.
+ */
+ IDN_GUNLOCK();
+
+ IDN_SYNC_LOCK();
+
+ sinfo.awol_domset = (ushort_t)idn.domset.ds_awol;
+ sinfo.conn_domset = (ushort_t)(idn.domset.ds_connected &
+ ~idn.domset.ds_trans_on);
+ DOMAINSET_ADD(sinfo.conn_domset, idn.localid);
+
+ count = 0;
+ for (d = 0; d < MAX_DOMAINS; d++) {
+ dp = &idn_domain[d];
+
+ if (dp->dcpu == IDN_NIL_DCPU)
+ continue;
+
+ IDN_DLOCK_SHARED(d);
+ if ((dp->dcpu == IDN_NIL_DCPU) ||
+ (dp->dstate == IDNDS_CLOSED)) {
+ IDN_DUNLOCK(d);
+ continue;
+ }
+
+ count++;
+ if (d == local_id) {
+ sinfo.local_index = (uchar_t)d;
+ sinfo.local_cpuid = (uchar_t)dp->dcpu;
+ PR_PROTO("%s: domid %d is LOCAL (cpuid = %d)\n",
+ proc, d, dp->dcpu);
+ }
+ if (d == master_id) {
+ sinfo.master_index = (uchar_t)d;
+ sinfo.master_cpuid = (uchar_t)dp->dcpu;
+ PR_PROTO("%s: domid %d is MASTER (cpuid = %d)\n",
+ proc, d, dp->dcpu);
+ }
+
+ sinfo.domain_boardset[d] = (ushort_t)dp->dhw.dh_boardset;
+
+ IDN_DUNLOCK(d);
+ }
+
+ IDN_SYNC_UNLOCK();
+
+ bcopy(&sinfo, sfp, sizeof (*sfp));
+
+ PR_PROTO("%s: Found %d domains within IDNnet\n", proc, count);
+
+ return (0);
+}
+
+/*
+ * ----------------------------------------------
+ * ndd param support routines.
+ * - Borrowed from tcp.
+ * ----------------------------------------------
+ */
+static void
+idn_param_cleanup()
+{
+ IDN_GLOCK_EXCL();
+ if (!idn.strup && idn_ndlist)
+ nd_free(&idn_ndlist);
+ IDN_GUNLOCK();
+}
+
+/*ARGSUSED*/
+static int
+idn_param_get(queue_t *q, mblk_t *mp, caddr_t cp, cred_t *cr)
+{
+ idnparam_t *idnpa = (idnparam_t *)cp;
+
+ /*
+ * lock grabbed before calling nd_getset.
+ */
+ ASSERT(IDN_GLOCK_IS_HELD());
+
+ (void) mi_mpprintf(mp, "%ld", idnpa->sp_val);
+
+ return (0);
+}
+
+/*ARGSUSED*/
+static int
+idn_param_set(queue_t *q, mblk_t *mp, char *value, caddr_t cp, cred_t *cr)
+{
+ char *end;
+ ulong_t new_value;
+ idnparam_t *idnpa = (idnparam_t *)cp;
+
+ /*
+ * lock grabbed before calling nd_getset.
+ */
+ ASSERT(IDN_GLOCK_IS_EXCL());
+
+ new_value = (ulong_t)mi_strtol(value, &end, 10);
+
+ if ((end == value) ||
+ (new_value < idnpa->sp_min) ||
+ (new_value > idnpa->sp_max))
+ return (EINVAL);
+
+ if (idn.enabled == 0) {
+#ifdef DEBUG
+ cmn_err(CE_NOTE,
+ "IDN: 102: driver disabled "
+ "- check OBP environment "
+ "(idn-smr-size)");
+#else /* DEBUG */
+ cmn_err(CE_NOTE,
+ "!IDN: 102: driver disabled "
+ "- check OBP environment "
+ "(idn-smr-size)");
+#endif /* DEBUG */
+ return (EACCES);
+ }
+
+ idnpa->sp_val = new_value;
+
+ return (0);
+}
+
+static int
+idn_param_register(register idnparam_t *idnpa, int count)
+{
+ ASSERT(IDN_GLOCK_IS_EXCL());
+
+ for (; count > 0; count--, idnpa++) {
+ if (idnpa->sp_name && idnpa->sp_name[0]) {
+ register int i;
+ ndsetf_t set_func;
+ char *p;
+ /*
+ * Don't advertise in non-DEBUG parameters.
+ */
+ for (i = 0; idn_param_debug_only[i]; i++) {
+ p = idn_param_debug_only[i];
+ if (strcmp(idnpa->sp_name, p) == 0)
+ break;
+ }
+ if (idn_param_debug_only[i])
+ continue;
+
+ /*
+ * Do not register a "set" function for
+ * read-only parameters.
+ */
+ for (i = 0; idn_param_read_only[i]; i++) {
+ p = idn_param_read_only[i];
+ if (strcmp(idnpa->sp_name, p) == 0)
+ break;
+ }
+ if (idn_param_read_only[i])
+ set_func = NULL;
+ else
+ set_func = idn_param_set;
+
+ if (!nd_load(&idn_ndlist, idnpa->sp_name,
+ idn_param_get, set_func,
+ (caddr_t)idnpa)) {
+ nd_free(&idn_ndlist);
+ return (-1);
+ }
+ }
+ }
+ if (!nd_load(&idn_ndlist, "idn_slabpool", idn_slabpool_report,
+ NULL, NULL)) {
+ nd_free(&idn_ndlist);
+ return (-1);
+ }
+ if (!nd_load(&idn_ndlist, "idn_buffers", idn_buffer_report,
+ NULL, NULL)) {
+ nd_free(&idn_ndlist);
+ return (-1);
+ }
+ if (!nd_load(&idn_ndlist, "idn_mboxtbl", idn_mboxtbl_report,
+ NULL, MBXTBL_PART_REPORT)) {
+ nd_free(&idn_ndlist);
+ return (-1);
+ }
+ if (!nd_load(&idn_ndlist, "idn_mboxtbl_all", idn_mboxtbl_report,
+ NULL, MBXTBL_FULL_REPORT)) {
+ nd_free(&idn_ndlist);
+ return (-1);
+ }
+ if (!nd_load(&idn_ndlist, "idn_mainmbox", idn_mainmbox_report,
+ NULL, NULL)) {
+ nd_free(&idn_ndlist);
+ return (-1);
+ }
+ if (!nd_load(&idn_ndlist, "idn_global", idn_global_report,
+ NULL, NULL)) {
+ nd_free(&idn_ndlist);
+ return (-1);
+ }
+ if (!nd_load(&idn_ndlist, "idn_domain", idn_domain_report,
+ NULL, (caddr_t)0)) {
+ nd_free(&idn_ndlist);
+ return (-1);
+ }
+ if (!nd_load(&idn_ndlist, "idn_domain_all", idn_domain_report,
+ NULL, (caddr_t)1)) {
+ nd_free(&idn_ndlist);
+ return (-1);
+ }
+ if (!nd_load(&idn_ndlist, "idn_bind_net", idn_get_net_binding,
+ idn_set_net_binding, NULL)) {
+ nd_free(&idn_ndlist);
+ return (-1);
+ }
+
+ return (0);
+}
+
+/*ARGSUSED*/
+static int
+idn_set_net_binding(queue_t *q, mblk_t *mp, char *value, caddr_t cp, cred_t *cr)
+{
+ char *end, *cpup;
+ long net;
+ processorid_t cpuid;
+
+ /*
+ * lock grabbed before calling nd_getset.
+ */
+ ASSERT(IDN_GLOCK_IS_EXCL());
+
+ if ((cpup = strchr(value, '=')) == NULL)
+ return (EINVAL);
+
+ *cpup++ = '\0';
+
+ net = mi_strtol(value, &end, 10);
+ if ((end == value) || (net < 0) || (net >= IDN_MAX_NETS) ||
+ !CHAN_IN_SET(idn.chanset, net))
+ return (EINVAL);
+
+ cpuid = (processorid_t)mi_strtol(cpup, &end, 10);
+ if ((end == cpup) || ((cpuid != -1) &&
+ (!VALID_CPUID(cpuid) ||
+ !CPU_IN_SET(cpu_ready_set, cpuid))))
+ return (EINVAL);
+
+ idn_chanserver_bind(net, cpuid);
+
+ return (0);
+}
+
+/*ARGSUSED*/
+static int
+idn_get_net_binding(queue_t *q, mblk_t *mp, caddr_t cp, cred_t *cr)
+{
+ int c;
+
+ /*
+ * lock grabbed before calling nd_getset.
+ */
+ ASSERT(IDN_GLOCK_IS_HELD());
+
+ (void) mi_mpprintf(mp,
+ "IDN network interfaces/channels active = %d",
+ idn.nchannels);
+
+ if (idn.nchannels == 0)
+ return (0);
+
+ mi_mpprintf(mp, "Net Cpu");
+
+ for (c = 0; c < IDN_MAX_NETS; c++) {
+ int bc;
+ idn_chansvr_t *csp;
+
+ if (!CHAN_IN_SET(idn.chanset, c))
+ continue;
+
+ csp = &idn.chan_servers[c];
+
+ if ((bc = csp->ch_bound_cpuid) == -1)
+ bc = csp->ch_bound_cpuid_pending;
+
+ if (c < 10)
+ mi_mpprintf(mp, " %d %d", c, bc);
+ else
+ mi_mpprintf(mp, " %d %d", c, bc);
+ }
+
+ return (0);
+}
+
+static int
+idnioc_link(idnop_t *idnop)
+{
+ int rv;
+ int pri;
+ idnsb_error_t err;
+ procname_t proc = "idnioc_link";
+
+ if (idnop->link.master < 0)
+ pri = IDNVOTE_MINPRI;
+ else if (idnop->link.master > 0)
+ pri = IDNVOTE_MAXPRI;
+ else
+ pri = IDNVOTE_DEFPRI;
+
+ PR_DRV("%s: domid = %d, cpuid = %d, pri = %d\n",
+ proc, idnop->link.domid, idnop->link.cpuid, pri);
+
+ rv = idn_link(idnop->link.domid, idnop->link.cpuid,
+ pri, idnop->link.wait, &err);
+
+ return (rv);
+}
+
+static int
+idnioc_unlink(idnop_t *idnop)
+{
+ int d, cpuid, domid, rv;
+ boardset_t idnset;
+ idn_fin_t fintype;
+ idn_domain_t *dp, *xdp;
+ idnsb_error_t err;
+ procname_t proc = "idnioc_unlink";
+
+ PR_DRV("%s: domid = %d, cpuid = %d, force = %d\n",
+ proc, idnop->unlink.domid, idnop->unlink.cpuid,
+ idnop->unlink.force);
+
+ idnset = BOARDSET_ALL;
+ domid = idnop->unlink.domid;
+ cpuid = idnop->unlink.cpuid;
+ dp = NULL;
+
+ if (domid == IDN_NIL_DOMID)
+ domid = idn.localid;
+
+ if (VALID_DOMAINID(domid)) {
+ dp = &idn_domain[domid];
+ if (VALID_CPUID(cpuid) && (dp->dcpu != IDN_NIL_DCPU) &&
+ !CPU_IN_SET(dp->dcpuset, cpuid)) {
+ dp = NULL;
+ PR_PROTO("%s: ERROR: invalid cpuid "
+ "(%d) for domain (%d) [cset = 0x%x.x%x]\n",
+ proc, cpuid, domid,
+ UPPER32_CPUMASK(dp->dcpuset),
+ LOWER32_CPUMASK(dp->dcpuset));
+ }
+ } else if (VALID_CPUID(cpuid)) {
+ for (d = 0; d < MAX_DOMAINS; d++) {
+ xdp = &idn_domain[d];
+
+ if (xdp->dcpu == IDN_NIL_DCPU)
+ continue;
+
+ if (CPU_IN_SET(xdp->dcpuset, cpuid))
+ break;
+ }
+ dp = (d == MAX_DOMAINS) ? NULL : xdp;
+ }
+
+ if ((dp == NULL) || (dp->dcpu == IDN_NIL_DCPU))
+ return (0);
+
+ domid = dp->domid;
+
+ switch (idnop->unlink.force) {
+ case SSIFORCE_OFF:
+ fintype = IDNFIN_NORMAL;
+ break;
+
+ case SSIFORCE_SOFT:
+ fintype = IDNFIN_FORCE_SOFT;
+ break;
+
+ case SSIFORCE_HARD:
+ fintype = IDNFIN_FORCE_HARD;
+ break;
+ default:
+ PR_PROTO("%s: invalid force parameter \"%d\"",
+ proc, idnop->unlink.force);
+ return (EINVAL);
+ }
+
+ rv = idn_unlink(domid, idnset, fintype, IDNFIN_OPT_UNLINK,
+ idnop->unlink.wait, &err);
+
+ return (rv);
+}
+
+static int
+idn_send_ping(idnop_t *idnop)
+{
+ int domid = idnop->ping.domid;
+ int cpuid = idnop->ping.cpuid;
+ int ocpuid;
+ idn_domain_t *dp;
+ idn_msgtype_t mt;
+ procname_t proc = "idn_send_ping";
+
+ if ((domid == IDN_NIL_DOMID) && (cpuid == IDN_NIL_DCPU)) {
+ cmn_err(CE_WARN,
+ "IDN: %s: no valid domain ID or CPU ID given",
+ proc);
+ return (EINVAL);
+ }
+ if (domid == IDN_NIL_DOMID)
+ domid = MAX_DOMAINS - 1;
+
+ dp = &idn_domain[domid];
+ IDN_DLOCK_EXCL(domid);
+ if ((dp->dcpu == IDN_NIL_DCPU) && (cpuid == IDN_NIL_DCPU)) {
+ cmn_err(CE_WARN,
+ "IDN: %s: no valid target CPU specified",
+ proc);
+ IDN_DUNLOCK(domid);
+ return (EINVAL);
+ }
+ if (cpuid == IDN_NIL_DCPU)
+ cpuid = dp->dcpu;
+
+ ocpuid = dp->dcpu;
+ dp->dcpu = cpuid;
+
+ /*
+ * XXX - Need a special PING IDN command.
+ */
+ mt.mt_mtype = IDNP_DATA | IDNP_ACK;
+ mt.mt_atype = 0;
+
+ (void) IDNXDC(domid, &mt, 0x100, 0x200, 0x300, 0x400);
+
+ dp->dcpu = ocpuid;
+ IDN_DUNLOCK(domid);
+
+ return (0);
+}
+
+/*
+ * ----------------------------------------------
+ */
+static void
+idn_dopers_init()
+{
+ int i;
+ dop_waitlist_t *dwl;
+
+ if (idn.dopers)
+ return;
+
+ idn.dopers = GETSTRUCT(struct dopers, 1);
+
+ bzero(idn.dopers, sizeof (struct dopers));
+
+ dwl = &idn.dopers->_dop_wcache[0];
+ for (i = 0; i < (IDNOP_CACHE_SIZE-1); i++)
+ dwl[i].dw_next = &dwl[i+1];
+ dwl[i].dw_next = NULL;
+
+ idn.dopers->dop_freelist = &idn.dopers->_dop_wcache[0];
+ idn.dopers->dop_waitcount = 0;
+ idn.dopers->dop_domset = 0;
+ idn.dopers->dop_waitlist = NULL;
+
+ cv_init(&idn.dopers->dop_cv, NULL, CV_DEFAULT, NULL);
+ mutex_init(&idn.dopers->dop_mutex, NULL, MUTEX_DEFAULT, NULL);
+}
+
+static void
+idn_dopers_deinit()
+{
+ dop_waitlist_t *dwl, *next_dwl;
+
+
+ if (idn.dopers == NULL)
+ return;
+
+ for (dwl = idn.dopers->dop_waitlist; dwl; dwl = next_dwl) {
+ next_dwl = dwl->dw_next;
+ if (!IDNOP_IN_CACHE(dwl))
+ FREESTRUCT(dwl, dop_waitlist_t, 1);
+ }
+
+ cv_destroy(&idn.dopers->dop_cv);
+ mutex_destroy(&idn.dopers->dop_mutex);
+
+ FREESTRUCT(idn.dopers, struct dopers, 1);
+ idn.dopers = NULL;
+}
+
+/*
+ * Reset the dop_errset field in preparation for an
+ * IDN operation attempt. This is only called from
+ * idn_link() and idn_unlink().
+ */
+void *
+idn_init_op(idn_opflag_t opflag, domainset_t domset, idnsb_error_t *sep)
+{
+ dop_waitlist_t *dwl;
+ /*
+ * Clear any outstanding error ops in preparation
+ * for an IDN (link/unlink) operation.
+ */
+ mutex_enter(&idn.dopers->dop_mutex);
+ if ((dwl = idn.dopers->dop_freelist) == NULL) {
+ dwl = GETSTRUCT(dop_waitlist_t, 1);
+ } else {
+ idn.dopers->dop_freelist = dwl->dw_next;
+ bzero(dwl, sizeof (*dwl));
+ }
+ dwl->dw_op = opflag;
+ dwl->dw_reqset = domset;
+ dwl->dw_idnerr = sep;
+ dwl->dw_next = idn.dopers->dop_waitlist;
+
+ idn.dopers->dop_waitlist = dwl;
+ idn.dopers->dop_waitcount++;
+ idn.dopers->dop_domset |= domset;
+ mutex_exit(&idn.dopers->dop_mutex);
+
+ return (dwl);
+}
+
+/*
+ * Anybody waiting on a opflag operation for any one
+ * of the domains in domset, needs to be updated to
+ * additionally wait for new domains in domset.
+ * This is used, for example, when needing to connect
+ * to more domains than known at the time of the
+ * original request.
+ */
+void
+idn_add_op(idn_opflag_t opflag, domainset_t domset)
+{
+ dop_waitlist_t *dwl;
+
+ mutex_enter(&idn.dopers->dop_mutex);
+ if ((idn.dopers->dop_waitcount == 0) ||
+ ((idn.dopers->dop_domset & domset) == 0)) {
+ mutex_exit(&idn.dopers->dop_mutex);
+ return;
+ }
+ for (dwl = idn.dopers->dop_waitlist; dwl; dwl = dwl->dw_next)
+ if ((dwl->dw_op == opflag) && (dwl->dw_reqset & domset))
+ dwl->dw_reqset |= domset;
+ mutex_exit(&idn.dopers->dop_mutex);
+}
+
+/*
+ * Mechanism to wakeup any potential users which may be waiting
+ * for a link/unlink operation to complete. If an error occurred
+ * don't update dop_errset unless there was no previous error.
+ */
+void
+idn_update_op(idn_opflag_t opflag, domainset_t domset, idnsb_error_t *sep)
+{
+ int do_wakeup = 0;
+ dop_waitlist_t *dw;
+ procname_t proc = "idn_update_op";
+
+ mutex_enter(&idn.dopers->dop_mutex);
+ /*
+ * If there are no waiters, or nobody is waiting for
+ * the particular domainset in question, then
+ * just bail.
+ */
+ if ((idn.dopers->dop_waitcount == 0) ||
+ ((idn.dopers->dop_domset & domset) == 0)) {
+ mutex_exit(&idn.dopers->dop_mutex);
+ PR_PROTO("%s: NO waiters exist (domset=0x%x)\n",
+ proc, domset);
+ return;
+ }
+
+ for (dw = idn.dopers->dop_waitlist; dw; dw = dw->dw_next) {
+ int d;
+ domainset_t dset, rset;
+
+ if ((dset = dw->dw_reqset & domset) == 0)
+ continue;
+
+ if (opflag == IDNOP_ERROR) {
+ dw->dw_errset |= dset;
+ if (sep) {
+ for (d = 0; d < MAX_DOMAINS; d++) {
+ if (!DOMAIN_IN_SET(dset, d))
+ continue;
+
+ dw->dw_errors[d] =
+ (short)GET_IDNKERR_ERRNO(sep);
+ }
+ bcopy(sep, dw->dw_idnerr, sizeof (*sep));
+ }
+ } else if (opflag == dw->dw_op) {
+ dw->dw_domset |= dset;
+ }
+
+ /*
+ * Check if all the domains are spoken for that
+ * a particular waiter may have been waiting for.
+ * If there's at least one, we'll need to broadcast.
+ */
+ rset = (dw->dw_errset | dw->dw_domset) & dw->dw_reqset;
+ if (rset == dw->dw_reqset)
+ do_wakeup++;
+ }
+
+ PR_PROTO("%s: found %d waiters ready for wakeup\n", proc, do_wakeup);
+
+ if (do_wakeup > 0)
+ cv_broadcast(&idn.dopers->dop_cv);
+
+ mutex_exit(&idn.dopers->dop_mutex);
+}
+
+void
+idn_deinit_op(void *cookie)
+{
+ domainset_t domset;
+ dop_waitlist_t *hw, *tw;
+ dop_waitlist_t *dwl = (dop_waitlist_t *)cookie;
+
+ mutex_enter(&idn.dopers->dop_mutex);
+
+ ASSERT(idn.dopers->dop_waitlist);
+
+ if (dwl == idn.dopers->dop_waitlist) {
+ idn.dopers->dop_waitlist = dwl->dw_next;
+ if (IDNOP_IN_CACHE(dwl)) {
+ dwl->dw_next = idn.dopers->dop_freelist;
+ idn.dopers->dop_freelist = dwl;
+ } else {
+ FREESTRUCT(dwl, dop_waitlist_t, 1);
+ }
+ } else {
+ for (tw = idn.dopers->dop_waitlist, hw = tw->dw_next;
+ hw;
+ tw = hw, hw = hw->dw_next) {
+ if (dwl == hw)
+ break;
+ }
+ ASSERT(hw);
+
+ tw->dw_next = hw->dw_next;
+ }
+
+ /*
+ * Recompute domainset for which waiters might be waiting.
+ * It's possible there may be other waiters waiting for
+ * the same domainset that the current waiter that's leaving
+ * may have been waiting for, so we can't simply delete
+ * the leaving waiter's domainset from dop_domset.
+ */
+ for (hw = idn.dopers->dop_waitlist, domset = 0; hw; hw = hw->dw_next)
+ domset |= hw->dw_reqset;
+
+ idn.dopers->dop_waitcount--;
+ idn.dopers->dop_domset = domset;
+
+ mutex_exit(&idn.dopers->dop_mutex);
+}
+
+/*
+ * Wait until the specified operation succeeds or fails with
+ * respect to the given domains. Note the function terminates
+ * if at least one error occurs.
+ * This process is necessary since link/unlink operations occur
+ * asynchronously and we need some way of waiting to find out
+ * if it indeed completed.
+ * Timeout value is received indirectly from the SSP and
+ * represents seconds.
+ */
+int
+idn_wait_op(void *cookie, domainset_t *domsetp, int wait_timeout)
+{
+ int d, rv, err = 0;
+ dop_waitlist_t *dwl;
+
+
+ dwl = (dop_waitlist_t *)cookie;
+
+ ASSERT(wait_timeout > 0);
+ ASSERT((dwl->dw_op == IDNOP_CONNECTED) ||
+ (dwl->dw_op == IDNOP_DISCONNECTED));
+
+ mutex_enter(&idn.dopers->dop_mutex);
+
+ while (((dwl->dw_domset | dwl->dw_errset) != dwl->dw_reqset) && !err) {
+ rv = cv_timedwait_sig(&idn.dopers->dop_cv,
+ &idn.dopers->dop_mutex,
+ lbolt + (wait_timeout * hz));
+
+ if ((dwl->dw_domset | dwl->dw_errset) == dwl->dw_reqset)
+ break;
+
+ switch (rv) {
+ case -1:
+ /*
+ * timed out
+ */
+ cmn_err(CE_WARN,
+ "!IDN: 129: %s operation timed out",
+ (dwl->dw_op == IDNOP_CONNECTED) ? "LINK" :
+ (dwl->dw_op == IDNOP_DISCONNECTED) ? "UNLINK" :
+ "UNKNOWN");
+ /*FALLTHROUGH*/
+ case 0:
+ /*
+ * signal, e.g. kill(2)
+ */
+ err = 1;
+ break;
+
+ default:
+ break;
+ }
+ }
+
+ if (dwl->dw_domset == dwl->dw_reqset) {
+ rv = 0;
+ } else {
+ /*
+ * Op failed for some domains or we were awakened.
+ */
+ for (d = rv = 0; (d < MAX_DOMAINS) && !rv; d++)
+ rv = dwl->dw_errors[d];
+ }
+ *domsetp = dwl->dw_domset;
+
+ mutex_exit(&idn.dopers->dop_mutex);
+
+ idn_deinit_op(cookie);
+
+ return (rv);
+}
+
+/*
+ * --------------------------------------------------
+ * Return any valid (& ready) cpuid for the given board based on
+ * the given cpuset.
+ * --------------------------------------------------
+ */
+int
+board_to_ready_cpu(int board, cpuset_t cpuset)
+{
+ int base_cpuid;
+ int ncpu_board = MAX_CPU_PER_BRD;
+
+ board *= ncpu_board;
+ for (base_cpuid = board;
+ base_cpuid < (board + ncpu_board);
+ base_cpuid++)
+ if (CPU_IN_SET(cpuset, base_cpuid))
+ return (base_cpuid);
+
+ return (-1);
+}
+
+void
+idn_domain_resetentry(idn_domain_t *dp)
+{
+ register int i;
+ procname_t proc = "idn_domain_resetentry";
+
+ ASSERT(dp);
+ ASSERT(dp->dstate == IDNDS_CLOSED);
+ ASSERT(IDN_DLOCK_IS_EXCL(dp->domid));
+ ASSERT(IDN_GLOCK_IS_EXCL());
+
+ ASSERT(dp->domid == (dp - &idn_domain[0]));
+
+ IDN_FSTATE_TRANSITION(dp, IDNFIN_OFF);
+ dp->dname[0] = '\0';
+ dp->dnetid = (ushort_t)-1;
+ dp->dmtu = 0;
+ dp->dbufsize = 0;
+ dp->dslabsize = 0;
+ dp->dnwrsize = 0;
+ dp->dncpus = 0;
+ dp->dcpuindex = 0;
+ CPUSET_ZERO(dp->dcpuset);
+ dp->dcpu = dp->dcpu_last = dp->dcpu_save = IDN_NIL_DCPU;
+ dp->dvote.ticket = 0;
+ dp->dslab = NULL;
+ dp->dslab_state = DSLAB_STATE_UNKNOWN;
+ dp->dnslabs = 0;
+ dp->dio = 0;
+ dp->dioerr = 0;
+ lock_clear(&dp->diowanted);
+ bzero(&dp->dhw, sizeof (dp->dhw));
+ dp->dxp = NULL;
+ IDN_XSTATE_TRANSITION(dp, IDNXS_NIL);
+ dp->dsync.s_cmd = IDNSYNC_NIL;
+ dp->dfin_sync = IDNFIN_SYNC_OFF;
+ IDN_RESET_COOKIES(dp->domid);
+ dp->dcookie_err = 0;
+ bzero(&dp->dawol, sizeof (dp->dawol));
+ dp->dtmp = -1;
+
+ if (dp->dtimerq.tq_queue != NULL) {
+ PR_PROTO("%s: WARNING: MSG timerq not empty (count = %d)\n",
+ proc, dp->dtimerq.tq_count);
+ IDN_MSGTIMER_STOP(dp->domid, 0, 0);
+ }
+
+ for (i = 0; i < NCPU; i++)
+ dp->dcpumap[i] = (uchar_t)-1;
+}
+
+int
+idn_open_domain(int domid, int cpuid, uint_t ticket)
+{
+ int c, new_cpuid;
+ idn_domain_t *dp, *ldp;
+ procname_t proc = "idn_open_domain";
+
+ ASSERT(IDN_SYNC_IS_LOCKED());
+ ASSERT(IDN_DLOCK_IS_EXCL(domid));
+
+ if (!VALID_DOMAINID(domid)) {
+ PR_PROTO("%s: INVALID domainid (%d) "
+ "[cpuid = %d, ticket = 0x%x]\n",
+ proc, domid, cpuid, ticket);
+ return (-1);
+ }
+
+ dp = &idn_domain[domid];
+ ldp = &idn_domain[idn.localid];
+
+ if (dp->dcpu >= 0) {
+ PR_PROTO("%s:%d: domain already OPEN (state = %s)\n",
+ proc, domid, idnds_str[dp->dstate]);
+ return (1);
+ }
+
+ if (DOMAIN_IN_SET(idn.domset.ds_relink, domid)) {
+ if (dp->dcpu_save == IDN_NIL_DCPU)
+ new_cpuid = cpuid;
+ else
+ new_cpuid = dp->dcpu_save;
+ } else {
+ new_cpuid = cpuid;
+ }
+
+ if (new_cpuid == IDN_NIL_DCPU) {
+ PR_PROTO("%s:%d: WARNING: invalid cpuid (%d) specified\n",
+ proc, domid, new_cpuid);
+ return (-1);
+ }
+
+ IDN_GLOCK_EXCL();
+
+ idn_domain_resetentry(dp);
+
+ PR_STATE("%s:%d: requested cpuid %d, assigning cpuid %d\n",
+ proc, domid, cpuid, new_cpuid);
+
+ idn_assign_cookie(domid);
+
+ dp->dcpu = dp->dcpu_save = new_cpuid;
+ dp->dvote.ticket = ticket;
+ CPUSET_ADD(dp->dcpuset, new_cpuid);
+ dp->dncpus = 1;
+ for (c = 0; c < NCPU; c++)
+ dp->dcpumap[c] = (uchar_t)new_cpuid;
+ dp->dhw.dh_nboards = 1;
+ dp->dhw.dh_boardset = BOARDSET(CPUID_TO_BOARDID(new_cpuid));
+
+ if (domid != idn.localid)
+ IDN_DLOCK_EXCL(idn.localid);
+
+ if (idn.ndomains == 1) {
+ struct hwconfig local_hw;
+
+ /*
+ * We're attempting to connect to our first domain.
+ * Recheck our local hardware configuration before
+ * we go any further in case it changed due to a DR,
+ * and update any structs dependent on this.
+ * ASSUMPTION:
+ * IDN is unlinked before performing any DRs.
+ */
+ PR_PROTO("%s: RECHECKING local HW config.\n", proc);
+ if (get_hw_config(&local_hw)) {
+ dp->dcpu = IDN_NIL_DCPU;
+ cmn_err(CE_WARN,
+ "IDN: 118: hardware config not appropriate");
+ if (domid != idn.localid)
+ IDN_DUNLOCK(idn.localid);
+ IDN_GUNLOCK();
+ return (-1);
+ }
+ (void) update_local_hw_config(ldp, &local_hw);
+ }
+
+ idn.ndomains++;
+
+ if (domid != idn.localid)
+ IDN_DUNLOCK(idn.localid);
+ IDN_GUNLOCK();
+
+ IDN_MBOX_LOCK(domid);
+ dp->dmbox.m_tbl = NULL;
+
+ if (domid != idn.localid) {
+ dp->dmbox.m_send = idn_mainmbox_init(domid,
+ IDNMMBOX_TYPE_SEND);
+ dp->dmbox.m_recv = idn_mainmbox_init(domid,
+ IDNMMBOX_TYPE_RECV);
+ } else {
+ /*
+ * The local domain does not need send/recv
+ * mailboxes in its idn_domain[] entry.
+ */
+ dp->dmbox.m_send = NULL;
+ dp->dmbox.m_recv = NULL;
+ }
+ IDN_MBOX_UNLOCK(domid);
+
+ PR_PROTO("%s:%d: new domain (cpu = %d, vote = 0x%x)\n",
+ proc, domid, dp->dcpu, dp->dvote.ticket);
+
+ return (0);
+}
+
+/*
+ * The local domain never "closes" itself unless the driver
+ * is doing a idndetach. It will be reopened during idnattach
+ * when idn_domains_init is called.
+ */
+void
+idn_close_domain(int domid)
+{
+ uint_t token;
+ idn_domain_t *dp;
+ procname_t proc = "idn_close_domain";
+
+ ASSERT(IDN_SYNC_IS_LOCKED());
+ ASSERT(IDN_DLOCK_IS_EXCL(domid));
+
+ dp = &idn_domain[domid];
+
+ ASSERT(dp->dstate == IDNDS_CLOSED);
+
+ if (dp->dcpu == IDN_NIL_DCPU) {
+ PR_PROTO("%s:%d: DOMAIN ALREADY CLOSED!\n",
+ proc, domid);
+ return;
+ }
+
+ token = IDN_RETRY_TOKEN(domid, IDN_RETRY_TYPEALL);
+
+ (void) idn_retry_terminate(token);
+
+ DOMAINSET_DEL(idn.domset.ds_trans_on, domid);
+ DOMAINSET_DEL(idn.domset.ds_ready_on, domid);
+ DOMAINSET_DEL(idn.domset.ds_connected, domid);
+ DOMAINSET_DEL(idn.domset.ds_trans_off, domid);
+ DOMAINSET_DEL(idn.domset.ds_ready_off, domid);
+ DOMAINSET_DEL(idn.domset.ds_hwlinked, domid);
+ DOMAINSET_DEL(idn.domset.ds_flush, domid);
+
+ idn_sync_exit(domid, IDNSYNC_CONNECT);
+ idn_sync_exit(domid, IDNSYNC_DISCONNECT);
+
+ IDN_GLOCK_EXCL();
+
+ if (DOMAIN_IN_SET(idn.domset.ds_awol, domid))
+ idn_clear_awol(domid);
+
+ idn.ndomains--;
+
+ IDN_GUNLOCK();
+
+ IDN_MBOX_LOCK(domid);
+ dp->dmbox.m_tbl = NULL;
+
+ if (dp->dmbox.m_send) {
+ idn_mainmbox_deinit(domid, dp->dmbox.m_send);
+ dp->dmbox.m_send = NULL;
+ }
+
+ if (dp->dmbox.m_recv) {
+ idn_mainmbox_deinit(domid, dp->dmbox.m_recv);
+ dp->dmbox.m_recv = NULL;
+ }
+
+ IDN_MBOX_UNLOCK(domid);
+
+ cmn_err(CE_NOTE,
+ "!IDN: 142: link (domain %d, CPU %d) disconnected",
+ dp->domid, dp->dcpu);
+
+ dp->dcpu = IDN_NIL_DCPU; /* ultimate demise */
+
+ IDN_RESET_COOKIES(domid);
+
+ ASSERT(dp->dio <= 0);
+ ASSERT(dp->dioerr == 0);
+ ASSERT(dp->dslab == NULL);
+ ASSERT(dp->dnslabs == 0);
+
+ IDN_GKSTAT_GLOBAL_EVENT(gk_unlinks, gk_unlink_last);
+}
+
+
+/*
+ * -----------------------------------------------------------------------
+ */
+static void
+idn_domains_init(struct hwconfig *local_hw)
+{
+ register int i, d;
+ idn_domain_t *ldp;
+ uchar_t *cpumap;
+
+ ASSERT(local_hw != NULL);
+
+ cpumap = GETSTRUCT(uchar_t, NCPU * MAX_DOMAINS);
+
+ for (d = 0; d < MAX_DOMAINS; d++) {
+ register idn_domain_t *dp;
+
+ dp = &idn_domain[d];
+
+ dp->domid = d;
+
+ rw_init(&dp->drwlock, NULL, RW_DEFAULT, NULL);
+
+ IDN_TIMERQ_INIT(&dp->dtimerq);
+
+ dp->dstate = IDNDS_CLOSED;
+
+ mutex_init(&dp->dmbox.m_mutex, NULL, MUTEX_DEFAULT, NULL);
+
+ dp->dcpumap = cpumap;
+
+ rw_init(&dp->dslab_rwlock, NULL, RW_DEFAULT, NULL);
+
+ IDN_DLOCK_EXCL(d);
+ IDN_GLOCK_EXCL();
+
+ idn_domain_resetentry(dp);
+
+ IDN_GUNLOCK();
+
+ IDNSB_DOMAIN_UPDATE(dp);
+
+ IDN_DUNLOCK(d);
+
+ cpumap += NCPU;
+ }
+
+ IDN_SYNC_LOCK();
+
+ /*
+ * Update local domain information.
+ */
+ ASSERT(idn.smr.locpfn);
+ ASSERT(local_hw->dh_nboards && local_hw->dh_boardset);
+
+ idn.ndomains = 0; /* note that open_domain will get us to 1 */
+
+ IDN_DLOCK_EXCL(idn.localid);
+ d = idn_open_domain(idn.localid, (int)CPU->cpu_id, 0);
+ ASSERT(d == 0);
+ IDN_GLOCK_EXCL();
+ IDN_SET_MASTERID(IDN_NIL_DOMID);
+ IDN_SET_NEW_MASTERID(IDN_NIL_DOMID);
+
+ ldp = &idn_domain[idn.localid];
+
+ strncpy(ldp->dname, utsname.nodename, MAXDNAME - 1);
+ ldp->dname[MAXDNAME-1] = '\0';
+ bcopy(local_hw, &ldp->dhw, sizeof (ldp->dhw));
+ ASSERT(idn.ndomains == 1);
+ ASSERT((ldp->dhw.dh_nboards > 0) &&
+ (ldp->dhw.dh_nboards <= MAX_BOARDS));
+ ldp->dnetid = IDN_DOMID2NETID(ldp->domid);
+ ldp->dmtu = IDN_MTU;
+ ldp->dbufsize = IDN_SMR_BUFSIZE;
+ ldp->dslabsize = (short)IDN_SLAB_BUFCOUNT;
+ ldp->dnwrsize = (short)IDN_NWR_SIZE;
+ ldp->dcpuset = cpu_ready_set;
+ ldp->dncpus = (short)ncpus;
+ ldp->dvote.ticket = IDNVOTE_INITIAL_TICKET;
+ ldp->dvote.v.master = 0;
+ ldp->dvote.v.nmembrds = ldp->dhw.dh_nmcadr - 1;
+ ldp->dvote.v.ncpus = (int)ldp->dncpus - 1;
+ ldp->dvote.v.board = CPUID_TO_BOARDID(ldp->dcpu);
+ i = -1;
+ for (d = 0; d < NCPU; d++) {
+ BUMP_INDEX(ldp->dcpuset, i);
+ ldp->dcpumap[d] = (uchar_t)i;
+ }
+
+ CPUSET_ZERO(idn.dc_cpuset);
+ CPUSET_OR(idn.dc_cpuset, ldp->dcpuset);
+ idn.dc_boardset = ldp->dhw.dh_boardset;
+
+ /*
+ * Setting the state for ourselves is only relevant
+ * for loopback performance testing. Anyway, it
+ * makes sense that we always have an established
+ * connection with ourself regardless of IDN :-o
+ */
+ IDN_DSTATE_TRANSITION(ldp, IDNDS_CONNECTED);
+
+ IDN_GUNLOCK();
+ IDN_DUNLOCK(idn.localid);
+ IDN_SYNC_UNLOCK();
+}
+
+static void
+idn_domains_deinit()
+{
+ register int d;
+
+ IDN_SYNC_LOCK();
+ IDN_DLOCK_EXCL(idn.localid);
+ IDN_DSTATE_TRANSITION(&idn_domain[idn.localid], IDNDS_CLOSED);
+ idn_close_domain(idn.localid);
+ IDN_DUNLOCK(idn.localid);
+ IDN_SYNC_UNLOCK();
+ idn.localid = IDN_NIL_DOMID;
+
+ FREESTRUCT(idn_domain[0].dcpumap, uchar_t, NCPU * MAX_DOMAINS);
+
+ for (d = 0; d < MAX_DOMAINS; d++) {
+ idn_domain_t *dp;
+
+ dp = &idn_domain[d];
+
+ rw_destroy(&dp->dslab_rwlock);
+ mutex_destroy(&dp->dmbox.m_mutex);
+ rw_destroy(&dp->drwlock);
+ IDN_TIMERQ_DEINIT(&dp->dtimerq);
+ dp->dcpumap = NULL;
+ }
+}
+
+/*
+ * -----------------------------------------------------------------------
+ */
+static void
+idn_retrytask_init()
+{
+ ASSERT(idn.retryqueue.rq_cache == NULL);
+
+ mutex_init(&idn.retryqueue.rq_mutex, NULL, MUTEX_DEFAULT, NULL);
+ idn.retryqueue.rq_cache = kmem_cache_create("idn_retryjob_cache",
+ sizeof (idn_retry_job_t),
+ 0, NULL, NULL, NULL,
+ NULL, NULL, 0);
+}
+
+static void
+idn_retrytask_deinit()
+{
+ if (idn.retryqueue.rq_cache == NULL)
+ return;
+
+ kmem_cache_destroy(idn.retryqueue.rq_cache);
+ mutex_destroy(&idn.retryqueue.rq_mutex);
+
+ bzero(&idn.retryqueue, sizeof (idn.retryqueue));
+}
+
+/*
+ * -----------------------------------------------------------------------
+ */
+static void
+idn_timercache_init()
+{
+ ASSERT(idn.timer_cache == NULL);
+
+ idn.timer_cache = kmem_cache_create("idn_timer_cache",
+ sizeof (idn_timer_t),
+ 0, NULL, NULL, NULL,
+ NULL, NULL, 0);
+}
+
+static void
+idn_timercache_deinit()
+{
+ if (idn.timer_cache == NULL)
+ return;
+
+ kmem_cache_destroy(idn.timer_cache);
+ idn.timer_cache = NULL;
+}
+
+idn_timer_t *
+idn_timer_alloc()
+{
+ idn_timer_t *tp;
+
+ tp = kmem_cache_alloc(idn.timer_cache, KM_SLEEP);
+ bzero(tp, sizeof (*tp));
+ tp->t_forw = tp->t_back = tp;
+
+ return (tp);
+}
+
+void
+idn_timer_free(idn_timer_t *tp)
+{
+ if (tp == NULL)
+ return;
+ kmem_cache_free(idn.timer_cache, tp);
+}
+
+void
+idn_timerq_init(idn_timerq_t *tq)
+{
+ mutex_init(&tq->tq_mutex, NULL, MUTEX_DEFAULT, NULL);
+ tq->tq_count = 0;
+ tq->tq_queue = NULL;
+}
+
+void
+idn_timerq_deinit(idn_timerq_t *tq)
+{
+ ASSERT(tq->tq_queue == NULL);
+ mutex_destroy(&tq->tq_mutex);
+}
+
+/*
+ * Dequeue all the timers of the given subtype from the
+ * given timerQ. If subtype is 0, then dequeue all the
+ * timers.
+ */
+idn_timer_t *
+idn_timer_get(idn_timerq_t *tq, int type, ushort_t tcookie)
+{
+ register idn_timer_t *tp, *tphead;
+
+ ASSERT(IDN_TIMERQ_IS_LOCKED(tq));
+
+ if ((tp = tq->tq_queue) == NULL)
+ return (NULL);
+
+ if (!type) {
+ tq->tq_queue = NULL;
+ tq->tq_count = 0;
+ tphead = tp;
+ } else {
+ int count;
+ idn_timer_t *tpnext;
+
+ tphead = NULL;
+ count = tq->tq_count;
+ do {
+ tpnext = tp->t_forw;
+ if ((tp->t_type == type) &&
+ (!tcookie || (tp->t_cookie == tcookie))) {
+ tp->t_forw->t_back = tp->t_back;
+ tp->t_back->t_forw = tp->t_forw;
+ if (tphead == NULL) {
+ tp->t_forw = tp->t_back = tp;
+ } else {
+ tp->t_forw = tphead;
+ tp->t_back = tphead->t_back;
+ tp->t_back->t_forw = tp;
+ tphead->t_back = tp;
+ }
+ tphead = tp;
+ if (--(tq->tq_count) == 0)
+ tq->tq_queue = NULL;
+ else if (tq->tq_queue == tp)
+ tq->tq_queue = tpnext;
+ }
+ tp = tpnext;
+ } while (--count > 0);
+ }
+
+ if (tphead) {
+ tphead->t_back->t_forw = NULL;
+
+ for (tp = tphead; tp; tp = tp->t_forw)
+ tp->t_onq = 0;
+ }
+
+ return (tphead);
+}
+
+ushort_t
+idn_timer_start(idn_timerq_t *tq, idn_timer_t *tp, clock_t tval)
+{
+ idn_timer_t *otp;
+ ushort_t tcookie;
+ procname_t proc = "idn_timer_start";
+ STRING(str);
+
+ ASSERT(tq && tp && (tval > 0));
+ ASSERT((tp->t_forw == tp) && (tp->t_back == tp));
+ ASSERT(tp->t_type != 0);
+
+ IDN_TIMERQ_LOCK(tq);
+ /*
+ * Assign a unique non-zero 8-bit cookie to this timer
+ * if the caller hasn't already preassigned one.
+ */
+ while ((tcookie = tp->t_cookie) == 0) {
+ tp->t_cookie = (tp->t_type & 0xf) |
+ ((++tq->tq_cookie & 0xf) << 4);
+ /*
+ * Calculated cookie must never conflict
+ * with the public timer cookie.
+ */
+ ASSERT(tp->t_cookie != IDN_TIMER_PUBLIC_COOKIE);
+ }
+
+ /*
+ * First have to remove old timers of the
+ * same type and cookie, and get rid of them.
+ */
+ otp = idn_timer_get(tq, tp->t_type, tcookie);
+
+ tq->tq_count++;
+
+ if (tq->tq_queue == NULL) {
+ tq->tq_queue = tp;
+ ASSERT((tp->t_forw == tp) && (tp->t_back == tp));
+ } else {
+ /*
+ * Put me at the end of the list.
+ */
+ tp->t_forw = tq->tq_queue;
+ tp->t_back = tq->tq_queue->t_back;
+ tp->t_back->t_forw = tp;
+ tp->t_forw->t_back = tp;
+ }
+
+ tp->t_onq = 1;
+ tp->t_q = tq;
+ tp->t_id = timeout(idn_timer_expired, (caddr_t)tp, tval);
+
+
+ INUM2STR(tp->t_type, str);
+ PR_TIMER("%s: started %s timer (domain = %d, cookie = 0x%x)\n",
+ proc, str, tp->t_domid, tcookie);
+
+ IDN_TIMERQ_UNLOCK(tq);
+
+ if (otp)
+ (void) idn_timer_stopall(otp);
+
+ return (tcookie);
+}
+
+/*
+ * Stop all timers of the given subtype.
+ * If subtype is 0, then stop all timers
+ * in this timerQ.
+ */
+void
+idn_timer_stop(idn_timerq_t *tq, int type, ushort_t tcookie)
+{
+ idn_timer_t *tphead;
+ procname_t proc = "idn_timer_stop";
+ STRING(str);
+
+ ASSERT(tq);
+
+ INUM2STR(type, str);
+
+ IDN_TIMERQ_LOCK(tq);
+
+ if (tq->tq_count == 0) {
+ PR_TIMER("%s: found no %s timers (count=0)\n", proc, str);
+ IDN_TIMERQ_UNLOCK(tq);
+ return;
+ }
+ tphead = idn_timer_get(tq, type, tcookie);
+#ifdef DEBUG
+ if (tphead == NULL)
+ PR_TIMER("%s: found no %s (cookie = 0x%x) "
+ "timers (count=%d)!!\n",
+ proc, str, tcookie, tq->tq_count);
+#endif /* DEBUG */
+ IDN_TIMERQ_UNLOCK(tq);
+
+ if (tphead)
+ (void) idn_timer_stopall(tphead);
+}
+
+int
+idn_timer_stopall(idn_timer_t *tp)
+{
+ int count = 0;
+ int nonactive;
+ uint_t type;
+ idn_timer_t *ntp;
+ procname_t proc = "idn_timer_stopall";
+ STRING(str);
+
+ nonactive = 0;
+
+ if (tp) {
+ /*
+ * Circle should have been broken.
+ */
+ ASSERT(tp->t_back->t_forw == NULL);
+ type = tp->t_type;
+ INUM2STR(type, str);
+ }
+
+ for (; tp; tp = ntp) {
+ ntp = tp->t_forw;
+ count++;
+ ASSERT(tp->t_id != (timeout_id_t)0);
+ if (untimeout(tp->t_id) < 0) {
+ nonactive++;
+ PR_TIMER("%s: bad %s untimeout (domain=%d)\n",
+ proc, str, tp->t_domid);
+ } else {
+ PR_TIMER("%s: good %s untimeout (domain=%d)\n",
+ proc, str, tp->t_domid);
+ }
+ /*
+ * There are two possible outcomes from
+ * the untimeout(). Each ultimately result
+ * in us having to free the timeout structure.
+ *
+ * 1. We successfully aborted a timeout call.
+ *
+ * 2. We failed to find the given timer. He
+ * probably just fired off.
+ */
+ idn_timer_free(tp);
+ }
+ PR_TIMER("%s: stopped %d of %d %s timers\n",
+ proc, count - nonactive, count, str);
+
+ return (count);
+}
+
+void
+idn_timer_dequeue(idn_timerq_t *tq, idn_timer_t *tp)
+{
+ ASSERT(tq && tp);
+ ASSERT(IDN_TIMERQ_IS_LOCKED(tq));
+
+ ASSERT(tp->t_q == tq);
+
+ if (tp->t_onq == 0) {
+ /*
+ * We've already been dequeued.
+ */
+ ASSERT(tp == tp->t_forw);
+ ASSERT(tp == tp->t_back);
+ } else {
+ /*
+ * We're still in the queue, get out.
+ */
+ if (tq->tq_queue == tp)
+ tq->tq_queue = tp->t_forw;
+ tp->t_forw->t_back = tp->t_back;
+ tp->t_back->t_forw = tp->t_forw;
+ tp->t_onq = 0;
+ if (--(tq->tq_count) == 0) {
+ ASSERT(tq->tq_queue == tp);
+ tq->tq_queue = NULL;
+ }
+ tp->t_forw = tp->t_back = tp;
+ }
+}
+
+/*
+ * -----------------------------------------------------------------------
+ */
+/*ARGSUSED*/
+static int
+idn_slabpool_report(queue_t *wq, mblk_t *mp, caddr_t cp, cred_t *cr)
+{
+ register int p, nfree;
+ char dsetstr[128];
+
+ ASSERT(IDN_GLOCK_IS_HELD());
+
+ if (idn.slabpool == NULL) {
+ (void) mi_mpprintf(mp,
+ "IDN slabpool not initialized (masterid = %d)",
+ IDN_GET_MASTERID());
+ return (0);
+ }
+
+ for (p = nfree = 0; p < idn.slabpool->npools; p++)
+ nfree += idn.slabpool->pool[p].nfree;
+
+ (void) mi_mpprintf(mp,
+ "IDN slabpool (ntotal_slabs = %d, nalloc = %d, "
+ "npools = %d)",
+ idn.slabpool->ntotslabs,
+ idn.slabpool->ntotslabs - nfree,
+ idn.slabpool->npools);
+
+ (void) mi_mpprintf(mp, "pool nslabs nfree domains");
+
+ for (p = 0; p < idn.slabpool->npools; p++) {
+ register int d, s;
+ uint_t domset;
+
+ domset = 0;
+ for (s = 0; s < idn.slabpool->pool[p].nslabs; s++) {
+ short dd;
+
+ dd = idn.slabpool->pool[p].sarray[s].sl_domid;
+ if (dd != (short)IDN_NIL_DOMID)
+ DOMAINSET_ADD(domset, dd);
+ }
+ dsetstr[0] = '\0';
+ if (domset) {
+ for (d = 0; d < MAX_DOMAINS; d++) {
+ if (!DOMAIN_IN_SET(domset, d))
+ continue;
+
+ if (dsetstr[0] == '\0')
+ (void) sprintf(dsetstr, "%d", d);
+ else
+ (void) sprintf(dsetstr, "%s %d",
+ dsetstr, d);
+ }
+ }
+
+ if (p < 10)
+ (void) mi_mpprintf(mp, " %d %d %d %s",
+ p, idn.slabpool->pool[p].nslabs,
+ idn.slabpool->pool[p].nfree,
+ dsetstr);
+ else
+ (void) mi_mpprintf(mp, " %d %d %d %s",
+ p, idn.slabpool->pool[p].nslabs,
+ idn.slabpool->pool[p].nfree,
+ dsetstr);
+ }
+ return (0);
+}
+
+/*ARGSUSED*/
+static int
+idn_buffer_report(queue_t *wq, mblk_t *mp, caddr_t cp, cred_t *cr)
+{
+ smr_slab_t *sp;
+ register int d, cnt;
+ int bufcount[MAX_DOMAINS];
+ int spl;
+
+ ASSERT(IDN_GLOCK_IS_HELD());
+
+ if (idn.localid == IDN_NIL_DOMID) {
+ (void) mi_mpprintf(mp, "IDN not initialized (localid = %d)",
+ idn.localid);
+ return (0);
+ }
+
+ (void) mi_mpprintf(mp, "Local domain has %d slabs allocated.",
+ idn_domain[idn.localid].dnslabs);
+
+ DSLAB_LOCK_SHARED(idn.localid);
+ if ((sp = idn_domain[idn.localid].dslab) == NULL) {
+ DSLAB_UNLOCK(idn.localid);
+ return (0);
+ }
+
+ bzero(bufcount, sizeof (bufcount));
+ cnt = 0;
+
+ spl = splhi();
+ for (; sp; sp = sp->sl_next) {
+ smr_slabbuf_t *bp;
+
+ while (!lock_try(&sp->sl_lock))
+ ;
+ for (bp = sp->sl_inuse; bp; bp = bp->sb_next) {
+ bufcount[bp->sb_domid]++;
+ cnt++;
+ }
+ lock_clear(&sp->sl_lock);
+ }
+ splx(spl);
+
+ DSLAB_UNLOCK(idn.localid);
+
+ (void) mi_mpprintf(mp, "Local domain has %d buffers outstanding.", cnt);
+ if (cnt == 0)
+ return (0);
+
+ (void) mi_mpprintf(mp, "Domain nbufs");
+ for (d = 0; d < MAX_DOMAINS; d++)
+ if (bufcount[d]) {
+ if (d < 10)
+ (void) mi_mpprintf(mp, " %d %d",
+ d, bufcount[d]);
+ else
+ (void) mi_mpprintf(mp, " %d %d",
+ d, bufcount[d]);
+ }
+
+ return (0);
+}
+
+static const char *
+_get_spaces(int w, int s, int W)
+{
+ static const char *const _spaces[] = {
+ "", /* 0 */
+ " ", /* 1 */
+ " ", /* 2 */
+ " ", /* 3 */
+ " ", /* 4 */
+ " ", /* 5 */
+ " ", /* 6 */
+ " ", /* 7 */
+ " ", /* 8 */
+ " ", /* 9 */
+ " ", /* 10 */
+ " ", /* 11 */
+ " ", /* 12 */
+ " ", /* 13 */
+ " ", /* 14 */
+ " ", /* 15 */
+ " ", /* 16 */
+ " ", /* 17 */
+ " ", /* 18 */
+ " ", /* 19 */
+ };
+ return (_spaces[w+s-W]);
+}
+
+#define _SSS(X, W, w, s) \
+ (((w) >= (W)) && (X)) ? _get_spaces((w), (s), (W))
+
+static const char *
+_hexspace(uint64_t v, int sz, int width, int padding)
+{
+ int maxnbl = 16;
+ int diff;
+ uchar_t *np;
+
+ diff = sizeof (uint64_t) - sz;
+ np = (uchar_t *)&v + diff;
+ maxnbl -= diff << 1;
+ while (sz-- > 0) {
+ if ((*np & 0xf0) && (width >= maxnbl))
+ return (_get_spaces(width, padding, maxnbl));
+ maxnbl--;
+ if ((*np & 0x0f) && (width >= maxnbl))
+ return (_get_spaces(width, padding, maxnbl));
+ maxnbl--;
+ np++;
+ }
+ return (_get_spaces(width, padding, 1));
+}
+
+#define HEXSPACE(v, t, w, s) _hexspace((uint64_t)(v), sizeof (t), (w), (s))
+
+#define DECSPACE(n, w, s) \
+ (_SSS((n) >= 10000000, 8, (w), (s)) : \
+ _SSS((n) >= 1000000, 7, (w), (s)) : \
+ _SSS((n) >= 100000, 6, (w), (s)) : \
+ _SSS((n) >= 10000, 5, (w), (s)) : \
+ _SSS((n) >= 1000, 4, (w), (s)) : \
+ _SSS((n) >= 100, 3, (w), (s)) : \
+ _SSS((n) >= 10, 2, (w), (s)) : \
+ _get_spaces((w), (s), 1))
+
+#define MBXINFO(mtp) \
+ &mtp->mt_header, \
+ HEXSPACE(&mtp->mt_header, &mtp->mt_header, 16, 2), \
+ mtp->mt_header.mh_svr_ready_ptr, \
+ HEXSPACE(mtp->mt_header.mh_svr_ready_ptr, \
+ mtp->mt_header.mh_svr_ready_ptr, 8, 1), \
+ mtp->mt_header.mh_svr_active_ptr, \
+ HEXSPACE(mtp->mt_header.mh_svr_active_ptr, \
+ mtp->mt_header.mh_svr_active_ptr, 8, 2), \
+ *(ushort_t *)(IDN_OFFSET2ADDR(mtp->mt_header.mh_svr_ready_ptr)), \
+ DECSPACE(*(ushort_t *) \
+ (IDN_OFFSET2ADDR(mtp->mt_header.mh_svr_ready_ptr)), \
+ 1, 1), \
+ *(ushort_t *)(IDN_OFFSET2ADDR(mtp->mt_header.mh_svr_active_ptr)), \
+ DECSPACE(*(ushort_t *) \
+ (IDN_OFFSET2ADDR(mtp->mt_header.mh_svr_active_ptr)), \
+ 1, 5), \
+ mtp->mt_header.mh_cookie, \
+ HEXSPACE(mtp->mt_header.mh_cookie, \
+ mtp->mt_header.mh_cookie, 8, 2), \
+ &mtp->mt_queue[0], \
+ HEXSPACE(&mtp->mt_queue[0], &mtp->mt_queue[0], 16, 2)
+
+/*ARGSUSED*/
+static int
+idn_mboxtbl_report(queue_t *wq, mblk_t *mp, caddr_t cp, cred_t *cr)
+{
+ register int c, n, domid, subdomid;
+ register idn_mboxtbl_t *mtp;
+ register idn_mboxmsg_t *msp;
+ idn_mboxtbl_t *map, *mtbasep;
+
+
+ ASSERT((cp == MBXTBL_PART_REPORT) || (cp == MBXTBL_FULL_REPORT));
+
+ if (IDN_GLOCK_TRY_SHARED() == 0) {
+ (void) mi_mpprintf(mp, "Local domain busy, try again.");
+ return (0);
+ }
+
+ if ((map = idn.mboxarea) == NULL) {
+ (void) mi_mpprintf(mp,
+ "WARNING: Local domain is not master, "
+ "ASSUMING idn.smr.vaddr.");
+ map = (idn_mboxtbl_t *)idn.smr.vaddr;
+ }
+
+ if (map) {
+ (void) mi_mpprintf(mp, "Mailbox Area starts @ 0x%p",
+ map);
+ } else {
+ (void) mi_mpprintf(mp, "Mailbox Area not found.");
+ goto repdone;
+ }
+
+ if (!idn.nchannels) {
+ (void) mi_mpprintf(mp, "No OPEN channels found");
+ goto repdone;
+ }
+
+ for (c = 0; c < IDN_MAX_NETS; c++) {
+
+ IDN_CHAN_LOCK_GLOBAL(&idn.chan_servers[c]);
+ if (!IDN_CHANNEL_IS_ATTACHED(&idn.chan_servers[c])) {
+ IDN_CHAN_UNLOCK_GLOBAL(&idn.chan_servers[c]);
+ continue;
+ }
+
+ (void) mi_mpprintf(mp,
+ "Channel %d ---------------------------"
+ "--------------------------"
+ "-----------------------------", c);
+ (void) mi_mpprintf(mp,
+ " Domain Header "
+ "Ready/Active Ptrs "
+ "rdy/actv cookie Queue "
+ "busy");
+
+ for (domid = 0; domid < MAX_DOMAINS; domid++) {
+ register int busy_count;
+
+ if ((cp == MBXTBL_PART_REPORT) &&
+ (idn_domain[domid].dcpu == IDN_NIL_DCPU))
+ continue;
+
+ mtbasep = IDN_MBOXAREA_BASE(map, domid);
+
+ for (subdomid = 0; subdomid < MAX_DOMAINS;
+ subdomid++) {
+ mtp = IDN_MBOXTBL_PTR(mtbasep, subdomid);
+ mtp = IDN_MBOXTBL_PTR_CHAN(mtp, c);
+
+ if (subdomid == domid) {
+ if (subdomid == 0)
+ (void) mi_mpprintf(mp,
+ " %x.%x-%d%s%s",
+ domid, subdomid, c,
+ /*CONSTCOND*/
+ DECSPACE(c, 2, 2),
+ "-- unused --");
+ else
+ (void) mi_mpprintf(mp,
+ " .%x-%d%s%s",
+ subdomid, c,
+ /*CONSTCOND*/
+ DECSPACE(c, 2, 2),
+ "-- unused --");
+ continue;
+ }
+ busy_count = 0;
+ msp = &mtp->mt_queue[0];
+ for (n = 0; n < IDN_MMBOX_NUMENTRIES; n++) {
+ if (msp[n].ms_owner)
+ busy_count++;
+ }
+ if (subdomid == 0) {
+ (void) mi_mpprintf(mp,
+ " %x.%x-%d%s%p%s%x%s/ %x%s"
+ "%d%s/ %d%s%x%s%p%s%d%s",
+ domid, subdomid, c,
+ /*CONSTCOND*/
+ DECSPACE(c, 2, 2),
+ MBXINFO(mtp), busy_count,
+ busy_count ? " <<<<<":"");
+ } else {
+ (void) mi_mpprintf(mp,
+ " .%x-%d%s%p%s%x%s/ %x%s"
+ "%d%s/ %d%s%x%s%p%s%d%s",
+ subdomid, c,
+ /*CONSTCOND*/
+ DECSPACE(c, 2, 2),
+ MBXINFO(mtp), busy_count,
+ busy_count ? " <<<<<":"");
+ }
+ }
+ }
+ IDN_CHAN_UNLOCK_GLOBAL(&idn.chan_servers[c]);
+ }
+
+repdone:
+ IDN_GUNLOCK();
+
+ return (0);
+}
+
+/*ARGSUSED*/
+static void
+idn_mainmbox_domain_report(queue_t *wq, mblk_t *mp, int domid,
+ idn_mainmbox_t *mmp, char *mbxtype)
+{
+ register int c;
+
+ if (mmp == NULL) {
+ (void) mi_mpprintf(mp, " %x.%s -- none --", domid, mbxtype);
+ return;
+ }
+
+ for (c = 0; c < IDN_MAX_NETS; mmp++, c++) {
+ int mm_count;
+
+ IDN_CHAN_LOCK_GLOBAL(&idn.chan_servers[c]);
+ if (IDN_CHANNEL_IS_DETACHED(&idn.chan_servers[c])) {
+ (void) mi_mpprintf(mp, " %x.%s %u -- not open --",
+ domid, mbxtype, (int)mmp->mm_channel);
+ IDN_CHAN_UNLOCK_GLOBAL(&idn.chan_servers[c]);
+ continue;
+ }
+
+ mm_count = ((mmp->mm_count < 0) ? 0 : mmp->mm_count) / 1000;
+
+ (void) mi_mpprintf(mp, " %x.%s %d%s%d%s%d%s%p%s%p%s%p%s%d/%d",
+ domid, mbxtype,
+ (int)mmp->mm_channel,
+ /*CONSTCOND*/
+ DECSPACE((int)mmp->mm_channel, 5, 2),
+ mm_count, DECSPACE(mm_count, 8, 2),
+ mmp->mm_dropped,
+ DECSPACE(mmp->mm_dropped, 8, 2),
+ mmp->mm_smr_mboxp,
+ HEXSPACE(mmp->mm_smr_mboxp,
+ mmp->mm_smr_mboxp, 16, 2),
+ mmp->mm_smr_readyp,
+ HEXSPACE(mmp->mm_smr_readyp,
+ mmp->mm_smr_readyp, 16, 2),
+ mmp->mm_smr_activep,
+ HEXSPACE(mmp->mm_smr_activep,
+ mmp->mm_smr_activep, 16, 2),
+ mmp->mm_qiget, mmp->mm_qiput);
+ IDN_CHAN_UNLOCK_GLOBAL(&idn.chan_servers[c]);
+ }
+}
+
+/*ARGSUSED2*/
+static int
+idn_mainmbox_report(queue_t *wq, mblk_t *mp, caddr_t cp, cred_t *cr)
+{
+ int domid;
+ int header = 0;
+
+ /*
+ * Domain 0 never has a send/recv mainmbox so
+ * don't bother printing him.
+ */
+ for (domid = 1; domid < MAX_DOMAINS; domid++) {
+ idn_domain_t *dp;
+
+ dp = &idn_domain[domid];
+
+ if (dp->dcpu == IDN_NIL_DCPU)
+ continue;
+ IDN_DLOCK_SHARED(domid);
+ if (dp->dcpu == IDN_NIL_DCPU) {
+ IDN_DUNLOCK(domid);
+ continue;
+ }
+ if (!header) {
+ (void) mi_mpprintf(mp,
+ "Domain Chan PktCntK "
+ "PktDrop SMRMbox "
+ "ReadyPtr "
+ "ActvPtr Miget/Miput");
+ header = 1;
+ }
+
+ mutex_enter(&dp->dmbox.m_mutex);
+ idn_mainmbox_domain_report(wq, mp, domid,
+ idn_domain[domid].dmbox.m_send,
+ "snd");
+ idn_mainmbox_domain_report(wq, mp, domid,
+ idn_domain[domid].dmbox.m_recv,
+ "rcv");
+ mutex_exit(&dp->dmbox.m_mutex);
+
+ IDN_DUNLOCK(domid);
+
+ (void) mi_mpprintf(mp,
+ " ---------------------------------------"
+ "------------------------"
+ "----------------------------");
+ }
+
+ if (!header)
+ (void) mi_mpprintf(mp, "No ACTIVE domain connections exist");
+
+ return (0);
+}
+
+/*ARGSUSED*/
+static int
+idn_global_report(queue_t *wq, mblk_t *mp, caddr_t cp, cred_t *cr)
+{
+ int i, nactive, masterid, nretry;
+ uint_t locpfn_upper, locpfn_lower,
+ rempfn_upper, rempfn_lower;
+ uint_t marea_upper, marea_lower,
+ iarea_upper, iarea_lower;
+ char alt_dbuffer[64];
+ idn_retry_job_t *rp;
+ domainset_t retryset;
+ domainset_t connected;
+ idn_synczone_t *zp;
+ idn_syncop_t *sp;
+ idn_domain_t *dp;
+ char *dbp, *dbuffer;
+
+ if (IDN_SYNC_TRYLOCK() == 0) {
+ (void) mi_mpprintf(mp, "Sync lock busy, try again.");
+ return (0);
+ }
+
+ if (IDN_GLOCK_TRY_SHARED() == 0) {
+ (void) mi_mpprintf(mp, "Local domain busy, try again.");
+ IDN_SYNC_UNLOCK();
+ return (0);
+ }
+ if ((dbp = dbuffer = ALLOC_DISPSTRING()) == NULL)
+ dbp = alt_dbuffer;
+
+ (void) mi_mpprintf(mp, "IDN\n Global State = %s (%d)",
+ idngs_str[idn.state], idn.state);
+
+ (void) mi_mpprintf(mp, "SMR");
+ (void) mi_mpprintf(mp, " vaddr ");
+ (void) mi_mpprintf(mp, " 0x%p", idn.smr.vaddr);
+
+ (void) mi_mpprintf(mp, " paddr-local paddr-remote");
+ masterid = IDN_GET_MASTERID();
+ locpfn_upper = (uint_t)(idn.smr.locpfn >> (32 - PAGESHIFT));
+ locpfn_lower = (uint_t)(idn.smr.locpfn << PAGESHIFT);
+ if (idn.smr.rempfn == PFN_INVALID) {
+ rempfn_upper = rempfn_lower = 0;
+ } else {
+ rempfn_upper = (uint_t)(idn.smr.rempfn >> (32 - PAGESHIFT));
+ rempfn_lower = (uint_t)(idn.smr.rempfn << PAGESHIFT);
+ }
+ (void) mi_mpprintf(mp, " 0x%x.%x%s0x%x.%x",
+ locpfn_upper, locpfn_lower,
+ HEXSPACE(locpfn_lower, locpfn_lower, 8,
+ (locpfn_upper < 0x10) ? 4 : 3),
+ rempfn_upper, rempfn_lower);
+
+ (void) mi_mpprintf(mp, " SMR length = %d MBytes", IDN_SMR_SIZE);
+ (void) mi_mpprintf(mp, " SMR bufsize = %d Bytes", IDN_SMR_BUFSIZE);
+ (void) mi_mpprintf(mp, " NWR length = %d MBytes", IDN_NWR_SIZE);
+ marea_upper = (uint_t)((uint64_t)IDN_MBOXAREA_SIZE >> 32);
+ marea_lower = (uint_t)((uint64_t)IDN_MBOXAREA_SIZE & 0xffffffff);
+ iarea_upper = (uint_t)((uint64_t)(MB2B(IDN_NWR_SIZE) -
+ (size_t)IDN_MBOXAREA_SIZE) >> 32);
+ iarea_lower = (uint_t)((MB2B(IDN_NWR_SIZE) -
+ (size_t)IDN_MBOXAREA_SIZE) & 0xffffffff);
+ (void) mi_mpprintf(mp,
+ " [ mbox area = 0x%x.%x Bytes, "
+ "iobuf area = 0x%x.%x Bytes ]",
+ marea_upper, marea_lower, iarea_upper, iarea_lower);
+
+ (void) mi_mpprintf(mp,
+ "\nIDNnet (local domain [id:%d] [name:%s] is %s)",
+ idn.localid,
+ idn_domain[idn.localid].dname,
+ (masterid == IDN_NIL_DOMID) ? "IDLE" :
+ (idn.localid == masterid) ? "MASTER" :
+ "SLAVE");
+ nactive = 0;
+ for (i = 0; i < IDN_MAX_NETS; i++) {
+ IDN_CHAN_LOCK_GLOBAL(&idn.chan_servers[i]);
+ if (IDN_CHANNEL_IS_ACTIVE(&idn.chan_servers[i]))
+ nactive++;
+ IDN_CHAN_UNLOCK_GLOBAL(&idn.chan_servers[i]);
+ }
+ (void) mi_mpprintf(mp, " I/O Networks: (Open = %d, "
+ "Active = %d, Max = %d)",
+ idn.nchannels, nactive, IDN_MAX_NETS);
+ (void) mi_mpprintf(mp, " Number of Domains = %d", idn.ndomains);
+ (void) mi_mpprintf(mp, " Number of AWOLs = %d", idn.nawols);
+ /*
+ * During connect domains can possibly be in ds_connected
+ * while still in ds_trans_on. Only once they leave ds_trans_on
+ * are they really connected.
+ */
+ connected = idn.domset.ds_connected & ~idn.domset.ds_trans_on;
+ DOMAINSET_ADD(connected, idn.localid);
+ boardset2str(connected, dbp);
+ (void) mi_mpprintf(mp, " Connected Domains = %s", dbp);
+ domainset2str(idn.domset.ds_trans_on, dbp);
+ (void) mi_mpprintf(mp, " Pending Domain Links = %s",
+ idn.domset.ds_trans_on ? dbp : "<>");
+ domainset2str(idn.domset.ds_trans_off, dbp);
+ (void) mi_mpprintf(mp, " Pending Domain Unlinks = %s",
+ idn.domset.ds_trans_off ? dbp : "<>");
+ mutex_enter(&idn.retryqueue.rq_mutex);
+ nretry = idn.retryqueue.rq_count;
+ retryset = 0;
+ for (i = 0, rp = idn.retryqueue.rq_jobs; i < nretry; i++,
+ rp = rp->rj_next) {
+ int domid;
+
+ domid = IDN_RETRY_TOKEN2DOMID(rp->rj_token);
+ if (VALID_DOMAINID(domid)) {
+ DOMAINSET_ADD(retryset, domid);
+ }
+ }
+ mutex_exit(&idn.retryqueue.rq_mutex);
+ domainset2str(retryset, dbp);
+ (void) mi_mpprintf(mp, " Retry Jobs:Domains = %d:%s",
+ nretry, retryset ? dbp : "<>");
+ domainset2str(idn.domset.ds_hitlist, dbp);
+ (void) mi_mpprintf(mp, " Hitlist Domains = %s",
+ idn.domset.ds_hitlist ? dbp : "<>");
+ domainset2str(idn.domset.ds_relink, dbp);
+ (void) mi_mpprintf(mp, " Reconfig Domains = %s",
+ idn.domset.ds_relink ? dbp : "<>");
+ if (idn.domset.ds_relink)
+ (void) mi_mpprintf(mp, " new master id = %d",
+ IDN_GET_NEW_MASTERID());
+ if (masterid == IDN_NIL_DOMID) {
+ (void) mi_mpprintf(mp, " Master Domain: no master");
+ } else {
+ idn_domain_t *mdp;
+
+ mdp = &idn_domain[masterid];
+
+ (void) mi_mpprintf(mp,
+ " Master Domain (id:name/brds - state):");
+
+ if (strlen(mdp->dname) > 0)
+ strcpy(dbp, mdp->dname);
+ else
+ boardset2str(mdp->dhw.dh_boardset, dbp);
+ if (masterid < 10)
+ (void) mi_mpprintf(mp, " %d: %s - %s",
+ masterid, dbp,
+ idnds_str[mdp->dstate]);
+ else
+ (void) mi_mpprintf(mp, " %d: %s - %s",
+ masterid, dbp,
+ idnds_str[mdp->dstate]);
+ }
+ if (idn.ndomains <= 1) {
+ (void) mi_mpprintf(mp, " Slave Domains: none");
+ } else {
+ int d;
+
+ (void) mi_mpprintf(mp,
+ " Slave Domains (id:name/brds - state):");
+ for (d = 0; d < MAX_DOMAINS; d++) {
+ dp = &idn_domain[d];
+
+ if ((dp->dcpu == IDN_NIL_DCPU) || (d == masterid))
+ continue;
+
+ if (strlen(dp->dname) > 0)
+ strcpy(dbp, dp->dname);
+ else
+ boardset2str(dp->dhw.dh_boardset, dbp);
+ if (d < 10)
+ (void) mi_mpprintf(mp, " %d: %s - %s",
+ d, dbp,
+ idnds_str[dp->dstate]);
+ else
+ (void) mi_mpprintf(mp, " %d: %s - %s",
+ d, dbp,
+ idnds_str[dp->dstate]);
+ }
+ }
+
+ if (idn.nawols == 0) {
+ (void) mi_mpprintf(mp, " AWOL Domains: none");
+ } else {
+ int d;
+
+ (void) mi_mpprintf(mp, " AWOL Domains (id:name/brds):");
+ for (d = 0; d < MAX_DOMAINS; d++) {
+ dp = &idn_domain[d];
+
+ if (!DOMAIN_IN_SET(idn.domset.ds_awol, d) ||
+ (dp->dcpu == IDN_NIL_DCPU))
+ continue;
+
+ if (strlen(dp->dname) > 0)
+ strcpy(dbp, dp->dname);
+ else
+ boardset2str(dp->dhw.dh_boardset, dbp);
+ if (d < 10)
+ (void) mi_mpprintf(mp, " %d: %s",
+ d, dbp);
+ else
+ (void) mi_mpprintf(mp, " %d: %s",
+ d, dbp);
+ }
+ }
+
+ /*CONSTCOND*/
+ i = IDN_SYNC_GETZONE(IDNSYNC_CONNECT);
+ zp = &idn.sync.sz_zone[i];
+ if (zp->sc_cnt == 0) {
+ (void) mi_mpprintf(mp, " Sync Zone (con): [empty]");
+ } else {
+ (void) mi_mpprintf(mp, " Sync Zone (con): [%d domains]",
+ zp->sc_cnt);
+ sp = zp->sc_op;
+ for (i = 0; (i < zp->sc_cnt) && sp; i++) {
+ (void) mi_mpprintf(mp,
+ " "
+ "%x: x_set =%s0x%x, r_set =%s0x%x",
+ sp->s_domid,
+ HEXSPACE(sp->s_set_exp,
+ sp->s_set_exp, 4, 1),
+ sp->s_set_exp,
+ HEXSPACE(sp->s_set_rdy,
+ sp->s_set_rdy, 4, 1),
+ sp->s_set_rdy);
+ sp = sp->s_next;
+ }
+ }
+ /*CONSTCOND*/
+ i = IDN_SYNC_GETZONE(IDNSYNC_DISCONNECT);
+ zp = &idn.sync.sz_zone[i];
+ if (zp->sc_cnt == 0) {
+ (void) mi_mpprintf(mp, " Sync Zone (dis): [empty]");
+ } else {
+ (void) mi_mpprintf(mp, " Sync Zone (dis): [%d domains]",
+ zp->sc_cnt);
+ sp = zp->sc_op;
+ for (i = 0; (i < zp->sc_cnt) && sp; i++) {
+ (void) mi_mpprintf(mp,
+ " "
+ "%x: x_set =%s0x%x, r_set =%s0x%x",
+ sp->s_domid,
+ HEXSPACE(sp->s_set_exp,
+ sp->s_set_exp, 4, 1),
+ sp->s_set_exp,
+ HEXSPACE(sp->s_set_rdy,
+ sp->s_set_rdy, 4, 1),
+ sp->s_set_rdy);
+ sp = sp->s_next;
+ }
+ }
+
+ IDN_GUNLOCK();
+ IDN_SYNC_UNLOCK();
+
+ if (dbuffer) {
+ FREE_DISPSTRING(dbuffer);
+ }
+
+ return (0);
+}
+
+/*ARGSUSED*/
+static int
+idn_domain_report(queue_t *wq, mblk_t *mp, caddr_t cp, cred_t *cr)
+{
+ int d, nchan;
+ uint_t domset;
+ idn_chanset_t chanset;
+ idn_domain_t *dp;
+ uint_t pset_upper, pset_lower;
+ char *dbuffer, *dbp;
+ char alt_dbuffer[64];
+
+
+ if (IDN_SYNC_TRYLOCK() == 0) {
+ (void) mi_mpprintf(mp, "Sync lock busy, try again.");
+ return (0);
+ }
+
+ if (IDN_GLOCK_TRY_SHARED() == 0) {
+ (void) mi_mpprintf(mp, "Local domain busy, try again.");
+ IDN_SYNC_UNLOCK();
+ return (0);
+ }
+
+ if ((dbp = dbuffer = ALLOC_DISPSTRING()) == NULL)
+ dbp = alt_dbuffer;
+
+ if ((int)cp == 0)
+ domset = DOMAINSET(idn.localid);
+ else
+ domset = DOMAINSET_ALL;
+
+ for (d = 0; d < MAX_DOMAINS; d++) {
+
+ if (DOMAIN_IN_SET(domset, d) == 0)
+ continue;
+
+ dp = &idn_domain[d];
+
+ if (dp->dcpu == IDN_NIL_DCPU)
+ continue;
+
+ if (IDN_DLOCK_TRY_SHARED(d) == 0) {
+ if (d < 10)
+ (void) mi_mpprintf(mp,
+ "Domain %d (0x%p) busy...",
+ d, dp);
+ else
+ (void) mi_mpprintf(mp,
+ "Domain %d (0x%p) busy...",
+ d, dp);
+ continue;
+ }
+ if (dp->dcpu == IDN_NIL_DCPU) {
+ IDN_DUNLOCK(d);
+ continue;
+ }
+ if (d < 10)
+ (void) mi_mpprintf(mp, "%sDomain %d (0x%p)",
+ (d && (idn.ndomains > 1)) ? "\n" : "",
+ d, dp);
+ else
+ (void) mi_mpprintf(mp, "%sDomain %d (0x%p)",
+ (d && (idn.ndomains > 1)) ? "\n" : "",
+ d, dp);
+
+ if (d == idn.localid)
+ (void) mi_mpprintf(mp, " (local) State = %s (%d)",
+ idnds_str[dp->dstate], dp->dstate);
+ else
+ (void) mi_mpprintf(mp, " State = %s (%d)",
+ idnds_str[dp->dstate], dp->dstate);
+ (void) mi_mpprintf(mp, " Name = %s, Netid = %d",
+ (strlen(dp->dname) > 0) ? dp->dname : "<>",
+ (int)dp->dnetid);
+
+ CHANSET_ZERO(chanset);
+ nchan = idn_domain_is_registered(d, -1, &chanset);
+ if (dbuffer)
+ mask2str(chanset, dbp, 32);
+ else
+ (void) sprintf(dbp, "0x%x", chanset);
+ (void) mi_mpprintf(mp, " Nchans = %d, Chanset = %s",
+ nchan, nchan ? dbp : "<>");
+ pset_upper = UPPER32_CPUMASK(dp->dcpuset);
+ pset_lower = LOWER32_CPUMASK(dp->dcpuset);
+ if (dbuffer)
+ boardset2str(dp->dhw.dh_boardset, dbp);
+ else
+ (void) sprintf(dbp, "0x%x", dp->dhw.dh_boardset);
+
+ (void) mi_mpprintf(mp, " Nboards = %d, Brdset = %s",
+ dp->dhw.dh_nboards,
+ dp->dhw.dh_nboards ? dbp : "<>");
+ (void) sprintf(dbp, "0x%x.%x", pset_upper, pset_lower);
+ (void) mi_mpprintf(mp, " Ncpus = %d, Cpuset = %s",
+ dp->dncpus, dp->dncpus ? dbp : "<>");
+ (void) mi_mpprintf(mp, " Nmcadr = %d",
+ dp->dhw.dh_nmcadr);
+ (void) mi_mpprintf(mp,
+ " MsgTimer = %s (cnt = %d)",
+ (dp->dtimerq.tq_count > 0)
+ ? "active" : "idle",
+ dp->dtimerq.tq_count);
+ (void) mi_mpprintf(mp, " Dcpu = %d "
+ "(lastcpu = %d, cpuindex = %d)",
+ dp->dcpu, dp->dcpu_last, dp->dcpuindex);
+ (void) mi_mpprintf(mp, " Dio = %d "
+ "(ioerr = %d, iochk = %d, iowanted = %d)",
+ dp->dio, dp->dioerr, dp->diocheck ? 1 : 0,
+ dp->diowanted ? 1 : 0);
+ if (dp->dsync.s_cmd == IDNSYNC_NIL) {
+ (void) mi_mpprintf(mp, " Dsync = %s",
+ idnsync_str[IDNSYNC_NIL]);
+ } else {
+ (void) mi_mpprintf(mp,
+ " Dsync = %s "
+ "(x_set = 0x%x, r_set = 0x%x)",
+ idnsync_str[dp->dsync.s_cmd],
+ (uint_t)dp->dsync.s_set_exp,
+ (uint_t)dp->dsync.s_set_rdy);
+ }
+ (void) mi_mpprintf(mp, " Dvote = 0x%x",
+ dp->dvote.ticket);
+ (void) mi_mpprintf(mp, " Dfin = %s (Sync = %s)",
+ idnfin_str[dp->dfin],
+ (dp->dfin_sync == IDNFIN_SYNC_OFF) ? "OFF" :
+ (dp->dfin_sync == IDNFIN_SYNC_YES) ? "YES" :
+ "NO");
+ (void) mi_mpprintf(mp, " Dcookie_err = %s (cnt = %d)",
+ dp->dcookie_err ? "YES" : "NO",
+ dp->dcookie_errcnt);
+ IDN_DUNLOCK(d);
+ }
+
+ IDN_GUNLOCK();
+
+ if (dbuffer) {
+ FREE_DISPSTRING(dbuffer);
+ }
+
+ IDN_SYNC_UNLOCK();
+
+ return (0);
+}
+
+#define SNOOP_ENTRIES 2048 /* power of 2 */
+
+struct snoop_buffer {
+/* 0 */ char io;
+/* 1 */ char board;
+/* 2 */ char trans[14];
+
+/* 10 */ uint_t xargs[4];
+} *snoop_data, snoop_buffer[SNOOP_ENTRIES+1];
+
+
+int snoop_index;
+kmutex_t snoop_mutex;
+static char _bd2hexascii[] = {
+ '0', '1', '2', '3', '4', '5', '6', '7',
+ '8', '9', 'A', 'B', 'C', 'D', 'E', 'F'
+};
+
+#define SNOOP_IDN(in, tr, bd, arg1, arg2, arg3, arg4) \
+{ \
+ if (idn_snoop) { \
+ mutex_enter(&snoop_mutex); \
+ if (snoop_data == NULL) { \
+ snoop_data = (struct snoop_buffer *) \
+ (((uint_t)snoop_buffer + 0xf) & ~0xf); \
+ } \
+ snoop_data[snoop_index].io = ((in) == 0) ? 'o' : 'i'; \
+ snoop_data[snoop_index].board = \
+ ((bd) == -1) ? 'X' : _bd2hexascii[bd]; \
+ (void) strncpy(snoop_data[snoop_index].trans, (tr), 14); \
+ snoop_data[snoop_index].xargs[0] = (arg1); \
+ snoop_data[snoop_index].xargs[1] = (arg2); \
+ snoop_data[snoop_index].xargs[2] = (arg3); \
+ snoop_data[snoop_index].xargs[3] = (arg4); \
+ snoop_index++; \
+ snoop_index &= SNOOP_ENTRIES - 1; \
+ mutex_exit(&snoop_mutex); \
+ } \
+}
+
+/*
+ * Allocate the circular buffers to be used for
+ * DMV interrupt processing.
+ */
+static int
+idn_init_handler()
+{
+ int i, c;
+ size_t len;
+ idn_dmv_msg_t *basep, *ivp;
+ uint32_t ivp_offset;
+ procname_t proc = "idn_init_handler";
+
+ if (idn.intr.dmv_data != NULL) {
+ cmn_err(CE_WARN,
+ "IDN: 130: IDN DMV handler already initialized");
+ return (-1);
+ }
+
+ /*
+ * This memory will be touched by the low-level
+ * DMV trap handler for IDN.
+ */
+ len = sizeof (idn_dmv_data_t);
+ len = roundup(len, sizeof (uint64_t));
+ len += NCPU * idn_dmv_pending_max * sizeof (idn_dmv_msg_t);
+ len = roundup(len, PAGESIZE);
+
+ PR_PROTO("%s: sizeof (idn_dmv_data_t) = %d\n",
+ proc, sizeof (idn_dmv_data_t));
+ PR_PROTO("%s: allocating %d bytes for dmv data area\n", proc, len);
+
+ idn.intr.dmv_data_len = len;
+ idn.intr.dmv_data = kmem_zalloc(len, KM_SLEEP);
+
+ PR_PROTO("%s: DMV data area = %p\n", proc, idn.intr.dmv_data);
+
+ idn_dmv_data = (idn_dmv_data_t *)idn.intr.dmv_data;
+ basep = (idn_dmv_msg_t *)roundup((size_t)idn.intr.dmv_data +
+ sizeof (idn_dmv_data_t),
+ sizeof (uint64_t));
+ idn_dmv_data->idn_dmv_qbase = (uint64_t)basep;
+
+ ivp = basep;
+ ivp_offset = 0;
+ /*
+ * The buffer queues are allocated per-cpu.
+ */
+ for (c = 0, ivp = basep; c < NCPU; ivp++, c++) {
+ idn_dmv_data->idn_dmv_cpu[c].idn_dmv_current = ivp_offset;
+ idn_iv_queue[c] = ivp;
+ ivp_offset += sizeof (idn_dmv_msg_t);
+ for (i = 1; i < idn_dmv_pending_max; ivp++, i++) {
+ ivp->iv_next = ivp_offset;
+ ivp->iv_ready = 0;
+ lock_set(&ivp->iv_ready);
+ ivp_offset += sizeof (idn_dmv_msg_t);
+ }
+ ivp->iv_next = idn_dmv_data->idn_dmv_cpu[c].idn_dmv_current;
+ ivp->iv_ready = 0;
+ lock_set(&ivp->iv_ready);
+ }
+
+ idn.intr.dmv_inum = STARFIRE_DMV_IDN_BASE;
+ idn.intr.soft_inum = add_softintr((uint_t)idn_pil, idn_handler, 0);
+ idn_dmv_data->idn_soft_inum = idn.intr.soft_inum;
+ /*
+ * Make sure everything is out there before
+ * we effectively set it free for use.
+ */
+ membar_stld_stst();
+
+ if (dmv_add_intr(idn.intr.dmv_inum, idn_dmv_handler,
+ (caddr_t)idn_dmv_data)) {
+ idn_deinit_handler();
+ cmn_err(CE_WARN, "IDN: 132: failed to add IDN DMV handler");
+ return (-1);
+ }
+
+ return (0);
+}
+
+static void
+idn_deinit_handler()
+{
+ if (idn.intr.dmv_data == NULL)
+ return;
+
+ (void) dmv_rem_intr(idn.intr.dmv_inum);
+ rem_softintr(idn.intr.soft_inum);
+ kmem_free(idn.intr.dmv_data, idn.intr.dmv_data_len);
+ idn.intr.dmv_data = NULL;
+}
+
+/*
+ * High-level (soft interrupt) handler for DMV interrupts
+ */
+/*ARGSUSED0*/
+static uint_t
+idn_handler(caddr_t unused, caddr_t unused2)
+{
+#ifdef DEBUG
+ int count = 0;
+#endif /* DEBUG */
+ int cpuid = (int)CPU->cpu_id;
+ ushort_t mtype, atype;
+ idn_dmv_msg_t *xp, *xplimit;
+ procname_t proc = "idn_handler";
+
+ ASSERT(getpil() >= idn_pil);
+ flush_windows();
+
+ /*
+ * Clear the synchronization flag to indicate that
+ * processing has started. As long as idn_dmv_active
+ * is non-zero, idn_dmv_handler will queue work without
+ * initiating a soft interrupt. Since we clear it
+ * first thing at most one pil-interrupt for IDN will
+ * queue up behind the currently active one. We don't
+ * want to clear this flag at the end because it leaves
+ * a window where an interrupt could get lost (unless it's
+ * pushed by a subsequent interrupt). The objective in
+ * doing this is to prevent exhausting a cpu's intr_req
+ * structures with interrupts of the same pil level.
+ */
+ lock_clear(&idn_dmv_data->idn_dmv_cpu[cpuid].idn_dmv_active);
+
+ xp = idn_iv_queue[cpuid];
+ xplimit = xp + idn_dmv_pending_max;
+ xp += idn_intr_index[cpuid];
+ /*
+ * As long as there's stuff that's READY in the
+ * queue, keep processing.
+ */
+ while (lock_try(&xp->iv_ready)) {
+
+ ASSERT(lock_try(&xp->iv_inuse) == 0);
+
+ mtype = (ushort_t)xp->iv_mtype;
+ mtype &= IDNP_MSGTYPE_MASK | IDNP_ACKNACK_MASK;
+ atype = (ushort_t)xp->iv_atype;
+
+ if (((int)xp->iv_ver == idn.version) && mtype) {
+ idn_protojob_t *jp;
+#ifdef DEBUG
+ STRING(mstr);
+ STRING(astr);
+
+ INUM2STR(mtype, mstr);
+ if ((mtype & IDNP_MSGTYPE_MASK) == 0) {
+ INUM2STR(atype, astr);
+ strcat(mstr, "/");
+ strcat(mstr, astr);
+ }
+
+ count++;
+
+ PR_XDC("%s:%d:%d RECV: scpu = %d, msg = 0x%x(%s)\n",
+ proc, (int)xp->iv_domid, count,
+ (int)xp->iv_cpuid, mtype, mstr);
+ PR_XDC("%s:%d:%d R-DATA: a0 = 0x%x, a1 = 0x%x\n",
+ proc, (int)xp->iv_domid, count,
+ xp->iv_xargs0, xp->iv_xargs1);
+ PR_XDC("%s:%d:%d R-DATA: a2 = 0x%x, a3 = 0x%x\n",
+ proc, (int)xp->iv_domid, count,
+ xp->iv_xargs2, xp->iv_xargs3);
+#endif /* DEBUG */
+
+ if (mtype == IDNP_DATA) {
+ jp = NULL;
+ /*
+ * The only time we receive pure
+ * data messages at this level is
+ * to wake up the channel server.
+ * Since this is often an urgent
+ * request we'll do it from here
+ * instead of waiting for a proto
+ * server to do it.
+ */
+ idn_signal_data_server((int)xp->iv_domid,
+ (ushort_t)xp->iv_xargs0);
+ } else {
+ jp = idn_protojob_alloc(KM_NOSLEEP);
+ /*
+ * If the allocation fails, just drop
+ * the message and get on with life.
+ * If memory pressure is this great then
+ * dropping this message is probably
+ * the least of our worries!
+ */
+ if (jp) {
+ jp->j_msg.m_domid = (int)xp->iv_domid;
+ jp->j_msg.m_cpuid = (int)xp->iv_cpuid;
+ jp->j_msg.m_msgtype = mtype;
+ jp->j_msg.m_acktype = atype;
+ jp->j_msg.m_cookie = xp->iv_cookie;
+ SET_XARGS(jp->j_msg.m_xargs,
+ xp->iv_xargs0, xp->iv_xargs1,
+ xp->iv_xargs2, xp->iv_xargs3);
+ }
+
+ }
+ membar_ldst_stst();
+
+ lock_clear(&xp->iv_inuse);
+
+ if (jp)
+ idn_protojob_submit(jp->j_msg.m_domid, jp);
+ } else {
+ membar_ldst_stst();
+ IDN_GKSTAT_INC(gk_dropped_intrs);
+ lock_clear(&xp->iv_inuse);
+ }
+
+ if (++xp == xplimit)
+ xp = idn_iv_queue[cpuid];
+ }
+
+ idn_intr_index[cpuid] = xp - idn_iv_queue[cpuid];
+
+ return (DDI_INTR_CLAIMED);
+}
+
+void
+idn_awol_event_set(boardset_t boardset)
+{
+ idnsb_event_t *sbp;
+ procname_t proc = "idn_awol_event_set";
+
+ ASSERT(IDN_GLOCK_IS_EXCL());
+
+ mutex_enter(&idn.idnsb_mutex);
+ sbp = idn.idnsb_eventp;
+ if (sbp == NULL) {
+ cmn_err(CE_WARN, "IDN: 133: sigblock event area missing");
+ cmn_err(CE_CONT,
+ "IDN: 134: unable to mark boardset (0x%x) AWOL\n",
+ boardset);
+ mutex_exit(&idn.idnsb_mutex);
+ return;
+ }
+
+ if (boardset == 0) {
+ PR_PROTO("%s: AWOL BOARDSET is 0, NO EVENT <<<<<<<<<<<<<<<\n",
+ proc);
+ mutex_exit(&idn.idnsb_mutex);
+ return;
+ } else {
+ PR_PROTO("%s: MARKING BOARDSET (0x%x) AWOL\n", proc, boardset);
+ }
+ SSIEVENT_ADD(sbp, SSIEVENT_AWOL, boardset);
+ mutex_exit(&idn.idnsb_mutex);
+}
+
+void
+idn_awol_event_clear(boardset_t boardset)
+{
+ idnsb_event_t *sbp;
+ procname_t proc = "idn_awol_event_clear";
+
+ ASSERT(IDN_GLOCK_IS_EXCL());
+
+ mutex_enter(&idn.idnsb_mutex);
+ sbp = idn.idnsb_eventp;
+ if (sbp == NULL) {
+ cmn_err(CE_WARN, "IDN: 133: sigblock event area missing");
+ cmn_err(CE_CONT,
+ "IDN: 134: unable to mark boardset (0x%x) AWOL\n",
+ boardset);
+ mutex_exit(&idn.idnsb_mutex);
+ return;
+ }
+
+ if (boardset == 0) {
+ PR_PROTO("%s: AWOL BOARDSET is 0, NO EVENT <<<<<<<<<<<<<<<\n",
+ proc);
+ mutex_exit(&idn.idnsb_mutex);
+ return;
+ } else {
+ PR_PROTO("%s: CLEARING BOARDSET (0x%x) AWOL\n", proc, boardset);
+ }
+ SSIEVENT_DEL(sbp, SSIEVENT_AWOL, boardset);
+ mutex_exit(&idn.idnsb_mutex);
+}
+
+static void
+idn_gkstat_init()
+{
+ struct kstat *ksp;
+ struct idn_gkstat_named *sgkp;
+
+#ifdef kstat
+ if ((ksp = kstat_create(IDNNAME, ddi_get_instance(idn.dip),
+ IDNNAME, "net", KSTAT_TYPE_NAMED,
+ sizeof (struct idn_gkstat_named) / sizeof (kstat_named_t),
+ KSTAT_FLAG_PERSISTENT)) == NULL) {
+#else
+ if ((ksp = kstat_create(IDNNAME, ddi_get_instance(idn.dip),
+ IDNNAME, "net", KSTAT_TYPE_NAMED,
+ sizeof (struct idn_gkstat_named) /
+ sizeof (kstat_named_t), 0)) == NULL) {
+#endif /* kstat */
+ cmn_err(CE_CONT, "IDN: 135: %s: %s\n",
+ IDNNAME, "kstat_create failed");
+ return;
+ }
+
+ idn.ksp = ksp;
+ sgkp = (struct idn_gkstat_named *)(ksp->ks_data);
+ kstat_named_init(&sgkp->sk_curtime, "curtime",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_reconfigs, "reconfigs",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_reconfig_last, "reconfig_last",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_reaps, "reaps",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_reap_last, "reap_last",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_links, "links",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_link_last, "link_last",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_unlinks, "unlinks",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_unlink_last, "unlink_last",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_buffail, "buf_fail",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_buffail_last, "buf_fail_last",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_slabfail, "slab_fail",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_slabfail_last, "slab_fail_last",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_slabfail_last, "slab_fail_last",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_reap_count, "reap_count",
+ KSTAT_DATA_ULONG);
+ kstat_named_init(&sgkp->sk_dropped_intrs, "dropped_intrs",
+ KSTAT_DATA_ULONG);
+ ksp->ks_update = idn_gkstat_update;
+ ksp->ks_private = (void *)NULL;
+ kstat_install(ksp);
+}
+
+static void
+idn_gkstat_deinit()
+{
+ if (idn.ksp)
+ kstat_delete(idn.ksp);
+ idn.ksp = NULL;
+}
+
+static int
+idn_gkstat_update(kstat_t *ksp, int rw)
+{
+ struct idn_gkstat_named *sgkp;
+
+ sgkp = (struct idn_gkstat_named *)ksp->ks_data;
+
+ if (rw == KSTAT_WRITE) {
+ sg_kstat.gk_reconfigs = sgkp->sk_reconfigs.value.ul;
+ sg_kstat.gk_reconfig_last = sgkp->sk_reconfig_last.value.ul;
+ sg_kstat.gk_reaps = sgkp->sk_reaps.value.ul;
+ sg_kstat.gk_reap_last = sgkp->sk_reap_last.value.ul;
+ sg_kstat.gk_links = sgkp->sk_links.value.ul;
+ sg_kstat.gk_link_last = sgkp->sk_link_last.value.ul;
+ sg_kstat.gk_unlinks = sgkp->sk_unlinks.value.ul;
+ sg_kstat.gk_unlink_last = sgkp->sk_unlink_last.value.ul;
+ sg_kstat.gk_buffail = sgkp->sk_buffail.value.ul;
+ sg_kstat.gk_buffail_last = sgkp->sk_buffail_last.value.ul;
+ sg_kstat.gk_slabfail = sgkp->sk_slabfail.value.ul;
+ sg_kstat.gk_slabfail_last = sgkp->sk_slabfail_last.value.ul;
+ sg_kstat.gk_reap_count = sgkp->sk_reap_count.value.ul;
+ sg_kstat.gk_dropped_intrs = sgkp->sk_dropped_intrs.value.ul;
+ } else {
+ sgkp->sk_curtime.value.ul = lbolt;
+ sgkp->sk_reconfigs.value.ul = sg_kstat.gk_reconfigs;
+ sgkp->sk_reconfig_last.value.ul = sg_kstat.gk_reconfig_last;
+ sgkp->sk_reaps.value.ul = sg_kstat.gk_reaps;
+ sgkp->sk_reap_last.value.ul = sg_kstat.gk_reap_last;
+ sgkp->sk_links.value.ul = sg_kstat.gk_links;
+ sgkp->sk_link_last.value.ul = sg_kstat.gk_link_last;
+ sgkp->sk_unlinks.value.ul = sg_kstat.gk_unlinks;
+ sgkp->sk_unlink_last.value.ul = sg_kstat.gk_unlink_last;
+ sgkp->sk_buffail.value.ul = sg_kstat.gk_buffail;
+ sgkp->sk_buffail_last.value.ul = sg_kstat.gk_buffail_last;
+ sgkp->sk_slabfail.value.ul = sg_kstat.gk_slabfail;
+ sgkp->sk_slabfail_last.value.ul = sg_kstat.gk_slabfail_last;
+ sgkp->sk_reap_count.value.ul = sg_kstat.gk_reap_count;
+ sgkp->sk_dropped_intrs.value.ul = sg_kstat.gk_dropped_intrs;
+ }
+
+ return (0);
+}
+
+#ifdef DEBUG
+#define RW_HISTORY 100
+static uint_t rw_history[NCPU][RW_HISTORY];
+static int rw_index[NCPU];
+#endif /* DEBUG */
+
+static int
+idn_rw_mem(idnop_t *idnop)
+{
+ uint_t lo_off, hi_off;
+ int rw, blksize, num;
+ int cpuid;
+ register int n, idx;
+ char *ibuf, *obuf;
+ char *smraddr;
+ struct seg *segp;
+ ulong_t randx;
+ kmutex_t slock;
+ kcondvar_t scv;
+ static int orig_gstate = IDNGS_IGNORE;
+ extern struct seg ktextseg;
+
+#define RANDOM_INIT() (randx = lbolt)
+#define RANDOM(a, b) \
+ (((a) >= (b)) ? \
+ (a) : (((randx = randx * 1103515245L + 12345) % ((b)-(a))) + (a)))
+
+ RANDOM_INIT();
+
+ lo_off = idnop->rwmem.lo_off;
+ hi_off = idnop->rwmem.hi_off;
+ blksize = idnop->rwmem.blksize;
+ num = idnop->rwmem.num;
+ rw = idnop->rwmem.rw; /* 0 = rd, 1 = wr, 2 = rd/wr */
+
+ if (((hi_off > (uint_t)MB2B(IDN_SMR_SIZE)) || (lo_off >= hi_off) ||
+ (blksize <= 0) || (blksize > (hi_off - lo_off)) || (num <= 0)) &&
+ (idnop->rwmem.goawol == -1)) {
+ return (EINVAL);
+ }
+
+ if (idnop->rwmem.goawol && (orig_gstate == IDNGS_IGNORE)) {
+ IDN_GLOCK_EXCL();
+ cmn_err(CE_WARN, "IDN: Local domain going into IGNORE MODE!!");
+ orig_gstate = idn.state;
+ IDN_GSTATE_TRANSITION(IDNGS_IGNORE);
+ IDN_GUNLOCK();
+
+ } else if (!idnop->rwmem.goawol && (orig_gstate != IDNGS_IGNORE)) {
+ IDN_GLOCK_EXCL();
+ cmn_err(CE_WARN,
+ "IDN: Local domain restoring original state %s(%d)",
+ idngs_str[orig_gstate], (int)orig_gstate);
+ IDN_GSTATE_TRANSITION(orig_gstate);
+ orig_gstate = IDNGS_IGNORE;
+ IDN_GUNLOCK();
+ }
+ /*
+ * Just requested AWOL.
+ */
+ if (num == 0)
+ return (0);
+ /*
+ * Default READ only.
+ */
+ ibuf = (char *)kmem_alloc(blksize, KM_SLEEP);
+ if (rw == 1) {
+ /*
+ * WRITE only.
+ */
+ obuf = ibuf;
+ ibuf = NULL;
+ } else if (rw == 2) {
+ /*
+ * READ/WRITE.
+ */
+ obuf = (char *)kmem_alloc(blksize, KM_SLEEP);
+ for (segp = &ktextseg; segp; segp = AS_SEGNEXT(&kas, segp)) {
+ if (segp->s_size >= blksize)
+ break;
+ }
+ if (segp == NULL) {
+ cmn_err(CE_WARN,
+ "IDN: blksize (%d) too large", blksize);
+ return (EINVAL);
+ }
+ bcopy(segp->s_base, obuf, blksize);
+ }
+
+ mutex_init(&slock, NULL, MUTEX_DEFAULT, NULL);
+ cv_init(&scv, NULL, CV_DEFAULT, NULL);
+
+ cmn_err(CE_NOTE,
+ "IDN: starting %s of %d blocks of %d bytes each...",
+ (rw == 1) ? "W-ONLY" : (rw == 2) ? "RW" : "R-ONLY",
+ num, blksize);
+
+ for (n = 0; n < num; n++) {
+ uint_t rpos;
+
+ if ((hi_off - lo_off) > blksize)
+ rpos = RANDOM(lo_off, (hi_off - blksize));
+ else
+ rpos = lo_off;
+ smraddr = IDN_OFFSET2ADDR(rpos);
+
+ cpuid = (int)CPU->cpu_id;
+#ifdef DEBUG
+ idx = rw_index[cpuid]++ % RW_HISTORY;
+ rw_history[cpuid][idx] = rpos;
+#endif /* DEBUG */
+
+ switch (rw) {
+ case 0:
+ bcopy(smraddr, ibuf, blksize);
+ break;
+ case 1:
+ bcopy(obuf, smraddr, blksize);
+ break;
+ case 2:
+ if (n & 1)
+ bcopy(obuf, smraddr, blksize);
+ else
+ bcopy(smraddr, ibuf, blksize);
+ break;
+ default:
+ break;
+ }
+ if (!(n % 1000)) {
+ int rv;
+
+ mutex_enter(&slock);
+ rv = cv_timedwait_sig(&scv, &slock, lbolt+hz);
+ mutex_exit(&slock);
+ if (rv == 0)
+ break;
+ }
+ }
+
+ cv_destroy(&scv);
+ mutex_destroy(&slock);
+
+ if (ibuf)
+ kmem_free(ibuf, blksize);
+ if (obuf)
+ kmem_free(obuf, blksize);
+
+ return (0);
+}
+
+void
+inum2str(uint_t inum, char str[])
+{
+ uint_t acknack;
+
+ str[0] = '\0';
+
+ acknack = (inum & IDNP_ACKNACK_MASK);
+ inum &= ~IDNP_ACKNACK_MASK;
+
+ if (!inum && !acknack) {
+ strcpy(str, idnm_str[0]);
+ return;
+ }
+
+ if (inum == 0) {
+ strcpy(str, (acknack & IDNP_ACK) ? "ack" : "nack");
+ } else {
+ if (inum < IDN_NUM_MSGTYPES)
+ strcpy(str, idnm_str[inum]);
+ else
+ sprintf(str, "0x%x?", inum);
+ if (acknack) {
+ if (acknack & IDNP_ACK)
+ strcat(str, "+ack");
+ else
+ strcat(str, "+nack");
+ }
+ }
+}
+
+boardset_t
+cpuset2boardset(cpuset_t portset)
+{
+ register int c;
+ register boardset_t bset;
+
+ bset = 0;
+ for (c = 0; c < NCPU; )
+ if (CPU_IN_SET(portset, c)) {
+ BOARDSET_ADD(bset, CPUID_TO_BOARDID(c));
+ c = (c + 4) & ~3;
+ } else {
+ c++;
+ }
+
+ return (bset);
+}
+
+void
+cpuset2str(cpuset_t cset, char buffer[])
+{
+ register int c, n;
+
+ buffer[0] = '\0';
+ for (c = n = 0; c < NCPU; c++) {
+ if (!CPU_IN_SET(cset, c))
+ continue;
+#ifdef DEBUG
+ if (strlen(buffer) >= _DSTRLEN) {
+ PR_PROTO("************* WARNING WARNING WARNING\n");
+ PR_PROTO("cpuset2str(cpu = %d) buffer "
+ "OVERFLOW <<<<<<\n", c);
+ PR_PROTO("*******************************\n");
+ (void) sprintf(&buffer[_DSTRLEN-6], "*OVER\0");
+ return;
+ }
+#endif /* DEBUG */
+ if (n == 0)
+ (void) sprintf(buffer, "%d\0", c);
+ else
+ (void) sprintf(buffer, "%s, %d\0", buffer, c);
+ n++;
+ }
+}
+
+void
+domainset2str(domainset_t dset, char buffer[])
+{
+ /*
+ * Since domainset_t and boardset_t are the
+ * same (max = MAX_DOMAINS = MAX_BOARDS) we
+ * can just overload boardset2str().
+ */
+ mask2str((uint_t)dset, buffer, MAX_DOMAINS);
+}
+
+void
+boardset2str(boardset_t bset, char buffer[])
+{
+ mask2str((uint_t)bset, buffer, MAX_BOARDS);
+}
+
+void
+mask2str(uint_t mask, char buffer[], int maxnum)
+{
+ int n, i;
+
+ buffer[0] = '\0';
+ for (i = n = 0; i < maxnum; i++) {
+ if ((mask & (1 << i)) == 0)
+ continue;
+ if (n == 0)
+ (void) sprintf(buffer, "%d\0", i);
+ else
+ (void) sprintf(buffer, "%s, %d\0", buffer, i);
+ n++;
+ }
+}
+
+int
+idnxdc(int domid, idn_msgtype_t *mtp,
+ uint_t arg1, uint_t arg2,
+ uint_t arg3, uint_t arg4)
+{
+ int rv, cpuid, tcpuid;
+ uint_t cookie;
+ uint64_t pdata;
+ uint64_t dmv_word0, dmv_word1, dmv_word2;
+ idn_domain_t *dp = &idn_domain[domid];
+ extern kmutex_t xc_sys_mutex;
+ extern int xc_spl_enter[];
+ procname_t proc = "idnxdc";
+
+
+ if (idn_snoop) {
+ int bd;
+ STRING(str);
+ STRING(mstr);
+ STRING(astr);
+
+ INUM2STR(mtp->mt_mtype, mstr);
+ if ((mtp->mt_mtype & IDNP_MSGTYPE_MASK) == 0) {
+ INUM2STR(arg1, astr);
+ sprintf(str, "%s/%s", mstr, astr);
+ } else {
+ strcpy(str, mstr);
+ }
+ if (dp->dcpu == IDN_NIL_DCPU)
+ bd = -1;
+ else
+ bd = CPUID_TO_BOARDID(dp->dcpu);
+ SNOOP_IDN(0, str, bd, arg1, arg2, arg3, arg4);
+ }
+
+ /*
+ * For NEGO messages we send the remote domain the cookie we
+ * expect it to use in subsequent messages that it sends
+ * to us (dcookie_recv).
+ * For other messages, we must use the cookie that the
+ * remote domain assigned to us for sending (dcookie_send).
+ */
+ if ((mtp->mt_mtype & IDNP_MSGTYPE_MASK) == IDNP_NEGO)
+ cookie = IDN_MAKE_COOKIE(dp->dcookie_recv, mtp->mt_cookie);
+ else
+ cookie = IDN_MAKE_COOKIE(dp->dcookie_send, mtp->mt_cookie);
+
+ pdata = IDN_MAKE_PDATA(mtp->mt_mtype, mtp->mt_atype, cookie);
+
+ dmv_word0 = DMV_MAKE_DMV(idn.intr.dmv_inum, pdata);
+ dmv_word1 = ((uint64_t)arg1 << 32) | (uint64_t)arg2;
+ dmv_word2 = ((uint64_t)arg3 << 32) | (uint64_t)arg4;
+
+ ASSERT((dp->dcpu != IDN_NIL_DCPU) ||
+ (dp->dcpu_last != IDN_NIL_DCPU));
+
+ tcpuid = (dp->dcpu == IDN_NIL_DCPU) ?
+ dp->dcpu_last : dp->dcpu;
+
+ if (tcpuid == IDN_NIL_DCPU) {
+ PR_PROTO("%s:%d: cpu/cpu_last == NIL_DCPU\n",
+ proc, domid);
+ return (-1);
+ }
+
+ mutex_enter(&xc_sys_mutex);
+ cpuid = (int)CPU->cpu_id;
+ xc_spl_enter[cpuid] = 1;
+
+ idnxf_init_mondo(dmv_word0, dmv_word1, dmv_word2);
+
+ rv = idnxf_send_mondo(STARFIRE_UPAID2HWMID(tcpuid));
+
+ xc_spl_enter[cpuid] = 0;
+ mutex_exit(&xc_sys_mutex);
+
+ return (rv);
+}
+
+void
+idnxdc_broadcast(domainset_t domset, idn_msgtype_t *mtp,
+ uint_t arg1, uint_t arg2,
+ uint_t arg3, uint_t arg4)
+{
+ int d;
+
+ for (d = 0; d < MAX_DOMAINS; d++) {
+ idn_domain_t *dp;
+
+ if (!DOMAIN_IN_SET(domset, d))
+ continue;
+
+ dp = &idn_domain[d];
+ if (dp->dcpu == IDN_NIL_DCPU)
+ continue;
+
+ (void) IDNXDC(d, mtp, arg1, arg2, arg3, arg4);
+ }
+}
+
+#define PROM_SMRSIZE 0x1
+#define PROM_SMRADDR 0x2
+#define PROM_SMRPROPS (PROM_SMRSIZE | PROM_SMRADDR)
+/*
+ * Locate the idn-smr-size property to determine the size of the SMR
+ * region for the SSI. Value inherently enables/disables SSI capability.
+ */
+static int
+idn_prom_getsmr(uint_t *smrsz, uint64_t *paddrp, uint64_t *sizep)
+{
+ dnode_t nodeid;
+ int found = 0;
+ int len;
+ uint_t smrsize = 0;
+ uint64_t obpaddr, obpsize;
+ struct smraddr {
+ uint32_t hi_addr;
+ uint32_t lo_addr;
+ uint32_t hi_size;
+ uint32_t lo_size;
+ } smraddr;
+ procname_t proc = "idn_prom_getsmr";
+
+ bzero(&smraddr, sizeof (smraddr));
+ /*
+ * idn-smr-size is a property of the "memory" node and
+ * is defined in megabytes.
+ */
+ nodeid = prom_finddevice("/memory");
+
+ if (nodeid != OBP_NONODE) {
+ len = prom_getproplen(nodeid, IDN_PROP_SMRSIZE);
+ if (len == sizeof (smrsize)) {
+ (void) prom_getprop(nodeid, IDN_PROP_SMRSIZE,
+ (caddr_t)&smrsize);
+ found |= PROM_SMRSIZE;
+ }
+ len = prom_getproplen(nodeid, IDN_PROP_SMRADDR);
+ if (len == sizeof (smraddr)) {
+ (void) prom_getprop(nodeid, IDN_PROP_SMRADDR,
+ (caddr_t)&smraddr);
+ found |= PROM_SMRADDR;
+ }
+ }
+
+ if (found != PROM_SMRPROPS) {
+ if ((found & PROM_SMRSIZE) == 0)
+ cmn_err(CE_WARN,
+ "IDN: 136: \"%s\" property not found, "
+ "disabling IDN",
+ IDN_PROP_SMRSIZE);
+ if (smrsize && ((found & PROM_SMRADDR) == 0))
+ cmn_err(CE_WARN,
+ "IDN: 136: \"%s\" property not found, "
+ "disabling IDN",
+ IDN_PROP_SMRADDR);
+ return (-1);
+ }
+
+ if (smrsize == 0) {
+ PR_SMR("%s: IDN DISABLED (idn_smr_size = 0)\n", proc);
+ cmn_err(CE_NOTE, "!IDN: 137: SMR size is 0, disabling IDN");
+
+ } else if (smrsize > IDN_SMR_MAXSIZE) {
+ PR_SMR("%s: IDN DISABLED (idn_smr_size too big %d > %d MB)\n",
+ proc, smrsize, IDN_SMR_MAXSIZE);
+ cmn_err(CE_WARN,
+ "!IDN: 138: SMR size (%dMB) is too big (max = %dMB), "
+ "disabling IDN",
+ smrsize, IDN_SMR_MAXSIZE);
+ smrsize = 0;
+ } else {
+ *smrsz = smrsize;
+ found &= ~PROM_SMRSIZE;
+ }
+
+ obpaddr = ((uint64_t)smraddr.hi_addr << 32) |
+ (uint64_t)smraddr.lo_addr;
+ obpsize = ((uint64_t)smraddr.hi_size << 32) |
+ (uint64_t)smraddr.lo_size;
+
+ if (obpsize == 0) {
+ if (smrsize > 0) {
+ cmn_err(CE_WARN, "!IDN: 139: OBP region for "
+ "SMR is 0 length");
+ }
+ } else if (obpsize < (uint64_t)MB2B(smrsize)) {
+ cmn_err(CE_WARN,
+ "!IDN: 140: OBP region (%lld B) smaller "
+ "than requested size (%ld B)",
+ obpsize, MB2B(smrsize));
+ } else if ((obpaddr & ((uint64_t)IDN_SMR_ALIGN - 1)) != 0) {
+ cmn_err(CE_WARN,
+ "!IDN: 141: OBP region (0x%llx) not on (0x%lx) "
+ "boundary", obpaddr, IDN_SMR_ALIGN);
+ } else {
+ *sizep = obpsize;
+ *paddrp = obpaddr;
+ found &= ~PROM_SMRADDR;
+ }
+
+ return (found ? -1 : 0);
+}
+
+void
+idn_init_autolink()
+{
+ idnsb_event_t *sbp;
+ procname_t proc = "idn_init_autolink";
+
+ mutex_enter(&idn.idnsb_mutex);
+ if ((sbp = idn.idnsb_eventp) == NULL) {
+ PR_PROTO("%s: IDN private sigb (event) area is NULL\n", proc);
+ mutex_exit(&idn.idnsb_mutex);
+ return;
+ }
+
+ PR_PROTO("%s: marking domain IDN ready.\n", proc);
+
+ bzero(sbp, sizeof (*sbp));
+
+ sbp->idn_version = (uchar_t)idn.version;
+ SSIEVENT_SET(sbp, SSIEVENT_BOOT, 0);
+ (void) strncpy(sbp->idn_cookie_str, SSIEVENT_COOKIE,
+ SSIEVENT_COOKIE_LEN);
+ mutex_exit(&idn.idnsb_mutex);
+}
+
+void
+idn_deinit_autolink()
+{
+ idnsb_event_t *sbp;
+ procname_t proc = "idn_deinit_autolink";
+
+ mutex_enter(&idn.idnsb_mutex);
+ if ((sbp = idn.idnsb_eventp) == NULL) {
+ PR_PROTO("%s: IDN private sigb (event) area is NULL\n", proc);
+ mutex_exit(&idn.idnsb_mutex);
+ return;
+ }
+
+ PR_PROTO("%s: marking domain IDN unavailable.\n", proc);
+
+ sbp->idn_version = (uchar_t)idn.version;
+ SSIEVENT_CLEAR(sbp, SSIEVENT_BOOT, 0);
+ (void) strncpy(sbp->idn_cookie_str, SSIEVENT_COOKIE,
+ SSIEVENT_COOKIE_LEN);
+ mutex_exit(&idn.idnsb_mutex);
+}
+
+void
+_make64cpumask(cpuset_t *csetp, uint_t upper, uint_t lower)
+{
+ int c;
+
+ CPUSET_ZERO(*csetp);
+
+ for (c = 0; c < 32; c++) {
+ if (lower & (1 << c)) {
+ CPUSET_ADD(*csetp, c);
+ }
+ if (upper & (1 << (c + 32))) {
+ CPUSET_ADD(*csetp, c + 32);
+ }
+ }
+}
+
+uint_t
+_lower32cpumask(cpuset_t cset)
+{
+ int c;
+ uint_t set = 0;
+
+ for (c = 0; c < 32; c++)
+ if (CPU_IN_SET(cset, c))
+ set |= 1 << c;
+
+ return (set);
+}
+
+uint_t
+_upper32cpumask(cpuset_t cset)
+{
+ int c;
+ uint_t set = 0;
+
+ for (c = 32; c < NCPU; c++)
+ if (CPU_IN_SET(cset, c))
+ set |= 1 << (c - 32);
+
+ return (set);
+}
+
+#ifdef DEBUG
+int
+debug_idnxdc(char *f, int domid, idn_msgtype_t *mtp,
+ uint_t a1, uint_t a2, uint_t a3, uint_t a4)
+{
+ idn_domain_t *dp = &idn_domain[domid];
+ int rv, cpuid, bd;
+ static int xx = 0;
+ STRING(str);
+ STRING(mstr);
+ STRING(astr);
+
+ xx++;
+ INUM2STR(mtp->mt_mtype, mstr);
+ if ((mtp->mt_mtype & IDNP_MSGTYPE_MASK) == 0) {
+ INUM2STR(a1, astr);
+ sprintf(str, "%s/%s", mstr, astr);
+ } else {
+ strcpy(str, mstr);
+ }
+
+ if ((cpuid = dp->dcpu) == IDN_NIL_DCPU)
+ bd = -1;
+ else
+ bd = CPUID_TO_BOARDID(cpuid);
+
+ SNOOP_IDN(0, str, bd, a1, a2, a3, a4);
+
+ PR_XDC("%s:%d:%d SENT: scpu = %d, msg = 0x%x(%s)\n",
+ f, domid, xx, cpuid, mtp->mt_mtype, str);
+ PR_XDC("%s:%d:%d S-DATA: a1 = 0x%x, a2 = 0x%x\n",
+ f, domid, xx, a1, a2);
+ PR_XDC("%s:%d:%d S-DATA: a3 = 0x%x, a4 = 0x%x\n",
+ f, domid, xx, a3, a4);
+
+ rv = idnxdc(domid, mtp, a1, a2, a3, a4);
+ if (rv != 0) {
+ PR_XDC("%s:%d:%d: WARNING: idnxdc(cpu %d) FAILED\n",
+ f, domid, xx, cpuid);
+ }
+
+ return (rv);
+}
+
+caddr_t
+_idn_getstruct(char *structname, int size)
+{
+ caddr_t ptr;
+ procname_t proc = "GETSTRUCT";
+
+ ptr = kmem_zalloc(size, KM_SLEEP);
+
+ PR_ALLOC("%s: ptr 0x%x, struct(%s), size = %d\n",
+ proc, (uint_t)ptr, structname, size);
+
+ return (ptr);
+}
+
+void
+_idn_freestruct(caddr_t ptr, char *structname, int size)
+{
+ procname_t proc = "FREESTRUCT";
+
+ PR_ALLOC("%s: ptr 0x%x, struct(%s), size = %d\n",
+ proc, (uint_t)ptr, structname, size);
+
+ ASSERT(ptr != NULL);
+ kmem_free(ptr, size);
+}
+#endif /* DEBUG */