summaryrefslogtreecommitdiff
path: root/src/pmdas/solaris/help
diff options
context:
space:
mode:
Diffstat (limited to 'src/pmdas/solaris/help')
-rw-r--r--src/pmdas/solaris/help729
1 files changed, 729 insertions, 0 deletions
diff --git a/src/pmdas/solaris/help b/src/pmdas/solaris/help
new file mode 100644
index 0000000..4940247
--- /dev/null
+++ b/src/pmdas/solaris/help
@@ -0,0 +1,729 @@
+#
+# Copyright (c) 2000-2004 Silicon Graphics, Inc. All Rights Reserved.
+#
+# This program is free software; you can redistribute it and/or modify it
+# under the terms of the GNU General Public License as published by the
+# Free Software Foundation; either version 2 of the License, or (at your
+# option) any later version.
+#
+# This program is distributed in the hope that it will be useful, but
+# WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
+# or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
+# for more details.
+#
+# You should have received a copy of the GNU General Public License along
+# with this program; if not, write to the Free Software Foundation, Inc.,
+# 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+#
+# Solaris PMDA help file in the ASCII format
+#
+# lines beginning with a # are ignored
+# lines beginning @ introduce a new entry of the form
+# @ metric_name oneline-text
+# help test goes
+# here over multiple lines
+# ...
+#
+# the metric_name is decoded against the default PMNS -- as a special case,
+# a name of the form NNN.MM (for numeric NNN and MM) is interpreted as an
+# instance domain identification, and the text describes the instance domain
+#
+# blank lines before the @ line are ignored
+#
+
+@ kernel.all.cpu.idle Amount of time CPUs were idle
+@ kernel.all.cpu.user Amount of time spent executing userspace tasks
+@ kernel.all.cpu.sys Amount of time spent executing kernel code
+@ kernel.all.cpu.wait.total Amount of time CPU spent waiting for events
+@ kernel.percpu.cpu.user Amount of time spent executing userspace tasks by each CPU
+@ kernel.percpu.cpu.idle Amount of time each CPU was idle
+@ kernel.percpu.cpu.sys Amount of time each CPU spent executing kernel code
+@ kernel.percpu.cpu.wait.total Amount of time each CPU spent waiting for events
+@ disk.all.read Number of read requests aggregated across all disks
+@ disk.all.write Number of write requests aggregated across all disks
+@ disk.all.total Number of IO requests aggregated across all disks
+@ disk.all.read_bytes Number of bytes read from all disks
+@ disk.all.write_bytes Number of bytes written to all disk
+@ disk.all.total_bytes Number of bytes transferred to and from all disks
+@ disk.dev.read Number of read requests for each individual disk
+@ disk.dev.write Number of write requests for each individual disk
+@ disk.dev.total Number of IO requests for each individual disk
+@ disk.dev.read_bytes Number of bytes read from each individual disk
+@ disk.dev.write_bytes Number of bytes written to each individual disk
+@ disk.dev.total_bytes Number of bytes transferred to and from each individual disk
+@ network.interface.mtu Maximum Transmission Unit of a network interface
+Maximum Transmision Unit is the largest size of IP datagram which can be
+transferred over the data link.
+@ network.interface.in.bytes Number of bytes received by a network interface
+@ network.interface.in.errors Number of receive errors per network interface
+Number of receive errors per network interface. The errors counted towards
+this metric are: IP header errors, packets larger then the link MTU, packets
+delivered to the unknown address, packets sent to the unknown IP protocol,
+truncated packets, packets discarded due to not having a route for
+the destination.
+@ network.interface.in.drops Number of packets droped by a network interface
+Number of packets discared due to lack of space during the input processing.
+@ network.interface.in.delivers Number of packets delivered to ULPs
+Number of packets delivered for further processing by the upper-layer
+protocols.
+@ network.interface.in.bcasts Number of broadcast packets received by a network interface
+@ network.interface.in.packets Number of IP packets received by a network interface
+@ network.interface.in.mcasts Number of multicast packets received by a network interface
+@ network.interface.out.packets Number of packets sent by a network interface
+@ network.interface.out.bytes Number of bytes sent by a network interface
+@ network.interface.out.errors Number of send errors per network interface
+@ network.interface.out.bcasts Number of broadcast packets sent by a network interface
+@ network.interface.out.mcasts Number of multicast packets sent by a network interface
+@ network.interface.out.drops Number of packets discared by a network interface
+Number of packets discared due to lack of space during the output processing.
+@ network.udp.ipackets Number of UDP packets received
+@ network.udp.opackets Number of UDP packets sent
+@ network.udp.ierrors Number of receive errors in UDP processing
+@ network.udp.oerrors Number of send erros in UDP processing
+@ network.udp.noports Number of UDP packets received on unknown UDP port
+Number of UDP packets received for which here is no port can be found.
+This counter is reported on the per-interface basis and aggregated by the PMDA.
+
+@ network.udp.overflows Number of UDP packets droped due to queue overflow
+Number of UDP packets droped due to queue overflow.
+This counter is reported on the per-interface basis and aggregated by the PMDA.
+
+@zpool.capacity Total capacity of a zpool in bytes
+@zpool.used Total space used on a pool
+@zpool.checksum_errors Number of checksum errors per zpool
+@zpool.self_healed Number of bytes healed
+@zpool.in.bytes Counter of bytes read from a zpool
+@zpool.out.bytes Counter of bytes written to a zpool
+@zpool.in.ops Counter of reads per zpool
+@zpool.out.ops Counter of writes per zpool
+@zpool.in.errors Counter of read errors per zpool
+@zpool.out.errors Counter of write errors per zpool
+@zpool.state Current state of zpool
+@zpool.state_int vs_aux << 8 | vs_state
+
+@zfs.available Amount of space available to the dataset
+The amount of space available to the dataset (a filesystem,
+a snapshot or a volume) and all its children. This is usually
+the amount of space available in the zpool which houses the
+dataset.
+
+@zfs.used.total Amount of space used by the dataset and its children
+The amount of space consumed by the filesystem, snapshot or
+volume and all its children. This amount does not include
+any reservation made by the dataset itself but do include
+the reservation of the children.
+
+@zfs.used.byme Amount of space used by the dataset itself.
+This amount exclude any space used by the children of this dataset
+or any of its snapshots.
+
+@zfs.used.bysnapshots Amount of space used by the snapshots of the dataset
+The amount of space consumed by the snapshots of this dataset.
+
+@zfs.used.bychildren Amount of space used by decendents of the dataset
+The amount of space consumed by all the decendants of this dataset.
+
+@zfs.quota Maximum amount of space a dataset can use
+Quotas are used to restrict the growth of the datasets. If
+the quota is set to 0 then the size of the dataset is limited only
+by the size of the pool which houses this dataset.
+
+@zfs.reservation Minimum amount of space guaranteed to a dataset
+The amount of space which dataset and its decendents are guaranteed
+to be available for them to use. This amount of taken off the quota
+of the parent of the dataset.
+
+@zfs.compression Compression ratio of the dataset
+Compression ratio is expressed as multiplier. To estimate how much data
+will be used by the uncompressed data multiply the amount of space used
+by the dataset by the compression ratio.
+
+@zfs.copies Number of redundant copies of data
+The number of redundant copies does not include any copies made as
+part of the pool redundancy.
+
+@zfs.recordsize Recommendend block size for files in filesystems
+By using recommended block size applications which deal with fixed size
+records can improve I/O performance.
+
+@zfs.used.byrefreservation Space used by refreservation
+The amount of space used by a refreservation set on this
+filesystem, which would be freed if the refreservation was
+removed.
+
+@zfs.refreservation Minimum amount of space guaranteed to a filesystem
+The minimum amount of space guaranteed to a dataset, not
+including its descendents. Unlike reservation refreservation is
+counted towards the total used space of a dataset.
+
+@zfs.refquota Amount of space a filesystem can consume
+The hard limit on the amount of space a filesystem but not its descendants
+can consume from the pool.
+
+@zfs.referenced Amount of space referenced by the filesystem
+The amount of data that is accessible by the filesystem. The data
+may be shared with other datasets in the pool.
+
+@zfs.nsnapshots Number of snapshots in the filesystem
+
+@zfs.snapshot.compression Compression ratio of the data in the snapshot
+Compression ratio is expressed as multiplier. To estimate how much data
+will be used by the uncompressed data multiply the amount of space used
+by the snapshot by the compression ratio.
+
+@zfs.snapshot.used Amount of space used by the snapshot
+
+@zfs.snapshot.referenced Amount of space referenced by the snapshot
+The amount of data that is accessible by the snapshot. The data
+may be shared with other datasets in the filesystem.
+
+
+@zpool.perdisk.state Current state per disk in zpool
+@zpool.perdisk.state_int vs_aux << 8 | vs_state
+@zpool.perdisk.checksum_errors Number of checksum errors per disk in zpool
+@zpool.perdisk.self_healed Number of bytes healed per disk in zpool
+@zpool.perdisk.in.errors Counter of read errors per disk in zpool
+@zpool.perdisk.out.errors Counter of write errors per disk in zpool
+
+@network.link.in.errors Number of input errors per link
+Counts input errors per link
+@network.link.in.packets Numbe of datagrams received by a link
+@network.link.in.bytes Number of bytes received by a link
+Counts number of bytes received by a link. For the physical links
+this is the raw counter of bytes received, for the aggregated links
+this is the number of bytes received by all links in the aggregation
+group
+@network.link.in.bcasts Number of broadcast datagrams received by a link
+@network.link.in.mcasts Number of multicast datagrams
+Counts multicast datagram recieved by a link.
+@network.link.in.nobufs Number of inpit packets discared
+Counts number of packets discared because of failure to allocate buffers
+@network.link.out.errors Number of output errors per link
+@network.link.out.packets Number of packets sent from a link
+@network.link.out.bytes Number of bytes sent from a link
+@network.link.out.bcasts Number of broadcast datagrams sent from a link
+@network.link.out.mcasts Number of multicast datagrams sent from a link
+@network.link.out.nobufs Number of output packets discared
+Counts number of packets discared because of failure to allocate buffers
+@network.link.collisions Number of collisions detected per link
+@network.link.state Link state
+1 - Link is up, 2 - Link is down, 0 - unknown state
+@network.link.duplex Link duplex
+1 - Half duplex, 2 - Full duplex
+@network.link.speed Link speed in bytes per second
+@hinv.pagesize Memory page size
+The memory page size of the running kernel in bytes.
+@hinv.physmem Total physical system memory
+Total physical system memory size rounded down to the nearest page size
+boundary
+@pmda.uname identity and type of current system
+Identity and type of current system. The concatenation of the values
+returned from utsname(2), also similar to uname -a.
+@kernel.fsflush.scanned Number of pages scanned by fsflush daemon
+@kernel.fsflush.examined Number of pages examined by fsflush daemon
+@kernel.fsflush.coalesced Number of pages coalesced into larger page
+@kernel.fsflush.modified Number of modified pages written to disk
+@kernel.fsflush.locked Number of pages locked by fsflush daemon
+Pages which were considered to be on interest for further examination
+are locked before deciding if they could be coalesced, released or flushed
+to disk.
+@kernel.fsflush.released Number of free pages released by fsflush daemon
+@kernel.fsflush.time Amount of time fsflush daemon spent doing its work
+@mem.physmem Total physical system memory
+Total physical system memory size rounded down to the nearest page size
+boundary. This metric is the same as hinv.physmem but uses different
+units.
+@mem.freemem Amount of free memory in the system
+@mem.lotsfree Paging theshold
+If freemem fails below the lostfree threshold then paging out daemon
+starts its activity. Default value for lotsfree is 1/64 of physical memory
+or 512K (which ever is larger).
+@mem.availrmem Amount of resident memory in the system
+
+@kernel.all.io.bread Physical block reads across all CPUs
+This metric is only updated if reading or writing to UFS mounted filesystems,
+reads and writes to ZFS do not update this metric.
+@kernel.all.io.bwrite Physical block writes across all CPUs
+This metric is only updated if reading or writing to UFS mounted filesystems,
+reads and writes to ZFS do not update this metric.
+@kernel.all.io.lread Logical block reads across all CPUs
+This metric is only updated if reading or writing to UFS mounted filesystems,
+reads and writes to ZFS do not update this metric.
+@kernel.all.io.lwrite Logical block writes across all CPUs
+This metric is only updated if reading or writing to UFS mounted filesystems,
+reads and writes to ZFS do not update this metric.
+@kernel.all.io.phread Raw I/O reads across all CPUs
+@kernel.all.io.phwrite Raw I/O writes across all CPUs
+@kernel.all.io.intr Device interrupts across all CPUs
+
+@kernel.percpu.io.bread Physical block reads
+This metric is only updated if reading or writing to UFS mounted filesystems,
+reads and writes to ZFS do not update this metric.
+@kernel.percpu.io.bwrite Physical block writes
+This metric is only updated if reading or writing to UFS mounted filesystems,
+reads and writes to ZFS do not update this metric.
+@kernel.percpu.io.lread Logical block reads
+This metric is only updated if reading or writing to UFS mounted filesystems,
+reads and writes to ZFS do not update this metric.
+@kernel.percpu.io.lwrite Logical block writes
+This metric is only updated if reading or writing to UFS mounted filesystems,
+reads and writes to ZFS do not update this metric.
+@kernel.percpu.io.phread Raw I/O reads
+@kernel.percpu.io.phwrite Raw I/O writes
+@kernel.percpu.io.intr Device interrupts
+
+@hinv.ncpu Number of CPUs in the system
+@hinv.ndisk Number of disks in the system
+
+@kernel.all.trap Traps across all CPUs
+@kernel.all.pswitch Context switches across all CPUs
+@kernel.all.syscall Total number of system calls across all CPUs
+@kernel.all.sysexec Total number of calls from exec(2) family across all CPUs
+@kernel.all.sysfork Total number of new processes created across all CPUs
+@kernel.all.sysvfork Total number of new processes created across all CPUs
+Unlike fork vfork does not copy all the virtual memory of the parent
+process into the child process and is mostly used to create new system context
+for execve(2). vfork(2) calls are not counted towards kernel.all.sysfork.
+@kernel.all.sysread Total number of system calls from read(2) family across all CPUs
+@kernel.all.syswrite Total number of system calls from write (2) family across all CPUs
+
+@kernel.percpu.trap Traps on each CPUs
+@kernel.percpu.pswitch Context switches on each CPUs
+@kernel.percpu.syscall Total number of system calls on each CPU
+@kernel.percpu.sysexec Total number of calls from exec(2) family on each CPU
+@kernel.percpu.sysfork Total number of new processes created on each CPU
+@kernel.percpu.sysvfork Total number of new processes created on each CPU
+Unlike fork vfork does not copy all the virtual memory of the parent
+process into the child process and is mostly used to create new system context
+for execve(2). vfork(2) calls are not counted towards kernel.percpu.sysfork.
+@kernel.percpu.sysread Total number of system calls from read(2) family on each CPU
+@kernel.percpu.syswrite Total number of system calls from write (2) family on each CPU
+
+@kernel.all.load Classic load average for 1, 5 and 15 minute intervals
+
+@kernel.all.cpu.wait.io Time spent waiting for I/O across all CPUs
+This metric is not updated by OpenSolaris kernel.
+@kernel.all.cpu.wait.pio Time spent wait for polled I/O across all CPUs
+This metric is not updated by OpenSolaris kernel.
+@kernel.all.cpu.wait.swap Time spent wait for swap across all CPUs
+This metric is not updated by OpenSolaris kernel.
+@kernel.percpu.cpu.wait.io Time spent waiting for I/O on per-CPU basis
+This metric is not updated by OpenSolaris kernel.
+@kernel.percpu.cpu.wait.pio Time spent waiting for polled I/O on per-CPU basis
+This metric is not updated by OpenSolaris kernel.
+@kernel.percpu.cpu.wait.swap Time spent waiting swap on per-CPU basis
+This metric is not updated by OpenSolaris kernel.
+
+@zfs.arc.size Total amount of memory used by ZFS ARC
+@zfs.arc.min_size Lower limit of them amount of memory for ZFS ARC
+@zfs.arc.max_size Upper limit of the amount of memory for ZFS ARC
+The default is to use 7/8 of total physical memory.
+@zfs.arc.mru_size Amount of memory used by the most recently used pages
+@zfs.arc.target_size "Ideal" size of the cached based on aging
+@zfs.arc.hits.total Number of times data is found in the cache
+@zfs.arc.hits.mfu Number of times data is found in the most frequently used buffers
+@zfs.arc.hits.mru Number of times data is found in the most recently used buffers
+@zfs.arc.hits.mfu_ghost Number of times MFU ghost buffer is accessed
+A ghost buffer is a buffer which is no longer cached but is still
+linked into the hash.
+@zfs.arc.hits.mru_ghost Number of times MRU ghost buffer is accessed
+A ghost buffer is a buffer which is no longer cached but is still
+linked into the hash.
+@zfs.arc.hits.demand_data Number of times file data is found in the cache
+ARC statistics provide separate counters for demand vs prefetch
+and data vs metadata accesses: demand is a result of the direct
+request for a particular data, prefetch is a result of speculative
+request for a particular data.
+@zfs.arc.hits.demand_metadata Number of times filesystem metadata is found in the cache
+ARC statistics provide separate counters for demand vs prefetch
+and data vs metadata accesses: demand is a result of the direct
+request for a particular data, prefetch is a result of speculative
+request for a particular data.
+@zfs.arc.hits.prefetch_data Number of times speculative request for data is satisfied from the cache
+ARC statistics provide separate counters for demand vs prefetch
+and data vs metadata accesses: demand is a result of the direct
+request for a particular data, prefetch is a result of speculative
+request for a particular data.
+@zfs.arc.hits.prefetch_metadata Number of times speculative request for metadata is satisfied from the cache
+ARC statistics provide separate counters for demand vs prefetch
+and data vs metadata accesses: demand is a result of the direct
+request for a particular data, prefetch is a result of speculative
+request for a particular data.
+@zfs.arc.misses.total Number of times the data is not found in the cache
+@zfs.arc.misses.demand_data Number of times file data is not found in the cache
+ARC statistics provide separate counters for demand vs prefetch
+and data vs metadata accesses: demand is a result of the direct
+request for a particular data, prefetch is a result of speculative
+request for a particular data.
+@zfs.arc.misses.demand_metadata Number of times filesystem metadata is not found in the cache
+ARC statistics provide separate counters for demand vs prefetch
+and data vs metadata accesses: demand is a result of the direct
+request for a particular data, prefetch is a result of speculative
+request for a particular data.
+@zfs.arc.misses.prefetch_data Number of times speculatively accessed file data is not found in the cache
+ARC statistics provide separate counters for demand vs prefetch
+and data vs metadata accesses: demand is a result of the direct
+request for a particular data, prefetch is a result of speculative
+request for a particular data.
+@zfs.arc.misses.prefetch_metadata Number of times speculatively accessed filesystem metadata is not found in the cache
+ARC statistics provide separate counters for demand vs prefetch
+and data vs metadata accesses: demand is a result of the direct
+request for a particular data, prefetch is a result of speculative
+request for a particular data.
+@pmda.prefetch.time Amount of time spent extracting information about a group of metrics
+Each metric belongs to a prefetch group. When a client asks for the metric
+to be fetched the information for the group must be extracted from the kernel.
+@pmda.prefetch.count Number of times each group of metrics was updated
+
+@pmda.metric.time Amount of time spent extracting information about individual metrics
+Requesting multiple instances of the same metrics counts against the metric
+itself and not against the individual instances
+@pmda.metric.count Number of times individual metrics have been fetched
+Requesting multiple instances of the same metrics counts as multiple hits
+against the metric itself
+
+@disk.all.wait.time Amount of time IO requests spent waiting for service
+Amount of time IO transactions spent waiting to be serviced, i.e. the
+transaction has been accepted for processing but for which the processing
+has not yet begun. Each transaction waiting for processing adds to
+to the total time which means that if multiple transactions are waiting then
+total time for the sampling interval may be larger then the interval.
+
+@disk.dev.wait.time Amount of time IO requests spent waiting for service
+Amount of time IO transactions spent waiting to be serviced, i.e. the
+transaction has been accepted for processing but for which the processing
+has not yet begun. Each transaction waiting for processing adds to
+to the total time which means that if multiple transactions are waiting then
+total time for the sampling interval may be larger then the interval.
+
+@disk.all.wait.count Number of transactions waiting to be serviced
+Number of transactions accepted for processing but for which the processing
+has not yet begun.
+@disk.dev.wait.count Number of transactions waiting to be serviced
+Number of transactions accepted for processing but for which the processing
+has not yet begun.
+
+@disk.all.run.time Amount of time spent processing IO requests
+@disk.dev.run.time Amount of time spent processing IO requests
+@disk.all.run.count Number of transactions being processed
+@disk.dev.run.count Number of transactions being processed
+
+
+# from i86pc/os/cpuid.c
+# /*
+# * 8bit APIC IDs on dual core Pentiums
+# * look like this:
+# *
+# * +-----------------------+------+------+
+# * | Physical Package ID | MC | HT |
+# * +-----------------------+------+------+
+# * <------- chipid -------->
+# * <------- coreid --------------->
+# * <--- clogid -->
+# * <------>
+# * pkgcoreid
+# *
+# * Where the number of bits necessary to
+# * represent MC and HT fields together equals
+# * to the minimum number of bits necessary to
+# * store the value of cpi->cpi_ncpu_per_chip.
+# * Of those bits, the MC part uses the number
+# * of bits necessary to store the value of
+# * cpi->cpi_ncore_per_chip.
+# */
+#
+@hinv.cpu.brand Marketing name of CPU
+@hinv.cpu.clock Current CPU clock frequency
+On CPUs which support dynamic clock frequency changes current clock frequency
+could differ from the nominal ("maximum") clock frequency specified by
+the manufacturer.
+@hinv.cpu.maxclock Maximum clock frequency supported by CPU
+Nominal CPU clock frequency as specified by the manufacturer.
+@hinv.cpu.frequencies List of clock frequencies supported by CPU
+@hinv.cpu.implementation Details of CPU implementation
+@hinv.cpu.chip_id Chip or Socket identifier of the CPU
+Logical CPUs can share single chip identifier
+@hinv.cpu.clog_id Logical core identifier
+Logical cores identifier combines identifiers of the CPU core and
+virtual CPU identifier (aka hyperthread identifier).
+@hinv.cpu.core_id CPU core identifier
+CPU core identifire combines chip identifier and per-chip core
+identifier. If cores support more the one virtual CPU per core
+then same core identifier is shared across several virtual
+CPUs.
+@hinv.cpu.pkg_core_id Per-chip core identifier
+This identifier is used to identify individual cores within the
+package. If a core support more the one virtual CPU per core
+then same core identifier is shared across several virtual
+CPUs.
+@hinv.cpu.cstate Current CPU idle state
+@hinv.cpu.maxcstates Maximum number of idle state supported by the CPU
+Information about cstate is available in kstat(1m).
+@hinv.cpu.ncores Number of CPU cores per physical chip
+@hinv.cpu.ncpus Number of virtual CPUs per physical chip
+
+@disk.dev.errors.soft Number of soft errors per device
+@disk.dev.errors.hard Number of hard errors per device
+@disk.dev.errors.transport Number of transport errors per device
+@disk.dev.errors.media Number of media errors per device
+@disk.dev.errors.recoverable Number of recoverable errors per device
+@disk.dev.errors.notready Number of times device reported as not ready
+@disk.dev.errors.nodevice Number of times device was found missing
+@disk.dev.errors.badrequest Number of illegal requests per device
+@disk.dev.errors.pfa Number of times failure prediction threshold has been exceeded
+@hinv.disk.vendor Device Vendor
+Can be reported as ATA if SATA device is behind SAS expander
+@hinv.disk.product Device name
+Vendor's device name (up-to 16 characters long)
+@hinv.disk.revision Device Revision
+@hinv.disk.serial Device Serial Number
+@hinv.disk.capacity Device Capacity
+For removable devices capacity of the media is reported.
+
+@kernel.fs.vnops.access Number of VOP_ACCESS calls per filesystem
+VOP_ACCESS is used by access(2) system call.
+@kernel.fs.vnops.addmap Number of VOP_ADDMAP calls per filesystem
+VOP_ADDMAP is used to manage reference counting of the vnode used by
+mmap(2) operations.
+@kernel.fs.vnops.close Number of VOP_CLOSE calls per filesystem
+VOP_CLOSE is called every time a close(2) system call is called
+@kernel.fs.vnops.cmp Number of VOP_CMP calls per filesystem
+VOP_CMP is used to check if two vnodes are "equal" to each other, i.e.
+both refer to the same filesystem object.
+@kernel.fs.vnops.create Number of VOP_CREATE calls per filesystem
+VOP_CREATE is used to create regular files and device or FIFO nodes.
+@kernel.fs.vnops.delmap Number of VOP_DELMAP calls
+VOP_DELMAP is used to destroy a previously created memory-mapped region
+of a file.
+@kernel.fs.vnops.dispose Number ot VOP_DISPOSE calls per filesystem
+VOP_DISPOSE is used to dispose(free or invalidate) of a page associated
+with a file.
+@kernel.fs.vnops.dump Number of VOP_DUMP calls per filesystem
+VOP_DUMP is used to transfer data from the frozen kernel directly
+to the dump device
+@kernel.fs.vnops.dumpctl Number of VOP_DUMPCTL calls per filesystem
+VOP_DUMPCTL sets up context used by VOP_DUMP call. It is used to
+allocate, free or search for data blocks on the dump device.
+@kernel.fs.vnops.fid Number of VOP_FID calls per filesystem
+VOP_FID is used to get file identifier which can be used instead of the
+file name in some operations. NFS server is one known user of this vnode
+operation.
+@kernel.fs.vnops.frlock Number of VOP_FRLOCK calls per filesystem
+VOP_FRLOCK is used to implement file record locking used by flock(2)
+@kernel.fs.vnops.fsync Number of VOP_FSYNC calls per filesystem
+VOP_FSYNC is used to implement fsync(2) system call which flushes
+data for a specific file to disk.
+@kernel.fs.vnops.getattr Number of VOP_GETATTR calls per filesystem
+VOP_GETATTR is used to extract vnode attributes. It use used as part of
+many system calls which manipulate file attributes, e.g. chmod(2), stat(2),
+utimes(2) etc.
+@kernel.fs.vnops.getpage Number of VOP_GETPAGE calls per filesystem
+VOP_GETPAGE is used to allocate pages (could be several at a time) to cover
+a region in a file.
+@kernel.fs.vnops.getsecattr Number of VOP_GETSECATTR calls per filesystem
+VOP_GETSECATTR used to extract ACL entires associated with a file.
+@kernel.fs.vnops.inactive Number of VOP_INACTIVE calls per filesystem
+VOP_INACTIVE is used to destroy vnode before it is removed from the
+cache or reused.
+@kernel.fs.vnops.ioctl Number of VOP_IOCTL calls per filesystem
+VOP_IOCTL is used to implement ioctl(2) system call.
+@kernel.fs.vnops.link Number of VOP_LINK calls per filesystem
+VOP_LINK is used to implement support for hard links
+@kernel.fs.vnops.lookup Number of VOP_LOOKUP calls per filesystem
+VOP_LOOKUP is used to translate filename to vnode.
+@kernel.fs.vnops.map Number of VOP_MAP calls per filesystem
+VOP_MAP is used to create a new memory-mapped region of a file
+@kernel.fs.vnops.mkdir Number of VOP_MKDIR calls per filesystem
+VOP_MKDIR is used to create directories
+@kernel.fs.vnops.open Number of VOP_OPEN calls per filesystem
+VOP_OPEN is called every time open(2) system call is called.
+@kernel.fs.vnops.pageio Number of VOP_PAGEIO calls per filesystem
+VOP_PAGEIO is similar to VOP_GETPAGE and VOP_PUTPAGE and can be used when
+either of the other two are less efficient, e.g. in the case when pages
+will be reused after the IO is done.
+@kernel.fs.vnops.pathconf Number of VOP_PATHCONF calls per filesystem
+VOP_PATHCONF is used to obtain information about filesystem's parameters
+reported by pathconf(2) system call
+@kernel.fs.vnops.poll Number of VOP_POLL calls per filesystem
+VOP_POLL is used to implement pool(2) system call
+@kernel.fs.vnops.putpage Number of VOP_PUTPAGE calls per filesystem
+VOP_PUTPAGE is used to release pages which have been used to hold
+data from a file
+@kernel.fs.vnops.read Number of VOP_READ calls per filesystem
+VOP_READ is used to implement read(2) system call
+@kernel.fs.vnops.readdir Number of VOP_READDIR calls per filesystem
+VOP_READDIR is used to read directory entries
+@kernel.fs.vnops.readlink Number of VOP_READLINK calls per filesystem
+VOP_READLINK is used to read the information about the target of the symbolic
+link
+@kernel.fs.vnops.realvp Number of VOP_REALVP calls per filesystem
+VOP_REALVP is used to traverse stacking filesystems and extract information
+about the vnode which refers to the "real" filesystem object.
+@kernel.fs.vnops.remove Number of VOP_REMOVE calls per filesystem
+VOP_REMOVE is used to remove entires from a directory.
+@kernel.fs.vnops.rename Number of VOP_RENAME calls per filesystem
+VOP_RENAME is used to implement rename(2) system call
+@kernel.fs.vnops.rmdir Number of VOP_RMDIR calls per filesystem
+VOP_RMDIR is used to implement rmdir(2) system call
+@kernel.fs.vnops.rwlock Number of VOP_RWLOCK calls per filesystem
+VOP_RWLOCK and VOP_RWUNLOCK are used to protect access to vnode data.
+@kernel.fs.vnops.rwunlock Number of VOP_RWUNLOCK calls per filesystem
+VOP_RWLOCK and VOP_RWUNLOCK are used to protect access to vnode data.
+@kernel.fs.vnops.seek Number of VOP_SEEK calls per filesystem
+VOP_SEEK is used by lseek(2). Because vnodes can be shared across multiple
+instances of vfile VOP_SEEK does not usually change the position of the
+file pointer, it instead used to verify the offset before it is changed.
+@kernel.fs.vnops.setattr Number of VOP_SETATTR calls per filesystem
+VOP_SETATTR is used to change vnode attributes which are modified by system
+calls like chmod(2), chown(2), utimes(2) etc.
+@kernel.fs.vnops.setfl Number of VOP_SETFL calls per filesystem
+VOP_SETFL is used to implement fcntl(2) F_SETFL option.
+Currently only sockfs pseudo filesystem is implementing this vnode operation.
+@kernel.fs.vnops.setsecattr Number of VOP_SETSECATTR calls per filesystem
+VOP_SETSECATTR is used to change ACL entries
+@kernel.fs.vnops.shrlock Number of VOP_SHRLOCK calls per filesystem
+VOP_SHRLOCK is usually used to implement CIFS and NLMv3 shared reservations.
+@kernel.fs.vnops.space Number of VOP_SPACE calls per filesystem
+VOP_SPACE is used to provide optimized support for growing and shrinking the files.
+F_FREESP option of fcntl(2) is using this vnode operation to implment ftruncate(3c)
+function.
+@kernel.fs.vnops.symlink Number of VOP_SYMLINK calls per filesystem
+VOP_SYMLINK is used to create symbolic links.
+@kernel.fs.vnops.vnevent Number of VOP_VNEVENT calls per filesystem
+VIP_VNEVENT is used to check if a filesystem support vnode event
+notifications for operations which change the names of the files.
+@kernel.fs.vnops.write Number of VOP_WRITE calls per filesystem
+VOP_WRITE is used to implement write(2) system call
+@kernel.fs.read_bytes Number of bytes read from a specific filesystem
+@kernel.fs.readdir_bytes Number of bytes containting directory entires read from a specific filesystem
+@kernel.fs.write_bytes Number of bytes written to a specific filesystem
+
+@kernel.fstype.vnops.access Number of VOP_ACCESS calls for all filesystems of a given type
+VOP_ACCESS is used by access(2) system call.
+@kernel.fstype.vnops.addmap Number of VOP_ADDMAP calls for all filesystems of a given type
+VOP_ADDMAP is used to manage reference counting of the vnode used by
+mmap(2) operations.
+@kernel.fstype.vnops.close Number of VOP_CLOSE calls for the specific filesystem
+VOP_CLOSE is called every time a close(2) system call is called
+@kernel.fstype.vnops.cmp Number of VOP_CMP calls for the specific filesystem
+VOP_CMP is used to check if two vnodes are "equal" to each other, i.e.
+both refer to the same filesystem object.
+@kernel.fstype.vnops.create Number of VOP_CREATE calls for all filesystems of a given type
+VOP_CREATE is used to create regular files and device or FIFO nodes.
+@kernel.fstype.vnops.delmap Number of VOP_DELMAP was called
+VOP_DELMAP is used to destroy a previously created memory-mapped region
+of a file.
+@kernel.fstype.vnops.dispose Number ot VOP_DISPOSE calls for all filesystems of a given type
+VOP_DISPOSE is used to dispose(free or invalidate) of a page associated
+with a file.
+@kernel.fstype.vnops.dump Number of VOP_DUMP calls for all filesystems of a given type
+VOP_DUMP is used to transfer data from the frozen kernel directly
+to the dump device
+@kernel.fstype.vnops.dumpctl Number of VOP_DUMPCTL calls for all filesystems of a given type
+VOP_DUMPCTL sets up context used by VOP_DUMP call. It is used to
+allocate, free or search for data blocks on the dump device.
+@kernel.fstype.vnops.fid Number of VOP_FID calls for all filesystems of a given type
+VOP_FID is used to get file identifier which can be used instead of the
+file name in some operations. NFS server is one known user of this vnode
+operation.
+@kernel.fstype.vnops.frlock Number of time VOP_FRLOCK calls for all filesystems of a given type
+VOP_FRLOCK is used to implement file record locking used by flock(2)
+@kernel.fstype.vnops.fsync Number of VOP_FSYNC calls for all filesystems of a given type
+VOP_FSYNC is used to implement fsync(2) system call which flushes
+data for a specific file to disk.
+@kernel.fstype.vnops.getattr Number of VOP_GETATTR calls for all filesystems of a given type
+VOP_GETATTR is used to extract vnode attributes. It use used as part of
+many system calls which manipulate file attributes, e.g. chmod(2), stat(2),
+utimes(2) etc.
+@kernel.fstype.vnops.getpage Number of VOP_GETPAGE calls for all filesystems of a given type
+VOP_GETPAGE is used to allocate pages (could be several at a time) to cover
+a region in a file.
+@kernel.fstype.vnops.getsecattr Number of VOP_GETSECATTR calls for all filesystems of a given type
+VOP_GETSECATTR used to extract ACL entires associated with a file.
+@kernel.fstype.vnops.inactive Number of VOP_INACTIVE calls for all filesystems of a given type
+VOP_INACTIVE is used to destroy vnode before it is removed from the
+cache or reused.
+@kernel.fstype.vnops.ioctl Number of VOP_IOCTL calls for all filesystems of a given type
+VOP_IOCTL is used to implement ioctl(2) system call.
+@kernel.fstype.vnops.link Number of VOP_LINK calls for all filesystems of a given type
+VOP_LINK is used to implement support for hard links
+@kernel.fstype.vnops.lookup Number of VOP_LOOKUP calls for all filesystems of a given type
+VOP_LOOKUP is used to translate filename to vnode.
+@kernel.fstype.vnops.map Number of VOP_MAP calls for all filesystems of a given type
+VOP_MAP is used to create a new memory-mapped region of a file
+@kernel.fstype.vnops.mkdir Number of VOP_MKDIR calls for all filesystems of a given type
+VOP_MKDIR is used to create directories
+@kernel.fstype.vnops.open Number of VOP_OPEN calls for all filesystems of a given type
+VOP_OPEN is called every time open(2) system call is called.
+@kernel.fstype.vnops.pageio Number of VOP_PAGEIO calls for all filesystems of a given type
+VOP_PAGEIO is similar to VOP_GETPAGE and VOP_PUTPAGE and can be used when
+either of the other two are less efficient, e.g. in the case when pages
+will be reused after the IO is done.
+@kernel.fstype.vnops.pathconf Number of VOP_PATHCONF calls for all filesystems of a given type
+VOP_PATHCONF is used to obtain information about filesystem's parameters
+reported by pathconf(2) system call
+@kernel.fstype.vnops.poll Number of VOP_POLL calls for all filesystems of a given type
+VOP_POLL is used to implement pool(2) system call
+@kernel.fstype.vnops.putpage Number of VOP_PUTPAGE calls for all filesystems of a given type
+VOP_PUTPAGE is used to release pages which have been used to hold
+data from a file
+@kernel.fstype.vnops.read Number of VOP_READ calls for all filesystems of a given type
+VOP_READ is used to implement read(2) system call
+@kernel.fstype.vnops.readdir Number of VOP_READDIR calls for all filesystems of a given type
+VOP_READDIR is used to read directory entries
+@kernel.fstype.vnops.readlink Number of VOP_READLINK calls for all filesystems of a given type
+VOP_READLINK is used to read the information about the target of the symbolic
+link
+@kernel.fstype.vnops.realvp Number of VOP_REALVP calls for all filesystems of a given type
+VOP_REALVP is used to traverse stacking filesystems and extract information
+about the vnode which refers to the "real" filesystem object.
+@kernel.fstype.vnops.remove Number of VOP_REMOVE calls for all filesystems of a given type
+VOP_REMOVE is used to remove entires from a directory.
+@kernel.fstype.vnops.rename Number of VOP_RENAME calls for all filesystems of a given type
+VOP_RENAME is used to implement rename(2) system call
+@kernel.fstype.vnops.rmdir Number of VOP_RMDIR calls for all filesystems of a given type
+VOP_RMDIR is used to implement rmdir(2) system call
+@kernel.fstype.vnops.rwlock Number of VOP_RWLOCK calls for all filesystems of a given type
+VOP_RWLOCK and VOP_RWUNLOCK are used to protect access to vnode data.
+@kernel.fstype.vnops.rwunlock Number of VOP_RWUNLOCK calls for all filesystems of a given type
+VOP_RWLOCK and VOP_RWUNLOCK are used to protect access to vnode data.
+@kernel.fstype.vnops.seek Number of VOP_SEEK calls for all filesystems of a given type
+VOP_SEEK is used by lseek(2). Because vnodes can be shared across multiple
+instances of vfile VOP_SEEK does not usually change the position of the
+file pointer, it instead used to verify the offset before it is changed.
+@kernel.fstype.vnops.setattr Number of VOP_SETATTR calls for all filesystems of a given type
+VOP_SETATTR is used to change vnode attributes which are modified by system
+calls like chmod(2), chown(2), utimes(2) etc.
+@kernel.fstype.vnops.setfl Number of VOP_SETFL calls for all filesystems of a given type
+VOP_SETFL is used to implement fcntl(2) F_SETFL option.
+Currently only sockfs pseudo filesystem is implementing this vnode operation.
+@kernel.fstype.vnops.setsecattr Number of VOP_SETSECATTR calls for all filesystems of a given type
+VOP_SETSECATTR is used to change ACL entries
+@kernel.fstype.vnops.shrlock Number of VOP_SHRLOCK calls for all filesystems of a given type
+VOP_SHRLOCK is usually used to implement CIFS and NLMv3 shared reservations.
+@kernel.fstype.vnops.space Number of VOP_SPACE calls for all filesystems of a given type
+VOP_SPACE is used to provide optimized support for growing and shrinking the files.
+F_FREESP option of fcntl(2) is using this vnode operation to implment ftruncate(3c)
+function.
+@kernel.fstype.vnops.symlink Number of VOP_SYMLINK calls for all filesystems of a given type
+VOP_SYMLINK is used to create symbolic links.
+@kernel.fstype.vnops.vnevent Number of VOP_VNEVENT calls for all filesystems of a given type
+VIP_VNEVENT is used to check if a filesystem support vnode event
+notifications for operations which change the names of the files.
+@kernel.fstype.vnops.write Number of VOP_WRITE calls for all filesystems of a given type
+VOP_WRITE is used to implement write(2) system call
+@kernel.fstype.read_bytes Bytes read from all filesystems of a given type
+@kernel.fstype.readdir_bytes Bytes read for directory entries from all filesystems of a given type
+@kernel.fstype.write_bytes Bytes written to all filesystems of a given type
+
+@hinv.disk.devlink Disk name in the descriptive format
+Solaris uses symbolic links under /dev to provide access to device nodes via
+"descriptive" names like /dev/dsk/cXtYdZsN. This metrics provides a
+translation from a "descriptive" name to instances in the disk instance
+domain.
+
+The name is always the name of the first minor device for a particular disk
+and includes the slice information.
+
+NOTE! Fetching this metric is expensive - several system calls are made
+ to fetch each instance.