# # Copyright (c) 2000,2004-2008 Silicon Graphics, Inc. All Rights Reserved. # Portions Copyright (c) International Business Machines Corp., 2002 # Portions Copyright (c) 2007-2009 Aconex. All Rights Reserved. # Portions Copyright (c) 2013 Red Hat. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the # Free Software Foundation; either version 2 of the License, or (at your # option) any later version. # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # Linux PMDA help file in the ASCII format # # lines beginning with a # are ignored # lines beginning @ introduce a new entry of the form # @ metric_name oneline-text # help test goes # here over multiple lines # ... # # the metric_name is decoded against the default PMNS -- as a special case, # a name of the form NNN.MM (for numeric NNN and MM) is interpreted as an # instance domain identification, and the text describes the instance domain # # blank lines before the @ line are ignored # @ kernel.uname.release release level of the running kernel Release level of the running kernel as reported via the release[] value returned from uname(2) or uname -r. See also pmda.uname. @ kernel.uname.version version level (build number) and build date of the running kernel Version level of the running kernel as reported by the version[] value returned from uname(2) or uname -v. Usually a build number followed by a build date. See also pmda.uname. @ kernel.uname.sysname name of the implementation of the operating system Name of the implementation of the running operating system as reported by the sysname[] value returned from uname(2) or uname -s. Usually "Linux". See also pmda.uname. @ kernel.uname.machine name of the hardware type the system is running on Name of the hardware type the system is running on as reported by the machine[] value returned from uname(2) or uname -m, e.g. "i686". See also pmda.uname. @ kernel.uname.nodename host name of this node on the network Name of this node on the network as reported by the nodename[] value returned from uname(2) or uname -n. Usually a synonym for the host name. See also pmda.uname. @ kernel.uname.distro Linux distribution name The Linux distribution name, as determined by a number of heuristics. For example: + on Fedora, the contents of /etc/fedora-release + on RedHat, the contents of /etc/redhat-release @ kernel.percpu.cpu.user percpu user CPU time metric from /proc/stat, including guest CPU time @ kernel.percpu.cpu.vuser percpu user CPU time metric from /proc/stat, excluding guest CPU time @ kernel.percpu.cpu.nice percpu nice user CPU time metric from /proc/stat @ kernel.percpu.cpu.sys percpu sys CPU time metric from /proc/stat @ kernel.percpu.cpu.idle percpu idle CPU time metric from /proc/stat @ kernel.percpu.cpu.wait.total percpu wait CPU time Per-CPU I/O wait CPU time - time spent with outstanding I/O requests. @ kernel.percpu.cpu.intr percpu interrupt CPU time Total time spent processing interrupts on each CPU (this includes both soft and hard interrupt processing time). @ kernel.percpu.cpu.irq.soft percpu soft interrupt CPU time Per-CPU soft interrupt CPU time (deferred interrupt handling code, not run in the initial interrupt handler). @ kernel.percpu.cpu.irq.hard percpu hard interrupt CPU time Per-CPU hard interrupt CPU time ("hard" interrupt handling code is the code run directly on receipt of the initial hardware interrupt, and does not include "soft" interrupt handling code which is deferred until later). @ kernel.percpu.cpu.steal percpu CPU steal time Per-CPU time when the CPU had a runnable process, but the hypervisor (virtualisation layer) chose to run something else instead. @ kernel.percpu.cpu.guest percpu guest CPU time Per-CPU time spent running (virtual) guest operating systems. @ kernel.all.interrupts.errors interrupt error count from /proc/interrupts This is a global counter (normally converted to a count/second) for any and all errors that occur while handling interrupts. @ disk.dev.read per-disk read operations Cumulative number of disk read operations since system boot time (subject to counter wrap). @ disk.dev.write per-disk write operations Cumulative number of disk write operations since system boot time (subject to counter wrap). @ disk.dev.total per-disk total (read+write) operations Cumulative number of disk read and write operations since system boot time (subject to counter wrap). @ disk.dev.blkread per-disk block read operations Cumulative number of disk block read operations since system boot time (subject to counter wrap). @ disk.dev.blkwrite per-disk block write operations Cumulative number of disk block write operations since system boot time (subject to counter wrap). @ disk.dev.blktotal per-disk total (read+write) block operations Cumulative number of disk block read and write operations since system boot time (subject to counter wrap). @ disk.dev.read_bytes per-disk count of bytes read @ disk.dev.write_bytes per-disk count of bytes written @ disk.dev.total_bytes per-disk count of total bytes read and written @ disk.dev.scheduler per-disk I/O scheduler The name of the I/O scheduler in use for each device. The scheduler is part of the block layer in the kernel, and attempts to optimise the I/O submission patterns using various techniques (typically, sorting and merging adjacent requests into larger ones to reduce seek activity, but certainly not limited to that). @ disk.dev.avactive per-disk count of active time When converted to a rate, this metric represents the average utilization of the disk during the sampling interval. A value of 0.5 (or 50%) means the disk was active (i.e. busy) half the time. @ disk.dev.read_rawactive per-disk raw count of active read time When converted to a rate, this metric represents the raw utilization of the disk during the sampling interval as a result of reads. Accounting for this metric is only done on I/O completion and can thus result in more than a second's worth of IO being accounted for within any one second, leading to >100% utilisation. It is suitable mainly for use in calculations with other metrics, e.g. mirroring the results from existing performance tools: iostat.dev.r_await = delta(disk.dev.read_rawactive) / delta(disk.dev.read) @ disk.dev.write_rawactive per-disk raw count of active write time When converted to a rate, this metric represents the raw utilization of the disk during the sampling interval as a result of writes. Accounting for this metric is only done on I/O completion and can thus result in more than a second's worth of IO being accounted for within any one second, leading to >100% utilisation. It is suitable mainly for use in calculations with other metrics, e.g. mirroring the results from existing performance tools: iostat.dev.w_await = delta(disk.dev.write_rawactive) / delta(disk.dev.write) @ disk.dev.aveq per-disk time averaged count of request queue length When converted to a rate, this metric represents the time averaged disk request queue length during the sampling interval. A value of 2.5 (or 250%) represents a time averaged queue length of 2.5 requests during the sampling interval. @ disk.dev.read_merge per-disk count of merged read requests Count of read requests that were merged with an already queued read request. @ disk.dev.write_merge per-disk count of merged write requests Count of write requests that were merged with an already queued write request. @ disk.dm.read per-device-mapper device read operations @ disk.dm.write per-device-mapper device write operations @ disk.dm.total per-device-mapper device total (read+write) operations @ disk.dm.blkread per-device-mapper device block read operations @ disk.dm.blkwrite per-device-mapper device block write operations @ disk.dm.blktotal per-device-mapper device total (read+write) block operations @ disk.dm.read_bytes per-device-mapper device count of bytes read @ disk.dm.write_bytes per-device-mapper device count of bytes written @ disk.dm.total_bytes per-device-mapper device count of total bytes read and written @ disk.dm.read_merge per-device-mapper device count of merged read requests @ disk.dm.write_merge per-device-mapper device count of merged write requests @ disk.dm.avactive per-device-mapper device count of active time @ disk.dm.aveq per-device-mapper device time averaged count of request queue length @ disk.dm.read_rawactive per-device-mapper raw count of active read time When converted to a rate, this metric represents the raw utilization of the device during the sampling interval as a result of reads. Accounting for this metric is only done on I/O completion and can thus result in more than a second's worth of IO being accounted for within any one second, leading to >100% utilisation. It is suitable mainly for use in calculations with other metrics, e.g. mirroring the results from existing performance tools: iostat.dm.r_await = delta(disk.dm.read_rawactive) / delta(disk.dm.read) @ disk.dm.write_rawactive per-device-mapper raw count of active write time When converted to a rate, this metric represents the raw utilization of the device during the sampling interval as a result of writes. Accounting for this metric is only done on I/O completion and can thus result in more than a second's worth of IO being accounted for within any one second, leading to >100% utilisation. It is suitable mainly for use in calculations with other metrics, e.g. mirroring the results from existing performance tools: iostat.dm.w_await = delta(disk.dm.write_rawactive) / delta(disk.dm.write) @ hinv.map.dmname per-device-mapper device persistent name mapping to dm-[0-9]* @ disk.all.read_merge total count of merged read requests, summed for all disks Total count of read requests that were merged with an already queued read request. @ disk.all.write_merge total count of merged write requests, summed for all disks Total count of write requests that were merged with an already queued write request. @ disk.all.avactive total count of active time, summed for all disks When converted to a rate, this metric represents the average utilization of all disks during the sampling interval. A value of 0.25 (or 25%) means that on average every disk was active (i.e. busy) one quarter of the time. @ disk.all.read_rawactive raw count of active read time, summed for all disks When converted to a rate, this metric represents the raw utilization of all disks during the sampling interval due to read requests. The accounting for this metric is only done on I/O completion and can thus result in more than a second's worth of IO being accounted for within any one second, leading to >100% utilisation. It is suitable mainly for use in calculations with other metrics, e.g. mirroring the results from existing performance tools: iostat.all.r_await = delta(disk.all.read_rawactive) / delta(disk.all.read) @ disk.all.write_rawactive raw count of active write time, summed for all disks When converted to a rate, this metric represents the raw utilization of all disks during the sampling interval due to write requests. The accounting for this metric is only done on I/O completion and can thus result in more than a second's worth of IO being accounted for within any one second, leading to >100% utilisation. It is suitable mainly for use in calculations with other metrics, e.g. mirroring the result from existing performance tools: iostat.all.w_await = delta(disk.all.write_rawactive) / delta(disk.all.write) @ disk.all.aveq total time averaged count of request queue length, summed for all disks When converted to a rate, this metric represents the average across all disks of the time averaged request queue length during the sampling interval. A value of 1.5 (or 150%) suggests that (on average) each all disk experienced a time averaged queue length of 1.5 requests during the sampling interval. @ disk.all.read total read operations, summed for all disks Cumulative number of disk read operations since system boot time (subject to counter wrap), summed over all disk devices. @ disk.all.write total write operations, summed for all disks Cumulative number of disk read operations since system boot time (subject to counter wrap), summed over all disk devices. @ disk.all.total total (read+write) operations, summed for all disks Cumulative number of disk read and write operations since system boot time (subject to counter wrap), summed over all disk devices. @ disk.all.blkread block read operations, summed for all disks Cumulative number of disk block read operations since system boot time (subject to counter wrap), summed over all disk devices. @ disk.all.blkwrite block write operations, summed for all disks Cumulative number of disk block write operations since system boot time (subject to counter wrap), summed over all disk devices. @ disk.all.blktotal total (read+write) block operations, summed for all disks Cumulative number of disk block read and write operations since system boot time (subject to counter wrap), summed over all disk devices. @ disk.all.read_bytes count of bytes read for all disk devices @ disk.all.write_bytes count of bytes written for all disk devices @ disk.all.total_bytes total count of bytes read and written for all disk devices @ disk.partitions.read read operations metric for storage partitions Cumulative number of disk read operations since system boot time (subject to counter wrap) for individual disk partitions or logical volumes. @ disk.partitions.write write operations metric for storage partitions Cumulative number of disk write operations since system boot time (subject to counter wrap) for individual disk partitions or logical volumes. @ disk.partitions.total total (read+write) I/O operations metric for storage partitions Cumulative number of disk read and write operations since system boot time (subject to counter wrap) for individual disk partitions or logical volumes. @ disk.partitions.blkread block read operations metric for storage partitions Cumulative number of disk block read operations since system boot time (subject to counter wrap) for individual disk partitions or logical volumes. @ disk.partitions.blkwrite block write operations metric for storage partitions Cumulative number of disk block write operations since system boot time (subject to counter wrap) for individual disk partitions or logical volumes. @ disk.partitions.blktotal total (read+write) block operations metric for storage partitions Cumulative number of disk block read and write operations since system boot time (subject to counter wrap) for individual disk partitions or logical volumes. @ disk.partitions.read_bytes number of bytes read for storage partitions Cumulative number of bytes read since system boot time (subject to counter wrap) for individual disk partitions or logical volumes. @ disk.partitions.write_bytes number of bytes written for storage partitions Cumulative number of bytes written since system boot time (subject to counter wrap) for individual disk partitions or logical volumes. @ disk.partitions.total_bytes total number of bytes read and written for storage partitions Cumulative number of bytes read and written since system boot time (subject to counter wrap) for individual disk partitions or logical volumes. @ swap.pagesin pages read from swap devices due to demand for physical memory @ swap.pagesout pages written to swap devices due to demand for physical memory @ swap.in number of swap in operations @ swap.out number of swap out operations @ kernel.all.pswitch context switches metric from /proc/stat @ kernel.all.sysfork fork rate metric from /proc/stat @ kernel.all.intr intrrupt rate metric from /proc/stat The value is the first value from the intr field in /proc/stat, which is a counter of the total number of interrupts processed. The value is normally converted to a rate (count/second). This counter usually increases by at least HZ/second, i.e. the clock interrupt rate, wehich is usually 100/second. See also kernel.percpu.interrupts to get a breakdown of interrupt rates by interrupt type and which CPU processed each one. @ mem.physmem total system memory metric reported by /proc/meminfo The value of this metric corresponds to the "MemTotal" field reported by /proc/meminfo. Note that this does not necessarily correspond to actual installed physical memory - there may be areas of the physical address space mapped as ROM in various peripheral devices and the bios may be mirroring certain ROMs in RAM. @ mem.freemem free system memory metric from /proc/meminfo @ mem.util.used used memory metric from /proc/meminfo Used memory is the difference between mem.physmem and mem.freemem. @ mem.util.free free memory metric from /proc/meminfo Alias for mem.freemem. @ mem.util.available available memory from /proc/meminfo The amount of memory that is available for a new workload, without pushing the system into swap. Estimated from MemFree, Active(file), Inactive(file), and SReclaimable, as well as the "low" watermarks from /proc/zoneinfo. @ mem.util.shared shared memory metric from /proc/meminfo Shared memory metric. Currently always zero on Linux 2.4 kernels and has been removed from 2.6 kernels. @ mem.util.bufmem I/O buffers metric from /proc/meminfo Memory allocated for buffer_heads. @ mem.util.cached page cache metric from /proc/meminfo Memory used by the page cache, including buffered file data. This is in-memory cache for files read from the disk (the pagecache) but doesn't include SwapCached. @ mem.util.other unaccounted memory Memory that is not free (i.e. has been referenced) and is not cached. mem.physmem - mem.util.free - mem.util.cached - mem.util.buffers @ mem.util.active Kbytes on the active page list (recently referenced pages) Memory that has been used more recently and usually not reclaimed unless absolutely necessary. @ mem.util.inactive Kbytes on the inactive page list (candidates for discarding) Memory which has been less recently used. It is more eligible to be reclaimed for other purposes @ mem.util.swapCached Kbytes in swap cache, from /proc/meminfo Memory that once was swapped out, is swapped back in but still also is in the swapfile (if memory is needed it doesn't need to be swapped out AGAIN because it is already in the swapfile. This saves I/O) @ mem.util.highTotal Kbytes in high memory, from /proc/meminfo This is apparently an i386 specific metric, and seems to be always zero on ia64 architecture (and possibly others). On i386 arch (at least), highmem is all memory above ~860MB of physical memory. Highmem areas are for use by userspace programs, or for the pagecache. The kernel must use tricks to access this memory, making it slower to access than lowmem. @ mem.util.highFree Kbytes free high memory, from /proc/meminfo See mem.util.highTotal. Not used on ia64 arch (and possibly others). @ mem.util.lowTotal Kbytes in low memory total, from /proc/meminfo Lowmem is memory which can be used for everything that highmem can be used for, but it is also availble for the kernel's use for its own data structures. Among many other things, it is where everything from the Slab is allocated. Bad things happen when you're out of lowmem. (this may only be true on i386 architectures). @ mem.util.lowFree Kbytes free low memory, from /proc/meminfo See mem.util.lowTotal @ mem.util.swapTotal Kbytes swap, from /proc/meminfo total amount of swap space available @ mem.util.swapFree Kbytes free swap, from /proc/meminfo Memory which has been evicted from RAM, and is temporarily on the disk @ mem.util.dirty Kbytes in dirty pages, from /proc/meminfo Memory which is waiting to get written back to the disk @ mem.util.writeback Kbytes in writeback pages, from /proc/meminfo Memory which is actively being written back to the disk @ mem.util.mapped Kbytes in mapped pages, from /proc/meminfo files which have been mmaped, such as libraries @ mem.util.slab Kbytes in slab memory, from /proc/meminfo in-kernel data structures cache @ mem.util.commitLimit Kbytes limit for address space commit, from /proc/meminfo The static total, in Kbytes, available for commitment to address spaces. Thus, mem.util.committed_AS may range up to this total. Normally the kernel overcommits memory, so this value may exceed mem.physmem @ mem.util.committed_AS Kbytes committed to address spaces, from /proc/meminfo An estimate of how much RAM you would need to make a 99.99% guarantee that there never is OOM (out of memory) for this workload. Normally the kernel will overcommit memory. That means, say you do a 1GB malloc, nothing happens, really. Only when you start USING that malloc memory you will get real memory on demand, and just as much as you use. @ mem.util.pageTables Kbytes in kernel page tables, from /proc/meminfo @ mem.util.reverseMaps Kbytes in reverse mapped pages, from /proc/meminfo @ mem.util.cache_clean Kbytes cached and not dirty or writeback, derived from /proc/meminfo @ mem.util.anonpages Kbytes in user pages not backed by files, from /proc/meminfo @ mem.util.bounce Kbytes in bounce buffers, from /proc/meminfo @ mem.util.NFS_Unstable Kbytes in NFS unstable memory, from /proc/meminfo @ mem.util.slabReclaimable Kbytes in reclaimable slab pages, from /proc/meminfo @ mem.util.slabUnreclaimable Kbytes in unreclaimable slab pages, from /proc/meminfo @ mem.util.active_anon anonymous Active list LRU memory @ mem.util.inactive_anon anonymous Inactive list LRU memory @ mem.util.active_file file-backed Active list LRU memory @ mem.util.inactive_file file-backed Inactive list LRU memory @ mem.util.unevictable kbytes of memory that is unevictable @ mem.util.mlocked kbytes of memory that is pinned via mlock() @ mem.util.shmem kbytes of shmem @ mem.util.kernelStack kbytes of memory used for kernel stacks @ mem.util.hugepagesTotal a count of total hugepages @ mem.util.hugepagesFree a count of free hugepages @ mem.util.hugepagesSurp a count of surplus hugepages @ mem.util.directMap4k amount of memory that is directly mapped in 4kB pages @ mem.util.directMap2M amount of memory that is directly mapped in 2MB pages @ mem.util.directMap1G amount of memory that is directly mapped in 1GB pages @ mem.util.vmallocTotal amount of kernel memory allocated via vmalloc @ mem.util.vmallocUsed amount of used vmalloc memory @ mem.util.vmallocChunk amount of vmalloc chunk memory @ mem.util.mmap_copy amount of mmap_copy space (non-MMU kernels only) @ mem.util.quicklists amount of memory in the per-CPU quicklists @ mem.util.corrupthardware amount of memory in hardware corrupted pages @ mem.util.anonhugepages amount of memory in anonymous huge pages User memory (Kbytes) in pages not backed by files, e.g. from malloc() @ mem.numa.util.total per-node total memory @ mem.numa.util.free per-node free memory @ mem.numa.util.used per-node used memory @ mem.numa.util.active per-node Active list LRU memory @ mem.numa.util.inactive per-node Inactive list LRU memory @ mem.numa.util.active_anon per-node anonymous Active list LRU memory @ mem.numa.util.inactive_anon per-node anonymous Inactive list LRU memory @ mem.numa.util.active_file per-node file-backed Active list LRU memory @ mem.numa.util.inactive_file per-node file-backed Inactive list LRU memory @ mem.numa.util.highTotal per-node highmem total @ mem.numa.util.highFree per-node highmem free @ mem.numa.util.lowTotal per-node lowmem total @ mem.numa.util.lowFree per-node lowmem free @ mem.numa.util.unevictable per-node Unevictable memory @ mem.numa.util.mlocked per-node count of Mlocked memory @ mem.numa.util.dirty per-node dirty memory @ mem.numa.util.writeback per-node count of memory locked for writeback to stable storage @ mem.numa.util.filePages per-node count of memory backed by files @ mem.numa.util.mapped per-node mapped memory @ mem.numa.util.anonpages per-node anonymous memory @ mem.numa.util.shmem per-node amount of shared memory @ mem.numa.util.kernelStack per-node memory used as kernel stacks @ mem.numa.util.pageTables per-node memory used for pagetables @ mem.numa.util.NFS_Unstable per-node memory holding NFS data that needs writeback @ mem.numa.util.bounce per-node memory used for bounce buffers @ mem.numa.util.writebackTmp per-node temporary memory used for writeback @ mem.numa.util.slab per-node memory used for slab objects @ mem.numa.util.slabReclaimable per-node memory used for slab objects that can be reclaimed @ mem.numa.util.slabUnreclaimable per-node memory used for slab objects that is unreclaimable @ mem.numa.util.hugepagesTotal per-node total count of hugepages @ mem.numa.util.hugepagesFree per-node count of free hugepages @ mem.numa.util.hugepagesSurp per-node count of surplus hugepages @ mem.numa.alloc.hit per-node count of times a task wanted alloc on local node and succeeded @ mem.numa.alloc.miss per-node count of times a task wanted alloc on local node but got another node @ mem.numa.alloc.foreign count of times a task on another node alloced on that node, but got this node @ mem.numa.alloc.interleave_hit count of times interleaving wanted to allocate on this node and succeeded @ mem.numa.alloc.local_node count of times a process ran on this node and got memory on this node @ mem.numa.alloc.other_node count of times a process ran on this node and got memory from another node @ mem.vmstat.nr_dirty number of pages in dirty state Instantaneous number of pages in dirty state, from /proc/vmstat @ mem.vmstat.nr_dirtied count of pages dirtied Count of pages entering dirty state, from /proc/vmstat @ mem.vmstat.nr_writeback number of pages in writeback state Instantaneous number of pages in writeback state, from /proc/vmstat @ mem.vmstat.nr_unstable number of pages in unstable state Instantaneous number of pages in unstable state, from /proc/vmstat @ mem.vmstat.nr_page_table_pages number of page table pages Instantaneous number of page table pages, from /proc/vmstat @ mem.vmstat.nr_mapped number of mapped pagecache pages Instantaneous number of mapped pagecache pages, from /proc/vmstat See also mem.vmstat.nr_anon for anonymous mapped pages. @ mem.vmstat.nr_slab number of slab pages Instantaneous number of slab pages, from /proc/vmstat This counter was retired in 2.6.18 kernels, and is now the sum of mem.vmstat.nr_slab_reclaimable and mem.vmstat.nr_slab_unreclaimable. @ mem.vmstat.nr_written count of pages written out Count of pages written out, from /proc/vmstat @ mem.vmstat.numa_foreign count of foreign NUMA zone allocations @ mem.vmstat.numa_hit count of successful allocations from preferred NUMA zone @ mem.vmstat.numa_interleave count of interleaved NUMA allocations @ mem.vmstat.numa_local count of successful allocations from local NUMA zone @ mem.vmstat.numa_miss count of unsuccessful allocations from preferred NUMA zona @ mem.vmstat.numa_other count of unsuccessful allocations from local NUMA zone @ mem.vmstat.pgpgin page in operations Count of page in operations since boot, from /proc/vmstat @ mem.vmstat.pgpgout page out operations Count of page out operations since boot, from /proc/vmstat @ mem.vmstat.pswpin pages swapped in Count of pages swapped in since boot, from /proc/vmstat @ mem.vmstat.pswpout pages swapped out Count of pages swapped out since boot, from /proc/vmstat @ mem.vmstat.pgalloc_high high mem page allocations Count of high mem page allocations since boot, from /proc/vmstat @ mem.vmstat.pgalloc_normal normal mem page allocations Count of normal mem page allocations since boot, from /proc/vmstat @ mem.vmstat.pgalloc_dma dma mem page allocations Count of dma mem page allocations since boot, from /proc/vmstat @ mem.vmstat.pgalloc_dma32 dma32 mem page allocations Count of dma32 mem page allocations since boot, from /proc/vmstat @ mem.vmstat.pgalloc_movable movable mem page allocations Count of movable mem page allocations since boot, from /proc/vmstat @ mem.vmstat.pgfree page free operations Count of page free operations since boot, from /proc/vmstat @ mem.vmstat.pgactivate pages moved from inactive to active Count of pages moved from inactive to active since boot, from /proc/vmstat @ mem.vmstat.pgdeactivate pages moved from active to inactive Count of pages moved from active to inactive since boot, from /proc/vmstat @ mem.vmstat.pgfault page major and minor fault operations Count of page major and minor fault operations since boot, from /proc/vmstat @ mem.vmstat.pgmajfault major page fault operations Count of major page fault operations since boot, from /proc/vmstat @ mem.vmstat.pgrefill_high high mem pages inspected in refill_inactive_zone Count of high mem pages inspected in refill_inactive_zone since boot, from /proc/vmstat @ mem.vmstat.pgrefill_normal normal mem pages inspected in refill_inactive_zone Count of normal mem pages inspected in refill_inactive_zone since boot, from /proc/vmstat @ mem.vmstat.pgrefill_dma dma mem pages inspected in refill_inactive_zone Count of dma mem pages inspected in refill_inactive_zone since boot, from /proc/vmstat @ mem.vmstat.pgrefill_dma32 dma32 mem pages inspected in refill_inactive_zone Count of dma32 mem pages inspected in refill_inactive_zone since boot, from /proc/vmstat @ mem.vmstat.pgrefill_movable movable mem pages inspected in refill_inactive_zone Count of movable mem pages inspected in refill_inactive_zone since boot, from /proc/vmstat @ mem.vmstat.pgsteal_high high mem pages reclaimed Count of high mem pages reclaimed since boot, from /proc/vmstat @ mem.vmstat.pgsteal_normal normal mem pages reclaimed Count of normal mem pages reclaimed since boot, from /proc/vmstat @ mem.vmstat.pgsteal_dma dma mem pages reclaimed Count of dma mem pages reclaimed since boot, from /proc/vmstat @ mem.vmstat.pgsteal_dma32 dma32 mem pages reclaimed Count of dma32 mem pages reclaimed since boot, from /proc/vmstat @ mem.vmstat.pgsteal_movable movable mem pages reclaimed Count of movable mem pages reclaimed since boot, from /proc/vmstat @ mem.vmstat.pgscan_kswapd_high high mem pages scanned by kswapd Count of high mem pages scanned by kswapd since boot, from /proc/vmstat @ mem.vmstat.pgscan_kswapd_normal normal mem pages scanned by kswapd Count of normal mem pages scanned by kswapd since boot, from /proc/vmstat @ mem.vmstat.pgscan_kswapd_dma dma mem pages scanned by kswapd Count of dma mem pages scanned by kswapd since boot, from /proc/vmstat @ mem.vmstat.pgscan_kswapd_dma32 dma32 mem pages scanned by kswapd Count of dma32 mem pages scanned by kswapd since boot, from /proc/vmstat @ mem.vmstat.pgscan_kswapd_movable movable mem pages scanned by kswapd Count of movable mem pages scanned by kswapd since boot, from /proc/vmstat @ mem.vmstat.pgscan_direct_high high mem pages scanned Count of high mem pages scanned since boot, from /proc/vmstat @ mem.vmstat.pgscan_direct_normal normal mem pages scanned Count of normal mem pages scanned since boot, from /proc/vmstat @ mem.vmstat.pgscan_direct_dma dma mem pages scanned Count of dma mem pages scanned since boot, from /proc/vmstat @ mem.vmstat.pgscan_direct_dma32 dma32 mem pages scanned Count of dma32 mem pages scanned since boot, from /proc/vmstat @ mem.vmstat.pgscan_direct_movable Count of movable mem pages scanned since boot, from /proc/vmstat @ mem.vmstat.pginodesteal pages reclaimed via inode freeing Count of pages reclaimed via inode freeing since boot, from /proc/vmstat @ mem.vmstat.slabs_scanned slab pages scanned Count of slab pages scanned since boot, from /proc/vmstat @ mem.vmstat.kswapd_steal pages reclaimed by kswapd Count of pages reclaimed by kswapd since boot, from /proc/vmstat @ mem.vmstat.kswapd_low_wmark_hit_quickly count of times low watermark reached quickly Count of times kswapd reached low watermark quickly, from /proc/vmstat @ mem.vmstat.kswapd_high_wmark_hit_quickly count of times high watermark reached quickly Count of times kswapd reached high watermark quickly, from /proc/vmstat @ mem.vmstat.kswapd_skip_congestion_wait count of times kswapd skipped waiting on device congestion Count of times kswapd skipped waiting due to device congestion as a result of being under the low watermark, from /proc/vmstat @ mem.vmstat.kswapd_inodesteal pages reclaimed via kswapd inode freeing Count of pages reclaimed via kswapd inode freeing since boot, from /proc/vmstat @ mem.vmstat.pageoutrun kswapd calls to page reclaim Count of kswapd calls to page reclaim since boot, from /proc/vmstat @ mem.vmstat.allocstall direct reclaim calls Count of direct reclaim calls since boot, from /proc/vmstat @ mem.vmstat.pgrotated pages rotated to tail of the LRU Count of pages rotated to tail of the LRU since boot, from /proc/vmstat @mem.vmstat.nr_anon_pages number of anonymous mapped pagecache pages Instantaneous number of anonymous mapped pagecache pages, from /proc/vmstat See also mem.vmstat.mapped for other mapped pages. @mem.vmstat.nr_anon_transparent_hugepages number of anonymous transparent huge pages Instantaneous number of anonymous transparent huge pages, from /proc/vmstat @mem.vmstat.nr_bounce number of bounce buffer pages Instantaneous number of bounce buffer pages, from /proc/vmstat @mem.vmstat.nr_slab_reclaimable reclaimable slab pages Instantaneous number of reclaimable slab pages, from /proc/vmstat. @mem.vmstat.nr_slab_unreclaimable unreclaimable slab pages Instantaneous number of unreclaimable slab pages, from /proc/vmstat. @mem.vmstat.nr_vmscan_write pages written by VM scanner from LRU Count of pages written from the LRU by the VM scanner, from /proc/vmstat. The VM is supposed to minimise the number of pages which get written from the LRU (for IO scheduling efficiency, and for high reclaim-success rates). @ mem.vmstat.htlb_buddy_alloc_fail huge TLB page buddy allocation failures Count of huge TLB page buddy allocation failures, from /proc/vmstat @ mem.vmstat.htlb_buddy_alloc_success huge TLB page buddy allocation successes Count of huge TLB page buddy allocation successes, from /proc/vmstat @ mem.vmstat.nr_active_anon number of active anonymous memory pages @ mem.vmstat.nr_active_file number of active file memory memory pages @ mem.vmstat.nr_free_pages number of free pages @ mem.vmstat.nr_inactive_anon number of inactive anonymous memory pages @ mem.vmstat.nr_inactive_file number of inactive file memory pages @ mem.vmstat.nr_isolated_anon number of isolated anonymous memory pages @ mem.vmstat.nr_isolated_file number of isolated file memory pages @ mem.vmstat.nr_kernel_stack number of pages of kernel stack @ mem.vmstat.nr_mlock number of pages under mlock @ mem.vmstat.nr_shmem number of shared memory pages @ mem.vmstat.nr_unevictable number of unevictable pages @ mem.vmstat.nr_writeback_temp number of temporary writeback pages @ mem.vmstat.compact_blocks_moved count of compact blocks moved @ mem.vmstat.compact_fail count of unsuccessful compactions for high order allocations @ mem.vmstat.compact_pagemigrate_failed count of pages unsuccessfully compacted @ mem.vmstat.compact_pages_moved count of pages successfully moved for compaction @ mem.vmstat.compact_stall count of failures to even start compacting @ mem.vmstat.compact_success count of successful compactions for high order allocations @ mem.vmstat.thp_fault_alloc transparent huge page fault allocations @ mem.vmstat.thp_fault_fallback transparent huge page fault fallbacks @ mem.vmstat.thp_collapse_alloc transparent huge page collapse allocations @ mem.vmstat.thp_collapse_alloc_failed transparent huge page collapse failures @ mem.vmstat.thp_split count of transparent huge page splits @ mem.vmstat.unevictable_pgs_cleared count of unevictable pages cleared @ mem.vmstat.unevictable_pgs_culled count of unevictable pages culled @ mem.vmstat.unevictable_pgs_mlocked count of mlocked unevictable pages @ mem.vmstat.unevictable_pgs_mlockfreed count of unevictable pages mlock freed @ mem.vmstat.unevictable_pgs_munlocked count of unevictable pages munlocked @ mem.vmstat.unevictable_pgs_rescued count of unevictable pages rescued @ mem.vmstat.unevictable_pgs_scanned count of unevictable pages scanned @ mem.vmstat.unevictable_pgs_stranded count of unevictable pages stranded @ mem.vmstat.zone_reclaim_failed number of zone reclaim failures @ swap.length total swap available metric from /proc/meminfo @ swap.used swap used metric from /proc/meminfo @ swap.free swap free metric from /proc/meminfo @ kernel.all.load 1, 5 and 15 minute load average @ kernel.all.cpu.user total user CPU time from /proc/stat for all CPUs, including guest CPU time @ kernel.all.cpu.vuser total user CPU time from /proc/stat for all CPUs, excluding guest CPU time @ kernel.all.cpu.intr total interrupt CPU time from /proc/stat for all CPUs Total time spent processing interrupts on all CPUs. This value includes both soft and hard interrupt processing time. @ kernel.all.cpu.wait.total total wait CPU time from /proc/stat for all CPUs @ kernel.all.cpu.nice total nice user CPU time from /proc/stat for all CPUs @ kernel.all.cpu.sys total sys CPU time from /proc/stat for all CPUs @ kernel.all.cpu.idle total idle CPU time from /proc/stat for all CPUs @ kernel.all.cpu.irq.soft soft interrupt CPU time from /proc/stat for all CPUs Total soft interrupt CPU time (deferred interrupt handling code, not run in the initial interrupt handler). @ kernel.all.cpu.irq.hard hard interrupt CPU time from /proc/stat for all CPUs Total hard interrupt CPU time ("hard" interrupt handling code is the code run directly on receipt of the initial hardware interrupt, and does not include "soft" interrupt handling code which is deferred until later). @ kernel.all.cpu.steal total virtualisation CPU steal time for all CPUs Total CPU time when a CPU had a runnable process, but the hypervisor (virtualisation layer) chose to run something else instead. @ kernel.all.cpu.guest total virtual guest CPU time for all CPUs Total CPU time spent running virtual guest operating systems. @ kernel.all.nusers number of user sessions on system @ hinv.ninterface number of active (up) network interfaces @ network.interface.in.bytes network recv read bytes from /proc/net/dev per network interface @ network.interface.in.packets network recv read packets from /proc/net/dev per network interface @ network.interface.in.errors network recv read errors from /proc/net/dev per network interface @ network.interface.in.drops network recv read drops from /proc/net/dev per network interface @ network.interface.in.mcasts network recv compressed from /proc/net/dev per network interface @ network.interface.in.fifo network recv read fifos from /proc/net/dev per network interface @ network.interface.in.frame network recv read frames from /proc/net/dev per network interface @ network.interface.in.compressed network recv compressed from /proc/net/dev per network interface @ network.interface.out.bytes network send bytes from /proc/net/dev per network interface @ network.interface.out.packets network send packets from /proc/net/dev per network interface @ network.interface.out.errors network send errors from /proc/net/dev per network interface @ network.interface.out.drops network send drops from /proc/net/dev per network interface @ network.interface.out.fifo network send fifos from /proc/net/dev per network interface @ network.interface.collisions network send collisions from /proc/net/dev per network interface @ network.interface.out.carrier network send carrier from /proc/net/dev per network interface @ network.interface.out.compressed network send compressed from /proc/net/dev per network interface @ network.interface.total.bytes network total (in+out) bytes from /proc/net/dev per network interface @ network.interface.total.packets network total (in+out) packets from /proc/net/dev per network interface @ network.interface.total.errors network total (in+out) errors from /proc/net/dev per network interface @ network.interface.total.drops network total (in+out) drops from /proc/net/dev per network interface @ network.interface.total.mcasts network total (in+out) mcasts from /proc/net/dev per network interface @ network.interface.mtu maximum transmission unit on network interface @ network.interface.speed interface speed in megabytes per second The linespeed on the network interface, as reported by the kernel, scaled from Megabits/second to Megabytes/second. See also network.interface.baudrate for the bytes/second value. @ network.interface.baudrate interface speed in bytes per second The linespeed on the network interface, as reported by the kernel, scaled up from Megabits/second to bits/second and divided by 8 to convert to bytes/second. See also network.interface.speed for the Megabytes/second value. @ network.interface.duplex value one for half or two for full duplex interface @ network.interface.up boolean for whether interface is currently up or down @ network.interface.running boolean for whether interface has resources allocated @ network.interface.inet_addr string INET interface address (ifconfig style) @ network.interface.ipv6_addr string IPv6 interface address (ifconfig style) @ network.interface.ipv6_scope string IPv6 interface scope (ifconfig style) @ network.interface.hw_addr hardware address (from sysfs) @ network.sockstat.tcp.inuse instantaneous number of tcp sockets currently in use @ network.sockstat.tcp.highest highest number of tcp sockets in use at any one time since boot @ network.sockstat.tcp.util instantaneous tcp socket utilization (100 * inuse/highest) @ network.sockstat.udp.inuse instantaneous number of udp sockets currently in use @ network.sockstat.udp.highest highest number of udp sockets in use at any one time since boot @ network.sockstat.udp.util instantaneous udp socket utilization (100 * inuse/highest) @ network.sockstat.raw.inuse instantaneous number of raw sockets currently in use @ network.sockstat.raw.highest highest number of raw sockets in use at any one time since boot @ network.sockstat.raw.util instantaneous raw socket utilization (100 * inuse/highest) @ hinv.physmem total system memory metric from /proc/meminfo @ hinv.pagesize Memory page size The memory page size of the running kernel in bytes. @ hinv.ncpu number of CPUs in the system @ hinv.ndisk number of disks in the system @ hinv.nfilesys number of (local) file systems currently mounted @ hinv.nnode number of NUMA nodes in the system @ hinv.map.scsi list of active SCSI devices There is one string value for each SCSI device active in the system, as extracted from /proc/scsi/scsi. The external instance name for each device is in the format scsiD:C:I:L where D is controller number, C is channel number, I is device ID and L is the SCSI LUN number for the device. The values for this metric are the actual device names (sd[a-z] are SCSI disks, st[0-9] are SCSI tapes and scd[0-9] are SCSI CD-ROMS. @ hinv.nlv number of logical volumes @ hinv.map.lvname mapping of logical volume names for devices Provides a logical-volume-name to device-name mapping for the device mapper subsystem. @ filesys.capacity Total capacity of mounted filesystem (Kbytes) @ filesys.used Total space used on mounted filesystem (Kbytes) @ filesys.free Total space free on mounted filesystem (Kbytes) @ filesys.maxfiles Inodes capacity of mounted filesystem @ filesys.usedfiles Number of inodes allocated on mounted filesystem @ filesys.freefiles Number of unallocated inodes on mounted filesystem @ filesys.mountdir File system mount point @ filesys.full Percentage of filesystem in use @ filesys.blocksize Size of each block on mounted filesystem (Bytes) @ filesys.avail Total space free to non-superusers on mounted filesystem (Kbytes) @ filesys.readonly Indicates whether a filesystem is mounted readonly @ tmpfs.capacity Total capacity of mounted tmpfs filesystem (Kbytes) @ tmpfs.used Total space used on mounted tmpfs filesystem (Kbytes) @ tmpfs.free Total space free on mounted tmpfs filesystem (Kbytes) @ tmpfs.maxfiles Inodes capacity of mounted tmpfs filesystem @ tmpfs.usedfiles Number of inodes allocated on mounted tmpfs filesystem @ tmpfs.freefiles Number of unallocated inodes on mounted tmpfs filesystem @ tmpfs.full Percentage of tmpfs filesystem in use @ swapdev.free physical swap free space @ swapdev.length physical swap size @ swapdev.maxswap maximum swap length (same as swapdev.length on Linux) @ swapdev.vlength virtual swap size (always zero on Linux) Virtual swap size (always zero on Linux since Linux does not support virtual swap). This metric is retained on Linux for interoperability with PCP monitor tools running on IRIX. @ swapdev.priority swap resource priority @ nfs.client.calls cumulative total of client NFSv2 requests @ nfs.client.reqs cumulative total of client NFSv2 requests by request type @ nfs.server.calls cumulative total of server NFSv2 requests @ nfs.server.reqs cumulative total of client NFSv2 requests by request type @ nfs3.client.calls cumulative total of client NFSv3 requests @ nfs3.client.reqs cumulative total of client NFSv3 requests by request type @ nfs3.server.calls cumulative total of server NFSv3 requests @ nfs3.server.reqs cumulative total of client NFSv3 requests by request type @ nfs4.client.calls cumulative total of client NFSv4 requests @ nfs4.client.reqs cumulative total for each client NFSv4 request type @ nfs4.server.calls cumulative total of server NFSv4 operations, plus NULL requests @ nfs4.server.reqs cumulative total for each server NFSv4 operation, and for NULL requests @ rpc.client.rpccnt cumulative total of client RPC requests @ rpc.client.rpcretrans cumulative total of client RPC retransmissions @ rpc.client.rpcauthrefresh cumulative total of client RPC auth refreshes @ rpc.client.netcnt cumulative total of client RPC network layer requests @ rpc.client.netudpcnt cumulative total of client RPC UDP network layer requests @ rpc.client.nettcpcnt cumulative total of client RPC TCP network layer requests @ rpc.client.nettcpconn cumulative total of client RPC TCP network layer connection requests @ rpc.server.rpccnt cumulative total of server RPC requests @ rpc.server.rpcerr cumulative total of server RPC errors @ rpc.server.rpcbadfmt cumulative total of server RPC bad format errors @ rpc.server.rpcbadauth cumulative total of server RPC bad auth errors @ rpc.server.rpcbadclnt cumulative total of server RPC bad client errors @ rpc.server.rchits cumulative total of request-reply-cache hits @ rpc.server.rcmisses cumulative total of request-reply-cache misses @ rpc.server.rcnocache cumulative total of uncached request-reply-cache requests @ rpc.server.fh_cached cumulative total of file handle cache requests @ rpc.server.fh_valid cumulative total of file handle cache validations @ rpc.server.fh_fixup cumulative total of file handle cache fixup validations @ rpc.server.fh_lookup cumulative total of file handle cache new lookups @ rpc.server.fh_stale cumulative total of stale file handle cache errors @ rpc.server.fh_concurrent cumulative total of concurrent file handle cache requests @ rpc.server.netcnt cumulative total of server RPC network layer requests @ rpc.server.netudpcnt cumulative total of server RPC UDP network layer requests @ rpc.server.nettcpcnt cumulative total of server RPC TCP network layer requests @ rpc.server.nettcpconn cumulative total of server RPC TCP network layer connection requests @ rpc.server.fh_anon cumulative total anonymous file dentries returned @ rpc.server.fh_nocache_dir count of directory file handles not found cached @ rpc.server.fh_nocache_nondir count of non-directory file handles not found cached @ rpc.server.io_read cumulative count of bytes returned from read requests @ rpc.server.io_write cumulative count of bytes passed into write requests @ rpc.server.th_cnt available nfsd threads @ rpc.server.th_fullcnt number of times the last free nfsd thread was used @ network.ip.forwarding count of ip forwarding @ network.ip.defaultttl count of ip defaultttl @ network.ip.inreceives count of ip inreceives @ network.ip.inhdrerrors count of ip inhdrerrors @ network.ip.inaddrerrors count of ip inaddrerrors @ network.ip.forwdatagrams count of ip forwdatagrams @ network.ip.inunknownprotos count of ip inunknownprotos @ network.ip.indiscards count of ip indiscards @ network.ip.indelivers count of ip indelivers @ network.ip.outrequests count of ip outrequests @ network.ip.outdiscards count of ip outdiscards @ network.ip.outnoroutes count of ip outnoroutes @ network.ip.reasmtimeout count of ip reasmtimeout @ network.ip.reasmreqds count of ip reasmreqds @ network.ip.reasmoks count of ip reasmoks @ network.ip.reasmfails count of ip reasmfails @ network.ip.fragoks count of ip fragoks @ network.ip.fragfails count of ip fragfails @ network.ip.fragcreates count of ip fragcreates @ network.icmp.inmsgs count of icmp inmsgs @ network.icmp.inerrors count of icmp inerrors @ network.icmp.indestunreachs count of icmp indestunreachs @ network.icmp.intimeexcds count of icmp intimeexcds @ network.icmp.inparmprobs count of icmp inparmprobs @ network.icmp.insrcquenchs count of icmp insrcquenchs @ network.icmp.inredirects count of icmp inredirects @ network.icmp.inechos count of icmp inechos @ network.icmp.inechoreps count of icmp inechoreps @ network.icmp.intimestamps count of icmp intimestamps @ network.icmp.intimestampreps count of icmp intimestampreps @ network.icmp.inaddrmasks count of icmp inaddrmasks @ network.icmp.inaddrmaskreps count of icmp inaddrmaskreps @ network.icmp.outmsgs count of icmp outmsgs @ network.icmp.outerrors count of icmp outerrors @ network.icmp.outdestunreachs count of icmp outdestunreachs @ network.icmp.outtimeexcds count of icmp outtimeexcds @ network.icmp.outparmprobs count of icmp outparmprobs @ network.icmp.outsrcquenchs count of icmp outsrcquenchs @ network.icmp.outredirects count of icmp outredirects @ network.icmp.outechos count of icmp outechos @ network.icmp.outechoreps count of icmp outechoreps @ network.icmp.outtimestamps count of icmp outtimestamps @ network.icmp.outtimestampreps count of icmp outtimestampreps @ network.icmp.outaddrmasks count of icmp outaddrmasks @ network.icmp.outaddrmaskreps count of icmp outaddrmaskreps @ network.icmp.incsumerrors count of icmp in checksum errors @ network.icmpmsg.intype count of icmp message types recvd @ network.icmpmsg.outtype count of icmp message types sent @ network.tcp.rtoalgorithm count of tcp rtoalgorithm @ network.tcp.rtomin count of tcp rtomin @ network.tcp.rtomax count of tcp rtomax @ network.tcp.maxconn count of tcp maxconn @ network.tcp.activeopens count of tcp activeopens @ network.tcp.passiveopens count of tcp passiveopens @ network.tcp.attemptfails count of tcp attemptfails @ network.tcp.estabresets count of tcp estabresets @ network.tcp.currestab count of tcp currestab @ network.tcp.insegs count of tcp insegs @ network.tcp.outsegs count of tcp outsegs @ network.tcp.retranssegs count of tcp retranssegs @ network.tcp.inerrs count of tcp inerrs @ network.tcp.outrsts count of tcp outrsts @ network.tcp.incsumerrors count of tcp in checksum errors @ network.tcpconn.established Number of established connections @ network.tcpconn.syn_sent Number of SYN_SENT connections @ network.tcpconn.syn_recv Number of SYN_RECV connections @ network.tcpconn.fin_wait1 Number of FIN_WAIT1 connections @ network.tcpconn.fin_wait2 Number of FIN_WAIT2 connections @ network.tcpconn.time_wait Number of TIME_WAIT connections @ network.tcpconn.close Number of CLOSE connections @ network.tcpconn.close_wait Number of CLOSE_WAIT connections @ network.tcpconn.last_ack Number of LAST_ACK connections @ network.tcpconn.listen Number of LISTEN connections @ network.tcpconn.closing Number of CLOSING connections @ network.udp.indatagrams count of udp indatagrams @ network.udp.noports count of udp noports @ network.udp.inerrors count of udp inerrors @ network.udp.outdatagrams count of udp outdatagrams @ network.udp.recvbuferrors count of udp receive buffer errors @ network.udp.sndbuferrors count of udp send buffer errors @ network.udp.incsumerrors count of udp in checksum errors @ network.udplite.indatagrams count of udplite indatagrams @ network.udplite.noports count of udplite noports @ network.udplite.inerrors count of udplite inerrors @ network.udplite.outdatagrams count of udplite outdatagrams @ network.udplite.recvbuferrors count of udplite receive buffer errors @ network.udplite.sndbuferrors count of udplite send buffer errors @ network.udplite.incsumerrors count of udplite in checksum errors @ network.ip.innoroutes Number of IP datagrams discarded due to no routes in forwarding path @ network.ip.intruncatedpkts Number of IP datagrams discarded due to frame not carrying enough data @ network.ip.inmcastpkts Number of received IP multicast datagrams @ network.ip.outmcastpkts Number of sent IP multicast datagrams @ network.ip.inbcastpkts Number of received IP broadcast datagrams @ network.ip.outbcastpkts Number of sent IP bradcast datagrams @ network.ip.inoctets Number of received octets @ network.ip.outoctets Number of sent octets @ network.ip.inmcastoctets Number of received IP multicast octets @ network.ip.outmcastoctets Number of sent IP multicast octets @ network.ip.inbcastoctets Number of received IP broadcast octets @ network.ip.outbcastoctets Number of sent IP broadcast octets @ network.ip.csumerrors Number of IP datagrams with checksum errors @ network.ip.noectpkts Number of packets received with NOECT @ network.ip.ect1pkts Number of packets received with ECT(1) @ network.ip.ect0pkts Number of packets received with ECT(0) @ network.ip.cepkts Number of packets received with Congestion Experimented @ network.tcp.syncookiessent Number of sent SYN cookies @ network.tcp.syncookiesrecv Number of received SYN cookies @ network.tcp.syncookiesfailed Number of failed SYN cookies @ network.tcp.embryonicrsts Number of resets received for embryonic SYN_RECV sockets @ network.tcp.prunecalled Number of packets pruned from receive queue because of socket buffer overrun @ network.tcp.rcvpruned Number of packets pruned from receive queue @ network.tcp.ofopruned Number of packets dropped from out-of-order queue because of socket buffer overrun @ network.tcp.outofwindowicmps Number of dropped out of window ICMPs @ network.tcp.lockdroppedicmps Number of dropped ICMP because socket was locked @ network.tcp.arpfilter Number of arp packets filtered @ network.tcp.timewaited Number of TCP sockets finished time wait in fast timer @ network.tcp.timewaitrecycled Number of time wait sockets recycled by time stamp @ network.tcp.timewaitkilled Number of TCP sockets finished time wait in slow timer @ network.tcp.pawspassiverejected Number of passive connections rejected because of timestamp @ network.tcp.pawsactiverejected Number of active connections rejected because of timestamp @ network.tcp.pawsestabrejected Number of packets rejects in established connections because of timestamp @ network.tcp.delayedacks Number of delayed acks sent @ network.tcp.delayedacklocked Number of delayed acks further delayed because of locked socket @ network.tcp.delayedacklost Number of times quick ack mode was activated times @ network.tcp.listenoverflows Number of times the listen queue of a socket overflowed @ network.tcp.listendrops Number of SYNs to LISTEN sockets dropped @ network.tcp.prequeued Number of packets directly queued to recvmsg prequeue @ network.tcp.directcopyfrombacklog Number of bytes directly in process context from backlog @ network.tcp.directcopyfromprequeue Number of bytes directly received in process context from prequeue @ network.tcp.prequeueddropped Number of packets dropped from prequeue @ network.tcp.hphits Number of packet headers predicted @ network.tcp.hphitstouser Number of packets header predicted and directly queued to user @ network.tcp.pureacks Number of acknowledgments not containing data payload received @ network.tcp.hpacks Number of predicted acknowledgments @ network.tcp.renorecovery Number of times recovered from packet loss due to fast retransmit @ network.tcp.sackrecovery Number of times recovered from packet loss by selective acknowledgements @ network.tcp.sackreneging Number of bad SACK blocks received @ network.tcp.fackreorder Number of times detected reordering using FACK @ network.tcp.sackreorder Number of times detected reordering using SACK @ network.tcp.renoreorder Number of times detected reordering using reno fast retransmit @ network.tcp.tsreorder Number of times detected reordering times using time stamp @ network.tcp.fullundo Number of congestion windows fully recovered without slow start @ network.tcp.partialundo Number of congestion windows partially recovered using Hoe heuristic @ network.tcp.dsackundo Number of congestion windows recovered without slow start using DSACK @ network.tcp.lossundo Number of congestion windows recovered without slow start after partial ack @ network.tcp.lostretransmit Number of retransmits lost @ network.tcp.renofailures Number of timeouts after reno fast retransmit @ network.tcp.sackfailures Number of timeouts after SACK recovery @ network.tcp.lossfailures Number of timeouts in loss state @ network.tcp.fastretrans Number of fast retransmits @ network.tcp.forwardretrans Number of forward retransmits @ network.tcp.slowstartretrans Number of retransmits in slow start @ network.tcp.timeouts Number of other TCP timeouts @ network.tcp.lossprobes Number of sent TCP loss probes @ network.tcp.lossproberecovery Number of TCP loss probe recoveries @ network.tcp.renorecoveryfail Number of reno fast retransmits failed @ network.tcp.sackrecoveryfail Number of SACK retransmits failed @ network.tcp.schedulerfail Number of times receiver scheduled too late for direct processing @ network.tcp.rcvcollapsed Number of packets collapsed in receive queue due to low socket buffer @ network.tcp.dsackoldsent Number of DSACKs sent for old packets @ network.tcp.dsackofosent Number of DSACKs sent for out of order packets @ network.tcp.dsackrecv Number of DSACKs received @ network.tcp.dsackoforecv Number of DSACKs for out of order packets received @ network.tcp.abortondata Number of connections reset due to unexpected data @ network.tcp.abortonclose Number of connections reset due to early user close @ network.tcp.abortonmemory Number of connections aborted due to memory pressure @ network.tcp.abortontimeout Number of connections aborted due to timeout @ network.tcp.abortonlinger Number of connections aborted after user close in linger timeout @ network.tcp.abortfailed Number of times unable to send RST due to no memory @ network.tcp.memorypressures Numer of times TCP ran low on memory @ network.tcp.sackdiscard Number of SACKs discarded @ network.tcp.dsackignoredold Number of ignored old duplicate SACKs @ network.tcp.dsackignorednoundo Number of ignored duplicate SACKs with undo_marker not set @ network.tcp.spuriousrtos Number of FRTO's successfully detected spurious RTOs @ network.tcp.md5notfound Number of times MD5 hash expected but not found @ network.tcp.md5unexpected Number of times MD5 hash unexpected but found @ network.tcp.sackshifted Number of SACKs shifted @ network.tcp.sackmerged Number of SACKs merged @ network.tcp.sackshiftfallback Number of SACKs fallbacks @ network.tcp.backlogdrop Number of frames dropped because of full backlog queue @ network.tcp.minttldrop Number of frames dropped when TTL is under the minimum @ network.tcp.deferacceptdrop Number of dropped ACK frames when socket is in SYN-RECV state Due to SYNACK retrans count lower than defer_accept value @ network.tcp.iprpfilter Number of packets dropped in input path because of rp_filter settings @ network.tcp.timewaitoverflow Number of occurences of time wait bucket overflow @ network.tcp.reqqfulldocookies Number of times a SYNCOOKIE was replied to client @ network.tcp.reqqfulldrop Number of times a SYN request was dropped due to disabled syncookies @ network.tcp.retransfail Number of failed tcp_retransmit_skb() calls @ network.tcp.rcvcoalesce Number of times tried to coalesce the receive queue @ network.tcp.ofoqueue Number of packets queued in OFO queue @ network.tcp.ofodrop Number of packets meant to be queued in OFO but dropped due to limits hit Number of packets meant to be queued in OFO but dropped because socket rcvbuf limit reached. @ network.tcp.ofomerge Number of packets in OFO that were merged with other packets @ network.tcp.challengeack Number of challenge ACKs sent (RFC 5961 3.2) @ network.tcp.synchallenge Number of challenge ACKs sent in response to SYN packets @ network.tcp.fastopenactive Number of successful active fast opens @ network.tcp.fastopenactivefail Number of fast open attempts failed due to remote not accepting it or time outs @ network.tcp.fastopenpassive Number of successful passive fast opens @ network.tcp.fastopenpassivefail Number of passive fast open attempts failed @ network.tcp.fastopenlistenoverflow Number of times the fastopen listen queue overflowed @ network.tcp.fastopencookiereqd Number of fast open cookies requested @ network.tcp.spuriousrtxhostqueues Number of times that the fast clone is not yet freed in tcp_transmit_skb() @ network.tcp.busypollrxpackets Number of low latency application-fetched packets @ network.tcp.autocorking Number of times stack detected skb was underused and its flush was deferred @ network.tcp.fromzerowindowadv Number of times window went from zero to non-zero @ network.tcp.tozerowindowadv Number of times window went from non-zero to zero @ network.tcp.wantzerowindowadv Number of times zero window announced @ network.tcp.synretrans Number of SYN-SYN/ACK retransmits Number of SYN-SYN/ACK retransmits to break down retransmissions in SYN, fast/timeout retransmits. @ network.tcp.origdatasent Number of outgoing packets with original data Excluding retransmission but including data-in-SYN). This counter is different from TcpOutSegs because TcpOutSegs also tracks pure ACKs. TCPOrigDataSent is more useful to track the TCP retransmission rate. @ pmda.uname identity and type of current system Identity and type of current system. The concatenation of the values returned from utsname(2), also similar to uname -a. See also the kernel.uname.* metrics @ pmda.version build version of Linux PMDA @ hinv.map.cpu_num logical to physical CPU mapping for each CPU @ hinv.map.cpu_node logical CPU to NUMA node mapping for each CPU @ hinv.machine machine name, IP35 if SGI SNIA, else simply linux @ hinv.cpu.clock clock rate in Mhz for each CPU as reported by /proc/cpuinfo @ hinv.cpu.vendor manafacturer of each CPU as reported by /proc/cpuinfo @ hinv.cpu.model model number of each CPU as reported by /proc/cpuinfo @ hinv.cpu.model_name model name of each CPU as reported by /proc/cpuinfo @ hinv.cpu.stepping stepping of each CPU as reported by /proc/cpuinfo @ hinv.cpu.cache primary cache size of each CPU as reported by /proc/cpuinfo @ hinv.cpu.bogomips bogo mips rating for each CPU as reported by /proc/cpuinfo @ kernel.all.hz value of HZ (jiffies/second) for the currently running kernel @ kernel.all.uptime time the current kernel has been running @ kernel.all.idletime time the current kernel has been idle since boot @ kernel.all.lastpid most recently allocated process id @ kernel.all.runnable total number of processes in the (per-CPU) run queues @ kernel.all.nprocs total number of processes (lightweight) @ mem.slabinfo.objects.active number of active objects in each cache @ mem.slabinfo.objects.total total number of objects in each cache @ mem.slabinfo.objects.size size of individual objects of each cache @ mem.slabinfo.slabs.active number of active slabs comprising each cache @ mem.slabinfo.slabs.total total number of slabs comprising each cache @ mem.slabinfo.slabs.pages_per_slab number of pages in each slab @ mem.slabinfo.slabs.objects_per_slab number of objects in each slab @ mem.slabinfo.slabs.total_size total number of bytes allocated for active objects in each slab @ ipc.sem.max_semmap maximum number of entries in a semaphore map (from semctl(..,IPC_INFO,..)) @ ipc.sem.max_semid maximum number of semaphore identifiers (from semctl(..,IPC_INFO,..)) @ ipc.sem.max_sem maximum number of semaphores in system (from semctl(..,IPC_INFO,..)) @ ipc.sem.num_undo number of undo structures in system (from semctl(..,IPC_INFO,..)) @ ipc.sem.max_perid maximum number of semaphores per identifier (from semctl(..,IPC_INFO,..)) @ ipc.sem.max_ops maximum number of operations per semop call (from semctl(..,IPC_INFO,..)) @ ipc.sem.max_undoent maximum number of undo entries per process (from semctl(..,IPC_INFO,..)) @ ipc.sem.sz_semundo size of struct sem_undo (from semctl(..,IPC_INFO,..)) @ ipc.sem.max_semval semaphore maximum value (from semctl(..,IPC_INFO,..)) @ ipc.sem.max_exit adjust on exit maximum value (from semctl(..,IPC_INFO,..)) @ ipc.msg.sz_pool size of message pool in kilobytes (from msgctl(..,IPC_INFO,..)) @ ipc.msg.mapent number of entries in a message map (from msgctl(..,IPC_INFO,..)) @ ipc.msg.max_msgsz maximum size of a message in bytes (from msgctl(..,IPC_INFO,..)) @ ipc.msg.max_defmsgq default maximum size of a message queue (from msgctl(..,IPC_INFO,..)) @ ipc.msg.max_msgqid maximum number of message queue identifiers (from msgctl(..,IPC_INFO,..)) @ ipc.msg.max_msgseg message segment size (from msgctl(..,IPC_INFO,..)) @ ipc.msg.num_smsghdr number of system message headers (from msgctl(..,IPC_INFO,..)) @ ipc.msg.max_seg maximum number of message segments (from msgctl(..,IPC_INFO,..)) @ ipc.shm.max_segsz maximum shared segment size in bytes (from shmctl(..,IPC_INFO,..)) @ ipc.shm.min_segsz minimum shared segment size in bytes (from shmctl(..,IPC_INFO,..)) @ ipc.shm.max_seg maximum number of shared segments in system (from shmctl(..,IPC_INFO,..)) @ ipc.shm.max_segproc maximum number of shared segments per process (from shmctl(..,IPC_INFO,..)) @ ipc.shm.max_shmsys maximum amount of shared memory in system in pages (from shmctl(..,IPC_INFO,..)) @ vfs.files.count number of in-use file structures @ vfs.files.free number of available file structures @ vfs.files.max hard maximum on number of file structures @ vfs.inodes.count number of in-use inode structures @ vfs.inodes.free number of available inode structures @ vfs.dentry.count number of in-use dentry structures @ vfs.dentry.free number of available dentry structures @ sysfs.kernel.uevent_seqnum counter of the number of uevents processed by the udev subsystem