Created
February 14, 2017 09:04
-
-
Save indigo423/a8b15b7594bb0b68f859e36cdda2260e to your computer and use it in GitHub Desktop.
Full config diff between Horizon 18.0.4-1 and 19.0.0-1
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
diff --git a/all.policy b/all.policy | |
index 751b326..7585f22 100644 | |
--- a/all.policy | |
+++ b/all.policy | |
@@ -19,4 +19,4 @@ | |
grant { | |
permission java.security.AllPermission; | |
-}; | |
\ No newline at end of file | |
+}; | |
diff --git a/collectd-configuration.xml b/collectd-configuration.xml | |
index a43ac1e..30a8bf2 100644 | |
--- a/collectd-configuration.xml | |
+++ b/collectd-configuration.xml | |
@@ -3,37 +3,29 @@ | |
<collectd-configuration | |
threads="50"> | |
- <package name="cassandra21x"> | |
- <filter><![CDATA[(IPADDR != '0.0.0.0') & (categoryName == 'Cassandra21x')]]></filter> | |
+ <package name="cassandra-via-jmx"> | |
+ <filter>IPADDR != '0.0.0.0'</filter> | |
<service name="JMX-Cassandra" interval="300000" user-defined="false" status="on"> | |
<parameter key="port" value="7199"/> | |
<parameter key="retry" value="2"/> | |
<parameter key="timeout" value="3000"/> | |
<parameter key="protocol" value="rmi"/> | |
<parameter key="urlPath" value="/jmxrmi"/> | |
- <parameter key="rrd-base-name" value="cassandra21x"/> | |
- <parameter key="ds-name" value="cassandra21x"/> | |
- <parameter key="friendly-name" value="cassandra21x"/> | |
- <parameter key="collection" value="cassandra21x"/> | |
+ <parameter key="collection" value="jmx-cassandra30x"/> | |
+ <parameter key="friendly-name" value="cassandra"/> | |
<parameter key="thresholding-enabled" value="true"/> | |
<parameter key="factory" value="PASSWORD-CLEAR"/> | |
<parameter key="username" value="cassandra-username"/> | |
<parameter key="password" value="cassandra-password"/> | |
</service> | |
- </package> | |
- | |
- <package name="cassandra21x-newts"> | |
- <filter><![CDATA[(IPADDR != '0.0.0.0') & (catincCassandra21x& catincNewts)]]></filter> | |
<service name="JMX-Cassandra-Newts" interval="300000" user-defined="false" status="on"> | |
<parameter key="port" value="7199"/> | |
<parameter key="retry" value="2"/> | |
<parameter key="timeout" value="3000"/> | |
<parameter key="protocol" value="rmi"/> | |
<parameter key="urlPath" value="/jmxrmi"/> | |
- <parameter key="rrd-base-name" value="cassandra21x-newts"/> | |
- <parameter key="ds-name" value="cassandra21x-newts"/> | |
- <parameter key="friendly-name" value="cassandra21x-newts"/> | |
- <parameter key="collection" value="cassandra21x-newts"/> | |
+ <parameter key="collection" value="jmx-cassandra30x-newts"/> | |
+ <parameter key="friendly-name" value="cassandra-newts"/> | |
<parameter key="thresholding-enabled" value="true"/> | |
<parameter key="factory" value="PASSWORD-CLEAR"/> | |
<parameter key="username" value="cassandra-username"/> | |
@@ -91,7 +83,22 @@ | |
<parameter key="thresholding-enabled" value="true"/> | |
</service> | |
</package> | |
+ <package name="vmware6"> | |
+ <filter><![CDATA[(IPADDR != '0.0.0.0') & (categoryName == 'VMware6')]]></filter> | |
+ <service name="VMware-VirtualMachine" interval="300000" user-defined="false" status="on"> | |
+ <parameter key="collection" value="default-VirtualMachine6"/> | |
+ <parameter key="thresholding-enabled" value="true"/> | |
+ </service> | |
+ <service name="VMware-HostSystem" interval="300000" user-defined="false" status="on"> | |
+ <parameter key="collection" value="default-HostSystem6"/> | |
+ <parameter key="thresholding-enabled" value="true"/> | |
+ </service> | |
+ <service name="VMwareCim-HostSystem" interval="300000" user-defined="false" status="on"> | |
+ <parameter key="collection" value="default-ESX-HostSystem"/> | |
+ <parameter key="thresholding-enabled" value="true"/> | |
+ </service> | |
+ </package> | |
<package name="example1"> | |
<filter>IPADDR != '0.0.0.0'</filter> | |
<include-range begin="1.1.1.1" end="254.254.254.254"/> | |
@@ -123,6 +130,32 @@ | |
<parameter key="friendly-name" value="opennms-jvm"/> | |
</service> | |
+ <service name="JMX-Minion" interval="300000" user-defined="false" status="on"> | |
+ <parameter key="port" value="1299"/> | |
+ <parameter key="retry" value="2"/> | |
+ <parameter key="timeout" value="3000"/> | |
+ <parameter key="urlPath" value="/karaf-minion"/> | |
+ <parameter key="factory" value="PASSWORD-CLEAR"/> | |
+ <parameter key="username" value="admin"/> | |
+ <parameter key="password" value="admin"/> | |
+ <parameter key="rrd-base-name" value="java"/> | |
+ <parameter key="collection" value="jmx-minion"/> | |
+ <parameter key="thresholding-enabled" value="true"/> | |
+ <parameter key="ds-name" value="jmx-minion"/> | |
+ <parameter key="friendly-name" value="jmx-minion"/> | |
+ </service> | |
+ | |
+ <service name="JMX-Kafka" interval="300000" user-defined="false" status="on"> | |
+ <parameter key="port" value="9999"/> | |
+ <parameter key="retry" value="2"/> | |
+ <parameter key="timeout" value="3000"/> | |
+ <parameter key="rrd-base-name" value="java"/> | |
+ <parameter key="collection" value="jmx-kafka"/> | |
+ <parameter key="thresholding-enabled" value="true"/> | |
+ <parameter key="ds-name" value="jmx-kafka"/> | |
+ <parameter key="friendly-name" value="jmx-kafka"/> | |
+ </service> | |
+ | |
<service name="PostgreSQL" interval="300000" user-defined="false" status="on"> | |
<parameter key="collection" value="PostgreSQL"/> | |
<parameter key="thresholding-enabled" value="true"/> | |
@@ -131,18 +164,26 @@ | |
<parameter key="password" value="postgres"/> | |
<parameter key="url" value="jdbc:postgresql://OPENNMS_JDBC_HOSTNAME:5432/opennms"/> | |
</service> | |
+ | |
+ <service name="Elasticsearch" interval="300000" user-defined="false" status="on"> | |
+ <parameter key="collection" value="xml-elasticsearch-cluster-stats" /> | |
+ <parameter key="handler-class" value="org.opennms.protocols.json.collector.DefaultJsonCollectionHandler"/> | |
+ </service> | |
</package> | |
- <collector service="JMX-Cassandra" class-name="org.opennms.netmgt.collectd.Jsr160Collector"/> | |
- <collector service="JMX-Cassandra-Newts" class-name="org.opennms.netmgt.collectd.Jsr160Collector"/> | |
+ <collector service="Elasticsearch" class-name="org.opennms.protocols.xml.collector.XmlCollector"/> | |
+ <collector service="PostgreSQL" class-name="org.opennms.netmgt.collectd.JdbcCollector"/> | |
<collector service="SNMP" class-name="org.opennms.netmgt.collectd.SnmpCollector"/> | |
<collector service="WMI" class-name="org.opennms.netmgt.collectd.WmiCollector"/> | |
<collector service="WS-Man" class-name="org.opennms.netmgt.collectd.WsManCollector"/> | |
- <collector service="OpenNMS-JVM" class-name="org.opennms.netmgt.collectd.Jsr160Collector"/> | |
<collector service="VMware-VirtualMachine" class-name="org.opennms.netmgt.collectd.VmwareCollector"/> | |
<collector service="VMware-HostSystem" class-name="org.opennms.netmgt.collectd.VmwareCollector"/> | |
<collector service="VMwareCim-HostSystem" class-name="org.opennms.netmgt.collectd.VmwareCimCollector"/> | |
- <collector service="PostgreSQL" class-name="org.opennms.netmgt.collectd.JdbcCollector"/> | |
-</collectd-configuration> | |
+ <collector service="OpenNMS-JVM" class-name="org.opennms.netmgt.collectd.Jsr160Collector"/> | |
+ <collector service="JMX-Minion" class-name="org.opennms.netmgt.collectd.Jsr160Collector"/> | |
+ <collector service="JMX-Cassandra" class-name="org.opennms.netmgt.collectd.Jsr160Collector"/> | |
+ <collector service="JMX-Cassandra-Newts" class-name="org.opennms.netmgt.collectd.Jsr160Collector"/> | |
+ <collector service="JMX-Kafka" class-name="org.opennms.netmgt.collectd.Jsr160Collector"/> | |
+</collectd-configuration> | |
diff --git a/config.properties b/config.properties | |
index 19f97c7..22f7b9b 100644 | |
--- a/config.properties | |
+++ b/config.properties | |
@@ -71,11 +71,10 @@ org.osgi.framework.system.packages= \ | |
org.osgi.service.packageadmin;uses:="org.osgi.framework";version="1.2",\ | |
org.osgi.service.url;version="1.0", \ | |
org.osgi.util.tracker;uses:="org.osgi.framework";version="1.5.1", \ | |
- org.apache.karaf.jaas.boot;version="2.4.0", \ | |
- org.apache.karaf.jaas.boot.principal;version="2.4.0", \ | |
- org.apache.karaf.management.boot;version="2.4.0", \ | |
- org.apache.karaf.version;version="2.4.0", \ | |
- org.apache.karaf.diagnostic.core;version="2.4.0", \ | |
+ org.apache.karaf.jaas.boot;version="2.4.3", \ | |
+ org.apache.karaf.jaas.boot.principal;version="2.4.3", \ | |
+ org.apache.karaf.version;version="2.4.3", \ | |
+ org.apache.karaf.diagnostic.core;version="2.4.3", \ | |
${jre-${java.specification.version}} | |
# | |
@@ -146,19 +145,19 @@ eecap-1.2= osgi.ee; osgi.ee="OSGi/Minimum"; version:List<Version>="1.0,1.1", \ | |
# | |
# javax.transaction is needed to avoid class loader constraint violation when using javax.sql | |
# | |
-org.osgi.framework.bootdelegation=org.apache.karaf.jaas.boot,org.apache.karaf.management.boot,sun.*,com.sun.*,javax.transaction,javax.transaction.*,javax.xml.crypto,javax.xml.crypto.*,org.apache.xerces.jaxp.datatype,org.apache.xerces.stax,org.apache.xerces.parsers,org.apache.xerces.jaxp,org.apache.xerces.jaxp.validation,org.apache.xerces.dom | |
+org.osgi.framework.bootdelegation=org.apache.karaf.jaas.boot,sun.*,com.sun.*,javax.transaction,javax.transaction.*,javax.xml.crypto,javax.xml.crypto.*,org.apache.xerces.jaxp.datatype,org.apache.xerces.stax,org.apache.xerces.parsers,org.apache.xerces.jaxp,org.apache.xerces.jaxp.validation,org.apache.xerces.dom | |
# jVisualVM support | |
# in order to use Karaf with jvisualvm, the org.osgi.framework.bootdelegation property has to contain the org.netbeans.lib.profiler.server package | |
# and, so, it should look like: | |
# | |
-# org.osgi.framework.bootdelegation=org.apache.karaf.jaas.boot,org.apache.karaf.jaas.boot.principal,org.apache.karaf.management.boot,sun.*,com.sun.*,javax.transaction,javax.transaction.*,javax.xml.crypto,javax.xml.crypto.*,org.apache.xerces.jaxp.datatype,org.apache.xerces.stax,org.apache.xerces.parsers,org.apache.xerces.jaxp,org.apache.xerces.jaxp.validation,org.apache.xerces.dom,org.netbeans.lib.profiler.server | |
+# org.osgi.framework.bootdelegation=org.apache.karaf.jaas.boot,org.apache.karaf.jaas.boot.principal,sun.*,com.sun.*,javax.transaction,javax.transaction.*,javax.xml.crypto,javax.xml.crypto.*,org.apache.xerces.jaxp.datatype,org.apache.xerces.stax,org.apache.xerces.parsers,org.apache.xerces.jaxp,org.apache.xerces.jaxp.validation,org.apache.xerces.dom,org.netbeans.lib.profiler.server | |
# | |
# YourKit support | |
# in order to use Karaf with YourKit, the org.osgi.framework.bootdelegation property has to contain the com.yourkit.* packages | |
# and, so, it should look like: | |
# | |
-# org.osgi.framework.bootdelegation=org.apache.karaf.jaas.boot,org.apache.karaf.jaas.boot.principal,org.apache.karaf.management.boot,sun.*,com.sun.*,javax.transaction,javax.transaction.*,javax.xml.crypto,javax.xml.crypto.*,org.apache.xerces.jaxp.datatype,org.apache.xerces.stax,org.apache.xerces.parsers,org.apache.xerces.jaxp,org.apache.xerces.jaxp.validation,org.apache.xerces.dom,com.yourkit.* | |
+# org.osgi.framework.bootdelegation=org.apache.karaf.jaas.boot,org.apache.karaf.jaas.boot.principal,sun.*,com.sun.*,javax.transaction,javax.transaction.*,javax.xml.crypto,javax.xml.crypto.*,org.apache.xerces.jaxp.datatype,org.apache.xerces.stax,org.apache.xerces.parsers,org.apache.xerces.jaxp,org.apache.xerces.jaxp.validation,org.apache.xerces.dom,com.yourkit.* | |
# | |
# | |
@@ -217,4 +216,4 @@ karaf.delay.console=false | |
# | |
# Set the Blueprint container in synchronous mode to avoid dependencies startup issue | |
# | |
-org.apache.aries.blueprint.synchronous=true | |
\ No newline at end of file | |
+org.apache.aries.blueprint.synchronous=true | |
diff --git a/create.sql b/create.sql | |
index 9cf2f5f..4ee24a8 100644 | |
--- a/create.sql | |
+++ b/create.sql | |
@@ -246,7 +246,7 @@ CREATE TABLE monitoringlocationspollingpackages ( | |
monitoringlocationid TEXT NOT NULL, | |
packagename TEXT NOT NULL, | |
- CONSTRAINT monitoringlocationspollingpackages_fkey FOREIGN KEY (monitoringlocationid) REFERENCES monitoringlocations (id) ON DELETE CASCADE | |
+ CONSTRAINT monitoringlocationspollingpackages_fkey FOREIGN KEY (monitoringlocationid) REFERENCES monitoringlocations (id) ON DELETE CASCADE ON UPDATE CASCADE | |
); | |
CREATE INDEX monitoringlocationspollingpackages_id_idx on monitoringlocationspollingpackages(monitoringlocationid); | |
@@ -257,7 +257,7 @@ CREATE TABLE monitoringlocationscollectionpackages ( | |
monitoringlocationid TEXT NOT NULL, | |
packagename TEXT NOT NULL, | |
- CONSTRAINT monitoringlocationscollectionpackages_fkey FOREIGN KEY (monitoringlocationid) REFERENCES monitoringlocations (id) ON DELETE CASCADE | |
+ CONSTRAINT monitoringlocationscollectionpackages_fkey FOREIGN KEY (monitoringlocationid) REFERENCES monitoringlocations (id) ON DELETE CASCADE ON UPDATE CASCADE | |
); | |
CREATE INDEX monitoringlocationscollectionpackages_id_idx on monitoringlocationscollectionpackages(monitoringlocationid); | |
@@ -268,12 +268,18 @@ CREATE TABLE monitoringlocationstags ( | |
monitoringlocationid TEXT NOT NULL, | |
tag TEXT NOT NULL, | |
- CONSTRAINT monitoringlocationstags_fkey FOREIGN KEY (monitoringlocationid) REFERENCES monitoringlocations (id) ON DELETE CASCADE | |
+ CONSTRAINT monitoringlocationstags_fkey FOREIGN KEY (monitoringlocationid) REFERENCES monitoringlocations (id) ON DELETE CASCADE ON UPDATE CASCADE | |
); | |
CREATE INDEX monitoringlocationstags_id_idx on monitoringlocationstags(monitoringlocationid); | |
CREATE UNIQUE INDEX monitoringlocationstags_id_pkg_idx on monitoringlocationstags(monitoringlocationid, tag); | |
+--################################################################## | |
+--# The following command adds the initial 'Default' entry to | |
+--# the 'monitoringlocations' table. | |
+--################################################################## | |
+INSERT INTO monitoringlocations (id, monitoringarea) values ('Default', 'Default'); | |
+ | |
--##################################################### | |
--# monitoringsystems Table - Contains a list of OpenNMS systems | |
@@ -317,7 +323,7 @@ CREATE UNIQUE INDEX monitoringsystemsproperties_id_property_idx on monitoringsys | |
--# The following command adds the initial localhost poller entry to | |
--# the 'monitoringsystems' table. | |
--################################################################## | |
-INSERT INTO monitoringsystems (id, label, location, type) values ('00000000-0000-0000-0000-000000000000', 'localhost', 'localhost', 'OpenNMS'); | |
+INSERT INTO monitoringsystems (id, label, location, type) values ('00000000-0000-0000-0000-000000000000', 'localhost', 'Default', 'OpenNMS'); | |
--##################################################### | |
@@ -339,7 +345,7 @@ CREATE TABLE scanreports ( | |
timestamp TIMESTAMP WITH TIME ZONE, | |
CONSTRAINT scanreports_pkey PRIMARY KEY (id), | |
- CONSTRAINT scanreports_monitoringlocations_fkey FOREIGN KEY (location) REFERENCES monitoringlocations (id) ON DELETE CASCADE | |
+ CONSTRAINT scanreports_monitoringlocations_fkey FOREIGN KEY (location) REFERENCES monitoringlocations (id) ON DELETE CASCADE ON UPDATE CASCADE | |
); | |
CREATE UNIQUE INDEX scanreports_id_idx on scanreport(id); | |
@@ -484,8 +490,10 @@ create table node ( | |
lastCapsdPoll timestamp with time zone, | |
foreignSource varchar(64), | |
foreignId varchar(64), | |
+ location text not null, | |
- constraint pk_nodeID primary key (nodeID) | |
+ constraint pk_nodeID primary key (nodeID), | |
+ constraint fk_node_location foreign key (location) references monitoringlocations (id) ON DELETE CASCADE ON UPDATE CASCADE | |
); | |
create index node_id_type_idx on node(nodeID, nodeType); | |
@@ -2350,6 +2358,7 @@ CREATE TABLE bsm_reduce ( | |
type character varying(32) NOT NULL, | |
threshold float, | |
threshold_severity integer, | |
+ base float, | |
CONSTRAINT bsm_reduce_pkey PRIMARY KEY (id) | |
); | |
@@ -2419,3 +2428,38 @@ CREATE TABLE bsm_service_children ( | |
CONSTRAINT fk_bsm_service_child_service_id FOREIGN KEY (bsm_service_child_id) | |
REFERENCES bsm_service (id) ON DELETE CASCADE | |
); | |
+ | |
+--################################################################## | |
+--# Topology tables | |
+--################################################################## | |
+ | |
+-- Layout table | |
+CREATE TABLE topo_layout ( | |
+ id varchar(255) NOT NULL, | |
+ created timestamp NOT NULL, | |
+ creator varchar(255) NOT NULL, | |
+ updated timestamp NOT NULL, | |
+ updator varchar(255) NOT NULL, | |
+ last_used timestamp, | |
+ CONSTRAINT topo_layout_pkey PRIMARY KEY (id) | |
+); | |
+ | |
+-- Layout coordinates of vertex | |
+CREATE TABLE topo_vertex_position ( | |
+ id integer NOT NULL, | |
+ x integer NOT NULL, | |
+ y integer NOT NULL, | |
+ vertex_namespace varchar(255) NULL, | |
+ vertex_id varchar(255) NULL, | |
+ CONSTRAINT topo_vertex_position_pkey PRIMARY KEY (id) | |
+); | |
+ | |
+-- Relation table (layout -> vertex positions) | |
+CREATE TABLE topo_layout_vertex_positions ( | |
+ vertex_position_id integer NOT NULL, | |
+ layout_id varchar(255) NOT NULL, | |
+ CONSTRAINT fk_topo_layout_vertex_positions_layout_id FOREIGN KEY (layout_id) | |
+ REFERENCES topo_layout (id) ON DELETE CASCADE, | |
+ CONSTRAINT fk_topo_layout_vertex_positions_vertex_position_id FOREIGN KEY (vertex_position_id) | |
+ REFERENCES topo_vertex_position (id) ON DELETE CASCADE | |
+); | |
\ No newline at end of file | |
diff --git a/custom.properties b/custom.properties | |
index 1a5cfb4..11b5df6 100644 | |
--- a/custom.properties | |
+++ b/custom.properties | |
@@ -32,7 +32,7 @@ org.osgi.framework.system.packages.extra=org.apache.karaf.branding,\ | |
sun.misc,\ | |
sun.net.spi.nameservice,\ | |
javax.jms;version=1.1.0,\ | |
- javax.servlet;javax.servlet.annotation;javax.servlet.descriptor;javax.servlet.http;javax.servlet.resources;version=2.6,\ | |
+ javax.servlet;javax.servlet.annotation;javax.servlet.descriptor;javax.servlet.http;javax.servlet.resources;version=3.1.0,\ | |
javax.persistence;version=2.0,\ | |
javax.validation.constraints,\ | |
javax.ws.rs;javax.ws.rs.client;javax.ws.rs.container;javax.ws.rs.core;javax.ws.rs.ext;version=2.1,\ | |
@@ -45,27 +45,21 @@ org.osgi.framework.system.packages.extra=org.apache.karaf.branding,\ | |
javax.wsdl.factory;version=1.6.3,\ | |
javax.wsdl.xml;version=1.6.3,\ | |
javax.wsdl;version=1.6.3,\ | |
- javax.xml.bind;version=2.1.7,\ | |
- javax.xml.bind.annotation;version=2.1.7,\ | |
- javax.xml.bind.annotation.adapters;version=2.1.7,\ | |
- javax.xml.bind.attachment;version=2.1.7,\ | |
- javax.xml.bind.helpers;version=2.1.7,\ | |
- javax.xml.bind.util;version=2.1.7,\ | |
antlr;version=2.7.7,\ | |
antlr.collections.impl;version=2.7.7,\ | |
com.codahale.metrics;version=3.1.2,\ | |
- com.google.common.annotations;version=17.0,\ | |
- com.google.common.base;version=17.0,\ | |
- com.google.common.cache;version=17.0,\ | |
- com.google.common.collect;version=17.0,\ | |
- com.google.common.eventbus;version=17.0,\ | |
- com.google.common.hash;version=17.0,\ | |
- com.google.common.io;version=17.0,\ | |
- com.google.common.math;version=17.0,\ | |
- com.google.common.net;version=17.0,\ | |
- com.google.common.primitives;version=17.0,\ | |
- com.google.common.reflect;version=17.0,\ | |
- com.google.common.util.concurrent;version=17.0,\ | |
+ com.google.common.annotations;version=18.0,\ | |
+ com.google.common.base;version=18.0,\ | |
+ com.google.common.cache;version=18.0,\ | |
+ com.google.common.collect;version=18.0,\ | |
+ com.google.common.eventbus;version=18.0,\ | |
+ com.google.common.hash;version=18.0,\ | |
+ com.google.common.io;version=18.0,\ | |
+ com.google.common.math;version=18.0,\ | |
+ com.google.common.net;version=18.0,\ | |
+ com.google.common.primitives;version=18.0,\ | |
+ com.google.common.reflect;version=18.0,\ | |
+ com.google.common.util.concurrent;version=18.0,\ | |
com.ibm.wsdl.extensions.http;version=1.6.3,\ | |
com.ibm.wsdl.extensions.mime;version=1.6.3,\ | |
com.ibm.wsdl.extensions.schema;version=1.6.3,\ | |
@@ -83,12 +77,12 @@ org.osgi.framework.system.packages.extra=org.apache.karaf.branding,\ | |
freemarker.template;version=2.3.21,\ | |
freemarker.template.utility;version=2.3.21,\ | |
org.apache.commons.beanutils;version=1.8.3,\ | |
- org.apache.commons.codec;version=1.9,\ | |
- org.apache.commons.codec.binary;version=1.9,\ | |
- org.apache.commons.codec.digest;version=1.9,\ | |
- org.apache.commons.codec.language;version=1.9,\ | |
- org.apache.commons.codec.language.bm;version=1.9,\ | |
- org.apache.commons.codec.net;version=1.9,\ | |
+ org.apache.commons.codec;version=1.10,\ | |
+ org.apache.commons.codec.binary;version=1.10,\ | |
+ org.apache.commons.codec.digest;version=1.10,\ | |
+ org.apache.commons.codec.language;version=1.10,\ | |
+ org.apache.commons.codec.language.bm;version=1.10,\ | |
+ org.apache.commons.codec.net;version=1.10,\ | |
org.apache.commons.collections;version=3.2.2,\ | |
org.apache.commons.collections.comparators;version=3.2.2,\ | |
org.apache.commons.collections.keyvalue;version=3.2.2,\ | |
@@ -116,97 +110,97 @@ org.osgi.framework.system.packages.extra=org.apache.karaf.branding,\ | |
org.apache.commons.lang.reflect;version=2.6,\ | |
org.apache.commons.lang.text;version=2.6,\ | |
org.apache.commons.lang.time;version=2.6,\ | |
- org.apache.cxf.annotations;version=3.1.5,\ | |
- org.apache.cxf.attachment;version=3.1.5,\ | |
- org.apache.cxf.binding;version=3.1.5,\ | |
- org.apache.cxf.bus.blueprint;version=3.1.5,\ | |
- org.apache.cxf.bus.extension;version=3.1.5,\ | |
- org.apache.cxf.bus.managers;version=3.1.5,\ | |
- org.apache.cxf.bus.osgi;version=3.1.5,\ | |
- org.apache.cxf.bus.resource;version=3.1.5,\ | |
- org.apache.cxf.bus.spring;version=3.1.5,\ | |
- org.apache.cxf.bus;version=3.1.5,\ | |
- org.apache.cxf.buslifecycle;version=3.1.5,\ | |
- org.apache.cxf.catalog;version=3.1.5,\ | |
- org.apache.cxf.common.annotation;version=3.1.5,\ | |
- org.apache.cxf.common.classloader;version=3.1.5,\ | |
- org.apache.cxf.common.i18n;version=3.1.5,\ | |
- org.apache.cxf.common.injection;version=3.1.5,\ | |
- org.apache.cxf.common.jaxb;version=3.1.5,\ | |
- org.apache.cxf.common.logging;version=3.1.5,\ | |
- org.apache.cxf.common.security;version=3.1.5,\ | |
- org.apache.cxf.common.util;version=3.1.5,\ | |
- org.apache.cxf.common.xmlschema;version=3.1.5,\ | |
- org.apache.cxf.common;version=3.1.5,\ | |
- org.apache.cxf.configuration.blueprint;version=3.1.5,\ | |
- org.apache.cxf.configuration.jsse;version=3.1.5,\ | |
- org.apache.cxf.configuration.security;version=3.1.5,\ | |
- org.apache.cxf.configuration.spring;version=3.1.5,\ | |
- org.apache.cxf.configuration;version=3.1.5,\ | |
- org.apache.cxf.continuations;version=3.1.5,\ | |
- org.apache.cxf.databinding.source.mime;version=3.1.5,\ | |
- org.apache.cxf.databinding.source;version=3.1.5,\ | |
- org.apache.cxf.databinding.stax;version=3.1.5,\ | |
- org.apache.cxf.databinding;version=3.1.5,\ | |
- org.apache.cxf.endpoint;version=3.1.5,\ | |
- org.apache.cxf.extension;version=3.1.5,\ | |
- org.apache.cxf.feature.transform;version=3.1.5,\ | |
- org.apache.cxf.feature.validation;version=3.1.5,\ | |
- org.apache.cxf.feature;version=3.1.5,\ | |
- org.apache.cxf.headers;version=3.1.5,\ | |
- org.apache.cxf.helpers;version=3.1.5,\ | |
- org.apache.cxf.interceptor.security.callback;version=3.1.5,\ | |
- org.apache.cxf.interceptor.security;version=3.1.5,\ | |
- org.apache.cxf.interceptor.transform;version=3.1.5,\ | |
- org.apache.cxf.interceptor;version=3.1.5,\ | |
- org.apache.cxf.io;version=3.1.5,\ | |
- org.apache.cxf.logging;version=3.1.5,\ | |
- org.apache.cxf.management.annotation;version=3.1.5,\ | |
- org.apache.cxf.management;version=3.1.5,\ | |
- org.apache.cxf.message;version=3.1.5,\ | |
- org.apache.cxf.phase;version=3.1.5,\ | |
- org.apache.cxf.policy;version=3.1.5,\ | |
- org.apache.cxf.resource;version=3.1.5,\ | |
- org.apache.cxf.security.claims.authorization;version=3.1.5,\ | |
- org.apache.cxf.security.transport;version=3.1.5,\ | |
- org.apache.cxf.security;version=3.1.5,\ | |
- org.apache.cxf.service.factory;version=3.1.5,\ | |
- org.apache.cxf.service.invoker.spring;version=3.1.5,\ | |
- org.apache.cxf.service.invoker;version=3.1.5,\ | |
- org.apache.cxf.service.model;version=3.1.5,\ | |
- org.apache.cxf.service;version=3.1.5,\ | |
- org.apache.cxf.staxutils.transform;version=3.1.5,\ | |
- org.apache.cxf.staxutils.validation;version=3.1.5,\ | |
- org.apache.cxf.staxutils;version=3.1.5,\ | |
- org.apache.cxf.transport.common.gzip;version=3.1.5,\ | |
- org.apache.cxf.transport.http.auth;version=3.1.5,\ | |
- org.apache.cxf.transport.http.blueprint;version=3.1.5,\ | |
- org.apache.cxf.transport.http.osgi;version=3.1.5,\ | |
- org.apache.cxf.transport.http.policy.impl;version=3.1.5,\ | |
- org.apache.cxf.transport.http.policy;version=3.1.5,\ | |
- org.apache.cxf.transport.http.spring;version=3.1.5,\ | |
- org.apache.cxf.transport.http;version=3.1.5,\ | |
- org.apache.cxf.transport.https.httpclient;version=3.1.5,\ | |
- org.apache.cxf.transport.https;version=3.1.5,\ | |
- org.apache.cxf.transport.servlet.blueprint;version=3.1.5,\ | |
- org.apache.cxf.transport.servlet.servicelist;version=3.1.5,\ | |
- org.apache.cxf.transport.servlet;version=3.1.5,\ | |
- org.apache.cxf.transport;version=3.1.5,\ | |
- org.apache.cxf.transports.http.configuration;version=3.1.5,\ | |
- org.apache.cxf.validation;version=3.1.5,\ | |
- org.apache.cxf.version;version=3.1.5,\ | |
- org.apache.cxf.workqueue;version=3.1.5,\ | |
- org.apache.cxf.ws.addressing.v200403;version=3.1.5,\ | |
- org.apache.cxf.ws.addressing.v200408;version=3.1.5,\ | |
- org.apache.cxf.ws.addressing.wsdl;version=3.1.5,\ | |
- org.apache.cxf.ws.addressing;version=3.1.5,\ | |
- org.apache.cxf.wsdl.binding;version=3.1.5,\ | |
- org.apache.cxf.wsdl.http;version=3.1.5,\ | |
- org.apache.cxf.wsdl.interceptors;version=3.1.5,\ | |
- org.apache.cxf.wsdl.service.factory;version=3.1.5,\ | |
- org.apache.cxf.wsdl11;version=3.1.5,\ | |
- org.apache.cxf.wsdl;version=3.1.5,\ | |
- org.apache.cxf;version=3.1.5,\ | |
+ org.apache.cxf.annotations;version=3.1.7,\ | |
+ org.apache.cxf.attachment;version=3.1.7,\ | |
+ org.apache.cxf.binding;version=3.1.7,\ | |
+ org.apache.cxf.bus.blueprint;version=3.1.7,\ | |
+ org.apache.cxf.bus.extension;version=3.1.7,\ | |
+ org.apache.cxf.bus.managers;version=3.1.7,\ | |
+ org.apache.cxf.bus.osgi;version=3.1.7,\ | |
+ org.apache.cxf.bus.resource;version=3.1.7,\ | |
+ org.apache.cxf.bus.spring;version=3.1.7,\ | |
+ org.apache.cxf.bus;version=3.1.7,\ | |
+ org.apache.cxf.buslifecycle;version=3.1.7,\ | |
+ org.apache.cxf.catalog;version=3.1.7,\ | |
+ org.apache.cxf.common.annotation;version=3.1.7,\ | |
+ org.apache.cxf.common.classloader;version=3.1.7,\ | |
+ org.apache.cxf.common.i18n;version=3.1.7,\ | |
+ org.apache.cxf.common.injection;version=3.1.7,\ | |
+ org.apache.cxf.common.jaxb;version=3.1.7,\ | |
+ org.apache.cxf.common.logging;version=3.1.7,\ | |
+ org.apache.cxf.common.security;version=3.1.7,\ | |
+ org.apache.cxf.common.util;version=3.1.7,\ | |
+ org.apache.cxf.common.xmlschema;version=3.1.7,\ | |
+ org.apache.cxf.common;version=3.1.7,\ | |
+ org.apache.cxf.configuration.blueprint;version=3.1.7,\ | |
+ org.apache.cxf.configuration.jsse;version=3.1.7,\ | |
+ org.apache.cxf.configuration.security;version=3.1.7,\ | |
+ org.apache.cxf.configuration.spring;version=3.1.7,\ | |
+ org.apache.cxf.configuration;version=3.1.7,\ | |
+ org.apache.cxf.continuations;version=3.1.7,\ | |
+ org.apache.cxf.databinding.source.mime;version=3.1.7,\ | |
+ org.apache.cxf.databinding.source;version=3.1.7,\ | |
+ org.apache.cxf.databinding.stax;version=3.1.7,\ | |
+ org.apache.cxf.databinding;version=3.1.7,\ | |
+ org.apache.cxf.endpoint;version=3.1.7,\ | |
+ org.apache.cxf.extension;version=3.1.7,\ | |
+ org.apache.cxf.feature.transform;version=3.1.7,\ | |
+ org.apache.cxf.feature.validation;version=3.1.7,\ | |
+ org.apache.cxf.feature;version=3.1.7,\ | |
+ org.apache.cxf.headers;version=3.1.7,\ | |
+ org.apache.cxf.helpers;version=3.1.7,\ | |
+ org.apache.cxf.interceptor.security.callback;version=3.1.7,\ | |
+ org.apache.cxf.interceptor.security;version=3.1.7,\ | |
+ org.apache.cxf.interceptor.transform;version=3.1.7,\ | |
+ org.apache.cxf.interceptor;version=3.1.7,\ | |
+ org.apache.cxf.io;version=3.1.7,\ | |
+ org.apache.cxf.logging;version=3.1.7,\ | |
+ org.apache.cxf.management.annotation;version=3.1.7,\ | |
+ org.apache.cxf.management;version=3.1.7,\ | |
+ org.apache.cxf.message;version=3.1.7,\ | |
+ org.apache.cxf.phase;version=3.1.7,\ | |
+ org.apache.cxf.policy;version=3.1.7,\ | |
+ org.apache.cxf.resource;version=3.1.7,\ | |
+ org.apache.cxf.security.claims.authorization;version=3.1.7,\ | |
+ org.apache.cxf.security.transport;version=3.1.7,\ | |
+ org.apache.cxf.security;version=3.1.7,\ | |
+ org.apache.cxf.service.factory;version=3.1.7,\ | |
+ org.apache.cxf.service.invoker.spring;version=3.1.7,\ | |
+ org.apache.cxf.service.invoker;version=3.1.7,\ | |
+ org.apache.cxf.service.model;version=3.1.7,\ | |
+ org.apache.cxf.service;version=3.1.7,\ | |
+ org.apache.cxf.staxutils.transform;version=3.1.7,\ | |
+ org.apache.cxf.staxutils.validation;version=3.1.7,\ | |
+ org.apache.cxf.staxutils;version=3.1.7,\ | |
+ org.apache.cxf.transport.common.gzip;version=3.1.7,\ | |
+ org.apache.cxf.transport.http.auth;version=3.1.7,\ | |
+ org.apache.cxf.transport.http.blueprint;version=3.1.7,\ | |
+ org.apache.cxf.transport.http.osgi;version=3.1.7,\ | |
+ org.apache.cxf.transport.http.policy.impl;version=3.1.7,\ | |
+ org.apache.cxf.transport.http.policy;version=3.1.7,\ | |
+ org.apache.cxf.transport.http.spring;version=3.1.7,\ | |
+ org.apache.cxf.transport.http;version=3.1.7,\ | |
+ org.apache.cxf.transport.https.httpclient;version=3.1.7,\ | |
+ org.apache.cxf.transport.https;version=3.1.7,\ | |
+ org.apache.cxf.transport.servlet.blueprint;version=3.1.7,\ | |
+ org.apache.cxf.transport.servlet.servicelist;version=3.1.7,\ | |
+ org.apache.cxf.transport.servlet;version=3.1.7,\ | |
+ org.apache.cxf.transport;version=3.1.7,\ | |
+ org.apache.cxf.transports.http.configuration;version=3.1.7,\ | |
+ org.apache.cxf.validation;version=3.1.7,\ | |
+ org.apache.cxf.version;version=3.1.7,\ | |
+ org.apache.cxf.workqueue;version=3.1.7,\ | |
+ org.apache.cxf.ws.addressing.v200403;version=3.1.7,\ | |
+ org.apache.cxf.ws.addressing.v200408;version=3.1.7,\ | |
+ org.apache.cxf.ws.addressing.wsdl;version=3.1.7,\ | |
+ org.apache.cxf.ws.addressing;version=3.1.7,\ | |
+ org.apache.cxf.wsdl.binding;version=3.1.7,\ | |
+ org.apache.cxf.wsdl.http;version=3.1.7,\ | |
+ org.apache.cxf.wsdl.interceptors;version=3.1.7,\ | |
+ org.apache.cxf.wsdl.service.factory;version=3.1.7,\ | |
+ org.apache.cxf.wsdl11;version=3.1.7,\ | |
+ org.apache.cxf.wsdl;version=3.1.7,\ | |
+ org.apache.cxf;version=3.1.7,\ | |
org.apache.http;version=4.3.3,\ | |
org.apache.http.annotation;version=4.3.3,\ | |
org.apache.http.concurrent;version=4.3.3,\ | |
@@ -311,285 +305,302 @@ org.osgi.framework.system.packages.extra=org.apache.karaf.branding,\ | |
org.joda.time.format;version=2.1,\ | |
org.owasp.encoder;version=1.2,\ | |
org.quartz;version=1.6.5,\ | |
- org.springframework;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop.aspectj;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop.aspectj.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop.aspectj.autoproxy;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop.config;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop.framework;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop.framework.adapter;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop.framework.autoproxy;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop.framework.autoproxy.target;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop.interceptor;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop.scope;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop.support.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop.target;version=4.0.5.RELEASE_1,\ | |
- org.springframework.aop.target.dynamic;version=4.0.5.RELEASE_1,\ | |
- org.springframework.asm;version=4.0.5.RELEASE_1,\ | |
- org.springframework.asm.commons;version=4.0.5.RELEASE_1,\ | |
- org.springframework.asm.signature;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans.factory;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans.factory.access;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans.factory.access.el;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans.factory.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans.factory.aspectj;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans.factory.config;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans.factory.parsing;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans.factory.serviceloader;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans.factory.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans.factory.wiring;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans.factory.xml;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans.propertyeditors;version=4.0.5.RELEASE_1,\ | |
- org.springframework.beans.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.cache;version=4.0.5.RELEASE_1,\ | |
- org.springframework.cache.ehcache;version=4.0.5.RELEASE_1,\ | |
- org.springframework.context;version=4.0.5.RELEASE_1,\ | |
- org.springframework.context.access;version=4.0.5.RELEASE_1,\ | |
- org.springframework.context.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.context.config;version=4.0.5.RELEASE_1,\ | |
- org.springframework.context.event;version=4.0.5.RELEASE_1,\ | |
- org.springframework.context.expression;version=4.0.5.RELEASE_1,\ | |
- org.springframework.context.i18n;version=4.0.5.RELEASE_1,\ | |
- org.springframework.context.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.context.weaving;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.convert;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.convert.converter;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.convert.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.enums;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.io;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.io.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.serializer;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.serializer.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.style;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.task;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.task.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.type;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.type.classreading;version=4.0.5.RELEASE_1,\ | |
- org.springframework.core.type.filter;version=4.0.5.RELEASE_1,\ | |
- org.springframework.dao;version=4.0.5.RELEASE_1,\ | |
- org.springframework.dao.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.dao.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.ejb;version=4.0.5.RELEASE_1,\ | |
- org.springframework.ejb.access;version=4.0.5.RELEASE_1,\ | |
- org.springframework.ejb.config;version=4.0.5.RELEASE_1,\ | |
- org.springframework.ejb.interceptor;version=4.0.5.RELEASE_1,\ | |
- org.springframework.ejb.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.expression;version=4.0.5.RELEASE_1,\ | |
- org.springframework.expression.common;version=4.0.5.RELEASE_1,\ | |
- org.springframework.expression.spel;version=4.0.5.RELEASE_1,\ | |
- org.springframework.expression.spel.ast;version=4.0.5.RELEASE_1,\ | |
- org.springframework.expression.spel.generated;version=4.0.5.RELEASE_1,\ | |
- org.springframework.expression.spel.standard;version=4.0.5.RELEASE_1,\ | |
- org.springframework.expression.spel.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.format;version=4.0.5.RELEASE_1,\ | |
- org.springframework.format.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.format.datetime;version=4.0.5.RELEASE_1,\ | |
- org.springframework.format.datetime.joda;version=4.0.5.RELEASE_1,\ | |
- org.springframework.format.number;version=4.0.5.RELEASE_1,\ | |
- org.springframework.format.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.http;version=4.0.5.RELEASE_1,\ | |
- org.springframework.http.client;version=4.0.5.RELEASE_1,\ | |
- org.springframework.http.client.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.http.converter;version=4.0.5.RELEASE_1,\ | |
- org.springframework.http.converter.feed;version=4.0.5.RELEASE_1,\ | |
- org.springframework.http.converter.json;version=4.0.5.RELEASE_1,\ | |
- org.springframework.http.converter.xml;version=4.0.5.RELEASE_1,\ | |
- org.springframework.http.server;version=4.0.5.RELEASE_1,\ | |
- org.springframework.instrument;version=4.0.5.RELEASE_1,\ | |
- org.springframework.instrument.classloading;version=4.0.5.RELEASE_1,\ | |
- org.springframework.instrument.classloading.glassfish;version=4.0.5.RELEASE_1,\ | |
- org.springframework.instrument.classloading.jboss;version=4.0.5.RELEASE_1,\ | |
- org.springframework.instrument.classloading.oc4j;version=4.0.5.RELEASE_1,\ | |
- org.springframework.instrument.classloading.weblogic;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jca;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jca.cci;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jca.cci.connection;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jca.cci.core;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jca.cci.core.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jca.cci.object;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jca.context;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jca.endpoint;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jca.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jca.work;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jca.work.glassfish;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jca.work.jboss;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.config;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.core;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.core.metadata;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.core.namedparam;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.core.simple;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.core.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.datasource;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.datasource.embedded;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.datasource.init;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.datasource.lookup;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.object;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.support.incrementer;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.support.lob;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.support.nativejdbc;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.support.rowset;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jdbc.support.xml;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jms;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jms.config;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jms.connection;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jms.core;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jms.core.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jms.listener;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jms.listener.adapter;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jms.listener.endpoint;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jms.remoting;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jms.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jms.support.converter;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jms.support.destination;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jmx;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jmx.access;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jmx.export;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jmx.export.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jmx.export.assembler;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jmx.export.metadata;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jmx.export.naming;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jmx.export.notification;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jmx.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jndi;version=4.0.5.RELEASE_1,\ | |
- org.springframework.jndi.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.mail;version=4.0.5.RELEASE_1,\ | |
- org.springframework.mail.javamail;version=4.0.5.RELEASE_1,\ | |
- org.springframework.mock;version=4.0.5.RELEASE_1,\ | |
- org.springframework.mock.jndi;version=4.0.5.RELEASE_1,\ | |
- org.springframework.mock.staticmock;version=4.0.5.RELEASE_1,\ | |
- org.springframework.mock.web;version=4.0.5.RELEASE_1,\ | |
- org.springframework.mock.web.portlet;version=4.0.5.RELEASE_1,\ | |
- org.springframework.orm;version=4.0.5.RELEASE_1,\ | |
- org.springframework.orm.hibernate3;version=4.0.5.RELEASE_1,\ | |
- org.springframework.orm.hibernate3.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.orm.hibernate3.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.orm.ibatis;version=4.0.5.RELEASE_1,\ | |
- org.springframework.orm.ibatis.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.orm.jdo;version=4.0.5.RELEASE_1,\ | |
- org.springframework.orm.jdo.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.orm.jpa;version=4.0.5.RELEASE_1,\ | |
- org.springframework.orm.jpa.aspectj;version=4.0.5.RELEASE_1,\ | |
- org.springframework.orm.jpa.persistenceunit;version=4.0.5.RELEASE_1,\ | |
- org.springframework.orm.jpa.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.orm.jpa.vendor;version=4.0.5.RELEASE_1,\ | |
- org.springframework.oxm;version=4.0.5.RELEASE_1,\ | |
- org.springframework.oxm.castor;version=4.0.5.RELEASE_1,\ | |
- org.springframework.oxm.config;version=4.0.5.RELEASE_1,\ | |
- org.springframework.oxm.jaxb;version=4.0.5.RELEASE_1,\ | |
- org.springframework.oxm.jibx;version=4.0.5.RELEASE_1,\ | |
- org.springframework.oxm.mime;version=4.0.5.RELEASE_1,\ | |
- org.springframework.oxm.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.oxm.xmlbeans;version=4.0.5.RELEASE_1,\ | |
- org.springframework.oxm.xstream;version=4.0.5.RELEASE_1,\ | |
- org.springframework.remoting;version=4.0.5.RELEASE_1,\ | |
- org.springframework.remoting.caucho;version=4.0.5.RELEASE_1,\ | |
- org.springframework.remoting.httpinvoker;version=4.0.5.RELEASE_1,\ | |
- org.springframework.remoting.jaxrpc;version=4.0.5.RELEASE_1,\ | |
- org.springframework.remoting.jaxws;version=4.0.5.RELEASE_1,\ | |
- org.springframework.remoting.rmi;version=4.0.5.RELEASE_1,\ | |
- org.springframework.remoting.soap;version=4.0.5.RELEASE_1,\ | |
- org.springframework.remoting.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scheduling;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scheduling.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scheduling.aspectj;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scheduling.backportconcurrent;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scheduling.commonj;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scheduling.concurrent;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scheduling.config;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scheduling.quartz;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scheduling.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scheduling.timer;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scripting;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scripting.bsh;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scripting.config;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scripting.groovy;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scripting.jruby;version=4.0.5.RELEASE_1,\ | |
- org.springframework.scripting.support;version=4.0.5.RELEASE_1,\ | |
+ org.snmp4j;version=2.4.3,\ | |
+ org.snmp4j.asn1;version=2.4.3,\ | |
+ org.snmp4j.event;version=2.4.3,\ | |
+ org.snmp4j.log;version=2.4.3,\ | |
+ org.snmp4j.mp;version=2.4.3,\ | |
+ org.snmp4j.security;version=2.4.3,\ | |
+ org.snmp4j.security.nonstandard;version=2.4.3,\ | |
+ org.snmp4j.smi;version=2.4.3,\ | |
+ org.snmp4j.test;version=2.4.3,\ | |
+ org.snmp4j.tools;version=2.4.3,\ | |
+ org.snmp4j.tools.console;version=2.4.3,\ | |
+ org.snmp4j.transport;version=2.4.3,\ | |
+ org.snmp4j.transport.ssh;version=2.4.3,\ | |
+ org.snmp4j.transport.tls;version=2.4.3,\ | |
+ org.snmp4j.uri;version=2.4.3,\ | |
+ org.snmp4j.util;version=2.4.3,\ | |
+ org.snmp4j.version;version=2.4.3,\ | |
+ org.springframework;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop.aspectj;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop.aspectj.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop.aspectj.autoproxy;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop.config;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop.framework;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop.framework.adapter;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop.framework.autoproxy;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop.framework.autoproxy.target;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop.interceptor;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop.scope;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop.support.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop.target;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.aop.target.dynamic;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.asm;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.asm.commons;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.asm.signature;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans.factory;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans.factory.access;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans.factory.access.el;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans.factory.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans.factory.aspectj;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans.factory.config;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans.factory.parsing;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans.factory.serviceloader;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans.factory.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans.factory.wiring;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans.factory.xml;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans.propertyeditors;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.beans.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.cache;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.cache.ehcache;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.context;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.context.access;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.context.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.context.config;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.context.event;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.context.expression;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.context.i18n;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.context.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.context.weaving;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.convert;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.convert.converter;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.convert.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.enums;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.io;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.io.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.serializer;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.serializer.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.style;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.task;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.task.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.type;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.type.classreading;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.core.type.filter;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.dao;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.dao.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.dao.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.ejb;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.ejb.access;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.ejb.config;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.ejb.interceptor;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.ejb.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.expression;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.expression.common;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.expression.spel;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.expression.spel.ast;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.expression.spel.generated;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.expression.spel.standard;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.expression.spel.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.format;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.format.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.format.datetime;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.format.datetime.joda;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.format.number;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.format.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.http;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.http.client;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.http.client.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.http.converter;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.http.converter.feed;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.http.converter.json;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.http.converter.xml;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.http.server;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.instrument;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.instrument.classloading;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.instrument.classloading.glassfish;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.instrument.classloading.jboss;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.instrument.classloading.oc4j;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.instrument.classloading.weblogic;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jca;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jca.cci;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jca.cci.connection;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jca.cci.core;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jca.cci.core.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jca.cci.object;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jca.context;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jca.endpoint;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jca.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jca.work;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jca.work.glassfish;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jca.work.jboss;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.config;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.core;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.core.metadata;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.core.namedparam;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.core.simple;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.core.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.datasource;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.datasource.embedded;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.datasource.init;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.datasource.lookup;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.object;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.support.incrementer;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.support.lob;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.support.nativejdbc;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.support.rowset;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jdbc.support.xml;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jms;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jms.config;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jms.connection;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jms.core;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jms.core.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jms.listener;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jms.listener.adapter;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jms.listener.endpoint;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jms.remoting;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jms.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jms.support.converter;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jms.support.destination;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jmx;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jmx.access;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jmx.export;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jmx.export.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jmx.export.assembler;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jmx.export.metadata;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jmx.export.naming;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jmx.export.notification;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jmx.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jndi;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.jndi.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.mail;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.mail.javamail;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.mock;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.mock.jndi;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.mock.staticmock;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.mock.web;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.mock.web.portlet;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.orm;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.orm.hibernate3;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.orm.hibernate3.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.orm.hibernate3.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.orm.ibatis;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.orm.ibatis.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.orm.jdo;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.orm.jdo.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.orm.jpa;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.orm.jpa.aspectj;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.orm.jpa.persistenceunit;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.orm.jpa.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.orm.jpa.vendor;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.oxm;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.oxm.castor;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.oxm.config;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.oxm.jaxb;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.oxm.jibx;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.oxm.mime;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.oxm.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.oxm.xmlbeans;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.oxm.xstream;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.remoting;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.remoting.caucho;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.remoting.httpinvoker;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.remoting.jaxrpc;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.remoting.jaxws;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.remoting.rmi;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.remoting.soap;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.remoting.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scheduling;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scheduling.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scheduling.aspectj;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scheduling.backportconcurrent;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scheduling.commonj;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scheduling.concurrent;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scheduling.config;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scheduling.quartz;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scheduling.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scheduling.timer;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scripting;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scripting.bsh;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scripting.config;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scripting.groovy;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scripting.jruby;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.scripting.support;version=4.0.7.RELEASE_1,\ | |
org.springframework.security.core;version=3.2.7.RELEASE,\ | |
org.springframework.security.core.context;version=3.2.7.RELEASE,\ | |
- org.springframework.stereotype;version=4.0.5.RELEASE_1,\ | |
- org.springframework.test;version=4.0.5.RELEASE_1,\ | |
- org.springframework.test.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.test.context;version=4.0.5.RELEASE_1,\ | |
- org.springframework.test.context.junit38;version=4.0.5.RELEASE_1,\ | |
- org.springframework.test.context.junit4;version=4.0.5.RELEASE_1,\ | |
- org.springframework.test.context.junit4.statements;version=4.0.5.RELEASE_1,\ | |
- org.springframework.test.context.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.test.context.testng;version=4.0.5.RELEASE_1,\ | |
- org.springframework.test.context.transaction;version=4.0.5.RELEASE_1,\ | |
- org.springframework.test.jdbc;version=4.0.5.RELEASE_1,\ | |
- org.springframework.test.jpa;version=4.0.5.RELEASE_1,\ | |
- org.springframework.test.util;version=4.0.5.RELEASE_1,\ | |
- org.springframework.test.web;version=4.0.5.RELEASE_1,\ | |
- org.springframework.transaction;version=4.0.5.RELEASE_1,\ | |
- org.springframework.transaction.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.transaction.aspectj;version=4.0.5.RELEASE_1,\ | |
- org.springframework.transaction.config;version=4.0.5.RELEASE_1,\ | |
- org.springframework.transaction.interceptor;version=4.0.5.RELEASE_1,\ | |
- org.springframework.transaction.jta;version=4.0.5.RELEASE_1,\ | |
- org.springframework.transaction.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.ui;version=4.0.5.RELEASE_1,\ | |
- org.springframework.ui.context;version=4.0.5.RELEASE_1,\ | |
- org.springframework.ui.context.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.ui.freemarker;version=4.0.5.RELEASE_1,\ | |
- org.springframework.ui.jasperreports;version=4.0.5.RELEASE_1,\ | |
- org.springframework.ui.velocity;version=4.0.5.RELEASE_1,\ | |
- org.springframework.util;version=4.0.5.RELEASE_1,\ | |
- org.springframework.util.comparator;version=4.0.5.RELEASE_1,\ | |
- org.springframework.util.xml;version=4.0.5.RELEASE_1,\ | |
- org.springframework.validation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.validation.beanvalidation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.validation.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.bind;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.bind.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.bind.annotation.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.bind.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.client;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.client.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.context;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.context.request;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.context.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.filter;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.jsf;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.jsf.el;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.multipart;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.multipart.commons;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.multipart.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.config;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.handler;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.i18n;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.mvc;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.mvc.annotation;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.mvc.multiaction;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.mvc.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.resource;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.support;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.tags;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.tags.form;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.theme;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.view;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.view.document;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.view.feed;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.view.freemarker;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.view.jasperreports;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.view.json;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.view.tiles;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.view.tiles2;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.view.velocity;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.view.xml;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.servlet.view.xslt;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.struts;version=4.0.5.RELEASE_1,\ | |
- org.springframework.web.util;version=4.0.5.RELEASE_1,\ | |
+ org.springframework.stereotype;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.test;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.test.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.test.context;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.test.context.junit38;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.test.context.junit4;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.test.context.junit4.statements;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.test.context.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.test.context.testng;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.test.context.transaction;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.test.jdbc;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.test.jpa;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.test.util;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.test.web;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.transaction;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.transaction.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.transaction.aspectj;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.transaction.config;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.transaction.interceptor;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.transaction.jta;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.transaction.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.ui;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.ui.context;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.ui.context.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.ui.freemarker;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.ui.jasperreports;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.ui.velocity;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.util;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.util.comparator;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.util.xml;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.validation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.validation.beanvalidation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.validation.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.bind;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.bind.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.bind.annotation.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.bind.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.client;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.client.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.context;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.context.request;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.context.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.filter;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.jsf;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.jsf.el;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.multipart;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.multipart.commons;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.multipart.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.config;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.handler;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.i18n;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.mvc;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.mvc.annotation;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.mvc.multiaction;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.mvc.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.resource;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.support;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.tags;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.tags.form;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.theme;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.view;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.view.document;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.view.feed;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.view.freemarker;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.view.jasperreports;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.view.json;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.view.tiles;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.view.tiles2;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.view.velocity;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.view.xml;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.servlet.view.xslt;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.struts;version=4.0.7.RELEASE_1,\ | |
+ org.springframework.web.util;version=4.0.7.RELEASE_1,\ | |
org.springframework.binding.collection;version=2.3.4.RELEASE,\ | |
org.springframework.binding.convert;version=2.3.4.RELEASE,\ | |
org.springframework.binding.convert.converters;version=2.3.4.RELEASE,\ | |
@@ -650,128 +661,129 @@ org.osgi.framework.system.packages.extra=org.apache.karaf.branding,\ | |
org.springframework.webflow.test.execution;version=2.3.4.RELEASE,\ | |
org.springframework.webflow.upgrade;version=2.3.4.RELEASE,\ | |
org.springframework.webflow.validation;version=2.3.4.RELEASE,\ | |
- org.opennms.netmgt;version=18.0.4,\ | |
- org.opennms.netmgt.charts;version=18.0.4,\ | |
- org.opennms.netmgt.collectd;version=18.0.4,\ | |
- org.opennms.netmgt.collection.support;version=18.0.4,\ | |
- org.opennms.netmgt.collection.support.builder;version=18.0.4,\ | |
- org.opennms.netmgt.collection.api;version=18.0.4,\ | |
- org.opennms.netmgt.config;version=18.0.4,\ | |
- org.opennms.netmgt.config.api;version=18.0.4,\ | |
- org.opennms.netmgt.config.categories;version=18.0.4,\ | |
- org.opennms.netmgt.config.charts;version=18.0.4,\ | |
- org.opennms.netmgt.config.datacollection;version=18.0.4,\ | |
- org.opennms.netmgt.config.groups;version=18.0.4,\ | |
- org.opennms.netmgt.config.kscReports;version=18.0.4,\ | |
- org.opennms.netmgt.config.opennmsDataSources;version=18.0.4,\ | |
- org.opennms.netmgt.config.siteStatusViews;version=18.0.4,\ | |
- org.opennms.netmgt.config.surveillanceViews;version=18.0.4,\ | |
- org.opennms.netmgt.config.users;version=18.0.4,\ | |
- org.opennms.netmgt.config.viewsdisplay;version=18.0.4,\ | |
- org.opennms.netmgt.config.webuiColors;version=18.0.4,\ | |
- org.opennms.api.integration.ticketing;version=18.0.4,\ | |
- org.opennms.api.reporting;version=18.0.4,\ | |
- org.opennms.api.reporting.parameter;version=18.0.4,\ | |
- org.opennms.container.web;version=18.0.4,\ | |
- org.opennms.core.config.api;version=18.0.4,\ | |
- org.opennms.core.db;version=18.0.4,\ | |
- org.opennms.core.network;version=18.0.4,\ | |
- org.opennms.core.criteria;version=18.0.4,\ | |
- org.opennms.core.criteria.restrictions;version=18.0.4,\ | |
- org.opennms.core.utils;version=18.0.4,\ | |
- org.opennms.core.logging;version=18.0.4,\ | |
- org.opennms.core.soa;version=18.0.4,\ | |
- org.opennms.core.soa.filter;version=18.0.4,\ | |
- org.opennms.core.soa.config;version=18.0.4,\ | |
- org.opennms.core.soa.support;version=18.0.4,\ | |
- org.opennms.core.spring;version=18.0.4,\ | |
- org.opennms.core.xml;version=18.0.4,\ | |
- org.opennms.core.concurrent;version=18.0.4,\ | |
- org.opennms.core.fiber;version=18.0.4,\ | |
- org.opennms.core.queue;version=18.0.4,\ | |
- org.opennms.core.resource;version=18.0.4,\ | |
- org.opennms.core.resource.db;version=18.0.4,\ | |
- org.opennms.core.utils.url;version=18.0.4,\ | |
- org.opennms.features.reporting.model;version=18.0.4,\ | |
- org.opennms.features.reporting.model.basicreport;version=18.0.4,\ | |
- org.opennms.features.reporting.model.jasper;version=18.0.4,\ | |
- org.opennms.features.reporting.model.jasperreport;version=18.0.4,\ | |
- org.opennms.features.reporting.model.remoterepository;version=18.0.4,\ | |
- org.opennms.features.reporting.repository;version=18.0.4,\ | |
- org.opennms.features.reporting.repository.global;version=18.0.4,\ | |
- org.opennms.features.reporting.repository.local;version=18.0.4,\ | |
- org.opennms.features.reporting.repository.remote;version=18.0.4,\ | |
- org.opennms.netmgt.bsm.service;version=18.0.4,\ | |
- org.opennms.netmgt.bsm.service.model;version=18.0.4,\ | |
- org.opennms.netmgt.bsm.service.model.edge;version=18.0.4,\ | |
- org.opennms.netmgt.bsm.service.model.edge.ro;version=18.0.4,\ | |
- org.opennms.netmgt.bsm.service.model.functions.map;version=18.0.4,\ | |
- org.opennms.netmgt.bsm.service.model.functions.reduce;version=18.0.4,\ | |
- org.opennms.netmgt.bsm.service.model.graph;version=18.0.4,\ | |
- org.opennms.netmgt.bsm.persistence.api;version=18.0.4,\ | |
- org.opennms.netmgt.bsm.service.internal;version=18.0.4,\ | |
- org.opennms.netmgt.dao;version=18.0.4,\ | |
- org.opennms.netmgt.dao.api;version=18.0.4,\ | |
- org.opennms.netmgt.dao.support;version=18.0.4,\ | |
- org.opennms.netmgt.events.api;version=18.0.4,\ | |
- org.opennms.netmgt.events.api.annotations;version=18.0.4,\ | |
- org.opennms.netmgt.events.api.support;version=18.0.4,\ | |
- org.opennms.netmgt.jmx.connection;version=18.0.4,\ | |
- org.opennms.netmgt.jmx.impl.connection.connectors;version=18.0.4,\ | |
- org.opennms.netmgt.model;version=18.0.4,\ | |
- org.opennms.netmgt.model.alarm;version=18.0.4,\ | |
- org.opennms.netmgt.alarmd.api;version=18.0.4,\ | |
- org.opennms.netmgt.alarmd.api.support;version=18.0.4,\ | |
- org.opennms.netmgt.model.capsd;version=18.0.4,\ | |
- org.opennms.netmgt.model.discovery;version=18.0.4,\ | |
- org.opennms.netmgt.model.events;version=18.0.4,\ | |
- org.opennms.netmgt.model.minion;version=18.0.4,\ | |
- org.opennms.netmgt.model.ncs;version=18.0.4,\ | |
- org.opennms.netmgt.model.notifd;version=18.0.4,\ | |
- org.opennms.netmgt.model.outage;version=18.0.4,\ | |
- org.opennms.netmgt.model.topology;version=18.0.4,\ | |
- org.opennms.netmgt.poller;version=18.0.4,\ | |
- org.opennms.netmgt.provision.persist;version=18.0.4,\ | |
- org.opennms.netmgt.provision.persist.foreignsource;version=18.0.4,\ | |
- org.opennms.netmgt.provision.persist.policies;version=18.0.4,\ | |
- org.opennms.netmgt.provision.persist.requisition;version=18.0.4,\ | |
- org.opennms.netmgt.rrd;version=18.0.4,\ | |
- org.opennms.netmgt.snmp;version=18.0.4,\ | |
- org.opennms.netmgt.syslogd;version=18.0.4,\ | |
- org.opennms.netmgt.ticketd;version=18.0.4,\ | |
- org.opennms.netmgt.xml.event;version=18.0.4,\ | |
- org.opennms.netmgt.xml.eventconf;version=18.0.4,\ | |
- org.opennms.netmgt.xml.rtc;version=18.0.4,\ | |
- org.opennms.reporting.core;version=18.0.4,\ | |
- org.opennms.reporting.core.svclayer;version=18.0.4,\ | |
- org.opennms.reporting.core.svclayer.support;version=18.0.4,\ | |
- org.opennms.web.api;version=18.0.4,\ | |
- org.opennms.web.category;version=18.0.4,\ | |
- org.opennms.web.charts;version=18.0.4,\ | |
- org.opennms.web.navigate;version=18.0.4,\ | |
- org.opennms.web.servlet;version=18.0.4,\ | |
- org.opennms.web.svclayer;version=18.0.4,\ | |
- org.opennms.web.svclayer.api;version=18.0.4,\ | |
- org.opennms.web.svclayer.dao;version=18.0.4,\ | |
- org.opennms.web.svclayer.dao.support;version=18.0.4,\ | |
- org.opennms.web.svclayer.model;version=18.0.4,\ | |
- org.opennms.web.svclayer.support;version=18.0.4,\ | |
- org.opennms.web.springframework.security;version=18.0.4,\ | |
- org.opennms.netmgt.icmp;version=18.0.4 | |
+ org.opennms.netmgt;version=19.0.0,\ | |
+ org.opennms.netmgt.charts;version=19.0.0,\ | |
+ org.opennms.netmgt.collectd;version=19.0.0,\ | |
+ org.opennms.netmgt.collection.support;version=19.0.0,\ | |
+ org.opennms.netmgt.collection.support.builder;version=19.0.0,\ | |
+ org.opennms.netmgt.collection.api;version=19.0.0,\ | |
+ org.opennms.netmgt.config;version=19.0.0,\ | |
+ org.opennms.netmgt.config.api;version=19.0.0,\ | |
+ org.opennms.netmgt.config.categories;version=19.0.0,\ | |
+ org.opennms.netmgt.config.charts;version=19.0.0,\ | |
+ org.opennms.netmgt.config.collectd;version=19.0.0,\ | |
+ org.opennms.netmgt.config.datacollection;version=19.0.0,\ | |
+ org.opennms.netmgt.config.groups;version=19.0.0,\ | |
+ org.opennms.netmgt.config.kscReports;version=19.0.0,\ | |
+ org.opennms.netmgt.config.opennmsDataSources;version=19.0.0,\ | |
+ org.opennms.netmgt.config.siteStatusViews;version=19.0.0,\ | |
+ org.opennms.netmgt.config.surveillanceViews;version=19.0.0,\ | |
+ org.opennms.netmgt.config.trapd;version=19.0.0,\ | |
+ org.opennms.netmgt.config.users;version=19.0.0,\ | |
+ org.opennms.netmgt.config.viewsdisplay;version=19.0.0,\ | |
+ org.opennms.netmgt.config.webuiColors;version=19.0.0,\ | |
+ org.opennms.api.integration.ticketing;version=19.0.0,\ | |
+ org.opennms.api.reporting;version=19.0.0,\ | |
+ org.opennms.api.reporting.parameter;version=19.0.0,\ | |
+ org.opennms.container.web;version=19.0.0,\ | |
+ org.opennms.core.config.api;version=19.0.0,\ | |
+ org.opennms.core.db;version=19.0.0,\ | |
+ org.opennms.core.network;version=19.0.0,\ | |
+ org.opennms.core.criteria;version=19.0.0,\ | |
+ org.opennms.core.criteria.restrictions;version=19.0.0,\ | |
+ org.opennms.core.utils;version=19.0.0,\ | |
+ org.opennms.core.logging;version=19.0.0,\ | |
+ org.opennms.core.soa;version=19.0.0,\ | |
+ org.opennms.core.soa.filter;version=19.0.0,\ | |
+ org.opennms.core.soa.config;version=19.0.0,\ | |
+ org.opennms.core.soa.support;version=19.0.0,\ | |
+ org.opennms.core.spring;version=19.0.0,\ | |
+ org.opennms.core.xml;version=19.0.0,\ | |
+ org.opennms.core.concurrent;version=19.0.0,\ | |
+ org.opennms.core.fiber;version=19.0.0,\ | |
+ org.opennms.core.queue;version=19.0.0,\ | |
+ org.opennms.core.resource;version=19.0.0,\ | |
+ org.opennms.core.resource.db;version=19.0.0,\ | |
+ org.opennms.core.rpc.api;version=19.0.0,\ | |
+ org.opennms.core.utils.url;version=19.0.0,\ | |
+ org.opennms.features.geolocation.api;version=19.0.0,\ | |
+ org.opennms.features.reporting.model;version=19.0.0,\ | |
+ org.opennms.features.reporting.model.basicreport;version=19.0.0,\ | |
+ org.opennms.features.reporting.model.jasper;version=19.0.0,\ | |
+ org.opennms.features.reporting.model.jasperreport;version=19.0.0,\ | |
+ org.opennms.features.reporting.model.remoterepository;version=19.0.0,\ | |
+ org.opennms.features.reporting.repository;version=19.0.0,\ | |
+ org.opennms.features.reporting.repository.global;version=19.0.0,\ | |
+ org.opennms.features.reporting.repository.local;version=19.0.0,\ | |
+ org.opennms.features.reporting.repository.remote;version=19.0.0,\ | |
+ org.opennms.netmgt.bsm.service;version=19.0.0,\ | |
+ org.opennms.netmgt.bsm.service.model;version=19.0.0,\ | |
+ org.opennms.netmgt.bsm.service.model.edge;version=19.0.0,\ | |
+ org.opennms.netmgt.bsm.service.model.edge.ro;version=19.0.0,\ | |
+ org.opennms.netmgt.bsm.service.model.functions.map;version=19.0.0,\ | |
+ org.opennms.netmgt.bsm.service.model.functions.reduce;version=19.0.0,\ | |
+ org.opennms.netmgt.bsm.service.model.graph;version=19.0.0,\ | |
+ org.opennms.netmgt.bsm.persistence.api;version=19.0.0,\ | |
+ org.opennms.netmgt.bsm.service.internal;version=19.0.0,\ | |
+ org.opennms.netmgt.dao;version=19.0.0,\ | |
+ org.opennms.netmgt.dao.api;version=19.0.0,\ | |
+ org.opennms.netmgt.dao.support;version=19.0.0,\ | |
+ org.opennms.netmgt.events.api;version=19.0.0,\ | |
+ org.opennms.netmgt.events.api.annotations;version=19.0.0,\ | |
+ org.opennms.netmgt.events.api.support;version=19.0.0,\ | |
+ org.opennms.netmgt.jmx.connection;version=19.0.0,\ | |
+ org.opennms.netmgt.jmx.impl.connection.connectors;version=19.0.0,\ | |
+ org.opennms.netmgt.model;version=19.0.0,\ | |
+ org.opennms.netmgt.model.alarm;version=19.0.0,\ | |
+ org.opennms.netmgt.alarmd.api;version=19.0.0,\ | |
+ org.opennms.netmgt.alarmd.api.support;version=19.0.0,\ | |
+ org.opennms.netmgt.measurements.api;version=19.0.0,\ | |
+ org.opennms.netmgt.measurements.api.exceptions;version=19.0.0,\ | |
+ org.opennms.netmgt.measurements.model;version=19.0.0,\ | |
+ org.opennms.netmgt.model.capsd;version=19.0.0,\ | |
+ org.opennms.netmgt.model.discovery;version=19.0.0,\ | |
+ org.opennms.netmgt.model.events;version=19.0.0,\ | |
+ org.opennms.netmgt.model.minion;version=19.0.0,\ | |
+ org.opennms.netmgt.model.monitoringLocations;version=19.0.0,\ | |
+ org.opennms.netmgt.model.ncs;version=19.0.0,\ | |
+ org.opennms.netmgt.model.notifd;version=19.0.0,\ | |
+ org.opennms.netmgt.model.outage;version=19.0.0,\ | |
+ org.opennms.netmgt.model.topology;version=19.0.0,\ | |
+ org.opennms.netmgt.poller;version=19.0.0,\ | |
+ org.opennms.netmgt.poller.support;version=19.0.0,\ | |
+ org.opennms.netmgt.provision;version=19.0.0,\ | |
+ org.opennms.netmgt.provision.detector.registry.api;version=19.0.0,\ | |
+ org.opennms.netmgt.provision.persist;version=19.0.0,\ | |
+ org.opennms.netmgt.provision.persist.foreignsource;version=19.0.0,\ | |
+ org.opennms.netmgt.provision.persist.policies;version=19.0.0,\ | |
+ org.opennms.netmgt.provision.persist.requisition;version=19.0.0,\ | |
+ org.opennms.netmgt.rrd;version=19.0.0,\ | |
+ org.opennms.netmgt.snmp;version=19.0.0,\ | |
+ org.opennms.netmgt.snmp.snmp4j;version=19.0.0,\ | |
+ org.opennms.netmgt.snmp.proxy;version=19.0.0,\ | |
+ org.opennms.netmgt.syslogd;version=19.0.0,\ | |
+ org.opennms.netmgt.ticketd;version=19.0.0,\ | |
+ org.opennms.netmgt.topology.persistence.api;version=19.0.0,\ | |
+ org.opennms.netmgt.trapd;version=19.0.0,\ | |
+ org.opennms.netmgt.xml.event;version=19.0.0,\ | |
+ org.opennms.netmgt.xml.eventconf;version=19.0.0,\ | |
+ org.opennms.netmgt.xml.rtc;version=19.0.0,\ | |
+ org.opennms.reporting.core;version=19.0.0,\ | |
+ org.opennms.reporting.core.svclayer;version=19.0.0,\ | |
+ org.opennms.reporting.core.svclayer.support;version=19.0.0,\ | |
+ org.opennms.web.api;version=19.0.0,\ | |
+ org.opennms.web.category;version=19.0.0,\ | |
+ org.opennms.web.charts;version=19.0.0,\ | |
+ org.opennms.web.navigate;version=19.0.0,\ | |
+ org.opennms.web.servlet;version=19.0.0,\ | |
+ org.opennms.web.springframework;version=19.0.0,\ | |
+ org.opennms.web.springframework.security;version=19.0.0,\ | |
+ org.opennms.web.svclayer;version=19.0.0,\ | |
+ org.opennms.web.svclayer.api;version=19.0.0,\ | |
+ org.opennms.web.svclayer.dao;version=19.0.0,\ | |
+ org.opennms.web.svclayer.dao.support;version=19.0.0,\ | |
+ org.opennms.web.svclayer.model;version=19.0.0,\ | |
+ org.opennms.web.svclayer.support;version=19.0.0,\ | |
+ org.opennms.netmgt.icmp;version=19.0.0,\ | |
+ org.opennms.netmgt.icmp.proxy;version=19.0.0 | |
- | |
- | |
-# org.snmp4j;version=1.11.1,\ | |
-# org.snmp4j.asn1;version=1.11.1,\ | |
-# org.snmp4j.doc-files;version=1.11.1,\ | |
-# org.snmp4j.event;version=1.11.1,\ | |
-# org.snmp4j.log;version=1.11.1,\ | |
-# org.snmp4j.mp;version=1.11.1,\ | |
-# org.snmp4j.security;version=1.11.1,\ | |
-# org.snmp4j.smi;version=1.11.1,\ | |
-# org.snmp4j.smi;version=1.11.1,\ | |
-# org.snmp4j.test;version=1.11.1,\ | |
-# org.snmp4j.tools.console;version=1.11.1,\ | |
-# org.snmp4j.transport;version=1.11.1,\ | |
-# org.snmp4j.util;version=1.11.1,\ | |
-# org.snmp4j.version;version=1.11.1 | |
diff --git a/database-reports.xml b/database-reports.xml | |
index 38cfa64..e3b9094 100644 | |
--- a/database-reports.xml | |
+++ b/database-reports.xml | |
@@ -7,35 +7,40 @@ | |
report-service="availabilityReportService" description="standard opennms report in calendar format" /> | |
<report id="defaultClassicReport" display-name="Default classic report" | |
report-service="availabilityReportService" description="standard opennms report in tabular format" /> | |
- <report id="Early-Morning-Report" display-name="Early morning report" online="true" | |
+ <report id="Early-Morning-Report" display-name="Early morning report" online="true" | |
report-service="jasperReportService" description="Global overview of outages, notifications and events in last 24 hours" /> | |
- <report id="Response-Time-Summary-Report" display-name="Response Time Summary for node" online="true" | |
+ <report id="Response-Time-Summary-Report" display-name="Response Time Summary for node" online="true" | |
report-service="jasperReportService" description="Response Time by node across one or more surveillance categories. Note: % can be used as a place holder for any string literal" /> | |
- <report id="Node-Availability-Report" display-name="Availability by node" online="true" | |
+ <report id="Node-Availability-Report" display-name="Availability by node" online="true" | |
report-service="jasperReportService" description="Availability by node across one or more surveillance categories. Note: % can be used as a place holder for any string literal" /> | |
- <report id="Availability-Summary-Report" display-name="Availability Summary -Default configuration for past 7 Days" online="true" | |
+ <report id="Availability-Summary-Report" display-name="Availability Summary -Default configuration for past 7 Days" online="true" | |
report-service="jasperReportService" description="Availability summary across one or more surveillance categories. Note: % can be used as a place holder for any string literal" /> | |
- <report id="Response-Time-Report" display-name="Response time by node" online="true" | |
+ <report id="Response-Time-Report" display-name="Response time by node" online="true" | |
report-service="jasperReportService" description="Response time by node across one or more surveillance categories. Note: % can be used as a place holder for any string literal" /> | |
- <report id="Serial-Interface-Utilization-Summary" display-name="Serial Interface Utilization Summary" online="true" | |
+ <report id="Serial-Interface-Utilization-Summary" display-name="Serial Interface Utilization Summary" online="true" | |
report-service="jasperReportService" description="Serial Interface Utilization Summary" /> | |
- <report id="Total-Bytes-Transferred-By-Interface" display-name="Total Bytes Transferred by Interface " online="true" | |
+ <report id="Total-Bytes-Transferred-By-Interface" display-name="Total Bytes Transferred by Interface " online="true" | |
report-service="jasperReportService" description="Total Bytes Transferred by Interface" /> | |
- <report id="Average-Peak-Traffic-Rates" display-name="Average and Peak Traffic rates for Nodes by Interface" online="true" | |
+ <report id="Average-Peak-Traffic-Rates" display-name="Average and Peak Traffic rates for Nodes by Interface" online="true" | |
report-service="jasperReportService" description="Average and Peak Traffic rates for Nodes by Interface" /> | |
- <report id="Interface-Availability-Report" display-name="Interface Availability Report" online="true" | |
+ <report id="Interface-Availability-Report" display-name="Interface Availability Report" online="true" | |
report-service="jasperReportService" description="Interface Availability Report, show interface availability for interfaces with outages within time range" /> | |
- <report id="Snmp-Interface-Oper-Availability" display-name="Snmp Interface Availability Report" online="true" | |
+ <report id="Snmp-Interface-Oper-Availability" display-name="Snmp Interface Availability Report" online="true" | |
report-service="jasperReportService" description="Snmp Interface Availability Report, shows availability for snmp interfaces with interfaceOperDown outages within the time range" /> | |
- <report id="AssetMangementMaintExpired" display-name="Maintenance contracts expired" online="true" | |
+ <report id="AssetMangementMaintExpired" display-name="Maintenance contracts expired" online="true" | |
report-service="jasperReportService" description="Asset management report shows all maintenance contracts expired." /> | |
- <report id="AssetMangementMaintStrategy" display-name="Maintenance contracts strategy" online="true" | |
+ <report id="AssetMangementMaintStrategy" display-name="Maintenance contracts strategy" online="true" | |
report-service="jasperReportService" description="Asset management report focused on maintenance strategy.Forecast and overview for 12 month and informations about age of nodes and maintenance contracts" /> | |
- <report id="Event-Analysis" display-name="Event Analysis report" online="true" | |
+ <report id="Event-Analysis" display-name="Event Analysis report" online="true" | |
report-service="jasperReportService" description="Analyse events based on events source and quantity by nodes." /> | |
<!-- The following report expects that storeByGroup is enabled --> | |
<!-- | |
- <report id="DiskUsageForCTX" display-name="C: Disk Usage for CTX servers" online="true" | |
+ <report id="DiskUsageForCTX" display-name="C: Disk Usage for CTX servers" online="true" | |
report-service="jasperReportService" description="C: Disk Usage for CTX servers" /> | |
--> | |
+<!-- The following report expects that storeByForeignSource is enabled --> | |
+<!-- | |
+ <report id="Top20-IOWait" display-name="TOP 20 nodes by I/O Wait" online="true" | |
+ report-service="jasperReportService" description="TOP 20 nodes by I/O Wait" /> | |
+--> | |
</database-reports> | |
diff --git a/database-schema.xml b/database-schema.xml | |
index f6994b6..6616dca 100644 | |
--- a/database-schema.xml | |
+++ b/database-schema.xml | |
@@ -1,20 +1,10 @@ | |
<?xml version="1.0"?> | |
<database-schema> | |
- <table name="distPoller" visible="false"> | |
- <column name="dpNumber"/> | |
- <column name="dpName"/> | |
- <column name="dpIP"/> | |
- <column name="dpComment"/> | |
- <column name="dpDiscLimit"/> | |
- <column name="dpAdminState"/> | |
- <column name="dpRunState"/> | |
- </table> | |
- | |
<table name="node"> | |
<join column="nodeID" table="ipInterface" table-column="nodeID"/> | |
<column name="nodeID"/> | |
- <column name="dpName" visible="false"/> | |
+ <column name="location"/> | |
<column name="nodeCreateTime"/> | |
<column name="nodeParentID"/> | |
<column name="nodeType"/> | |
@@ -24,6 +14,10 @@ | |
<column name="nodeSysLocation"/> | |
<column name="nodeSysContact"/> | |
<column name="nodeLabel"/> | |
+ <column name="nodeLabelSource"/> | |
+ <column name="nodeNetbiosName"/> | |
+ <column name="nodeDomainName"/> | |
+ <column name="operatingSystem"/> | |
<column name="foreignSource"/> | |
<column name="foreignID"/> | |
</table> | |
@@ -57,7 +51,6 @@ | |
<join column="id" table="ipInterface" table-column="snmpinterfaceid"/> | |
<column name="id" visible="false"/> | |
<column name="nodeID" visible="false"/> | |
- <column name="ipAddr" visible="false"/> | |
<column name="snmpIpAdEntNetMask"/> | |
<column name="snmpPhysAddr"/> | |
<column name="snmpIfIndex"/> | |
@@ -67,7 +60,8 @@ | |
<column name="snmpIfAlias"/> | |
<column name="snmpIfAdminStatus"/> | |
<column name="snmpIfOperStatus"/> | |
- <column name="snmpcollect"/> | |
+ <column name="snmpCollect"/> | |
+ <column name="snmpPoll"/> | |
</table> | |
<table name="service"> | |
diff --git a/datacollection/fortinet-fortigate-application-v5.2.xml b/datacollection/fortinet-fortigate-application-v5.2.xml | |
index 2cf392d..af11eba 100644 | |
--- a/datacollection/fortinet-fortigate-application-v5.2.xml | |
+++ b/datacollection/fortinet-fortigate-application-v5.2.xml | |
@@ -160,15 +160,15 @@ | |
<mibObj oid=".1.3.6.1.4.1.12356.101.10.111.3.1.1" instance="fgApFTPStatsEntry" alias="fgApFTPReqProcessed" type="Counter32" /> | |
</group> | |
<group name="fgApFTPConnections" ifType="ignore"> | |
- <mibObj oid=".1.3.6.1.4.1.12356.101.10.111.4" instance="0" alias="fgApFTPConnections" type="Integer32" /> | |
- <mibObj oid=".1.3.6.1.4.1.12356.101.10.111.5" instance="0" alias="fgApFTPMaxConnections" type="Integer32" /> | |
+ <mibObj oid=".1.3.6.1.4.1.12356.101.10.111.4" instance="0" alias="fgApFTPConns" type="Integer32" /> | |
+ <mibObj oid=".1.3.6.1.4.1.12356.101.10.111.5" instance="0" alias="fgApFTPMaxConns" type="Integer32" /> | |
</group> | |
<group name="fgWebCacheDiskStatsTable" ifType="all"> | |
<mibObj oid=".1.3.6.1.4.1.12356.101.10.113.2.1.1" instance="fgWebChDskStsEntry" alias="fgWebCacheDisk" type="String" /> | |
<mibObj oid=".1.3.6.1.4.1.12356.101.10.113.2.1.2" instance="fgWebChDskStsEntry" alias="fgWebCacheDiskLimit" type="Gauge64" /> | |
<mibObj oid=".1.3.6.1.4.1.12356.101.10.113.2.1.3" instance="fgWebChDskStsEntry" alias="fgWebCacheDiskUsage" type="Gauge64" /> | |
<mibObj oid=".1.3.6.1.4.1.12356.101.10.113.2.1.4" instance="fgWebChDskStsEntry" alias="fgWebCacheDiskHits" type="Counter32" /> | |
- <mibObj oid=".1.3.6.1.4.1.12356.101.10.113.2.1.5" instance="fgWebChDskStsEntry" alias="fgWebCacheDiskMisses" type="Counter32" /> | |
+ <mibObj oid=".1.3.6.1.4.1.12356.101.10.113.2.1.5" instance="fgWebChDskStsEntry" alias="fgWebCacheDiskMiss" type="Counter32" /> | |
</group> | |
<systemDef name="Fortinet-Fortigate-Application-v5.2"> | |
<sysoidMask>.1.3.6.1.4.1.12356.</sysoidMask> | |
diff --git a/datacollection/netapp.xml b/datacollection/netapp.xml | |
index 53e1c16..6339aaf 100644 | |
--- a/datacollection/netapp.xml | |
+++ b/datacollection/netapp.xml | |
@@ -46,8 +46,8 @@ | |
</group> | |
<group name="netapp-misc" ifType="ignore"> | |
- <mibObj oid=".1.3.6.1.4.1.789.1.2.2.16" instance="0" alias="naMiscLowDiskReadBytes" type="Counter" /> | |
- <mibObj oid=".1.3.6.1.4.1.789.1.2.2.18" instance="0" alias="naMiscLowDiskWriteBytes" type="Counter" /> | |
+ <mibObj oid=".1.3.6.1.4.1.789.1.2.2.16" instance="0" alias="naMscLowDiskRdBytes" type="Counter" /> | |
+ <mibObj oid=".1.3.6.1.4.1.789.1.2.2.18" instance="0" alias="naMscLowDiskWrBytes" type="Counter" /> | |
</group> | |
<group name="netapp-sis" ifType="all"> | |
diff --git a/datacollection/vmware6.xml b/datacollection/vmware6.xml | |
new file mode 100644 | |
index 0000000..3c7393a | |
--- /dev/null | |
+++ b/datacollection/vmware6.xml | |
@@ -0,0 +1,61 @@ | |
+<?xml version="1.0"?> | |
+ | |
+<!-- | |
+Configuration file generated for: | |
+ | |
+Full name.......: VMware vCenter Server 6.0.0 build-3339084 | |
+API type........: VirtualCenter | |
+API version.....: 6.0 | |
+Product name....: VMware VirtualCenter Server | |
+Product version.: 6.0 | |
+OS type.........: linux-x64 | |
+--> | |
+ | |
+<datacollection-group name="VMware6"> | |
+ | |
+ <resourceType name="vmware6Cpu" label="VMware v6 Cpu" resourceLabel="${vmware6CpuName}"> | |
+ <persistenceSelectorStrategy class="org.opennms.netmgt.collection.support.PersistAllSelectorStrategy"/> | |
+ <storageStrategy class="org.opennms.netmgt.collection.support.IndexStorageStrategy"/> | |
+ </resourceType> | |
+ | |
+ <resourceType name="vmware6DaSt" label="VMware v6 DaSt" resourceLabel="${vmware6DaStName}"> | |
+ <persistenceSelectorStrategy class="org.opennms.netmgt.collection.support.PersistAllSelectorStrategy"/> | |
+ <storageStrategy class="org.opennms.netmgt.collection.support.IndexStorageStrategy"/> | |
+ </resourceType> | |
+ | |
+ <resourceType name="vmware6Disk" label="VMware v6 Disk" resourceLabel="${vmware6DiskName}"> | |
+ <persistenceSelectorStrategy class="org.opennms.netmgt.collection.support.PersistAllSelectorStrategy"/> | |
+ <storageStrategy class="org.opennms.netmgt.collection.support.IndexStorageStrategy"/> | |
+ </resourceType> | |
+ | |
+ <resourceType name="vmware6Net" label="VMware v6 Net" resourceLabel="${vmware6NetName}"> | |
+ <persistenceSelectorStrategy class="org.opennms.netmgt.collection.support.PersistAllSelectorStrategy"/> | |
+ <storageStrategy class="org.opennms.netmgt.collection.support.IndexStorageStrategy"/> | |
+ </resourceType> | |
+ | |
+ <resourceType name="vmware6StAdptr" label="VMware v6 StAdptr" resourceLabel="${vmware6StAdptrName}"> | |
+ <persistenceSelectorStrategy class="org.opennms.netmgt.collection.support.PersistAllSelectorStrategy"/> | |
+ <storageStrategy class="org.opennms.netmgt.collection.support.IndexStorageStrategy"/> | |
+ </resourceType> | |
+ | |
+ <resourceType name="vmware6StPth" label="VMware v6 StPth" resourceLabel="${vmware6StPthName}"> | |
+ <persistenceSelectorStrategy class="org.opennms.netmgt.collection.support.PersistAllSelectorStrategy"/> | |
+ <storageStrategy class="org.opennms.netmgt.collection.support.IndexStorageStrategy"/> | |
+ </resourceType> | |
+ | |
+ <resourceType name="vmware6Sys" label="VMware v6 Sys" resourceLabel="${vmware6SysName}"> | |
+ <persistenceSelectorStrategy class="org.opennms.netmgt.collection.support.PersistAllSelectorStrategy"/> | |
+ <storageStrategy class="org.opennms.netmgt.collection.support.IndexStorageStrategy"/> | |
+ </resourceType> | |
+ | |
+ <resourceType name="vmware6VrtDisk" label="VMware v6 VrtDisk" resourceLabel="${vmware6VrtDiskName}"> | |
+ <persistenceSelectorStrategy class="org.opennms.netmgt.collection.support.PersistAllSelectorStrategy"/> | |
+ <storageStrategy class="org.opennms.netmgt.collection.support.IndexStorageStrategy"/> | |
+ </resourceType> | |
+ | |
+ <resourceType name="vmware6vflashModule" label="VMware v6 vflashModule" resourceLabel="${vmware6vflashModuleName}"> | |
+ <persistenceSelectorStrategy class="org.opennms.netmgt.collection.support.PersistAllSelectorStrategy"/> | |
+ <storageStrategy class="org.opennms.netmgt.collection.support.IndexStorageStrategy"/> | |
+ </resourceType> | |
+ | |
+</datacollection-group> | |
diff --git a/events/opennms.events.xml b/events/opennms.events.xml | |
index 565582d..6946cfa 100644 | |
--- a/events/opennms.events.xml | |
+++ b/events/opennms.events.xml | |
@@ -914,13 +914,15 @@ | |
<p>This event is generated when node outage processing | |
determines that the critical path IP address/service for | |
this node is not responding..</p> | |
+ <p>More information on the affected node(s) can be found at | |
+ the <a href="opennms/pathOutage/index.jsp">Path Outages</a> page.</p> | |
</descr> | |
<logmsg dest="logndisplay"> | |
%parm[nodelabel]% path outage. Critical path = | |
%parm[criticalPathIp]% %parm[criticalPathServiceName]% | |
</logmsg> | |
<severity>Major</severity> | |
- <alarm-data reduction-key="%uei%:%dpname%:%nodeid%" alarm-type="3" auto-clean="false"/> | |
+ <alarm-data reduction-key="%uei%:%dpname%:%parm[criticalPathIp]%:%parm[criticalPathServiceName]%" alarm-type="3" auto-clean="false"/> | |
</event> | |
<event> | |
<uei>uei.opennms.org/nodes/nodeGainedInterface</uei> | |
@@ -2078,4 +2080,26 @@ | |
<update-field field-name="severity" update-on-reduction="true"/> | |
</alarm-data> | |
</event> | |
+ | |
+ <event> | |
+ <uei>uei.opennms.org/internal/monitoringSystemAdded</uei> | |
+ <event-label>Monitoring system Added</event-label> | |
+ <descr> A new monitoring system has been added </descr> | |
+ <logmsg dest="logndisplay">A new monitoring system of type '%parm[monitoringSystemType]%' has been added with ID '%parm[monitoringSystemId]%' at location '%parm[monitoringSystemLocation]%'. </logmsg> | |
+ <severity>Normal</severity> | |
+ </event> | |
+ <event> | |
+ <uei>uei.opennms.org/internal/monitoringSystemLocationChanged</uei> | |
+ <event-label>Monitoring system Location Changed</event-label> | |
+ <descr> Monitoring system location changed</descr> | |
+ <logmsg dest="logndisplay"> Monitoring system of type '%parm[monitoringSystemType]%' with ID '%parm[monitoringSystemId]%' has changed its location from '%parm[monitoringSystemPreviousLocation]%' to '%parm[monitoringSystemLocation]%'. </logmsg> | |
+ <severity>Normal</severity> | |
+ </event> | |
+ <event> | |
+ <uei>uei.opennms.org/internal/monitoringSystemDeleted</uei> | |
+ <event-label>Monitoring system Deleted</event-label> | |
+ <descr> Monitoring system Deleted</descr> | |
+ <logmsg dest="logndisplay"> Monitoring system of type '%parm[monitoringSystemType]%' with ID '%parm[monitoringSystemId]%' at location '%parm[monitoringSystemLocation]%' has been deleted. </logmsg> | |
+ <severity>Normal</severity> | |
+ </event> | |
</events> | |
diff --git a/graphml-edge-status/.readme b/graphml-edge-status/.readme | |
new file mode 100644 | |
index 0000000..51f5fa5 | |
--- /dev/null | |
+++ b/graphml-edge-status/.readme | |
@@ -0,0 +1 @@ | |
+Add your custom .groovy files to this folder. | |
diff --git a/infopanel/.readme b/infopanel/.readme | |
new file mode 100644 | |
index 0000000..d1e91d8 | |
--- /dev/null | |
+++ b/infopanel/.readme | |
@@ -0,0 +1 @@ | |
+Add your custom info panel item files (*.HTML) to this folder. | |
\ No newline at end of file | |
diff --git a/jasper-reports.xml b/jasper-reports.xml | |
index e9d9e62..d7c187a 100644 | |
--- a/jasper-reports.xml | |
+++ b/jasper-reports.xml | |
@@ -16,5 +16,6 @@ | |
<report id="AssetMangementMaintExpired" template="AssetManagementMaintExpired.jrxml" engine="jdbc" /> | |
<report id="AssetMangementMaintStrategy" template="AssetManagementMaintStrategy.jrxml" engine="jdbc" /> | |
<report id="Event-Analysis" template="EventAnalysis.jrxml" engine="jdbc" /> | |
- <report id="DiskUsageForCTX" template="DiskUsageForCTX.jrxml" engine="jdbc" /> | |
+ <report id="DiskUsageForCTX" template="DiskUsageForCTX.jrxml" engine="jdbc" /> | |
+ <report id="Top20-IOWait" template="TopIOWait.jrxml" engine="jdbc" /> | |
</jasper-reports> | |
diff --git a/java.util.logging.properties b/java.util.logging.properties | |
index 6b21730..b7b50c5 100644 | |
--- a/java.util.logging.properties | |
+++ b/java.util.logging.properties | |
@@ -20,3 +20,4 @@ | |
# Empty java.util.logging.properties to prevent the log to stderr, so that | |
# all logs will be delegated to pax logging JUL handler only | |
+ | |
diff --git a/jmx-datacollection-config.d/cassandra30x-newts.xml b/jmx-datacollection-config.d/cassandra30x-newts.xml | |
new file mode 100644 | |
index 0000000..17255a7 | |
--- /dev/null | |
+++ b/jmx-datacollection-config.d/cassandra30x-newts.xml | |
@@ -0,0 +1,118 @@ | |
+<?xml version="1.0"?> | |
+<jmx-datacollection-config> | |
+ <jmx-collection name="jmx-cassandra30x-newts"> | |
+ <rrd step="300"> | |
+ <rra>RRA:AVERAGE:0.5:1:2016</rra> | |
+ <rra>RRA:AVERAGE:0.5:12:1488</rra> | |
+ <rra>RRA:AVERAGE:0.5:288:366</rra> | |
+ <rra>RRA:MAX:0.5:288:366</rra> | |
+ <rra>RRA:MIN:0.5:288:366</rra> | |
+ </rrd> | |
+ <mbeans> | |
+ <!-- Newts :: AllMemmtables --> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=AllMemtablesLiveDataSize"> | |
+ <attrib name="Value" alias="alMemTblLiDaSi" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=AllMemtablesOffHeapDataSize"> | |
+ <attrib name="Value" alias="alMemTblOffHeapDaSi" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=AllMemtablesOnHeapDataSize"> | |
+ <attrib name="Value" alias="alMemTblOnHeapDaSi" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <!-- Memtable :: Count --> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=MemtableSwitchCount"> | |
+ <attrib name="Value" alias="memTblSwitchCount" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=MemtableColumnsCount"> | |
+ <attrib name="Value" alias="memTblColumnsCnt" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <!-- Memtable :: Sizes --> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=MemtableLiveDataSize"> | |
+ <attrib name="Value" alias="memTblLiveDaSi" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=MemtableOffHeapDataSize"> | |
+ <attrib name="Value" alias="memTblOffHeapDaSi" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=MemtableOnHeapDataSize"> | |
+ <attrib name="Value" alias="memTblOnHeapDaSi" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <!-- Latency --> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=ReadTotalLatency"> | |
+ <attrib name="Count" alias="readTotLtncy" type="counter"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=RangeLatency"> | |
+ <attrib name="99thPercentile" alias="rangeLtncy99" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=WriteTotalLatency"> | |
+ <attrib name="Count" alias="writeTotLtncy" type="counter"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=CasCommitTotalLatency"> | |
+ <attrib name="Count" alias="casCommitTotLtncy" type="counter"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=CasPrepareTotalLatency"> | |
+ <attrib name="Count" alias="casPrepareTotLtncy" type="counter"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=CasProposeTotalLatency"> | |
+ <attrib name="Count" alias="casProposeTotLtncy" type="counter"/> | |
+ </mbean> | |
+ | |
+ <!-- Bloom Filter --> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=BloomFilterDiskSpaceUsed"> | |
+ <attrib name="Value" alias="blmFltrDskSpcUsed" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=BloomFilterOffHeapMemoryUsed"> | |
+ <attrib name="Value" alias="blmFltrOffHeapMemUs" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <!-- Memory Used --> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=CompressionMetadataOffHeapMemoryUsed"> | |
+ <attrib name="Value" alias="cmpMetaOffHeapMemUs" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=IndexSummaryOffHeapMemoryUsed"> | |
+ <attrib name="Value" alias="idxSumOffHeapMemUs" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <!-- Pending --> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=PendingCompactions"> | |
+ <attrib name="Value" alias="pendingCompactions" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=PendingFlushes"> | |
+ <attrib name="Value" alias="pendingFlushes" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <!-- Disk Space --> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=TotalDiskSpaceUsed"> | |
+ <attrib name="Value" alias="totalDiskSpaceUsed" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Keyspace" | |
+ objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=LiveDiskSpaceUsed"> | |
+ <attrib name="Value" alias="liveDiskSpaceUsed" type="gauge"/> | |
+ </mbean> | |
+ </mbeans> | |
+ </jmx-collection> | |
+</jmx-datacollection-config> | |
diff --git a/jmx-datacollection-config.d/cassandra30x.xml b/jmx-datacollection-config.d/cassandra30x.xml | |
new file mode 100644 | |
index 0000000..897745b | |
--- /dev/null | |
+++ b/jmx-datacollection-config.d/cassandra30x.xml | |
@@ -0,0 +1,254 @@ | |
+<?xml version="1.0"?> | |
+<jmx-datacollection-config> | |
+ <jmx-collection name="jmx-cassandra30x"> | |
+ <rrd step="300"> | |
+ <rra>RRA:AVERAGE:0.5:1:2016</rra> | |
+ <rra>RRA:AVERAGE:0.5:12:1488</rra> | |
+ <rra>RRA:AVERAGE:0.5:288:366</rra> | |
+ <rra>RRA:MAX:0.5:288:366</rra> | |
+ <rra>RRA:MIN:0.5:288:366</rra> | |
+ </rrd> | |
+ <mbeans> | |
+ | |
+ <!-- Clients --> | |
+ <mbean name="org.apache.cassandra.metrics.Client" | |
+ objectname="org.apache.cassandra.metrics:type=Client,name=connectedNativeClients"> | |
+ <attrib name="Value" alias="clntConNativeClnts" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Client" | |
+ objectname="org.apache.cassandra.metrics:type=Client,name=connectedThriftClients"> | |
+ <attrib name="Value" alias="clntConThriftClnts" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <!-- Compaction --> | |
+ <mbean name="org.apache.cassandra.metrics.Compaction" | |
+ objectname="org.apache.cassandra.metrics:type=Compaction,name=BytesCompacted"> | |
+ <attrib name="Count" alias="cpctBytesCompacted" type="counter"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Compaction" | |
+ objectname="org.apache.cassandra.metrics:type=Compaction,name=CompletedTasks"> | |
+ <attrib name="Value" alias="cpctCompletedTasks" type="counter"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Compaction" | |
+ objectname="org.apache.cassandra.metrics:type=Compaction,name=PendingTasks"> | |
+ <attrib name="Value" alias="cpctPendingTasks" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <!-- Storage --> | |
+ <mbean name="org.apache.cassandra.metrics.Storage" | |
+ objectname="org.apache.cassandra.metrics:type=Storage,name=Load"> | |
+ <attrib name="Count" alias="strgLoad" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.Storage" | |
+ objectname="org.apache.cassandra.metrics:type=Storage,name=Exceptions"> | |
+ <attrib name="Count" alias="strgExceptions" type="counter"/> | |
+ </mbean> | |
+ | |
+ <!-- Dropped Messages --> | |
+ <mbean name="org.apache.cassandra.metrics.DroppedMessage" | |
+ objectname="org.apache.cassandra.metrics:type=DroppedMessage,scope=READ,name=Dropped"> | |
+ <attrib name="Count" alias="drpdMsgRead" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.DroppedMessage" | |
+ objectname="org.apache.cassandra.metrics:type=DroppedMessage,scope=READ_REPAIR,name=Dropped"> | |
+ <attrib name="Count" alias="drpdMsgReadRepair" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.DroppedMessage" | |
+ objectname="org.apache.cassandra.metrics:type=DroppedMessage,scope=REQUEST_RESPONSE,name=Dropped"> | |
+ <attrib name="Count" alias="drpdMsgReqResp" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.DroppedMessage" | |
+ objectname="org.apache.cassandra.metrics:type=DroppedMessage,scope=RANGE_SLICE,name=Dropped"> | |
+ <attrib name="Count" alias="drpdMsgRangeSlice" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.DroppedMessage" | |
+ objectname="org.apache.cassandra.metrics:type=DroppedMessage,scope=MUTATION,name=Dropped"> | |
+ <attrib name="Count" alias="drpdMsgMutation" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.DroppedMessage" | |
+ objectname="org.apache.cassandra.metrics:type=DroppedMessage,scope=PAGED_RANGE,name=Dropped"> | |
+ <attrib name="Count" alias="drpdMsgPagedRange" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <!-- ThreadPools :: MemtableFlushWriter --> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtableFlushWriter,name=ActiveTasks"> | |
+ <attrib name="Value" alias="tpIntMemTblFlsWrAt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtableFlushWriter,name=CurrentlyBlockedTasks"> | |
+ <attrib name="Count" alias="tpIntMemTblFlsWrCbt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtableFlushWriter,name=PendingTasks"> | |
+ <attrib name="Value" alias="tpIntMemTblFlsWrPt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtableFlushWriter,name=CompletedTasks"> | |
+ <attrib name="Count" alias="tpIntMemTblFlsWrCt" type="counter"/> | |
+ </mbean> | |
+ | |
+ <!-- ThreadPools :: MemtablePostFlush --> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtablePostFlush,name=ActiveTasks"> | |
+ <attrib name="Value" alias="tpIntMemTblPoFlsAt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtablePostFlush,name=CurrentlyBlockedTasks"> | |
+ <attrib name="Count" alias="tpIntMemTblPoFlsCbt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtablePostFlush,name=PendingTasks"> | |
+ <attrib name="Value" alias="tpIntMemTblPoFlsPt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtablePostFlush,name=CompletedTasks"> | |
+ <attrib name="Count" alias="tpIntMemTblPoFlsCt" type="counter"/> | |
+ </mbean> | |
+ | |
+ <!-- ThreadPools :: AntiEntropyStage --> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=AntiEntropyStage,name=ActiveTasks"> | |
+ <attrib name="Value" alias="tpIntAntiEntStgeAt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=AntiEntropyStage,name=CurrentlyBlockedTasks"> | |
+ <attrib name="Count" alias="tpIntAntiEntStgeCbt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=AntiEntropyStage,name=PendingTasks"> | |
+ <attrib name="Value" alias="tpIntAntiEntStgePt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=AntiEntropyStage,name=CompletedTasks"> | |
+ <attrib name="Value" alias="tpIntAntiEntStgeCt" type="counter"/> | |
+ </mbean> | |
+ | |
+ <!-- ThreadPools :: GossipStage --> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=GossipStage,name=ActiveTasks"> | |
+ <attrib name="Value" alias="tpIntGosStgeAt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=GossipStage,name=CurrentlyBlockedTasks"> | |
+ <attrib name="Count" alias="tpIntGosStgeCbt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=GossipStage,name=PendingTasks"> | |
+ <attrib name="Value" alias="tpIntGosStgePt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=GossipStage,name=CompletedTasks"> | |
+ <attrib name="Value" alias="tpIntGosStgeCt" type="counter"/> | |
+ </mbean> | |
+ | |
+ <!-- ThreadPools :: MigrationStage --> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MigrationStage,name=ActiveTasks"> | |
+ <attrib name="Value" alias="tpIntMigStgeAt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MigrationStage,name=CurrentlyBlockedTasks"> | |
+ <attrib name="Count" alias="tpIntMigStgeCbt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MigrationStage,name=PendingTasks"> | |
+ <attrib name="Value" alias="tpIntMigStgePt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MigrationStage,name=CompletedTasks"> | |
+ <attrib name="Value" alias="tpIntMigStgeCt" type="counter"/> | |
+ </mbean> | |
+ | |
+ <!-- ThreadPools :: MiscStage --> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MiscStage,name=ActiveTasks"> | |
+ <attrib name="Value" alias="tpIntMiscStgeAt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MiscStage,name=CurrentlyBlockedTasks"> | |
+ <attrib name="Count" alias="tpIntMiscStgeCbt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MiscStage,name=PendingTasks"> | |
+ <attrib name="Value" alias="tpIntMiscStgePt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MiscStage,name=CompletedTasks"> | |
+ <attrib name="Value" alias="tpIntMiscStgeCt" type="counter"/> | |
+ </mbean> | |
+ | |
+ <!-- ThreadPools :: MutationStage --> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=MutationStage,name=ActiveTasks"> | |
+ <attrib name="Value" alias="tpMutStgeAt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=MutationStage,name=CurrentlyBlockedTasks"> | |
+ <attrib name="Value" alias="tpMutStgeCbt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=MutationStage,name=PendingTasks"> | |
+ <attrib name="Value" alias="tpMutStgePt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=MutationStage,name=CompletedTasks"> | |
+ <attrib name="Value" alias="tpMutStgeCt" type="counter"/> | |
+ </mbean> | |
+ | |
+ <!-- ThreadPools :: ReadStage --> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadStage,name=ActiveTasks"> | |
+ <attrib name="Value" alias="tpReadStageAt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadStage,name=CurrentlyBlockedTasks"> | |
+ <attrib name="Value" alias="tpReadStageCbt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadStage,name=PendingTasks"> | |
+ <attrib name="Value" alias="tpReadStagePt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadStage,name=CompletedTasks"> | |
+ <attrib name="Value" alias="tpReadStageCt" type="counter"/> | |
+ </mbean> | |
+ | |
+ <!-- ThreadPools :: RequestResponseStage --> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=RequestResponseStage,name=ActiveTasks"> | |
+ <attrib name="Value" alias="tpReqRespStgeAt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=RequestResponseStage,name=CurrentlyBlockedTasks"> | |
+ <attrib name="Value" alias="tpReqRespStgeCbt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=RequestResponseStage,name=PendingTasks"> | |
+ <attrib name="Value" alias="tpReqRespStgePt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=RequestResponseStage,name=CompletedTasks"> | |
+ <attrib name="Value" alias="tpReqRespStgeCt" type="counter"/> | |
+ </mbean> | |
+ | |
+ <!-- ThreadPools :: ReadRepairStage --> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadRepairStage,name=ActiveTasks"> | |
+ <attrib name="Value" alias="tpReadRepairStgeAt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadRepairStage,name=CurrentlyBlockedTasks"> | |
+ <attrib name="Count" alias="tpReadRepairStgeCbt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadRepairStage,name=PendingTasks"> | |
+ <attrib name="Value" alias="tpReadRepairStgePt" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
+ objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadRepairStage,name=CompletedTasks"> | |
+ <attrib name="Value" alias="tpReadRepairStgeCt" type="counter"/> | |
+ </mbean> | |
+ </mbeans> | |
+ </jmx-collection> | |
+</jmx-datacollection-config> | |
diff --git a/jmx-datacollection-config.d/jboss.xml b/jmx-datacollection-config.d/jboss.xml | |
new file mode 100644 | |
index 0000000..3595bde | |
--- /dev/null | |
+++ b/jmx-datacollection-config.d/jboss.xml | |
@@ -0,0 +1,72 @@ | |
+<?xml version="1.0"?> | |
+<jmx-datacollection-config> | |
+ <jmx-collection name="jmx-jboss"> | |
+ <rrd step="300"> | |
+ <rra>RRA:AVERAGE:0.5:1:2016</rra> | |
+ <rra>RRA:AVERAGE:0.5:12:1488</rra> | |
+ <rra>RRA:AVERAGE:0.5:288:366</rra> | |
+ <rra>RRA:MAX:0.5:288:366</rra> | |
+ <rra>RRA:MIN:0.5:288:366</rra> | |
+ </rrd> | |
+ <mbeans> | |
+ <mbean name="SystemInfo" objectname="jboss.system:type=ServerInfo"> | |
+ <attrib name="FreeMemory" alias="FreeMemory" type="gauge"/> | |
+ <attrib name="TotalMemory" alias="TotalMemory" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="DefaultDSManagedConnectionPool" | |
+ objectname="jboss.jca:service=ManagedConnectionPool,name=DefaultDS"> | |
+ <attrib name="AvailableConnectionCount" alias="DefDS_AvailConns" type="gauge"/> | |
+ <attrib name="ConnectionCount" alias="DefDS_Conns" type="gauge"/> | |
+ <attrib name="InUseConnectionCount" alias="DefDS_InUseConns" type="gauge"/> | |
+ <attrib name="ConnectionCreatedCount" alias="DefDS_CreatedConns" type="counter"/> | |
+ <attrib name="ConnectionDestroyedCount" alias="DefDS_DestroyConns" type="counter"/> | |
+ </mbean> | |
+ <!-- JmsXA Connector Pool | |
+ <mbean name="JmsXAManagedConnectionPool" objectname="jboss.jca:service=ManagedConnectionPool,name=JmsXA"> | |
+ <attrib name="AvailableConnectionCount" alias="JmsXA_AvailConns" type="gauge"/> | |
+ <attrib name="ConnectionCount" alias="JmsXA_Conns" type="gauge"/> | |
+ <attrib name="InUseConnectionCount" alias="JmsXA_InUseConns" type="gauge"/> | |
+ <attrib name="ConnectionCreatedCount" alias="JmsXA_CreatedConns" type="gauge"/> | |
+ <attrib name="ConnectionDestroyedCount" alias="JmsXA_DestroyConns" type="gauge"/> | |
+ </mbean--> | |
+ <!-- Destination Manager | |
+ <mbean name="MQ_DestinationManager" objectname="jboss.mq:service=DestinationManager"> | |
+ <attrib name="ClientCount" alias="MQ_ClientCount" type="gauge"/> | |
+ </mbean--> | |
+ <!-- an example of a Queue | |
+ <mbean name="MQ_DestinationQueueA" objectname="jboss.mq.destination:service=Queue,name=A"> | |
+ <attrib name="QueueDepth" alias="A_QueueDepth" type="gauge"/> | |
+ <attrib name="ReceiversCount" alias="A_QueueRecv" type="gauge"/> | |
+ <attrib name="MaxDepth" alias="A_QueueMaxDepth" type="gauge"/> | |
+ </mbean--> | |
+ <!-- an example of a Topic | |
+ <mbean name="SecuredTopic" objectname="jboss.mq.destination:service=Topic,name=securedTopic"> | |
+ <attrib name="DurableMessageCount" alias="ST_DurMsg" type="gauge"/> | |
+ <attrib name="NonDurableMessageCount" alias="ST_NonDurMsg" type="gauge"/> | |
+ <attrib name="NonDurableSubscriptionCount" alias="ST_NonDurSub" type="gauge"/> | |
+ <attrib name="DurableSubscriptionCount" alias="ST_DurSub" type="gauge"/> | |
+ <attrib name="AllMessageCount" alias="ST_AllMsg" type="gauge"/> | |
+ <attrib name="MaxDepth" alias="ST_MaxDepth" type="gauge"/> | |
+ <attrib name="AllSubscriptionsCount" alias="ST_AllSub" type="gauge"/> | |
+ </mbean--> | |
+ <!-- Global Request Processor --> | |
+ <mbean name="GlobalRequestProcessor" | |
+ objectname="jboss.web:type=GlobalRequestProcessor,name=http-0.0.0.0-8080"> | |
+ <attrib name="requestCount" alias="GRP_requests" type="counter"/> | |
+ <attrib name="maxTime" alias="GRP_maxTime" type="gauge"/> | |
+ <attrib name="bytesSent" alias="GRP_bytesSent" type="counter"/> | |
+ <attrib name="bytesReceived" alias="GRP_bytesRec" type="counter"/> | |
+ <attrib name="processingTime" alias="GRP_procTime" type="counter"/> | |
+ <attrib name="errorCount" alias="GRP_errors" type="counter"/> | |
+ </mbean> | |
+ <!-- Thread Pool --> | |
+ <mbean name="ThreadPool" objectname="jboss.web:type=ThreadPool,name=http-0.0.0.0-8080"> | |
+ <attrib name="currentThreadsBusy" alias="BusyThreads" type="gauge"/> | |
+ <attrib name="currentThreadCount" alias="Threads" type="gauge"/> | |
+ <attrib name="minSpareThreads" alias="MinSpareThreads" type="gauge"/> | |
+ <attrib name="maxSpareThreads" alias="MaxSpareThreads" type="gauge"/> | |
+ <attrib name="maxThreads" alias="MaxThreads" type="gauge"/> | |
+ </mbean> | |
+ </mbeans> | |
+ </jmx-collection> | |
+</jmx-datacollection-config> | |
diff --git a/jmx-datacollection-config.d/kafka.xml b/jmx-datacollection-config.d/kafka.xml | |
new file mode 100644 | |
index 0000000..38d91e6 | |
--- /dev/null | |
+++ b/jmx-datacollection-config.d/kafka.xml | |
@@ -0,0 +1,240 @@ | |
+<?xml version="1.0"?> | |
+<jmx-datacollection-config rrdRepository="/opt/opennms/share/rrd/snmp/"> | |
+ <jmx-collection name="jmx-kafka"> | |
+ <rrd step="300"> | |
+ <rra>RRA:AVERAGE:0.5:1:2016</rra> | |
+ <rra>RRA:AVERAGE:0.5:12:1488</rra> | |
+ <rra>RRA:AVERAGE:0.5:288:366</rra> | |
+ <rra>RRA:MAX:0.5:288:366</rra> | |
+ <rra>RRA:MIN:0.5:288:366</rra> | |
+ </rrd> | |
+ <mbeans> | |
+ | |
+ <mbean name="JVM Memory" objectname="java.lang:type=OperatingSystem"> | |
+ <attrib name="FreePhysicalMemorySize" alias="FreeMemory" type="gauge"/> | |
+ <attrib name="TotalPhysicalMemorySize" alias="TotalMemory" type="gauge"/> | |
+ <attrib name="FreeSwapSpaceSize" alias="FreeSwapSpace" type="gauge"/> | |
+ <attrib name="TotalSwapSpaceSize" alias="TotalSwapSpace" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="JVM Threading" objectname="java.lang:type=Threading"> | |
+ <attrib name="ThreadCount" alias="ThreadCount" type="gauge"/> | |
+ <attrib name="PeakThreadCount" alias="PeakThreadCount" type="gauge"/> | |
+ <attrib name="DaemonThreadCount" alias="DaemonThreadCount" type="gauge"/> | |
+ <attrib name="CurrentThreadCpuTime" alias="CurThreadCpuTime" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="JVM GarbageCollector:MarkSweepCompact" objectname="java.lang:type=GarbageCollector,name=MarkSweepCompact"> | |
+ <attrib name="CollectionCount" alias="MSCCollCnt" type="counter"/> | |
+ <attrib name="CollectionTime" alias="MSCCollTime" type="counter"/> | |
+ <comp-attrib name="LastGcInfo" type="Composite" alias="MSCLastGcInfo"> | |
+ <comp-member name="GcThreadCount" type="gauge" alias="MSCGcThreadCnt"/> | |
+ <comp-member name="duration" type="gauge" alias="MSCDuration"/> | |
+ <comp-member name="endTime" type="gauge" alias="MSCEndTime"/> | |
+ </comp-attrib> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- ================== --> | |
+ <!-- BrokerTopicMetrics --> | |
+ <!-- ================== --> | |
+ <!-- https://kafka.apache.org/090/ops.html --> | |
+ <!-- https://www.datadoghq.com/blog/monitoring-kafka-performance-metrics/ --> | |
+ | |
+ <mbean name="Messages In Per Second" objectname="kafka.server:type=BrokerTopicMetrics,name=MessagesInPerSec"> | |
+ <attrib name="Count" alias="msgInPerSec" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Bytes In Per Second" objectname="kafka.server:type=BrokerTopicMetrics,name=BytesInPerSec"> | |
+ <attrib name="Count" alias="bytesInPerSec" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Bytes Out Per Second" objectname="kafka.server:type=BrokerTopicMetrics,name=BytesOutPerSec"> | |
+ <attrib name="Count" alias="bytesOutPerSec" type="gauge"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- ============== --> | |
+ <!-- ReplicaManager --> | |
+ <!-- ============== --> | |
+ <!-- https://kafka.apache.org/090/ops.html --> | |
+ | |
+ <!-- https://www.datadoghq.com/blog/monitoring-kafka-performance-metrics/ --> | |
+ <mbean name="Under-Replicated Partitions" objectname="kafka.server:type=ReplicaManager,name=UnderReplicatedPartitions"> | |
+ <attrib name="Value" alias="underReplPart" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Partitions" objectname="kafka.server:type=ReplicaManager,name=PartitionCount"> | |
+ <attrib name="Value" alias="partitions" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Leaders" objectname="kafka.server:type=ReplicaManager,name=LeaderCount"> | |
+ <attrib name="Value" alias="leaders" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <!-- https://www.datadoghq.com/blog/monitoring-kafka-performance-metrics/ --> | |
+ <mbean name="In-Sync Replica Shrinks Per Second" objectname="kafka.server:type=ReplicaManager,name=IsrShrinksPerSec"> | |
+ <attrib name="Count" alias="isrShrinksPerSec" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <!-- https://www.datadoghq.com/blog/monitoring-kafka-performance-metrics/ --> | |
+ <mbean name="In-Sync Replica Expands Per Second" objectname="kafka.server:type=ReplicaManager,name=IsrExpandsPerSec"> | |
+ <attrib name="Count" alias="isrExpandsPerSec" type="gauge"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- ===================== --> | |
+ <!-- ReplicaFetcherManager --> | |
+ <!-- ===================== --> | |
+ <!-- https://kafka.apache.org/090/ops.html --> | |
+ | |
+ <mbean name="Maximum Lag Between Replicas" objectname="kafka.server:type=ReplicaFetcherManager,name=MaxLag,clientId=Replica"> | |
+ <attrib name="Value" alias="replicaMaxLag" type="gauge"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- =============== --> | |
+ <!-- KafkaController --> | |
+ <!-- =============== --> | |
+ <!-- https://kafka.apache.org/090/ops.html --> | |
+ <!-- https://www.datadoghq.com/blog/monitoring-kafka-performance-metrics/ --> | |
+ | |
+ <mbean name="Active Controllers" objectname="kafka.controller:type=KafkaController,name=ActiveControllerCount"> | |
+ <attrib name="Value" alias="activeControllers" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Offline Partitions" objectname="kafka.controller:type=KafkaController,name=OfflinePartitionsCount"> | |
+ <attrib name="Value" alias="offlinePartitions" type="gauge"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- =============== --> | |
+ <!-- ControllerStats --> | |
+ <!-- =============== --> | |
+ <!-- https://kafka.apache.org/090/ops.html --> | |
+ <!-- https://www.datadoghq.com/blog/monitoring-kafka-performance-metrics/ --> | |
+ | |
+ <mbean name="Leader Election Rate and Time (ms)" objectname="kafka.controller:type=ControllerStats,name=LeaderElectionRateAndTimeMs"> | |
+ <attrib name="Count" alias="leaderEps" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Unclean Leader Elections Per Second" objectname="kafka.controller:type=ControllerStats,name=UncleanLeaderElectionsPerSec"> | |
+ <attrib name="Count" alias="uncleanLeaderEps" type="gauge"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- ========================= --> | |
+ <!-- DelayedOperationPurgatory --> | |
+ <!-- ========================= --> | |
+ <!-- https://kafka.apache.org/090/ops.html --> | |
+ <!-- https://www.datadoghq.com/blog/monitoring-kafka-performance-metrics/ --> | |
+ | |
+ <mbean name="Purgatory Size: Produce" objectname="kafka.server:type=DelayedOperationPurgatory,name=PurgatorySize,delayedOperation=Produce"> | |
+ <attrib name="Value" alias="purgatoryProduce" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Purgatory Size: Fetch" objectname="kafka.server:type=DelayedOperationPurgatory,name=PurgatorySize,delayedOperation=Fetch"> | |
+ <attrib name="Value" alias="purgatoryFetch" type="gauge"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- ============ --> | |
+ <!-- SocketServer --> | |
+ <!-- ============ --> | |
+ <!-- https://kafka.apache.org/090/ops.html --> | |
+ | |
+ <mbean name="Network Processor Average Idle Percentage" objectname="kafka.network:type=SocketServer,name=NetworkProcessorAvgIdlePercent"> | |
+ <attrib name="Value" alias="netProcAvgIdle" type="gauge"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- ======================= --> | |
+ <!-- KafkaRequestHandlerPool --> | |
+ <!-- ======================= --> | |
+ <!-- https://kafka.apache.org/090/ops.html --> | |
+ | |
+ <mbean name="Request Handler Average Idle Percentage" objectname="kafka.server:type=KafkaRequestHandlerPool,name=RequestHandlerAvgIdlePercent"> | |
+ <attrib name="Value" alias="reqHandAvgIdle" type="gauge"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- ============== --> | |
+ <!-- RequestMetrics --> | |
+ <!-- ============== --> | |
+ <!-- https://kafka.apache.org/090/ops.html --> | |
+ <!-- https://www.datadoghq.com/blog/monitoring-kafka-performance-metrics/ --> | |
+ | |
+ <mbean name="Requests Per Second: Produce" objectname="kafka.network:type=RequestMetrics,name=RequestsPerSec,request=Produce"> | |
+ <attrib name="Count" alias="reqSecProduce" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Requests Per Second: FetchConsumer" objectname="kafka.network:type=RequestMetrics,name=RequestsPerSec,request=FetchConsumer"> | |
+ <attrib name="Count" alias="reqSecConsumer" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Requests Per Second: FetchFollower" objectname="kafka.network:type=RequestMetrics,name=RequestsPerSec,request=FetchFollower"> | |
+ <attrib name="Count" alias="reqSecFollower" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Total Time: Produce" objectname="kafka.network:type=RequestMetrics,name=TotalTimeMs,request=Produce"> | |
+ <attrib name="Count" alias="totalTimeProduce" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Total Time: FetchConsumer" objectname="kafka.network:type=RequestMetrics,name=TotalTimeMs,request=FetchConsumer"> | |
+ <attrib name="Count" alias="totalTimeConsumer" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Total Time: FetchFollower" objectname="kafka.network:type=RequestMetrics,name=TotalTimeMs,request=FetchFollower"> | |
+ <attrib name="Count" alias="totalTimeFollower" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Queue Time: Produce" objectname="kafka.network:type=RequestMetrics,name=RequestQueueTimeMs,request=Produce"> | |
+ <attrib name="Count" alias="queueTimeProduce" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Queue Time: FetchConsumer" objectname="kafka.network:type=RequestMetrics,name=RequestQueueTimeMs,request=FetchConsumer"> | |
+ <attrib name="Count" alias="queueTimeConsumer" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Queue Time: FetchFollower" objectname="kafka.network:type=RequestMetrics,name=RequestQueueTimeMs,request=FetchFollower"> | |
+ <attrib name="Count" alias="queueTimeFollower" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Local Time: Produce" objectname="kafka.network:type=RequestMetrics,name=LocalTimeMs,request=Produce"> | |
+ <attrib name="Count" alias="localTimeProduce" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Local Time: FetchConsumer" objectname="kafka.network:type=RequestMetrics,name=LocalTimeMs,request=FetchConsumer"> | |
+ <attrib name="Count" alias="localTimeConsumer" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Local Time: FetchFollower" objectname="kafka.network:type=RequestMetrics,name=LocalTimeMs,request=FetchFollower"> | |
+ <attrib name="Count" alias="localTimeFollower" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Remote Time: Produce" objectname="kafka.network:type=RequestMetrics,name=RemoteTimeMs,request=Produce"> | |
+ <attrib name="Count" alias="remoteTimeProduce" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Remote Time: FetchConsumer" objectname="kafka.network:type=RequestMetrics,name=RemoteTimeMs,request=FetchConsumer"> | |
+ <attrib name="Count" alias="remoteTimeConsumer" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Remote Time: FetchFollower" objectname="kafka.network:type=RequestMetrics,name=RemoteTimeMs,request=FetchFollower"> | |
+ <attrib name="Count" alias="remoteTimeFollower" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Response Send Time: Produce" objectname="kafka.network:type=RequestMetrics,name=ResponseSendTimeMs,request=Produce"> | |
+ <attrib name="Count" alias="sendTimeProduce" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Response Send Time: FetchConsumer" objectname="kafka.network:type=RequestMetrics,name=ResponseSendTimeMs,request=FetchConsumer"> | |
+ <attrib name="Count" alias="sendTimeConsumer" type="gauge"/> | |
+ </mbean> | |
+ | |
+ <mbean name="Response Send Time: FetchFollower" objectname="kafka.network:type=RequestMetrics,name=ResponseSendTimeMs,request=FetchFollower"> | |
+ <attrib name="Count" alias="sendTimeFollower" type="gauge"/> | |
+ </mbean> | |
+ | |
+ </mbeans> | |
+ | |
+ </jmx-collection> | |
+</jmx-datacollection-config> | |
diff --git a/jmx-datacollection-config.d/minion.xml b/jmx-datacollection-config.d/minion.xml | |
new file mode 100644 | |
index 0000000..e8f0f88 | |
--- /dev/null | |
+++ b/jmx-datacollection-config.d/minion.xml | |
@@ -0,0 +1,135 @@ | |
+<?xml version="1.0"?> | |
+<jmx-datacollection-config rrdRepository="/opt/opennms/share/rrd/snmp/"> | |
+ <jmx-collection name="jmx-minion"> | |
+ <rrd step="300"> | |
+ <rra>RRA:AVERAGE:0.5:1:2016</rra> | |
+ <rra>RRA:AVERAGE:0.5:12:1488</rra> | |
+ <rra>RRA:AVERAGE:0.5:288:366</rra> | |
+ <rra>RRA:MAX:0.5:288:366</rra> | |
+ <rra>RRA:MIN:0.5:288:366</rra> | |
+ </rrd> | |
+ <mbeans> | |
+ <mbean name="JVM Memory" objectname="java.lang:type=OperatingSystem"> | |
+ <attrib name="FreePhysicalMemorySize" alias="FreeMemory" type="gauge"/> | |
+ <attrib name="TotalPhysicalMemorySize" alias="TotalMemory" type="gauge"/> | |
+ <attrib name="FreeSwapSpaceSize" alias="FreeSwapSpace" type="gauge"/> | |
+ <attrib name="TotalSwapSpaceSize" alias="TotalSwapSpace" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="JVM Threading" objectname="java.lang:type=Threading"> | |
+ <attrib name="ThreadCount" alias="ThreadCount" type="gauge"/> | |
+ <attrib name="PeakThreadCount" alias="PeakThreadCount" type="gauge"/> | |
+ <attrib name="DaemonThreadCount" alias="DaemonThreadCount" type="gauge"/> | |
+ <attrib name="CurrentThreadCpuTime" alias="CurThreadCpuTime" type="gauge"/> | |
+ </mbean> | |
+ <mbean name="JVM GarbageCollector:MarkSweepCompact" objectname="java.lang:type=GarbageCollector,name=MarkSweepCompact"> | |
+ <attrib name="CollectionCount" alias="MSCCollCnt" type="counter"/> | |
+ <attrib name="CollectionTime" alias="MSCCollTime" type="counter"/> | |
+ <comp-attrib name="LastGcInfo" type="Composite" alias="MSCLastGcInfo"> | |
+ <comp-member name="GcThreadCount" type="gauge" alias="MSCGcThreadCnt"/> | |
+ <comp-member name="duration" type="gauge" alias="MSCDuration"/> | |
+ <comp-member name="endTime" type="gauge" alias="MSCEndTime"/> | |
+ </comp-attrib> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- Route stats for syslogListen --> | |
+ <mbean name="Syslog Listener" objectname="org.apache.camel:context=syslogdListenerCamelNettyContext,type=routes,name="syslogListen""> | |
+ <attrib name="ExchangesCompleted" alias="SlogListComplete" type="counter"/> | |
+ <attrib name="ExchangesFailed" alias="SlogListFailed" type="counter"/> | |
+ <attrib name="ExchangesTotal" alias="SlogListTotal" type="counter"/> | |
+ <attrib name="MaxProcessingTime" alias="SlogListMaxProc" type="gauge"/> | |
+ <attrib name="MeanProcessingTime" alias="SlogListMeanProc" type="gauge"/> | |
+ <attrib name="MinProcessingTime" alias="SlogListMinProc" type="gauge"/> | |
+ <attrib name="LastProcessingTime" alias="SlogListLastProc" type="gauge"/> | |
+ <attrib name="TotalProcessingTime" alias="SlogListTotProc" type="counter"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- Route stats for RPC.Server.Detect --> | |
+ <mbean name="Provisioning Detectors RPC Server" objectname="org.apache.camel:context=org.opennms.core.ipc.rpc.camel-impl,type=routes,name="RPC.Server.Detect""> | |
+ <attrib name="ExchangesCompleted" alias="DetectComplete" type="counter"/> | |
+ <attrib name="ExchangesFailed" alias="DetectFailed" type="counter"/> | |
+ <attrib name="ExchangesTotal" alias="DetectTotal" type="counter"/> | |
+ <attrib name="MaxProcessingTime" alias="DetectMaxProc" type="gauge"/> | |
+ <attrib name="MeanProcessingTime" alias="DetectMeanProc" type="gauge"/> | |
+ <attrib name="MinProcessingTime" alias="DetectMinProc" type="gauge"/> | |
+ <attrib name="LastProcessingTime" alias="DetectLastProc" type="gauge"/> | |
+ <attrib name="TotalProcessingTime" alias="DetectTotProc" type="counter"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- Route stats for RPC.Server.DNS --> | |
+ <mbean name="DNS RPC Server" objectname="org.apache.camel:context=org.opennms.core.ipc.rpc.camel-impl,type=routes,name="RPC.Server.DNS""> | |
+ <attrib name="ExchangesCompleted" alias="DnsComplete" type="counter"/> | |
+ <attrib name="ExchangesFailed" alias="DnsFailed" type="counter"/> | |
+ <attrib name="ExchangesTotal" alias="DnsTotal" type="counter"/> | |
+ <attrib name="MaxProcessingTime" alias="DnsMaxProc" type="gauge"/> | |
+ <attrib name="MeanProcessingTime" alias="DnsMeanProc" type="gauge"/> | |
+ <attrib name="MinProcessingTime" alias="DnsMinProc" type="gauge"/> | |
+ <attrib name="LastProcessingTime" alias="DnsLastProc" type="gauge"/> | |
+ <attrib name="TotalProcessingTime" alias="DnsTotProc" type="counter"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- Route stats for RPC.Server.PING --> | |
+ <mbean name="Ping RPC Server" objectname="org.apache.camel:context=org.opennms.core.ipc.rpc.camel-impl,type=routes,name="RPC.Server.PING""> | |
+ <attrib name="ExchangesCompleted" alias="PingComplete" type="counter"/> | |
+ <attrib name="ExchangesFailed" alias="PingFailed" type="counter"/> | |
+ <attrib name="ExchangesTotal" alias="PingTotal" type="counter"/> | |
+ <attrib name="MaxProcessingTime" alias="PingMaxProc" type="gauge"/> | |
+ <attrib name="MeanProcessingTime" alias="PingMeanProc" type="gauge"/> | |
+ <attrib name="MinProcessingTime" alias="PingMinProc" type="gauge"/> | |
+ <attrib name="LastProcessingTime" alias="PingLastProc" type="gauge"/> | |
+ <attrib name="TotalProcessingTime" alias="PingTotProc" type="counter"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- Route stats for RPC.Server.PING-SWEEP --> | |
+ <mbean name="Ping Sweep RPC Server" objectname="org.apache.camel:context=org.opennms.core.ipc.rpc.camel-impl,type=routes,name="RPC.Server.PING-SWEEP""> | |
+ <attrib name="ExchangesCompleted" alias="SweepComplete" type="counter"/> | |
+ <attrib name="ExchangesFailed" alias="SweepFailed" type="counter"/> | |
+ <attrib name="ExchangesTotal" alias="SweepTotal" type="counter"/> | |
+ <attrib name="MaxProcessingTime" alias="SweepMaxProc" type="gauge"/> | |
+ <attrib name="MeanProcessingTime" alias="SweepMeanProc" type="gauge"/> | |
+ <attrib name="MinProcessingTime" alias="SweepMinProc" type="gauge"/> | |
+ <attrib name="LastProcessingTime" alias="SweepLastProc" type="gauge"/> | |
+ <attrib name="TotalProcessingTime" alias="SweepTotProc" type="counter"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- Route stats for RPC.Server.Poller --> | |
+ <mbean name="Poller RPC Server" objectname="org.apache.camel:context=org.opennms.core.ipc.rpc.camel-impl,type=routes,name="RPC.Server.Poller""> | |
+ <attrib name="ExchangesCompleted" alias="PollComplete" type="counter"/> | |
+ <attrib name="ExchangesFailed" alias="PollFailed" type="counter"/> | |
+ <attrib name="ExchangesTotal" alias="PollTotal" type="counter"/> | |
+ <attrib name="MaxProcessingTime" alias="PollMaxProc" type="gauge"/> | |
+ <attrib name="MeanProcessingTime" alias="PollMeanProc" type="gauge"/> | |
+ <attrib name="MinProcessingTime" alias="PollMinProc" type="gauge"/> | |
+ <attrib name="LastProcessingTime" alias="PollLastProc" type="gauge"/> | |
+ <attrib name="TotalProcessingTime" alias="PollTotProc" type="counter"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- Route stats for RPC.Server.SNMP --> | |
+ <mbean name="SNMP RPC Server" objectname="org.apache.camel:context=org.opennms.core.ipc.rpc.camel-impl,type=routes,name="RPC.Server.SNMP""> | |
+ <attrib name="ExchangesCompleted" alias="SnmpComplete" type="counter"/> | |
+ <attrib name="ExchangesFailed" alias="SnmpFailed" type="counter"/> | |
+ <attrib name="ExchangesTotal" alias="SnmpTotal" type="counter"/> | |
+ <attrib name="MaxProcessingTime" alias="SnmpMaxProc" type="gauge"/> | |
+ <attrib name="MeanProcessingTime" alias="SnmpMeanProc" type="gauge"/> | |
+ <attrib name="MinProcessingTime" alias="SnmpMinProc" type="gauge"/> | |
+ <attrib name="LastProcessingTime" alias="SnmpLastProc" type="gauge"/> | |
+ <attrib name="TotalProcessingTime" alias="SnmpTotProc" type="counter"/> | |
+ </mbean> | |
+ | |
+ | |
+ <!-- | |
+ Example of SEDA endpoint collection that won't work well because the objectname is based on the URI which might change | |
+ | |
+ <mbean name="Syslog Marshaller Queue" objectname="org.apache.camel:context=syslogdHandlerMinionContext,type=endpoints,name="seda://handleMessage?concurrentConsumers=4""> | |
+ <attrib name="CurrentQueueSize" alias="SlogMarQueue" type="gauge"/> | |
+ </mbean> | |
+ --> | |
+ </mbeans> | |
+ </jmx-collection> | |
+</jmx-datacollection-config> | |
diff --git a/jmx-datacollection-config.xml b/jmx-datacollection-config.xml | |
index b49c44b..44c0afc 100644 | |
--- a/jmx-datacollection-config.xml | |
+++ b/jmx-datacollection-config.xml | |
@@ -1,74 +1,5 @@ | |
<?xml version="1.0"?> | |
<jmx-datacollection-config rrdRepository="/opt/opennms/share/rrd/snmp/"> | |
- <jmx-collection name="jboss"> | |
- <rrd step="300"> | |
- <rra>RRA:AVERAGE:0.5:1:2016</rra> | |
- <rra>RRA:AVERAGE:0.5:12:1488</rra> | |
- <rra>RRA:AVERAGE:0.5:288:366</rra> | |
- <rra>RRA:MAX:0.5:288:366</rra> | |
- <rra>RRA:MIN:0.5:288:366</rra> | |
- </rrd> | |
- <mbeans> | |
- <mbean name="SystemInfo" objectname="jboss.system:type=ServerInfo"> | |
- <attrib name="FreeMemory" alias="FreeMemory" type="gauge"/> | |
- <attrib name="TotalMemory" alias="TotalMemory" type="gauge"/> | |
- </mbean> | |
- <mbean name="DefaultDSManagedConnectionPool" | |
- objectname="jboss.jca:service=ManagedConnectionPool,name=DefaultDS"> | |
- <attrib name="AvailableConnectionCount" alias="DefDS_AvailConns" type="gauge"/> | |
- <attrib name="ConnectionCount" alias="DefDS_Conns" type="gauge"/> | |
- <attrib name="InUseConnectionCount" alias="DefDS_InUseConns" type="gauge"/> | |
- <attrib name="ConnectionCreatedCount" alias="DefDS_CreatedConns" type="counter"/> | |
- <attrib name="ConnectionDestroyedCount" alias="DefDS_DestroyConns" type="counter"/> | |
- </mbean> | |
- <!-- JmsXA Connector Pool | |
- <mbean name="JmsXAManagedConnectionPool" objectname="jboss.jca:service=ManagedConnectionPool,name=JmsXA"> | |
- <attrib name="AvailableConnectionCount" alias="JmsXA_AvailConns" type="gauge"/> | |
- <attrib name="ConnectionCount" alias="JmsXA_Conns" type="gauge"/> | |
- <attrib name="InUseConnectionCount" alias="JmsXA_InUseConns" type="gauge"/> | |
- <attrib name="ConnectionCreatedCount" alias="JmsXA_CreatedConns" type="gauge"/> | |
- <attrib name="ConnectionDestroyedCount" alias="JmsXA_DestroyConns" type="gauge"/> | |
- </mbean--> | |
- <!-- Destination Manager | |
- <mbean name="MQ_DestinationManager" objectname="jboss.mq:service=DestinationManager"> | |
- <attrib name="ClientCount" alias="MQ_ClientCount" type="gauge"/> | |
- </mbean--> | |
- <!-- an example of a Queue | |
- <mbean name="MQ_DestinationQueueA" objectname="jboss.mq.destination:service=Queue,name=A"> | |
- <attrib name="QueueDepth" alias="A_QueueDepth" type="gauge"/> | |
- <attrib name="ReceiversCount" alias="A_QueueRecv" type="gauge"/> | |
- <attrib name="MaxDepth" alias="A_QueueMaxDepth" type="gauge"/> | |
- </mbean--> | |
- <!-- an example of a Topic | |
- <mbean name="SecuredTopic" objectname="jboss.mq.destination:service=Topic,name=securedTopic"> | |
- <attrib name="DurableMessageCount" alias="ST_DurMsg" type="gauge"/> | |
- <attrib name="NonDurableMessageCount" alias="ST_NonDurMsg" type="gauge"/> | |
- <attrib name="NonDurableSubscriptionCount" alias="ST_NonDurSub" type="gauge"/> | |
- <attrib name="DurableSubscriptionCount" alias="ST_DurSub" type="gauge"/> | |
- <attrib name="AllMessageCount" alias="ST_AllMsg" type="gauge"/> | |
- <attrib name="MaxDepth" alias="ST_MaxDepth" type="gauge"/> | |
- <attrib name="AllSubscriptionsCount" alias="ST_AllSub" type="gauge"/> | |
- </mbean--> | |
- <!-- Global Request Processor --> | |
- <mbean name="GlobalRequestProcessor" | |
- objectname="jboss.web:type=GlobalRequestProcessor,name=http-0.0.0.0-8080"> | |
- <attrib name="requestCount" alias="GRP_requests" type="counter"/> | |
- <attrib name="maxTime" alias="GRP_maxTime" type="gauge"/> | |
- <attrib name="bytesSent" alias="GRP_bytesSent" type="counter"/> | |
- <attrib name="bytesReceived" alias="GRP_bytesRec" type="counter"/> | |
- <attrib name="processingTime" alias="GRP_procTime" type="counter"/> | |
- <attrib name="errorCount" alias="GRP_errors" type="counter"/> | |
- </mbean> | |
- <!-- Thread Pool --> | |
- <mbean name="ThreadPool" objectname="jboss.web:type=ThreadPool,name=http-0.0.0.0-8080"> | |
- <attrib name="currentThreadsBusy" alias="BusyThreads" type="gauge"/> | |
- <attrib name="currentThreadCount" alias="Threads" type="gauge"/> | |
- <attrib name="minSpareThreads" alias="MinSpareThreads" type="gauge"/> | |
- <attrib name="maxSpareThreads" alias="MaxSpareThreads" type="gauge"/> | |
- <attrib name="maxThreads" alias="MaxThreads" type="gauge"/> | |
- </mbean> | |
- </mbeans> | |
- </jmx-collection> | |
<jmx-collection name="jsr160"> | |
<rrd step="300"> | |
<rra>RRA:AVERAGE:0.5:1:2016</rra> | |
@@ -368,375 +299,23 @@ | |
<attrib name="999thPercentile" alias="CasCluster1Req999" type="gauge"/> | |
<attrib name="Count" alias="CasCluster1ReqCnt" type="counter"/> | |
</mbean> | |
- </mbeans> | |
- </jmx-collection> | |
- | |
- <jmx-collection name="cassandra21x"> | |
- <rrd step="300"> | |
- <rra>RRA:AVERAGE:0.5:1:2016</rra> | |
- <rra>RRA:AVERAGE:0.5:12:1488</rra> | |
- <rra>RRA:AVERAGE:0.5:288:366</rra> | |
- <rra>RRA:MAX:0.5:288:366</rra> | |
- <rra>RRA:MIN:0.5:288:366</rra> | |
- </rrd> | |
- <mbeans> | |
- | |
- <!-- Clients --> | |
- <mbean name="org.apache.cassandra.metrics.Client" | |
- objectname="org.apache.cassandra.metrics:type=Client,name=connectedNativeClients"> | |
- <attrib name="Value" alias="clntConNativeClnts" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Client" | |
- objectname="org.apache.cassandra.metrics:type=Client,name=connectedThriftClients"> | |
- <attrib name="Value" alias="clntConThriftClnts" type="gauge"/> | |
- </mbean> | |
- | |
- <!-- Compaction --> | |
- <mbean name="org.apache.cassandra.metrics.Compaction" | |
- objectname="org.apache.cassandra.metrics:type=Compaction,name=BytesCompacted"> | |
- <attrib name="Count" alias="cpctBytesCompacted" type="counter"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Compaction" | |
- objectname="org.apache.cassandra.metrics:type=Compaction,name=CompletedTasks"> | |
- <attrib name="Value" alias="cpctCompletedTasks" type="counter"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Compaction" | |
- objectname="org.apache.cassandra.metrics:type=Compaction,name=PendingTasks"> | |
- <attrib name="Value" alias="cpctPendingTasks" type="gauge"/> | |
- </mbean> | |
- | |
- <!-- Storage --> | |
- <mbean name="org.apache.cassandra.metrics.Storage" | |
- objectname="org.apache.cassandra.metrics:type=Storage,name=Load"> | |
- <attrib name="Count" alias="strgLoad" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Storage" | |
- objectname="org.apache.cassandra.metrics:type=Storage,name=Exceptions"> | |
- <attrib name="Count" alias="strgExceptions" type="counter"/> | |
- </mbean> | |
- | |
- <!-- Dropped Messages --> | |
- <mbean name="org.apache.cassandra.metrics.DroppedMessage" | |
- objectname="org.apache.cassandra.metrics:type=DroppedMessage,scope=READ,name=Dropped"> | |
- <attrib name="Count" alias="drpdMsgRead" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.DroppedMessage" | |
- objectname="org.apache.cassandra.metrics:type=DroppedMessage,scope=READ_REPAIR,name=Dropped"> | |
- <attrib name="Count" alias="drpdMsgReadRepair" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.DroppedMessage" | |
- objectname="org.apache.cassandra.metrics:type=DroppedMessage,scope=REQUEST_RESPONSE,name=Dropped"> | |
- <attrib name="Count" alias="drpdMsgReqResp" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.DroppedMessage" | |
- objectname="org.apache.cassandra.metrics:type=DroppedMessage,scope=RANGE_SLICE,name=Dropped"> | |
- <attrib name="Count" alias="drpdMsgRangeSlice" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.DroppedMessage" | |
- objectname="org.apache.cassandra.metrics:type=DroppedMessage,scope=MUTATION,name=Dropped"> | |
- <attrib name="Count" alias="drpdMsgMutation" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.DroppedMessage" | |
- objectname="org.apache.cassandra.metrics:type=DroppedMessage,scope=PAGED_RANGE,name=Dropped"> | |
- <attrib name="Count" alias="drpdMsgPagedRange" type="gauge"/> | |
- </mbean> | |
- | |
- <!-- ThreadPools :: MemtableFlushWriter --> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtableFlushWriter,name=ActiveTasks"> | |
- <attrib name="Value" alias="tpIntMemTblFlsWrAt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtableFlushWriter,name=CurrentlyBlockedTasks"> | |
- <attrib name="Count" alias="tpIntMemTblFlsWrCbt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtableFlushWriter,name=PendingTasks"> | |
- <attrib name="Value" alias="tpIntMemTblFlsWrPt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtableFlushWriter,name=CompletedTasks"> | |
- <attrib name="Count" alias="tpIntMemTblFlsWrCt" type="counter"/> | |
- </mbean> | |
- | |
- <!-- ThreadPools :: MemtablePostFlush --> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtablePostFlush,name=ActiveTasks"> | |
- <attrib name="Value" alias="tpIntMemTblPoFlsAt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtablePostFlush,name=CurrentlyBlockedTasks"> | |
- <attrib name="Count" alias="tpIntMemTblPoFlsCbt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtablePostFlush,name=PendingTasks"> | |
- <attrib name="Value" alias="tpIntMemTblPoFlsPt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MemtablePostFlush,name=CompletedTasks"> | |
- <attrib name="Count" alias="tpIntMemTblPoFlsCt" type="counter"/> | |
- </mbean> | |
- | |
- <!-- ThreadPools :: AntiEntropyStage --> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=AntiEntropyStage,name=ActiveTasks"> | |
- <attrib name="Value" alias="tpIntAntiEntStgeAt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=AntiEntropyStage,name=CurrentlyBlockedTasks"> | |
- <attrib name="Count" alias="tpIntAntiEntStgeCbt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=AntiEntropyStage,name=PendingTasks"> | |
- <attrib name="Value" alias="tpIntAntiEntStgePt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=AntiEntropyStage,name=CompletedTasks"> | |
- <attrib name="Value" alias="tpIntAntiEntStgeCt" type="counter"/> | |
- </mbean> | |
- | |
- <!-- ThreadPools :: GossipStage --> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=GossipStage,name=ActiveTasks"> | |
- <attrib name="Value" alias="tpIntGosStgeAt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=GossipStage,name=CurrentlyBlockedTasks"> | |
- <attrib name="Count" alias="tpIntGosStgeCbt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=GossipStage,name=PendingTasks"> | |
- <attrib name="Value" alias="tpIntGosStgePt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=GossipStage,name=CompletedTasks"> | |
- <attrib name="Value" alias="tpIntGosStgeCt" type="counter"/> | |
- </mbean> | |
- | |
- <!-- ThreadPools :: MigrationStage --> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MigrationStage,name=ActiveTasks"> | |
- <attrib name="Value" alias="tpIntMigStgeAt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MigrationStage,name=CurrentlyBlockedTasks"> | |
- <attrib name="Count" alias="tpIntMigStgeCbt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MigrationStage,name=PendingTasks"> | |
- <attrib name="Value" alias="tpIntMigStgePt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MigrationStage,name=CompletedTasks"> | |
- <attrib name="Value" alias="tpIntMigStgeCt" type="counter"/> | |
- </mbean> | |
- | |
- <!-- ThreadPools :: MiscStage --> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MiscStage,name=ActiveTasks"> | |
- <attrib name="Value" alias="tpIntMiscStgeAt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MiscStage,name=CurrentlyBlockedTasks"> | |
- <attrib name="Count" alias="tpIntMiscStgeCbt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MiscStage,name=PendingTasks"> | |
- <attrib name="Value" alias="tpIntMiscStgePt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=internal,scope=MiscStage,name=CompletedTasks"> | |
- <attrib name="Value" alias="tpIntMiscStgeCt" type="counter"/> | |
- </mbean> | |
- | |
- <!-- ThreadPools :: MutationStage --> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=MutationStage,name=ActiveTasks"> | |
- <attrib name="Value" alias="tpMutStgeAt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=MutationStage,name=CurrentlyBlockedTasks"> | |
- <attrib name="Value" alias="tpMutStgeCbt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=MutationStage,name=PendingTasks"> | |
- <attrib name="Value" alias="tpMutStgePt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=MutationStage,name=CompletedTasks"> | |
- <attrib name="Value" alias="tpMutStgeCt" type="counter"/> | |
- </mbean> | |
- | |
- <!-- ThreadPools :: ReadStage --> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadStage,name=ActiveTasks"> | |
- <attrib name="Value" alias="tpReadStageAt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadStage,name=CurrentlyBlockedTasks"> | |
- <attrib name="Value" alias="tpReadStageCbt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadStage,name=PendingTasks"> | |
- <attrib name="Value" alias="tpReadStagePt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadStage,name=CompletedTasks"> | |
- <attrib name="Value" alias="tpReadStageCt" type="counter"/> | |
- </mbean> | |
- | |
- <!-- ThreadPools :: RequestResponseStage --> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=RequestResponseStage,name=ActiveTasks"> | |
- <attrib name="Value" alias="tpReqRespStgeAt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=RequestResponseStage,name=CurrentlyBlockedTasks"> | |
- <attrib name="Value" alias="tpReqRespStgeCbt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=RequestResponseStage,name=PendingTasks"> | |
- <attrib name="Value" alias="tpReqRespStgePt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=RequestResponseStage,name=CompletedTasks"> | |
- <attrib name="Value" alias="tpReqRespStgeCt" type="counter"/> | |
- </mbean> | |
- | |
- <!-- ThreadPools :: ReadRepairStage --> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadRepairStage,name=ActiveTasks"> | |
- <attrib name="Value" alias="tpReadRepairStgeAt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadRepairStage,name=CurrentlyBlockedTasks"> | |
- <attrib name="Count" alias="tpReadRepairStgeCbt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadRepairStage,name=PendingTasks"> | |
- <attrib name="Value" alias="tpReadRepairStgePt" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.ThreadPools" | |
- objectname="org.apache.cassandra.metrics:type=ThreadPools,path=request,scope=ReadRepairStage,name=CompletedTasks"> | |
- <attrib name="Value" alias="tpReadRepairStgeCt" type="counter"/> | |
+ <mbean name="org.apache.activemq" objectname="org.apache.activemq:type=Broker,brokerName=localhost"> | |
+ <attrib name="TotalConnectionsCount" alias="TtlConCnt" type="gauge"/> | |
+ <attrib name="TotalEnqueueCount" alias="TtlEnqCnt" type="gauge"/> | |
+ <attrib name="TotalDequeueCount" alias="TtlDeqCnt" type="gauge"/> | |
+ <attrib name="TotalConsumerCount" alias="TtlConsumerCnt" type="gauge"/> | |
+ <attrib name="TotalProducerCount" alias="TtlProdCnt" type="gauge"/> | |
+ <attrib name="TotalMessageCount" alias="TtlMsgCnt" type="gauge"/> | |
+ <attrib name="AverageMessageSize" alias="AvgMsgSize" type="gauge"/> | |
+ <attrib name="MaxMessageSize" alias="MaxMsgSize" type="gauge"/> | |
+ <attrib name="MinMessageSize" alias="MinMsgSize" type="gauge"/> | |
+ <attrib name="MemoryLimit" alias="MemLimit" type="gauge"/> | |
+ <attrib name="MemoryPercentUsage" alias="MemPctUsage" type="gauge"/> | |
+ <attrib name="StoreLimit" alias="StoreLimit" type="gauge"/> | |
+ <attrib name="StorePercentUsage" alias="StorePctUsage" type="gauge"/> | |
+ <attrib name="TempLimit" alias="TempLimit" type="gauge"/> | |
+ <attrib name="TempPercentUsage" alias="TempPctUsage" type="gauge"/> | |
</mbean> | |
</mbeans> | |
</jmx-collection> | |
- | |
- <jmx-collection name="cassandra21x-newts"> | |
- <rrd step="300"> | |
- <rra>RRA:AVERAGE:0.5:1:2016</rra> | |
- <rra>RRA:AVERAGE:0.5:12:1488</rra> | |
- <rra>RRA:AVERAGE:0.5:288:366</rra> | |
- <rra>RRA:MAX:0.5:288:366</rra> | |
- <rra>RRA:MIN:0.5:288:366</rra> | |
- </rrd> | |
- <mbeans> | |
- <!-- Newts :: AllMemmtables --> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=AllMemtablesLiveDataSize"> | |
- <attrib name="Value" alias="alMemTblLiDaSi" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=AllMemtablesOffHeapDataSize"> | |
- <attrib name="Value" alias="alMemTblOffHeapDaSi" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=AllMemtablesOnHeapDataSize"> | |
- <attrib name="Value" alias="alMemTblOnHeapDaSi" type="gauge"/> | |
- </mbean> | |
- | |
- <!-- Memtable :: Count --> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=MemtableSwitchCount"> | |
- <attrib name="Value" alias="memTblSwitchCount" type="gauge"/> | |
- </mbean> | |
- | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=MemtableColumnsCount"> | |
- <attrib name="Value" alias="memTblColumnsCnt" type="gauge"/> | |
- </mbean> | |
- | |
- <!-- Memtable :: Sizes --> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=MemtableLiveDataSize"> | |
- <attrib name="Value" alias="memTblLiveDaSi" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=MemtableOffHeapDataSize"> | |
- <attrib name="Value" alias="memTblOffHeapDaSi" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=MemtableOnHeapDataSize"> | |
- <attrib name="Value" alias="memTblOnHeapDaSi" type="gauge"/> | |
- </mbean> | |
- | |
- <!-- Latency --> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=ReadTotalLatency"> | |
- <attrib name="Count" alias="readTotLtncy" type="counter"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=RangeLatency"> | |
- <attrib name="99thPercentile" alias="rangeLtncy99" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=WriteTotalLatency"> | |
- <attrib name="Count" alias="writeTotLtncy" type="counter"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=CasCommitTotalLatency"> | |
- <attrib name="Count" alias="casCommitTotLtncy" type="counter"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=CasPrepareTotalLatency"> | |
- <attrib name="Count" alias="casPrepareTotLtncy" type="counter"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=CasProposeTotalLatency"> | |
- <attrib name="Count" alias="casProposeTotLtncy" type="counter"/> | |
- </mbean> | |
- | |
- <!-- Bloom Filter --> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=BloomFilterDiskSpaceUsed"> | |
- <attrib name="Value" alias="blmFltrDskSpcUsed" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=BloomFilterOffHeapMemoryUsed"> | |
- <attrib name="Value" alias="blmFltrOffHeapMemUs" type="gauge"/> | |
- </mbean> | |
- | |
- <!-- Memory Used --> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=CompressionMetadataOffHeapMemoryUsed"> | |
- <attrib name="Value" alias="cmpMetaOffHeapMemUs" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=IndexSummaryOffHeapMemoryUsed"> | |
- <attrib name="Value" alias="idxSumOffHeapMemUs" type="gauge"/> | |
- </mbean> | |
- | |
- <!-- Pending --> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=PendingCompactions"> | |
- <attrib name="Value" alias="pendingCompactions" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=PendingFlushes"> | |
- <attrib name="Value" alias="pendingFlushes" type="gauge"/> | |
- </mbean> | |
- | |
- <!-- Disk Space --> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=TotalDiskSpaceUsed"> | |
- <attrib name="Value" alias="totalDiskSpaceUsed" type="gauge"/> | |
- </mbean> | |
- <mbean name="org.apache.cassandra.metrics.Keyspace" | |
- objectname="org.apache.cassandra.metrics:type=Keyspace,keyspace=newts,name=LiveDiskSpaceUsed"> | |
- <attrib name="Value" alias="liveDiskSpaceUsed" type="gauge"/> | |
- </mbean> | |
- </mbeans> | |
- </jmx-collection> | |
- | |
</jmx-datacollection-config> | |
diff --git a/jmx.acl.cfg b/jmx.acl.cfg | |
index 531d8c4..ec42f3d 100644 | |
--- a/jmx.acl.cfg | |
+++ b/jmx.acl.cfg | |
@@ -70,4 +70,4 @@ list* = viewer | |
get* = viewer | |
is* = viewer | |
set* = admin | |
-* = admin | |
\ No newline at end of file | |
+* = admin | |
diff --git a/jmx.acl.java.lang.Memory.cfg b/jmx.acl.java.lang.Memory.cfg | |
index 60e6275..a58bcf5 100644 | |
--- a/jmx.acl.java.lang.Memory.cfg | |
+++ b/jmx.acl.java.lang.Memory.cfg | |
@@ -22,4 +22,4 @@ | |
# | |
# For a description of the format of this file, see jmx.acl.cfg | |
# | |
-gc = manager | |
\ No newline at end of file | |
+gc = manager | |
diff --git a/jmx.acl.org.apache.karaf.bundle.cfg b/jmx.acl.org.apache.karaf.bundle.cfg | |
index dd318d8..b8ba95a 100644 | |
--- a/jmx.acl.org.apache.karaf.bundle.cfg | |
+++ b/jmx.acl.org.apache.karaf.bundle.cfg | |
@@ -37,4 +37,4 @@ uninstall(java.lang.String)["0"] = #this is a comment, no roles can perform this | |
uninstall = admin | |
update(java.lang.String)[/([1-4])?[0-9]/] = admin | |
update(java.lang.String,java.lang.String)[/([1-4])?[0-9]/,/.*/] = admin | |
-update = manager | |
\ No newline at end of file | |
+update = manager | |
diff --git a/jmx.acl.org.apache.karaf.config.cfg b/jmx.acl.org.apache.karaf.config.cfg | |
index a597112..e6df22b 100644 | |
--- a/jmx.acl.org.apache.karaf.config.cfg | |
+++ b/jmx.acl.org.apache.karaf.config.cfg | |
@@ -49,4 +49,4 @@ setProperty(java.lang.String,java.lang.String,java.lang.String) = manager | |
update(java.lang.String,java.util.Map)[/jmx[.]acl.*/,/.*/] = admin | |
update(java.lang.String,java.util.Map)[/org[.]apache[.]karaf[.]command[.]acl.+/,/.*/] = admin | |
update(java.lang.String,java.util.Map)[/org[.]apache[.]karaf[.]service[.]acl.+/,/.*/] = admin | |
-update(java.lang.String,java.util.Map) = manager | |
\ No newline at end of file | |
+update(java.lang.String,java.util.Map) = manager | |
diff --git a/jmx.acl.org.apache.karaf.security.jmx.cfg b/jmx.acl.org.apache.karaf.security.jmx.cfg | |
index 0af2c96..22471e5 100644 | |
--- a/jmx.acl.org.apache.karaf.security.jmx.cfg | |
+++ b/jmx.acl.org.apache.karaf.security.jmx.cfg | |
@@ -24,4 +24,4 @@ | |
# | |
# For a description of the format of this file, see jmx.acl.cfg | |
# | |
-canInvoke = viewer | |
\ No newline at end of file | |
+canInvoke = viewer | |
diff --git a/jmx.acl.osgi.compendium.cm.cfg b/jmx.acl.osgi.compendium.cm.cfg | |
index 2d9045d..345d7b9 100644 | |
--- a/jmx.acl.osgi.compendium.cm.cfg | |
+++ b/jmx.acl.osgi.compendium.cm.cfg | |
@@ -50,4 +50,4 @@ update(java.lang.String,javax.management.openmbean.TabularData) = manager | |
updateForLocation(java.lang.String,java.lang.String,javax.management.openmbean.TabularData)[/jmx[.]acl.*/,/.*/,/.*/] = admin | |
updateForLocation(java.lang.String,java.lang.String,javax.management.openmbean.TabularData)[/org[.]apache[.]karaf[.]command[.]acl[.].+/,/.*/,/.*/] = admin | |
updateForLocation(java.lang.String,java.lang.String,javax.management.openmbean.TabularData)[/org[.]apache[.]karaf[.]service[.]acl[.].+/,/.*/,/.*/] = admin | |
-updateForLocation(java.lang.String,java.lang.String,javax.management.openmbean.TabularData) = manager | |
\ No newline at end of file | |
+updateForLocation(java.lang.String,java.lang.String,javax.management.openmbean.TabularData) = manager | |
diff --git a/jmxremote.access b/jmxremote.access | |
index 6254ae3..55aad04 100644 | |
--- a/jmxremote.access | |
+++ b/jmxremote.access | |
@@ -1 +1,2 @@ | |
admin readwrite | |
+jmx readonly | |
diff --git a/jre.properties b/jre.properties | |
index ea0d6d7..94da9db 100644 | |
--- a/jre.properties | |
+++ b/jre.properties | |
@@ -539,4 +539,4 @@ jre-1.8= \ | |
org.w3c.dom.xpath, \ | |
org.xml.sax, \ | |
org.xml.sax.ext, \ | |
- org.xml.sax.helpers | |
\ No newline at end of file | |
+ org.xml.sax.helpers | |
diff --git a/keys.properties b/keys.properties | |
index d51e633..a13d3e6 100644 | |
--- a/keys.properties | |
+++ b/keys.properties | |
@@ -32,4 +32,4 @@ | |
# The user guide describes how to generate/update the key. | |
# | |
#karaf=AAAAB3NzaC1kc3MAAACBAP1/U4EddRIpUt9KnC7s5Of2EbdSPO9EAMMeP4C2USZpRV1AIlH7WT2NWPq/xfW6MPbLm1Vs14E7gB00b/JmYLdrmVClpJ+f6AR7ECLCT7up1/63xhv4O1fnxqimFQ8E+4P208UewwI1VBNaFpEy9nXzrith1yrv8iIDGZ3RSAHHAAAAFQCXYFCPFSMLzLKSuYKi64QL8Fgc9QAAAIEA9+GghdabPd7LvKtcNrhXuXmUr7v6OuqC+VdMCz0HgmdRWVeOutRZT+ZxBxCBgLRJFnEj6EwoFhO3zwkyjMim4TwWeotUfI0o4KOuHiuzpnWRbqN/C/ohNWLx+2J6ASQ7zKTxvqhRkImog9/hWuWfBpKLZl6Ae1UlZAFMO/7PSSoAAACBAKKSU2PFl/qOLxIwmBZPPIcJshVe7bVUpFvyl3BbJDow8rXfskl8wO63OzP/qLmcJM0+JbcRU/53JjTuyk31drV2qxhIOsLDC9dGCWj47Y7TyhPdXh/0dthTRBy6bqGtRPxGa7gJov1xm/UuYYXPIUR/3x9MAZvZ5xvE0kYXO+rx,_g_:admingroup | |
-_g_\:admingroup = group,admin,manager,viewer | |
\ No newline at end of file | |
+_g_\:admingroup = group,admin,manager,viewer | |
diff --git a/libraries.properties b/libraries.properties | |
index cdc8199..99f1e82 100644 | |
--- a/libraries.properties | |
+++ b/libraries.properties | |
@@ -1,4 +1,4 @@ | |
-#Tue Feb 14 08:52:48 UTC 2017 | |
+#Tue Feb 14 08:33:34 UTC 2017 | |
opennms.library.jicmp=/usr/lib64/libjicmp.so | |
opennms.library.jrrd2=/usr/lib64/libjrrd2.so | |
opennms.library.jicmp6=/usr/lib64/libjicmp6.so | |
diff --git a/log4j2.xml b/log4j2.xml | |
index 78a3628..de3368b 100644 | |
--- a/log4j2.xml | |
+++ b/log4j2.xml | |
@@ -99,6 +99,9 @@ | |
<logger name="org.apache.bsf" additivity="false" level="INFO"> | |
<appender-ref ref="RoutingAppender"/> | |
</logger> | |
+ <logger name="org.apache.camel.component.jms" additivity="false" level="INFO"> | |
+ <appender-ref ref="RoutingAppender"/> | |
+ </logger> | |
<logger name="org.apache.commons" additivity="false" level="WARN"> | |
<appender-ref ref="RoutingAppender"/> | |
</logger> | |
@@ -166,44 +169,45 @@ | |
<DynamicThresholdFilter key="prefix" defaultThreshold="DEBUG"> | |
<!-- always leave instrumentation logging at INFO --> | |
<KeyValuePair key="instrumentation" value="INFO" /> | |
- <KeyValuePair key="access-point-monitor" value="WARN" /> | |
- <KeyValuePair key="ackd" value="WARN" /> | |
- <KeyValuePair key="actiond" value="WARN" /> | |
- <KeyValuePair key="alarmd" value="WARN" /> | |
- <KeyValuePair key="asterisk-gateway" value="WARN" /> | |
- <KeyValuePair key="archiver" value="WARN" /> | |
- <KeyValuePair key="bsmd" value="WARN" /> | |
- <KeyValuePair key="collectd" value="WARN" /> | |
- <KeyValuePair key="correlator" value="WARN" /> | |
- <KeyValuePair key="dhcpd" value="WARN" /> | |
- <KeyValuePair key="discovery" value="WARN" /> | |
- <KeyValuePair key="eventd" value="WARN" /> | |
- <KeyValuePair key="event-translator" value="WARN" /> | |
- <KeyValuePair key="icmp" value="WARN" /> | |
- <KeyValuePair key="jetty-server" value="WARN" /> | |
- <KeyValuePair key="enlinkd" value="WARN" /> | |
+ <KeyValuePair key="access-point-monitor" value="DEBUG" /> | |
+ <KeyValuePair key="ackd" value="DEBUG" /> | |
+ <KeyValuePair key="actiond" value="DEBUG" /> | |
+ <KeyValuePair key="alarmd" value="DEBUG" /> | |
+ <KeyValuePair key="asterisk-gateway" value="DEBUG" /> | |
+ <KeyValuePair key="archiver" value="DEBUG" /> | |
+ <KeyValuePair key="bsmd" value="DEBUG" /> | |
+ <KeyValuePair key="collectd" value="DEBUG" /> | |
+ <KeyValuePair key="correlator" value="DEBUG" /> | |
+ <KeyValuePair key="dhcpd" value="DEBUG" /> | |
+ <KeyValuePair key="discovery" value="DEBUG" /> | |
+ <KeyValuePair key="eventd" value="DEBUG" /> | |
+ <KeyValuePair key="event-translator" value="DEBUG" /> | |
+ <KeyValuePair key="icmp" value="DEBUG" /> | |
+ <KeyValuePair key="ipc" value="DEBUG" /> | |
+ <KeyValuePair key="jetty-server" value="DEBUG" /> | |
+ <KeyValuePair key="enlinkd" value="DEBUG" /> | |
<KeyValuePair key="manager" value="DEBUG" /> | |
- <KeyValuePair key="map" value="WARN" /> | |
- <KeyValuePair key="notifd" value="WARN" /> | |
- <KeyValuePair key="oss-qosd" value="WARN" /> | |
- <KeyValuePair key="oss-qosdrx" value="WARN" /> | |
- <KeyValuePair key="passive" value="WARN" /> | |
- <KeyValuePair key="poller" value="WARN" /> | |
- <KeyValuePair key="provisiond" value="WARN" /> | |
- <KeyValuePair key="queued" value="WARN" /> | |
- <KeyValuePair key="reportd" value="WARN" /> | |
- <KeyValuePair key="reports" value="WARN" /> | |
- <KeyValuePair key="rtc" value="WARN" /> | |
- <KeyValuePair key="statsd" value="WARN" /> | |
- <KeyValuePair key="scriptd" value="WARN" /> | |
- <KeyValuePair key="snmp-poller" value="WARN" /> | |
- <KeyValuePair key="syslogd" value="WARN" /> | |
- <KeyValuePair key="threshd" value="WARN" /> | |
- <KeyValuePair key="tl1d" value="WARN" /> | |
- <KeyValuePair key="trapd" value="WARN" /> | |
- <KeyValuePair key="trouble-ticketer" value="WARN" /> | |
- <KeyValuePair key="vacuumd" value="WARN" /> | |
- <KeyValuePair key="web" value="WARN" /> | |
+ <KeyValuePair key="map" value="DEBUG" /> | |
+ <KeyValuePair key="notifd" value="DEBUG" /> | |
+ <KeyValuePair key="oss-qosd" value="DEBUG" /> | |
+ <KeyValuePair key="oss-qosdrx" value="DEBUG" /> | |
+ <KeyValuePair key="passive" value="DEBUG" /> | |
+ <KeyValuePair key="poller" value="DEBUG" /> | |
+ <KeyValuePair key="provisiond" value="DEBUG" /> | |
+ <KeyValuePair key="queued" value="DEBUG" /> | |
+ <KeyValuePair key="reportd" value="DEBUG" /> | |
+ <KeyValuePair key="reports" value="DEBUG" /> | |
+ <KeyValuePair key="rtc" value="DEBUG" /> | |
+ <KeyValuePair key="statsd" value="DEBUG" /> | |
+ <KeyValuePair key="scriptd" value="DEBUG" /> | |
+ <KeyValuePair key="snmp-poller" value="DEBUG" /> | |
+ <KeyValuePair key="syslogd" value="DEBUG" /> | |
+ <KeyValuePair key="threshd" value="DEBUG" /> | |
+ <KeyValuePair key="tl1d" value="DEBUG" /> | |
+ <KeyValuePair key="trapd" value="DEBUG" /> | |
+ <KeyValuePair key="trouble-ticketer" value="DEBUG" /> | |
+ <KeyValuePair key="vacuumd" value="DEBUG" /> | |
+ <KeyValuePair key="web" value="DEBUG" /> | |
</DynamicThresholdFilter> | |
<appender-ref ref="RoutingAppender"/> | |
diff --git a/magic-users.properties b/magic-users.properties | |
deleted file mode 100644 | |
index 1461809..0000000 | |
--- a/magic-users.properties | |
+++ /dev/null | |
@@ -1,75 +0,0 @@ | |
-############################################################################ | |
-# magic-users.properties | |
-#-------------------------------------------------------------------------- | |
-# This file defines the secondary authentication and authorization info for | |
-# the Web UI. | |
-# | |
-# The "magic" users set up here will take priority over users created in the | |
-# users.xml. For example, if you create a user in users.xml called "rtc", | |
-# its password will be ignored in favor of the password here, et cetera. | |
-# | |
-# The authorization roles set up here are all the roles available to the | |
-# system except for the default role: OpenNMS User. All users belong to | |
-# the OpenNMS User role, even the "magic" users defined in this file, unless | |
-# they are a member of one or more roles in this file *and* all of the roles | |
-# they are a member of have "notInDefaultGroup" set to true. All | |
-# other roles specify their own user list within this file. | |
-# | |
-########################################################################### | |
- | |
- | |
-########################################################################### | |
-## U S E R S | |
-########################################################################### | |
- | |
-# A comma-separated list of user keys. A user.{KEY}.username and | |
-# user.{KEY}.password property must be set for each key in this property. | |
-users=rtc | |
- | |
-# The RTC View Control Manager daemon uses this user to authenticate itself | |
-# while sending RTC data posts. | |
-user.rtc.username=rtc | |
-user.rtc.password=rtc | |
- | |
-########################################################################### | |
-## R O L E S | |
-########################################################################### | |
- | |
-# A comma-separated list of role keys. A role.{KEY}.name and | |
-# role.{KEY}.users property must be set for each key in this property. | |
-roles=rtc, admin, rouser, dashboard, provision, remoting, rest, asset, mobile | |
- | |
-# This role allows a user to make RTC data posts. | |
-role.rtc.name=OpenNMS RTC Daemon | |
-role.rtc.users=rtc | |
-role.rtc.notInDefaultGroup=true | |
- | |
-# This role allows users access to configuration and | |
-# administrative web pages. | |
-role.admin.name=OpenNMS Administrator | |
-role.admin.users=admin | |
- | |
-# This role disallows user write access | |
-role.rouser.name=OpenNMS Read-Only User | |
-role.rouser.users= | |
- | |
-# This role allows access to the dashboard only | |
-role.dashboard.name=OpenNMS Dashboard User | |
-role.dashboard.users= | |
-role.dashboard.notInDefaultGroup=true | |
- | |
-# This role allows edit access to the Asset Editor GUI | |
-role.asset.name=OpenNMS Asset Editor | |
-role.asset.users= | |
- | |
-role.provision.name=OpenNMS Provision User | |
-role.provision.users= | |
- | |
-role.remoting.name=OpenNMS Remote Poller User | |
-role.remoting.users= | |
- | |
-role.rest.name=OpenNMS REST User | |
-role.rest.users=iphone | |
- | |
-role.mobile.name=OpenNMS Mobile User | |
-role.mobile.users= | |
diff --git a/opennms-activemq.xml b/opennms-activemq.xml | |
index 3cf0384..72f99dc 100644 | |
--- a/opennms-activemq.xml | |
+++ b/opennms-activemq.xml | |
@@ -36,6 +36,28 @@ | |
The <broker> element is used to configure the ActiveMQ broker. | |
--> | |
<broker xmlns="http://activemq.apache.org/schema/core" brokerName="localhost" dataDirectory="${activemq.data}"> | |
+ <plugins> | |
+ <bean id="openNMSJaasBrokerPlugin" class="org.opennms.netmgt.activemq.auth.OpenNMSJaasBrokerPlugin" xmlns="http://www.springframework.org/schema/beans" /> | |
+ | |
+ <authorizationPlugin> | |
+ <map> | |
+ <authorizationMap> | |
+ <authorizationEntries> | |
+ <!-- Users in the admin role can read/write/create any queue/topic --> | |
+ <authorizationEntry queue=">" read="admin" write="admin" admin="admin" /> | |
+ <authorizationEntry topic=">" read="admin" write="admin" admin="admin"/> | |
+ <!-- Users in the minion role can write/create queues that are not keyed by location --> | |
+ <authorizationEntry queue="OpenNMS.*.*" write="minion" admin="minion" /> | |
+ <!-- Users in the minion role can read/create from queues that are keyed by location --> | |
+ <authorizationEntry queue="OpenNMS.*.*.*" read="minion" admin="minion" /> | |
+ <!-- Users in the minion role can read/write/create advisory topics --> | |
+ <authorizationEntry topic="ActiveMQ.Advisory.>" read="minion" write="minion" admin="minion"/> | |
+ </authorizationEntries> | |
+ <!-- Allow all users to read/write/create temporary destinations (by omitting a <tempDestinationAuthorizationEntry>) --> | |
+ </authorizationMap> | |
+ </map> | |
+ </authorizationPlugin> | |
+ </plugins> | |
<!-- | |
For better performances use VM cursor and small memory limit. | |
@@ -135,7 +157,7 @@ | |
<!-- Uncomment this line to allow external TCP connections --> | |
<!-- | |
WARNING: Access to port 61616 should be firewalled to prevent unauthorized injection | |
- tof data into OpenNMS when this port is open. | |
+ of data into OpenNMS when this port is open. | |
--> | |
<!-- <transportConnector name="openwire" uri="tcp://0.0.0.0:61616?useJmx=false&maximumConnections=1000&wireformat.maxFrameSize=104857600"/> --> | |
diff --git a/opennms-upgrade-status.properties b/opennms-upgrade-status.properties | |
index a6ee1c9..1b6ab34 100644 | |
--- a/opennms-upgrade-status.properties | |
+++ b/opennms-upgrade-status.properties | |
@@ -1,12 +1,14 @@ | |
-#Tue Feb 14 08:52:59 UTC 2017 | |
-JettyConfigMigratorOffline=Tue Feb 14 08\:52\:57 UTC 2017 | |
-DataCollectionConfigMigrator17Offline=Tue Feb 14 08\:52\:58 UTC 2017 | |
-EOLServiceConfigMigratorOffline=Tue Feb 14 08\:52\:59 UTC 2017 | |
-RequisitionsMigratorOffline=Tue Feb 14 08\:52\:56 UTC 2017 | |
-KscReportsMigrator=Tue Feb 14 08\:52\:57 UTC 2017 | |
-DiscoveryConfigurationMigratorOffline=Tue Feb 14 08\:52\:59 UTC 2017 | |
-ServiceConfig1701MigratorOffline=Tue Feb 14 08\:52\:57 UTC 2017 | |
-DataCollectionConfigMigratorOffline=Tue Feb 14 08\:52\:57 UTC 2017 | |
-ServiceConfigMigratorOffline=Tue Feb 14 08\:52\:57 UTC 2017 | |
-JmxRrdMigratorOffline=Tue Feb 14 08\:52\:57 UTC 2017 | |
-MonitoringLocationsMigratorOffline=Tue Feb 14 08\:52\:57 UTC 2017 | |
+#Tue Feb 14 08:33:44 UTC 2017 | |
+JettyConfigMigratorOffline=Tue Feb 14 08\:33\:41 UTC 2017 | |
+DataCollectionConfigMigrator17Offline=Tue Feb 14 08\:33\:43 UTC 2017 | |
+EOLServiceConfigMigratorOffline=Tue Feb 14 08\:33\:43 UTC 2017 | |
+RequisitionsMigratorOffline=Tue Feb 14 08\:33\:41 UTC 2017 | |
+DiscoveryConfigurationLocationMigratorOffline=Tue Feb 14 08\:33\:44 UTC 2017 | |
+KscReportsMigrator=Tue Feb 14 08\:33\:42 UTC 2017 | |
+MagicUsersMigratorOffline=Tue Feb 14 08\:33\:43 UTC 2017 | |
+DiscoveryConfigurationMigratorOffline=Tue Feb 14 08\:33\:43 UTC 2017 | |
+ServiceConfig1701MigratorOffline=Tue Feb 14 08\:33\:42 UTC 2017 | |
+DataCollectionConfigMigratorOffline=Tue Feb 14 08\:33\:42 UTC 2017 | |
+ServiceConfigMigratorOffline=Tue Feb 14 08\:33\:42 UTC 2017 | |
+JmxRrdMigratorOffline=Tue Feb 14 08\:33\:41 UTC 2017 | |
+MonitoringLocationsMigratorOffline=Tue Feb 14 08\:33\:42 UTC 2017 | |
diff --git a/opennms.properties b/opennms.properties | |
index a0fe8e7..9813cff 100644 | |
--- a/opennms.properties | |
+++ b/opennms.properties | |
@@ -28,15 +28,18 @@ | |
# the default ICMP implementation used in the remote poller, since it does | |
# not rely on any external native code to be installed outside of the JVM. | |
# | |
-# To use the JNI ICMPv4 interface only, use the following property setting: | |
+# To use the JNI ICMPv4/ICMPv6 implementation, use the following property: | |
+#org.opennms.netmgt.icmp.pingerClass=org.opennms.netmgt.icmp.jni6.Jni6Pinger | |
+# | |
+# To use the JNI ICMPv4 interface only, use the following property: | |
#org.opennms.netmgt.icmp.pingerClass=org.opennms.netmgt.icmp.jni.JniPinger | |
# | |
# To use the JNA ICMPv4/ICMPv6 implementation, use the following property: | |
#org.opennms.netmgt.icmp.pingerClass=org.opennms.netmgt.icmp.jna.JnaPinger | |
# | |
-# The default is set to use the JNI ICMPv4/ICMPv6 interface like so: | |
-#org.opennms.netmgt.icmp.pingerClass=org.opennms.netmgt.icmp.jni6.Jni6Pinger | |
- | |
+# If no pingerClass is set, OpenNMS will attempt to choose the best | |
+# available pinger automatically. | |
+# | |
# By default, OpenNMS will start up if either ICMPv4 *or* ICMPv6 are | |
# available and initialize properly. If you wish to force IPv4 or IPv6 | |
# explicitly, set one or both of these properties to true. | |
@@ -219,10 +222,10 @@ opennms.ticketer.plugin=org.opennms.netmgt.ticketd.NullTicketerPlugin | |
#opennms.alarmTroubleTicketLinkTemplate = <a href="http://172.20.0.76:8180/arsys/servlet/ViewFormServlet?form=HPD:Help%20Desk&server=itts3h&qual='Incident ID*%2B'=%22${id}%22">${id}</a> | |
# Enable this flag to ignore 'uei.opennms.org/troubleTicket/create' events against alarms with a severity of 'Cleared' | |
-#opennms.ticketer.skipCreateWhenCleared = false | |
+#opennms.ticketer.skipCreateWhenCleared = true | |
# Enable this flag to ignore 'uei.opennms.org/troubleTicket/close' events against alarms with a severity other than 'Cleared' | |
-#opennms.ticketer.skipCloseWhenNotCleared = false | |
+#opennms.ticketer.skipCloseWhenNotCleared = true | |
###### MISCELLANEOUS ###### | |
@@ -342,17 +345,6 @@ org.eclipse.jetty.server.Request.maxFormKeys=2000 | |
## in the keystore specified by the https-keystore property), uncomment and | |
## change this property. | |
#org.opennms.netmgt.jetty.https-keypassword = changeit | |
-## To specify a particular SSL certificate alias in the keystore, set this | |
-## property. Otherwise, the first certificate that is found will be used. | |
-#org.opennms.netmgt.jetty.https-cert-alias = opennms-jetty-certificate | |
-## To exclude specific SSL/TLS cipher suites from use, set this property to a | |
-## colon-separated list of suite names. Whitespace surrounding colons is OK. | |
-#org.opennms.netmgt.jetty.https-exclude-cipher-suites=SSL_DHE_DSS_WITH_DES_CBC_SHA: \ | |
-# SSL_DHE_RSA_EXPORT_WITH_DES40_CBC_SHA:SSL_DHE_DSS_EXPORT_WITH_DES40_CBC_SHA: \ | |
-# SSL_DHE_RSA_WITH_DES_CBC_SHA:SSL_RSA_EXPORT_WITH_DES40_CBC_SHA: \ | |
-# SSL_RSA_EXPORT_WITH_RC4_40_MD5:SSL_RSA_WITH_3DES_EDE_CBC_SHA: \ | |
-# SSL_RSA_WITH_DES_CBC_SHA:TLS_DHE_RSA_EXPORT_WITH_DES40_CBC_SHA: \ | |
-# TLS_RSA_EXPORT_WITH_DES40_CBC_SHA:TLS_RSA_WITH_DES_CBC_SHA | |
# If OpenNMS is setting an incorrect HTML <base> tag, you can override | |
# its idea of the local URL here. The URL must end with a '/'. The following | |
@@ -441,6 +433,10 @@ gnu.io.SerialPorts=/dev/ttyACM0:/dev/ttyACM1:/dev/ttyACM2:/dev/ttyACM3:/dev/ttyA | |
# you will get a phone timed out exception | |
smslib.serial.polling=true | |
+###### EVENTD OPTIONS ###### | |
+# This property is used to define the size of the event parsing cache. The size must be >= 0, where 0 disables caching. | |
+#org.opennms.eventd.eventTemplateCacheSize = 1000 | |
+ | |
###### PROVISIOND OPTIONS ###### | |
# | |
# This property is used to enable/disable the handling of new suspect events | |
@@ -604,7 +600,7 @@ opennms.eventlist.showCount=false | |
###### JasperReports Properties ###### | |
# Defines the Version of the Jasperreports library | |
-org.opennms.jasperReportsVersion=6.1.1 | |
+org.opennms.jasperReportsVersion=6.3.0 | |
# Define if duplicates are ignored when using pie charts within a JasperReport template (*.jrxml) file. | |
# See http://jasperreports.sourceforge.net/config.reference.html for more details. | |
@@ -612,7 +608,7 @@ net.sf.jasperreports.chart.pie.ignore.duplicated.key=true | |
###### Web Console - Front Page ###### | |
# This value controls the content that will be displayed in the middle box of the front page. | |
-# The default is the view of SLM/RTC categories: /includes/categories-box.jsp. | |
+# The default is the view of SLM/RTC categories and the geographical map: /includes/categories-box.jsp,/geomap/map-box.jsp. | |
# You can also use a comma-seperated list of files to include more than one content file. | |
# Uncomment the following line to display the widget for the surveillance view from the dashboard. | |
# (It uses the same rules for the dashboard) | |
@@ -628,14 +624,6 @@ net.sf.jasperreports.chart.pie.ignore.duplicated.key=true | |
# services, you may need to add them to this list. | |
excludeServiceMonitorsFromRemotePoller=DHCP,NSClient,RadiusAuth,XMP | |
-###### DASHBOARD/SURVEILLANCE VIEW IMPLEMENTATION ###### | |
-# OpenNMS provides two different dashboard/surveillance view implementations. The GWT | |
-# variant is the original. Later, the UI was rewitten using the VAADIN framework. So, the | |
-# two valid options fpr this option are 'vaadin' or 'gwt'. The VAADIN implementation is | |
-# the default one. Please note that the GWT version is deprecated and will be removed in | |
-# future versions. | |
-#org.opennms.dashboard.implementation=gwt | |
- | |
###### DASHBOARD LANDING PAGE ###### | |
# This setting controls whether users will be redirected to the dashboard page after | |
# a successful login. The two valid options for this are 'true' or 'false' which is | |
@@ -731,6 +719,15 @@ excludeServiceMonitorsFromRemotePoller=DHCP,NSClient,RadiusAuth,XMP | |
# account when generating the alarm-based heatmap. | |
#org.opennms.heatmap.onlyUnacknowledged=false | |
+# ###### GEOMAP BOX ###### | |
+# This setting controls how each node's status is calculated. | |
+# Valid options are Alarms or Outages | |
+#org.opennms.geomap.defaultStrategy=Alarms | |
+ | |
+# This setting controls the node's severity to show up. | |
+# By default all nodes with a severity >= Normal are shown. | |
+#org.opennms.geomap.defaultSeverity=Normal | |
+ | |
# ###### GRAFANA BOX ###### | |
# This setting controls whether a grafana box showing the available dashboards is | |
# placed on the landing page. The two valid options for this are 'true' or 'false' | |
@@ -757,3 +754,31 @@ excludeServiceMonitorsFromRemotePoller=DHCP,NSClient,RadiusAuth,XMP | |
# Timeouts for contacting the grafana server | |
#org.opennms.grafanaBox.connectionTimeout=500 | |
#org.opennms.grafanaBox.soTimeout=500 | |
+ | |
+# ###### ActiveMQ Settings ###### | |
+# These settings are used to control which ActiveMQ broker will be used. | |
+# By default, we use an embedded broker. | |
+# | |
+#org.opennms.activemq.broker.disable=false | |
+#org.opennms.activemq.broker.url=vm://localhost?create=false | |
+#org.opennms.activemq.broker.username= | |
+#org.opennms.activemq.broker.password= | |
+#org.opennms.activemq.client.max-connections=8 | |
+#org.opennms.activemq.client.concurrent-consumers=10 | |
+ | |
+# ###### Minion provisioning ###### | |
+# These settings control the automatic provisioning of minions. | |
+# | |
+# Enables the provisioning mechanism | |
+#opennms.minion.provisioning=true | |
+# | |
+# The pattern is used to name the foreign source used to provision the minions. | |
+# The pattern can contain a single '%s' placeholder which is replaced with the | |
+# minions locations. | |
+#opennms.minion.provisioning.foreignSourcePattern=Minions | |
+ | |
+# ###### JMS Timeout ###### | |
+# Various OpenNMS components communicate via a message queue. These messages require a request timeout value to | |
+# be set. In many cases OpenNMS computes a proper timeout value for its operations. However, if a value cannot be | |
+# determined this default value (ms) will be used. | |
+#org.opennms.jms.timeout = 20000 | |
diff --git a/org.apache.felix.fileinstall-deploy.cfg b/org.apache.felix.fileinstall-deploy.cfg | |
index b2aedac..7f9f0ca 100644 | |
--- a/org.apache.felix.fileinstall-deploy.cfg | |
+++ b/org.apache.felix.fileinstall-deploy.cfg | |
@@ -22,4 +22,4 @@ felix.fileinstall.tmpdir = ${karaf.data}/generated-bundles | |
felix.fileinstall.poll = 1000 | |
felix.fileinstall.start.level = 80 | |
felix.fileinstall.active.level = 80 | |
-felix.fileinstall.log.level = 3 | |
\ No newline at end of file | |
+felix.fileinstall.log.level = 3 | |
diff --git a/org.apache.karaf.command.acl.config.cfg b/org.apache.karaf.command.acl.config.cfg | |
index 542b15f..69a4705 100644 | |
--- a/org.apache.karaf.command.acl.config.cfg | |
+++ b/org.apache.karaf.command.acl.config.cfg | |
@@ -42,4 +42,4 @@ propset = manager | |
propset[/.*jmx[.]acl.*/] = admin | |
propset[/.*org[.]apache[.]karaf[.]command[.]acl[.].+/] = admin | |
propset[/.*org[.]apache[.]karaf[.]service[.]acl[.].+/] = admin | |
-update = manager | |
\ No newline at end of file | |
+update = manager | |
diff --git a/org.apache.karaf.command.acl.features.cfg b/org.apache.karaf.command.acl.features.cfg | |
index dfe4483..175fdf9 100644 | |
--- a/org.apache.karaf.command.acl.features.cfg | |
+++ b/org.apache.karaf.command.acl.features.cfg | |
@@ -24,4 +24,4 @@ | |
# org.apache.karaf.command.acl.osgi.cfg | |
# | |
install = admin | |
-uninstall = admin | |
\ No newline at end of file | |
+uninstall = admin | |
diff --git a/org.apache.karaf.command.acl.jaas.cfg b/org.apache.karaf.command.acl.jaas.cfg | |
index 5ab39d7..5713dea 100644 | |
--- a/org.apache.karaf.command.acl.jaas.cfg | |
+++ b/org.apache.karaf.command.acl.jaas.cfg | |
@@ -24,4 +24,4 @@ | |
# org.apache.karaf.command.acl.osgi.cfg | |
# | |
# Jaas commands commands have no effect until update is called. | |
-update = admin | |
\ No newline at end of file | |
+update = admin | |
diff --git a/org.apache.karaf.command.acl.osgi.cfg b/org.apache.karaf.command.acl.osgi.cfg | |
index bbb2cdd..7852cf8 100644 | |
--- a/org.apache.karaf.command.acl.osgi.cfg | |
+++ b/org.apache.karaf.command.acl.osgi.cfg | |
@@ -67,4 +67,4 @@ watch = admin | |
shutdown = admin | |
start-level[/.*[0-9][0-9][0-9]+.*/] = manager # manager can set startlevels above 100 | |
start-level[/[^0-9]*/] = viewer # viewer can obtain the current start level | |
-start-level = admin # admin can set any start level, including < 100 | |
\ No newline at end of file | |
+start-level = admin # admin can set any start level, including < 100 | |
diff --git a/org.apache.karaf.command.acl.scope_bundle.cfg b/org.apache.karaf.command.acl.scope_bundle.cfg | |
index 588e3f9..5e2621f 100644 | |
--- a/org.apache.karaf.command.acl.scope_bundle.cfg | |
+++ b/org.apache.karaf.command.acl.scope_bundle.cfg | |
@@ -31,4 +31,4 @@ log=org.apache.karaf.shell.log | |
packages=org.apache.karaf.shell.packages | |
config=org.apache.karaf.shell.config | |
ssh=org.apache.karaf.shell.ssh | |
-shell=org.apache.karaf.shell.commands | |
\ No newline at end of file | |
+shell=org.apache.karaf.shell.commands | |
diff --git a/org.apache.karaf.command.acl.shell.cfg b/org.apache.karaf.command.acl.shell.cfg | |
index 1f94e28..c47f6f5 100644 | |
--- a/org.apache.karaf.command.acl.shell.cfg | |
+++ b/org.apache.karaf.command.acl.shell.cfg | |
@@ -26,4 +26,4 @@ | |
edit = admin | |
exec = admin | |
new = admin | |
-java = admin | |
\ No newline at end of file | |
+java = admin | |
diff --git a/org.apache.karaf.features.cfg b/org.apache.karaf.features.cfg | |
index 733563b..cd42b32 100644 | |
--- a/org.apache.karaf.features.cfg | |
+++ b/org.apache.karaf.features.cfg | |
@@ -1,7 +1,7 @@ | |
# | |
# Comma separated list of features repositories to register by default | |
# | |
-featuresRepositories=mvn:org.opennms.container/karaf/18.0.4/xml/features,mvn:org.opennms.karaf/opennms/18.0.4/xml/features | |
+featuresRepositories=mvn:org.opennms.container/org.opennms.container.karaf/19.0.0/xml/features,mvn:org.opennms.karaf/opennms/19.0.0/xml/features | |
# | |
# Comma separated list of features to install at startup | |
@@ -15,12 +15,17 @@ featuresBoot=karaf-framework,ssh,config,features,management,\ | |
opennms-jaas-login-module,\ | |
datachoices, \ | |
opennms-collection-commands, \ | |
+ opennms-events-commands, \ | |
+ opennms-icmp-commands, \ | |
+ opennms-snmp-commands, \ | |
opennms-topology-runtime-browsers,\ | |
opennms-topology-runtime-linkd,\ | |
- opennms-topology-runtime-simple,\ | |
opennms-topology-runtime-vmware,\ | |
opennms-topology-runtime-application,\ | |
opennms-topology-runtime-bsm,\ | |
+ opennms-provisioning-shell,\ | |
+ opennms-poller-shell,\ | |
+ opennms-topology-runtime-graphml,\ | |
osgi-nrtg-local,\ | |
vaadin-node-maps,\ | |
vaadin-snmp-events-and-metrics, \ | |
@@ -42,4 +47,5 @@ featuresBoot=karaf-framework,ssh,config,features,management,\ | |
vaadin-opennms-pluginmanager, \ | |
vaadin-adminpage, \ | |
org.opennms.features.bsm.shell-commands, \ | |
- internal-plugins-descriptor | |
+ internal-plugins-descriptor, \ | |
+ geolocation | |
diff --git a/org.apache.karaf.features.obr.cfg b/org.apache.karaf.features.obr.cfg | |
index 12ba4cd..7d47c7d 100644 | |
--- a/org.apache.karaf.features.obr.cfg | |
+++ b/org.apache.karaf.features.obr.cfg | |
@@ -35,4 +35,4 @@ startByDefault = true | |
# | |
# Defines the start level for resolved bundles. The default is 80. | |
# | |
-startLevel = 80 | |
\ No newline at end of file | |
+startLevel = 80 | |
diff --git a/org.apache.karaf.features.repos.cfg b/org.apache.karaf.features.repos.cfg | |
index a8cc7d4..46fa539 100644 | |
--- a/org.apache.karaf.features.repos.cfg | |
+++ b/org.apache.karaf.features.repos.cfg | |
@@ -20,8 +20,10 @@ | |
# | |
# This file describes the features repository URL for some projects | |
# | |
-cellar=mvn:org.apache.karaf.cellar/apache-karaf-cellar/[2,3)/xml/features | |
-cave=mvn:org.apache.karaf.cave/apache-karaf-cave/[2,3)/xml/features | |
+enterprise=mvn:org.apache.karaf.assemblies.features/enterprise/LATEST/xml/features | |
+spring=mvn:org.apache.karaf.assemblies.features/spring/LATEST/xml/features | |
+cellar=mvn:org.apache.karaf.cellar/apache-karaf-cellar/LATEST/xml/features | |
+cave=mvn:org.apache.karaf.cave/apache-karaf-cave/LATEST/xml/features | |
camel=mvn:org.apache.camel.karaf/apache-camel/LATEST/xml/features | |
camel-extras=mvn:org.apache-extras.camel-extra.karaf/camel-extra/LATEST/xml/features | |
cxf=mvn:org.apache.cxf.karaf/apache-cxf/LATEST/xml/features | |
@@ -30,4 +32,11 @@ activemq=mvn:org.apache.activemq/activemq-karaf/LATEST/xml/features | |
jclouds=mvn:org.apache.jclouds.karaf/jclouds-karaf/LATEST/xml/features | |
openejb=mvn:org.apache.openejb/openejb-feature/LATEST/xml/features | |
wicket=mvn:org.ops4j.pax.wicket/features/LATEST/xml/features | |
-hawtio=mvn:io.hawt/hawtio-karaf/LATEST/xml/features | |
\ No newline at end of file | |
+hawtio=mvn:io.hawt/hawtio-karaf/LATEST/xml/features | |
+pax-cdi=mvn:org.ops4j.pax.cdi/pax-cdi-features/LATEST/xml/features | |
+pax-jdbc=mvn:org.ops4j.pax.jdbc/pax-jdbc-features/LATEST/xml/features | |
+pax-jpa=mvn:org.ops4j.pax.jpa/pax-jpa-features/LATEST/xml/features | |
+pax-web=mvn:org.ops4j.pax.web/pax-web-features/LATEST/xml/features | |
+pax-wicket=mvn:org.ops4j.pax.wicket/pax-wicket-features/LATEST/xml/features | |
+ecf=http://download.eclipse.org/rt/ecf/latest/site.p2/karaf-features.xml | |
+decanter=mvn:org.apache.karaf.decanter/apache-karaf-decanter/LATEST/xml/features | |
diff --git a/org.apache.karaf.jaas.cfg b/org.apache.karaf.jaas.cfg | |
index 8ac5af6..19f3fa0 100644 | |
--- a/org.apache.karaf.jaas.cfg | |
+++ b/org.apache.karaf.jaas.cfg | |
@@ -58,4 +58,4 @@ encryption.algorithm = MD5 | |
# hexadecimal | |
# base64 | |
# | |
-encryption.encoding = hexadecimal | |
\ No newline at end of file | |
+encryption.encoding = hexadecimal | |
diff --git a/org.apache.karaf.kar.cfg b/org.apache.karaf.kar.cfg | |
index 6c7cf16..0b71af2 100644 | |
--- a/org.apache.karaf.kar.cfg | |
+++ b/org.apache.karaf.kar.cfg | |
@@ -21,4 +21,4 @@ | |
# Enable or disable the refresh of the bundles when installing | |
# the features contained in a KAR file | |
# | |
-noAutoRefreshBundles=false | |
\ No newline at end of file | |
+noAutoRefreshBundles=false | |
diff --git a/org.apache.karaf.log.cfg b/org.apache.karaf.log.cfg | |
index 20ceafc..340f572 100644 | |
--- a/org.apache.karaf.log.cfg | |
+++ b/org.apache.karaf.log.cfg | |
@@ -33,4 +33,4 @@ size = 500 | |
# The pattern used to format the log statement when using log:display. This pattern is according | |
# to the log4j layout. You can override this parameter at runtime using log:display with -p. | |
# | |
-pattern = %d{ISO8601} | %-5.5p | %-16.16t | %-32.32c{1} | %-32.32C %4L | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n | |
\ No newline at end of file | |
+pattern = %d{ISO8601} | %-5.5p | %-16.16t | %-32.32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n | |
diff --git a/org.apache.karaf.shell.cfg b/org.apache.karaf.shell.cfg | |
index 5798ad9..03d2e07 100644 | |
--- a/org.apache.karaf.shell.cfg | |
+++ b/org.apache.karaf.shell.cfg | |
@@ -24,39 +24,38 @@ | |
# | |
# Via sshPort and sshHost you define the address you can login into Karaf. | |
# | |
-sshPort=8101 | |
-sshHost=0.0.0.0 | |
+sshPort = 8101 | |
+sshHost = 127.0.0.1 | |
# | |
# The sshIdleTimeout defines the inactivity timeout to logout the SSH session. | |
# The sshIdleTimeout is in milliseconds, and the default is set to 30 minutes. | |
# | |
-sshIdleTimeout=1800000 | |
+sshIdleTimeout = 1800000 | |
# | |
# sshRealm defines which JAAS domain to use for password authentication. | |
# | |
-sshRealm=karaf | |
+sshRealm = karaf | |
# | |
# The location of the hostKey file defines where the private/public key of the server | |
# is located. If no file is at the defined location it will be ignored. | |
# | |
-hostKey=${karaf.base}/etc/host.key | |
- | |
-# | |
-# Role name used for SSH access authorization | |
-# If not set, this defaults to the ${karaf.admin.role} configured in etc/system.properties | |
-# | |
-# sshRole=admin | |
+hostKey = ${karaf.etc}/host.key | |
# | |
# Self defined key size in 1024, 2048, 3072, or 4096 | |
# If not set, this defaults to 1024. | |
# | |
-# keySize=1024 | |
+# keySize = 1024 | |
# | |
# Specify host key algorithm, defaults to DSA | |
# | |
-# algorithm=DSA | |
+# algorithm = DSA | |
+ | |
+# Specify an additional welcome banner to be displayed when a user logs into the server. | |
+# | |
+# welcomeBanner = | |
+ | |
diff --git a/org.opennms.features.geocoder.google.cfg b/org.opennms.features.geocoder.google.cfg | |
index 84b564b..6f59a70 100644 | |
--- a/org.opennms.features.geocoder.google.cfg | |
+++ b/org.opennms.features.geocoder.google.cfg | |
@@ -1,2 +1,3 @@ | |
clientId= | |
clientKey= | |
+timeout=500 | |
diff --git a/org.opennms.features.topology.app.cfg b/org.opennms.features.topology.app.cfg | |
index 3467e11..9f020a1 100644 | |
--- a/org.opennms.features.topology.app.cfg | |
+++ b/org.opennms.features.topology.app.cfg | |
@@ -1,7 +1,5 @@ | |
-servletAlias = /topology | |
org.apache.karaf.features.configKey = org.opennms.features.topology.app | |
-theme = topo_default | |
-widgetset = org.opennms.features.topology.widgetset.gwt.TopologyWidgetset | |
showHeader = true | |
autoRefresh.enabled = false | |
autoRefresh.interval = 60 | |
+resolveCoordinatesFromAddressString = true | |
diff --git a/org.opennms.features.topology.app.icons.application.cfg b/org.opennms.features.topology.app.icons.application.cfg | |
index 1dec86b..34a4bf1 100644 | |
--- a/org.opennms.features.topology.app.icons.application.cfg | |
+++ b/org.opennms.features.topology.app.icons.application.cfg | |
@@ -1,3 +1,3 @@ | |
# Application Topology | |
application.application = business_service | |
-application.monitored-service = IP_service | |
\ No newline at end of file | |
+application.monitored-service = IP_service | |
diff --git a/org.opennms.features.topology.app.icons.bsm.cfg b/org.opennms.features.topology.app.icons.bsm.cfg | |
index 8c948d6..abcd2af 100644 | |
--- a/org.opennms.features.topology.app.icons.bsm.cfg | |
+++ b/org.opennms.features.topology.app.icons.bsm.cfg | |
@@ -1,4 +1,4 @@ | |
# Business Service Topology | |
bsm.business-service = business_service | |
bsm.ip-service = IP_service | |
-bsm.reduction-key = reduction_key | |
\ No newline at end of file | |
+bsm.reduction-key = reduction_key | |
diff --git a/org.opennms.features.topology.app.icons.list b/org.opennms.features.topology.app.icons.list | |
index 4fb412c..4344248 100644 | |
--- a/org.opennms.features.topology.app.icons.list | |
+++ b/org.opennms.features.topology.app.icons.list | |
@@ -13,24 +13,36 @@ linux_file_server | |
opennms_server | |
cloud | |
+# atlas icons (since 19.0.0) | |
+microwave_backhaul_1 | |
+microwave_backhaul_2 | |
+region_1 | |
+region_2 | |
+market_1 | |
+market_2 | |
+site_1 | |
+site_2 | |
+site_3 | |
+ | |
# BSM Icons | |
IP_service | |
business_service | |
reduction_key | |
# vmware_icons | |
+vmware-datacenter | |
vmware-cluster | |
-vmware-datastore | |
+vmware-hostsystem-on | |
+vmware-hostsystem-off | |
+vmware-hostsystem-standby | |
vmware-hostsystem-unknown | |
-vmware-virtualmachine-off | |
+vmware-network | |
+vmware-datastore | |
vmware-virtualmachine-on | |
-vmware-hostsystem-off | |
+vmware-virtualmachine-off | |
vmware-virtualmachine-suspended | |
-vmware-hostsystem-standby | |
-vmware-hostsystem-on | |
vmware-virtualmachine-unknown | |
-vmware-datacenter | |
-vmware-network | |
+ | |
# default_icons (legacy, before 18.0.0) | |
legacy_linux_file_server | |
diff --git a/org.opennms.features.topology.app.icons.sfree.cfg b/org.opennms.features.topology.app.icons.sfree.cfg | |
index fa8f401..d3718d9 100644 | |
--- a/org.opennms.features.topology.app.icons.sfree.cfg | |
+++ b/org.opennms.features.topology.app.icons.sfree.cfg | |
@@ -1,3 +1,3 @@ | |
# Scale Free Topology | |
sfree.group = cloud | |
-sfree.system = vmware-network | |
\ No newline at end of file | |
+sfree.system = vmware-network | |
diff --git a/org.opennms.features.topology.app.icons.vmware.cfg b/org.opennms.features.topology.app.icons.vmware.cfg | |
index 7a7a977..90e5d9e 100644 | |
--- a/org.opennms.features.topology.app.icons.vmware.cfg | |
+++ b/org.opennms.features.topology.app.icons.vmware.cfg | |
@@ -9,4 +9,4 @@ vmware.VIRTUALMACHINE_ICON_OFF = vmware-virtualmachine-off | |
vmware.VIRTUALMACHINE_ICON_SUSPENDED = vmware-virtualmachine-suspended | |
vmware.VIRTUALMACHINE_ICON_UNKNOWN = vmware-virtualmachine-unknown | |
vmware.DATASTORE_ICON = vmware-datastore | |
-vmware.DATACENTER_ICON = vmware-datacenter | |
\ No newline at end of file | |
+vmware.DATACENTER_ICON = vmware-datacenter | |
diff --git a/org.ops4j.pax.logging.cfg b/org.ops4j.pax.logging.cfg | |
index b9acb61..daf8961 100644 | |
--- a/org.ops4j.pax.logging.cfg | |
+++ b/org.ops4j.pax.logging.cfg | |
@@ -21,14 +21,14 @@ | |
log4j.rootLogger=INFO, out, osgi:* | |
log4j.throwableRenderer=org.apache.log4j.OsgiThrowableRenderer | |
-# To avoid flooding the log when using WARN level on an ssh connection and doing log:tail | |
+# To avoid flooding the log when using DEBUG level on an ssh connection and doing log:tail | |
log4j.logger.org.apache.sshd.server.channel.ChannelSession = INFO | |
# Without this, a TON of topology logs go to output.log | |
log4j.logger.org.opennms.features.topology = WARN | |
-# Display all WARN logs for our code | |
-log4j.category.org.opennms=WARN, out, osgi:* | |
+# Display all DEBUG logs for our code | |
+log4j.category.org.opennms=DEBUG, out, osgi:* | |
log4j.additivity.org.opennms=false | |
# CONSOLE appender not used by default | |
@@ -40,7 +40,7 @@ log4j.appender.stdout.layout.ConversionPattern=%d %-5p %X{bundle.name}:%X{bundle | |
log4j.appender.out=org.apache.log4j.RollingFileAppender | |
log4j.appender.out.layout=org.apache.log4j.PatternLayout | |
log4j.appender.out.layout.ConversionPattern=%d %-5p %X{bundle.name}:%X{bundle.version}(%X{bundle.id}) [%t] %c: %m%n | |
-log4j.appender.out.file=${karaf.data}/log/karaf.log | |
+log4j.appender.out.file=${karaf.base}/logs/karaf.log | |
log4j.appender.out.append=true | |
log4j.appender.out.maxFileSize=1MB | |
log4j.appender.out.maxBackupIndex=10 | |
@@ -52,6 +52,6 @@ log4j.appender.sift.default=karaf | |
log4j.appender.sift.appender=org.apache.log4j.FileAppender | |
log4j.appender.sift.appender.layout=org.apache.log4j.PatternLayout | |
log4j.appender.sift.appender.layout.ConversionPattern=%d %-5p [%t] %c: %m%n | |
-log4j.appender.sift.appender.file=${karaf.data}/log/$\\{bundle.name\\}.log | |
+log4j.appender.sift.appender.file=${karaf.base}/logs/$\\{bundle.name\\}.log | |
log4j.appender.sift.appender.append=true | |
diff --git a/org.ops4j.pax.url.mvn.cfg b/org.ops4j.pax.url.mvn.cfg | |
index 5b431c4..c92ecac 100644 | |
--- a/org.ops4j.pax.url.mvn.cfg | |
+++ b/org.ops4j.pax.url.mvn.cfg | |
@@ -101,3 +101,4 @@ org.ops4j.pax.url.mvn.repositories= \ | |
http://repository.springsource.com/maven/bundles/release@id=springsource.release, \ | |
http://repository.springsource.com/maven/bundles/external@id=springsource.external, \ | |
https://oss.sonatype.org/content/repositories/releases/@id=sonatype | |
+ | |
diff --git a/pluginManifestData.xml b/pluginManifestData.xml | |
new file mode 100644 | |
index 0000000..0328a51 | |
--- /dev/null | |
+++ b/pluginManifestData.xml | |
@@ -0,0 +1,9 @@ | |
+<?xml version="1.0" encoding="UTF-8" standalone="yes"?> | |
+<pluginManagerData> | |
+ <pluginServerPassword>admin</pluginServerPassword> | |
+ <pluginServerUsername>admin</pluginServerUsername> | |
+ <pluginServerUrl>http://localhost:8980/opennms</pluginServerUrl> | |
+ <licenceShoppingCartUrl>http://opennms.org</licenceShoppingCartUrl> | |
+ <karafDataMap/> | |
+ <karafManifestEntryMap/> | |
+</pluginManagerData> | |
diff --git a/poller-config.properties b/poller-config.properties | |
deleted file mode 100644 | |
index 7f3b4b3..0000000 | |
--- a/poller-config.properties | |
+++ /dev/null | |
@@ -1,75 +0,0 @@ | |
-services=SMTP,FTP,IMAP,POP3,TCP,HTTP,HTTPS,NTP | |
-interval=300000 | |
-timeout=3000 | |
-banner=* | |
- | |
-# JBoss | |
-service.JBoss4.monitor=org.opennms.netmgt.poller.monitors.JBoss4Monitor | |
-service. JBoss4.capsd-class=org.opennms.netmgt.capsd.plugins.JBoss4Plugin | |
-service. JBoss4.protocol=JBoss4 | |
-service. JBoss4.port=1099 | |
- | |
-service.JBoss32.monitor=org.opennms.netmgt.poller.monitors.JBoss32Monitor | |
-service. JBoss32.capsd-class=org.opennms.netmgt.capsd.plugins.JBoss32Plugin | |
-service. JBoss32.protocol=JBoss32 | |
-service. JBoss32.port=1099 | |
- | |
-# SMTP | |
-service.SMTP.monitor=org.opennms.netmgt.poller.monitors.SmtpMonitor | |
-service.SMTP.capsd-class=org.opennms.netmgt.capsd.plugins.SmtpPlugin | |
-service.SMTP.protocol=SMTP | |
-service.SMTP.port=25 | |
- | |
-# FTP | |
-service.FTP.monitor=org.opennms.netmgt.poller.monitors.FtpMonitor | |
-service.FTP.capsd-class=org.opennms.netmgt.capsd.plugins.FtpPlugin | |
-service.FTP.protocol=FTP | |
-service.FTP.port=21 | |
- | |
-# Postgres | |
-service.Postgres.monitor=org.opennms.netmgt.poller.monitors.TcpMonitor | |
-service.Postgres.capsd-class=org.opennms.netmgt.capsd.plugins.TcpPlugin | |
-service.Postgres.protocol=Postgres | |
-service.Postgres.port=5432 | |
- | |
-# MySQL | |
-#service.MySQL.monitor=org.opennms.netmgt.poller.monitors.TcpMonitor | |
-#service.MySQL.capsd-class=org.opennms.netmgt.capsd.plugins.TcpPlugin | |
-#service.MySQL.protocol=MySQL | |
-#service.MySQL.port= | |
- | |
-# IMAP | |
-service.IMAP.monitor=org.opennms.netmgt.poller.monitors.ImapMonitor | |
-service.IMAP.capsd-class=org.opennms.netmgt.capsd.plugins.ImapPlugin | |
-service.IMAP.protocol=IMAP | |
-service.IMAP.port=143 | |
- | |
-# POP3 | |
-service.POP3.monitor=org.opennms.netmgt.poller.monitors.Pop3Montior | |
-service.POP3.capsd-class=org.opennms.netmgt.capsd.plugins.Pop3Plugin | |
-service.POP3.protocol=POP3 | |
-service.POP3.port=110 | |
- | |
-# TCP | |
-service.TCP.monitor=org.opennms.netmgt.poller.monitors.TcpMonitor | |
-service.TCP.capsd-class=org.opennms.netmgt.capsd.plugins.TcpPlugin | |
-service.TCP.protocol=TCP | |
-service.TCP.port= | |
- | |
-# HTTP | |
-service.HTTP.monitor=org.opennms.netmgt.poller.monitors.HttpMonitor | |
-service.HTTP.capsd-class=org.opennms.netmgt.capsd.plugins.HttpPlugin | |
-service.HTTP.protocol=HTTP | |
-service.HTTP.port=80:8080:8088 | |
- | |
-# HTTPS | |
-service.HTTPS.monitor=org.opennms.netmgt.poller.monitors.HttpsMonitor | |
-service.HTTPS.capsd-class=org.opennms.netmgt.capsd.plugins.HttpsPlugin | |
-service.HTTPS.protocol=HTTPS | |
-service.HTTPS.port= | |
- | |
-# NTP | |
-service.NTP.monitor=org.opennms.netmgt.poller.monitors.NtpMonitor | |
-service.NTP.capsd-class=org.opennms.netmgt.capsd.plugins.NtpPlugin | |
-service.NTP.protocol=NTP | |
-service.NTP.port=123 | |
diff --git a/poller-configuration.xml b/poller-configuration.xml | |
index 6e7a452..47d8dbf 100644 | |
--- a/poller-configuration.xml | |
+++ b/poller-configuration.xml | |
@@ -8,8 +8,8 @@ | |
<critical-service name="ICMP" /> | |
</node-outage> | |
- <package name="cassandra21x"> | |
- <filter><![CDATA[(IPADDR != '0.0.0.0') & (categoryName == 'Cassandra21x')]]></filter> | |
+ <package name="cassandra-via-jmx"> | |
+ <filter>IPADDR != '0.0.0.0'</filter> | |
<rrd step="300"> | |
<rra>RRA:AVERAGE:0.5:1:2016</rra> | |
<rra>RRA:AVERAGE:0.5:12:1488</rra> | |
@@ -23,9 +23,8 @@ | |
<parameter key="timeout" value="3000"/> | |
<parameter key="protocol" value="rmi"/> | |
<parameter key="urlPath" value="/jmxrmi"/> | |
- <parameter key="ds-name" value="cassandra21x"/> | |
- <parameter key="friendly-name" value="cassandra21x"/> | |
- <parameter key="collection" value="cassandra21x"/> | |
+ <parameter key="rrd-base-name" value="jmx-cassandra"/> | |
+ <parameter key="ds-name" value="jmx-cassandra"/> | |
<parameter key="thresholding-enabled" value="true"/> | |
<parameter key="factory" value="PASSWORD-CLEAR"/> | |
<parameter key="username" value="cassandra-username"/> | |
@@ -36,31 +35,14 @@ | |
<parameter key="tests.joined" value="storage.Joined"/> | |
<parameter key="tests.unreachables" value="empty(storage.UnreachableNodes)"/> | |
</service> | |
- <downtime interval="30000" begin="0" end="300000" /><!-- 30s, 0, 5m --> | |
- <downtime interval="300000" begin="300000" end="43200000" /><!-- 5m, 5m, 12h --> | |
- <downtime interval="600000" begin="43200000" end="432000000" /><!-- 10m, 12h, 5d --> | |
- <downtime begin="432000000" delete="true" /><!-- anything after 5 days delete --> | |
- </package> | |
- | |
- <package name="cassandra21x-newts"> | |
- <filter><![CDATA[(IPADDR != '0.0.0.0') & (catincCassandra21x & catincNewts)]]></filter> | |
- <rrd step="300"> | |
- <rra>RRA:AVERAGE:0.5:1:2016</rra> | |
- <rra>RRA:AVERAGE:0.5:12:1488</rra> | |
- <rra>RRA:AVERAGE:0.5:288:366</rra> | |
- <rra>RRA:MAX:0.5:288:366</rra> | |
- <rra>RRA:MIN:0.5:288:366</rra> | |
- </rrd> | |
<service name="JMX-Cassandra-Newts" interval="300000" user-defined="false" status="on"> | |
<parameter key="port" value="7199"/> | |
<parameter key="retry" value="2"/> | |
<parameter key="timeout" value="3000"/> | |
<parameter key="protocol" value="rmi"/> | |
<parameter key="urlPath" value="/jmxrmi"/> | |
- <parameter key="rrd-base-name" value="cassandra21x-newts"/> | |
- <parameter key="ds-name" value="cassandra21x-newts"/> | |
- <parameter key="friendly-name" value="cassandra21x-newts"/> | |
- <parameter key="collection" value="cassandra21x-newts"/> | |
+ <parameter key="rrd-base-name" value="jmx-cassandra-newts"/> | |
+ <parameter key="ds-name" value="jmx-cassandra-newts"/> | |
<parameter key="thresholding-enabled" value="true"/> | |
<parameter key="factory" value="PASSWORD-CLEAR"/> | |
<parameter key="username" value="cassandra-username"/> | |
@@ -109,6 +91,14 @@ | |
<parameter key="rrd-base-name" value="dns" /> | |
<parameter key="ds-name" value="dns" /> | |
</service> | |
+ <service name="Elasticsearch" interval="300000" user-defined="false" status="on"> | |
+ <parameter key="retry" value="1"/> | |
+ <parameter key="timeout" value="3000"/> | |
+ <parameter key="port" value="9200"/> | |
+ <parameter key="url" value="/_cluster/stats"/> | |
+ <parameter key="response" value="200-202,299"/> | |
+ <parameter key="response-text" value="~.*status.:.green.*"/> | |
+ </service> | |
<service name="SMTP" interval="300000" user-defined="false" status="on"> | |
<parameter key="retry" value="1" /> | |
<parameter key="timeout" value="3000" /> | |
@@ -258,6 +248,25 @@ | |
<parameter key="timeout" value="3000"/> | |
<parameter key="rrd-repository" value="/opt/opennms/share/rrd/response" /> | |
</service> | |
+ <service name="JMX-Minion" interval="300000" user-defined="false" status="on"> | |
+ <parameter key="port" value="1299"/> | |
+ <parameter key="retry" value="2"/> | |
+ <parameter key="timeout" value="3000"/> | |
+ <parameter key="urlPath" value="/karaf-minion"/> | |
+ <parameter key="factory" value="PASSWORD-CLEAR"/> | |
+ <parameter key="username" value="admin"/> | |
+ <parameter key="password" value="admin"/> | |
+ <parameter key="rrd-repository" value="/opt/opennms/share/rrd/response" /> | |
+ </service> | |
+ <service name="JMX-Kafka" interval="300000" user-defined="false" status="on"> | |
+ <parameter key="port" value="9999"/> | |
+ <parameter key="retry" value="2"/> | |
+ <parameter key="timeout" value="3000"/> | |
+ <parameter key="factory" value="PASSWORD-CLEAR"/> | |
+ <parameter key="username" value="admin"/> | |
+ <parameter key="password" value="admin"/> | |
+ <parameter key="rrd-repository" value="/opt/opennms/share/rrd/response" /> | |
+ </service> | |
<service name="VMwareCim-HostSystem" interval="300000" user-defined="false" status="on"> | |
<parameter key="retry" value="2"/> | |
<parameter key="timeout" value="3000"/> | |
@@ -274,6 +283,9 @@ | |
<parameter key="port" value="3389" /> | |
<parameter key="timeout" value="3000" /> | |
</service> | |
+ <service name="Minion-Heartbeat" interval="30000" user-defined="false" status="on"> | |
+ <parameter key="period" value="30000" /> <!-- Service interval should be same as period --> | |
+ </service> | |
<downtime interval="30000" begin="0" end="300000" /><!-- 30s, 0, 5m --> | |
<downtime interval="300000" begin="300000" end="43200000" /><!-- 5m, 5m, 12h --> | |
@@ -324,10 +336,12 @@ | |
<monitor service="HypericHQ" class-name="org.opennms.netmgt.poller.monitors.PageSequenceMonitor" /> | |
<monitor service="SMTP" class-name="org.opennms.netmgt.poller.monitors.SmtpMonitor" /> | |
<monitor service="DNS" class-name="org.opennms.netmgt.poller.monitors.DnsMonitor" /> | |
+ <monitor service="Elasticsearch" class-name="org.opennms.netmgt.poller.monitors.HttpMonitor" /> | |
<monitor service="FTP" class-name="org.opennms.netmgt.poller.monitors.FtpMonitor" /> | |
<monitor service="SNMP" class-name="org.opennms.netmgt.poller.monitors.SnmpMonitor" /> | |
<monitor service="Oracle" class-name="org.opennms.netmgt.poller.monitors.TcpMonitor" /> | |
<monitor service="Postgres" class-name="org.opennms.netmgt.poller.monitors.TcpMonitor" /> | |
+ <monitor service="Minion-Heartbeat" class-name="org.opennms.netmgt.poller.monitors.MinionHeartbeatMonitor" /> | |
<monitor service="MySQL" class-name="org.opennms.netmgt.poller.monitors.TcpMonitor" /> | |
<monitor service="SQLServer" class-name="org.opennms.netmgt.poller.monitors.TcpMonitor" /> | |
<monitor service="SSH" class-name="org.opennms.netmgt.poller.monitors.SshMonitor" /> | |
@@ -337,6 +351,8 @@ | |
<monitor service="NRPE-NoSSL" class-name="org.opennms.netmgt.poller.monitors.NrpeMonitor" /> | |
<monitor service="Windows-Task-Scheduler" class-name="org.opennms.netmgt.poller.monitors.Win32ServiceMonitor" /> | |
<monitor service="OpenNMS-JVM" class-name="org.opennms.netmgt.poller.monitors.Jsr160Monitor" /> | |
+ <monitor service="JMX-Minion" class-name="org.opennms.netmgt.poller.monitors.Jsr160Monitor" /> | |
+ <monitor service="JMX-Kafka" class-name="org.opennms.netmgt.poller.monitors.Jsr160Monitor" /> | |
<monitor service="VMwareCim-HostSystem" class-name="org.opennms.netmgt.poller.monitors.VmwareCimMonitor"/> | |
<monitor service="VMware-ManagedEntity" class-name="org.opennms.netmgt.poller.monitors.VmwareMonitor"/> | |
<monitor service="MS-RDP" class-name="org.opennms.netmgt.poller.monitors.TcpMonitor" /> | |
diff --git a/report-templates/AssetManagementMaintExpired.jasper b/report-templates/AssetManagementMaintExpired.jasper | |
index 3ed64ae..00ef70a 100644 | |
Binary files a/report-templates/AssetManagementMaintExpired.jasper and b/report-templates/AssetManagementMaintExpired.jasper differ | |
diff --git a/report-templates/AssetManagementMaintStrategy.jasper b/report-templates/AssetManagementMaintStrategy.jasper | |
index 8c1094a..cdfa4e0 100644 | |
Binary files a/report-templates/AssetManagementMaintStrategy.jasper and b/report-templates/AssetManagementMaintStrategy.jasper differ | |
diff --git a/report-templates/AvailabilitySummary.jasper b/report-templates/AvailabilitySummary.jasper | |
index abc12a5..50b024b 100644 | |
Binary files a/report-templates/AvailabilitySummary.jasper and b/report-templates/AvailabilitySummary.jasper differ | |
diff --git a/report-templates/AveragePeakTrafficRates.jasper b/report-templates/AveragePeakTrafficRates.jasper | |
index 561aa8e..1262858 100644 | |
Binary files a/report-templates/AveragePeakTrafficRates.jasper and b/report-templates/AveragePeakTrafficRates.jasper differ | |
diff --git a/report-templates/DiskUsageForCTX.jasper b/report-templates/DiskUsageForCTX.jasper | |
index d9fbcbe..1024ad1 100644 | |
Binary files a/report-templates/DiskUsageForCTX.jasper and b/report-templates/DiskUsageForCTX.jasper differ | |
diff --git a/report-templates/Early-Morning-Report.jasper b/report-templates/Early-Morning-Report.jasper | |
index 30983a8..0148fbb 100644 | |
Binary files a/report-templates/Early-Morning-Report.jasper and b/report-templates/Early-Morning-Report.jasper differ | |
diff --git a/report-templates/EventAnalysis.jasper b/report-templates/EventAnalysis.jasper | |
index ef016f6..888cfad 100644 | |
Binary files a/report-templates/EventAnalysis.jasper and b/report-templates/EventAnalysis.jasper differ | |
diff --git a/report-templates/InterfaceAvailabilityReport.jasper b/report-templates/InterfaceAvailabilityReport.jasper | |
index ad4857f..12b02d3 100644 | |
Binary files a/report-templates/InterfaceAvailabilityReport.jasper and b/report-templates/InterfaceAvailabilityReport.jasper differ | |
diff --git a/report-templates/NodeAvailabilityReport.jasper b/report-templates/NodeAvailabilityReport.jasper | |
index ed762f3..e8c2f07 100644 | |
Binary files a/report-templates/NodeAvailabilityReport.jasper and b/report-templates/NodeAvailabilityReport.jasper differ | |
diff --git a/report-templates/ResponseTime.jasper b/report-templates/ResponseTime.jasper | |
index d70d0fb..d1be959 100644 | |
Binary files a/report-templates/ResponseTime.jasper and b/report-templates/ResponseTime.jasper differ | |
diff --git a/report-templates/ResponseTimeCharts.jasper b/report-templates/ResponseTimeCharts.jasper | |
index 0e558ed..3a8f627 100644 | |
Binary files a/report-templates/ResponseTimeCharts.jasper and b/report-templates/ResponseTimeCharts.jasper differ | |
diff --git a/report-templates/ResponseTimeSummary.jasper b/report-templates/ResponseTimeSummary.jasper | |
index 16afc5c..c85473d 100644 | |
Binary files a/report-templates/ResponseTimeSummary.jasper and b/report-templates/ResponseTimeSummary.jasper differ | |
diff --git a/report-templates/SerialInterfaceUtilizationSummary.jasper b/report-templates/SerialInterfaceUtilizationSummary.jasper | |
index 78a47c1..1dfa18d 100644 | |
Binary files a/report-templates/SerialInterfaceUtilizationSummary.jasper and b/report-templates/SerialInterfaceUtilizationSummary.jasper differ | |
diff --git a/report-templates/SnmpInterfaceOperAvailabilityReport.jasper b/report-templates/SnmpInterfaceOperAvailabilityReport.jasper | |
index 2172fb3..81759a9 100644 | |
Binary files a/report-templates/SnmpInterfaceOperAvailabilityReport.jasper and b/report-templates/SnmpInterfaceOperAvailabilityReport.jasper differ | |
diff --git a/report-templates/TopIOWait.jasper b/report-templates/TopIOWait.jasper | |
new file mode 100644 | |
index 0000000..f7a0f66 | |
Binary files /dev/null and b/report-templates/TopIOWait.jasper differ | |
diff --git a/report-templates/TopIOWait.jrxml b/report-templates/TopIOWait.jrxml | |
new file mode 100644 | |
index 0000000..71b194b | |
--- /dev/null | |
+++ b/report-templates/TopIOWait.jrxml | |
@@ -0,0 +1,271 @@ | |
+<?xml version="1.0" encoding="UTF-8"?> | |
+<!-- Created with Jaspersoft Studio version 6.1.1.final using JasperReports Library version 6.1.1 --> | |
+<!-- 2016-09-14T12:01:14 --> | |
+<jasperReport xmlns="http://jasperreports.sourceforge.net/jasperreports" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://jasperreports.sourceforge.net/jasperreports http://jasperreports.sourceforge.net/xsd/jasperreport.xsd" name="Average and Peak Traffic Rates" pageWidth="595" pageHeight="842" whenNoDataType="NoDataSection" columnWidth="555" leftMargin="20" rightMargin="20" topMargin="20" bottomMargin="20" uuid="dd0c0584-c0b6-4860-a6fd-62182c4d11a5"> | |
+ <property name="ireport.zoom" value="1.2100000000000006"/> | |
+ <property name="ireport.x" value="0"/> | |
+ <property name="ireport.y" value="0"/> | |
+ <property name="com.jaspersoft.studio.unit." value="pixel"/> | |
+ <property name="com.jaspersoft.studio.data.sql.tables" value=""/> | |
+ <property name="com.jaspersoft.studio.data.defaultdataadapter" value="opennms"/> | |
+ <template><![CDATA[$P{ONMS_REPORT_DIR} + "/assets/styles/defaultStyles.jrtx"]]></template> | |
+ <style name="Report_Title" forecolor="#000000" fontSize="20"/> | |
+ <style name="Report_Subtitle" forecolor="#000000" vTextAlign="Middle" vImageAlign="Middle" fontSize="10" isBold="false" isItalic="true" isUnderline="false" isStrikeThrough="false"/> | |
+ <style name="Table_Detail" hTextAlign="Left" hImageAlign="Left" vTextAlign="Middle" vImageAlign="Middle" fontName="SansSerif" fontSize="10" isBold="false" isItalic="false" isUnderline="false" isStrikeThrough="false"/> | |
+ <style name="Table_Grid" mode="Transparent" forecolor="#FFFFFF" isBold="false" isItalic="false" isUnderline="false" isStrikeThrough="false"> | |
+ <pen lineWidth="0.0" lineColor="#FFFFFF"/> | |
+ <box> | |
+ <bottomPen lineWidth="1.0"/> | |
+ </box> | |
+ <conditionalStyle> | |
+ <conditionExpression><![CDATA[new Boolean($V{style_helper_COUNT}%new Integer("5") == new Integer("0"))]]></conditionExpression> | |
+ <style mode="Opaque" forecolor="#999999"> | |
+ <box> | |
+ <bottomPen lineWidth="1.0"/> | |
+ </box> | |
+ </style> | |
+ </conditionalStyle> | |
+ </style> | |
+ <style name="Page_Footer" fontSize="10" isBold="false" isItalic="false" isUnderline="false" isStrikeThrough="false"/> | |
+ <style name="Table_Header" isBold="true" isItalic="false" isUnderline="false" isStrikeThrough="false"/> | |
+ <style name="Table_Header_BG" mode="Opaque" backcolor="#CCFFCC"/> | |
+ <style name="Table_Surveillance_Cat" forecolor="#000000" vTextAlign="Middle" vImageAlign="Middle" fontSize="10" isBold="true" isItalic="false" isUnderline="false" isStrikeThrough="false"/> | |
+ <style name="Table_Surveillance_Cat_BG" mode="Opaque" backcolor="#DFDFDF"/> | |
+ <style name="Table_Surveillance_Cat_Footer_BG" mode="Transparent" backcolor="#FFFFFF"> | |
+ <pen lineWidth="1.0" lineStyle="Double"/> | |
+ </style> | |
+ <style name="Table_Surveillance_Cat_Footer" hTextAlign="Center" hImageAlign="Center" vTextAlign="Middle" vImageAlign="Middle" fontSize="10" isBold="true" isItalic="false" isUnderline="false" isStrikeThrough="false"> | |
+ <box> | |
+ <bottomPen lineWidth="0.0" lineStyle="Double" lineColor="#000000"/> | |
+ </box> | |
+ </style> | |
+ <style name="Table_Surveillance_Cat_Footer_Line" hTextAlign="Center" hImageAlign="Center" vTextAlign="Middle" vImageAlign="Middle" isBold="true" isItalic="false" isUnderline="false" isStrikeThrough="false"> | |
+ <box> | |
+ <bottomPen lineWidth="2.0" lineStyle="Double" lineColor="#000000"/> | |
+ </box> | |
+ </style> | |
+ <style name="Surveillance_Category_Group" mode="Opaque" backcolor="#CCFFCC" hTextAlign="Left" hImageAlign="Left" vTextAlign="Middle" vImageAlign="Middle" fontSize="12" isBold="true" isItalic="false" isUnderline="false" isStrikeThrough="false"/> | |
+ <style name="Node_Group" mode="Opaque" backcolor="#DFDFDF" hTextAlign="Left" hImageAlign="Left" vTextAlign="Middle" vImageAlign="Middle" fontSize="10" isBold="false" isItalic="false" isUnderline="false" isStrikeThrough="false"/> | |
+ <style name="table"> | |
+ <box> | |
+ <pen lineWidth="1.0" lineColor="#000000"/> | |
+ </box> | |
+ </style> | |
+ <style name="table_TH" mode="Opaque" backcolor="#F0F8FF"> | |
+ <box> | |
+ <topPen lineWidth="0.5" lineColor="#000000"/> | |
+ <bottomPen lineWidth="0.5" lineColor="#000000"/> | |
+ </box> | |
+ </style> | |
+ <style name="table_CH" mode="Opaque" backcolor="#BFE1FF"> | |
+ <box> | |
+ <topPen lineWidth="0.5" lineColor="#000000"/> | |
+ <bottomPen lineWidth="0.5" lineColor="#000000"/> | |
+ </box> | |
+ </style> | |
+ <style name="table_TD" mode="Opaque" backcolor="#FFFFFF"> | |
+ <box> | |
+ <topPen lineWidth="0.5" lineColor="#000000"/> | |
+ <bottomPen lineWidth="0.5" lineColor="#000000"/> | |
+ </box> | |
+ </style> | |
+ <style name="style1"/> | |
+ <style name="Interface_Header" hTextAlign="Center" hImageAlign="Center" vTextAlign="Middle" vImageAlign="Middle" isBold="true" isItalic="false" isUnderline="false" isStrikeThrough="false"/> | |
+ <parameter name="ONMS_REPORT_DIR" class="java.lang.String" isForPrompting="false"> | |
+ <parameterDescription><![CDATA[The directory where all reports can be found]]></parameterDescription> | |
+ <defaultValueExpression><![CDATA["/opt/opennms/etc/report-templates"]]></defaultValueExpression> | |
+ </parameter> | |
+ <parameter name="COMPANY_LOGO" class="java.lang.String" isForPrompting="false"> | |
+ <defaultValueExpression><![CDATA[$P{ONMS_REPORT_DIR} + "/assets/images/company-logo.png"]]></defaultValueExpression> | |
+ </parameter> | |
+ <parameter name="SUBREPORT_DIR" class="java.lang.String" isForPrompting="false"> | |
+ <defaultValueExpression><![CDATA[$P{ONMS_REPORT_DIR} + "/subreports/"]]></defaultValueExpression> | |
+ </parameter> | |
+ <queryString> | |
+ <![CDATA[select | |
+n.nodeid, n.nodelabel, n.nodesyscontact, n.nodesysdescription, n.nodesyslocation, d.value | |
+from | |
+ node n, | |
+ (select * from statisticsreport order by id desc limit 1) s, | |
+ statisticsreportdata d, | |
+ resourcereference r | |
+where | |
+ s.id=d.reportid | |
+ and s.name='TopN_IOWait' | |
+ and d.resourceid=r.id | |
+ and r.resourceid='nodeSource['||n.foreignsource||'%3A'||n.foreignid||'].nodeSnmp[]' | |
+order by d.value desc]]> | |
+ </queryString> | |
+ <field name="nodeid" class="java.lang.Integer"/> | |
+ <field name="nodelabel" class="java.lang.String"/> | |
+ <field name="nodesyscontact" class="java.lang.String"/> | |
+ <field name="nodesysdescription" class="java.lang.String"/> | |
+ <field name="nodesyslocation" class="java.lang.String"/> | |
+ <field name="value" class="java.lang.Double"/> | |
+ <background> | |
+ <band splitType="Stretch"/> | |
+ </background> | |
+ <title> | |
+ <band height="4" splitType="Stretch"/> | |
+ </title> | |
+ <pageHeader> | |
+ <band height="80" splitType="Stretch"> | |
+ <staticText> | |
+ <reportElement style="Title" x="0" y="0" width="355" height="30" uuid="e2bbd5aa-7bb3-4da8-9975-db0696053b3e"> | |
+ <property name="local_mesure_unitwidth" value="pixel"/> | |
+ <property name="com.jaspersoft.studio.unit.width" value="px"/> | |
+ <property name="local_mesure_unitheight" value="pixel"/> | |
+ <property name="com.jaspersoft.studio.unit.height" value="px"/> | |
+ </reportElement> | |
+ <textElement verticalAlignment="Middle"> | |
+ <font size="20" isBold="true"/> | |
+ </textElement> | |
+ <text><![CDATA[Top 20 nodes by I/O Wait]]></text> | |
+ </staticText> | |
+ <image> | |
+ <reportElement x="360" y="0" width="194" height="50" uuid="7a2e4fd9-7cb3-4ee0-a739-39549ded3164"> | |
+ <property name="local_mesure_unitwidth" value="pixel"/> | |
+ <property name="com.jaspersoft.studio.unit.width" value="px"/> | |
+ <property name="local_mesure_unitheight" value="pixel"/> | |
+ <property name="com.jaspersoft.studio.unit.height" value="px"/> | |
+ </reportElement> | |
+ <imageExpression><![CDATA[$P{COMPANY_LOGO}]]></imageExpression> | |
+ </image> | |
+ <staticText> | |
+ <reportElement x="0" y="30" width="354" height="30" uuid="49028b4e-cd3b-4a4f-b33c-f70c20e4a543"/> | |
+ <text><![CDATA[Average for the last 24 hours.]]></text> | |
+ </staticText> | |
+ </band> | |
+ </pageHeader> | |
+ <columnHeader> | |
+ <band height="19" splitType="Stretch"> | |
+ <line> | |
+ <reportElement x="0" y="0" width="555" height="1" uuid="c5cc50fc-34f3-4f66-a76a-c648728b32a8"> | |
+ <property name="local_mesure_unitx" value="pixel"/> | |
+ <property name="com.jaspersoft.studio.unit.x" value="px"/> | |
+ <property name="local_mesure_unitwidth" value="pixel"/> | |
+ <property name="com.jaspersoft.studio.unit.width" value="px"/> | |
+ </reportElement> | |
+ <graphicElement> | |
+ <pen lineWidth="1.5"/> | |
+ </graphicElement> | |
+ </line> | |
+ <staticText> | |
+ <reportElement x="0" y="1" width="140" height="18" uuid="405bc934-e220-4d23-a4bd-64377740d5ca"/> | |
+ <textElement textAlignment="Center"> | |
+ <font isBold="true"/> | |
+ </textElement> | |
+ <text><![CDATA[Node Label]]></text> | |
+ </staticText> | |
+ <staticText> | |
+ <reportElement x="140" y="1" width="280" height="18" uuid="42f8b403-c7e6-4aa3-9936-749fac94cb66"/> | |
+ <textElement textAlignment="Center"> | |
+ <font isBold="true"/> | |
+ </textElement> | |
+ <text><![CDATA[Node details]]></text> | |
+ </staticText> | |
+ <staticText> | |
+ <reportElement x="440" y="1" width="114" height="18" uuid="b68e8040-9386-4658-b857-dc354a53b2dc"/> | |
+ <textElement textAlignment="Right"> | |
+ <font isBold="true"/> | |
+ </textElement> | |
+ <text><![CDATA[ I/O Wait (raw)]]></text> | |
+ </staticText> | |
+ </band> | |
+ </columnHeader> | |
+ <detail> | |
+ <band height="22" splitType="Stretch"> | |
+ <textField> | |
+ <reportElement x="0" y="0" width="140" height="16" uuid="27a3911b-2cb4-4f8c-8113-cd87a3b7ddca"/> | |
+ <textFieldExpression><![CDATA[$F{nodelabel}]]></textFieldExpression> | |
+ </textField> | |
+ <textField> | |
+ <reportElement x="140" y="0" width="280" height="16" forecolor="#9C9C9C" uuid="ee591f75-53af-49d4-96aa-ef1022246013"/> | |
+ <textFieldExpression><![CDATA[$F{nodesysdescription}]]></textFieldExpression> | |
+ </textField> | |
+ <textField pattern="#0"> | |
+ <reportElement x="440" y="0" width="114" height="16" uuid="127ebdb6-ffa0-4571-bfb9-ce004b14b246"/> | |
+ <textElement textAlignment="Right"/> | |
+ <textFieldExpression><![CDATA[$F{value}]]></textFieldExpression> | |
+ </textField> | |
+ </band> | |
+ </detail> | |
+ <columnFooter> | |
+ <band height="259" splitType="Stretch"> | |
+ <property name="com.jaspersoft.studio.unit.height" value="pixel"/> | |
+ <lineChart> | |
+ <chart evaluationTime="Report"> | |
+ <reportElement x="0" y="0" width="554" height="258" uuid="1418a549-330a-4a9e-87fb-0d2448131ef4"/> | |
+ <chartTitle/> | |
+ <chartSubtitle/> | |
+ <chartLegend/> | |
+ </chart> | |
+ <categoryDataset> | |
+ <categorySeries> | |
+ <seriesExpression><![CDATA["I/O WAIT"]]></seriesExpression> | |
+ <categoryExpression><![CDATA[$F{nodelabel}]]></categoryExpression> | |
+ <valueExpression><![CDATA[$F{value}]]></valueExpression> | |
+ <labelExpression><![CDATA[$F{nodelabel}]]></labelExpression> | |
+ </categorySeries> | |
+ </categoryDataset> | |
+ <linePlot> | |
+ <plot labelRotation="45.0"/> | |
+ <categoryAxisFormat labelRotation="45.0"> | |
+ <axisFormat/> | |
+ </categoryAxisFormat> | |
+ <valueAxisFormat> | |
+ <axisFormat/> | |
+ </valueAxisFormat> | |
+ </linePlot> | |
+ </lineChart> | |
+ </band> | |
+ </columnFooter> | |
+ <pageFooter> | |
+ <band height="35" splitType="Stretch"> | |
+ <property name="local_mesure_unitheight" value="pixel"/> | |
+ <property name="com.jaspersoft.studio.unit.height" value="px"/> | |
+ <line> | |
+ <reportElement x="0" y="10" width="554" height="1" uuid="ec45ea9b-d9e4-4343-9cf9-512b9db8e881"/> | |
+ <graphicElement> | |
+ <pen lineWidth="1.5"/> | |
+ </graphicElement> | |
+ </line> | |
+ <textField> | |
+ <reportElement style="Paging-Footer" x="451" y="13" width="80" height="20" uuid="eb684bd0-d9e6-4a31-a516-b4edf17125d7"/> | |
+ <textElement textAlignment="Right"/> | |
+ <textFieldExpression><![CDATA["Page "+$V{PAGE_NUMBER}+" of"]]></textFieldExpression> | |
+ </textField> | |
+ <textField evaluationTime="Report"> | |
+ <reportElement style="Paging-Footer" x="531" y="13" width="24" height="20" uuid="647f6ab1-e658-4c3c-b0d7-356b2366cba4"/> | |
+ <textFieldExpression><![CDATA[" " + $V{PAGE_NUMBER}]]></textFieldExpression> | |
+ </textField> | |
+ <textField pattern="yyyy/MM/dd HH:mm:ss"> | |
+ <reportElement style="Creation-Date" x="0" y="13" width="355" height="20" uuid="25086ab7-3068-458e-8feb-e35e07b2b9e5"> | |
+ <property name="local_mesure_unitwidth" value="pixel"/> | |
+ <property name="com.jaspersoft.studio.unit.width" value="px"/> | |
+ </reportElement> | |
+ <textFieldExpression><![CDATA[new java.util.Date()]]></textFieldExpression> | |
+ </textField> | |
+ </band> | |
+ </pageFooter> | |
+ <noData> | |
+ <band height="155"> | |
+ <line> | |
+ <reportElement x="0" y="80" width="555" height="1" uuid="3904dc8d-463e-4824-b19c-5663db0bc38a"> | |
+ <property name="local_mesure_unity" value="pixel"/> | |
+ <property name="com.jaspersoft.studio.unit.y" value="px"/> | |
+ </reportElement> | |
+ <graphicElement> | |
+ <pen lineWidth="1.5"/> | |
+ </graphicElement> | |
+ </line> | |
+ <textField> | |
+ <reportElement x="0" y="85" width="555" height="59" uuid="5fa4caf7-f9b2-4ab4-8570-ecd5b55057fa"> | |
+ <property name="local_mesure_unity" value="pixel"/> | |
+ <property name="com.jaspersoft.studio.unit.y" value="px"/> | |
+ </reportElement> | |
+ <textFieldExpression><![CDATA["There is no data for this report yet."]]></textFieldExpression> | |
+ </textField> | |
+ </band> | |
+ </noData> | |
+</jasperReport> | |
diff --git a/report-templates/TotalBytesTransferredByInterface.jasper b/report-templates/TotalBytesTransferredByInterface.jasper | |
index ee677ff..ad3afff 100644 | |
Binary files a/report-templates/TotalBytesTransferredByInterface.jasper and b/report-templates/TotalBytesTransferredByInterface.jasper differ | |
diff --git a/report-templates/sample-report.jasper b/report-templates/sample-report.jasper | |
index fabefe3..2f2f826 100644 | |
Binary files a/report-templates/sample-report.jasper and b/report-templates/sample-report.jasper differ | |
diff --git a/report-templates/subreports/95thPercentileTrafficRate_subreport.jasper b/report-templates/subreports/95thPercentileTrafficRate_subreport.jasper | |
index b7504cc..55ae735 100644 | |
Binary files a/report-templates/subreports/95thPercentileTrafficRate_subreport.jasper and b/report-templates/subreports/95thPercentileTrafficRate_subreport.jasper differ | |
diff --git a/report-templates/subreports/AvailabilitySummaryChart_subreport.jasper b/report-templates/subreports/AvailabilitySummaryChart_subreport.jasper | |
index c207f8a..cfa17af 100644 | |
Binary files a/report-templates/subreports/AvailabilitySummaryChart_subreport.jasper and b/report-templates/subreports/AvailabilitySummaryChart_subreport.jasper differ | |
diff --git a/report-templates/subreports/DiskUsageForCTXServers_subreport1.jasper b/report-templates/subreports/DiskUsageForCTXServers_subreport1.jasper | |
index c78b23a..26d0c91 100644 | |
Binary files a/report-templates/subreports/DiskUsageForCTXServers_subreport1.jasper and b/report-templates/subreports/DiskUsageForCTXServers_subreport1.jasper differ | |
diff --git a/report-templates/subreports/InterfaceAvailabilityReport_subreport1.jasper b/report-templates/subreports/InterfaceAvailabilityReport_subreport1.jasper | |
index 8fdfd63..0657413 100644 | |
Binary files a/report-templates/subreports/InterfaceAvailabilityReport_subreport1.jasper and b/report-templates/subreports/InterfaceAvailabilityReport_subreport1.jasper differ | |
diff --git a/report-templates/subreports/NodeId_to_NodeLabel_subreport.jasper b/report-templates/subreports/NodeId_to_NodeLabel_subreport.jasper | |
index c7a2fe1..3f6f89a 100644 | |
Binary files a/report-templates/subreports/NodeId_to_NodeLabel_subreport.jasper and b/report-templates/subreports/NodeId_to_NodeLabel_subreport.jasper differ | |
diff --git a/report-templates/subreports/PeakTraffic_subreport.jasper b/report-templates/subreports/PeakTraffic_subreport.jasper | |
index ee62ff1..a4b951a 100644 | |
Binary files a/report-templates/subreports/PeakTraffic_subreport.jasper and b/report-templates/subreports/PeakTraffic_subreport.jasper differ | |
diff --git a/report-templates/subreports/ResponseTimeSummary_Availability_Offenders_subreport.jasper b/report-templates/subreports/ResponseTimeSummary_Availability_Offenders_subreport.jasper | |
index 76c351b..32c1336 100644 | |
Binary files a/report-templates/subreports/ResponseTimeSummary_Availability_Offenders_subreport.jasper and b/report-templates/subreports/ResponseTimeSummary_Availability_Offenders_subreport.jasper differ | |
diff --git a/report-templates/subreports/ResponseTimeSummary_Availability_Offenders_subreport.jrxml b/report-templates/subreports/ResponseTimeSummary_Availability_Offenders_subreport.jrxml | |
index db8005f..e1a9165 100644 | |
--- a/report-templates/subreports/ResponseTimeSummary_Availability_Offenders_subreport.jrxml | |
+++ b/report-templates/subreports/ResponseTimeSummary_Availability_Offenders_subreport.jrxml | |
@@ -98,7 +98,7 @@ FROM | |
outages, service, events, ifservices | |
WHERE | |
(iflostservice, COALESCE(ifregainedservice,'$P!{END_TIME_STRING}'::TIMESTAMP)) OVERLAPS ('$P!{START_TIME_STRING}'::TIMESTAMP, '$P!{START_TIME_STRING}'::TIMESTAMP + '$P!{TIME_RANGE}'::INTERVAL) AND | |
- outages.serviceid = service.serviceid AND | |
+ outages.ifserviceid = service.serviceid AND | |
service.servicename = 'ICMP' AND | |
outages.svclosteventid = events.eventid AND | |
events.eventuei = 'uei.opennms.org/nodes/nodeDown') | |
diff --git a/report-templates/subreports/ResponseTimeSummary_Availability_subreport.jasper b/report-templates/subreports/ResponseTimeSummary_Availability_subreport.jasper | |
index 2079e9e..730caf9 100644 | |
Binary files a/report-templates/subreports/ResponseTimeSummary_Availability_subreport.jasper and b/report-templates/subreports/ResponseTimeSummary_Availability_subreport.jasper differ | |
diff --git a/report-templates/subreports/ResponseTimeSummary_Response_Offenders_subreport.jasper b/report-templates/subreports/ResponseTimeSummary_Response_Offenders_subreport.jasper | |
index 3554356..58f66a9 100644 | |
Binary files a/report-templates/subreports/ResponseTimeSummary_Response_Offenders_subreport.jasper and b/report-templates/subreports/ResponseTimeSummary_Response_Offenders_subreport.jasper differ | |
diff --git a/report-templates/subreports/ResponseTimeSummary_subreport.jasper b/report-templates/subreports/ResponseTimeSummary_subreport.jasper | |
index 9b17688..284cd26 100644 | |
Binary files a/report-templates/subreports/ResponseTimeSummary_subreport.jasper and b/report-templates/subreports/ResponseTimeSummary_subreport.jasper differ | |
diff --git a/report-templates/subreports/ResponseTime_subreport1.jasper b/report-templates/subreports/ResponseTime_subreport1.jasper | |
index 8201404..9f678b4 100644 | |
Binary files a/report-templates/subreports/ResponseTime_subreport1.jasper and b/report-templates/subreports/ResponseTime_subreport1.jasper differ | |
diff --git a/report-templates/subreports/Top25PercentDown_subreport.jasper b/report-templates/subreports/Top25PercentDown_subreport.jasper | |
index aaca5f9..875434d 100644 | |
Binary files a/report-templates/subreports/Top25PercentDown_subreport.jasper and b/report-templates/subreports/Top25PercentDown_subreport.jasper differ | |
diff --git a/report-templates/subreports/TotalBytesTransferredByInterface_subreport1.jasper b/report-templates/subreports/TotalBytesTransferredByInterface_subreport1.jasper | |
index 1cc45dd..e1721d3 100644 | |
Binary files a/report-templates/subreports/TotalBytesTransferredByInterface_subreport1.jasper and b/report-templates/subreports/TotalBytesTransferredByInterface_subreport1.jasper differ | |
diff --git a/shell.init.script b/shell.init.script | |
index edcd769..74adbde 100644 | |
--- a/shell.init.script | |
+++ b/shell.init.script | |
@@ -124,4 +124,4 @@ enable-3x-aliases = { | |
echo "Karaf 3.x aliases enabled" | |
-} | |
\ No newline at end of file | |
+} | |
diff --git a/snmp-graph.properties.d/activemq-graph.properties b/snmp-graph.properties.d/activemq-graph.properties | |
new file mode 100644 | |
index 0000000..672c165 | |
--- /dev/null | |
+++ b/snmp-graph.properties.d/activemq-graph.properties | |
@@ -0,0 +1,141 @@ | |
+reports=activemq.total.consumer.count, \ | |
+activemq.total.connection.count, \ | |
+activemq.message.count, \ | |
+activemq.consumer.producer.count, \ | |
+activemq.storage.limit, \ | |
+activemq.storage.percentage, \ | |
+activemq.temp.storage.limit, \ | |
+activemq.temp.storage.percentage | |
+ | |
+# | |
+# Consumer Count | |
+# | |
+report.activemq.total.consumer.count.name=ActiveMQ Consumer Count | |
+report.activemq.total.consumer.count.columns=TtlConsumerCnt | |
+report.activemq.total.consumer.count.type=interfaceSnmp | |
+report.activemq.total.consumer.count.command=--title="ActiveMQ Consumer Count" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Consumer Count" \ | |
+ DEF:val1={rrd1}:TtlConsumerCnt:AVERAGE \ | |
+ AREA:val1#c4a000 \ | |
+ LINE2:val1#000000:"Consumer Count " \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+# | |
+# Total Connection Count | |
+# | |
+report.activemq.total.connection.count.name=ActiveMQ Total Connection Count | |
+report.activemq.total.connection.count.columns=TtlConCnt | |
+report.activemq.total.connection.count.type=interfaceSnmp | |
+report.activemq.total.connection.count.command=--title="ActiveMQ Connection Count" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Connection Count" \ | |
+ DEF:val1={rrd1}:TtlConCnt:AVERAGE \ | |
+ AREA:val1#4e9a06 \ | |
+ LINE2:val1#000000:"Connection Count " \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+# | |
+# Message Count | |
+# | |
+report.activemq.message.count.name=ActiveMQ Message Size | |
+report.activemq.message.count.columns=MinMsgSize, AvgMsgSize, MaxMsgSize | |
+report.activemq.message.count.type=interfaceSnmp | |
+report.activemq.message.count.command=--title="ActiveMQ Message Size" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Message Size" \ | |
+ DEF:val1={rrd1}:MinMsgSize:AVERAGE \ | |
+ DEF:val2={rrd2}:AvgMsgSize:AVERAGE \ | |
+ DEF:val3={rrd3}:MaxMsgSize:AVERAGE \ | |
+ COMMENT:"Message Size\\n" \ | |
+ LINE2:val1#ad7fa8:"Min " \ | |
+ GPRINT:val1:MIN:"%10.2lf\\n" \ | |
+ LINE2:val2#729fcf:"Max " \ | |
+ GPRINT:val2:MAX:"%10.2lf\\n" \ | |
+ LINE2:val3#c17d11:"Avg " \ | |
+ GPRINT:val3:AVERAGE:"%10.2lf\\n" | |
+ | |
+# | |
+# Prducer-Consumer Count | |
+# | |
+report.activemq.consumer.producer.count.name=ActiveMQ Producer-Consumer Count | |
+report.activemq.consumer.producer.count.columns=TtlProdCnt,TtlConsumerCnt | |
+report.activemq.consumer.producer.count.type=interfaceSnmp | |
+report.activemq.consumer.producer.count.command=--title="ActiveMQ Producer-Consumer Count" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Producer-Consumer Count" \ | |
+ DEF:val1={rrd1}:TtlProdCnt:AVERAGE \ | |
+ DEF:val2={rrd2}:TtlConsumerCnt:AVERAGE \ | |
+ COMMENT:"Producer Consumer Count\\n" \ | |
+ LINE2:val1#ad7fa8:"Min " \ | |
+ GPRINT:val1:MIN:"%10.2lf\\n" \ | |
+ LINE2:val2#729fcf:"Max " \ | |
+ GPRINT:val2:MAX:"%10.2lf\\n" | |
+ | |
+# | |
+# Storage Limit | |
+# | |
+report.activemq.storage.limit.name=ActiveMQ Storage Limit | |
+report.activemq.storage.limit.columns=StoreLimit | |
+report.activemq.storage.limit.type=interfaceSnmp | |
+report.activemq.storage.limit.command=--title="ActiveMQ Storage Limit" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Storage Limit" \ | |
+ DEF:val1={rrd1}:StoreLimit:AVERAGE \ | |
+ AREA:val1#ce5c00 \ | |
+ LINE2:val1#000000:"Storage Limit" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+# | |
+# Percentage Usage | |
+# | |
+report.activemq.storage.percentage.name=ActiveMQ Storage Percentage Usage | |
+report.activemq.storage.percentage.columns=StorePctUsage | |
+report.activemq.storage.percentage.type=interfaceSnmp | |
+report.activemq.storage.percentage.command=--title="ActiveMQ Storage Percentage Usage" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Storage Percentage Usage" \ | |
+ DEF:val1={rrd1}:StorePctUsage:AVERAGE \ | |
+ AREA:val1#8f5902 \ | |
+ LINE2:val1#000000:"Storage Percentage Usage" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+# | |
+# Temp Storage Limit | |
+# | |
+report.activemq.temp.storage.limit.name=ActiveMQ Temp Storage Limit | |
+report.activemq.temp.storage.limit.columns=TempLimit | |
+report.activemq.temp.storage.limit.type=interfaceSnmp | |
+report.activemq.temp.storage.limit.command=--title\="ActiveMQ Temp Storage Limit" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Temp Storage Limit" \ | |
+ DEF:val1={rrd1}:TempLimit:AVERAGE \ | |
+ AREA:val1#ce5c00 \ | |
+ LINE2:val1#000000:"Temp Storage Limit" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+# | |
+# Temp Percentage Storage | |
+# | |
+report.activemq.temp.storage.percentage.name=ActiveMQ Temp Storage Percentage Used | |
+report.activemq.temp.storage.percentage.columns=TempPctUsage | |
+report.activemq.temp.storage.percentage.type=interfaceSnmp | |
+report.activemq.temp.storage.percentage.command=--title\="ActiveMQ Temp Storage Percentage Used" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Temp Storage Percentage Used" \ | |
+ DEF:val1={rrd1}:TempPctUsage:AVERAGE \ | |
+ AREA:val1#8f5902 \ | |
+ LINE2:val1#000000:"Temp Percentage Storage Used" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
diff --git a/snmp-graph.properties.d/bluecoat-sgproxy-graph.properties b/snmp-graph.properties.d/bluecoat-sgproxy-graph.properties | |
index d668ba9..f653de8 100644 | |
--- a/snmp-graph.properties.d/bluecoat-sgproxy-graph.properties | |
+++ b/snmp-graph.properties.d/bluecoat-sgproxy-graph.properties | |
@@ -69,7 +69,7 @@ report.sgProxy.workers.command=--title="Client-Server Workers" \ | |
STACK:ServerConnections#0000ff:"ServerConnections" \ | |
GPRINT:ServerConnections:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:ServerConnections:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:ServerConnections:MAX:"Max \\: %8.2lf %s\\n" | |
+ GPRINT:ServerConnections:MAX:"Max \\: %8.2lf %s\\n" | |
report.sgProxy.client.connections.name=ProxySG Client Workers | |
report.sgProxy.client.connections.columns=ClientConnections,ClientConnectionsAc,ClientConnectionsId | |
@@ -92,7 +92,7 @@ report.sgProxy.client.connections.command=--title="Client Workers" \ | |
STACK:ClientConnectionsId#ff0000:"ClientConnectionsId" \ | |
GPRINT:ClientConnectionsId:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:ClientConnectionsId:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:ClientConnectionsId:MAX:"Max \\: %8.2lf %s\\n" | |
+ GPRINT:ClientConnectionsId:MAX:"Max \\: %8.2lf %s\\n" | |
report.sgProxy.server.connections.name=ProxySG Server Workers | |
report.sgProxy.server.connections.columns=ServerConnections,ServerConnectionsAc,ServerConnectionsId | |
@@ -133,7 +133,7 @@ report.sgProxy.cpu.command=--title="CPU Usage" \ | |
STACK:CpuIdlePerCent#00ff00:"CpuIdlePerCent" \ | |
GPRINT:CpuIdlePerCent:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:CpuIdlePerCent:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:CpuIdlePerCent:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:CpuIdlePerCent:MAX:"Max \\: %8.2lf %s\\n" | |
report.sgProxy.cache.name=ProxySG Cache | |
report.sgProxy.cache.columns=ByteRateHit,ByteRatePartialHit,ByteRateMiss | |
diff --git a/snmp-graph.properties.d/cassandra-graph.properties b/snmp-graph.properties.d/cassandra-graph.properties | |
new file mode 100644 | |
index 0000000..2263b27 | |
--- /dev/null | |
+++ b/snmp-graph.properties.d/cassandra-graph.properties | |
@@ -0,0 +1,361 @@ | |
+reports=cassandra.metrics.Client, \ | |
+cassandra.metrics.Compaction.Bytes, \ | |
+cassandra.metrics.Compaction.Tasks, \ | |
+cassandra.metrics.Storage.Load, \ | |
+cassandra.metrics.Storage.Exceptions, \ | |
+cassandra.metrics.DroppedMessages, \ | |
+cassandra.metrics.ThreadPools.internal.MemtableFlushWriter, \ | |
+cassandra.metrics.ThreadPools.internal.MemtablePostFlush, \ | |
+cassandra.metrics.ThreadPools.internal.AntiEntropyStage, \ | |
+cassandra.metrics.ThreadPools.internal.GossipStage, \ | |
+cassandra.metrics.ThreadPools.internal.MigrationStage, \ | |
+cassandra.metrics.ThreadPools.internal.MiscStage, \ | |
+cassandra.metrics.ThreadPools.MutationStage, \ | |
+cassandra.metrics.ThreadPools.request.ReadStage, \ | |
+cassandra.metrics.ThreadPools.RequestResponseStage, \ | |
+cassandra.metrics.ThreadPools.ReadRepairStage | |
+ | |
+report.cassandra.metrics.Client.name=Cassandra Client Connections | |
+report.cassandra.metrics.Client.columns=clntConNativeClnts, clntConThriftClnts | |
+report.cassandra.metrics.Client.type=interfaceSnmp | |
+report.cassandra.metrics.Client.command=--title="Cassandra Client Connections" \ | |
+ --vertical-label="Clients" \ | |
+ DEF:val1={rrd1}:clntConNativeClnts:AVERAGE \ | |
+ DEF:val2={rrd2}:clntConThriftClnts:AVERAGE \ | |
+ AREA:val1#cc0000:"Connected Native Clients" \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ STACK:val2#f57900:"Connected Thrift Clients" \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.Compaction.Bytes.name=Cassandra Compaction | |
+report.cassandra.metrics.Compaction.Bytes.columns=cpctBytesCompacted | |
+report.cassandra.metrics.Compaction.Bytes.type=interfaceSnmp | |
+report.cassandra.metrics.Compaction.Bytes.command=--title="Cassandra Compaction" \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:cpctBytesCompacted:AVERAGE \ | |
+ AREA:val1#babdb6 \ | |
+ LINE1.5:val1#888a85:"Bytes Compacted" \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.Compaction.Tasks.name=Cassandra Compaction Tasks | |
+report.cassandra.metrics.Compaction.Tasks.columns=cpctPendingTasks, cpctCompletedTasks | |
+report.cassandra.metrics.Compaction.Tasks.type=interfaceSnmp | |
+report.cassandra.metrics.Compaction.Tasks.command=--title="Cassandra Compaction Tasks" \ | |
+ --vertical-label="Tasks" \ | |
+ DEF:val1={rrd1}:cpctPendingTasks:AVERAGE \ | |
+ DEF:val2={rrd2}:cpctCompletedTasks:AVERAGE \ | |
+ AREA:val1#cc0000:"Compaction Tasks Pending " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ STACK:val2#f57900:"Compaction Tasks Completed" \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.Storage.Load.name=Cassandra Storage Load | |
+report.cassandra.metrics.Storage.Load.columns=strgLoad | |
+report.cassandra.metrics.Storage.Load.type=interfaceSnmp | |
+report.cassandra.metrics.Storage.Load.command=--title="Cassandra Storage Load" \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:strgLoad:AVERAGE \ | |
+ AREA:val1#babdb6 \ | |
+ LINE1.5:val1#888a85:"Storage Load " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.Storage.Exceptions.name=Cassandra Exceptions | |
+report.cassandra.metrics.Storage.Exceptions.columns=strgExceptions | |
+report.cassandra.metrics.Storage.Exceptions.type=interfaceSnmp | |
+report.cassandra.metrics.Storage.Exceptions.command=--title="Cassandra Exceptions" \ | |
+ --vertical-label="Exceptions" \ | |
+ DEF:val1={rrd1}:strgExceptions:AVERAGE \ | |
+ LINE1.5:val1#3465a4:"Unhandled Exceptions " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.DroppedMessages.name=Cassandra Dropped Messages | |
+report.cassandra.metrics.DroppedMessages.columns=drpdMsgRead, drpdMsgReadRepair, drpdMsgReqResp, drpdMsgRangeSlice, drpdMsgMutation | |
+report.cassandra.metrics.DroppedMessages.type=interfaceSnmp | |
+report.cassandra.metrics.DroppedMessages.command=--title="Cassandra Dropped Messages" \ | |
+ --vertical-label="Dropped Messages" \ | |
+ DEF:val1={rrd1}:drpdMsgRead:AVERAGE \ | |
+ DEF:val2={rrd2}:drpdMsgReadRepair:AVERAGE \ | |
+ DEF:val3={rrd3}:drpdMsgReqResp:AVERAGE \ | |
+ DEF:val4={rrd4}:drpdMsgRangeSlice:AVERAGE \ | |
+ DEF:val5={rrd5}:drpdMsgMutation:AVERAGE \ | |
+ LINE1.5:val1#f57900:"Read " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#cc0000:"Read Repair " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val3#4e9a06:"Request Response " \ | |
+ GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val4#3465a4:"Range Slice " \ | |
+ GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val5#5c3566:"Message Mutation " \ | |
+ GPRINT:val5:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val5:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val5:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.ThreadPools.internal.MemtableFlushWriter.name=Cassandra Thread Pool Memtable Flush Writer | |
+report.cassandra.metrics.ThreadPools.internal.MemtableFlushWriter.columns=tpIntMemTblFlsWrAt, tpIntMemTblFlsWrCbt | |
+report.cassandra.metrics.ThreadPools.internal.MemtableFlushWriter.type=interfaceSnmp | |
+report.cassandra.metrics.ThreadPools.internal.MemtableFlushWriter.command=--title="Cassandra Thread Pool Memtable Flush Writer" \ | |
+ --vertical-label="Tasks" \ | |
+ DEF:val1={rrd1}:tpIntMemTblFlsWrAt:AVERAGE \ | |
+ DEF:val2={rrd2}:tpIntMemTblFlsWrCbt:AVERAGE \ | |
+ LINE1.5:val1#cc0000:"Active Tasks " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.ThreadPools.internal.MemtablePostFlush.name=Cassandra Thread Pool Memtable Post Flush Writer | |
+report.cassandra.metrics.ThreadPools.internal.MemtablePostFlush.columns=tpIntMemTblPoFlsAt, tpIntMemTblPoFlsCbt, tpIntMemTblPoFlsPt | |
+report.cassandra.metrics.ThreadPools.internal.MemtablePostFlush.type=interfaceSnmp | |
+report.cassandra.metrics.ThreadPools.internal.MemtablePostFlush.command=--title="Cassandra Thread Pool Memtable Post Flush Writer" \ | |
+ --vertical-label="Tasks" \ | |
+ DEF:val1={rrd1}:tpIntMemTblPoFlsAt:AVERAGE \ | |
+ DEF:val2={rrd2}:tpIntMemTblPoFlsCbt:AVERAGE \ | |
+ DEF:val3={rrd3}:tpIntMemTblPoFlsPt:AVERAGE \ | |
+ LINE1.5:val1#cc0000:"Active Tasks " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val3#3465a4:"Pending Tasks " \ | |
+ GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.ThreadPools.internal.AntiEntropyStage.name=Thread Pool Internal Anti-Entropy Stage | |
+report.cassandra.metrics.ThreadPools.internal.AntiEntropyStage.columns=tpIntAntiEntStgeAt, tpIntAntiEntStgeCbt, tpIntAntiEntStgePt, tpIntAntiEntStgeCt | |
+report.cassandra.metrics.ThreadPools.internal.AntiEntropyStage.type=interfaceSnmp | |
+report.cassandra.metrics.ThreadPools.internal.AntiEntropyStage.command=--title="Thread Pool Internal Anti-Entropy Stage" \ | |
+ --vertical-label="Tasks" \ | |
+ DEF:val1={rrd1}:tpIntAntiEntStgeAt:AVERAGE \ | |
+ DEF:val2={rrd2}:tpIntAntiEntStgeCbt:AVERAGE \ | |
+ DEF:val3={rrd3}:tpIntAntiEntStgePt:AVERAGE \ | |
+ DEF:val4={rrd4}:tpIntAntiEntStgeCt:AVERAGE \ | |
+ LINE1.5:val1#cc0000:"Active Tasks " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val3#3465a4:"Pending Tasks " \ | |
+ GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
+ GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.ThreadPools.internal.GossipStage.name=Thread Pool Internal Gossip Stage | |
+report.cassandra.metrics.ThreadPools.internal.GossipStage.columns=tpIntGosStgeAt, tpIntGosStgeCbt, tpIntGosStgePt, tpIntGosStgeCt | |
+report.cassandra.metrics.ThreadPools.internal.GossipStage.type=interfaceSnmp | |
+report.cassandra.metrics.ThreadPools.internal.GossipStage.command=--title="Thread Pool Internal Gossip Stage" \ | |
+ --vertical-label="Tasks" \ | |
+ DEF:val1={rrd1}:tpIntGosStgeAt:AVERAGE \ | |
+ DEF:val2={rrd2}:tpIntGosStgeCbt:AVERAGE \ | |
+ DEF:val3={rrd3}:tpIntGosStgePt:AVERAGE \ | |
+ DEF:val4={rrd4}:tpIntGosStgeCt:AVERAGE \ | |
+ LINE1.5:val1#cc0000:"Active Tasks " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val3#3465a4:"Pending Tasks " \ | |
+ GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
+ GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.ThreadPools.internal.MigrationStage.name=Thread Pool Internal Migration Stage | |
+report.cassandra.metrics.ThreadPools.internal.MigrationStage.columns=tpIntMigStgeAt, tpIntMigStgeCbt, tpIntMigStgePt, tpIntMigStgeCt | |
+report.cassandra.metrics.ThreadPools.internal.MigrationStage.type=interfaceSnmp | |
+report.cassandra.metrics.ThreadPools.internal.MigrationStage.command=--title="Thread Pool Internal Migration Stage" \ | |
+ --vertical-label="Tasks" \ | |
+ DEF:val1={rrd1}:tpIntMigStgeAt:AVERAGE \ | |
+ DEF:val2={rrd2}:tpIntMigStgeCbt:AVERAGE \ | |
+ DEF:val3={rrd3}:tpIntMigStgePt:AVERAGE \ | |
+ DEF:val4={rrd4}:tpIntMigStgeCt:AVERAGE \ | |
+ LINE1.5:val1#cc0000:"Active Tasks " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val3#3465a4:"Pending Tasks " \ | |
+ GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
+ GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.ThreadPools.internal.MiscStage.name=Thread Pool Internal Misc Stage | |
+report.cassandra.metrics.ThreadPools.internal.MiscStage.columns=tpIntMiscStgeAt, tpIntMiscStgeCbt, tpIntMiscStgePt, tpIntMiscStgeCt | |
+report.cassandra.metrics.ThreadPools.internal.MiscStage.type=interfaceSnmp | |
+report.cassandra.metrics.ThreadPools.internal.MiscStage.command=--title="Thread Pool Internal Misc Stage" \ | |
+ --vertical-label="Tasks" \ | |
+ DEF:val1={rrd1}:tpIntMiscStgeAt:AVERAGE \ | |
+ DEF:val2={rrd2}:tpIntMiscStgeCbt:AVERAGE \ | |
+ DEF:val3={rrd3}:tpIntMiscStgePt:AVERAGE \ | |
+ DEF:val4={rrd4}:tpIntMiscStgeCt:AVERAGE \ | |
+ LINE1.5:val1#cc0000:"Active Tasks " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val3#3465a4:"Pending Tasks " \ | |
+ GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
+ GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.ThreadPools.MutationStage.name=Thread Pool Mutation Stage | |
+report.cassandra.metrics.ThreadPools.MutationStage.columns=tpMutStgeAt, tpMutStgeCbt, tpMutStgePt, tpMutStgeCt | |
+report.cassandra.metrics.ThreadPools.MutationStage.type=interfaceSnmp | |
+report.cassandra.metrics.ThreadPools.MutationStage.command=--title="Thread Pool Mutation Stage" \ | |
+ --vertical-label="Tasks" \ | |
+ DEF:val1={rrd1}:tpMutStgeAt:AVERAGE \ | |
+ DEF:val2={rrd2}:tpMutStgeCbt:AVERAGE \ | |
+ DEF:val3={rrd3}:tpMutStgePt:AVERAGE \ | |
+ DEF:val4={rrd4}:tpMutStgeCt:AVERAGE \ | |
+ LINE1.5:val1#cc0000:"Active Tasks " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val3#3465a4:"Pending Tasks " \ | |
+ GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
+ GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.ThreadPools.request.ReadStage.name=Thread Pool Read Stage | |
+report.cassandra.metrics.ThreadPools.request.ReadStage.columns=tpReadStageAt, tpReadStageCbt, tpReadStagePt, tpReadStageCt | |
+report.cassandra.metrics.ThreadPools.request.ReadStage.type=interfaceSnmp | |
+report.cassandra.metrics.ThreadPools.request.ReadStage.command=--title="Thread Pool Read Stage" \ | |
+ --vertical-label="Tasks" \ | |
+ DEF:val1={rrd1}:tpReadStageAt:AVERAGE \ | |
+ DEF:val2={rrd2}:tpReadStageCbt:AVERAGE \ | |
+ DEF:val3={rrd3}:tpReadStagePt:AVERAGE \ | |
+ DEF:val4={rrd4}:tpReadStageCt:AVERAGE \ | |
+ LINE1.5:val1#cc0000:"Active Tasks " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val3#3465a4:"Pending Tasks " \ | |
+ GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
+ GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.ThreadPools.RequestResponseStage.name=Thread Pool Request Response Stage | |
+report.cassandra.metrics.ThreadPools.RequestResponseStage.columns=tpReqRespStgeAt, tpReqRespStgeCbt, tpReqRespStgePt, tpReqRespStgeCt | |
+report.cassandra.metrics.ThreadPools.RequestResponseStage.type=interfaceSnmp | |
+report.cassandra.metrics.ThreadPools.RequestResponseStage.command=--title="Thread Pool Request Response Stage" \ | |
+ --vertical-label="Tasks" \ | |
+ DEF:val1={rrd1}:tpReqRespStgeAt:AVERAGE \ | |
+ DEF:val2={rrd2}:tpReqRespStgeCbt:AVERAGE \ | |
+ DEF:val3={rrd3}:tpReqRespStgePt:AVERAGE \ | |
+ DEF:val4={rrd4}:tpReqRespStgeCt:AVERAGE \ | |
+ LINE1.5:val1#cc0000:"Active Tasks " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val3#3465a4:"Pending Tasks " \ | |
+ GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
+ GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.ThreadPools.ReadRepairStage.name=Thread Pool Read Repair Stage | |
+report.cassandra.metrics.ThreadPools.ReadRepairStage.columns=tpReadRepairStgeAt, tpReadRepairStgeCbt, tpReadRepairStgePt, tpReadRepairStgeCt | |
+report.cassandra.metrics.ThreadPools.ReadRepairStage.type=interfaceSnmp | |
+report.cassandra.metrics.ThreadPools.ReadRepairStage.command=--title="Thread Pool Read Repair Stage" \ | |
+ --vertical-label="Tasks" \ | |
+ DEF:val1={rrd1}:tpReadRepairStgeAt:AVERAGE \ | |
+ DEF:val2={rrd2}:tpReadRepairStgeCbt:AVERAGE \ | |
+ DEF:val3={rrd3}:tpReadRepairStgePt:AVERAGE \ | |
+ DEF:val4={rrd4}:tpReadRepairStgeCt:AVERAGE \ | |
+ LINE1.5:val1#cc0000:"Active Tasks " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val3#3465a4:"Pending Tasks " \ | |
+ GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
+ GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
diff --git a/snmp-graph.properties.d/cassandra-newts-graph.properties b/snmp-graph.properties.d/cassandra-newts-graph.properties | |
new file mode 100644 | |
index 0000000..427ce1a | |
--- /dev/null | |
+++ b/snmp-graph.properties.d/cassandra-newts-graph.properties | |
@@ -0,0 +1,196 @@ | |
+reports=cassandra.metrics.keyspace.newts.AllMemtables.DataSize, \ | |
+cassandra.metrics.keyspace.newts.Memtables.Switch.Counter, \ | |
+cassandra.metrics.keyspace.newts.Memtables.Columns.Counter, \ | |
+cassandra.metrics.keyspace.newts.Memtable.DataSize, \ | |
+cassandra.metrics.keyspace.newts.rwLatency, \ | |
+cassandra.metrics.keyspace.newts.RangeLatency.99th, \ | |
+cassandra.metrics.keyspace.newts.Latency, \ | |
+cassandra.metrics.keyspace.newts.Bloom.Disk, \ | |
+cassandra.metrics.keyspace.newts.Bloom.Memory, \ | |
+cassandra.metrics.keyspace.newts.MemoryUsed, \ | |
+cassandra.metrics.keyspace.newts.pending, \ | |
+cassandra.metrics.keyspace.newts.DiskSpace | |
+ | |
+report.cassandra.metrics.keyspace.newts.AllMemtables.DataSize.name=All Memtables Data Size | |
+report.cassandra.metrics.keyspace.newts.AllMemtables.DataSize.columns=alMemTblLiDaSi, alMemTblOffHeapDaSi, alMemTblOnHeapDaSi | |
+report.cassandra.metrics.keyspace.newts.AllMemtables.DataSize.type=interfaceSnmp | |
+report.cassandra.metrics.keyspace.newts.AllMemtables.DataSize.command=--title="All Memtables Data Size" \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:alMemTblLiDaSi:AVERAGE \ | |
+ DEF:val2={rrd2}:alMemTblOffHeapDaSi:AVERAGE \ | |
+ DEF:val3={rrd3}:alMemTblOnHeapDaSi:AVERAGE \ | |
+ LINE1.5:val1#3465a4:"Live Data Size " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#75507b:"Off-Heap Data Size " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val3#c17d11:"On-Heap Data Size " \ | |
+ GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.keyspace.newts.Memtables.Switch.Counter.name=All Memtables Switch Counter | |
+report.cassandra.metrics.keyspace.newts.Memtables.Switch.Counter.columns=memTblSwitchCount | |
+report.cassandra.metrics.keyspace.newts.Memtables.Switch.Counter.type=interfaceSnmp | |
+report.cassandra.metrics.keyspace.newts.Memtables.Switch.Counter.command=--title="All Memtables Switch Counter" \ | |
+ --vertical-label="Number of Times" \ | |
+ DEF:val1={rrd1}:memTblSwitchCount:AVERAGE \ | |
+ LINE1.5:val1#3465a4:"Switch Counter " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.keyspace.newts.Memtables.Columns.Counter.name=All Memtables Columns Counter | |
+report.cassandra.metrics.keyspace.newts.Memtables.Columns.Counter.columns=memTblColumnsCnt | |
+report.cassandra.metrics.keyspace.newts.Memtables.Columns.Counter.type=interfaceSnmp | |
+report.cassandra.metrics.keyspace.newts.Memtables.Columns.Counter.command=--title="All Memtables Columns Counter" \ | |
+ --vertical-label="Columns" \ | |
+ DEF:val1={rrd1}:memTblColumnsCnt:AVERAGE \ | |
+ LINE1.5:val1#3465a4:"Columns " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.keyspace.newts.Memtable.DataSize.name=Newts Memtable Data Size | |
+report.cassandra.metrics.keyspace.newts.Memtable.DataSize.columns=memTblLiveDaSi, memTblOffHeapDaSi, memTblOnHeapDaSi | |
+report.cassandra.metrics.keyspace.newts.Memtable.DataSize.type=interfaceSnmp | |
+report.cassandra.metrics.keyspace.newts.Memtable.DataSize.command=--title="Newts Memtable Data Size" \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:memTblLiveDaSi:AVERAGE \ | |
+ DEF:val2={rrd2}:memTblOffHeapDaSi:AVERAGE \ | |
+ DEF:val3={rrd3}:memTblOnHeapDaSi:AVERAGE \ | |
+ LINE1.5:val1#3465a4:"Live Data Size " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#75507b:"Off-Heap Data Size " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#c17d11:"On-Heap Data Size " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.keyspace.newts.rwLatency.name=Newts Read and Write Latency | |
+report.cassandra.metrics.keyspace.newts.rwLatency.columns=readTotLtncy, writeTotLtncy | |
+report.cassandra.metrics.keyspace.newts.rwLatency.type=interfaceSnmp | |
+report.cassandra.metrics.keyspace.newts.rwLatency.command=--title="Newts Read and Write Latency" \ | |
+ --vertical-label="micro seconds" \ | |
+ DEF:val1={rrd1}:readTotLtncy:AVERAGE \ | |
+ DEF:val2={rrd2}:writeTotLtncy:AVERAGE \ | |
+ LINE1.5:val1#73d216:"Read Total Latency " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#3465a4:"Write Total Latency " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.keyspace.newts.RangeLatency.99th.name=Newts Range Latency 99th Percentile | |
+report.cassandra.metrics.keyspace.newts.RangeLatency.99th.columns=rangeLtncy99 | |
+report.cassandra.metrics.keyspace.newts.RangeLatency.99th.type=interfaceSnmp | |
+report.cassandra.metrics.keyspace.newts.RangeLatency.99th.command=--title="Newts Range Latency 99th Percentile" \ | |
+ --vertical-label="micro seconds" \ | |
+ DEF:val1={rrd1}:rangeLtncy99:AVERAGE \ | |
+ LINE1.5:val1#3465a4:"Range Latency 99 Percentile " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.keyspace.newts.Latency.name=Newts Latency | |
+report.cassandra.metrics.keyspace.newts.Latency.columns=casCommitTotLtncy, casPrepareTotLtncy, casProposeTotLtncy | |
+report.cassandra.metrics.keyspace.newts.Latency.type=interfaceSnmp | |
+report.cassandra.metrics.keyspace.newts.Latency.command=--title="Newts Latency" \ | |
+ --vertical-label="micro seconds" \ | |
+ DEF:val1={rrd1}:casCommitTotLtncy:AVERAGE \ | |
+ DEF:val2={rrd2}:casPrepareTotLtncy:AVERAGE \ | |
+ DEF:val3={rrd3}:casProposeTotLtncy:AVERAGE \ | |
+ LINE1.5:val1#f57900:"Commit Total Latency " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ STACK:val2#3465a4:"Preprare Total Latency " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ STACK:val3#75507b:"Propose Total Latency " \ | |
+ GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.keyspace.newts.Bloom.Disk.name=Bloom Filter Disk Usage | |
+report.cassandra.metrics.keyspace.newts.Bloom.Disk.columns=blmFltrDskSpcUsed | |
+report.cassandra.metrics.keyspace.newts.Bloom.Disk.type=interfaceSnmp | |
+report.cassandra.metrics.keyspace.newts.Bloom.Disk.command=--title="Bloom Filter Disk Usage" \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:blmFltrDskSpcUsed:AVERAGE \ | |
+ AREA:val1#babdb6 \ | |
+ LINE1.5:val1#888a85:"Disk Space Used " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.keyspace.newts.Bloom.Memory.name=Bloom Filter Memory Usage | |
+report.cassandra.metrics.keyspace.newts.Bloom.Memory.columns=blmFltrOffHeapMemUs | |
+report.cassandra.metrics.keyspace.newts.Bloom.Memory.type=interfaceSnmp | |
+report.cassandra.metrics.keyspace.newts.Bloom.Memory.command=--title="Bloom Filter Memory Usage" \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:blmFltrOffHeapMemUs:AVERAGE \ | |
+ AREA:val1#babdb6 \ | |
+ LINE1.5:val1#888a85:"Off-Heap Memory Used " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.keyspace.newts.MemoryUsed.name=Newts Memory Used | |
+report.cassandra.metrics.keyspace.newts.MemoryUsed.columns=cmpMetaOffHeapMemUs, idxSumOffHeapMemUs | |
+report.cassandra.metrics.keyspace.newts.MemoryUsed.type=interfaceSnmp | |
+report.cassandra.metrics.keyspace.newts.MemoryUsed.command=--title="Newts Memory Used" \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:cmpMetaOffHeapMemUs:AVERAGE \ | |
+ DEF:val2={rrd2}:idxSumOffHeapMemUs:AVERAGE \ | |
+ LINE1.5:val1#f57900:"Compression Metadata Off-Heap Memory Used " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#3465a4:"Index Summary Off-Heap Memory Used " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.keyspace.newts.pending.name=Newts Pending | |
+report.cassandra.metrics.keyspace.newts.pending.columns=pendingCompactions, pendingFlushes | |
+report.cassandra.metrics.keyspace.newts.pending.type=interfaceSnmp | |
+report.cassandra.metrics.keyspace.newts.pending.command=--title="Newts Pending" \ | |
+ --vertical-label="Tasks" \ | |
+ DEF:val1={rrd1}:pendingCompactions:AVERAGE \ | |
+ DEF:val2={rrd2}:pendingFlushes:AVERAGE \ | |
+ LINE1.5:val1#f57900:"Pending Compactions " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#3465a4:"Pending Flushes " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.cassandra.metrics.keyspace.newts.DiskSpace.name=Newts Disk Space | |
+report.cassandra.metrics.keyspace.newts.DiskSpace.columns=totalDiskSpaceUsed, liveDiskSpaceUsed | |
+report.cassandra.metrics.keyspace.newts.DiskSpace.type=interfaceSnmp | |
+report.cassandra.metrics.keyspace.newts.DiskSpace.command=--title="Newts Disk Space" \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:totalDiskSpaceUsed:AVERAGE \ | |
+ DEF:val2={rrd2}:liveDiskSpaceUsed:AVERAGE \ | |
+ LINE1.5:val1#f57900:"Total Disk Space Used " \ | |
+ GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ LINE1.5:val2#3465a4:"Live Disk Space Used " \ | |
+ GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
diff --git a/snmp-graph.properties.d/cassandra21x-graph.properties b/snmp-graph.properties.d/cassandra21x-graph.properties | |
deleted file mode 100644 | |
index 2263b27..0000000 | |
--- a/snmp-graph.properties.d/cassandra21x-graph.properties | |
+++ /dev/null | |
@@ -1,361 +0,0 @@ | |
-reports=cassandra.metrics.Client, \ | |
-cassandra.metrics.Compaction.Bytes, \ | |
-cassandra.metrics.Compaction.Tasks, \ | |
-cassandra.metrics.Storage.Load, \ | |
-cassandra.metrics.Storage.Exceptions, \ | |
-cassandra.metrics.DroppedMessages, \ | |
-cassandra.metrics.ThreadPools.internal.MemtableFlushWriter, \ | |
-cassandra.metrics.ThreadPools.internal.MemtablePostFlush, \ | |
-cassandra.metrics.ThreadPools.internal.AntiEntropyStage, \ | |
-cassandra.metrics.ThreadPools.internal.GossipStage, \ | |
-cassandra.metrics.ThreadPools.internal.MigrationStage, \ | |
-cassandra.metrics.ThreadPools.internal.MiscStage, \ | |
-cassandra.metrics.ThreadPools.MutationStage, \ | |
-cassandra.metrics.ThreadPools.request.ReadStage, \ | |
-cassandra.metrics.ThreadPools.RequestResponseStage, \ | |
-cassandra.metrics.ThreadPools.ReadRepairStage | |
- | |
-report.cassandra.metrics.Client.name=Cassandra Client Connections | |
-report.cassandra.metrics.Client.columns=clntConNativeClnts, clntConThriftClnts | |
-report.cassandra.metrics.Client.type=interfaceSnmp | |
-report.cassandra.metrics.Client.command=--title="Cassandra Client Connections" \ | |
- --vertical-label="Clients" \ | |
- DEF:val1={rrd1}:clntConNativeClnts:AVERAGE \ | |
- DEF:val2={rrd2}:clntConThriftClnts:AVERAGE \ | |
- AREA:val1#cc0000:"Connected Native Clients" \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- STACK:val2#f57900:"Connected Thrift Clients" \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.Compaction.Bytes.name=Cassandra Compaction | |
-report.cassandra.metrics.Compaction.Bytes.columns=cpctBytesCompacted | |
-report.cassandra.metrics.Compaction.Bytes.type=interfaceSnmp | |
-report.cassandra.metrics.Compaction.Bytes.command=--title="Cassandra Compaction" \ | |
- --vertical-label="Bytes" \ | |
- DEF:val1={rrd1}:cpctBytesCompacted:AVERAGE \ | |
- AREA:val1#babdb6 \ | |
- LINE1.5:val1#888a85:"Bytes Compacted" \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.Compaction.Tasks.name=Cassandra Compaction Tasks | |
-report.cassandra.metrics.Compaction.Tasks.columns=cpctPendingTasks, cpctCompletedTasks | |
-report.cassandra.metrics.Compaction.Tasks.type=interfaceSnmp | |
-report.cassandra.metrics.Compaction.Tasks.command=--title="Cassandra Compaction Tasks" \ | |
- --vertical-label="Tasks" \ | |
- DEF:val1={rrd1}:cpctPendingTasks:AVERAGE \ | |
- DEF:val2={rrd2}:cpctCompletedTasks:AVERAGE \ | |
- AREA:val1#cc0000:"Compaction Tasks Pending " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- STACK:val2#f57900:"Compaction Tasks Completed" \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.Storage.Load.name=Cassandra Storage Load | |
-report.cassandra.metrics.Storage.Load.columns=strgLoad | |
-report.cassandra.metrics.Storage.Load.type=interfaceSnmp | |
-report.cassandra.metrics.Storage.Load.command=--title="Cassandra Storage Load" \ | |
- --vertical-label="Bytes" \ | |
- DEF:val1={rrd1}:strgLoad:AVERAGE \ | |
- AREA:val1#babdb6 \ | |
- LINE1.5:val1#888a85:"Storage Load " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.Storage.Exceptions.name=Cassandra Exceptions | |
-report.cassandra.metrics.Storage.Exceptions.columns=strgExceptions | |
-report.cassandra.metrics.Storage.Exceptions.type=interfaceSnmp | |
-report.cassandra.metrics.Storage.Exceptions.command=--title="Cassandra Exceptions" \ | |
- --vertical-label="Exceptions" \ | |
- DEF:val1={rrd1}:strgExceptions:AVERAGE \ | |
- LINE1.5:val1#3465a4:"Unhandled Exceptions " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.DroppedMessages.name=Cassandra Dropped Messages | |
-report.cassandra.metrics.DroppedMessages.columns=drpdMsgRead, drpdMsgReadRepair, drpdMsgReqResp, drpdMsgRangeSlice, drpdMsgMutation | |
-report.cassandra.metrics.DroppedMessages.type=interfaceSnmp | |
-report.cassandra.metrics.DroppedMessages.command=--title="Cassandra Dropped Messages" \ | |
- --vertical-label="Dropped Messages" \ | |
- DEF:val1={rrd1}:drpdMsgRead:AVERAGE \ | |
- DEF:val2={rrd2}:drpdMsgReadRepair:AVERAGE \ | |
- DEF:val3={rrd3}:drpdMsgReqResp:AVERAGE \ | |
- DEF:val4={rrd4}:drpdMsgRangeSlice:AVERAGE \ | |
- DEF:val5={rrd5}:drpdMsgMutation:AVERAGE \ | |
- LINE1.5:val1#f57900:"Read " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#cc0000:"Read Repair " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val3#4e9a06:"Request Response " \ | |
- GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val4#3465a4:"Range Slice " \ | |
- GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val5#5c3566:"Message Mutation " \ | |
- GPRINT:val5:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val5:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val5:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.ThreadPools.internal.MemtableFlushWriter.name=Cassandra Thread Pool Memtable Flush Writer | |
-report.cassandra.metrics.ThreadPools.internal.MemtableFlushWriter.columns=tpIntMemTblFlsWrAt, tpIntMemTblFlsWrCbt | |
-report.cassandra.metrics.ThreadPools.internal.MemtableFlushWriter.type=interfaceSnmp | |
-report.cassandra.metrics.ThreadPools.internal.MemtableFlushWriter.command=--title="Cassandra Thread Pool Memtable Flush Writer" \ | |
- --vertical-label="Tasks" \ | |
- DEF:val1={rrd1}:tpIntMemTblFlsWrAt:AVERAGE \ | |
- DEF:val2={rrd2}:tpIntMemTblFlsWrCbt:AVERAGE \ | |
- LINE1.5:val1#cc0000:"Active Tasks " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.ThreadPools.internal.MemtablePostFlush.name=Cassandra Thread Pool Memtable Post Flush Writer | |
-report.cassandra.metrics.ThreadPools.internal.MemtablePostFlush.columns=tpIntMemTblPoFlsAt, tpIntMemTblPoFlsCbt, tpIntMemTblPoFlsPt | |
-report.cassandra.metrics.ThreadPools.internal.MemtablePostFlush.type=interfaceSnmp | |
-report.cassandra.metrics.ThreadPools.internal.MemtablePostFlush.command=--title="Cassandra Thread Pool Memtable Post Flush Writer" \ | |
- --vertical-label="Tasks" \ | |
- DEF:val1={rrd1}:tpIntMemTblPoFlsAt:AVERAGE \ | |
- DEF:val2={rrd2}:tpIntMemTblPoFlsCbt:AVERAGE \ | |
- DEF:val3={rrd3}:tpIntMemTblPoFlsPt:AVERAGE \ | |
- LINE1.5:val1#cc0000:"Active Tasks " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val3#3465a4:"Pending Tasks " \ | |
- GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.ThreadPools.internal.AntiEntropyStage.name=Thread Pool Internal Anti-Entropy Stage | |
-report.cassandra.metrics.ThreadPools.internal.AntiEntropyStage.columns=tpIntAntiEntStgeAt, tpIntAntiEntStgeCbt, tpIntAntiEntStgePt, tpIntAntiEntStgeCt | |
-report.cassandra.metrics.ThreadPools.internal.AntiEntropyStage.type=interfaceSnmp | |
-report.cassandra.metrics.ThreadPools.internal.AntiEntropyStage.command=--title="Thread Pool Internal Anti-Entropy Stage" \ | |
- --vertical-label="Tasks" \ | |
- DEF:val1={rrd1}:tpIntAntiEntStgeAt:AVERAGE \ | |
- DEF:val2={rrd2}:tpIntAntiEntStgeCbt:AVERAGE \ | |
- DEF:val3={rrd3}:tpIntAntiEntStgePt:AVERAGE \ | |
- DEF:val4={rrd4}:tpIntAntiEntStgeCt:AVERAGE \ | |
- LINE1.5:val1#cc0000:"Active Tasks " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val3#3465a4:"Pending Tasks " \ | |
- GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
- GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.ThreadPools.internal.GossipStage.name=Thread Pool Internal Gossip Stage | |
-report.cassandra.metrics.ThreadPools.internal.GossipStage.columns=tpIntGosStgeAt, tpIntGosStgeCbt, tpIntGosStgePt, tpIntGosStgeCt | |
-report.cassandra.metrics.ThreadPools.internal.GossipStage.type=interfaceSnmp | |
-report.cassandra.metrics.ThreadPools.internal.GossipStage.command=--title="Thread Pool Internal Gossip Stage" \ | |
- --vertical-label="Tasks" \ | |
- DEF:val1={rrd1}:tpIntGosStgeAt:AVERAGE \ | |
- DEF:val2={rrd2}:tpIntGosStgeCbt:AVERAGE \ | |
- DEF:val3={rrd3}:tpIntGosStgePt:AVERAGE \ | |
- DEF:val4={rrd4}:tpIntGosStgeCt:AVERAGE \ | |
- LINE1.5:val1#cc0000:"Active Tasks " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val3#3465a4:"Pending Tasks " \ | |
- GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
- GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.ThreadPools.internal.MigrationStage.name=Thread Pool Internal Migration Stage | |
-report.cassandra.metrics.ThreadPools.internal.MigrationStage.columns=tpIntMigStgeAt, tpIntMigStgeCbt, tpIntMigStgePt, tpIntMigStgeCt | |
-report.cassandra.metrics.ThreadPools.internal.MigrationStage.type=interfaceSnmp | |
-report.cassandra.metrics.ThreadPools.internal.MigrationStage.command=--title="Thread Pool Internal Migration Stage" \ | |
- --vertical-label="Tasks" \ | |
- DEF:val1={rrd1}:tpIntMigStgeAt:AVERAGE \ | |
- DEF:val2={rrd2}:tpIntMigStgeCbt:AVERAGE \ | |
- DEF:val3={rrd3}:tpIntMigStgePt:AVERAGE \ | |
- DEF:val4={rrd4}:tpIntMigStgeCt:AVERAGE \ | |
- LINE1.5:val1#cc0000:"Active Tasks " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val3#3465a4:"Pending Tasks " \ | |
- GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
- GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.ThreadPools.internal.MiscStage.name=Thread Pool Internal Misc Stage | |
-report.cassandra.metrics.ThreadPools.internal.MiscStage.columns=tpIntMiscStgeAt, tpIntMiscStgeCbt, tpIntMiscStgePt, tpIntMiscStgeCt | |
-report.cassandra.metrics.ThreadPools.internal.MiscStage.type=interfaceSnmp | |
-report.cassandra.metrics.ThreadPools.internal.MiscStage.command=--title="Thread Pool Internal Misc Stage" \ | |
- --vertical-label="Tasks" \ | |
- DEF:val1={rrd1}:tpIntMiscStgeAt:AVERAGE \ | |
- DEF:val2={rrd2}:tpIntMiscStgeCbt:AVERAGE \ | |
- DEF:val3={rrd3}:tpIntMiscStgePt:AVERAGE \ | |
- DEF:val4={rrd4}:tpIntMiscStgeCt:AVERAGE \ | |
- LINE1.5:val1#cc0000:"Active Tasks " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val3#3465a4:"Pending Tasks " \ | |
- GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
- GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.ThreadPools.MutationStage.name=Thread Pool Mutation Stage | |
-report.cassandra.metrics.ThreadPools.MutationStage.columns=tpMutStgeAt, tpMutStgeCbt, tpMutStgePt, tpMutStgeCt | |
-report.cassandra.metrics.ThreadPools.MutationStage.type=interfaceSnmp | |
-report.cassandra.metrics.ThreadPools.MutationStage.command=--title="Thread Pool Mutation Stage" \ | |
- --vertical-label="Tasks" \ | |
- DEF:val1={rrd1}:tpMutStgeAt:AVERAGE \ | |
- DEF:val2={rrd2}:tpMutStgeCbt:AVERAGE \ | |
- DEF:val3={rrd3}:tpMutStgePt:AVERAGE \ | |
- DEF:val4={rrd4}:tpMutStgeCt:AVERAGE \ | |
- LINE1.5:val1#cc0000:"Active Tasks " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val3#3465a4:"Pending Tasks " \ | |
- GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
- GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.ThreadPools.request.ReadStage.name=Thread Pool Read Stage | |
-report.cassandra.metrics.ThreadPools.request.ReadStage.columns=tpReadStageAt, tpReadStageCbt, tpReadStagePt, tpReadStageCt | |
-report.cassandra.metrics.ThreadPools.request.ReadStage.type=interfaceSnmp | |
-report.cassandra.metrics.ThreadPools.request.ReadStage.command=--title="Thread Pool Read Stage" \ | |
- --vertical-label="Tasks" \ | |
- DEF:val1={rrd1}:tpReadStageAt:AVERAGE \ | |
- DEF:val2={rrd2}:tpReadStageCbt:AVERAGE \ | |
- DEF:val3={rrd3}:tpReadStagePt:AVERAGE \ | |
- DEF:val4={rrd4}:tpReadStageCt:AVERAGE \ | |
- LINE1.5:val1#cc0000:"Active Tasks " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val3#3465a4:"Pending Tasks " \ | |
- GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
- GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.ThreadPools.RequestResponseStage.name=Thread Pool Request Response Stage | |
-report.cassandra.metrics.ThreadPools.RequestResponseStage.columns=tpReqRespStgeAt, tpReqRespStgeCbt, tpReqRespStgePt, tpReqRespStgeCt | |
-report.cassandra.metrics.ThreadPools.RequestResponseStage.type=interfaceSnmp | |
-report.cassandra.metrics.ThreadPools.RequestResponseStage.command=--title="Thread Pool Request Response Stage" \ | |
- --vertical-label="Tasks" \ | |
- DEF:val1={rrd1}:tpReqRespStgeAt:AVERAGE \ | |
- DEF:val2={rrd2}:tpReqRespStgeCbt:AVERAGE \ | |
- DEF:val3={rrd3}:tpReqRespStgePt:AVERAGE \ | |
- DEF:val4={rrd4}:tpReqRespStgeCt:AVERAGE \ | |
- LINE1.5:val1#cc0000:"Active Tasks " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val3#3465a4:"Pending Tasks " \ | |
- GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
- GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.ThreadPools.ReadRepairStage.name=Thread Pool Read Repair Stage | |
-report.cassandra.metrics.ThreadPools.ReadRepairStage.columns=tpReadRepairStgeAt, tpReadRepairStgeCbt, tpReadRepairStgePt, tpReadRepairStgeCt | |
-report.cassandra.metrics.ThreadPools.ReadRepairStage.type=interfaceSnmp | |
-report.cassandra.metrics.ThreadPools.ReadRepairStage.command=--title="Thread Pool Read Repair Stage" \ | |
- --vertical-label="Tasks" \ | |
- DEF:val1={rrd1}:tpReadRepairStgeAt:AVERAGE \ | |
- DEF:val2={rrd2}:tpReadRepairStgeCbt:AVERAGE \ | |
- DEF:val3={rrd3}:tpReadRepairStgePt:AVERAGE \ | |
- DEF:val4={rrd4}:tpReadRepairStgeCt:AVERAGE \ | |
- LINE1.5:val1#cc0000:"Active Tasks " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#f57900:"Currently Blocked Tasks " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val3#3465a4:"Pending Tasks " \ | |
- GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val4#4e9a06:"Completed Tasks " \ | |
- GPRINT:val4:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val4:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val4:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
- | |
diff --git a/snmp-graph.properties.d/cassandra21x-newts-graph.properties b/snmp-graph.properties.d/cassandra21x-newts-graph.properties | |
deleted file mode 100644 | |
index 427ce1a..0000000 | |
--- a/snmp-graph.properties.d/cassandra21x-newts-graph.properties | |
+++ /dev/null | |
@@ -1,196 +0,0 @@ | |
-reports=cassandra.metrics.keyspace.newts.AllMemtables.DataSize, \ | |
-cassandra.metrics.keyspace.newts.Memtables.Switch.Counter, \ | |
-cassandra.metrics.keyspace.newts.Memtables.Columns.Counter, \ | |
-cassandra.metrics.keyspace.newts.Memtable.DataSize, \ | |
-cassandra.metrics.keyspace.newts.rwLatency, \ | |
-cassandra.metrics.keyspace.newts.RangeLatency.99th, \ | |
-cassandra.metrics.keyspace.newts.Latency, \ | |
-cassandra.metrics.keyspace.newts.Bloom.Disk, \ | |
-cassandra.metrics.keyspace.newts.Bloom.Memory, \ | |
-cassandra.metrics.keyspace.newts.MemoryUsed, \ | |
-cassandra.metrics.keyspace.newts.pending, \ | |
-cassandra.metrics.keyspace.newts.DiskSpace | |
- | |
-report.cassandra.metrics.keyspace.newts.AllMemtables.DataSize.name=All Memtables Data Size | |
-report.cassandra.metrics.keyspace.newts.AllMemtables.DataSize.columns=alMemTblLiDaSi, alMemTblOffHeapDaSi, alMemTblOnHeapDaSi | |
-report.cassandra.metrics.keyspace.newts.AllMemtables.DataSize.type=interfaceSnmp | |
-report.cassandra.metrics.keyspace.newts.AllMemtables.DataSize.command=--title="All Memtables Data Size" \ | |
- --vertical-label="Bytes" \ | |
- DEF:val1={rrd1}:alMemTblLiDaSi:AVERAGE \ | |
- DEF:val2={rrd2}:alMemTblOffHeapDaSi:AVERAGE \ | |
- DEF:val3={rrd3}:alMemTblOnHeapDaSi:AVERAGE \ | |
- LINE1.5:val1#3465a4:"Live Data Size " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#75507b:"Off-Heap Data Size " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val3#c17d11:"On-Heap Data Size " \ | |
- GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.keyspace.newts.Memtables.Switch.Counter.name=All Memtables Switch Counter | |
-report.cassandra.metrics.keyspace.newts.Memtables.Switch.Counter.columns=memTblSwitchCount | |
-report.cassandra.metrics.keyspace.newts.Memtables.Switch.Counter.type=interfaceSnmp | |
-report.cassandra.metrics.keyspace.newts.Memtables.Switch.Counter.command=--title="All Memtables Switch Counter" \ | |
- --vertical-label="Number of Times" \ | |
- DEF:val1={rrd1}:memTblSwitchCount:AVERAGE \ | |
- LINE1.5:val1#3465a4:"Switch Counter " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.keyspace.newts.Memtables.Columns.Counter.name=All Memtables Columns Counter | |
-report.cassandra.metrics.keyspace.newts.Memtables.Columns.Counter.columns=memTblColumnsCnt | |
-report.cassandra.metrics.keyspace.newts.Memtables.Columns.Counter.type=interfaceSnmp | |
-report.cassandra.metrics.keyspace.newts.Memtables.Columns.Counter.command=--title="All Memtables Columns Counter" \ | |
- --vertical-label="Columns" \ | |
- DEF:val1={rrd1}:memTblColumnsCnt:AVERAGE \ | |
- LINE1.5:val1#3465a4:"Columns " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.keyspace.newts.Memtable.DataSize.name=Newts Memtable Data Size | |
-report.cassandra.metrics.keyspace.newts.Memtable.DataSize.columns=memTblLiveDaSi, memTblOffHeapDaSi, memTblOnHeapDaSi | |
-report.cassandra.metrics.keyspace.newts.Memtable.DataSize.type=interfaceSnmp | |
-report.cassandra.metrics.keyspace.newts.Memtable.DataSize.command=--title="Newts Memtable Data Size" \ | |
- --vertical-label="Bytes" \ | |
- DEF:val1={rrd1}:memTblLiveDaSi:AVERAGE \ | |
- DEF:val2={rrd2}:memTblOffHeapDaSi:AVERAGE \ | |
- DEF:val3={rrd3}:memTblOnHeapDaSi:AVERAGE \ | |
- LINE1.5:val1#3465a4:"Live Data Size " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#75507b:"Off-Heap Data Size " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#c17d11:"On-Heap Data Size " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.keyspace.newts.rwLatency.name=Newts Read and Write Latency | |
-report.cassandra.metrics.keyspace.newts.rwLatency.columns=readTotLtncy, writeTotLtncy | |
-report.cassandra.metrics.keyspace.newts.rwLatency.type=interfaceSnmp | |
-report.cassandra.metrics.keyspace.newts.rwLatency.command=--title="Newts Read and Write Latency" \ | |
- --vertical-label="micro seconds" \ | |
- DEF:val1={rrd1}:readTotLtncy:AVERAGE \ | |
- DEF:val2={rrd2}:writeTotLtncy:AVERAGE \ | |
- LINE1.5:val1#73d216:"Read Total Latency " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#3465a4:"Write Total Latency " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.keyspace.newts.RangeLatency.99th.name=Newts Range Latency 99th Percentile | |
-report.cassandra.metrics.keyspace.newts.RangeLatency.99th.columns=rangeLtncy99 | |
-report.cassandra.metrics.keyspace.newts.RangeLatency.99th.type=interfaceSnmp | |
-report.cassandra.metrics.keyspace.newts.RangeLatency.99th.command=--title="Newts Range Latency 99th Percentile" \ | |
- --vertical-label="micro seconds" \ | |
- DEF:val1={rrd1}:rangeLtncy99:AVERAGE \ | |
- LINE1.5:val1#3465a4:"Range Latency 99 Percentile " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.keyspace.newts.Latency.name=Newts Latency | |
-report.cassandra.metrics.keyspace.newts.Latency.columns=casCommitTotLtncy, casPrepareTotLtncy, casProposeTotLtncy | |
-report.cassandra.metrics.keyspace.newts.Latency.type=interfaceSnmp | |
-report.cassandra.metrics.keyspace.newts.Latency.command=--title="Newts Latency" \ | |
- --vertical-label="micro seconds" \ | |
- DEF:val1={rrd1}:casCommitTotLtncy:AVERAGE \ | |
- DEF:val2={rrd2}:casPrepareTotLtncy:AVERAGE \ | |
- DEF:val3={rrd3}:casProposeTotLtncy:AVERAGE \ | |
- LINE1.5:val1#f57900:"Commit Total Latency " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- STACK:val2#3465a4:"Preprare Total Latency " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" \ | |
- STACK:val3#75507b:"Propose Total Latency " \ | |
- GPRINT:val3:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val3:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val3:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.keyspace.newts.Bloom.Disk.name=Bloom Filter Disk Usage | |
-report.cassandra.metrics.keyspace.newts.Bloom.Disk.columns=blmFltrDskSpcUsed | |
-report.cassandra.metrics.keyspace.newts.Bloom.Disk.type=interfaceSnmp | |
-report.cassandra.metrics.keyspace.newts.Bloom.Disk.command=--title="Bloom Filter Disk Usage" \ | |
- --vertical-label="Bytes" \ | |
- DEF:val1={rrd1}:blmFltrDskSpcUsed:AVERAGE \ | |
- AREA:val1#babdb6 \ | |
- LINE1.5:val1#888a85:"Disk Space Used " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.keyspace.newts.Bloom.Memory.name=Bloom Filter Memory Usage | |
-report.cassandra.metrics.keyspace.newts.Bloom.Memory.columns=blmFltrOffHeapMemUs | |
-report.cassandra.metrics.keyspace.newts.Bloom.Memory.type=interfaceSnmp | |
-report.cassandra.metrics.keyspace.newts.Bloom.Memory.command=--title="Bloom Filter Memory Usage" \ | |
- --vertical-label="Bytes" \ | |
- DEF:val1={rrd1}:blmFltrOffHeapMemUs:AVERAGE \ | |
- AREA:val1#babdb6 \ | |
- LINE1.5:val1#888a85:"Off-Heap Memory Used " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.keyspace.newts.MemoryUsed.name=Newts Memory Used | |
-report.cassandra.metrics.keyspace.newts.MemoryUsed.columns=cmpMetaOffHeapMemUs, idxSumOffHeapMemUs | |
-report.cassandra.metrics.keyspace.newts.MemoryUsed.type=interfaceSnmp | |
-report.cassandra.metrics.keyspace.newts.MemoryUsed.command=--title="Newts Memory Used" \ | |
- --vertical-label="Bytes" \ | |
- DEF:val1={rrd1}:cmpMetaOffHeapMemUs:AVERAGE \ | |
- DEF:val2={rrd2}:idxSumOffHeapMemUs:AVERAGE \ | |
- LINE1.5:val1#f57900:"Compression Metadata Off-Heap Memory Used " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#3465a4:"Index Summary Off-Heap Memory Used " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.keyspace.newts.pending.name=Newts Pending | |
-report.cassandra.metrics.keyspace.newts.pending.columns=pendingCompactions, pendingFlushes | |
-report.cassandra.metrics.keyspace.newts.pending.type=interfaceSnmp | |
-report.cassandra.metrics.keyspace.newts.pending.command=--title="Newts Pending" \ | |
- --vertical-label="Tasks" \ | |
- DEF:val1={rrd1}:pendingCompactions:AVERAGE \ | |
- DEF:val2={rrd2}:pendingFlushes:AVERAGE \ | |
- LINE1.5:val1#f57900:"Pending Compactions " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#3465a4:"Pending Flushes " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
- | |
-report.cassandra.metrics.keyspace.newts.DiskSpace.name=Newts Disk Space | |
-report.cassandra.metrics.keyspace.newts.DiskSpace.columns=totalDiskSpaceUsed, liveDiskSpaceUsed | |
-report.cassandra.metrics.keyspace.newts.DiskSpace.type=interfaceSnmp | |
-report.cassandra.metrics.keyspace.newts.DiskSpace.command=--title="Newts Disk Space" \ | |
- --vertical-label="Bytes" \ | |
- DEF:val1={rrd1}:totalDiskSpaceUsed:AVERAGE \ | |
- DEF:val2={rrd2}:liveDiskSpaceUsed:AVERAGE \ | |
- LINE1.5:val1#f57900:"Total Disk Space Used " \ | |
- GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val1:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val1:MAX:" Max \\: %8.2lf %s\\n" \ | |
- LINE1.5:val2#3465a4:"Live Disk Space Used " \ | |
- GPRINT:val2:AVERAGE:" Avg \\: %8.2lf %s" \ | |
- GPRINT:val2:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:val2:MAX:" Max \\: %8.2lf %s\\n" | |
diff --git a/snmp-graph.properties.d/checkpoint-graph.properties b/snmp-graph.properties.d/checkpoint-graph.properties | |
index 985a0a9..522f82d 100644 | |
--- a/snmp-graph.properties.d/checkpoint-graph.properties | |
+++ b/snmp-graph.properties.d/checkpoint-graph.properties | |
@@ -38,7 +38,7 @@ report.checkpoint.pktsAccepted.command=--title="Packets Accepted" \ | |
LINE2:accepted#0000ff:"Accepted" \ | |
GPRINT:accepted:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:accepted:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:accepted:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:accepted:MAX:"Max \\: %8.2lf %s\\n" | |
report.checkpoint.pktsDropped.name=Packets Dropped (CheckPoint) | |
report.checkpoint.pktsDropped.columns=pktsDropped | |
@@ -49,7 +49,7 @@ report.checkpoint.pktsDropped.command=--title="Packets Dropped" \ | |
LINE2:dropped#0000ff:"Dropped" \ | |
GPRINT:dropped:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:dropped:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:dropped:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:dropped:MAX:"Max \\: %8.2lf %s\\n" | |
report.checkpoint.pktsLogged.name=Packets Logged (CheckPoint) | |
report.checkpoint.pktsLogged.columns=pktsLogged | |
@@ -60,7 +60,7 @@ report.checkpoint.pktsLogged.command=--title="Packets Logged" \ | |
LINE2:logged#0000ff:"Logged" \ | |
GPRINT:logged:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:logged:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:logged:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:logged:MAX:"Max \\: %8.2lf %s\\n" | |
report.checkpoint.pktsRejected.name=Packets Rejected (CheckPoint) | |
report.checkpoint.pktsRejected.columns=pktsRejected | |
@@ -71,7 +71,7 @@ report.checkpoint.pktsRejected.command=--title="Packets Rejected" \ | |
LINE2:rejected#0000ff:"Rejected" \ | |
GPRINT:rejected:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:rejected:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:rejected:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:rejected:MAX:"Max \\: %8.2lf %s\\n" | |
report.checkpoint.pktsEncrypted.name=Packets Encrypted and Decrypted (CheckPoint) | |
report.checkpoint.pktsEncrypted.columns=cpvEncPackets, cpvDecPackets | |
diff --git a/snmp-graph.properties.d/cisco-graph.properties b/snmp-graph.properties.d/cisco-graph.properties | |
index 3c28ddb..1bf9fd1 100644 | |
--- a/snmp-graph.properties.d/cisco-graph.properties | |
+++ b/snmp-graph.properties.d/cisco-graph.properties | |
@@ -702,7 +702,7 @@ report.cisco.css.hits.command=--title="Hits per Second" --height 150 --width 600 | |
LINE1:fixedHits#0000ff:"Hits " \ | |
GPRINT:fixedHits:AVERAGE:"Avg\\: %8.2lf %s" \ | |
GPRINT:fixedHits:MIN:"Min\\: %8.2lf %s" \ | |
- GPRINT:fixedHits:MAX:"Max\\: %8.2lf %s\\n" | |
+ GPRINT:fixedHits:MAX:"Max\\: %8.2lf %s\\n" | |
report.cisco.css.bytes.name=CSS Bytes | |
report.cisco.css.bytes.height=150 | |
@@ -716,7 +716,7 @@ report.cisco.css.bytes.command=--title="Bytes per Second" --height 150 --width 6 | |
LINE1:Bytes#0000ff:"Bytes/second " \ | |
GPRINT:Bytes:AVERAGE:"Avg\\: %8.2lf %s" \ | |
GPRINT:Bytes:MIN:"Min\\: %8.2lf %s" \ | |
- GPRINT:Bytes:MAX:"Max\\: %8.2lf %s\\n" | |
+ GPRINT:Bytes:MAX:"Max\\: %8.2lf %s\\n" | |
report.cisco.css.redirects.name=CSS Redirects | |
report.cisco.css.redirects.height=150 | |
@@ -729,7 +729,7 @@ report.cisco.css.redirects.command=--title="Redirects per Second" --height 150 - | |
LINE1:Redirs#0000ff:"Redirects " \ | |
GPRINT:Redirs:AVERAGE:"Avg\\: %8.2lf %s" \ | |
GPRINT:Redirs:MIN:"Min\\: %8.2lf %s" \ | |
- GPRINT:Redirs:MAX:"Max\\: %8.2lf %s\\n" | |
+ GPRINT:Redirs:MAX:"Max\\: %8.2lf %s\\n" | |
report.cisco.css.sorries.name=CSS Sorries | |
report.cisco.css.sorries.height=150 | |
@@ -742,7 +742,7 @@ report.cisco.css.sorries.command=--title="Sorries per Second" --height 150 --wid | |
LINE1:Redirs#0000ff:"Sorries " \ | |
GPRINT:Redirs:AVERAGE:"Avg\\: %8.2lf %s" \ | |
GPRINT:Redirs:MIN:"Min\\: %8.2lf %s" \ | |
- GPRINT:Redirs:MAX:"Max\\: %8.2lf %s\\n" | |
+ GPRINT:Redirs:MAX:"Max\\: %8.2lf %s\\n" | |
report.cisco.docs.macchan.name=Cable Modem (MacChannel) Users | |
report.cisco.docs.macchan.columns=MacCmTotal,MacCmActive,MacCmReg | |
diff --git a/snmp-graph.properties.d/ciscoNexus-graph.properties b/snmp-graph.properties.d/ciscoNexus-graph.properties | |
index b98e341..b465dd9 100644 | |
--- a/snmp-graph.properties.d/ciscoNexus-graph.properties | |
+++ b/snmp-graph.properties.d/ciscoNexus-graph.properties | |
@@ -142,4 +142,5 @@ report.cisco.nexus.env.command=--title="Environmental Status of {entPhysicalName | |
LINE1:entSensor#f57900:"Temperature" \ | |
GPRINT:entSensor:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:entSensor:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:entSensor:MAX:"Max \\: %8.2lf %s\\n" | |
+ GPRINT:entSensor:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
diff --git a/snmp-graph.properties.d/clavister-graph.properties b/snmp-graph.properties.d/clavister-graph.properties | |
index c8f87b0..0269e7a 100644 | |
--- a/snmp-graph.properties.d/clavister-graph.properties | |
+++ b/snmp-graph.properties.d/clavister-graph.properties | |
@@ -294,7 +294,7 @@ report.clavister.ipsec.info.command=--title="Clavister IPsec Informational Excha | |
LINE2:clvIPsecInfoFailed#a40000:"Failed" \ | |
GPRINT:clvIPsecInfoFailed:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:clvIPsecInfoFailed:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:clvIPsecInfoFailed:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:clvIPsecInfoFailed:MAX:"Max \\: %8.2lf %s\\n" | |
report.clavister.ipsec.bits.name=Clavister IPsec Bits In/Out | |
report.clavister.ipsec.bits.columns=clvIPsecInOctComp,clvIPsecInOctUncomp,clvIPsecOutOctComp,clvIPsecOutOctUncom | |
@@ -798,7 +798,7 @@ report.clavister.vlan.untaggedpkts.command=--title="Clavister VLAN Untagged Pack | |
COMMENT:" Total" \ | |
GPRINT:total:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:total:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:total:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:total:MAX:"Max \\: %8.2lf %s\\n" | |
report.clavister.vlan.untaggedocts.name=Clavister VLAN Untagged Bytes | |
report.clavister.vlan.untaggedocts.columns=clvIfVlUntInOctets,clvIfVlUntOutOctets,clvIfVlUntTotOctets | |
@@ -823,7 +823,7 @@ report.clavister.vlan.untaggedocts.command=--title="Clavister VLAN Untagged Byte | |
COMMENT:" Total" \ | |
GPRINT:total:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:total:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:total:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:total:MAX:"Max \\: %8.2lf %s\\n" | |
# | |
#clvDHCPRelayRuleTable | |
@@ -1172,7 +1172,7 @@ report.clavister.system.connpersec.command=--title="Clavister System Connections | |
LINE1:closedInv#3465a4:"Closed" \ | |
GPRINT:closed:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:closed:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:closed:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:closed:MAX:"Max \\: %8.2lf %s\\n" | |
report.clavister.system.hcfwdbits.name=Clavister System Bits Forwarded (High Speed) | |
report.clavister.system.hcfwdbits.columns=clvSysHCFwdBits | |
diff --git a/snmp-graph.properties.d/ejn-graph.properties b/snmp-graph.properties.d/ejn-graph.properties | |
index 72e96f6..ecb0bce 100644 | |
--- a/snmp-graph.properties.d/ejn-graph.properties | |
+++ b/snmp-graph.properties.d/ejn-graph.properties | |
@@ -80,7 +80,7 @@ report.ejnggsn.apn.users.command=--title="APN {ApnName} Active PDP Contexts" \ | |
LINE2:active#0000ff:"PDP Contexts" \ | |
GPRINT:active:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:active:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:active:MAX:"Max \\: %8.2lf %s\\n" | |
+ GPRINT:active:MAX:"Max \\: %8.2lf %s\\n" | |
report.ejnggsn.apn.ippool.name=GGSN APN Available IPs | |
report.ejnggsn.apn.ippool.columns=ApnFreePoolIps | |
diff --git a/snmp-graph.properties.d/elasticsearch-graph.properties b/snmp-graph.properties.d/elasticsearch-graph.properties | |
new file mode 100644 | |
index 0000000..2731291 | |
--- /dev/null | |
+++ b/snmp-graph.properties.d/elasticsearch-graph.properties | |
@@ -0,0 +1,581 @@ | |
+reports=elasticsearch.cluster.indices.count, \ | |
+ elasticsearch.cluster.indices.shards.index.shards, \ | |
+ elasticsearch.cluster.indices.shards.index.primaries, \ | |
+ elasticsearch.cluster.indices.shards.index.replication, \ | |
+ elasticsearch.cluster.indices.docs, \ | |
+ elasticsearch.cluster.indices.store, \ | |
+ elasticsearch.cluster.indices.throttle.time, \ | |
+ elasticsearch.cluster.indices.fielddata.size, \ | |
+ elasticsearch.cluster.indices.fielddata.evictions, \ | |
+ elasticsearch.cluster.indices.filtercache.size, \ | |
+ elasticsearch.cluster.indices.filtercache.evictions, \ | |
+ elasticsearch.cluster.indices.idcache.size, \ | |
+ elasticsearch.cluster.indices.completion.size, \ | |
+ elasticsearch.cluster.indices.segments.count, \ | |
+ elasticsearch.cluster.indices.segments.memory, \ | |
+ elasticsearch.cluster.indices.perculate.total, \ | |
+ elasticsearch.cluster.indices.perculate.time, \ | |
+ elasticsearch.cluster.indices.perculate.current, \ | |
+ elasticsearch.cluster.indices.perculate.size, \ | |
+ elasticsearch.cluster.indices.perculate.queries, \ | |
+ elasticsearch.cluster.nodes, \ | |
+ elasticsearch.cluster.nodes.os.processors, \ | |
+ elasticsearch.cluster.nodes.os.memory.total, \ | |
+ elasticsearch.cluster.nodes.os.cpu, \ | |
+ elasticsearch.cluster.nodes.open.filedescriptors, \ | |
+ elasticsearch.cluster.node.jvm.uptime, \ | |
+ elasticsearch.cluster.node.jvm.memory, \ | |
+ elasticsearch.cluster.node.jvm.threads, \ | |
+ elasticsearch.cluster.node.fs.size, \ | |
+ elasticsearch.cluster.node.fs.dsk.io.ops, \ | |
+ elasticsearch.cluster.node.fs.dsk.io.size, \ | |
+ elasticsearch.cluster.node.fs.dsk.queue, \ | |
+ elasticsearch.cluster.node.fs.dsk.svc.time | |
+ | |
+report.elasticsearch.cluster.indices.count.name=ES Cluster Indices | |
+report.elasticsearch.cluster.indices.count.columns=indicesCount | |
+report.elasticsearch.cluster.indices.count.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.count.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.count.command=--title="ES Cluster Indices: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Count" \ | |
+ DEF:val1={rrd1}:indicesCount:AVERAGE \ | |
+ AREA:val1#babdb6 \ | |
+ LINE2:val1#888a85:"Indices " \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.shards.index.shards.name=ES Cluster Index Shards | |
+report.elasticsearch.cluster.indices.shards.index.shards.columns=shardsMin, shardsMax, shardsAvg | |
+report.elasticsearch.cluster.indices.shards.index.shards.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.shards.index.shards.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.shards.index.shards.command=--title="ES Cluster Index Shards: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Shards" \ | |
+ DEF:val1={rrd1}:shardsMin:AVERAGE \ | |
+ DEF:val2={rrd2}:shardsMax:AVERAGE \ | |
+ DEF:val3={rrd3}:shardsAvg:AVERAGE \ | |
+ COMMENT:"Index Shards\\n" \ | |
+ LINE2:val1#ad7fa8:"Min " \ | |
+ GPRINT:val1:MIN:"%10.2lf\\n" \ | |
+ LINE2:val2#729fcf:"Max " \ | |
+ GPRINT:val2:MAX:"%10.2lf\\n" \ | |
+ LINE2:val3#c17d11:"Avg " \ | |
+ GPRINT:val3:AVERAGE:"%10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.shards.index.primaries.name=ES Cluster Index Primaries | |
+report.elasticsearch.cluster.indices.shards.index.primaries.columns=primariesMin, primariesMax, primariesAvg | |
+report.elasticsearch.cluster.indices.shards.index.primaries.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.shards.index.primaries.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.shards.index.primaries.command=--title="ES Cluster Index Primaries: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Primaries" \ | |
+ DEF:val1={rrd1}:primariesMin:AVERAGE \ | |
+ DEF:val2={rrd2}:primariesMax:AVERAGE \ | |
+ DEF:val3={rrd3}:primariesAvg:AVERAGE \ | |
+ COMMENT:"Index Primaries\\n" \ | |
+ LINE2:val1#ad7fa8:"Min " \ | |
+ GPRINT:val1:MIN:"%10.2lf\\n" \ | |
+ LINE2:val2#729fcf:"Max " \ | |
+ GPRINT:val2:MAX:"%10.2lf\\n" \ | |
+ LINE2:val3#c17d11:"Avg " \ | |
+ GPRINT:val3:AVERAGE:"%10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.shards.index.replication.name=ES Cluster Index Replication | |
+report.elasticsearch.cluster.indices.shards.index.replication.columns=replicationMin, replicationMax, replicationAvg | |
+report.elasticsearch.cluster.indices.shards.index.replication.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.shards.index.replication.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.shards.index.replication.command=--title="ES Cluster Index Replication: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Replication" \ | |
+ DEF:val1={rrd1}:replicationMin:AVERAGE \ | |
+ DEF:val2={rrd2}:replicationMax:AVERAGE \ | |
+ DEF:val3={rrd3}:replicationAvg:AVERAGE \ | |
+ COMMENT:"Index Replication\\n" \ | |
+ LINE2:val1#ad7fa8:"Min " \ | |
+ GPRINT:val1:MIN:"%10.2lf\\n" \ | |
+ LINE2:val2#729fcf:"Max " \ | |
+ GPRINT:val2:MAX:"%10.2lf\\n" \ | |
+ LINE2:val3#c17d11:"Avg " \ | |
+ GPRINT:val3:AVERAGE:"%10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.docs.name=ES Cluster Indices Documents | |
+report.elasticsearch.cluster.indices.docs.columns=docsCount, docsDeleted | |
+report.elasticsearch.cluster.indices.docs.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.docs.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.docs.command=--title="ES Cluster Indices Documents: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Documents" \ | |
+ DEF:val1={rrd1}:docsCount:AVERAGE \ | |
+ DEF:val2={rrd2}:docsDeleted:AVERAGE \ | |
+ LINE2:val1#729fcf:"Count " \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val2#cc0000:"Deleted " \ | |
+ GPRINT:val2:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val2:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val2:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.store.name=ES Cluster Indices Store Size | |
+report.elasticsearch.cluster.indices.store.columns=storeSizeBytes | |
+report.elasticsearch.cluster.indices.store.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.store.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.store.command=--title="ES Cluster Indices Store Size: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:storeSizeBytes:AVERAGE \ | |
+ AREA:val1#729fcf \ | |
+ LINE2:val1#3465a4:"Store Size" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.throttle.time.name=ES Cluster Indices Store Throttle Time | |
+report.elasticsearch.cluster.indices.throttle.time.columns=throttleTimeMillis | |
+report.elasticsearch.cluster.indices.throttle.time.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.throttle.time.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.throttle.time.command=--title="ES Cluster Indices Store Throttle Time: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Milliseconds" \ | |
+ DEF:val1={rrd1}:throttleTimeMillis:AVERAGE \ | |
+ AREA:val1#ad7fa8 \ | |
+ LINE2:val1#75507b:"Time in ms" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.fielddata.size.name=ES Cluster Indices Field Data Size | |
+report.elasticsearch.cluster.indices.fielddata.size.columns=memorySizeBytes | |
+report.elasticsearch.cluster.indices.fielddata.size.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.fielddata.size.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.fielddata.size.command=--title="ES Cluster Indices Field Data Size: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:memorySizeBytes:AVERAGE \ | |
+ AREA:val1#729fcf \ | |
+ LINE2:val1#3465a4:"Field Data Size" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.fielddata.evictions.name=ES Cluster Indices Field Data Evictions | |
+report.elasticsearch.cluster.indices.fielddata.evictions.columns=memoryEvictions | |
+report.elasticsearch.cluster.indices.fielddata.evictions.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.fielddata.evictions.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.fielddata.evictions.command=--title="ES Cluster Indices Field Data Evictions: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Evictions" \ | |
+ DEF:val1={rrd1}:memoryEvictions:AVERAGE \ | |
+ AREA:val1#e9b96e \ | |
+ LINE2:val1#c17d11:"Field Data Evictions" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.filtercache.size.name=ES Cluster Indices Filter Cache Size | |
+report.elasticsearch.cluster.indices.filtercache.size.columns=fltrCacheBytes | |
+report.elasticsearch.cluster.indices.filtercache.size.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.filtercache.size.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.filtercache.size.command=--title="ES Cluster Indices Filter Cache Size: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:fltrCacheBytes:AVERAGE \ | |
+ AREA:val1#729fcf \ | |
+ LINE2:val1#3465a4:"Filter Cache Size" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.filtercache.evictions.name=ES Cluster Indices Filter Cache Evictions | |
+report.elasticsearch.cluster.indices.filtercache.evictions.columns=fltrCacheEvictions | |
+report.elasticsearch.cluster.indices.filtercache.evictions.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.filtercache.evictions.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.filtercache.evictions.command=--title="ES Cluster Indices Filter Cache Evictions: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Evictions" \ | |
+ DEF:val1={rrd1}:fltrCacheEvictions:AVERAGE \ | |
+ AREA:val1#e9b96e \ | |
+ LINE2:val1#c17d11:"Field Cache Evictions" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.idcache.size.name=ES Cluster Indices ID Cache Size | |
+report.elasticsearch.cluster.indices.idcache.size.columns=idCacheBytes | |
+report.elasticsearch.cluster.indices.idcache.size.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.idcache.size.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.idcache.size.command=--title="ES Cluster Indices ID Cache Size: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:idCacheBytes:AVERAGE \ | |
+ AREA:val1#729fcf \ | |
+ LINE2:val1#3465a4:"ID Cache Size" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.completion.size.name=ES Cluster Indices Completion Size | |
+report.elasticsearch.cluster.indices.completion.size.columns=completionSizeBytes | |
+report.elasticsearch.cluster.indices.completion.size.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.completion.size.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.completion.size.command=--title="ES Cluster Indices Completion Count: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:completionSizeBytes:AVERAGE \ | |
+ AREA:val1#729fcf \ | |
+ LINE2:val1#3465a4:"Completion Size" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.segments.count.name=ES Cluster Indices Segments Count | |
+report.elasticsearch.cluster.indices.segments.count.columns=segCount | |
+report.elasticsearch.cluster.indices.segments.count.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.segments.count.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.segments.count.command=--title="ES Cluster Indices Segments Count: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Count" \ | |
+ DEF:val1={rrd1}:segCount:AVERAGE \ | |
+ AREA:val1#ad7fa8 \ | |
+ LINE2:val1#75507b:"Segments" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.segments.memory.name=ES Cluster Indices Segments Memory | |
+report.elasticsearch.cluster.indices.segments.memory.columns=segMemory, segWriterMem, segWriterMemMax, segVerMapMem, segFixBitSetMem | |
+report.elasticsearch.cluster.indices.segments.memory.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.segments.memory.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.segments.memory.command=--title="ES Cluster Indices Segments Memory: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:segMemory:AVERAGE \ | |
+ DEF:val2={rrd2}:segWriterMem:AVERAGE \ | |
+ DEF:val3={rrd3}:segWriterMemMax:AVERAGE \ | |
+ DEF:val4={rrd4}:segVerMapMem:AVERAGE \ | |
+ DEF:val5={rrd5}:segFixBitSetMem:AVERAGE \ | |
+ LINE2:val1#75507b:"Memory " \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val2#729fcf:"Writer Memory " \ | |
+ GPRINT:val2:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val2:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val2:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val3#204a87:"Writer Memory Max " \ | |
+ GPRINT:val3:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val3:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val3:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val4#c17d11:"Version Map Memory " \ | |
+ GPRINT:val4:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val4:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val4:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val5#888a85:"Fixed Bit Set Memory " \ | |
+ GPRINT:val5:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val5:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val5:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.perculate.total.name=ES Cluster Indices Perculate Total | |
+report.elasticsearch.cluster.indices.perculate.total.columns=percTotal | |
+report.elasticsearch.cluster.indices.perculate.total.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.perculate.total.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.perculate.total.command=--title="ES Cluster Indices Perculate Total: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Total" \ | |
+ DEF:val1={rrd1}:percTotal:AVERAGE \ | |
+ LINE2:val1#75507b:"Perculate " \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.perculate.time.name=ES Cluster Indices Perculate Time | |
+report.elasticsearch.cluster.indices.perculate.time.columns=percTimeInMillis | |
+report.elasticsearch.cluster.indices.perculate.time.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.perculate.time.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.perculate.time.command=--title="ES Cluster Indices Perculate Time: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Milliseconds" \ | |
+ DEF:val1={rrd1}:percTimeInMillis:AVERAGE \ | |
+ AREA:val1#ad7fa8 \ | |
+ LINE2:val1#75507b:"Time in ms" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.perculate.current.name=ES Cluster Indices Perculate Current | |
+report.elasticsearch.cluster.indices.perculate.current.columns=percCurrent | |
+report.elasticsearch.cluster.indices.perculate.current.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.perculate.current.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.perculate.current.command=--title="ES Cluster Indices Perculate Current: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Current" \ | |
+ DEF:val1={rrd1}:percCurrent:AVERAGE \ | |
+ AREA:val1#ad7fa8 \ | |
+ LINE2:val1#75507b:"Current" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.perculate.size.name=ES Cluster Indices Perculate Size | |
+report.elasticsearch.cluster.indices.perculate.size.columns=percMemSizeBytes | |
+report.elasticsearch.cluster.indices.perculate.size.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.perculate.size.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.perculate.size.command=--title="ES Cluster Indices Perculate Size: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:percMemSizeBytes:AVERAGE \ | |
+ AREA:val1#729fcf \ | |
+ LINE2:val1#3465a4:"Size" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.indices.perculate.queries.name=ES Cluster Indices Perculate Queries | |
+report.elasticsearch.cluster.indices.perculate.queries.columns=percQueries | |
+report.elasticsearch.cluster.indices.perculate.queries.type=nodeSnmp | |
+report.elasticsearch.cluster.indices.perculate.queries.propertiesValues=clusterName | |
+report.elasticsearch.cluster.indices.perculate.queries.command=--title="ES Cluster Indices Perculate Queries: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Queries" \ | |
+ DEF:val1={rrd1}:percQueries:AVERAGE \ | |
+ LINE2:val1#75507b:"Queries" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.nodes.name=ES Cluster Nodes | |
+report.elasticsearch.cluster.nodes.columns=nodesCntTotal, nodesCntMstOnly, nodesCntDataOnly, nodesCntMasterData, nodesClient | |
+report.elasticsearch.cluster.nodes.type=nodeSnmp | |
+report.elasticsearch.cluster.nodes.propertiesValues=clusterName | |
+report.elasticsearch.cluster.nodes.command=--title="ES Cluster Indices Perculate Queries: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Nodes" \ | |
+ DEF:val1={rrd1}:nodesCntTotal:AVERAGE \ | |
+ DEF:val2={rrd2}:nodesCntMstOnly:AVERAGE \ | |
+ DEF:val3={rrd3}:nodesCntDataOnly:AVERAGE \ | |
+ DEF:val4={rrd4}:nodesCntMasterData:AVERAGE \ | |
+ DEF:val5={rrd5}:nodesClient:AVERAGE \ | |
+ LINE2:val1#75507b:"Total " \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val2#c17d11:"Master Only " \ | |
+ GPRINT:val2:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val2:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val2:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val3#75507b:"Data Only " \ | |
+ GPRINT:val3:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val3:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val3:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val4#3465a4:"Master Data " \ | |
+ GPRINT:val4:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val4:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val4:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val5#888a85:"Clients " \ | |
+ GPRINT:val5:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val5:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val5:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.nodes.os.processors.name=ES Cluster Nodes Processors | |
+report.elasticsearch.cluster.nodes.os.processors.columns=osAvailProc | |
+report.elasticsearch.cluster.nodes.os.processors.type=nodeSnmp | |
+report.elasticsearch.cluster.nodes.os.processors.propertiesValues=clusterName | |
+report.elasticsearch.cluster.nodes.os.processors.command=--title="ES Cluster Node Processors: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Processors" \ | |
+ DEF:val1={rrd1}:osAvailProc:AVERAGE \ | |
+ LINE2:val1#75507b:"Processors" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.nodes.os.memory.total.name=ES Cluster Nodes OS Memory Total | |
+report.elasticsearch.cluster.nodes.os.memory.total.columns=osMemTotalBytes | |
+report.elasticsearch.cluster.nodes.os.memory.total.type=nodeSnmp | |
+report.elasticsearch.cluster.nodes.os.memory.total.propertiesValues=clusterName | |
+report.elasticsearch.cluster.nodes.os.memory.total.command=--title="ES Cluster Node OS Memory Total: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:osMemTotalBytes:AVERAGE \ | |
+ AREA:val1#729fcf \ | |
+ LINE2:val1#3465a4:"OS Memory Total" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.nodes.os.cpu.name=ES Cluster Nodes Process CPU | |
+report.elasticsearch.cluster.nodes.os.cpu.columns=procCpuPercent | |
+report.elasticsearch.cluster.nodes.os.cpu.type=nodeSnmp | |
+report.elasticsearch.cluster.nodes.os.cpu.propertiesValues=clusterName | |
+report.elasticsearch.cluster.nodes.os.cpu.command=--title="ES Cluster Node Process CPU: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Percent" \ | |
+ DEF:val1={rrd1}:procCpuPercent:AVERAGE \ | |
+ AREA:val1#ad7fa8 \ | |
+ LINE2:val1#75507b:"Process CPU" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.nodes.open.filedescriptors.name=ES Cluster Nodes Open File Descriptors | |
+report.elasticsearch.cluster.nodes.open.filedescriptors.columns=openFileDescMin, openFileDescMax, openFileDescAvg | |
+report.elasticsearch.cluster.nodes.open.filedescriptors.type=nodeSnmp | |
+report.elasticsearch.cluster.nodes.open.filedescriptors.propertiesValues=clusterName | |
+report.elasticsearch.cluster.nodes.open.filedescriptors.command=--title="ES Cluster Node Open File Descriptors: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Open File Descriptors" \ | |
+ DEF:val1={rrd1}:openFileDescMin:AVERAGE \ | |
+ DEF:val2={rrd2}:openFileDescMax:AVERAGE \ | |
+ DEF:val3={rrd3}:openFileDescAvg:AVERAGE \ | |
+ COMMENT:"Open File Descriptors\\n" \ | |
+ LINE2:val1#ad7fa8:"Min " \ | |
+ GPRINT:val1:MIN:"%10.2lf\\n" \ | |
+ LINE2:val2#729fcf:"Max " \ | |
+ GPRINT:val2:MAX:"%10.2lf\\n" \ | |
+ LINE2:val3#c17d11:"Avg " \ | |
+ GPRINT:val3:AVERAGE:"%10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.node.jvm.uptime.name=ES Cluster Node JVM Uptime | |
+report.elasticsearch.cluster.node.jvm.uptime.columns=jvmMaxUptimeMillis | |
+report.elasticsearch.cluster.node.jvm.uptime.type=nodeSnmp | |
+report.elasticsearch.cluster.node.jvm.uptime.propertiesValues=clusterName | |
+report.elasticsearch.cluster.node.jvm.uptime.command=--title="ES Cluster Node JVM Uptime: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Milliseconds" \ | |
+ DEF:val1={rrd1}:jvmMaxUptimeMillis:AVERAGE \ | |
+ AREA:val1#ad7fa8 \ | |
+ LINE2:val1#75507b:"Time in ms" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.node.jvm.memory.name=ES Cluster Node JVM Memory | |
+report.elasticsearch.cluster.node.jvm.memory.columns=jvmMemHeapBytes, jvmMemMaxBytes | |
+report.elasticsearch.cluster.node.jvm.memory.type=nodeSnmp | |
+report.elasticsearch.cluster.node.jvm.memory.propertiesValues=clusterName | |
+report.elasticsearch.cluster.node.jvm.memory.command=--title="ES Cluster Node JVM Memory: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:jvmMemHeapBytes:AVERAGE \ | |
+ DEF:val2={rrd2}:jvmMemMaxBytes:AVERAGE \ | |
+ LINE2:val1#f57900:"JVM Heap Used" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val2#cc0000:"JVM Heap Max " \ | |
+ GPRINT:val2:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val2:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val2:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.node.jvm.threads.name=ES Cluster Node JVM Threads | |
+report.elasticsearch.cluster.node.jvm.threads.columns=jvmThreads | |
+report.elasticsearch.cluster.node.jvm.threads.type=nodeSnmp | |
+report.elasticsearch.cluster.node.jvm.threads.propertiesValues=clusterName | |
+report.elasticsearch.cluster.node.jvm.threads.command=--title="ES Cluster Node JVM Threads: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Threads" \ | |
+ DEF:val1={rrd1}:jvmThreads:AVERAGE \ | |
+ AREA:val1#ad7fa8 \ | |
+ LINE2:val1#75507b:"JVM Threads" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.node.fs.size.name=ES Cluster Node File System Size | |
+report.elasticsearch.cluster.node.fs.size.columns=fsTotalBytes, fsFreeBytes, fsAvailBytes | |
+report.elasticsearch.cluster.node.fs.size.type=nodeSnmp | |
+report.elasticsearch.cluster.node.fs.size.propertiesValues=clusterName | |
+report.elasticsearch.cluster.node.fs.size.command=--title="ES Cluster File System Size: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:fsTotalBytes:AVERAGE \ | |
+ DEF:val2={rrd2}:fsFreeBytes:AVERAGE \ | |
+ DEF:val3={rrd3}:fsAvailBytes:AVERAGE \ | |
+ LINE2:val1#3465a4:"File System Total " \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val2#4e9a06:"File System Free " \ | |
+ GPRINT:val2:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val2:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val2:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val3#555753:"File System Available " \ | |
+ GPRINT:val3:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val3:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val3:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.node.fs.dsk.io.ops.name=ES Cluster Node File System I/O Ops | |
+report.elasticsearch.cluster.node.fs.dsk.io.ops.columns=fsDskReads, fsDskWrites, fsDskIoOp | |
+report.elasticsearch.cluster.node.fs.dsk.io.ops.type=nodeSnmp | |
+report.elasticsearch.cluster.node.fs.dsk.io.ops.propertiesValues=clusterName | |
+report.elasticsearch.cluster.node.fs.dsk.io.ops.command=--title="ES Cluster File System I/O Ops: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="I/O Ops" \ | |
+ DEF:val1={rrd1}:fsDskReads:AVERAGE \ | |
+ DEF:val2={rrd2}:fsDskWrites:AVERAGE \ | |
+ DEF:val3={rrd3}:fsDskIoOp:AVERAGE \ | |
+ LINE2:val1#4e9a06:"Disk I/O Read Ops " \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val2#3465a4:"Disk I/O Write Ops" \ | |
+ GPRINT:val2:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val2:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val2:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val3#f57900:"Disk I/O Ops Total" \ | |
+ GPRINT:val3:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val3:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val3:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.node.fs.dsk.io.size.name=ES Cluster Node File System I/O Size | |
+report.elasticsearch.cluster.node.fs.dsk.io.size.columns=fsDskReadSizeBytes, fsDskWriteSizeBytes, fsDskIoSizeBytes | |
+report.elasticsearch.cluster.node.fs.dsk.io.size.type=nodeSnmp | |
+report.elasticsearch.cluster.node.fs.dsk.io.size.propertiesValues=clusterName | |
+report.elasticsearch.cluster.node.fs.dsk.io.size.command=--title="ES Cluster File System I/O Size: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Bytes" \ | |
+ DEF:val1={rrd1}:fsDskReadSizeBytes:AVERAGE \ | |
+ DEF:val2={rrd2}:fsDskWriteSizeBytes:AVERAGE \ | |
+ DEF:val3={rrd3}:fsDskIoSizeBytes:AVERAGE \ | |
+ LINE2:val1#4e9a06:"Disk I/O Read size " \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val2#3465a4:"Disk I/O Writes " \ | |
+ GPRINT:val2:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val2:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val2:MAX:"Max \\: %10.2lf\\n" \ | |
+ LINE2:val3#f57900:"Disk I/O Size Total" \ | |
+ GPRINT:val3:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val3:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val3:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.node.fs.dsk.queue.name=ES Cluster Node File System Disk Queue | |
+report.elasticsearch.cluster.node.fs.dsk.queue.columns=fsDskQueue | |
+report.elasticsearch.cluster.node.fs.dsk.queue.type=nodeSnmp | |
+report.elasticsearch.cluster.node.fs.dsk.queue.propertiesValues=clusterName | |
+report.elasticsearch.cluster.node.fs.dsk.queue.command=--title="ES Cluster Node File System Disk Queue: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Queue" \ | |
+ DEF:val1={rrd1}:fsDskQueue:AVERAGE \ | |
+ AREA:val1#fcaf3e \ | |
+ LINE2:val1#f57900:"Disk Queue" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
+ | |
+report.elasticsearch.cluster.node.fs.dsk.svc.time.name=ES Cluster Node File System Disk Service Time | |
+report.elasticsearch.cluster.node.fs.dsk.svc.time.columns=fsDskSvcTime | |
+report.elasticsearch.cluster.node.fs.dsk.svc.time.type=nodeSnmp | |
+report.elasticsearch.cluster.node.fs.dsk.svc.time.propertiesValues=clusterName | |
+report.elasticsearch.cluster.node.fs.dsk.svc.time.command=--title="ES Cluster Node File System Disk Service Time: {clusterName}" \ | |
+ --units-exponent=0 \ | |
+ --vertical-label="Service Time" \ | |
+ DEF:val1={rrd1}:fsDskSvcTime:AVERAGE \ | |
+ AREA:val1#fcaf3e \ | |
+ LINE2:val1#f57900:"Disk Queue" \ | |
+ GPRINT:val1:AVERAGE:"Avg \\: %10.2lf" \ | |
+ GPRINT:val1:MIN:"Min \\: %10.2lf" \ | |
+ GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
\ No newline at end of file | |
diff --git a/snmp-graph.properties.d/f5-graph.properties b/snmp-graph.properties.d/f5-graph.properties | |
index 2e5ad73..c6ae650 100644 | |
--- a/snmp-graph.properties.d/f5-graph.properties | |
+++ b/snmp-graph.properties.d/f5-graph.properties | |
@@ -130,7 +130,7 @@ report.bigip.lvsconns.command=--title="Virtual Server Current Connections (F5)" | |
LINE2:curConns#4e9a06:"Current " \ | |
GPRINT:curConns:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:curConns:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:curConns:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:curConns:MAX:"Max \\: %8.2lf %s\\n" | |
report.bigip.nonodeerrs.name=LTM Virtual Server No Node Errors | |
report.bigip.nonodeerrs.columns=vsNoNodeErrs | |
diff --git a/snmp-graph.properties.d/fortinet-fortigate-application-v5.2-graph.properties b/snmp-graph.properties.d/fortinet-fortigate-application-v5.2-graph.properties | |
index 4cfc452..b5592d4 100644 | |
--- a/snmp-graph.properties.d/fortinet-fortigate-application-v5.2-graph.properties | |
+++ b/snmp-graph.properties.d/fortinet-fortigate-application-v5.2-graph.properties | |
@@ -377,12 +377,12 @@ report.fortinet.fgApFTPStatsEntry.stats.command=--title="Fortigate FTP Proxy Sta | |
GPRINT:val1:MAX:"Max \\: %10.2lf\\n" | |
report.fortinet.fgApFTPStatsEntry.connection.name=Fortigate FTP Proxy Connections Statistics | |
-report.fortinet.fgApFTPStatsEntry.connection.columns=fgApFTPMaxConnections, fgApFTPConnections | |
+report.fortinet.fgApFTPStatsEntry.connection.columns=fgApFTPMaxConns, fgApFTPConns | |
report.fortinet.fgApFTPStatsEntry.connection.type=nodeSnmp | |
report.fortinet.fgApFTPStatsEntry.connection.command=--title="Fortigate FTP Proxy Connections Statistics" \ | |
--vertical-label="number" \ | |
- DEF:val1={rrd1}:fgApFTPMaxConnections:AVERAGE \ | |
- DEF:val2={rrd2}:fgApFTPConnections:AVERAGE \ | |
+ DEF:val1={rrd1}:fgApFTPMaxConns:AVERAGE \ | |
+ DEF:val2={rrd2}:fgApFTPConns:AVERAGE \ | |
LINE1:val1#cc0000:"max Connections" \ | |
GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:val1:MIN:"Min \\: %8.2lf %s" \ | |
@@ -410,13 +410,13 @@ report.fortinet.fgWebCacheDiskUsage.command=--title="Fortigate Web Cache Usage f | |
GPRINT:val2:MAX:"Max \\: %8.2lf %s\\n" | |
report.fortinet.fgWebChDskStsEntry.name=Fortigate Web Cache Disk Statistics | |
-report.fortinet.fgWebChDskStsEntry.columns=fgWebCacheDiskHits, fgWebCacheDiskMisses | |
+report.fortinet.fgWebChDskStsEntry.columns=fgWebCacheDiskHits, fgWebCacheDiskMiss | |
report.fortinet.fgWebChDskStsEntry.type=fgWebChDskStsEntry | |
report.fortinet.fgWebChDskStsEntry.propertiesValues=fgWebCacheDisk | |
report.fortinet.fgWebChDskStsEntry.command=--title="Fortigate Web Cache Disk Statistics for Disk: {fgWebCacheDisk}" \ | |
--vertical-label="number" \ | |
DEF:val1={rrd1}:fgWebCacheDiskHits:AVERAGE \ | |
- DEF:val2={rrd2}:fgWebCacheDiskMisses:AVERAGE \ | |
+ DEF:val2={rrd2}:fgWebCacheDiskMiss:AVERAGE \ | |
AREA:val1#cc0000:"Hits " \ | |
GPRINT:val1:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:val1:MIN:"Min \\: %8.2lf %s" \ | |
diff --git a/snmp-graph.properties.d/hwg-graph.properties b/snmp-graph.properties.d/hwg-graph.properties | |
index 766074c..5fd10b5 100644 | |
--- a/snmp-graph.properties.d/hwg-graph.properties | |
+++ b/snmp-graph.properties.d/hwg-graph.properties | |
@@ -36,13 +36,12 @@ report.sensTable.sensValue.command=--title="The Value Reported by {sensName} in | |
LINE2:var#000000:"Value" \ | |
GPRINT:var:AVERAGE:"Avg\\: %8.2lf %s" \ | |
GPRINT:var:MIN:"Min\\: %8.2lf %s" \ | |
- GPRINT:var:MAX:"Max\\: %8.2lf %s\n" \ | |
+ GPRINT:var:MAX:"Max\\: %8.2lf %s\\n" \ | |
LINE2:max#A00000:"Max " \ | |
GPRINT:max:AVERAGE:"Avg\\: %8.2lf %s" \ | |
GPRINT:max:MIN:"Min\\: %8.2lf %s" \ | |
- GPRINT:max:MAX:"Max\\: %8.2lf %s\n" \ | |
+ GPRINT:max:MAX:"Max\\: %8.2lf %s\\n" \ | |
LINE2:min#0000A0:"Min " \ | |
GPRINT:min:AVERAGE:"Avg\\: %8.2lf %s" \ | |
GPRINT:min:MIN:"Min\\: %8.2lf %s" \ | |
- GPRINT:min:MAX:"Max\\: %8.2lf %s\n | |
- | |
+ GPRINT:min:MAX:"Max\\: %8.2lf %s\\n" | |
\ No newline at end of file | |
diff --git a/snmp-graph.properties.d/ipunity-graph.properties b/snmp-graph.properties.d/ipunity-graph.properties | |
index 6fa331f..e616566 100644 | |
--- a/snmp-graph.properties.d/ipunity-graph.properties | |
+++ b/snmp-graph.properties.d/ipunity-graph.properties | |
@@ -589,7 +589,7 @@ report.ipunity.sip.methodDetail.command=--title="SIP Method Detail ({applDescrip | |
STACK:infoOutInv#ff7200:"INFO " \ | |
GPRINT:infoOut:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:infoOut:MIN:" Min \\: %8.2lf %s" \ | |
- GPRINT:infoOut:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:infoOut:MAX:" Max \\: %8.2lf %s\\n" | |
report.ipunity.sip.statusCodeDetail.name=SIP Status Detail (IP Unity) | |
report.ipunity.sip.statusCodeDetail.columns=ipuSIPInfoClsIn,ipuSIPInfoClsOut,ipuSIPSuccClsIn,ipuSIPSuccClsOut,ipuSIPRedirClsIn,ipuSIPRedirClsOut,ipuSIPReqFailClsIn,ipuSIPReqFailClsOut,ipuSIPSrvFailClsIn,ipuSIPSrvFailClsOut,ipuSIPGblFailClsIn,ipuSIPGblFailClsOut | |
diff --git a/snmp-graph.properties.d/jboss-graph.properties b/snmp-graph.properties.d/jboss-graph.properties | |
index 0eb2b2b..ba63953 100644 | |
--- a/snmp-graph.properties.d/jboss-graph.properties | |
+++ b/snmp-graph.properties.d/jboss-graph.properties | |
@@ -119,7 +119,7 @@ report.jboss.grp.time.command=--title="HTTP Global Request Processor - Time" \ | |
LINE2:proc#0000ff:"ProcessTime" \ | |
GPRINT:proc:AVERAGE:" Avg \\: %6.2lf %s" \ | |
GPRINT:proc:MIN:"Min \\: %6.2lf %s" \ | |
- GPRINT:proc:MAX:"Max \\: %6.2lf %s\\n" \ | |
+ GPRINT:proc:MAX:"Max \\: %6.2lf %s\\n" | |
report.jboss.http.tp.name=Http Thread Pool | |
report.jboss.http.tp.columns=BusyThreads, Threads | |
diff --git a/snmp-graph.properties.d/juniper-graph.properties b/snmp-graph.properties.d/juniper-graph.properties | |
index 6d69707..6684f3d 100644 | |
--- a/snmp-graph.properties.d/juniper-graph.properties | |
+++ b/snmp-graph.properties.d/juniper-graph.properties | |
@@ -387,5 +387,5 @@ report.ive.connections.command=--title="Juniper IVE Users" \ | |
LINE2:iveConcurrentUsers#ff0000:"iveConcurrentUsers" \ | |
GPRINT:iveConcurrentUsers:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:iveConcurrentUsers:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:iveConcurrentUsers:MAX:"Max \\: %8.2lf %s\\n" | |
+ GPRINT:iveConcurrentUsers:MAX:"Max \\: %8.2lf %s\\n" | |
diff --git a/snmp-graph.properties.d/kafka-graph.properties b/snmp-graph.properties.d/kafka-graph.properties | |
new file mode 100644 | |
index 0000000..7ebfc6c | |
--- /dev/null | |
+++ b/snmp-graph.properties.d/kafka-graph.properties | |
@@ -0,0 +1,581 @@ | |
+reports=\ | |
+kafka.syslogInBps, \ | |
+kafka.syslogMps, \ | |
+kafka.syslogOutBps, \ | |
+kafka.trapsInBps, \ | |
+kafka.trapsMps, \ | |
+kafka.trapsOutBps, \ | |
+kafka.activeControllers, \ | |
+kafka.bytesInPerSec, \ | |
+kafka.bytesOutPerSec, \ | |
+kafka.isrExpandsPerSec, \ | |
+kafka.isrShrinksPerSec, \ | |
+kafka.leaderEps, \ | |
+kafka.leaders, \ | |
+kafka.localTimeConsumer, \ | |
+kafka.localTimeFollower, \ | |
+kafka.localTimeProduce, \ | |
+kafka.msgInPerSec, \ | |
+kafka.netProcAvgIdle, \ | |
+kafka.offlinePartitions, \ | |
+kafka.partitions, \ | |
+kafka.purgatoryFetch, \ | |
+kafka.purgatoryProduce, \ | |
+kafka.queueTimeConsumer, \ | |
+kafka.queueTimeFollower, \ | |
+kafka.queueTimeProduce, \ | |
+kafka.remoteTimeConsumer, \ | |
+kafka.remoteTimeFollower, \ | |
+kafka.remoteTimeProduce, \ | |
+kafka.replicaMaxLag, \ | |
+kafka.reqHandAvgIdle, \ | |
+kafka.reqSecConsumer, \ | |
+kafka.reqSecFollower, \ | |
+kafka.reqSecProduce, \ | |
+kafka.sendTimeConsumer, \ | |
+kafka.sendTimeFollower, \ | |
+kafka.sendTimeProduce, \ | |
+kafka.totalTimeConsumer, \ | |
+kafka.totalTimeFollower, \ | |
+kafka.totalTimeProduce, \ | |
+kafka.uncleanLeaderEps, \ | |
+kafka.underReplPart | |
+ | |
+ | |
+report.kafka.syslogInBps.name=Syslog Bytes In Per Second | |
+report.kafka.syslogInBps.columns=syslogInBps | |
+report.kafka.syslogInBps.type=interfaceSnmp | |
+report.kafka.syslogInBps.command=--title="Syslog Bytes In Per Second" \ | |
+ --vertical-label="Bytes per second" \ | |
+ DEF:value={rrd1}:syslogInBps:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Bytes In" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.syslogOutBps.name=Syslog Bytes Out Per Second | |
+report.kafka.syslogOutBps.columns=syslogOutBps | |
+report.kafka.syslogOutBps.type=interfaceSnmp | |
+report.kafka.syslogOutBps.command=--title="Syslog Bytes Out Per Second" \ | |
+ --vertical-label="Bytes per second" \ | |
+ DEF:value={rrd1}:syslogOutBps:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Bytes Out" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.syslogMps.name=Syslog Messages In Per Second | |
+report.kafka.syslogMps.columns=syslogMps | |
+report.kafka.syslogMps.type=interfaceSnmp | |
+report.kafka.syslogMps.command=--title="Syslog Messages In Per Second" \ | |
+ --vertical-label="Messages per second" \ | |
+ DEF:value={rrd1}:syslogMps:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Messages" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.trapsInBps.name=Trap Bytes In Per Second | |
+report.kafka.trapsInBps.columns=trapsInBps | |
+report.kafka.trapsInBps.type=interfaceSnmp | |
+report.kafka.trapsInBps.command=--title="Trap Bytes In Per Second" \ | |
+ --vertical-label="Bytes per second" \ | |
+ DEF:value={rrd1}:trapsInBps:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Bytes In" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.trapsOutBps.name=Trap Bytes Out Per Second | |
+report.kafka.trapsOutBps.columns=trapsOutBps | |
+report.kafka.trapsOutBps.type=interfaceSnmp | |
+report.kafka.trapsOutBps.command=--title="Trap Bytes Out Per Second" \ | |
+ --vertical-label="Bytes per second" \ | |
+ DEF:value={rrd1}:trapsOutBps:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Bytes Out" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.trapsMps.name=Trap Messages In Per Second | |
+report.kafka.trapsMps.columns=trapsMps | |
+report.kafka.trapsMps.type=interfaceSnmp | |
+report.kafka.trapsMps.command=--title="Trap Messages In Per Second" \ | |
+ --vertical-label="Messages per second" \ | |
+ DEF:value={rrd1}:trapsMps:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Messages" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+ | |
+report.kafka.msgInPerSec.name=Messages In Per Second | |
+report.kafka.msgInPerSec.columns=msgInPerSec | |
+report.kafka.msgInPerSec.type=interfaceSnmp | |
+report.kafka.msgInPerSec.command=--title="Messages In Per Second" \ | |
+ --vertical-label="Messages per second" \ | |
+ DEF:value={rrd1}:msgInPerSec:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Messages In Per Second" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.bytesInPerSec.name=Bytes In Per Second | |
+report.kafka.bytesInPerSec.columns=bytesInPerSec | |
+report.kafka.bytesInPerSec.type=interfaceSnmp | |
+report.kafka.bytesInPerSec.command=--title="Bytes In Per Second" \ | |
+ --vertical-label="Bytes per second" \ | |
+ DEF:value={rrd1}:bytesInPerSec:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Bytes In Per Second" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.bytesOutPerSec.name=Bytes Out Per Second | |
+report.kafka.bytesOutPerSec.columns=bytesOutPerSec | |
+report.kafka.bytesOutPerSec.type=interfaceSnmp | |
+report.kafka.bytesOutPerSec.command=--title="Bytes Out Per Second" \ | |
+ --vertical-label="Bytes per second" \ | |
+ DEF:value={rrd1}:bytesOutPerSec:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Bytes Out Per Second" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+report.kafka.underReplPart.name=Under-Replicated Partitions | |
+report.kafka.underReplPart.columns=underReplPart | |
+report.kafka.underReplPart.type=interfaceSnmp | |
+report.kafka.underReplPart.command=--title="Under-Replicated Partitions" \ | |
+ --vertical-label="Partitions" \ | |
+ DEF:value={rrd1}:underReplPart:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Under-Replicated Partitions" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+report.kafka.activeControllers.name=Active Controllers | |
+report.kafka.activeControllers.columns=activeControllers | |
+report.kafka.activeControllers.type=interfaceSnmp | |
+report.kafka.activeControllers.command=--title="Active Controllers" \ | |
+ --vertical-label="Controllers" \ | |
+ DEF:value={rrd1}:activeControllers:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Active Controllers" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.offlinePartitions.name=Offline Partitions | |
+report.kafka.offlinePartitions.columns=offlinePartitions | |
+report.kafka.offlinePartitions.type=interfaceSnmp | |
+report.kafka.offlinePartitions.command=--title="Offline Partitions" \ | |
+ --vertical-label="Partitions" \ | |
+ DEF:value={rrd1}:offlinePartitions:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Offline Partitions" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+report.kafka.leaderEps.name=Leader Election Rate and Time | |
+report.kafka.leaderEps.columns=leaderEps | |
+report.kafka.leaderEps.type=interfaceSnmp | |
+report.kafka.leaderEps.command=--title="Leader Election Rate and Time" \ | |
+ --vertical-label="Elections per second" \ | |
+ DEF:value={rrd1}:leaderEps:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Elections per second" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.uncleanLeaderEps.name=Unclean Leader Elections Per Second | |
+report.kafka.uncleanLeaderEps.columns=uncleanLeaderEps | |
+report.kafka.uncleanLeaderEps.type=interfaceSnmp | |
+report.kafka.uncleanLeaderEps.command=--title="Unclean Leader Elections Per Second" \ | |
+ --vertical-label="Elections per second" \ | |
+ DEF:value={rrd1}:uncleanLeaderEps:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Elections per second" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+report.kafka.partitions.name=Partitions | |
+report.kafka.partitions.columns=partitions | |
+report.kafka.partitions.type=interfaceSnmp | |
+report.kafka.partitions.command=--title="Partitions" \ | |
+ --vertical-label="Partitions" \ | |
+ DEF:value={rrd1}:partitions:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Partitions" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.leaders.name=Leaders | |
+report.kafka.leaders.columns=leaders | |
+report.kafka.leaders.type=interfaceSnmp | |
+report.kafka.leaders.command=--title="Leaders" \ | |
+ --vertical-label="Leaders" \ | |
+ DEF:value={rrd1}:leaders:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Leaders" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+report.kafka.isrShrinksPerSec.name=In-Sync Replica Shrinks Per Second | |
+report.kafka.isrShrinksPerSec.columns=isrShrinksPerSec | |
+report.kafka.isrShrinksPerSec.type=interfaceSnmp | |
+report.kafka.isrShrinksPerSec.command=--title="In-Sync Replica Shrinks Per Second" \ | |
+ --vertical-label="Shrinks per second" \ | |
+ DEF:value={rrd1}:isrShrinksPerSec:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Shrinks per second" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.isrExpandsPerSec.name=In-Sync Replica Expansions Per Second | |
+report.kafka.isrExpandsPerSec.columns=isrExpandsPerSec | |
+report.kafka.isrExpandsPerSec.type=interfaceSnmp | |
+report.kafka.isrExpandsPerSec.command=--title="In-Sync Replica Expansions Per Second" \ | |
+ --vertical-label="Expansions per second" \ | |
+ DEF:value={rrd1}:isrExpandsPerSec:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Expansions per second" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+report.kafka.replicaMaxLag.name=Maximum Lag Between Replicas | |
+report.kafka.replicaMaxLag.columns=replicaMaxLag | |
+report.kafka.replicaMaxLag.type=interfaceSnmp | |
+report.kafka.replicaMaxLag.command=--title="Maximum Lag Between Replicas" \ | |
+ --vertical-label="Messages" \ | |
+ DEF:value={rrd1}:replicaMaxLag:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Maximum Lag" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+report.kafka.purgatoryProduce.name=Purgatory Size: Produce | |
+report.kafka.purgatoryProduce.columns=purgatoryProduce | |
+report.kafka.purgatoryProduce.type=interfaceSnmp | |
+report.kafka.purgatoryProduce.command=--title="Purgatory Size: Produce" \ | |
+ --vertical-label="Requests" \ | |
+ DEF:value={rrd1}:purgatoryProduce:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Producer requests" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.purgatoryFetch.name=Purgatory Size: Fetch | |
+report.kafka.purgatoryFetch.columns=purgatoryFetch | |
+report.kafka.purgatoryFetch.type=interfaceSnmp | |
+report.kafka.purgatoryFetch.command=--title="Purgatory Size: Fetch" \ | |
+ --vertical-label="Requests" \ | |
+ DEF:value={rrd1}:purgatoryFetch:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Fetch requests" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+# NOTE: This report reverses the value from idle percentage to busy percentage | |
+# The original value is between 0 and 1. | |
+# | |
+report.kafka.netProcAvgIdle.name=Network Processor Average Busy Percentage | |
+report.kafka.netProcAvgIdle.columns=netProcAvgIdle | |
+report.kafka.netProcAvgIdle.type=interfaceSnmp | |
+report.kafka.netProcAvgIdle.command=--title="Network Processor Average Busy Percentage" \ | |
+ --vertical-label="Percent Busy" \ | |
+ DEF:value={rrd1}:netProcAvgIdle:AVERAGE \ | |
+ CDEF:percent=1,value,-,100,* \ | |
+ AREA:percent#edd400 \ | |
+ LINE2:percent#c4a000:"Percent Busy" \ | |
+ GPRINT:percent:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:percent:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:percent:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+# NOTE: This report reverses the value from idle percentage to busy percentage. | |
+# The original value is between 0 and 1. | |
+# | |
+report.kafka.reqHandAvgIdle.name=Request Handler Average Busy Percentage | |
+report.kafka.reqHandAvgIdle.columns=reqHandAvgIdle | |
+report.kafka.reqHandAvgIdle.type=interfaceSnmp | |
+report.kafka.reqHandAvgIdle.command=--title="Request Handler Average Busy Percentage" \ | |
+ --vertical-label="Percent Busy" \ | |
+ DEF:value={rrd1}:reqHandAvgIdle:AVERAGE \ | |
+ CDEF:percent=1,value,-,100,* \ | |
+ AREA:percent#edd400 \ | |
+ LINE2:percent#c4a000:"Percent Busy" \ | |
+ GPRINT:percent:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:percent:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:percent:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+# reqSec* | |
+report.kafka.reqSecProduce.name=Requests per Second: Produce | |
+report.kafka.reqSecProduce.columns=reqSecProduce | |
+report.kafka.reqSecProduce.type=interfaceSnmp | |
+report.kafka.reqSecProduce.command=--title="Requests per Second: Produce" \ | |
+ --vertical-label="Requests per second" \ | |
+ DEF:value={rrd1}:reqSecProduce:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Requests per Second" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.reqSecConsumer.name=Requests per Second: FetchConsumer | |
+report.kafka.reqSecConsumer.columns=reqSecConsumer | |
+report.kafka.reqSecConsumer.type=interfaceSnmp | |
+report.kafka.reqSecConsumer.command=--title="Requests per Second: FetchConsumer" \ | |
+ --vertical-label="Requests per second" \ | |
+ DEF:value={rrd1}:reqSecConsumer:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Requests per Second" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.reqSecFollower.name=Requests per Second: FetchFollower | |
+report.kafka.reqSecFollower.columns=reqSecFollower | |
+report.kafka.reqSecFollower.type=interfaceSnmp | |
+report.kafka.reqSecFollower.command=--title="Requests per Second: FetchFollower" \ | |
+ --vertical-label="Requests per second" \ | |
+ DEF:value={rrd1}:reqSecFollower:AVERAGE \ | |
+ AREA:value#edd400 \ | |
+ LINE2:value#c4a000:"Requests per Second" \ | |
+ GPRINT:value:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:value:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:value:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+# totalTime* | |
+report.kafka.totalTimeProduce.name=Total Time: Produce | |
+report.kafka.totalTimeProduce.columns=totalTimeProduce | |
+report.kafka.totalTimeProduce.type=interfaceSnmp | |
+report.kafka.totalTimeProduce.command=--title="Total Time: Produce" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:totalTimeProduce:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Total Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.totalTimeConsumer.name=Total Time: FetchConsumer | |
+report.kafka.totalTimeConsumer.columns=totalTimeConsumer | |
+report.kafka.totalTimeConsumer.type=interfaceSnmp | |
+report.kafka.totalTimeConsumer.command=--title="Total Time: FetchConsumer" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:totalTimeConsumer:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Total Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.totalTimeFollower.name=Total Time: FetchFollower | |
+report.kafka.totalTimeFollower.columns=totalTimeFollower | |
+report.kafka.totalTimeFollower.type=interfaceSnmp | |
+report.kafka.totalTimeFollower.command=--title="Total Time: FetchFollower" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:totalTimeFollower:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Total Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+# queueTime* | |
+report.kafka.queueTimeProduce.name=Queue Time: Produce | |
+report.kafka.queueTimeProduce.columns=queueTimeProduce | |
+report.kafka.queueTimeProduce.type=interfaceSnmp | |
+report.kafka.queueTimeProduce.command=--title="Queue Time: Produce" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:queueTimeProduce:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Queue Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.queueTimeConsumer.name=Queue Time: FetchConsumer | |
+report.kafka.queueTimeConsumer.columns=queueTimeConsumer | |
+report.kafka.queueTimeConsumer.type=interfaceSnmp | |
+report.kafka.queueTimeConsumer.command=--title="Queue Time: FetchConsumer" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:queueTimeConsumer:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Queue Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.queueTimeFollower.name=Queue Time: FetchFollower | |
+report.kafka.queueTimeFollower.columns=queueTimeFollower | |
+report.kafka.queueTimeFollower.type=interfaceSnmp | |
+report.kafka.queueTimeFollower.command=--title="Queue Time: FetchFollower" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:queueTimeFollower:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Queue Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+# localTime* | |
+report.kafka.localTimeProduce.name=Local Time: Produce | |
+report.kafka.localTimeProduce.columns=localTimeProduce | |
+report.kafka.localTimeProduce.type=interfaceSnmp | |
+report.kafka.localTimeProduce.command=--title="Local Time: Produce" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:localTimeProduce:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Local Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.localTimeConsumer.name=Local Time: FetchConsumer | |
+report.kafka.localTimeConsumer.columns=localTimeConsumer | |
+report.kafka.localTimeConsumer.type=interfaceSnmp | |
+report.kafka.localTimeConsumer.command=--title="Local Time: FetchConsumer" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:localTimeConsumer:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Local Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.localTimeFollower.name=Local Time: FetchFollower | |
+report.kafka.localTimeFollower.columns=localTimeFollower | |
+report.kafka.localTimeFollower.type=interfaceSnmp | |
+report.kafka.localTimeFollower.command=--title="Local Time: FetchFollower" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:localTimeFollower:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Local Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+# remoteTime* | |
+report.kafka.remoteTimeProduce.name=Remote Time: Produce | |
+report.kafka.remoteTimeProduce.columns=remoteTimeProduce | |
+report.kafka.remoteTimeProduce.type=interfaceSnmp | |
+report.kafka.remoteTimeProduce.command=--title="Remote Time: Produce" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:remoteTimeProduce:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Remote Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.remoteTimeConsumer.name=Remote Time: FetchConsumer | |
+report.kafka.remoteTimeConsumer.columns=remoteTimeConsumer | |
+report.kafka.remoteTimeConsumer.type=interfaceSnmp | |
+report.kafka.remoteTimeConsumer.command=--title="Remote Time: FetchConsumer" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:remoteTimeConsumer:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Remote Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.remoteTimeFollower.name=Remote Time: FetchFollower | |
+report.kafka.remoteTimeFollower.columns=remoteTimeFollower | |
+report.kafka.remoteTimeFollower.type=interfaceSnmp | |
+report.kafka.remoteTimeFollower.command=--title="Remote Time: FetchFollower" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:remoteTimeFollower:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Remote Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+# sendTime* | |
+report.kafka.sendTimeProduce.name=Response Send Time: Produce | |
+report.kafka.sendTimeProduce.columns=sendTimeProduce | |
+report.kafka.sendTimeProduce.type=interfaceSnmp | |
+report.kafka.sendTimeProduce.command=--title="Response Send Time: Produce" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:sendTimeProduce:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Response Send Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.sendTimeConsumer.name=Response Send Time: FetchConsumer | |
+report.kafka.sendTimeConsumer.columns=sendTimeConsumer | |
+report.kafka.sendTimeConsumer.type=interfaceSnmp | |
+report.kafka.sendTimeConsumer.command=--title="Response Send Time: FetchConsumer" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:sendTimeConsumer:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Response Send Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+report.kafka.sendTimeFollower.name=Response Send Time: FetchFollower | |
+report.kafka.sendTimeFollower.columns=sendTimeFollower | |
+report.kafka.sendTimeFollower.type=interfaceSnmp | |
+report.kafka.sendTimeFollower.command=--title="Response Send Time: FetchFollower" \ | |
+ --vertical-label="Seconds" \ | |
+ DEF:value={rrd1}:sendTimeFollower:AVERAGE \ | |
+ CDEF:seconds=value,1000,/ \ | |
+ AREA:seconds#edd400 \ | |
+ LINE2:seconds#c4a000:"Response Send Time" \ | |
+ GPRINT:seconds:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:seconds:MAX:" Max \\: %8.2lf %s\\n" | |
diff --git a/snmp-graph.properties.d/lmsensors-graph.properties b/snmp-graph.properties.d/lmsensors-graph.properties | |
index 36e9350..1fa6869 100644 | |
--- a/snmp-graph.properties.d/lmsensors-graph.properties | |
+++ b/snmp-graph.properties.d/lmsensors-graph.properties | |
@@ -20,7 +20,7 @@ report.lmsensors.temp.command=--title="Temperature on {lms-tempdevice}" \ | |
LINE1:btemp#f57900:"Temperature\\:" \ | |
GPRINT:btemp:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:btemp:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:btemp:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:btemp:MAX:"Max \\: %8.2lf %s\\n" | |
report.lmsensors.fan.name=lmSensors Fan Sensor | |
report.lmsensors.fan.columns=lms-fan | |
@@ -31,7 +31,7 @@ report.lmsensors.fan.command=--title="Fan Speed on {lms-fandevice}" \ | |
LINE2:dfan#0000ff:"Fan Speed\\:" \ | |
GPRINT:dfan:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:dfan:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:dfan:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:dfan:MAX:"Max \\: %8.2lf %s\\n" | |
report.lmsensors.volt.name=lmSensors Volt Sensor | |
report.lmsensors.volt.columns=lms-volt | |
@@ -43,5 +43,5 @@ report.lmsensors.volt.command=--title="Volt on {lms-voltdevice}" \ | |
LINE2:bvolt#0000ff:"Volt Speed\\:" \ | |
GPRINT:bvolt:AVERAGE:" Avg \\: %8.2lf %s" \ | |
GPRINT:bvolt:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:bvolt:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:bvolt:MAX:"Max \\: %8.2lf %s\\n" | |
diff --git a/snmp-graph.properties.d/microsoft-sql-graph.properties b/snmp-graph.properties.d/microsoft-sql-graph.properties | |
index 8651d39..f8af9d4 100644 | |
--- a/snmp-graph.properties.d/microsoft-sql-graph.properties | |
+++ b/snmp-graph.properties.d/microsoft-sql-graph.properties | |
@@ -71,7 +71,7 @@ report.mssqlhitratios.command=--title="MSSQL Hit Ratios" \ | |
LINE2:logcache#FF0000:"Log Cache \\:" \ | |
GPRINT:logcache:AVERAGE:"Avg\\:%4.1lf" \ | |
GPRINT:logcache:MAX:"Max\\:%4.1lf" \ | |
- GPRINT:logcache:MIN:"Min\\:%4.1lf" \ | |
+ GPRINT:logcache:MIN:"Min\\:%4.1lf" | |
report.mssqllockwaittime.name=MSSQL Lock Wait Time | |
report.mssqllockwaittime.columns=sqllockavgwaittime | |
diff --git a/snmp-graph.properties.d/mikrotik-graph.properties b/snmp-graph.properties.d/mikrotik-graph.properties | |
index 04c9332..8be5b49 100644 | |
--- a/snmp-graph.properties.d/mikrotik-graph.properties | |
+++ b/snmp-graph.properties.d/mikrotik-graph.properties | |
@@ -91,7 +91,7 @@ report.mikrotik.wlstatrssi.command=--title="Wireless Station RSSI" \ | |
AREA:rssi#00ff00:"SigLevel " \ | |
GPRINT:rssi:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:rssi:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:rssi:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:rssi:MAX:"Max \\: %8.2lf %s\\n" | |
report.mikrotik.wlrtabrssi.name=Mikrotik Remote Station Signal Level | |
report.mikrotik.wlrtabrssi.columns=mtxrWlRtabStrength | |
@@ -102,7 +102,7 @@ report.mikrotik.wlrtabrssi.command=--title="Wireless Station RSSI" \ | |
AREA:rssi#00ff00:"SigLevel " \ | |
GPRINT:rssi:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:rssi:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:rssi:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:rssi:MAX:"Max \\: %8.2lf %s\\n" | |
report.mikrotik.wlrtabbit.name=Mikrotik Remote Wls Link Rate | |
report.mikrotik.wlrtabbit.columns=mtxrWlRtabRxRate,mtxrWlRtabTxRate | |
diff --git a/snmp-graph.properties.d/mysql-graph.properties b/snmp-graph.properties.d/mysql-graph.properties | |
index 639098c..986de80 100644 | |
--- a/snmp-graph.properties.d/mysql-graph.properties | |
+++ b/snmp-graph.properties.d/mysql-graph.properties | |
@@ -116,7 +116,7 @@ report.mysql.slow.queries.command=--title="MySQL Slow Queries" \ | |
LINE2:slow#0000ff:"Slow Queries" \ | |
GPRINT:slow:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:slow:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:slow:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:slow:MAX:"Max \\: %8.2lf %s\\n" | |
report.mysql.queries.name=MySQL Queries | |
report.mysql.queries.columns=MyComDelete,MyComDeleteMulti,MyComInsert,MyComInsertSelect,MyComUpdate,MyComUpdateMulti,MyComSelect,MyQuestions | |
@@ -162,7 +162,7 @@ report.mysql.queries.command=--title="MySQL Queries" \ | |
LINE2:questions#000000:"Questions " \ | |
GPRINT:questions:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:questions:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:questions:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:questions:MAX:"Max \\: %8.2lf %s\\n" | |
report.mysql.key.reads.name=MySQL Key Reads | |
report.mysql.key.reads.columns=MyKeyReads,MyKeyReadReqs | |
@@ -178,7 +178,7 @@ report.mysql.key.reads.command=--title="MySQL Key Reads" \ | |
LINE2:readreqs#000000:"Read Requests " \ | |
GPRINT:readreqs:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:readreqs:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:readreqs:MAX:"Max \\: %8.2lf %s\\n" \ | |
+ GPRINT:readreqs:MAX:"Max \\: %8.2lf %s\\n" | |
report.mysql.key.writes.name=MySQL Key Writes | |
report.mysql.key.writes.columns=MyKeyWrites,MyKeyWriteReqs | |
diff --git a/snmp-graph.properties.d/netapp-graph.properties b/snmp-graph.properties.d/netapp-graph.properties | |
index f005c16..d498ea1 100644 | |
--- a/snmp-graph.properties.d/netapp-graph.properties | |
+++ b/snmp-graph.properties.d/netapp-graph.properties | |
@@ -232,12 +232,12 @@ report.netapp.sis.command=--title="NetApp {naDfFileSys} SIS percent savings" \ | |
GPRINT:naSisPercent:MAX:"Max \\: %8.2lf %s\\n" | |
report.netapp.diskio.name=NetApp Disk IO | |
-report.netapp.diskio.columns=naMiscLowDiskReadBytes, naMiscLowDiskWriteBytes | |
+report.netapp.diskio.columns=naMscLowDiskRdBytes,naMscLowDiskWrBytes | |
report.netapp.diskio.type=nodeSnmp | |
report.netapp.diskio.command=--title="NetApp Disk IO Bytes" \ | |
--vertical-label operations \ | |
- DEF:naMiscLowDiskReadBytes={rrd1}:naMiscLowDiskReadBy:AVERAGE \ | |
- DEF:naMiscLowDiskWriteBytes={rrd2}:naMiscLowDiskWriteB:AVERAGE \ | |
+ DEF:naMiscLowDiskReadBytes={rrd1}:naMscLowDiskRdBytes:AVERAGE \ | |
+ DEF:naMiscLowDiskWriteBytes={rrd2}:naMscLowDiskWrBytes:AVERAGE \ | |
CDEF:naMiscLowDiskWriteBytesNeg=0,naMiscLowDiskWriteBytes,- \ | |
LINE1:naMiscLowDiskReadBytes#0000ff:"IO reads Bytes" \ | |
GPRINT:naMiscLowDiskReadBytes:AVERAGE:" Avg \\: %8.2lf %s" \ | |
diff --git a/snmp-graph.properties.d/opennms-graph.properties b/snmp-graph.properties.d/opennms-graph.properties | |
index e915a2b..82356f9 100644 | |
--- a/snmp-graph.properties.d/opennms-graph.properties | |
+++ b/snmp-graph.properties.d/opennms-graph.properties | |
@@ -164,7 +164,7 @@ report.onms.collectd.threadpool.command=--title="OpenNMS Collectd ThreadPool" \ | |
LINE2:max#9A27F1:"Maximum Active" \ | |
GPRINT:max:AVERAGE:"Avg\\: %5.0lf\" \ | |
GPRINT:max:MIN:"Min\\: %5.0lf" \ | |
- GPRINT:max:MAX:"Max\\: %5.0lf\\n" \ | |
+ GPRINT:max:MAX:"Max\\: %5.0lf\\n" | |
report.onms.collectd.completedRatio.name=OpenNMS Collectd Task Completion Ratio | |
report.onms.collectd.completedRatio.columns=ONMSCollectTasksTot,ONMSCollectTasksCpt | |
diff --git a/snmp-graph.properties.d/opennms-minion-graph.properties b/snmp-graph.properties.d/opennms-minion-graph.properties | |
new file mode 100644 | |
index 0000000..b01b4af | |
--- /dev/null | |
+++ b/snmp-graph.properties.d/opennms-minion-graph.properties | |
@@ -0,0 +1,271 @@ | |
+reports=\ | |
+OpenNMS.Minion.RPC.Server.Detect.Exchanges, \ | |
+OpenNMS.Minion.RPC.Server.Detect.ProcessingTime, \ | |
+OpenNMS.Minion.RPC.Server.DNS.Exchanges, \ | |
+OpenNMS.Minion.RPC.Server.DNS.ProcessingTime, \ | |
+OpenNMS.Minion.RPC.Server.PING.Exchanges, \ | |
+OpenNMS.Minion.RPC.Server.PING.ProcessingTime, \ | |
+OpenNMS.Minion.RPC.Server.PING-SWEEP.Exchanges, \ | |
+OpenNMS.Minion.RPC.Server.PING-SWEEP.ProcessingTime, \ | |
+OpenNMS.Minion.RPC.Server.Poller.Exchanges, \ | |
+OpenNMS.Minion.RPC.Server.Poller.ProcessingTime, \ | |
+OpenNMS.Minion.RPC.Server.SNMP.Exchanges, \ | |
+OpenNMS.Minion.RPC.Server.SNMP.ProcessingTime, \ | |
+OpenNMS.Minion.Syslogd.Listener.Exchanges | |
+ | |
+ | |
+########################################### | |
+## OpenNMS.REPORT_NAME.Exchanges | |
+########################################### | |
+#report.REPORT_NAME.name=METRIC Exchanges | |
+#report.REPORT_NAME.columns=METRICComplete, METRICFailed | |
+#report.REPORT_NAME.type=interfaceSnmp | |
+#report.REPORT_NAME.command=--title="METRIC Exchanges" \ | |
+# --vertical-label="Messages per second" \ | |
+# DEF:complete={rrd1}:METRICComplete:AVERAGE \ | |
+# DEF:failed={rrd2}:METRICFailed:AVERAGE \ | |
+# AREA:failed#EF343B:"Failed Messages " \ | |
+# GPRINT:failed:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+# GPRINT:failed:MIN:" Min \\: %8.2lf %s" \ | |
+# GPRINT:failed:MAX:" Max \\: %8.2lf %s\\n" \ | |
+# STACK:complete#8DC63F:"Successful Messages" \ | |
+# GPRINT:complete:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+# GPRINT:complete:MIN:" Min \\: %8.2lf %s" \ | |
+# GPRINT:complete:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+########################################### | |
+## OpenNMS.Minion.RPC.Server.Detect.Exchanges | |
+########################################### | |
+report.OpenNMS.Minion.RPC.Server.Detect.Exchanges.name=Provisioning Detection Messages Received | |
+report.OpenNMS.Minion.RPC.Server.Detect.Exchanges.columns=DetectComplete, DetectFailed | |
+report.OpenNMS.Minion.RPC.Server.Detect.Exchanges.type=interfaceSnmp | |
+report.OpenNMS.Minion.RPC.Server.Detect.Exchanges.command=--title="Provisioning Detection Messages Received" \ | |
+ --vertical-label="Messages per second" \ | |
+ DEF:complete={rrd1}:DetectComplete:AVERAGE \ | |
+ DEF:failed={rrd2}:DetectFailed:AVERAGE \ | |
+ AREA:failed#EF343B:"Failed Messages " \ | |
+ GPRINT:failed:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:failed:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:failed:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ STACK:complete#8DC63F:"Successful Messages" \ | |
+ GPRINT:complete:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:complete:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:complete:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+########################################### | |
+## OpenNMS.Minion.RPC.Server.Detect.ProcessingTime | |
+########################################### | |
+report.OpenNMS.Minion.RPC.Server.Detect.ProcessingTime.name=Provisioning Detection Processing Time | |
+report.OpenNMS.Minion.RPC.Server.Detect.ProcessingTime.columns=DetectLastProc, DetectMeanProc | |
+report.OpenNMS.Minion.RPC.Server.Detect.ProcessingTime.type=interfaceSnmp | |
+report.OpenNMS.Minion.RPC.Server.Detect.ProcessingTime.command=--title="Provisioning Detection Processing Time" \ | |
+ --vertical-label="Seconds per message" \ | |
+ DEF:mqLast={rrd1}:DetectLastProc:AVERAGE \ | |
+ DEF:mqMean={rrd2}:DetectMeanProc:AVERAGE \ | |
+ CDEF:mqLastSec=mqLast,1000,/ \ | |
+ AREA:mqLastSec#73d216:"Process via JMS" \ | |
+ GPRINT:mqLastSec:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:mqLastSec:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:mqLastSec:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+########################################### | |
+## OpenNMS.Minion.RPC.Server.DNS.Exchanges | |
+########################################### | |
+report.OpenNMS.Minion.RPC.Server.DNS.Exchanges.name=DNS Messages Received | |
+report.OpenNMS.Minion.RPC.Server.DNS.Exchanges.columns=DnsComplete, DnsFailed | |
+report.OpenNMS.Minion.RPC.Server.DNS.Exchanges.type=interfaceSnmp | |
+report.OpenNMS.Minion.RPC.Server.DNS.Exchanges.command=--title="DNS Messages Received" \ | |
+ --vertical-label="Messages per second" \ | |
+ DEF:complete={rrd1}:DnsComplete:AVERAGE \ | |
+ DEF:failed={rrd2}:DnsFailed:AVERAGE \ | |
+ AREA:failed#EF343B:"Failed Messages " \ | |
+ GPRINT:failed:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:failed:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:failed:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ STACK:complete#8DC63F:"Successful Messages" \ | |
+ GPRINT:complete:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:complete:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:complete:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+########################################### | |
+## OpenNMS.Minion.RPC.Server.DNS.ProcessingTime | |
+########################################### | |
+report.OpenNMS.Minion.RPC.Server.DNS.ProcessingTime.name=DNS Processing Time | |
+report.OpenNMS.Minion.RPC.Server.DNS.ProcessingTime.columns=DnsLastProc, DnsMeanProc | |
+report.OpenNMS.Minion.RPC.Server.DNS.ProcessingTime.type=interfaceSnmp | |
+report.OpenNMS.Minion.RPC.Server.DNS.ProcessingTime.command=--title="DNS Processing Time" \ | |
+ --vertical-label="Seconds per message" \ | |
+ DEF:mqLast={rrd1}:DnsLastProc:AVERAGE \ | |
+ DEF:mqMean={rrd2}:DnsMeanProc:AVERAGE \ | |
+ CDEF:mqLastSec=mqLast,1000,/ \ | |
+ AREA:mqLastSec#73d216:"Process via JMS" \ | |
+ GPRINT:mqLastSec:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:mqLastSec:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:mqLastSec:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+########################################### | |
+## OpenNMS.Minion.RPC.Server.PING.Exchanges | |
+########################################### | |
+report.OpenNMS.Minion.RPC.Server.PING.Exchanges.name=Ping Messages Received | |
+report.OpenNMS.Minion.RPC.Server.PING.Exchanges.columns=PingComplete, PingFailed | |
+report.OpenNMS.Minion.RPC.Server.PING.Exchanges.type=interfaceSnmp | |
+report.OpenNMS.Minion.RPC.Server.PING.Exchanges.command=--title="Ping Messages Received" \ | |
+ --vertical-label="Messages per second" \ | |
+ DEF:complete={rrd1}:PingComplete:AVERAGE \ | |
+ DEF:failed={rrd2}:PingFailed:AVERAGE \ | |
+ AREA:failed#EF343B:"Failed Messages " \ | |
+ GPRINT:failed:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:failed:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:failed:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ STACK:complete#8DC63F:"Successful Messages" \ | |
+ GPRINT:complete:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:complete:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:complete:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+########################################### | |
+## OpenNMS.Minion.RPC.Server.PING.ProcessingTime | |
+########################################### | |
+report.OpenNMS.Minion.RPC.Server.PING.ProcessingTime.name=Ping Processing Time | |
+report.OpenNMS.Minion.RPC.Server.PING.ProcessingTime.columns=PingLastProc, PingMeanProc | |
+report.OpenNMS.Minion.RPC.Server.PING.ProcessingTime.type=interfaceSnmp | |
+report.OpenNMS.Minion.RPC.Server.PING.ProcessingTime.command=--title="Ping Processing Time" \ | |
+ --vertical-label="Seconds per message" \ | |
+ DEF:mqLast={rrd1}:PingLastProc:AVERAGE \ | |
+ DEF:mqMean={rrd2}:PingMeanProc:AVERAGE \ | |
+ CDEF:mqLastSec=mqLast,1000,/ \ | |
+ AREA:mqLastSec#73d216:"Process via JMS" \ | |
+ GPRINT:mqLastSec:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:mqLastSec:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:mqLastSec:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+########################################### | |
+## OpenNMS.Minion.RPC.Server.PING-SWEEP.Exchanges | |
+########################################### | |
+report.OpenNMS.Minion.RPC.Server.PING-SWEEP.Exchanges.name=Ping Sweep Messages Received | |
+report.OpenNMS.Minion.RPC.Server.PING-SWEEP.Exchanges.columns=SweepComplete, SweepFailed | |
+report.OpenNMS.Minion.RPC.Server.PING-SWEEP.Exchanges.type=interfaceSnmp | |
+report.OpenNMS.Minion.RPC.Server.PING-SWEEP.Exchanges.command=--title="Ping Sweep Messages Received" \ | |
+ --vertical-label="Messages per second" \ | |
+ DEF:complete={rrd1}:SweepComplete:AVERAGE \ | |
+ DEF:failed={rrd2}:SweepFailed:AVERAGE \ | |
+ AREA:failed#EF343B:"Failed Messages " \ | |
+ GPRINT:failed:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:failed:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:failed:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ STACK:complete#8DC63F:"Successful Messages" \ | |
+ GPRINT:complete:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:complete:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:complete:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+########################################### | |
+## OpenNMS.Minion.RPC.Server.PING-SWEEP.ProcessingTime | |
+########################################### | |
+report.OpenNMS.Minion.RPC.Server.PING-SWEEP.ProcessingTime.name=Ping Sweep Processing Time | |
+report.OpenNMS.Minion.RPC.Server.PING-SWEEP.ProcessingTime.columns=SweepLastProc, SweepMeanProc | |
+report.OpenNMS.Minion.RPC.Server.PING-SWEEP.ProcessingTime.type=interfaceSnmp | |
+report.OpenNMS.Minion.RPC.Server.PING-SWEEP.ProcessingTime.command=--title="Ping Sweep Processing Time" \ | |
+ --vertical-label="Seconds per message" \ | |
+ DEF:mqLast={rrd1}:SweepLastProc:AVERAGE \ | |
+ DEF:mqMean={rrd2}:SweepMeanProc:AVERAGE \ | |
+ CDEF:mqLastSec=mqLast,1000,/ \ | |
+ AREA:mqLastSec#73d216:"Process via JMS" \ | |
+ GPRINT:mqLastSec:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:mqLastSec:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:mqLastSec:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+########################################### | |
+## OpenNMS.Minion.RPC.Server.Poller.Exchanges | |
+########################################### | |
+report.OpenNMS.Minion.RPC.Server.Poller.Exchanges.name=Poller Monitor Messages Received | |
+report.OpenNMS.Minion.RPC.Server.Poller.Exchanges.columns=PollComplete, PollFailed | |
+report.OpenNMS.Minion.RPC.Server.Poller.Exchanges.type=interfaceSnmp | |
+report.OpenNMS.Minion.RPC.Server.Poller.Exchanges.command=--title="Poller Monitor Messages Received" \ | |
+ --vertical-label="Messages per second" \ | |
+ DEF:complete={rrd1}:PollComplete:AVERAGE \ | |
+ DEF:failed={rrd2}:PollFailed:AVERAGE \ | |
+ AREA:failed#EF343B:"Failed Messages " \ | |
+ GPRINT:failed:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:failed:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:failed:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ STACK:complete#8DC63F:"Successful Messages" \ | |
+ GPRINT:complete:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:complete:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:complete:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+########################################### | |
+## OpenNMS.Minion.RPC.Server.Poller.ProcessingTime | |
+########################################### | |
+report.OpenNMS.Minion.RPC.Server.Poller.ProcessingTime.name=Poller Monitor Processing Time | |
+report.OpenNMS.Minion.RPC.Server.Poller.ProcessingTime.columns=PollLastProc, PollMeanProc | |
+report.OpenNMS.Minion.RPC.Server.Poller.ProcessingTime.type=interfaceSnmp | |
+report.OpenNMS.Minion.RPC.Server.Poller.ProcessingTime.command=--title="Poller Monitor Processing Time" \ | |
+ --vertical-label="Seconds per message" \ | |
+ DEF:mqLast={rrd1}:PollLastProc:AVERAGE \ | |
+ DEF:mqMean={rrd2}:PollMeanProc:AVERAGE \ | |
+ CDEF:mqLastSec=mqLast,1000,/ \ | |
+ AREA:mqLastSec#73d216:"Process via JMS" \ | |
+ GPRINT:mqLastSec:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:mqLastSec:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:mqLastSec:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+########################################### | |
+## OpenNMS.Minion.RPC.Server.SNMP.Exchanges | |
+########################################### | |
+report.OpenNMS.Minion.RPC.Server.SNMP.Exchanges.name=SNMP Messages Received | |
+report.OpenNMS.Minion.RPC.Server.SNMP.Exchanges.columns=SnmpComplete, SnmpFailed | |
+report.OpenNMS.Minion.RPC.Server.SNMP.Exchanges.type=interfaceSnmp | |
+report.OpenNMS.Minion.RPC.Server.SNMP.Exchanges.command=--title="SNMP Messages Received" \ | |
+ --vertical-label="Messages per second" \ | |
+ DEF:complete={rrd1}:SnmpComplete:AVERAGE \ | |
+ DEF:failed={rrd2}:SnmpFailed:AVERAGE \ | |
+ AREA:failed#EF343B:"Failed Messages " \ | |
+ GPRINT:failed:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:failed:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:failed:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ STACK:complete#8DC63F:"Successful Messages" \ | |
+ GPRINT:complete:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:complete:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:complete:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+########################################### | |
+## OpenNMS.Minion.RPC.Server.SNMP.ProcessingTime | |
+########################################### | |
+report.OpenNMS.Minion.RPC.Server.SNMP.ProcessingTime.name=SNMP Processing Time | |
+report.OpenNMS.Minion.RPC.Server.SNMP.ProcessingTime.columns=SnmpLastProc, SnmpMeanProc | |
+report.OpenNMS.Minion.RPC.Server.SNMP.ProcessingTime.type=interfaceSnmp | |
+report.OpenNMS.Minion.RPC.Server.SNMP.ProcessingTime.command=--title="SNMP Processing Time" \ | |
+ --vertical-label="Seconds per message" \ | |
+ DEF:mqLast={rrd1}:SnmpLastProc:AVERAGE \ | |
+ DEF:mqMean={rrd2}:SnmpMeanProc:AVERAGE \ | |
+ CDEF:mqLastSec=mqLast,1000,/ \ | |
+ AREA:mqLastSec#73d216:"Process via JMS" \ | |
+ GPRINT:mqLastSec:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:mqLastSec:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:mqLastSec:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
+ | |
+########################################### | |
+## OpenNMS.Minion.Syslogd.Listener.Exchanges | |
+########################################### | |
+report.OpenNMS.Minion.Syslogd.Listener.Exchanges.name=Syslog Messages Received | |
+report.OpenNMS.Minion.Syslogd.Listener.Exchanges.columns=SlogListComplete, SlogListFailed | |
+report.OpenNMS.Minion.Syslogd.Listener.Exchanges.type=interfaceSnmp | |
+report.OpenNMS.Minion.Syslogd.Listener.Exchanges.command=--title="Syslog Messages Received" \ | |
+ --vertical-label="Messages per second" \ | |
+ DEF:complete={rrd1}:SlogListComplete:AVERAGE \ | |
+ DEF:failed={rrd2}:SlogListFailed:AVERAGE \ | |
+ AREA:failed#EF343B:"Failed Messages " \ | |
+ GPRINT:failed:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:failed:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:failed:MAX:" Max \\: %8.2lf %s\\n" \ | |
+ STACK:complete#8DC63F:"Successful Messages" \ | |
+ GPRINT:complete:AVERAGE:" Avg \\: %8.2lf %s" \ | |
+ GPRINT:complete:MIN:" Min \\: %8.2lf %s" \ | |
+ GPRINT:complete:MAX:" Max \\: %8.2lf %s\\n" | |
+ | |
diff --git a/snmp-graph.properties.d/paloalto-graph.properties b/snmp-graph.properties.d/paloalto-graph.properties | |
index 9da2e5e..a95017c 100644 | |
--- a/snmp-graph.properties.d/paloalto-graph.properties | |
+++ b/snmp-graph.properties.d/paloalto-graph.properties | |
@@ -32,7 +32,7 @@ report.paloalto.envt.command=--title="Environmental Status of {panentPhysicalNam | |
COMMENT:"\\n" \ | |
GPRINT:currentSessions:AVERAGE:"Avg \\: %1.0lf" \ | |
GPRINT:currentSessions:MIN:"Min \\: %1.0lf" \ | |
- GPRINT:currentSessions:MAX:"Max \\: %1.0lf \\n" | |
+ GPRINT:currentSessions:MAX:"Max \\: %1.0lf \\n" | |
report.paloalto.session.util.name=Chassis Sessions (Palo Alto) | |
report.paloalto.session.util.columns=panSessionUtil | |
@@ -153,7 +153,7 @@ report.paloalto.lc.vdisk.command=--title="Log Collector Disk Usage" \ | |
LINE2:vdisk4MB#f0000f:"Disk 4" \ | |
GPRINT:vdisk4MB:AVERAGE:"Avg \\: %1.0lf" \ | |
GPRINT:vdisk4MB:MIN:"Min \\: %1.0lf" \ | |
- GPRINT:vdisk4MB:MAX:"Max \\: %1.0lf \\n" | |
+ GPRINT:vdisk4MB:MAX:"Max \\: %1.0lf \\n" | |
report.paloalto.lc.age.name=PaloAlto Log Collector Log Age | |
diff --git a/snmp-graph.properties.d/postgresql-graph.properties b/snmp-graph.properties.d/postgresql-graph.properties | |
index 2f328a3..6dde90e 100644 | |
--- a/snmp-graph.properties.d/postgresql-graph.properties | |
+++ b/snmp-graph.properties.d/postgresql-graph.properties | |
@@ -60,7 +60,7 @@ report.pgsql.dbsize.command=--title="PostgreSQL DB Size - {datname}" \ | |
AREA:size#7EE600:"DB Size" \ | |
GPRINT:size:AVERAGE:" Avg\\: %8.2lf %s" \ | |
GPRINT:size:MIN:"Min\\: %8.2lf %s" \ | |
- GPRINT:size:MAX:"Max\\: %8.2lf %s\\n" | |
+ GPRINT:size:MAX:"Max\\: %8.2lf %s\\n" | |
report.pgsql.dbbackends.name=PostgreSQL DB Backends | |
report.pgsql.dbbackends.type=pgDatabase | |
@@ -76,7 +76,7 @@ report.pgsql.dbbackends.command=--title="PostgreSQL DB Backends - {datname}" \ | |
AREA:size#7EE600:"Backends" \ | |
GPRINT:size:AVERAGE:" Avg\\: %8.2lf %s" \ | |
GPRINT:size:MIN:"Min\\: %8.2lf %s" \ | |
- GPRINT:size:MAX:"Max\\: %8.2lf %s\\n" | |
+ GPRINT:size:MAX:"Max\\: %8.2lf %s\\n" | |
report.pgsql.tssize.name=PostgreSQL Tablespace Size | |
@@ -93,5 +93,5 @@ report.pgsql.tssize.command=--title="PostgreSQL Tablespace Size - {spcname}" \ | |
AREA:size#7EE600:"Tablespace Size" \ | |
GPRINT:size:AVERAGE:" Avg\\: %8.2lf %s" \ | |
GPRINT:size:MIN:"Min\\: %8.2lf %s" \ | |
- GPRINT:size:MAX:"Max\\: %8.2lf %s\\n" | |
+ GPRINT:size:MAX:"Max\\: %8.2lf %s\\n" | |
diff --git a/snmp-graph.properties.d/riverbed-steelhead-graph.properties b/snmp-graph.properties.d/riverbed-steelhead-graph.properties | |
index 0d3ee1b..cad0ca7 100644 | |
--- a/snmp-graph.properties.d/riverbed-steelhead-graph.properties | |
+++ b/snmp-graph.properties.d/riverbed-steelhead-graph.properties | |
@@ -260,7 +260,7 @@ report.riverbed.steelhead.cpuStats.command=--title="Riverbed Steelhead CPU Stats | |
STACK:idle#a0ffa0:"Idle " \ | |
GPRINT:idle:AVERAGE:"Avg \\: %8.2lf" \ | |
GPRINT:idle:MIN:"Min \\: %8.2lf" \ | |
- GPRINT:idle:MAX:"Max \\: %8.2lf" \ | |
+ GPRINT:idle:MAX:"Max \\: %8.2lf" | |
report.riverbed.steelhead.portBandwidth.name=Riverbed Steelhead Port Bandwidth | |
report.riverbed.steelhead.portBandwidth.columns=rbshBwPortInLan,rbshBwPortInWan,rbshBwPortOutLan,rbshBwPortOutWan | |
diff --git a/snmp-graph.properties.d/sofaware-embeddedngx-graph.properties b/snmp-graph.properties.d/sofaware-embeddedngx-graph.properties | |
index bc6c42b..c6f78b9 100644 | |
--- a/snmp-graph.properties.d/sofaware-embeddedngx-graph.properties | |
+++ b/snmp-graph.properties.d/sofaware-embeddedngx-graph.properties | |
@@ -78,7 +78,7 @@ report.sofaware.embeddedngx.storageFirm.command=--title="SofaWare Embedded NGX F | |
LINE2:firmTot#000000:"Total" \ | |
GPRINT:firmTot:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:firmTot:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:firmTot:MAX:"Max \\: %8.2lf %s" | |
+ GPRINT:firmTot:MAX:"Max \\: %8.2lf %s" | |
report.sofaware.embeddedngx.storageCF.name=SofaWare Embedded NGX CF Storage | |
report.sofaware.embeddedngx.storageCF.columns=swStorageCFTot,swStorageCFFree | |
@@ -96,7 +96,7 @@ report.sofaware.embeddedngx.storageCF.command=--title="SofaWare Embedded NGX CF | |
LINE2:cfTot#000000:"Total" \ | |
GPRINT:cfTot:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:cfTot:MIN:"Min \\: %8.2lf %s" \ | |
- GPRINT:cfTot:MAX:"Max \\: %8.2lf %s" | |
+ GPRINT:cfTot:MAX:"Max \\: %8.2lf %s" | |
report.sofaware.embeddedngx.licenses.name=SofaWare Embedded NGX Licenses | |
report.sofaware.embeddedngx.licenses.columns=swLicenseUsedNodes | |
diff --git a/snmp-graph.properties.d/vmware-cim-graph-simple.properties b/snmp-graph.properties.d/vmware-cim-graph-simple.properties | |
index 6c0b8bd..1e1a4c5 100644 | |
--- a/snmp-graph.properties.d/vmware-cim-graph-simple.properties | |
+++ b/snmp-graph.properties.d/vmware-cim-graph-simple.properties | |
@@ -25,7 +25,7 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorOther.name=SensorOther | |
report.vmwareCim.SensorOther.columns=CurrentReading | |
@@ -37,7 +37,7 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorTemperature.name=SensorTemperature | |
report.vmwareCim.SensorTemperature.columns=CurrentReading | |
@@ -49,7 +49,7 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorVoltage.name=SensorVoltage | |
report.vmwareCim.SensorVoltage.columns=CurrentReading | |
@@ -61,7 +61,7 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorCurrent.name=SensorCurrent | |
report.vmwareCim.SensorCurrent.columns=CurrentReading | |
@@ -73,7 +73,7 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorCounter.name=SensorCounter | |
report.vmwareCim.SensorCounter.columns=CurrentReading | |
@@ -87,7 +87,7 @@ AREA:xxx#729fcf \ | |
LINE1:xxx#3465a4:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorTachometer.name=SensorTachometer | |
report.vmwareCim.SensorTachometer.columns=CurrentReading | |
@@ -99,7 +99,7 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorSwitch.name=SensorSwitch | |
report.vmwareCim.SensorSwitch.columns=CurrentReading | |
@@ -111,7 +111,7 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorLock.name=SensorLock | |
report.vmwareCim.SensorLock.columns=CurrentReading | |
@@ -123,7 +123,7 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorHumidity.name=SensorHumidity | |
report.vmwareCim.SensorHumidity.columns=CurrentReading | |
@@ -135,7 +135,7 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorSmokeDetection.name=SensorSmokeDetection | |
report.vmwareCim.SensorSmokeDetection.columns=CurrentReading | |
@@ -147,7 +147,7 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorPresence.name=SensorPresence | |
report.vmwareCim.SensorPresence.columns=CurrentReading | |
@@ -159,7 +159,7 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorAirFlow.name=SensorAirFlow | |
report.vmwareCim.SensorAirFlow.columns=CurrentReading | |
@@ -171,7 +171,7 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorPowerConsumption.name=SensorPowerConsumption | |
report.vmwareCim.SensorPowerConsumption.columns=CurrentReading | |
@@ -183,7 +183,7 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorPowerProduction.name=SensorPowerProduction | |
report.vmwareCim.SensorPowerProduction.columns=CurrentReading | |
@@ -195,7 +195,7 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmwareCim.SensorPressure.name=SensorPressure | |
report.vmwareCim.SensorPressure.columns=CurrentReading | |
@@ -207,5 +207,5 @@ DEF:xxx={rrd1}:CurrentReading:AVERAGE \ | |
LINE2:xxx#0000ff:"CurrentReading" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
diff --git a/snmp-graph.properties.d/vmware3-graph-simple.properties b/snmp-graph.properties.d/vmware3-graph-simple.properties | |
index f99b3a8..d8565a8 100644 | |
--- a/snmp-graph.properties.d/vmware3-graph-simple.properties | |
+++ b/snmp-graph.properties.d/vmware3-graph-simple.properties | |
@@ -136,7 +136,7 @@ DEF:xxx={rrd1}:CpuIdleSum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuIdleSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.CpuUsageNon.name=CpuUsageNon | |
report.vmware3.CpuUsageNon.columns=CpuUsageNon | |
@@ -148,7 +148,7 @@ DEF:xxx={rrd1}:CpuUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.CpuUsedSum.name=CpuUsedSum | |
report.vmware3.CpuUsedSum.columns=CpuUsedSum | |
@@ -160,7 +160,7 @@ DEF:xxx={rrd1}:CpuUsedSum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsedSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.CpuRdCyAvg.name=CpuRdCyAvg | |
report.vmware3.CpuRdCyAvg.columns=CpuRdCyAvg | |
@@ -171,7 +171,7 @@ DEF:xxx={rrd1}:CpuRdCyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuRdCyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.CpuUsageNon.name=CpuUsageNon | |
report.vmware3.CpuUsageNon.columns=CpuUsageNon | |
@@ -182,7 +182,7 @@ DEF:xxx={rrd1}:CpuUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.CpuUsagemhzNon.name=CpuUsagemhzNon | |
report.vmware3.CpuUsagemhzNon.columns=CpuUsagemhzNon | |
@@ -193,7 +193,7 @@ DEF:xxx={rrd1}:CpuUsagemhzNon:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsagemhzNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskUsageNon.name=DiskUsageNon | |
report.vmware3.DiskUsageNon.columns=DiskUsageNon | |
@@ -204,7 +204,7 @@ DEF:xxx={rrd1}:DiskUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemActiveNon.name=MemActiveNon | |
report.vmware3.MemActiveNon.columns=MemActiveNon | |
@@ -215,7 +215,7 @@ DEF:xxx={rrd1}:MemActiveNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemActiveNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemConsumedNon.name=MemConsumedNon | |
report.vmware3.MemConsumedNon.columns=MemConsumedNon | |
@@ -226,7 +226,7 @@ DEF:xxx={rrd1}:MemConsumedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemConsumedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemGrantedNon.name=MemGrantedNon | |
report.vmware3.MemGrantedNon.columns=MemGrantedNon | |
@@ -237,7 +237,7 @@ DEF:xxx={rrd1}:MemGrantedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemGrantedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemHeapNon.name=MemHeapNon | |
report.vmware3.MemHeapNon.columns=MemHeapNon | |
@@ -248,7 +248,7 @@ DEF:xxx={rrd1}:MemHeapNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemHeapNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemHeapfreeNon.name=MemHeapfreeNon | |
report.vmware3.MemHeapfreeNon.columns=MemHeapfreeNon | |
@@ -259,7 +259,7 @@ DEF:xxx={rrd1}:MemHeapfreeNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemHeapfreeNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemOdNon.name=MemOdNon | |
report.vmware3.MemOdNon.columns=MemOdNon | |
@@ -270,7 +270,7 @@ DEF:xxx={rrd1}:MemOdNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemOdNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemRdCyAvg.name=MemRdCyAvg | |
report.vmware3.MemRdCyAvg.columns=MemRdCyAvg | |
@@ -281,7 +281,7 @@ DEF:xxx={rrd1}:MemRdCyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"MemRdCyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemSharedNon.name=MemSharedNon | |
report.vmware3.MemSharedNon.columns=MemSharedNon | |
@@ -292,7 +292,7 @@ DEF:xxx={rrd1}:MemSharedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSharedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemSharedcommonNon.name=MemSharedcommonNon | |
report.vmware3.MemSharedcommonNon.columns=MemSharedcommonNon | |
@@ -303,7 +303,7 @@ DEF:xxx={rrd1}:MemSharedcommonNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSharedcommonNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemSpinNon.name=MemSpinNon | |
report.vmware3.MemSpinNon.columns=MemSpinNon | |
@@ -314,7 +314,7 @@ DEF:xxx={rrd1}:MemSpinNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpinNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemSpoutNon.name=MemSpoutNon | |
report.vmware3.MemSpoutNon.columns=MemSpoutNon | |
@@ -325,7 +325,7 @@ DEF:xxx={rrd1}:MemSpoutNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpoutNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemSpusedNon.name=MemSpusedNon | |
report.vmware3.MemSpusedNon.columns=MemSpusedNon | |
@@ -336,7 +336,7 @@ DEF:xxx={rrd1}:MemSpusedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpusedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemStateLat.name=MemStateLat | |
report.vmware3.MemStateLat.columns=MemStateLat | |
@@ -347,7 +347,7 @@ DEF:xxx={rrd1}:MemStateLat:AVERAGE \ | |
LINE2:xxx#0000ff:"MemStateLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemSysUsageNon.name=MemSysUsageNon | |
report.vmware3.MemSysUsageNon.columns=MemSysUsageNon | |
@@ -358,7 +358,7 @@ DEF:xxx={rrd1}:MemSysUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSysUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemUdNon.name=MemUdNon | |
report.vmware3.MemUdNon.columns=MemUdNon | |
@@ -369,7 +369,7 @@ DEF:xxx={rrd1}:MemUdNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemUdNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemUsageNon.name=MemUsageNon | |
report.vmware3.MemUsageNon.columns=MemUsageNon | |
@@ -380,7 +380,7 @@ DEF:xxx={rrd1}:MemUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemVmmemctlNon.name=MemVmmemctlNon | |
report.vmware3.MemVmmemctlNon.columns=MemVmmemctlNon | |
@@ -391,7 +391,7 @@ DEF:xxx={rrd1}:MemVmmemctlNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemVmmemctlNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemZeroNon.name=MemZeroNon | |
report.vmware3.MemZeroNon.columns=MemZeroNon | |
@@ -402,7 +402,7 @@ DEF:xxx={rrd1}:MemZeroNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemZeroNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.NetUsageNon.name=NetUsageNon | |
report.vmware3.NetUsageNon.columns=NetUsageNon | |
@@ -413,7 +413,7 @@ DEF:xxx={rrd1}:NetUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"NetUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuActav15Lat.name=ResCpuActav15Lat | |
report.vmware3.ResCpuActav15Lat.columns=ResCpuActav15Lat | |
@@ -424,7 +424,7 @@ DEF:xxx={rrd1}:ResCpuActav15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActav15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuActav1Lat.name=ResCpuActav1Lat | |
report.vmware3.ResCpuActav1Lat.columns=ResCpuActav1Lat | |
@@ -435,7 +435,7 @@ DEF:xxx={rrd1}:ResCpuActav1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActav1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuActav5Lat.name=ResCpuActav5Lat | |
report.vmware3.ResCpuActav5Lat.columns=ResCpuActav5Lat | |
@@ -446,7 +446,7 @@ DEF:xxx={rrd1}:ResCpuActav5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActav5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuActpk15Lat.name=ResCpuActpk15Lat | |
report.vmware3.ResCpuActpk15Lat.columns=ResCpuActpk15Lat | |
@@ -457,7 +457,7 @@ DEF:xxx={rrd1}:ResCpuActpk15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActpk15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuActpk1Lat.name=ResCpuActpk1Lat | |
report.vmware3.ResCpuActpk1Lat.columns=ResCpuActpk1Lat | |
@@ -468,7 +468,7 @@ DEF:xxx={rrd1}:ResCpuActpk1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActpk1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuActpk5Lat.name=ResCpuActpk5Lat | |
report.vmware3.ResCpuActpk5Lat.columns=ResCpuActpk5Lat | |
@@ -479,7 +479,7 @@ DEF:xxx={rrd1}:ResCpuActpk5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActpk5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuMaxLd15Lat.name=ResCpuMaxLd15Lat | |
report.vmware3.ResCpuMaxLd15Lat.columns=ResCpuMaxLd15Lat | |
@@ -490,7 +490,7 @@ DEF:xxx={rrd1}:ResCpuMaxLd15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuMaxLd15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuMaxLd1Lat.name=ResCpuMaxLd1Lat | |
report.vmware3.ResCpuMaxLd1Lat.columns=ResCpuMaxLd1Lat | |
@@ -501,7 +501,7 @@ DEF:xxx={rrd1}:ResCpuMaxLd1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuMaxLd1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuMaxLd5Lat.name=ResCpuMaxLd5Lat | |
report.vmware3.ResCpuMaxLd5Lat.columns=ResCpuMaxLd5Lat | |
@@ -512,7 +512,7 @@ DEF:xxx={rrd1}:ResCpuMaxLd5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuMaxLd5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuRunav15Lat.name=ResCpuRunav15Lat | |
report.vmware3.ResCpuRunav15Lat.columns=ResCpuRunav15Lat | |
@@ -523,7 +523,7 @@ DEF:xxx={rrd1}:ResCpuRunav15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunav15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuRunav1Lat.name=ResCpuRunav1Lat | |
report.vmware3.ResCpuRunav1Lat.columns=ResCpuRunav1Lat | |
@@ -534,7 +534,7 @@ DEF:xxx={rrd1}:ResCpuRunav1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunav1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuRunav5Lat.name=ResCpuRunav5Lat | |
report.vmware3.ResCpuRunav5Lat.columns=ResCpuRunav5Lat | |
@@ -545,7 +545,7 @@ DEF:xxx={rrd1}:ResCpuRunav5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunav5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuRunpk15Lat.name=ResCpuRunpk15Lat | |
report.vmware3.ResCpuRunpk15Lat.columns=ResCpuRunpk15Lat | |
@@ -556,7 +556,7 @@ DEF:xxx={rrd1}:ResCpuRunpk15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunpk15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuRunpk1Lat.name=ResCpuRunpk1Lat | |
report.vmware3.ResCpuRunpk1Lat.columns=ResCpuRunpk1Lat | |
@@ -567,7 +567,7 @@ DEF:xxx={rrd1}:ResCpuRunpk1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunpk1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuRunpk5Lat.name=ResCpuRunpk5Lat | |
report.vmware3.ResCpuRunpk5Lat.columns=ResCpuRunpk5Lat | |
@@ -578,7 +578,7 @@ DEF:xxx={rrd1}:ResCpuRunpk5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunpk5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuSeCtLat.name=ResCpuSeCtLat | |
report.vmware3.ResCpuSeCtLat.columns=ResCpuSeCtLat | |
@@ -589,7 +589,7 @@ DEF:xxx={rrd1}:ResCpuSeCtLat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuSeCtLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuSePeriodLat.name=ResCpuSePeriodLat | |
report.vmware3.ResCpuSePeriodLat.columns=ResCpuSePeriodLat | |
@@ -600,7 +600,7 @@ DEF:xxx={rrd1}:ResCpuSePeriodLat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuSePeriodLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.SysUptimeLat.name=SysUptimeLat | |
report.vmware3.SysUptimeLat.columns=SysUptimeLat | |
@@ -611,7 +611,7 @@ DEF:xxx={rrd1}:SysUptimeLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysUptimeLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MgtAgtMemUsedAvg.name=MgtAgtMemUsedAvg | |
report.vmware3.MgtAgtMemUsedAvg.columns=MgtAgtMemUsedAvg | |
@@ -623,7 +623,7 @@ DEF:xxx={rrd1}:MgtAgtMemUsedAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"MgtAgtMemUsedAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MgtAgtSpInAvg.name=MgtAgtSpInAvg | |
report.vmware3.MgtAgtSpInAvg.columns=MgtAgtSpInAvg | |
@@ -635,7 +635,7 @@ DEF:xxx={rrd1}:MgtAgtSpInAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"MgtAgtSpInAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MgtAgtSpOutAvg.name=MgtAgtSpOutAvg | |
report.vmware3.MgtAgtSpOutAvg.columns=MgtAgtSpOutAvg | |
@@ -647,7 +647,7 @@ DEF:xxx={rrd1}:MgtAgtSpOutAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"MgtAgtSpOutAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MgtAgtSpUsedAvg.name=MgtAgtSpUsedAvg | |
report.vmware3.MgtAgtSpUsedAvg.columns=MgtAgtSpUsedAvg | |
@@ -659,7 +659,7 @@ DEF:xxx={rrd1}:MgtAgtSpUsedAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"MgtAgtSpUsedAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.NetPacketsRxSum.name=NetPacketsRxSum | |
report.vmware3.NetPacketsRxSum.columns=NetPacketsRxSum | |
@@ -671,7 +671,7 @@ DEF:xxx={rrd1}:NetPacketsRxSum:AVERAGE \ | |
LINE2:xxx#0000ff:"NetPacketsRxSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.NetPacketsTxSum.name=NetPacketsTxSum | |
report.vmware3.NetPacketsTxSum.columns=NetPacketsTxSum | |
@@ -683,7 +683,7 @@ DEF:xxx={rrd1}:NetPacketsTxSum:AVERAGE \ | |
LINE2:xxx#0000ff:"NetPacketsTxSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.NetReceivedAvg.name=NetReceivedAvg | |
report.vmware3.NetReceivedAvg.columns=NetReceivedAvg | |
@@ -695,7 +695,7 @@ DEF:xxx={rrd1}:NetReceivedAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"NetReceivedAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.NetTransmittedAvg.name=NetTransmittedAvg | |
report.vmware3.NetTransmittedAvg.columns=NetTransmittedAvg | |
@@ -707,7 +707,7 @@ DEF:xxx={rrd1}:NetTransmittedAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"NetTransmittedAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskBusResetsSum.name=DiskBusResetsSum | |
report.vmware3.DiskBusResetsSum.columns=DiskBusResetsSum | |
@@ -719,7 +719,7 @@ DEF:xxx={rrd1}:DiskBusResetsSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskBusResetsSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskCsAdSum.name=DiskCsAdSum | |
report.vmware3.DiskCsAdSum.columns=DiskCsAdSum | |
@@ -731,7 +731,7 @@ DEF:xxx={rrd1}:DiskCsAdSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskCsAdSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskCsSum.name=DiskCsSum | |
report.vmware3.DiskCsSum.columns=DiskCsSum | |
@@ -743,7 +743,7 @@ DEF:xxx={rrd1}:DiskCsSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskCsSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskDeLyAvg.name=DiskDeLyAvg | |
report.vmware3.DiskDeLyAvg.columns=DiskDeLyAvg | |
@@ -755,7 +755,7 @@ DEF:xxx={rrd1}:DiskDeLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskDeLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskDeRdLyAvg.name=DiskDeRdLyAvg | |
report.vmware3.DiskDeRdLyAvg.columns=DiskDeRdLyAvg | |
@@ -767,7 +767,7 @@ DEF:xxx={rrd1}:DiskDeRdLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskDeRdLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskDeWeLyAvg.name=DiskDeWeLyAvg | |
report.vmware3.DiskDeWeLyAvg.columns=DiskDeWeLyAvg | |
@@ -779,7 +779,7 @@ DEF:xxx={rrd1}:DiskDeWeLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskDeWeLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskKlLyAvg.name=DiskKlLyAvg | |
report.vmware3.DiskKlLyAvg.columns=DiskKlLyAvg | |
@@ -791,7 +791,7 @@ DEF:xxx={rrd1}:DiskKlLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskKlLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskKlRdLyAvg.name=DiskKlRdLyAvg | |
report.vmware3.DiskKlRdLyAvg.columns=DiskKlRdLyAvg | |
@@ -803,7 +803,7 @@ DEF:xxx={rrd1}:DiskKlRdLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskKlRdLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskKlWeLyAvg.name=DiskKlWeLyAvg | |
report.vmware3.DiskKlWeLyAvg.columns=DiskKlWeLyAvg | |
@@ -815,7 +815,7 @@ DEF:xxx={rrd1}:DiskKlWeLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskKlWeLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskNrRdSum.name=DiskNrRdSum | |
report.vmware3.DiskNrRdSum.columns=DiskNrRdSum | |
@@ -827,7 +827,7 @@ DEF:xxx={rrd1}:DiskNrRdSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskNrRdSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskNrWeSum.name=DiskNrWeSum | |
report.vmware3.DiskNrWeSum.columns=DiskNrWeSum | |
@@ -839,7 +839,7 @@ DEF:xxx={rrd1}:DiskNrWeSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskNrWeSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskQeLyAvg.name=DiskQeLyAvg | |
report.vmware3.DiskQeLyAvg.columns=DiskQeLyAvg | |
@@ -851,7 +851,7 @@ DEF:xxx={rrd1}:DiskQeLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskQeLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskQeRdLyAvg.name=DiskQeRdLyAvg | |
report.vmware3.DiskQeRdLyAvg.columns=DiskQeRdLyAvg | |
@@ -863,7 +863,7 @@ DEF:xxx={rrd1}:DiskQeRdLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskQeRdLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskQeWeLyAvg.name=DiskQeWeLyAvg | |
report.vmware3.DiskQeWeLyAvg.columns=DiskQeWeLyAvg | |
@@ -875,7 +875,7 @@ DEF:xxx={rrd1}:DiskQeWeLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskQeWeLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskRdAvg.name=DiskRdAvg | |
report.vmware3.DiskRdAvg.columns=DiskRdAvg | |
@@ -887,7 +887,7 @@ DEF:xxx={rrd1}:DiskRdAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskRdAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskTlLyAvg.name=DiskTlLyAvg | |
report.vmware3.DiskTlLyAvg.columns=DiskTlLyAvg | |
@@ -899,7 +899,7 @@ DEF:xxx={rrd1}:DiskTlLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskTlLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskTlRdLyAvg.name=DiskTlRdLyAvg | |
report.vmware3.DiskTlRdLyAvg.columns=DiskTlRdLyAvg | |
@@ -911,7 +911,7 @@ DEF:xxx={rrd1}:DiskTlRdLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskTlRdLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskTlWeLyAvg.name=DiskTlWeLyAvg | |
report.vmware3.DiskTlWeLyAvg.columns=DiskTlWeLyAvg | |
@@ -923,7 +923,7 @@ DEF:xxx={rrd1}:DiskTlWeLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskTlWeLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskWeAvg.name=DiskWeAvg | |
report.vmware3.DiskWeAvg.columns=DiskWeAvg | |
@@ -935,7 +935,7 @@ DEF:xxx={rrd1}:DiskWeAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskWeAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.SysReCpuUsageNon.name=SysReCpuUsageNon | |
report.vmware3.SysReCpuUsageNon.columns=SysReCpuUsageNon | |
@@ -947,7 +947,7 @@ DEF:xxx={rrd1}:SysReCpuUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReCpuUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.CpuExtraSum.name=CpuExtraSum | |
report.vmware3.CpuExtraSum.columns=CpuExtraSum | |
@@ -959,7 +959,7 @@ DEF:xxx={rrd1}:CpuExtraSum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuExtraSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.CpuGuaranteedLat.name=CpuGuaranteedLat | |
report.vmware3.CpuGuaranteedLat.columns=CpuGuaranteedLat | |
@@ -971,7 +971,7 @@ DEF:xxx={rrd1}:CpuGuaranteedLat:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuGuaranteedLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.CpuRdySum.name=CpuRdySum | |
report.vmware3.CpuRdySum.columns=CpuRdySum | |
@@ -983,7 +983,7 @@ DEF:xxx={rrd1}:CpuRdySum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuRdySum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.CpuSystemSum.name=CpuSystemSum | |
report.vmware3.CpuSystemSum.columns=CpuSystemSum | |
@@ -995,7 +995,7 @@ DEF:xxx={rrd1}:CpuSystemSum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuSystemSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.CpuUsagemhzNon.name=CpuUsagemhzNon | |
report.vmware3.CpuUsagemhzNon.columns=CpuUsagemhzNon | |
@@ -1007,7 +1007,7 @@ DEF:xxx={rrd1}:CpuUsagemhzNon:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsagemhzNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.CpuUsedSum.name=CpuUsedSum | |
report.vmware3.CpuUsedSum.columns=CpuUsedSum | |
@@ -1019,7 +1019,7 @@ DEF:xxx={rrd1}:CpuUsedSum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsedSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.CpuWaitSum.name=CpuWaitSum | |
report.vmware3.CpuWaitSum.columns=CpuWaitSum | |
@@ -1031,7 +1031,7 @@ DEF:xxx={rrd1}:CpuWaitSum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuWaitSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.CpuUsageNon.name=CpuUsageNon | |
report.vmware3.CpuUsageNon.columns=CpuUsageNon | |
@@ -1042,7 +1042,7 @@ DEF:xxx={rrd1}:CpuUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.CpuUsagemhzNon.name=CpuUsagemhzNon | |
report.vmware3.CpuUsagemhzNon.columns=CpuUsagemhzNon | |
@@ -1053,7 +1053,7 @@ DEF:xxx={rrd1}:CpuUsagemhzNon:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsagemhzNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskUsageNon.name=DiskUsageNon | |
report.vmware3.DiskUsageNon.columns=DiskUsageNon | |
@@ -1064,7 +1064,7 @@ DEF:xxx={rrd1}:DiskUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemActiveNon.name=MemActiveNon | |
report.vmware3.MemActiveNon.columns=MemActiveNon | |
@@ -1075,7 +1075,7 @@ DEF:xxx={rrd1}:MemActiveNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemActiveNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemConsumedNon.name=MemConsumedNon | |
report.vmware3.MemConsumedNon.columns=MemConsumedNon | |
@@ -1086,7 +1086,7 @@ DEF:xxx={rrd1}:MemConsumedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemConsumedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemGrantedNon.name=MemGrantedNon | |
report.vmware3.MemGrantedNon.columns=MemGrantedNon | |
@@ -1097,7 +1097,7 @@ DEF:xxx={rrd1}:MemGrantedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemGrantedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemOdNon.name=MemOdNon | |
report.vmware3.MemOdNon.columns=MemOdNon | |
@@ -1108,7 +1108,7 @@ DEF:xxx={rrd1}:MemOdNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemOdNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemSharedNon.name=MemSharedNon | |
report.vmware3.MemSharedNon.columns=MemSharedNon | |
@@ -1119,7 +1119,7 @@ DEF:xxx={rrd1}:MemSharedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSharedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemSpTtNon.name=MemSpTtNon | |
report.vmware3.MemSpTtNon.columns=MemSpTtNon | |
@@ -1130,7 +1130,7 @@ DEF:xxx={rrd1}:MemSpTtNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpTtNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemSpinNon.name=MemSpinNon | |
report.vmware3.MemSpinNon.columns=MemSpinNon | |
@@ -1141,7 +1141,7 @@ DEF:xxx={rrd1}:MemSpinNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpinNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemSpoutNon.name=MemSpoutNon | |
report.vmware3.MemSpoutNon.columns=MemSpoutNon | |
@@ -1152,7 +1152,7 @@ DEF:xxx={rrd1}:MemSpoutNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpoutNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemSppedNon.name=MemSppedNon | |
report.vmware3.MemSppedNon.columns=MemSppedNon | |
@@ -1163,7 +1163,7 @@ DEF:xxx={rrd1}:MemSppedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSppedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemUsageNon.name=MemUsageNon | |
report.vmware3.MemUsageNon.columns=MemUsageNon | |
@@ -1174,7 +1174,7 @@ DEF:xxx={rrd1}:MemUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemVmmemctlNon.name=MemVmmemctlNon | |
report.vmware3.MemVmmemctlNon.columns=MemVmmemctlNon | |
@@ -1185,7 +1185,7 @@ DEF:xxx={rrd1}:MemVmmemctlNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemVmmemctlNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemVmmemctlTtNon.name=MemVmmemctlTtNon | |
report.vmware3.MemVmmemctlTtNon.columns=MemVmmemctlTtNon | |
@@ -1196,7 +1196,7 @@ DEF:xxx={rrd1}:MemVmmemctlTtNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemVmmemctlTtNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.MemZeroNon.name=MemZeroNon | |
report.vmware3.MemZeroNon.columns=MemZeroNon | |
@@ -1207,7 +1207,7 @@ DEF:xxx={rrd1}:MemZeroNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemZeroNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.NetUsageNon.name=NetUsageNon | |
report.vmware3.NetUsageNon.columns=NetUsageNon | |
@@ -1218,7 +1218,7 @@ DEF:xxx={rrd1}:NetUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"NetUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuActav15Lat.name=ResCpuActav15Lat | |
report.vmware3.ResCpuActav15Lat.columns=ResCpuActav15Lat | |
@@ -1229,7 +1229,7 @@ DEF:xxx={rrd1}:ResCpuActav15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActav15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuActav1Lat.name=ResCpuActav1Lat | |
report.vmware3.ResCpuActav1Lat.columns=ResCpuActav1Lat | |
@@ -1240,7 +1240,7 @@ DEF:xxx={rrd1}:ResCpuActav1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActav1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuActav5Lat.name=ResCpuActav5Lat | |
report.vmware3.ResCpuActav5Lat.columns=ResCpuActav5Lat | |
@@ -1251,7 +1251,7 @@ DEF:xxx={rrd1}:ResCpuActav5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActav5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuActpk15Lat.name=ResCpuActpk15Lat | |
report.vmware3.ResCpuActpk15Lat.columns=ResCpuActpk15Lat | |
@@ -1262,7 +1262,7 @@ DEF:xxx={rrd1}:ResCpuActpk15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActpk15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuActpk1Lat.name=ResCpuActpk1Lat | |
report.vmware3.ResCpuActpk1Lat.columns=ResCpuActpk1Lat | |
@@ -1273,7 +1273,7 @@ DEF:xxx={rrd1}:ResCpuActpk1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActpk1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuActpk5Lat.name=ResCpuActpk5Lat | |
report.vmware3.ResCpuActpk5Lat.columns=ResCpuActpk5Lat | |
@@ -1284,7 +1284,7 @@ DEF:xxx={rrd1}:ResCpuActpk5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActpk5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuMaxLd15Lat.name=ResCpuMaxLd15Lat | |
report.vmware3.ResCpuMaxLd15Lat.columns=ResCpuMaxLd15Lat | |
@@ -1295,7 +1295,7 @@ DEF:xxx={rrd1}:ResCpuMaxLd15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuMaxLd15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuMaxLd1Lat.name=ResCpuMaxLd1Lat | |
report.vmware3.ResCpuMaxLd1Lat.columns=ResCpuMaxLd1Lat | |
@@ -1306,7 +1306,7 @@ DEF:xxx={rrd1}:ResCpuMaxLd1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuMaxLd1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuMaxLd5Lat.name=ResCpuMaxLd5Lat | |
report.vmware3.ResCpuMaxLd5Lat.columns=ResCpuMaxLd5Lat | |
@@ -1317,7 +1317,7 @@ DEF:xxx={rrd1}:ResCpuMaxLd5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuMaxLd5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuRunav15Lat.name=ResCpuRunav15Lat | |
report.vmware3.ResCpuRunav15Lat.columns=ResCpuRunav15Lat | |
@@ -1328,7 +1328,7 @@ DEF:xxx={rrd1}:ResCpuRunav15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunav15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuRunav1Lat.name=ResCpuRunav1Lat | |
report.vmware3.ResCpuRunav1Lat.columns=ResCpuRunav1Lat | |
@@ -1339,7 +1339,7 @@ DEF:xxx={rrd1}:ResCpuRunav1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunav1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuRunav5Lat.name=ResCpuRunav5Lat | |
report.vmware3.ResCpuRunav5Lat.columns=ResCpuRunav5Lat | |
@@ -1350,7 +1350,7 @@ DEF:xxx={rrd1}:ResCpuRunav5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunav5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuRunpk15Lat.name=ResCpuRunpk15Lat | |
report.vmware3.ResCpuRunpk15Lat.columns=ResCpuRunpk15Lat | |
@@ -1361,7 +1361,7 @@ DEF:xxx={rrd1}:ResCpuRunpk15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunpk15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuRunpk1Lat.name=ResCpuRunpk1Lat | |
report.vmware3.ResCpuRunpk1Lat.columns=ResCpuRunpk1Lat | |
@@ -1372,7 +1372,7 @@ DEF:xxx={rrd1}:ResCpuRunpk1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunpk1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuRunpk5Lat.name=ResCpuRunpk5Lat | |
report.vmware3.ResCpuRunpk5Lat.columns=ResCpuRunpk5Lat | |
@@ -1383,7 +1383,7 @@ DEF:xxx={rrd1}:ResCpuRunpk5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunpk5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuSeCtLat.name=ResCpuSeCtLat | |
report.vmware3.ResCpuSeCtLat.columns=ResCpuSeCtLat | |
@@ -1394,7 +1394,7 @@ DEF:xxx={rrd1}:ResCpuSeCtLat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuSeCtLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.ResCpuSePeriodLat.name=ResCpuSePeriodLat | |
report.vmware3.ResCpuSePeriodLat.columns=ResCpuSePeriodLat | |
@@ -1405,7 +1405,7 @@ DEF:xxx={rrd1}:ResCpuSePeriodLat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuSePeriodLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.SysHeartbeatSum.name=SysHeartbeatSum | |
report.vmware3.SysHeartbeatSum.columns=SysHeartbeatSum | |
@@ -1416,7 +1416,7 @@ DEF:xxx={rrd1}:SysHeartbeatSum:AVERAGE \ | |
LINE2:xxx#0000ff:"SysHeartbeatSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.SysUptimeLat.name=SysUptimeLat | |
report.vmware3.SysUptimeLat.columns=SysUptimeLat | |
@@ -1427,7 +1427,7 @@ DEF:xxx={rrd1}:SysUptimeLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysUptimeLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.NetPacketsRxSum.name=NetPacketsRxSum | |
report.vmware3.NetPacketsRxSum.columns=NetPacketsRxSum | |
@@ -1439,7 +1439,7 @@ DEF:xxx={rrd1}:NetPacketsRxSum:AVERAGE \ | |
LINE2:xxx#0000ff:"NetPacketsRxSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.NetPacketsTxSum.name=NetPacketsTxSum | |
report.vmware3.NetPacketsTxSum.columns=NetPacketsTxSum | |
@@ -1451,7 +1451,7 @@ DEF:xxx={rrd1}:NetPacketsTxSum:AVERAGE \ | |
LINE2:xxx#0000ff:"NetPacketsTxSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.NetReceivedAvg.name=NetReceivedAvg | |
report.vmware3.NetReceivedAvg.columns=NetReceivedAvg | |
@@ -1463,7 +1463,7 @@ DEF:xxx={rrd1}:NetReceivedAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"NetReceivedAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.NetTransmittedAvg.name=NetTransmittedAvg | |
report.vmware3.NetTransmittedAvg.columns=NetTransmittedAvg | |
@@ -1475,7 +1475,7 @@ DEF:xxx={rrd1}:NetTransmittedAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"NetTransmittedAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskBusResetsSum.name=DiskBusResetsSum | |
report.vmware3.DiskBusResetsSum.columns=DiskBusResetsSum | |
@@ -1487,7 +1487,7 @@ DEF:xxx={rrd1}:DiskBusResetsSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskBusResetsSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskCsAdSum.name=DiskCsAdSum | |
report.vmware3.DiskCsAdSum.columns=DiskCsAdSum | |
@@ -1499,7 +1499,7 @@ DEF:xxx={rrd1}:DiskCsAdSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskCsAdSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskCsSum.name=DiskCsSum | |
report.vmware3.DiskCsSum.columns=DiskCsSum | |
@@ -1511,7 +1511,7 @@ DEF:xxx={rrd1}:DiskCsSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskCsSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskNrRdSum.name=DiskNrRdSum | |
report.vmware3.DiskNrRdSum.columns=DiskNrRdSum | |
@@ -1523,7 +1523,7 @@ DEF:xxx={rrd1}:DiskNrRdSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskNrRdSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskNrWeSum.name=DiskNrWeSum | |
report.vmware3.DiskNrWeSum.columns=DiskNrWeSum | |
@@ -1535,7 +1535,7 @@ DEF:xxx={rrd1}:DiskNrWeSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskNrWeSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskRdAvg.name=DiskRdAvg | |
report.vmware3.DiskRdAvg.columns=DiskRdAvg | |
@@ -1547,7 +1547,7 @@ DEF:xxx={rrd1}:DiskRdAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskRdAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware3.DiskWeAvg.name=DiskWeAvg | |
report.vmware3.DiskWeAvg.columns=DiskWeAvg | |
@@ -1559,5 +1559,5 @@ DEF:xxx={rrd1}:DiskWeAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskWeAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
diff --git a/snmp-graph.properties.d/vmware4-graph-simple.properties b/snmp-graph.properties.d/vmware4-graph-simple.properties | |
index ffdd6f8..3211f0e 100644 | |
--- a/snmp-graph.properties.d/vmware4-graph-simple.properties | |
+++ b/snmp-graph.properties.d/vmware4-graph-simple.properties | |
@@ -165,7 +165,7 @@ DEF:xxx={rrd1}:CpuIdleSum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuIdleSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.CpuUsageNon.name=CpuUsageNon | |
report.vmware4.CpuUsageNon.columns=CpuUsageNon | |
@@ -177,7 +177,7 @@ DEF:xxx={rrd1}:CpuUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.CpuUsedSum.name=CpuUsedSum | |
report.vmware4.CpuUsedSum.columns=CpuUsedSum | |
@@ -189,7 +189,7 @@ DEF:xxx={rrd1}:CpuUsedSum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsedSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.CpuRdCyAvg.name=CpuRdCyAvg | |
report.vmware4.CpuRdCyAvg.columns=CpuRdCyAvg | |
@@ -200,7 +200,7 @@ DEF:xxx={rrd1}:CpuRdCyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuRdCyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.CpuUsagemhzNon.name=CpuUsagemhzNon | |
report.vmware4.CpuUsagemhzNon.columns=CpuUsagemhzNon | |
@@ -211,7 +211,7 @@ DEF:xxx={rrd1}:CpuUsagemhzNon:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsagemhzNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskMaxTlLyLat.name=DiskMaxTlLyLat | |
report.vmware4.DiskMaxTlLyLat.columns=DiskMaxTlLyLat | |
@@ -222,7 +222,7 @@ DEF:xxx={rrd1}:DiskMaxTlLyLat:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskMaxTlLyLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskUsageNon.name=DiskUsageNon | |
report.vmware4.DiskUsageNon.columns=DiskUsageNon | |
@@ -233,7 +233,7 @@ DEF:xxx={rrd1}:DiskUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemActiveNon.name=MemActiveNon | |
report.vmware4.MemActiveNon.columns=MemActiveNon | |
@@ -244,7 +244,7 @@ DEF:xxx={rrd1}:MemActiveNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemActiveNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemConsumedNon.name=MemConsumedNon | |
report.vmware4.MemConsumedNon.columns=MemConsumedNon | |
@@ -255,7 +255,7 @@ DEF:xxx={rrd1}:MemConsumedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemConsumedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemGrantedNon.name=MemGrantedNon | |
report.vmware4.MemGrantedNon.columns=MemGrantedNon | |
@@ -266,7 +266,7 @@ DEF:xxx={rrd1}:MemGrantedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemGrantedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemHeapNon.name=MemHeapNon | |
report.vmware4.MemHeapNon.columns=MemHeapNon | |
@@ -277,7 +277,7 @@ DEF:xxx={rrd1}:MemHeapNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemHeapNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemHeapfreeNon.name=MemHeapfreeNon | |
report.vmware4.MemHeapfreeNon.columns=MemHeapfreeNon | |
@@ -288,7 +288,7 @@ DEF:xxx={rrd1}:MemHeapfreeNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemHeapfreeNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemOdNon.name=MemOdNon | |
report.vmware4.MemOdNon.columns=MemOdNon | |
@@ -299,7 +299,7 @@ DEF:xxx={rrd1}:MemOdNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemOdNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemRdCyAvg.name=MemRdCyAvg | |
report.vmware4.MemRdCyAvg.columns=MemRdCyAvg | |
@@ -310,7 +310,7 @@ DEF:xxx={rrd1}:MemRdCyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"MemRdCyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSharedNon.name=MemSharedNon | |
report.vmware4.MemSharedNon.columns=MemSharedNon | |
@@ -321,7 +321,7 @@ DEF:xxx={rrd1}:MemSharedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSharedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSharedcommonNon.name=MemSharedcommonNon | |
report.vmware4.MemSharedcommonNon.columns=MemSharedcommonNon | |
@@ -332,7 +332,7 @@ DEF:xxx={rrd1}:MemSharedcommonNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSharedcommonNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSpinNon.name=MemSpinNon | |
report.vmware4.MemSpinNon.columns=MemSpinNon | |
@@ -343,7 +343,7 @@ DEF:xxx={rrd1}:MemSpinNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpinNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSpinReAvg.name=MemSpinReAvg | |
report.vmware4.MemSpinReAvg.columns=MemSpinReAvg | |
@@ -354,7 +354,7 @@ DEF:xxx={rrd1}:MemSpinReAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpinReAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSpoutNon.name=MemSpoutNon | |
report.vmware4.MemSpoutNon.columns=MemSpoutNon | |
@@ -365,7 +365,7 @@ DEF:xxx={rrd1}:MemSpoutNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpoutNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSpoutReAvg.name=MemSpoutReAvg | |
report.vmware4.MemSpoutReAvg.columns=MemSpoutReAvg | |
@@ -376,7 +376,7 @@ DEF:xxx={rrd1}:MemSpoutReAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpoutReAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSpusedNon.name=MemSpusedNon | |
report.vmware4.MemSpusedNon.columns=MemSpusedNon | |
@@ -387,7 +387,7 @@ DEF:xxx={rrd1}:MemSpusedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpusedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemStateLat.name=MemStateLat | |
report.vmware4.MemStateLat.columns=MemStateLat | |
@@ -398,7 +398,7 @@ DEF:xxx={rrd1}:MemStateLat:AVERAGE \ | |
LINE2:xxx#0000ff:"MemStateLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSysUsageNon.name=MemSysUsageNon | |
report.vmware4.MemSysUsageNon.columns=MemSysUsageNon | |
@@ -409,7 +409,7 @@ DEF:xxx={rrd1}:MemSysUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSysUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemUdNon.name=MemUdNon | |
report.vmware4.MemUdNon.columns=MemUdNon | |
@@ -420,7 +420,7 @@ DEF:xxx={rrd1}:MemUdNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemUdNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemUsageNon.name=MemUsageNon | |
report.vmware4.MemUsageNon.columns=MemUsageNon | |
@@ -431,7 +431,7 @@ DEF:xxx={rrd1}:MemUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemVmmemctlNon.name=MemVmmemctlNon | |
report.vmware4.MemVmmemctlNon.columns=MemVmmemctlNon | |
@@ -442,7 +442,7 @@ DEF:xxx={rrd1}:MemVmmemctlNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemVmmemctlNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemZeroNon.name=MemZeroNon | |
report.vmware4.MemZeroNon.columns=MemZeroNon | |
@@ -453,7 +453,7 @@ DEF:xxx={rrd1}:MemZeroNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemZeroNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.NetTransmittedAvg.name=NetTransmittedAvg | |
report.vmware4.NetTransmittedAvg.columns=NetTransmittedAvg | |
@@ -464,7 +464,7 @@ DEF:xxx={rrd1}:NetTransmittedAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"NetTransmittedAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.NetUsageNon.name=NetUsageNon | |
report.vmware4.NetUsageNon.columns=NetUsageNon | |
@@ -475,7 +475,7 @@ DEF:xxx={rrd1}:NetUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"NetUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuActav15Lat.name=ResCpuActav15Lat | |
report.vmware4.ResCpuActav15Lat.columns=ResCpuActav15Lat | |
@@ -486,7 +486,7 @@ DEF:xxx={rrd1}:ResCpuActav15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActav15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuActav1Lat.name=ResCpuActav1Lat | |
report.vmware4.ResCpuActav1Lat.columns=ResCpuActav1Lat | |
@@ -497,7 +497,7 @@ DEF:xxx={rrd1}:ResCpuActav1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActav1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuActav5Lat.name=ResCpuActav5Lat | |
report.vmware4.ResCpuActav5Lat.columns=ResCpuActav5Lat | |
@@ -508,7 +508,7 @@ DEF:xxx={rrd1}:ResCpuActav5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActav5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuActpk15Lat.name=ResCpuActpk15Lat | |
report.vmware4.ResCpuActpk15Lat.columns=ResCpuActpk15Lat | |
@@ -519,7 +519,7 @@ DEF:xxx={rrd1}:ResCpuActpk15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActpk15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuActpk1Lat.name=ResCpuActpk1Lat | |
report.vmware4.ResCpuActpk1Lat.columns=ResCpuActpk1Lat | |
@@ -530,7 +530,7 @@ DEF:xxx={rrd1}:ResCpuActpk1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActpk1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuActpk5Lat.name=ResCpuActpk5Lat | |
report.vmware4.ResCpuActpk5Lat.columns=ResCpuActpk5Lat | |
@@ -541,7 +541,7 @@ DEF:xxx={rrd1}:ResCpuActpk5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActpk5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuMaxLd15Lat.name=ResCpuMaxLd15Lat | |
report.vmware4.ResCpuMaxLd15Lat.columns=ResCpuMaxLd15Lat | |
@@ -552,7 +552,7 @@ DEF:xxx={rrd1}:ResCpuMaxLd15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuMaxLd15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuMaxLd1Lat.name=ResCpuMaxLd1Lat | |
report.vmware4.ResCpuMaxLd1Lat.columns=ResCpuMaxLd1Lat | |
@@ -563,7 +563,7 @@ DEF:xxx={rrd1}:ResCpuMaxLd1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuMaxLd1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuMaxLd5Lat.name=ResCpuMaxLd5Lat | |
report.vmware4.ResCpuMaxLd5Lat.columns=ResCpuMaxLd5Lat | |
@@ -574,7 +574,7 @@ DEF:xxx={rrd1}:ResCpuMaxLd5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuMaxLd5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuRunav15Lat.name=ResCpuRunav15Lat | |
report.vmware4.ResCpuRunav15Lat.columns=ResCpuRunav15Lat | |
@@ -585,7 +585,7 @@ DEF:xxx={rrd1}:ResCpuRunav15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunav15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuRunav1Lat.name=ResCpuRunav1Lat | |
report.vmware4.ResCpuRunav1Lat.columns=ResCpuRunav1Lat | |
@@ -596,7 +596,7 @@ DEF:xxx={rrd1}:ResCpuRunav1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunav1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuRunav5Lat.name=ResCpuRunav5Lat | |
report.vmware4.ResCpuRunav5Lat.columns=ResCpuRunav5Lat | |
@@ -607,7 +607,7 @@ DEF:xxx={rrd1}:ResCpuRunav5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunav5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuRunpk15Lat.name=ResCpuRunpk15Lat | |
report.vmware4.ResCpuRunpk15Lat.columns=ResCpuRunpk15Lat | |
@@ -618,7 +618,7 @@ DEF:xxx={rrd1}:ResCpuRunpk15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunpk15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuRunpk1Lat.name=ResCpuRunpk1Lat | |
report.vmware4.ResCpuRunpk1Lat.columns=ResCpuRunpk1Lat | |
@@ -629,7 +629,7 @@ DEF:xxx={rrd1}:ResCpuRunpk1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunpk1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuRunpk5Lat.name=ResCpuRunpk5Lat | |
report.vmware4.ResCpuRunpk5Lat.columns=ResCpuRunpk5Lat | |
@@ -640,7 +640,7 @@ DEF:xxx={rrd1}:ResCpuRunpk5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunpk5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuSeCtLat.name=ResCpuSeCtLat | |
report.vmware4.ResCpuSeCtLat.columns=ResCpuSeCtLat | |
@@ -651,7 +651,7 @@ DEF:xxx={rrd1}:ResCpuSeCtLat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuSeCtLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuSePeriodLat.name=ResCpuSePeriodLat | |
report.vmware4.ResCpuSePeriodLat.columns=ResCpuSePeriodLat | |
@@ -662,7 +662,7 @@ DEF:xxx={rrd1}:ResCpuSePeriodLat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuSePeriodLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysUptimeLat.name=SysUptimeLat | |
report.vmware4.SysUptimeLat.columns=SysUptimeLat | |
@@ -673,7 +673,7 @@ DEF:xxx={rrd1}:SysUptimeLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysUptimeLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MgtAgtMemUsedAvg.name=MgtAgtMemUsedAvg | |
report.vmware4.MgtAgtMemUsedAvg.columns=MgtAgtMemUsedAvg | |
@@ -685,7 +685,7 @@ DEF:xxx={rrd1}:MgtAgtMemUsedAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"MgtAgtMemUsedAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MgtAgtSpInAvg.name=MgtAgtSpInAvg | |
report.vmware4.MgtAgtSpInAvg.columns=MgtAgtSpInAvg | |
@@ -697,7 +697,7 @@ DEF:xxx={rrd1}:MgtAgtSpInAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"MgtAgtSpInAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MgtAgtSpOutAvg.name=MgtAgtSpOutAvg | |
report.vmware4.MgtAgtSpOutAvg.columns=MgtAgtSpOutAvg | |
@@ -709,7 +709,7 @@ DEF:xxx={rrd1}:MgtAgtSpOutAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"MgtAgtSpOutAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MgtAgtSpUsedAvg.name=MgtAgtSpUsedAvg | |
report.vmware4.MgtAgtSpUsedAvg.columns=MgtAgtSpUsedAvg | |
@@ -721,7 +721,7 @@ DEF:xxx={rrd1}:MgtAgtSpUsedAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"MgtAgtSpUsedAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.NetDroppedRxSum.name=NetDroppedRxSum | |
report.vmware4.NetDroppedRxSum.columns=NetDroppedRxSum | |
@@ -733,7 +733,7 @@ DEF:xxx={rrd1}:NetDroppedRxSum:AVERAGE \ | |
LINE2:xxx#0000ff:"NetDroppedRxSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.NetDroppedTxSum.name=NetDroppedTxSum | |
report.vmware4.NetDroppedTxSum.columns=NetDroppedTxSum | |
@@ -745,7 +745,7 @@ DEF:xxx={rrd1}:NetDroppedTxSum:AVERAGE \ | |
LINE2:xxx#0000ff:"NetDroppedTxSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.NetPacketsRxSum.name=NetPacketsRxSum | |
report.vmware4.NetPacketsRxSum.columns=NetPacketsRxSum | |
@@ -757,7 +757,7 @@ DEF:xxx={rrd1}:NetPacketsRxSum:AVERAGE \ | |
LINE2:xxx#0000ff:"NetPacketsRxSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.NetPacketsTxSum.name=NetPacketsTxSum | |
report.vmware4.NetPacketsTxSum.columns=NetPacketsTxSum | |
@@ -769,7 +769,7 @@ DEF:xxx={rrd1}:NetPacketsTxSum:AVERAGE \ | |
LINE2:xxx#0000ff:"NetPacketsTxSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.NetReceivedAvg.name=NetReceivedAvg | |
report.vmware4.NetReceivedAvg.columns=NetReceivedAvg | |
@@ -781,7 +781,7 @@ DEF:xxx={rrd1}:NetReceivedAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"NetReceivedAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.NetTransmittedAvg.name=NetTransmittedAvg | |
report.vmware4.NetTransmittedAvg.columns=NetTransmittedAvg | |
@@ -793,7 +793,7 @@ DEF:xxx={rrd1}:NetTransmittedAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"NetTransmittedAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskBusResetsSum.name=DiskBusResetsSum | |
report.vmware4.DiskBusResetsSum.columns=DiskBusResetsSum | |
@@ -805,7 +805,7 @@ DEF:xxx={rrd1}:DiskBusResetsSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskBusResetsSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskCsAdSum.name=DiskCsAdSum | |
report.vmware4.DiskCsAdSum.columns=DiskCsAdSum | |
@@ -817,7 +817,7 @@ DEF:xxx={rrd1}:DiskCsAdSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskCsAdSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskCsSum.name=DiskCsSum | |
report.vmware4.DiskCsSum.columns=DiskCsSum | |
@@ -829,7 +829,7 @@ DEF:xxx={rrd1}:DiskCsSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskCsSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskDeLyAvg.name=DiskDeLyAvg | |
report.vmware4.DiskDeLyAvg.columns=DiskDeLyAvg | |
@@ -841,7 +841,7 @@ DEF:xxx={rrd1}:DiskDeLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskDeLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskDeRdLyAvg.name=DiskDeRdLyAvg | |
report.vmware4.DiskDeRdLyAvg.columns=DiskDeRdLyAvg | |
@@ -853,7 +853,7 @@ DEF:xxx={rrd1}:DiskDeRdLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskDeRdLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskDeWeLyAvg.name=DiskDeWeLyAvg | |
report.vmware4.DiskDeWeLyAvg.columns=DiskDeWeLyAvg | |
@@ -865,7 +865,7 @@ DEF:xxx={rrd1}:DiskDeWeLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskDeWeLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskKlLyAvg.name=DiskKlLyAvg | |
report.vmware4.DiskKlLyAvg.columns=DiskKlLyAvg | |
@@ -877,7 +877,7 @@ DEF:xxx={rrd1}:DiskKlLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskKlLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskKlRdLyAvg.name=DiskKlRdLyAvg | |
report.vmware4.DiskKlRdLyAvg.columns=DiskKlRdLyAvg | |
@@ -889,7 +889,7 @@ DEF:xxx={rrd1}:DiskKlRdLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskKlRdLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskKlWeLyAvg.name=DiskKlWeLyAvg | |
report.vmware4.DiskKlWeLyAvg.columns=DiskKlWeLyAvg | |
@@ -901,7 +901,7 @@ DEF:xxx={rrd1}:DiskKlWeLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskKlWeLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskNrRdSum.name=DiskNrRdSum | |
report.vmware4.DiskNrRdSum.columns=DiskNrRdSum | |
@@ -913,7 +913,7 @@ DEF:xxx={rrd1}:DiskNrRdSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskNrRdSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskNrWeSum.name=DiskNrWeSum | |
report.vmware4.DiskNrWeSum.columns=DiskNrWeSum | |
@@ -925,7 +925,7 @@ DEF:xxx={rrd1}:DiskNrWeSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskNrWeSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskQeLyAvg.name=DiskQeLyAvg | |
report.vmware4.DiskQeLyAvg.columns=DiskQeLyAvg | |
@@ -937,7 +937,7 @@ DEF:xxx={rrd1}:DiskQeLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskQeLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskQeRdLyAvg.name=DiskQeRdLyAvg | |
report.vmware4.DiskQeRdLyAvg.columns=DiskQeRdLyAvg | |
@@ -949,7 +949,7 @@ DEF:xxx={rrd1}:DiskQeRdLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskQeRdLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskQeWeLyAvg.name=DiskQeWeLyAvg | |
report.vmware4.DiskQeWeLyAvg.columns=DiskQeWeLyAvg | |
@@ -961,7 +961,7 @@ DEF:xxx={rrd1}:DiskQeWeLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskQeWeLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskRdAvg.name=DiskRdAvg | |
report.vmware4.DiskRdAvg.columns=DiskRdAvg | |
@@ -973,7 +973,7 @@ DEF:xxx={rrd1}:DiskRdAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskRdAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskTlLyAvg.name=DiskTlLyAvg | |
report.vmware4.DiskTlLyAvg.columns=DiskTlLyAvg | |
@@ -985,7 +985,7 @@ DEF:xxx={rrd1}:DiskTlLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskTlLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskTlRdLyAvg.name=DiskTlRdLyAvg | |
report.vmware4.DiskTlRdLyAvg.columns=DiskTlRdLyAvg | |
@@ -997,7 +997,7 @@ DEF:xxx={rrd1}:DiskTlRdLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskTlRdLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskTlWeLyAvg.name=DiskTlWeLyAvg | |
report.vmware4.DiskTlWeLyAvg.columns=DiskTlWeLyAvg | |
@@ -1009,7 +1009,7 @@ DEF:xxx={rrd1}:DiskTlWeLyAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskTlWeLyAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskWeAvg.name=DiskWeAvg | |
report.vmware4.DiskWeAvg.columns=DiskWeAvg | |
@@ -1021,7 +1021,7 @@ DEF:xxx={rrd1}:DiskWeAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskWeAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysDiskUsageLat.name=SysDiskUsageLat | |
report.vmware4.SysDiskUsageLat.columns=SysDiskUsageLat | |
@@ -1033,7 +1033,7 @@ DEF:xxx={rrd1}:SysDiskUsageLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysDiskUsageLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReCpuAcMaxLat.name=SysReCpuAcMaxLat | |
report.vmware4.SysReCpuAcMaxLat.columns=SysReCpuAcMaxLat | |
@@ -1045,7 +1045,7 @@ DEF:xxx={rrd1}:SysReCpuAcMaxLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReCpuAcMaxLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReCpuAcMinLat.name=SysReCpuAcMinLat | |
report.vmware4.SysReCpuAcMinLat.columns=SysReCpuAcMinLat | |
@@ -1057,7 +1057,7 @@ DEF:xxx={rrd1}:SysReCpuAcMinLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReCpuAcMinLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReCpuAcSsLat.name=SysReCpuAcSsLat | |
report.vmware4.SysReCpuAcSsLat.columns=SysReCpuAcSsLat | |
@@ -1069,7 +1069,7 @@ DEF:xxx={rrd1}:SysReCpuAcSsLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReCpuAcSsLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReCpuAct1Lat.name=SysReCpuAct1Lat | |
report.vmware4.SysReCpuAct1Lat.columns=SysReCpuAct1Lat | |
@@ -1081,7 +1081,7 @@ DEF:xxx={rrd1}:SysReCpuAct1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReCpuAct1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReCpuAct5Lat.name=SysReCpuAct5Lat | |
report.vmware4.SysReCpuAct5Lat.columns=SysReCpuAct5Lat | |
@@ -1093,7 +1093,7 @@ DEF:xxx={rrd1}:SysReCpuAct5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReCpuAct5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReCpuMaxLd1Lat.name=SysReCpuMaxLd1Lat | |
report.vmware4.SysReCpuMaxLd1Lat.columns=SysReCpuMaxLd1Lat | |
@@ -1105,7 +1105,7 @@ DEF:xxx={rrd1}:SysReCpuMaxLd1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReCpuMaxLd1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReCpuMaxLd5Lat.name=SysReCpuMaxLd5Lat | |
report.vmware4.SysReCpuMaxLd5Lat.columns=SysReCpuMaxLd5Lat | |
@@ -1117,7 +1117,7 @@ DEF:xxx={rrd1}:SysReCpuMaxLd5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReCpuMaxLd5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReCpuRun1Lat.name=SysReCpuRun1Lat | |
report.vmware4.SysReCpuRun1Lat.columns=SysReCpuRun1Lat | |
@@ -1129,7 +1129,7 @@ DEF:xxx={rrd1}:SysReCpuRun1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReCpuRun1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReCpuRun5Lat.name=SysReCpuRun5Lat | |
report.vmware4.SysReCpuRun5Lat.columns=SysReCpuRun5Lat | |
@@ -1141,7 +1141,7 @@ DEF:xxx={rrd1}:SysReCpuRun5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReCpuRun5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReCpuUsageNon.name=SysReCpuUsageNon | |
report.vmware4.SysReCpuUsageNon.columns=SysReCpuUsageNon | |
@@ -1153,7 +1153,7 @@ DEF:xxx={rrd1}:SysReCpuUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReCpuUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReMemAcMaxLat.name=SysReMemAcMaxLat | |
report.vmware4.SysReMemAcMaxLat.columns=SysReMemAcMaxLat | |
@@ -1165,7 +1165,7 @@ DEF:xxx={rrd1}:SysReMemAcMaxLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReMemAcMaxLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReMemAcMinLat.name=SysReMemAcMinLat | |
report.vmware4.SysReMemAcMinLat.columns=SysReMemAcMinLat | |
@@ -1177,7 +1177,7 @@ DEF:xxx={rrd1}:SysReMemAcMinLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReMemAcMinLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReMemAcSsLat.name=SysReMemAcSsLat | |
report.vmware4.SysReMemAcSsLat.columns=SysReMemAcSsLat | |
@@ -1189,7 +1189,7 @@ DEF:xxx={rrd1}:SysReMemAcSsLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReMemAcSsLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReMemCowLat.name=SysReMemCowLat | |
report.vmware4.SysReMemCowLat.columns=SysReMemCowLat | |
@@ -1201,7 +1201,7 @@ DEF:xxx={rrd1}:SysReMemCowLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReMemCowLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReMemMappedLat.name=SysReMemMappedLat | |
report.vmware4.SysReMemMappedLat.columns=SysReMemMappedLat | |
@@ -1213,7 +1213,7 @@ DEF:xxx={rrd1}:SysReMemMappedLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReMemMappedLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReMemOdLat.name=SysReMemOdLat | |
report.vmware4.SysReMemOdLat.columns=SysReMemOdLat | |
@@ -1225,7 +1225,7 @@ DEF:xxx={rrd1}:SysReMemOdLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReMemOdLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReMemSharedLat.name=SysReMemSharedLat | |
report.vmware4.SysReMemSharedLat.columns=SysReMemSharedLat | |
@@ -1237,7 +1237,7 @@ DEF:xxx={rrd1}:SysReMemSharedLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReMemSharedLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReMemSppedLat.name=SysReMemSppedLat | |
report.vmware4.SysReMemSppedLat.columns=SysReMemSppedLat | |
@@ -1249,7 +1249,7 @@ DEF:xxx={rrd1}:SysReMemSppedLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReMemSppedLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReMemTdLat.name=SysReMemTdLat | |
report.vmware4.SysReMemTdLat.columns=SysReMemTdLat | |
@@ -1261,7 +1261,7 @@ DEF:xxx={rrd1}:SysReMemTdLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReMemTdLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysReMemZeroLat.name=SysReMemZeroLat | |
report.vmware4.SysReMemZeroLat.columns=SysReMemZeroLat | |
@@ -1273,7 +1273,7 @@ DEF:xxx={rrd1}:SysReMemZeroLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysReMemZeroLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.CpuRdySum.name=CpuRdySum | |
report.vmware4.CpuRdySum.columns=CpuRdySum | |
@@ -1285,7 +1285,7 @@ DEF:xxx={rrd1}:CpuRdySum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuRdySum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.CpuSpwaitSum.name=CpuSpwaitSum | |
report.vmware4.CpuSpwaitSum.columns=CpuSpwaitSum | |
@@ -1297,7 +1297,7 @@ DEF:xxx={rrd1}:CpuSpwaitSum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuSpwaitSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.CpuSystemSum.name=CpuSystemSum | |
report.vmware4.CpuSystemSum.columns=CpuSystemSum | |
@@ -1309,7 +1309,7 @@ DEF:xxx={rrd1}:CpuSystemSum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuSystemSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.CpuUsagemhzNon.name=CpuUsagemhzNon | |
report.vmware4.CpuUsagemhzNon.columns=CpuUsagemhzNon | |
@@ -1321,7 +1321,7 @@ DEF:xxx={rrd1}:CpuUsagemhzNon:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsagemhzNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.CpuUsedSum.name=CpuUsedSum | |
report.vmware4.CpuUsedSum.columns=CpuUsedSum | |
@@ -1333,7 +1333,7 @@ DEF:xxx={rrd1}:CpuUsedSum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsedSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.CpuWaitSum.name=CpuWaitSum | |
report.vmware4.CpuWaitSum.columns=CpuWaitSum | |
@@ -1345,7 +1345,7 @@ DEF:xxx={rrd1}:CpuWaitSum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuWaitSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.CpuRdySum.name=CpuRdySum | |
report.vmware4.CpuRdySum.columns=CpuRdySum | |
@@ -1356,7 +1356,7 @@ DEF:xxx={rrd1}:CpuRdySum:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuRdySum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.CpuUsageNon.name=CpuUsageNon | |
report.vmware4.CpuUsageNon.columns=CpuUsageNon | |
@@ -1367,7 +1367,7 @@ DEF:xxx={rrd1}:CpuUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.CpuUsagemhzNon.name=CpuUsagemhzNon | |
report.vmware4.CpuUsagemhzNon.columns=CpuUsagemhzNon | |
@@ -1378,7 +1378,7 @@ DEF:xxx={rrd1}:CpuUsagemhzNon:AVERAGE \ | |
LINE2:xxx#0000ff:"CpuUsagemhzNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskRdAvg.name=DiskRdAvg | |
report.vmware4.DiskRdAvg.columns=DiskRdAvg | |
@@ -1389,7 +1389,7 @@ DEF:xxx={rrd1}:DiskRdAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskRdAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskUsageNon.name=DiskUsageNon | |
report.vmware4.DiskUsageNon.columns=DiskUsageNon | |
@@ -1400,7 +1400,7 @@ DEF:xxx={rrd1}:DiskUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemActiveNon.name=MemActiveNon | |
report.vmware4.MemActiveNon.columns=MemActiveNon | |
@@ -1411,7 +1411,7 @@ DEF:xxx={rrd1}:MemActiveNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemActiveNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemConsumedNon.name=MemConsumedNon | |
report.vmware4.MemConsumedNon.columns=MemConsumedNon | |
@@ -1422,7 +1422,7 @@ DEF:xxx={rrd1}:MemConsumedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemConsumedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemGrantedNon.name=MemGrantedNon | |
report.vmware4.MemGrantedNon.columns=MemGrantedNon | |
@@ -1433,7 +1433,7 @@ DEF:xxx={rrd1}:MemGrantedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemGrantedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemOdNon.name=MemOdNon | |
report.vmware4.MemOdNon.columns=MemOdNon | |
@@ -1444,7 +1444,7 @@ DEF:xxx={rrd1}:MemOdNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemOdNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSharedNon.name=MemSharedNon | |
report.vmware4.MemSharedNon.columns=MemSharedNon | |
@@ -1455,7 +1455,7 @@ DEF:xxx={rrd1}:MemSharedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSharedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSpTtNon.name=MemSpTtNon | |
report.vmware4.MemSpTtNon.columns=MemSpTtNon | |
@@ -1466,7 +1466,7 @@ DEF:xxx={rrd1}:MemSpTtNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpTtNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSpinNon.name=MemSpinNon | |
report.vmware4.MemSpinNon.columns=MemSpinNon | |
@@ -1477,7 +1477,7 @@ DEF:xxx={rrd1}:MemSpinNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpinNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSpinReAvg.name=MemSpinReAvg | |
report.vmware4.MemSpinReAvg.columns=MemSpinReAvg | |
@@ -1488,7 +1488,7 @@ DEF:xxx={rrd1}:MemSpinReAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpinReAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSpoutNon.name=MemSpoutNon | |
report.vmware4.MemSpoutNon.columns=MemSpoutNon | |
@@ -1499,7 +1499,7 @@ DEF:xxx={rrd1}:MemSpoutNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpoutNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSpoutReAvg.name=MemSpoutReAvg | |
report.vmware4.MemSpoutReAvg.columns=MemSpoutReAvg | |
@@ -1510,7 +1510,7 @@ DEF:xxx={rrd1}:MemSpoutReAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSpoutReAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemSppedNon.name=MemSppedNon | |
report.vmware4.MemSppedNon.columns=MemSppedNon | |
@@ -1521,7 +1521,7 @@ DEF:xxx={rrd1}:MemSppedNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemSppedNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemUsageNon.name=MemUsageNon | |
report.vmware4.MemUsageNon.columns=MemUsageNon | |
@@ -1532,7 +1532,7 @@ DEF:xxx={rrd1}:MemUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemVmmemctlNon.name=MemVmmemctlNon | |
report.vmware4.MemVmmemctlNon.columns=MemVmmemctlNon | |
@@ -1543,7 +1543,7 @@ DEF:xxx={rrd1}:MemVmmemctlNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemVmmemctlNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemVmmemctlTtNon.name=MemVmmemctlTtNon | |
report.vmware4.MemVmmemctlTtNon.columns=MemVmmemctlTtNon | |
@@ -1554,7 +1554,7 @@ DEF:xxx={rrd1}:MemVmmemctlTtNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemVmmemctlTtNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.MemZeroNon.name=MemZeroNon | |
report.vmware4.MemZeroNon.columns=MemZeroNon | |
@@ -1565,7 +1565,7 @@ DEF:xxx={rrd1}:MemZeroNon:AVERAGE \ | |
LINE2:xxx#0000ff:"MemZeroNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.NetReceivedAvg.name=NetReceivedAvg | |
report.vmware4.NetReceivedAvg.columns=NetReceivedAvg | |
@@ -1576,7 +1576,7 @@ DEF:xxx={rrd1}:NetReceivedAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"NetReceivedAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.NetUsageNon.name=NetUsageNon | |
report.vmware4.NetUsageNon.columns=NetUsageNon | |
@@ -1587,7 +1587,7 @@ DEF:xxx={rrd1}:NetUsageNon:AVERAGE \ | |
LINE2:xxx#0000ff:"NetUsageNon" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuActav15Lat.name=ResCpuActav15Lat | |
report.vmware4.ResCpuActav15Lat.columns=ResCpuActav15Lat | |
@@ -1598,7 +1598,7 @@ DEF:xxx={rrd1}:ResCpuActav15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActav15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuActav1Lat.name=ResCpuActav1Lat | |
report.vmware4.ResCpuActav1Lat.columns=ResCpuActav1Lat | |
@@ -1609,7 +1609,7 @@ DEF:xxx={rrd1}:ResCpuActav1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActav1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuActav5Lat.name=ResCpuActav5Lat | |
report.vmware4.ResCpuActav5Lat.columns=ResCpuActav5Lat | |
@@ -1620,7 +1620,7 @@ DEF:xxx={rrd1}:ResCpuActav5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActav5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuActpk15Lat.name=ResCpuActpk15Lat | |
report.vmware4.ResCpuActpk15Lat.columns=ResCpuActpk15Lat | |
@@ -1631,7 +1631,7 @@ DEF:xxx={rrd1}:ResCpuActpk15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActpk15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuActpk1Lat.name=ResCpuActpk1Lat | |
report.vmware4.ResCpuActpk1Lat.columns=ResCpuActpk1Lat | |
@@ -1642,7 +1642,7 @@ DEF:xxx={rrd1}:ResCpuActpk1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActpk1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuActpk5Lat.name=ResCpuActpk5Lat | |
report.vmware4.ResCpuActpk5Lat.columns=ResCpuActpk5Lat | |
@@ -1653,7 +1653,7 @@ DEF:xxx={rrd1}:ResCpuActpk5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuActpk5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuMaxLd15Lat.name=ResCpuMaxLd15Lat | |
report.vmware4.ResCpuMaxLd15Lat.columns=ResCpuMaxLd15Lat | |
@@ -1664,7 +1664,7 @@ DEF:xxx={rrd1}:ResCpuMaxLd15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuMaxLd15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuMaxLd1Lat.name=ResCpuMaxLd1Lat | |
report.vmware4.ResCpuMaxLd1Lat.columns=ResCpuMaxLd1Lat | |
@@ -1675,7 +1675,7 @@ DEF:xxx={rrd1}:ResCpuMaxLd1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuMaxLd1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuMaxLd5Lat.name=ResCpuMaxLd5Lat | |
report.vmware4.ResCpuMaxLd5Lat.columns=ResCpuMaxLd5Lat | |
@@ -1686,7 +1686,7 @@ DEF:xxx={rrd1}:ResCpuMaxLd5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuMaxLd5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuRunav15Lat.name=ResCpuRunav15Lat | |
report.vmware4.ResCpuRunav15Lat.columns=ResCpuRunav15Lat | |
@@ -1697,7 +1697,7 @@ DEF:xxx={rrd1}:ResCpuRunav15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunav15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuRunav1Lat.name=ResCpuRunav1Lat | |
report.vmware4.ResCpuRunav1Lat.columns=ResCpuRunav1Lat | |
@@ -1708,7 +1708,7 @@ DEF:xxx={rrd1}:ResCpuRunav1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunav1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuRunav5Lat.name=ResCpuRunav5Lat | |
report.vmware4.ResCpuRunav5Lat.columns=ResCpuRunav5Lat | |
@@ -1719,7 +1719,7 @@ DEF:xxx={rrd1}:ResCpuRunav5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunav5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuRunpk15Lat.name=ResCpuRunpk15Lat | |
report.vmware4.ResCpuRunpk15Lat.columns=ResCpuRunpk15Lat | |
@@ -1730,7 +1730,7 @@ DEF:xxx={rrd1}:ResCpuRunpk15Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunpk15Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuRunpk1Lat.name=ResCpuRunpk1Lat | |
report.vmware4.ResCpuRunpk1Lat.columns=ResCpuRunpk1Lat | |
@@ -1741,7 +1741,7 @@ DEF:xxx={rrd1}:ResCpuRunpk1Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunpk1Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuRunpk5Lat.name=ResCpuRunpk5Lat | |
report.vmware4.ResCpuRunpk5Lat.columns=ResCpuRunpk5Lat | |
@@ -1752,7 +1752,7 @@ DEF:xxx={rrd1}:ResCpuRunpk5Lat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuRunpk5Lat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuSeCtLat.name=ResCpuSeCtLat | |
report.vmware4.ResCpuSeCtLat.columns=ResCpuSeCtLat | |
@@ -1763,7 +1763,7 @@ DEF:xxx={rrd1}:ResCpuSeCtLat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuSeCtLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.ResCpuSePeriodLat.name=ResCpuSePeriodLat | |
report.vmware4.ResCpuSePeriodLat.columns=ResCpuSePeriodLat | |
@@ -1774,7 +1774,7 @@ DEF:xxx={rrd1}:ResCpuSePeriodLat:AVERAGE \ | |
LINE2:xxx#0000ff:"ResCpuSePeriodLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysHeartbeatSum.name=SysHeartbeatSum | |
report.vmware4.SysHeartbeatSum.columns=SysHeartbeatSum | |
@@ -1785,7 +1785,7 @@ DEF:xxx={rrd1}:SysHeartbeatSum:AVERAGE \ | |
LINE2:xxx#0000ff:"SysHeartbeatSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.SysUptimeLat.name=SysUptimeLat | |
report.vmware4.SysUptimeLat.columns=SysUptimeLat | |
@@ -1796,7 +1796,7 @@ DEF:xxx={rrd1}:SysUptimeLat:AVERAGE \ | |
LINE2:xxx#0000ff:"SysUptimeLat" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.NetPacketsRxSum.name=NetPacketsRxSum | |
report.vmware4.NetPacketsRxSum.columns=NetPacketsRxSum | |
@@ -1808,7 +1808,7 @@ DEF:xxx={rrd1}:NetPacketsRxSum:AVERAGE \ | |
LINE2:xxx#0000ff:"NetPacketsRxSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.NetPacketsTxSum.name=NetPacketsTxSum | |
report.vmware4.NetPacketsTxSum.columns=NetPacketsTxSum | |
@@ -1820,7 +1820,7 @@ DEF:xxx={rrd1}:NetPacketsTxSum:AVERAGE \ | |
LINE2:xxx#0000ff:"NetPacketsTxSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.NetReceivedAvg.name=NetReceivedAvg | |
report.vmware4.NetReceivedAvg.columns=NetReceivedAvg | |
@@ -1832,7 +1832,7 @@ DEF:xxx={rrd1}:NetReceivedAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"NetReceivedAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.NetTransmittedAvg.name=NetTransmittedAvg | |
report.vmware4.NetTransmittedAvg.columns=NetTransmittedAvg | |
@@ -1844,7 +1844,7 @@ DEF:xxx={rrd1}:NetTransmittedAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"NetTransmittedAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskBusResetsSum.name=DiskBusResetsSum | |
report.vmware4.DiskBusResetsSum.columns=DiskBusResetsSum | |
@@ -1856,7 +1856,7 @@ DEF:xxx={rrd1}:DiskBusResetsSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskBusResetsSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskCsAdSum.name=DiskCsAdSum | |
report.vmware4.DiskCsAdSum.columns=DiskCsAdSum | |
@@ -1868,7 +1868,7 @@ DEF:xxx={rrd1}:DiskCsAdSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskCsAdSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskCsSum.name=DiskCsSum | |
report.vmware4.DiskCsSum.columns=DiskCsSum | |
@@ -1880,7 +1880,7 @@ DEF:xxx={rrd1}:DiskCsSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskCsSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskNrRdSum.name=DiskNrRdSum | |
report.vmware4.DiskNrRdSum.columns=DiskNrRdSum | |
@@ -1892,7 +1892,7 @@ DEF:xxx={rrd1}:DiskNrRdSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskNrRdSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskNrWeSum.name=DiskNrWeSum | |
report.vmware4.DiskNrWeSum.columns=DiskNrWeSum | |
@@ -1904,7 +1904,7 @@ DEF:xxx={rrd1}:DiskNrWeSum:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskNrWeSum" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskRdAvg.name=DiskRdAvg | |
report.vmware4.DiskRdAvg.columns=DiskRdAvg | |
@@ -1916,7 +1916,7 @@ DEF:xxx={rrd1}:DiskRdAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskRdAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
report.vmware4.DiskWeAvg.name=DiskWeAvg | |
report.vmware4.DiskWeAvg.columns=DiskWeAvg | |
@@ -1928,5 +1928,5 @@ DEF:xxx={rrd1}:DiskWeAvg:AVERAGE \ | |
LINE2:xxx#0000ff:"DiskWeAvg" \ | |
GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
-GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
diff --git a/snmp-graph.properties.d/vmware6-graph-simple.properties b/snmp-graph.properties.d/vmware6-graph-simple.properties | |
new file mode 100644 | |
index 0000000..6268027 | |
--- /dev/null | |
+++ b/snmp-graph.properties.d/vmware6-graph-simple.properties | |
@@ -0,0 +1,2644 @@ | |
+reports=vmware6.VrtDiskLeSsLat, \ | |
+vmware6.VrtDiskMmSsLat, \ | |
+vmware6.VrtDiskNrRdAdAvg, \ | |
+vmware6.VrtDiskNrWeAdAvg, \ | |
+vmware6.VrtDiskRdAvg, \ | |
+vmware6.VrtDiskRdIOSizeLat, \ | |
+vmware6.VrtDiskRdLdMcLat, \ | |
+vmware6.VrtDiskRdLyUSLat, \ | |
+vmware6.VrtDiskRdOIOLat, \ | |
+vmware6.VrtDiskSlSsLat, \ | |
+vmware6.VrtDiskTlRdLyAvg, \ | |
+vmware6.VrtDiskTlWeLyAvg, \ | |
+vmware6.VrtDiskWeAvg, \ | |
+vmware6.VrtDiskWeIOSizeLat, \ | |
+vmware6.VrtDiskWeLdMcLat, \ | |
+vmware6.VrtDiskWeLyUSLat, \ | |
+vmware6.VrtDiskWeOIOLat, \ | |
+vmware6.CpuDemandAvg, \ | |
+vmware6.CpuDmdEntRatioLat, \ | |
+vmware6.CpuEntitlementLat, \ | |
+vmware6.CpuLyAvg, \ | |
+vmware6.CpuOverlapSum, \ | |
+vmware6.CpuRdinessAvg, \ | |
+vmware6.CpuSpwaitSum, \ | |
+vmware6.CpuUsageAvg, \ | |
+vmware6.CpuUsagemhzAvg, \ | |
+vmware6.CpuUsedSum, \ | |
+vmware6.CpuWaitSum, \ | |
+vmware6.DaStMaxTlLyLat, \ | |
+vmware6.DiskMaxTlLyLat, \ | |
+vmware6.DiskRdAvg, \ | |
+vmware6.DiskUsageAvg, \ | |
+vmware6.DiskWeAvg, \ | |
+vmware6.MemAeAvg, \ | |
+vmware6.MemAeWeAvg, \ | |
+vmware6.MemCdAvg, \ | |
+vmware6.MemCnReAvg, \ | |
+vmware6.MemCompressedAvg, \ | |
+vmware6.MemDnReAvg, \ | |
+vmware6.MemEntitlementAvg, \ | |
+vmware6.MemGrantedAvg, \ | |
+vmware6.MemLlSpInReAvg, \ | |
+vmware6.MemLlSpOutReAvg, \ | |
+vmware6.MemLlSpUsedAvg, \ | |
+vmware6.MemLyAvg, \ | |
+vmware6.MemOdAvg, \ | |
+vmware6.MemOdMaxAvg, \ | |
+vmware6.MemOdTdAvg, \ | |
+vmware6.MemSharedAvg, \ | |
+vmware6.MemSpTtAvg, \ | |
+vmware6.MemSpinAvg, \ | |
+vmware6.MemSpinReAvg, \ | |
+vmware6.MemSpoutAvg, \ | |
+vmware6.MemSpoutReAvg, \ | |
+vmware6.MemSppedAvg, \ | |
+vmware6.MemUsageAvg, \ | |
+vmware6.MemVmmemctlAvg, \ | |
+vmware6.MemVmmemctlTtAvg, \ | |
+vmware6.MemZeroAvg, \ | |
+vmware6.MemZipSavedLat, \ | |
+vmware6.MemZippedLat, \ | |
+vmware6.NetBroadcastRxSum, \ | |
+vmware6.NetBroadcastTxSum, \ | |
+vmware6.NetDroppedRxSum, \ | |
+vmware6.NetDroppedTxSum, \ | |
+vmware6.NetPacketsTxSum, \ | |
+vmware6.NetReceivedAvg, \ | |
+vmware6.NetTransmittedAvg, \ | |
+vmware6.PowerEnergySum, \ | |
+vmware6.PowerPowerAvg, \ | |
+vmware6.ResCpuActav15Lat, \ | |
+vmware6.ResCpuActav1Lat, \ | |
+vmware6.ResCpuActav5Lat, \ | |
+vmware6.ResCpuActpk15Lat, \ | |
+vmware6.ResCpuActpk1Lat, \ | |
+vmware6.ResCpuActpk5Lat, \ | |
+vmware6.ResCpuMaxLd15Lat, \ | |
+vmware6.ResCpuMaxLd1Lat, \ | |
+vmware6.ResCpuMaxLd5Lat, \ | |
+vmware6.ResCpuRunav15Lat, \ | |
+vmware6.ResCpuRunav1Lat, \ | |
+vmware6.ResCpuRunav5Lat, \ | |
+vmware6.ResCpuRunpk15Lat, \ | |
+vmware6.ResCpuRunpk1Lat, \ | |
+vmware6.ResCpuRunpk5Lat, \ | |
+vmware6.ResCpuSeCtLat, \ | |
+vmware6.ResCpuSePeriodLat, \ | |
+vmware6.SysHeartbeatLat, \ | |
+vmware6.SysOsUpTeLat, \ | |
+vmware6.SysUpTeLat, \ | |
+vmware6.CpuCostopSum, \ | |
+vmware6.CpuIdleSum, \ | |
+vmware6.CpuMaxLdSum, \ | |
+vmware6.CpuRdySum, \ | |
+vmware6.CpuRunSum, \ | |
+vmware6.CpuSystemSum, \ | |
+vmware6.NetBytesRxAvg, \ | |
+vmware6.NetBytesTxAvg, \ | |
+vmware6.NetMulticastRxSum, \ | |
+vmware6.NetMulticastTxSum, \ | |
+vmware6.NetPacketsRxSum, \ | |
+vmware6.NetUsageAvg, \ | |
+vmware6.DiskBusResetsSum, \ | |
+vmware6.DiskCsAdAvg, \ | |
+vmware6.DiskCsAdSum, \ | |
+vmware6.DiskCsSum, \ | |
+vmware6.DiskNrRdAdAvg, \ | |
+vmware6.DiskNrRdSum, \ | |
+vmware6.DiskNrWeAdAvg, \ | |
+vmware6.DiskNrWeSum, \ | |
+vmware6.DaStNrRdAdAvg, \ | |
+vmware6.DaStNrWeAdAvg, \ | |
+vmware6.DaStRdAvg, \ | |
+vmware6.DaStTlRdLyAvg, \ | |
+vmware6.DaStTlWeLyAvg, \ | |
+vmware6.DaStWeAvg, \ | |
+vmware6.StAdptrCsAdAvg, \ | |
+vmware6.StAdptrNrRdAdAvg, \ | |
+vmware6.StAdptrNrWeAdAvg, \ | |
+vmware6.StAdptrRdAvg, \ | |
+vmware6.StAdptrTlRdLyAvg, \ | |
+vmware6.StAdptrTlWeLyAvg, \ | |
+vmware6.StAdptrWeAvg, \ | |
+vmware6.StPthCsAdAvg, \ | |
+vmware6.StPthNrRdAdAvg, \ | |
+vmware6.StPthNrWeAdAvg, \ | |
+vmware6.StPthRdAvg, \ | |
+vmware6.StPthTlRdLyAvg, \ | |
+vmware6.StPthTlWeLyAvg, \ | |
+vmware6.StPthWeAvg, \ | |
+vmware6.CpuRdCyAvg, \ | |
+vmware6.CpuTlCyAvg, \ | |
+vmware6.HbrHbrNetRxAvg, \ | |
+vmware6.HbrHbrNetTxAvg, \ | |
+vmware6.HbrHbrNumVmsAvg, \ | |
+vmware6.MemHeapAvg, \ | |
+vmware6.MemHeapfreeAvg, \ | |
+vmware6.MemLlSpInAvg, \ | |
+vmware6.MemLlSpOutAvg, \ | |
+vmware6.MemLowfreeTdAvg, \ | |
+vmware6.MemRdCyAvg, \ | |
+vmware6.MemSharedcommonAvg, \ | |
+vmware6.MemSpusedAvg, \ | |
+vmware6.MemStateLat, \ | |
+vmware6.MemSysUsageAvg, \ | |
+vmware6.MemTlCyAvg, \ | |
+vmware6.MemUdAvg, \ | |
+vmware6.MemVmfs.pbc.OdLat, \ | |
+vmware6.MemVmfsPbcCpMsRtiLt, \ | |
+vmware6.MemVmfs.pbc.sizeLat, \ | |
+vmware6.MemVmfsPbcSizMaxLat, \ | |
+vmware6.MemVmfsPbcWrkSetLat, \ | |
+vmware6.MemVmfsPbcWrkStMxLt, \ | |
+vmware6.PowerPowerCapAvg, \ | |
+vmware6.StAdptrMaxTlLyLat, \ | |
+vmware6.StPthMaxTlLyLat, \ | |
+vmware6.CpuCoreUnAvg, \ | |
+vmware6.CpuUnAvg, \ | |
+vmware6.SysReCpuAcMinLat, \ | |
+vmware6.SysReCpuAcSsLat, \ | |
+vmware6.SysReCpuAct1Lat, \ | |
+vmware6.SysReCpuAct5Lat, \ | |
+vmware6.SysReCpuMaxLd1Lat, \ | |
+vmware6.SysReCpuMaxLd5Lat, \ | |
+vmware6.SysReCpuRun1Lat, \ | |
+vmware6.SysReCpuRun5Lat, \ | |
+vmware6.SysReCpuUsageAvg, \ | |
+vmware6.SysReFdUsageLat, \ | |
+vmware6.SysReMemAcMaxLat, \ | |
+vmware6.SysReMemAcMinLat, \ | |
+vmware6.SysReMemAcSsLat, \ | |
+vmware6.SysReMemCdLat, \ | |
+vmware6.SysReMemCowLat, \ | |
+vmware6.SysReMemMappedLat, \ | |
+vmware6.SysReMemOdLat, \ | |
+vmware6.SysReMemSharedLat, \ | |
+vmware6.SysReMemSppedLat, \ | |
+vmware6.SysReMemTdLat, \ | |
+vmware6.SysReMemZeroLat, \ | |
+vmware6.NetErrorsRxSum, \ | |
+vmware6.NetErrorsTxSum, \ | |
+vmware6.NetUnknownPsSum, \ | |
+vmware6.DiskDeLyAvg, \ | |
+vmware6.DiskDeRdLyAvg, \ | |
+vmware6.DiskDeWeLyAvg, \ | |
+vmware6.DiskKlLyAvg, \ | |
+vmware6.DiskKlRdLyAvg, \ | |
+vmware6.DiskKlWeLyAvg, \ | |
+vmware6.DiskMaxQeDhAvg, \ | |
+vmware6.DiskQeLyAvg, \ | |
+vmware6.DiskQeRdLyAvg, \ | |
+vmware6.DiskQeWeLyAvg, \ | |
+vmware6.DiskTlLyAvg, \ | |
+vmware6.DiskTlRdLyAvg, \ | |
+vmware6.DiskTlWeLyAvg, \ | |
+vmware6.vflModNumAeVMDKsLat, \ | |
+vmware6.DaStDeIopsAvg, \ | |
+vmware6.DaStDeMaxQeDhLat, \ | |
+vmware6.DaStDeNlRdLyLat, \ | |
+vmware6.DaStDeNlWeLyLat, \ | |
+vmware6.DaStDeRdBytesLat, \ | |
+vmware6.DaStDeRdIopsLat, \ | |
+vmware6.DaStDeRdLdMcLat, \ | |
+vmware6.DaStDeRdOIOLat, \ | |
+vmware6.DaStDeVMOdLyLat, \ | |
+vmware6.DaStDeWeBytesLat, \ | |
+vmware6.DaStDeWeIopsLat, \ | |
+vmware6.DaStDeWeLdMcLat, \ | |
+vmware6.DaStDeWeOIOLat, \ | |
+vmware6.DaStSiocAeTePeAvg, \ | |
+vmware6.DaStSizeNdDeLyAvg | |
+ | |
+report.vmware6.VrtDiskLeSsLat.name=vmware6.virtualDisk.largeSeeks.latest | |
+report.vmware6.VrtDiskLeSsLat.columns=VrtDiskLeSsLat | |
+report.vmware6.VrtDiskLeSsLat.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskLeSsLat.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskLeSsLat.command=--title="VMware6 virtualDisk.largeSeeks.latest {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskLeSsLat" \ | |
+DEF:xxx={rrd1}:VrtDiskLeSsLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskLeSsLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskMmSsLat.name=vmware6.virtualDisk.mediumSeeks.latest | |
+report.vmware6.VrtDiskMmSsLat.columns=VrtDiskMmSsLat | |
+report.vmware6.VrtDiskMmSsLat.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskMmSsLat.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskMmSsLat.command=--title="VMware6 virtualDisk.mediumSeeks.latest {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskMmSsLat" \ | |
+DEF:xxx={rrd1}:VrtDiskMmSsLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskMmSsLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskNrRdAdAvg.name=vmware6.virtualDisk.numberReadAveraged.average | |
+report.vmware6.VrtDiskNrRdAdAvg.columns=VrtDiskNrRdAdAvg | |
+report.vmware6.VrtDiskNrRdAdAvg.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskNrRdAdAvg.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskNrRdAdAvg.command=--title="VMware6 virtualDisk.numberReadAveraged.average {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskNrRdAdAvg" \ | |
+DEF:xxx={rrd1}:VrtDiskNrRdAdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskNrRdAdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskNrWeAdAvg.name=vmware6.virtualDisk.numberWriteAveraged.average | |
+report.vmware6.VrtDiskNrWeAdAvg.columns=VrtDiskNrWeAdAvg | |
+report.vmware6.VrtDiskNrWeAdAvg.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskNrWeAdAvg.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskNrWeAdAvg.command=--title="VMware6 virtualDisk.numberWriteAveraged.average {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskNrWeAdAvg" \ | |
+DEF:xxx={rrd1}:VrtDiskNrWeAdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskNrWeAdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskRdAvg.name=vmware6.virtualDisk.read.average | |
+report.vmware6.VrtDiskRdAvg.columns=VrtDiskRdAvg | |
+report.vmware6.VrtDiskRdAvg.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskRdAvg.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskRdAvg.command=--title="VMware6 virtualDisk.read.average {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskRdAvg" \ | |
+DEF:xxx={rrd1}:VrtDiskRdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskRdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskRdIOSizeLat.name=vmware6.virtualDisk.readIOSize.latest | |
+report.vmware6.VrtDiskRdIOSizeLat.columns=VrtDiskRdIOSizeLat | |
+report.vmware6.VrtDiskRdIOSizeLat.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskRdIOSizeLat.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskRdIOSizeLat.command=--title="VMware6 virtualDisk.readIOSize.latest {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskRdIOSizeLat" \ | |
+DEF:xxx={rrd1}:VrtDiskRdIOSizeLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskRdIOSizeLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskRdLdMcLat.name=vmware6.virtualDisk.readLoadMetric.latest | |
+report.vmware6.VrtDiskRdLdMcLat.columns=VrtDiskRdLdMcLat | |
+report.vmware6.VrtDiskRdLdMcLat.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskRdLdMcLat.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskRdLdMcLat.command=--title="VMware6 virtualDisk.readLoadMetric.latest {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskRdLdMcLat" \ | |
+DEF:xxx={rrd1}:VrtDiskRdLdMcLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskRdLdMcLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskRdLyUSLat.name=vmware6.virtualDisk.readLatencyUS.latest | |
+report.vmware6.VrtDiskRdLyUSLat.columns=VrtDiskRdLyUSLat | |
+report.vmware6.VrtDiskRdLyUSLat.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskRdLyUSLat.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskRdLyUSLat.command=--title="VMware6 virtualDisk.readLatencyUS.latest {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskRdLyUSLat" \ | |
+DEF:xxx={rrd1}:VrtDiskRdLyUSLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskRdLyUSLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskRdOIOLat.name=vmware6.virtualDisk.readOIO.latest | |
+report.vmware6.VrtDiskRdOIOLat.columns=VrtDiskRdOIOLat | |
+report.vmware6.VrtDiskRdOIOLat.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskRdOIOLat.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskRdOIOLat.command=--title="VMware6 virtualDisk.readOIO.latest {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskRdOIOLat" \ | |
+DEF:xxx={rrd1}:VrtDiskRdOIOLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskRdOIOLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskSlSsLat.name=vmware6.virtualDisk.smallSeeks.latest | |
+report.vmware6.VrtDiskSlSsLat.columns=VrtDiskSlSsLat | |
+report.vmware6.VrtDiskSlSsLat.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskSlSsLat.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskSlSsLat.command=--title="VMware6 virtualDisk.smallSeeks.latest {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskSlSsLat" \ | |
+DEF:xxx={rrd1}:VrtDiskSlSsLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskSlSsLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskTlRdLyAvg.name=vmware6.virtualDisk.totalReadLatency.average | |
+report.vmware6.VrtDiskTlRdLyAvg.columns=VrtDiskTlRdLyAvg | |
+report.vmware6.VrtDiskTlRdLyAvg.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskTlRdLyAvg.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskTlRdLyAvg.command=--title="VMware6 virtualDisk.totalReadLatency.average {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskTlRdLyAvg" \ | |
+DEF:xxx={rrd1}:VrtDiskTlRdLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskTlRdLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskTlWeLyAvg.name=vmware6.virtualDisk.totalWriteLatency.average | |
+report.vmware6.VrtDiskTlWeLyAvg.columns=VrtDiskTlWeLyAvg | |
+report.vmware6.VrtDiskTlWeLyAvg.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskTlWeLyAvg.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskTlWeLyAvg.command=--title="VMware6 virtualDisk.totalWriteLatency.average {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskTlWeLyAvg" \ | |
+DEF:xxx={rrd1}:VrtDiskTlWeLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskTlWeLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskWeAvg.name=vmware6.virtualDisk.write.average | |
+report.vmware6.VrtDiskWeAvg.columns=VrtDiskWeAvg | |
+report.vmware6.VrtDiskWeAvg.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskWeAvg.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskWeAvg.command=--title="VMware6 virtualDisk.write.average {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskWeAvg" \ | |
+DEF:xxx={rrd1}:VrtDiskWeAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskWeAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskWeIOSizeLat.name=vmware6.virtualDisk.writeIOSize.latest | |
+report.vmware6.VrtDiskWeIOSizeLat.columns=VrtDiskWeIOSizeLat | |
+report.vmware6.VrtDiskWeIOSizeLat.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskWeIOSizeLat.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskWeIOSizeLat.command=--title="VMware6 virtualDisk.writeIOSize.latest {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskWeIOSizeLat" \ | |
+DEF:xxx={rrd1}:VrtDiskWeIOSizeLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskWeIOSizeLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskWeLdMcLat.name=vmware6.virtualDisk.writeLoadMetric.latest | |
+report.vmware6.VrtDiskWeLdMcLat.columns=VrtDiskWeLdMcLat | |
+report.vmware6.VrtDiskWeLdMcLat.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskWeLdMcLat.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskWeLdMcLat.command=--title="VMware6 virtualDisk.writeLoadMetric.latest {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskWeLdMcLat" \ | |
+DEF:xxx={rrd1}:VrtDiskWeLdMcLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskWeLdMcLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskWeLyUSLat.name=vmware6.virtualDisk.writeLatencyUS.latest | |
+report.vmware6.VrtDiskWeLyUSLat.columns=VrtDiskWeLyUSLat | |
+report.vmware6.VrtDiskWeLyUSLat.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskWeLyUSLat.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskWeLyUSLat.command=--title="VMware6 virtualDisk.writeLatencyUS.latest {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskWeLyUSLat" \ | |
+DEF:xxx={rrd1}:VrtDiskWeLyUSLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskWeLyUSLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.VrtDiskWeOIOLat.name=vmware6.virtualDisk.writeOIO.latest | |
+report.vmware6.VrtDiskWeOIOLat.columns=VrtDiskWeOIOLat | |
+report.vmware6.VrtDiskWeOIOLat.propertiesValues=vmware6VrtDiskName | |
+report.vmware6.VrtDiskWeOIOLat.type=vmware6VrtDisk | |
+report.vmware6.VrtDiskWeOIOLat.command=--title="VMware6 virtualDisk.writeOIO.latest {vmware6VrtDiskName}" \ | |
+--vertical-label="VrtDiskWeOIOLat" \ | |
+DEF:xxx={rrd1}:VrtDiskWeOIOLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"VrtDiskWeOIOLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuDemandAvg.name=vmware6.cpu.demand.average | |
+report.vmware6.CpuDemandAvg.columns=CpuDemandAvg | |
+report.vmware6.CpuDemandAvg.type=nodeSnmp | |
+report.vmware6.CpuDemandAvg.command=--title="VMware6 cpu.demand.average" \ | |
+--vertical-label="CpuDemandAvg" \ | |
+DEF:xxx={rrd1}:CpuDemandAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuDemandAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuDmdEntRatioLat.name=vmware6.cpu.demandEntitlementRatio.latest | |
+report.vmware6.CpuDmdEntRatioLat.columns=CpuDmdEntRatioLat | |
+report.vmware6.CpuDmdEntRatioLat.type=nodeSnmp | |
+report.vmware6.CpuDmdEntRatioLat.command=--title="VMware6 cpu.demandEntitlementRatio.latest" \ | |
+--vertical-label="CpuDmdEntRatioLat" \ | |
+DEF:xxx={rrd1}:CpuDmdEntRatioLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuDmdEntRatioLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuEntitlementLat.name=vmware6.cpu.entitlement.latest | |
+report.vmware6.CpuEntitlementLat.columns=CpuEntitlementLat | |
+report.vmware6.CpuEntitlementLat.type=nodeSnmp | |
+report.vmware6.CpuEntitlementLat.command=--title="VMware6 cpu.entitlement.latest" \ | |
+--vertical-label="CpuEntitlementLat" \ | |
+DEF:xxx={rrd1}:CpuEntitlementLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuEntitlementLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuLyAvg.name=vmware6.cpu.latency.average | |
+report.vmware6.CpuLyAvg.columns=CpuLyAvg | |
+report.vmware6.CpuLyAvg.type=nodeSnmp | |
+report.vmware6.CpuLyAvg.command=--title="VMware6 cpu.latency.average" \ | |
+--vertical-label="CpuLyAvg" \ | |
+DEF:xxx={rrd1}:CpuLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuOverlapSum.name=vmware6.cpu.overlap.summation | |
+report.vmware6.CpuOverlapSum.columns=CpuOverlapSum | |
+report.vmware6.CpuOverlapSum.type=nodeSnmp | |
+report.vmware6.CpuOverlapSum.command=--title="VMware6 cpu.overlap.summation" \ | |
+--vertical-label="CpuOverlapSum" \ | |
+DEF:xxx={rrd1}:CpuOverlapSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuOverlapSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuRdinessAvg.name=vmware6.cpu.readiness.average | |
+report.vmware6.CpuRdinessAvg.columns=CpuRdinessAvg | |
+report.vmware6.CpuRdinessAvg.type=nodeSnmp | |
+report.vmware6.CpuRdinessAvg.command=--title="VMware6 cpu.readiness.average" \ | |
+--vertical-label="CpuRdinessAvg" \ | |
+DEF:xxx={rrd1}:CpuRdinessAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuRdinessAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuSpwaitSum.name=vmware6.cpu.swapwait.summation | |
+report.vmware6.CpuSpwaitSum.columns=CpuSpwaitSum | |
+report.vmware6.CpuSpwaitSum.type=nodeSnmp | |
+report.vmware6.CpuSpwaitSum.command=--title="VMware6 cpu.swapwait.summation" \ | |
+--vertical-label="CpuSpwaitSum" \ | |
+DEF:xxx={rrd1}:CpuSpwaitSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuSpwaitSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuUsageAvg.name=vmware6.cpu.usage.average | |
+report.vmware6.CpuUsageAvg.columns=CpuUsageAvg | |
+report.vmware6.CpuUsageAvg.type=nodeSnmp | |
+report.vmware6.CpuUsageAvg.command=--title="VMware6 cpu.usage.average" \ | |
+--vertical-label="CpuUsageAvg" \ | |
+DEF:xxx={rrd1}:CpuUsageAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuUsageAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuUsagemhzAvg.name=vmware6.cpu.usagemhz.average | |
+report.vmware6.CpuUsagemhzAvg.columns=CpuUsagemhzAvg | |
+report.vmware6.CpuUsagemhzAvg.type=nodeSnmp | |
+report.vmware6.CpuUsagemhzAvg.command=--title="VMware6 cpu.usagemhz.average" \ | |
+--vertical-label="CpuUsagemhzAvg" \ | |
+DEF:xxx={rrd1}:CpuUsagemhzAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuUsagemhzAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuUsedSum.name=vmware6.cpu.used.summation | |
+report.vmware6.CpuUsedSum.columns=CpuUsedSum | |
+report.vmware6.CpuUsedSum.type=nodeSnmp | |
+report.vmware6.CpuUsedSum.command=--title="VMware6 cpu.used.summation" \ | |
+--vertical-label="CpuUsedSum" \ | |
+DEF:xxx={rrd1}:CpuUsedSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuUsedSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuWaitSum.name=vmware6.cpu.wait.summation | |
+report.vmware6.CpuWaitSum.columns=CpuWaitSum | |
+report.vmware6.CpuWaitSum.type=nodeSnmp | |
+report.vmware6.CpuWaitSum.command=--title="VMware6 cpu.wait.summation" \ | |
+--vertical-label="CpuWaitSum" \ | |
+DEF:xxx={rrd1}:CpuWaitSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuWaitSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStMaxTlLyLat.name=vmware6.datastore.maxTotalLatency.latest | |
+report.vmware6.DaStMaxTlLyLat.columns=DaStMaxTlLyLat | |
+report.vmware6.DaStMaxTlLyLat.type=nodeSnmp | |
+report.vmware6.DaStMaxTlLyLat.command=--title="VMware6 datastore.maxTotalLatency.latest" \ | |
+--vertical-label="DaStMaxTlLyLat" \ | |
+DEF:xxx={rrd1}:DaStMaxTlLyLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStMaxTlLyLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskMaxTlLyLat.name=vmware6.disk.maxTotalLatency.latest | |
+report.vmware6.DiskMaxTlLyLat.columns=DiskMaxTlLyLat | |
+report.vmware6.DiskMaxTlLyLat.type=nodeSnmp | |
+report.vmware6.DiskMaxTlLyLat.command=--title="VMware6 disk.maxTotalLatency.latest" \ | |
+--vertical-label="DiskMaxTlLyLat" \ | |
+DEF:xxx={rrd1}:DiskMaxTlLyLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskMaxTlLyLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskRdAvg.name=vmware6.disk.read.average | |
+report.vmware6.DiskRdAvg.columns=DiskRdAvg | |
+report.vmware6.DiskRdAvg.type=nodeSnmp | |
+report.vmware6.DiskRdAvg.command=--title="VMware6 disk.read.average" \ | |
+--vertical-label="DiskRdAvg" \ | |
+DEF:xxx={rrd1}:DiskRdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskRdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskUsageAvg.name=vmware6.disk.usage.average | |
+report.vmware6.DiskUsageAvg.columns=DiskUsageAvg | |
+report.vmware6.DiskUsageAvg.type=nodeSnmp | |
+report.vmware6.DiskUsageAvg.command=--title="VMware6 disk.usage.average" \ | |
+--vertical-label="DiskUsageAvg" \ | |
+DEF:xxx={rrd1}:DiskUsageAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskUsageAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskWeAvg.name=vmware6.disk.write.average | |
+report.vmware6.DiskWeAvg.columns=DiskWeAvg | |
+report.vmware6.DiskWeAvg.type=nodeSnmp | |
+report.vmware6.DiskWeAvg.command=--title="VMware6 disk.write.average" \ | |
+--vertical-label="DiskWeAvg" \ | |
+DEF:xxx={rrd1}:DiskWeAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskWeAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemAeAvg.name=vmware6.mem.active.average | |
+report.vmware6.MemAeAvg.columns=MemAeAvg | |
+report.vmware6.MemAeAvg.type=nodeSnmp | |
+report.vmware6.MemAeAvg.command=--title="VMware6 mem.active.average" \ | |
+--vertical-label="MemAeAvg" \ | |
+DEF:xxx={rrd1}:MemAeAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemAeAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemAeWeAvg.name=vmware6.mem.activewrite.average | |
+report.vmware6.MemAeWeAvg.columns=MemAeWeAvg | |
+report.vmware6.MemAeWeAvg.type=nodeSnmp | |
+report.vmware6.MemAeWeAvg.command=--title="VMware6 mem.activewrite.average" \ | |
+--vertical-label="MemAeWeAvg" \ | |
+DEF:xxx={rrd1}:MemAeWeAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemAeWeAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemCdAvg.name=vmware6.mem.consumed.average | |
+report.vmware6.MemCdAvg.columns=MemCdAvg | |
+report.vmware6.MemCdAvg.type=nodeSnmp | |
+report.vmware6.MemCdAvg.command=--title="VMware6 mem.consumed.average" \ | |
+--vertical-label="MemCdAvg" \ | |
+DEF:xxx={rrd1}:MemCdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemCdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemCnReAvg.name=vmware6.mem.compressionRate.average | |
+report.vmware6.MemCnReAvg.columns=MemCnReAvg | |
+report.vmware6.MemCnReAvg.type=nodeSnmp | |
+report.vmware6.MemCnReAvg.command=--title="VMware6 mem.compressionRate.average" \ | |
+--vertical-label="MemCnReAvg" \ | |
+DEF:xxx={rrd1}:MemCnReAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemCnReAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemCompressedAvg.name=vmware6.mem.compressed.average | |
+report.vmware6.MemCompressedAvg.columns=MemCompressedAvg | |
+report.vmware6.MemCompressedAvg.type=nodeSnmp | |
+report.vmware6.MemCompressedAvg.command=--title="VMware6 mem.compressed.average" \ | |
+--vertical-label="MemCompressedAvg" \ | |
+DEF:xxx={rrd1}:MemCompressedAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemCompressedAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemDnReAvg.name=vmware6.mem.decompressionRate.average | |
+report.vmware6.MemDnReAvg.columns=MemDnReAvg | |
+report.vmware6.MemDnReAvg.type=nodeSnmp | |
+report.vmware6.MemDnReAvg.command=--title="VMware6 mem.decompressionRate.average" \ | |
+--vertical-label="MemDnReAvg" \ | |
+DEF:xxx={rrd1}:MemDnReAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemDnReAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemEntitlementAvg.name=vmware6.mem.entitlement.average | |
+report.vmware6.MemEntitlementAvg.columns=MemEntitlementAvg | |
+report.vmware6.MemEntitlementAvg.type=nodeSnmp | |
+report.vmware6.MemEntitlementAvg.command=--title="VMware6 mem.entitlement.average" \ | |
+--vertical-label="MemEntitlementAvg" \ | |
+DEF:xxx={rrd1}:MemEntitlementAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemEntitlementAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemGrantedAvg.name=vmware6.mem.granted.average | |
+report.vmware6.MemGrantedAvg.columns=MemGrantedAvg | |
+report.vmware6.MemGrantedAvg.type=nodeSnmp | |
+report.vmware6.MemGrantedAvg.command=--title="VMware6 mem.granted.average" \ | |
+--vertical-label="MemGrantedAvg" \ | |
+DEF:xxx={rrd1}:MemGrantedAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemGrantedAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemLlSpInReAvg.name=vmware6.mem.llSwapInRate.average | |
+report.vmware6.MemLlSpInReAvg.columns=MemLlSpInReAvg | |
+report.vmware6.MemLlSpInReAvg.type=nodeSnmp | |
+report.vmware6.MemLlSpInReAvg.command=--title="VMware6 mem.llSwapInRate.average" \ | |
+--vertical-label="MemLlSpInReAvg" \ | |
+DEF:xxx={rrd1}:MemLlSpInReAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemLlSpInReAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemLlSpOutReAvg.name=vmware6.mem.llSwapOutRate.average | |
+report.vmware6.MemLlSpOutReAvg.columns=MemLlSpOutReAvg | |
+report.vmware6.MemLlSpOutReAvg.type=nodeSnmp | |
+report.vmware6.MemLlSpOutReAvg.command=--title="VMware6 mem.llSwapOutRate.average" \ | |
+--vertical-label="MemLlSpOutReAvg" \ | |
+DEF:xxx={rrd1}:MemLlSpOutReAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemLlSpOutReAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemLlSpUsedAvg.name=vmware6.mem.llSwapUsed.average | |
+report.vmware6.MemLlSpUsedAvg.columns=MemLlSpUsedAvg | |
+report.vmware6.MemLlSpUsedAvg.type=nodeSnmp | |
+report.vmware6.MemLlSpUsedAvg.command=--title="VMware6 mem.llSwapUsed.average" \ | |
+--vertical-label="MemLlSpUsedAvg" \ | |
+DEF:xxx={rrd1}:MemLlSpUsedAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemLlSpUsedAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemLyAvg.name=vmware6.mem.latency.average | |
+report.vmware6.MemLyAvg.columns=MemLyAvg | |
+report.vmware6.MemLyAvg.type=nodeSnmp | |
+report.vmware6.MemLyAvg.command=--title="VMware6 mem.latency.average" \ | |
+--vertical-label="MemLyAvg" \ | |
+DEF:xxx={rrd1}:MemLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemOdAvg.name=vmware6.mem.overhead.average | |
+report.vmware6.MemOdAvg.columns=MemOdAvg | |
+report.vmware6.MemOdAvg.type=nodeSnmp | |
+report.vmware6.MemOdAvg.command=--title="VMware6 mem.overhead.average" \ | |
+--vertical-label="MemOdAvg" \ | |
+DEF:xxx={rrd1}:MemOdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemOdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemOdMaxAvg.name=vmware6.mem.overheadMax.average | |
+report.vmware6.MemOdMaxAvg.columns=MemOdMaxAvg | |
+report.vmware6.MemOdMaxAvg.type=nodeSnmp | |
+report.vmware6.MemOdMaxAvg.command=--title="VMware6 mem.overheadMax.average" \ | |
+--vertical-label="MemOdMaxAvg" \ | |
+DEF:xxx={rrd1}:MemOdMaxAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemOdMaxAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemOdTdAvg.name=vmware6.mem.overheadTouched.average | |
+report.vmware6.MemOdTdAvg.columns=MemOdTdAvg | |
+report.vmware6.MemOdTdAvg.type=nodeSnmp | |
+report.vmware6.MemOdTdAvg.command=--title="VMware6 mem.overheadTouched.average" \ | |
+--vertical-label="MemOdTdAvg" \ | |
+DEF:xxx={rrd1}:MemOdTdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemOdTdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemSharedAvg.name=vmware6.mem.shared.average | |
+report.vmware6.MemSharedAvg.columns=MemSharedAvg | |
+report.vmware6.MemSharedAvg.type=nodeSnmp | |
+report.vmware6.MemSharedAvg.command=--title="VMware6 mem.shared.average" \ | |
+--vertical-label="MemSharedAvg" \ | |
+DEF:xxx={rrd1}:MemSharedAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemSharedAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemSpTtAvg.name=vmware6.mem.swaptarget.average | |
+report.vmware6.MemSpTtAvg.columns=MemSpTtAvg | |
+report.vmware6.MemSpTtAvg.type=nodeSnmp | |
+report.vmware6.MemSpTtAvg.command=--title="VMware6 mem.swaptarget.average" \ | |
+--vertical-label="MemSpTtAvg" \ | |
+DEF:xxx={rrd1}:MemSpTtAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemSpTtAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemSpinAvg.name=vmware6.mem.swapin.average | |
+report.vmware6.MemSpinAvg.columns=MemSpinAvg | |
+report.vmware6.MemSpinAvg.type=nodeSnmp | |
+report.vmware6.MemSpinAvg.command=--title="VMware6 mem.swapin.average" \ | |
+--vertical-label="MemSpinAvg" \ | |
+DEF:xxx={rrd1}:MemSpinAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemSpinAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemSpinReAvg.name=vmware6.mem.swapinRate.average | |
+report.vmware6.MemSpinReAvg.columns=MemSpinReAvg | |
+report.vmware6.MemSpinReAvg.type=nodeSnmp | |
+report.vmware6.MemSpinReAvg.command=--title="VMware6 mem.swapinRate.average" \ | |
+--vertical-label="MemSpinReAvg" \ | |
+DEF:xxx={rrd1}:MemSpinReAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemSpinReAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemSpoutAvg.name=vmware6.mem.swapout.average | |
+report.vmware6.MemSpoutAvg.columns=MemSpoutAvg | |
+report.vmware6.MemSpoutAvg.type=nodeSnmp | |
+report.vmware6.MemSpoutAvg.command=--title="VMware6 mem.swapout.average" \ | |
+--vertical-label="MemSpoutAvg" \ | |
+DEF:xxx={rrd1}:MemSpoutAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemSpoutAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemSpoutReAvg.name=vmware6.mem.swapoutRate.average | |
+report.vmware6.MemSpoutReAvg.columns=MemSpoutReAvg | |
+report.vmware6.MemSpoutReAvg.type=nodeSnmp | |
+report.vmware6.MemSpoutReAvg.command=--title="VMware6 mem.swapoutRate.average" \ | |
+--vertical-label="MemSpoutReAvg" \ | |
+DEF:xxx={rrd1}:MemSpoutReAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemSpoutReAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemSppedAvg.name=vmware6.mem.swapped.average | |
+report.vmware6.MemSppedAvg.columns=MemSppedAvg | |
+report.vmware6.MemSppedAvg.type=nodeSnmp | |
+report.vmware6.MemSppedAvg.command=--title="VMware6 mem.swapped.average" \ | |
+--vertical-label="MemSppedAvg" \ | |
+DEF:xxx={rrd1}:MemSppedAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemSppedAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemUsageAvg.name=vmware6.mem.usage.average | |
+report.vmware6.MemUsageAvg.columns=MemUsageAvg | |
+report.vmware6.MemUsageAvg.type=nodeSnmp | |
+report.vmware6.MemUsageAvg.command=--title="VMware6 mem.usage.average" \ | |
+--vertical-label="MemUsageAvg" \ | |
+DEF:xxx={rrd1}:MemUsageAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemUsageAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemVmmemctlAvg.name=vmware6.mem.vmmemctl.average | |
+report.vmware6.MemVmmemctlAvg.columns=MemVmmemctlAvg | |
+report.vmware6.MemVmmemctlAvg.type=nodeSnmp | |
+report.vmware6.MemVmmemctlAvg.command=--title="VMware6 mem.vmmemctl.average" \ | |
+--vertical-label="MemVmmemctlAvg" \ | |
+DEF:xxx={rrd1}:MemVmmemctlAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemVmmemctlAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemVmmemctlTtAvg.name=vmware6.mem.vmmemctltarget.average | |
+report.vmware6.MemVmmemctlTtAvg.columns=MemVmmemctlTtAvg | |
+report.vmware6.MemVmmemctlTtAvg.type=nodeSnmp | |
+report.vmware6.MemVmmemctlTtAvg.command=--title="VMware6 mem.vmmemctltarget.average" \ | |
+--vertical-label="MemVmmemctlTtAvg" \ | |
+DEF:xxx={rrd1}:MemVmmemctlTtAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemVmmemctlTtAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemZeroAvg.name=vmware6.mem.zero.average | |
+report.vmware6.MemZeroAvg.columns=MemZeroAvg | |
+report.vmware6.MemZeroAvg.type=nodeSnmp | |
+report.vmware6.MemZeroAvg.command=--title="VMware6 mem.zero.average" \ | |
+--vertical-label="MemZeroAvg" \ | |
+DEF:xxx={rrd1}:MemZeroAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemZeroAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemZipSavedLat.name=vmware6.mem.zipSaved.latest | |
+report.vmware6.MemZipSavedLat.columns=MemZipSavedLat | |
+report.vmware6.MemZipSavedLat.type=nodeSnmp | |
+report.vmware6.MemZipSavedLat.command=--title="VMware6 mem.zipSaved.latest" \ | |
+--vertical-label="MemZipSavedLat" \ | |
+DEF:xxx={rrd1}:MemZipSavedLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemZipSavedLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemZippedLat.name=vmware6.mem.zipped.latest | |
+report.vmware6.MemZippedLat.columns=MemZippedLat | |
+report.vmware6.MemZippedLat.type=nodeSnmp | |
+report.vmware6.MemZippedLat.command=--title="VMware6 mem.zipped.latest" \ | |
+--vertical-label="MemZippedLat" \ | |
+DEF:xxx={rrd1}:MemZippedLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemZippedLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetBroadcastRxSum.name=vmware6.net.broadcastRx.summation | |
+report.vmware6.NetBroadcastRxSum.columns=NetBroadcastRxSum | |
+report.vmware6.NetBroadcastRxSum.type=nodeSnmp | |
+report.vmware6.NetBroadcastRxSum.command=--title="VMware6 net.broadcastRx.summation" \ | |
+--vertical-label="NetBroadcastRxSum" \ | |
+DEF:xxx={rrd1}:NetBroadcastRxSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetBroadcastRxSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetBroadcastTxSum.name=vmware6.net.broadcastTx.summation | |
+report.vmware6.NetBroadcastTxSum.columns=NetBroadcastTxSum | |
+report.vmware6.NetBroadcastTxSum.type=nodeSnmp | |
+report.vmware6.NetBroadcastTxSum.command=--title="VMware6 net.broadcastTx.summation" \ | |
+--vertical-label="NetBroadcastTxSum" \ | |
+DEF:xxx={rrd1}:NetBroadcastTxSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetBroadcastTxSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetDroppedRxSum.name=vmware6.net.droppedRx.summation | |
+report.vmware6.NetDroppedRxSum.columns=NetDroppedRxSum | |
+report.vmware6.NetDroppedRxSum.type=nodeSnmp | |
+report.vmware6.NetDroppedRxSum.command=--title="VMware6 net.droppedRx.summation" \ | |
+--vertical-label="NetDroppedRxSum" \ | |
+DEF:xxx={rrd1}:NetDroppedRxSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetDroppedRxSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetDroppedTxSum.name=vmware6.net.droppedTx.summation | |
+report.vmware6.NetDroppedTxSum.columns=NetDroppedTxSum | |
+report.vmware6.NetDroppedTxSum.type=nodeSnmp | |
+report.vmware6.NetDroppedTxSum.command=--title="VMware6 net.droppedTx.summation" \ | |
+--vertical-label="NetDroppedTxSum" \ | |
+DEF:xxx={rrd1}:NetDroppedTxSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetDroppedTxSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetPacketsTxSum.name=vmware6.net.packetsTx.summation | |
+report.vmware6.NetPacketsTxSum.columns=NetPacketsTxSum | |
+report.vmware6.NetPacketsTxSum.type=nodeSnmp | |
+report.vmware6.NetPacketsTxSum.command=--title="VMware6 net.packetsTx.summation" \ | |
+--vertical-label="NetPacketsTxSum" \ | |
+DEF:xxx={rrd1}:NetPacketsTxSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetPacketsTxSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetReceivedAvg.name=vmware6.net.received.average | |
+report.vmware6.NetReceivedAvg.columns=NetReceivedAvg | |
+report.vmware6.NetReceivedAvg.type=nodeSnmp | |
+report.vmware6.NetReceivedAvg.command=--title="VMware6 net.received.average" \ | |
+--vertical-label="NetReceivedAvg" \ | |
+DEF:xxx={rrd1}:NetReceivedAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetReceivedAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetTransmittedAvg.name=vmware6.net.transmitted.average | |
+report.vmware6.NetTransmittedAvg.columns=NetTransmittedAvg | |
+report.vmware6.NetTransmittedAvg.type=nodeSnmp | |
+report.vmware6.NetTransmittedAvg.command=--title="VMware6 net.transmitted.average" \ | |
+--vertical-label="NetTransmittedAvg" \ | |
+DEF:xxx={rrd1}:NetTransmittedAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetTransmittedAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.PowerEnergySum.name=vmware6.power.energy.summation | |
+report.vmware6.PowerEnergySum.columns=PowerEnergySum | |
+report.vmware6.PowerEnergySum.type=nodeSnmp | |
+report.vmware6.PowerEnergySum.command=--title="VMware6 power.energy.summation" \ | |
+--vertical-label="PowerEnergySum" \ | |
+DEF:xxx={rrd1}:PowerEnergySum:AVERAGE \ | |
+LINE2:xxx#0000ff:"PowerEnergySum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.PowerPowerAvg.name=vmware6.power.power.average | |
+report.vmware6.PowerPowerAvg.columns=PowerPowerAvg | |
+report.vmware6.PowerPowerAvg.type=nodeSnmp | |
+report.vmware6.PowerPowerAvg.command=--title="VMware6 power.power.average" \ | |
+--vertical-label="PowerPowerAvg" \ | |
+DEF:xxx={rrd1}:PowerPowerAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"PowerPowerAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuActav15Lat.name=vmware6.rescpu.actav15.latest | |
+report.vmware6.ResCpuActav15Lat.columns=ResCpuActav15Lat | |
+report.vmware6.ResCpuActav15Lat.type=nodeSnmp | |
+report.vmware6.ResCpuActav15Lat.command=--title="VMware6 rescpu.actav15.latest" \ | |
+--vertical-label="ResCpuActav15Lat" \ | |
+DEF:xxx={rrd1}:ResCpuActav15Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuActav15Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuActav1Lat.name=vmware6.rescpu.actav1.latest | |
+report.vmware6.ResCpuActav1Lat.columns=ResCpuActav1Lat | |
+report.vmware6.ResCpuActav1Lat.type=nodeSnmp | |
+report.vmware6.ResCpuActav1Lat.command=--title="VMware6 rescpu.actav1.latest" \ | |
+--vertical-label="ResCpuActav1Lat" \ | |
+DEF:xxx={rrd1}:ResCpuActav1Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuActav1Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuActav5Lat.name=vmware6.rescpu.actav5.latest | |
+report.vmware6.ResCpuActav5Lat.columns=ResCpuActav5Lat | |
+report.vmware6.ResCpuActav5Lat.type=nodeSnmp | |
+report.vmware6.ResCpuActav5Lat.command=--title="VMware6 rescpu.actav5.latest" \ | |
+--vertical-label="ResCpuActav5Lat" \ | |
+DEF:xxx={rrd1}:ResCpuActav5Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuActav5Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuActpk15Lat.name=vmware6.rescpu.actpk15.latest | |
+report.vmware6.ResCpuActpk15Lat.columns=ResCpuActpk15Lat | |
+report.vmware6.ResCpuActpk15Lat.type=nodeSnmp | |
+report.vmware6.ResCpuActpk15Lat.command=--title="VMware6 rescpu.actpk15.latest" \ | |
+--vertical-label="ResCpuActpk15Lat" \ | |
+DEF:xxx={rrd1}:ResCpuActpk15Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuActpk15Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuActpk1Lat.name=vmware6.rescpu.actpk1.latest | |
+report.vmware6.ResCpuActpk1Lat.columns=ResCpuActpk1Lat | |
+report.vmware6.ResCpuActpk1Lat.type=nodeSnmp | |
+report.vmware6.ResCpuActpk1Lat.command=--title="VMware6 rescpu.actpk1.latest" \ | |
+--vertical-label="ResCpuActpk1Lat" \ | |
+DEF:xxx={rrd1}:ResCpuActpk1Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuActpk1Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuActpk5Lat.name=vmware6.rescpu.actpk5.latest | |
+report.vmware6.ResCpuActpk5Lat.columns=ResCpuActpk5Lat | |
+report.vmware6.ResCpuActpk5Lat.type=nodeSnmp | |
+report.vmware6.ResCpuActpk5Lat.command=--title="VMware6 rescpu.actpk5.latest" \ | |
+--vertical-label="ResCpuActpk5Lat" \ | |
+DEF:xxx={rrd1}:ResCpuActpk5Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuActpk5Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuMaxLd15Lat.name=vmware6.rescpu.maxLimited15.latest | |
+report.vmware6.ResCpuMaxLd15Lat.columns=ResCpuMaxLd15Lat | |
+report.vmware6.ResCpuMaxLd15Lat.type=nodeSnmp | |
+report.vmware6.ResCpuMaxLd15Lat.command=--title="VMware6 rescpu.maxLimited15.latest" \ | |
+--vertical-label="ResCpuMaxLd15Lat" \ | |
+DEF:xxx={rrd1}:ResCpuMaxLd15Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuMaxLd15Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuMaxLd1Lat.name=vmware6.rescpu.maxLimited1.latest | |
+report.vmware6.ResCpuMaxLd1Lat.columns=ResCpuMaxLd1Lat | |
+report.vmware6.ResCpuMaxLd1Lat.type=nodeSnmp | |
+report.vmware6.ResCpuMaxLd1Lat.command=--title="VMware6 rescpu.maxLimited1.latest" \ | |
+--vertical-label="ResCpuMaxLd1Lat" \ | |
+DEF:xxx={rrd1}:ResCpuMaxLd1Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuMaxLd1Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuMaxLd5Lat.name=vmware6.rescpu.maxLimited5.latest | |
+report.vmware6.ResCpuMaxLd5Lat.columns=ResCpuMaxLd5Lat | |
+report.vmware6.ResCpuMaxLd5Lat.type=nodeSnmp | |
+report.vmware6.ResCpuMaxLd5Lat.command=--title="VMware6 rescpu.maxLimited5.latest" \ | |
+--vertical-label="ResCpuMaxLd5Lat" \ | |
+DEF:xxx={rrd1}:ResCpuMaxLd5Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuMaxLd5Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuRunav15Lat.name=vmware6.rescpu.runav15.latest | |
+report.vmware6.ResCpuRunav15Lat.columns=ResCpuRunav15Lat | |
+report.vmware6.ResCpuRunav15Lat.type=nodeSnmp | |
+report.vmware6.ResCpuRunav15Lat.command=--title="VMware6 rescpu.runav15.latest" \ | |
+--vertical-label="ResCpuRunav15Lat" \ | |
+DEF:xxx={rrd1}:ResCpuRunav15Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuRunav15Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuRunav1Lat.name=vmware6.rescpu.runav1.latest | |
+report.vmware6.ResCpuRunav1Lat.columns=ResCpuRunav1Lat | |
+report.vmware6.ResCpuRunav1Lat.type=nodeSnmp | |
+report.vmware6.ResCpuRunav1Lat.command=--title="VMware6 rescpu.runav1.latest" \ | |
+--vertical-label="ResCpuRunav1Lat" \ | |
+DEF:xxx={rrd1}:ResCpuRunav1Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuRunav1Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuRunav5Lat.name=vmware6.rescpu.runav5.latest | |
+report.vmware6.ResCpuRunav5Lat.columns=ResCpuRunav5Lat | |
+report.vmware6.ResCpuRunav5Lat.type=nodeSnmp | |
+report.vmware6.ResCpuRunav5Lat.command=--title="VMware6 rescpu.runav5.latest" \ | |
+--vertical-label="ResCpuRunav5Lat" \ | |
+DEF:xxx={rrd1}:ResCpuRunav5Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuRunav5Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuRunpk15Lat.name=vmware6.rescpu.runpk15.latest | |
+report.vmware6.ResCpuRunpk15Lat.columns=ResCpuRunpk15Lat | |
+report.vmware6.ResCpuRunpk15Lat.type=nodeSnmp | |
+report.vmware6.ResCpuRunpk15Lat.command=--title="VMware6 rescpu.runpk15.latest" \ | |
+--vertical-label="ResCpuRunpk15Lat" \ | |
+DEF:xxx={rrd1}:ResCpuRunpk15Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuRunpk15Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuRunpk1Lat.name=vmware6.rescpu.runpk1.latest | |
+report.vmware6.ResCpuRunpk1Lat.columns=ResCpuRunpk1Lat | |
+report.vmware6.ResCpuRunpk1Lat.type=nodeSnmp | |
+report.vmware6.ResCpuRunpk1Lat.command=--title="VMware6 rescpu.runpk1.latest" \ | |
+--vertical-label="ResCpuRunpk1Lat" \ | |
+DEF:xxx={rrd1}:ResCpuRunpk1Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuRunpk1Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuRunpk5Lat.name=vmware6.rescpu.runpk5.latest | |
+report.vmware6.ResCpuRunpk5Lat.columns=ResCpuRunpk5Lat | |
+report.vmware6.ResCpuRunpk5Lat.type=nodeSnmp | |
+report.vmware6.ResCpuRunpk5Lat.command=--title="VMware6 rescpu.runpk5.latest" \ | |
+--vertical-label="ResCpuRunpk5Lat" \ | |
+DEF:xxx={rrd1}:ResCpuRunpk5Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuRunpk5Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuSeCtLat.name=vmware6.rescpu.sampleCount.latest | |
+report.vmware6.ResCpuSeCtLat.columns=ResCpuSeCtLat | |
+report.vmware6.ResCpuSeCtLat.type=nodeSnmp | |
+report.vmware6.ResCpuSeCtLat.command=--title="VMware6 rescpu.sampleCount.latest" \ | |
+--vertical-label="ResCpuSeCtLat" \ | |
+DEF:xxx={rrd1}:ResCpuSeCtLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuSeCtLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.ResCpuSePeriodLat.name=vmware6.rescpu.samplePeriod.latest | |
+report.vmware6.ResCpuSePeriodLat.columns=ResCpuSePeriodLat | |
+report.vmware6.ResCpuSePeriodLat.type=nodeSnmp | |
+report.vmware6.ResCpuSePeriodLat.command=--title="VMware6 rescpu.samplePeriod.latest" \ | |
+--vertical-label="ResCpuSePeriodLat" \ | |
+DEF:xxx={rrd1}:ResCpuSePeriodLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"ResCpuSePeriodLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysHeartbeatLat.name=vmware6.sys.heartbeat.latest | |
+report.vmware6.SysHeartbeatLat.columns=SysHeartbeatLat | |
+report.vmware6.SysHeartbeatLat.type=nodeSnmp | |
+report.vmware6.SysHeartbeatLat.command=--title="VMware6 sys.heartbeat.latest" \ | |
+--vertical-label="SysHeartbeatLat" \ | |
+DEF:xxx={rrd1}:SysHeartbeatLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysHeartbeatLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysOsUpTeLat.name=vmware6.sys.osUptime.latest | |
+report.vmware6.SysOsUpTeLat.columns=SysOsUpTeLat | |
+report.vmware6.SysOsUpTeLat.type=nodeSnmp | |
+report.vmware6.SysOsUpTeLat.command=--title="VMware6 sys.osUptime.latest" \ | |
+--vertical-label="SysOsUpTeLat" \ | |
+DEF:xxx={rrd1}:SysOsUpTeLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysOsUpTeLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysUpTeLat.name=vmware6.sys.uptime.latest | |
+report.vmware6.SysUpTeLat.columns=SysUpTeLat | |
+report.vmware6.SysUpTeLat.type=nodeSnmp | |
+report.vmware6.SysUpTeLat.command=--title="VMware6 sys.uptime.latest" \ | |
+--vertical-label="SysUpTeLat" \ | |
+DEF:xxx={rrd1}:SysUpTeLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysUpTeLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuCostopSum.name=vmware6.cpu.costop.summation | |
+report.vmware6.CpuCostopSum.columns=CpuCostopSum | |
+report.vmware6.CpuCostopSum.propertiesValues=vmware6CpuName | |
+report.vmware6.CpuCostopSum.type=vmware6Cpu | |
+report.vmware6.CpuCostopSum.command=--title="VMware6 cpu.costop.summation {vmware6CpuName}" \ | |
+--vertical-label="CpuCostopSum" \ | |
+DEF:xxx={rrd1}:CpuCostopSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuCostopSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuIdleSum.name=vmware6.cpu.idle.summation | |
+report.vmware6.CpuIdleSum.columns=CpuIdleSum | |
+report.vmware6.CpuIdleSum.propertiesValues=vmware6CpuName | |
+report.vmware6.CpuIdleSum.type=vmware6Cpu | |
+report.vmware6.CpuIdleSum.command=--title="VMware6 cpu.idle.summation {vmware6CpuName}" \ | |
+--vertical-label="CpuIdleSum" \ | |
+DEF:xxx={rrd1}:CpuIdleSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuIdleSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuMaxLdSum.name=vmware6.cpu.maxlimited.summation | |
+report.vmware6.CpuMaxLdSum.columns=CpuMaxLdSum | |
+report.vmware6.CpuMaxLdSum.propertiesValues=vmware6CpuName | |
+report.vmware6.CpuMaxLdSum.type=vmware6Cpu | |
+report.vmware6.CpuMaxLdSum.command=--title="VMware6 cpu.maxlimited.summation {vmware6CpuName}" \ | |
+--vertical-label="CpuMaxLdSum" \ | |
+DEF:xxx={rrd1}:CpuMaxLdSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuMaxLdSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuRdySum.name=vmware6.cpu.ready.summation | |
+report.vmware6.CpuRdySum.columns=CpuRdySum | |
+report.vmware6.CpuRdySum.propertiesValues=vmware6CpuName | |
+report.vmware6.CpuRdySum.type=vmware6Cpu | |
+report.vmware6.CpuRdySum.command=--title="VMware6 cpu.ready.summation {vmware6CpuName}" \ | |
+--vertical-label="CpuRdySum" \ | |
+DEF:xxx={rrd1}:CpuRdySum:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuRdySum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuRunSum.name=vmware6.cpu.run.summation | |
+report.vmware6.CpuRunSum.columns=CpuRunSum | |
+report.vmware6.CpuRunSum.propertiesValues=vmware6CpuName | |
+report.vmware6.CpuRunSum.type=vmware6Cpu | |
+report.vmware6.CpuRunSum.command=--title="VMware6 cpu.run.summation {vmware6CpuName}" \ | |
+--vertical-label="CpuRunSum" \ | |
+DEF:xxx={rrd1}:CpuRunSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuRunSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuSystemSum.name=vmware6.cpu.system.summation | |
+report.vmware6.CpuSystemSum.columns=CpuSystemSum | |
+report.vmware6.CpuSystemSum.propertiesValues=vmware6CpuName | |
+report.vmware6.CpuSystemSum.type=vmware6Cpu | |
+report.vmware6.CpuSystemSum.command=--title="VMware6 cpu.system.summation {vmware6CpuName}" \ | |
+--vertical-label="CpuSystemSum" \ | |
+DEF:xxx={rrd1}:CpuSystemSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuSystemSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetBytesRxAvg.name=vmware6.net.bytesRx.average | |
+report.vmware6.NetBytesRxAvg.columns=NetBytesRxAvg | |
+report.vmware6.NetBytesRxAvg.propertiesValues=vmware6NetName | |
+report.vmware6.NetBytesRxAvg.type=vmware6Net | |
+report.vmware6.NetBytesRxAvg.command=--title="VMware6 net.bytesRx.average {vmware6NetName}" \ | |
+--vertical-label="NetBytesRxAvg" \ | |
+DEF:xxx={rrd1}:NetBytesRxAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetBytesRxAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetBytesTxAvg.name=vmware6.net.bytesTx.average | |
+report.vmware6.NetBytesTxAvg.columns=NetBytesTxAvg | |
+report.vmware6.NetBytesTxAvg.propertiesValues=vmware6NetName | |
+report.vmware6.NetBytesTxAvg.type=vmware6Net | |
+report.vmware6.NetBytesTxAvg.command=--title="VMware6 net.bytesTx.average {vmware6NetName}" \ | |
+--vertical-label="NetBytesTxAvg" \ | |
+DEF:xxx={rrd1}:NetBytesTxAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetBytesTxAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetMulticastRxSum.name=vmware6.net.multicastRx.summation | |
+report.vmware6.NetMulticastRxSum.columns=NetMulticastRxSum | |
+report.vmware6.NetMulticastRxSum.propertiesValues=vmware6NetName | |
+report.vmware6.NetMulticastRxSum.type=vmware6Net | |
+report.vmware6.NetMulticastRxSum.command=--title="VMware6 net.multicastRx.summation {vmware6NetName}" \ | |
+--vertical-label="NetMulticastRxSum" \ | |
+DEF:xxx={rrd1}:NetMulticastRxSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetMulticastRxSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetMulticastTxSum.name=vmware6.net.multicastTx.summation | |
+report.vmware6.NetMulticastTxSum.columns=NetMulticastTxSum | |
+report.vmware6.NetMulticastTxSum.propertiesValues=vmware6NetName | |
+report.vmware6.NetMulticastTxSum.type=vmware6Net | |
+report.vmware6.NetMulticastTxSum.command=--title="VMware6 net.multicastTx.summation {vmware6NetName}" \ | |
+--vertical-label="NetMulticastTxSum" \ | |
+DEF:xxx={rrd1}:NetMulticastTxSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetMulticastTxSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetPacketsRxSum.name=vmware6.net.packetsRx.summation | |
+report.vmware6.NetPacketsRxSum.columns=NetPacketsRxSum | |
+report.vmware6.NetPacketsRxSum.propertiesValues=vmware6NetName | |
+report.vmware6.NetPacketsRxSum.type=vmware6Net | |
+report.vmware6.NetPacketsRxSum.command=--title="VMware6 net.packetsRx.summation {vmware6NetName}" \ | |
+--vertical-label="NetPacketsRxSum" \ | |
+DEF:xxx={rrd1}:NetPacketsRxSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetPacketsRxSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetUsageAvg.name=vmware6.net.usage.average | |
+report.vmware6.NetUsageAvg.columns=NetUsageAvg | |
+report.vmware6.NetUsageAvg.propertiesValues=vmware6NetName | |
+report.vmware6.NetUsageAvg.type=vmware6Net | |
+report.vmware6.NetUsageAvg.command=--title="VMware6 net.usage.average {vmware6NetName}" \ | |
+--vertical-label="NetUsageAvg" \ | |
+DEF:xxx={rrd1}:NetUsageAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetUsageAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskBusResetsSum.name=vmware6.disk.busResets.summation | |
+report.vmware6.DiskBusResetsSum.columns=DiskBusResetsSum | |
+report.vmware6.DiskBusResetsSum.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskBusResetsSum.type=vmware6Disk | |
+report.vmware6.DiskBusResetsSum.command=--title="VMware6 disk.busResets.summation {vmware6DiskName}" \ | |
+--vertical-label="DiskBusResetsSum" \ | |
+DEF:xxx={rrd1}:DiskBusResetsSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskBusResetsSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskCsAdAvg.name=vmware6.disk.commandsAveraged.average | |
+report.vmware6.DiskCsAdAvg.columns=DiskCsAdAvg | |
+report.vmware6.DiskCsAdAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskCsAdAvg.type=vmware6Disk | |
+report.vmware6.DiskCsAdAvg.command=--title="VMware6 disk.commandsAveraged.average {vmware6DiskName}" \ | |
+--vertical-label="DiskCsAdAvg" \ | |
+DEF:xxx={rrd1}:DiskCsAdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskCsAdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskCsAdSum.name=vmware6.disk.commandsAborted.summation | |
+report.vmware6.DiskCsAdSum.columns=DiskCsAdSum | |
+report.vmware6.DiskCsAdSum.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskCsAdSum.type=vmware6Disk | |
+report.vmware6.DiskCsAdSum.command=--title="VMware6 disk.commandsAborted.summation {vmware6DiskName}" \ | |
+--vertical-label="DiskCsAdSum" \ | |
+DEF:xxx={rrd1}:DiskCsAdSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskCsAdSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskCsSum.name=vmware6.disk.commands.summation | |
+report.vmware6.DiskCsSum.columns=DiskCsSum | |
+report.vmware6.DiskCsSum.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskCsSum.type=vmware6Disk | |
+report.vmware6.DiskCsSum.command=--title="VMware6 disk.commands.summation {vmware6DiskName}" \ | |
+--vertical-label="DiskCsSum" \ | |
+DEF:xxx={rrd1}:DiskCsSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskCsSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskNrRdAdAvg.name=vmware6.disk.numberReadAveraged.average | |
+report.vmware6.DiskNrRdAdAvg.columns=DiskNrRdAdAvg | |
+report.vmware6.DiskNrRdAdAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskNrRdAdAvg.type=vmware6Disk | |
+report.vmware6.DiskNrRdAdAvg.command=--title="VMware6 disk.numberReadAveraged.average {vmware6DiskName}" \ | |
+--vertical-label="DiskNrRdAdAvg" \ | |
+DEF:xxx={rrd1}:DiskNrRdAdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskNrRdAdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskNrRdSum.name=vmware6.disk.numberRead.summation | |
+report.vmware6.DiskNrRdSum.columns=DiskNrRdSum | |
+report.vmware6.DiskNrRdSum.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskNrRdSum.type=vmware6Disk | |
+report.vmware6.DiskNrRdSum.command=--title="VMware6 disk.numberRead.summation {vmware6DiskName}" \ | |
+--vertical-label="DiskNrRdSum" \ | |
+DEF:xxx={rrd1}:DiskNrRdSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskNrRdSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskNrWeAdAvg.name=vmware6.disk.numberWriteAveraged.average | |
+report.vmware6.DiskNrWeAdAvg.columns=DiskNrWeAdAvg | |
+report.vmware6.DiskNrWeAdAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskNrWeAdAvg.type=vmware6Disk | |
+report.vmware6.DiskNrWeAdAvg.command=--title="VMware6 disk.numberWriteAveraged.average {vmware6DiskName}" \ | |
+--vertical-label="DiskNrWeAdAvg" \ | |
+DEF:xxx={rrd1}:DiskNrWeAdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskNrWeAdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskNrWeSum.name=vmware6.disk.numberWrite.summation | |
+report.vmware6.DiskNrWeSum.columns=DiskNrWeSum | |
+report.vmware6.DiskNrWeSum.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskNrWeSum.type=vmware6Disk | |
+report.vmware6.DiskNrWeSum.command=--title="VMware6 disk.numberWrite.summation {vmware6DiskName}" \ | |
+--vertical-label="DiskNrWeSum" \ | |
+DEF:xxx={rrd1}:DiskNrWeSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskNrWeSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStNrRdAdAvg.name=vmware6.datastore.numberReadAveraged.average | |
+report.vmware6.DaStNrRdAdAvg.columns=DaStNrRdAdAvg | |
+report.vmware6.DaStNrRdAdAvg.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStNrRdAdAvg.type=vmware6DaSt | |
+report.vmware6.DaStNrRdAdAvg.command=--title="VMware6 datastore.numberReadAveraged.average {vmware6DaStName}" \ | |
+--vertical-label="DaStNrRdAdAvg" \ | |
+DEF:xxx={rrd1}:DaStNrRdAdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStNrRdAdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStNrWeAdAvg.name=vmware6.datastore.numberWriteAveraged.average | |
+report.vmware6.DaStNrWeAdAvg.columns=DaStNrWeAdAvg | |
+report.vmware6.DaStNrWeAdAvg.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStNrWeAdAvg.type=vmware6DaSt | |
+report.vmware6.DaStNrWeAdAvg.command=--title="VMware6 datastore.numberWriteAveraged.average {vmware6DaStName}" \ | |
+--vertical-label="DaStNrWeAdAvg" \ | |
+DEF:xxx={rrd1}:DaStNrWeAdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStNrWeAdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStRdAvg.name=vmware6.datastore.read.average | |
+report.vmware6.DaStRdAvg.columns=DaStRdAvg | |
+report.vmware6.DaStRdAvg.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStRdAvg.type=vmware6DaSt | |
+report.vmware6.DaStRdAvg.command=--title="VMware6 datastore.read.average {vmware6DaStName}" \ | |
+--vertical-label="DaStRdAvg" \ | |
+DEF:xxx={rrd1}:DaStRdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStRdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStTlRdLyAvg.name=vmware6.datastore.totalReadLatency.average | |
+report.vmware6.DaStTlRdLyAvg.columns=DaStTlRdLyAvg | |
+report.vmware6.DaStTlRdLyAvg.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStTlRdLyAvg.type=vmware6DaSt | |
+report.vmware6.DaStTlRdLyAvg.command=--title="VMware6 datastore.totalReadLatency.average {vmware6DaStName}" \ | |
+--vertical-label="DaStTlRdLyAvg" \ | |
+DEF:xxx={rrd1}:DaStTlRdLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStTlRdLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStTlWeLyAvg.name=vmware6.datastore.totalWriteLatency.average | |
+report.vmware6.DaStTlWeLyAvg.columns=DaStTlWeLyAvg | |
+report.vmware6.DaStTlWeLyAvg.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStTlWeLyAvg.type=vmware6DaSt | |
+report.vmware6.DaStTlWeLyAvg.command=--title="VMware6 datastore.totalWriteLatency.average {vmware6DaStName}" \ | |
+--vertical-label="DaStTlWeLyAvg" \ | |
+DEF:xxx={rrd1}:DaStTlWeLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStTlWeLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStWeAvg.name=vmware6.datastore.write.average | |
+report.vmware6.DaStWeAvg.columns=DaStWeAvg | |
+report.vmware6.DaStWeAvg.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStWeAvg.type=vmware6DaSt | |
+report.vmware6.DaStWeAvg.command=--title="VMware6 datastore.write.average {vmware6DaStName}" \ | |
+--vertical-label="DaStWeAvg" \ | |
+DEF:xxx={rrd1}:DaStWeAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStWeAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StAdptrCsAdAvg.name=vmware6.storageAdapter.commandsAveraged.average | |
+report.vmware6.StAdptrCsAdAvg.columns=StAdptrCsAdAvg | |
+report.vmware6.StAdptrCsAdAvg.propertiesValues=vmware6StAdptrName | |
+report.vmware6.StAdptrCsAdAvg.type=vmware6StAdptr | |
+report.vmware6.StAdptrCsAdAvg.command=--title="VMware6 storageAdapter.commandsAveraged.average {vmware6StAdptrName}" \ | |
+--vertical-label="StAdptrCsAdAvg" \ | |
+DEF:xxx={rrd1}:StAdptrCsAdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"StAdptrCsAdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StAdptrNrRdAdAvg.name=vmware6.storageAdapter.numberReadAveraged.average | |
+report.vmware6.StAdptrNrRdAdAvg.columns=StAdptrNrRdAdAvg | |
+report.vmware6.StAdptrNrRdAdAvg.propertiesValues=vmware6StAdptrName | |
+report.vmware6.StAdptrNrRdAdAvg.type=vmware6StAdptr | |
+report.vmware6.StAdptrNrRdAdAvg.command=--title="VMware6 storageAdapter.numberReadAveraged.average {vmware6StAdptrName}" \ | |
+--vertical-label="StAdptrNrRdAdAvg" \ | |
+DEF:xxx={rrd1}:StAdptrNrRdAdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"StAdptrNrRdAdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StAdptrNrWeAdAvg.name=vmware6.storageAdapter.numberWriteAveraged.average | |
+report.vmware6.StAdptrNrWeAdAvg.columns=StAdptrNrWeAdAvg | |
+report.vmware6.StAdptrNrWeAdAvg.propertiesValues=vmware6StAdptrName | |
+report.vmware6.StAdptrNrWeAdAvg.type=vmware6StAdptr | |
+report.vmware6.StAdptrNrWeAdAvg.command=--title="VMware6 storageAdapter.numberWriteAveraged.average {vmware6StAdptrName}" \ | |
+--vertical-label="StAdptrNrWeAdAvg" \ | |
+DEF:xxx={rrd1}:StAdptrNrWeAdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"StAdptrNrWeAdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StAdptrRdAvg.name=vmware6.storageAdapter.read.average | |
+report.vmware6.StAdptrRdAvg.columns=StAdptrRdAvg | |
+report.vmware6.StAdptrRdAvg.propertiesValues=vmware6StAdptrName | |
+report.vmware6.StAdptrRdAvg.type=vmware6StAdptr | |
+report.vmware6.StAdptrRdAvg.command=--title="VMware6 storageAdapter.read.average {vmware6StAdptrName}" \ | |
+--vertical-label="StAdptrRdAvg" \ | |
+DEF:xxx={rrd1}:StAdptrRdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"StAdptrRdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StAdptrTlRdLyAvg.name=vmware6.storageAdapter.totalReadLatency.average | |
+report.vmware6.StAdptrTlRdLyAvg.columns=StAdptrTlRdLyAvg | |
+report.vmware6.StAdptrTlRdLyAvg.propertiesValues=vmware6StAdptrName | |
+report.vmware6.StAdptrTlRdLyAvg.type=vmware6StAdptr | |
+report.vmware6.StAdptrTlRdLyAvg.command=--title="VMware6 storageAdapter.totalReadLatency.average {vmware6StAdptrName}" \ | |
+--vertical-label="StAdptrTlRdLyAvg" \ | |
+DEF:xxx={rrd1}:StAdptrTlRdLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"StAdptrTlRdLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StAdptrTlWeLyAvg.name=vmware6.storageAdapter.totalWriteLatency.average | |
+report.vmware6.StAdptrTlWeLyAvg.columns=StAdptrTlWeLyAvg | |
+report.vmware6.StAdptrTlWeLyAvg.propertiesValues=vmware6StAdptrName | |
+report.vmware6.StAdptrTlWeLyAvg.type=vmware6StAdptr | |
+report.vmware6.StAdptrTlWeLyAvg.command=--title="VMware6 storageAdapter.totalWriteLatency.average {vmware6StAdptrName}" \ | |
+--vertical-label="StAdptrTlWeLyAvg" \ | |
+DEF:xxx={rrd1}:StAdptrTlWeLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"StAdptrTlWeLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StAdptrWeAvg.name=vmware6.storageAdapter.write.average | |
+report.vmware6.StAdptrWeAvg.columns=StAdptrWeAvg | |
+report.vmware6.StAdptrWeAvg.propertiesValues=vmware6StAdptrName | |
+report.vmware6.StAdptrWeAvg.type=vmware6StAdptr | |
+report.vmware6.StAdptrWeAvg.command=--title="VMware6 storageAdapter.write.average {vmware6StAdptrName}" \ | |
+--vertical-label="StAdptrWeAvg" \ | |
+DEF:xxx={rrd1}:StAdptrWeAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"StAdptrWeAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StPthCsAdAvg.name=vmware6.storagePath.commandsAveraged.average | |
+report.vmware6.StPthCsAdAvg.columns=StPthCsAdAvg | |
+report.vmware6.StPthCsAdAvg.propertiesValues=vmware6StPthName | |
+report.vmware6.StPthCsAdAvg.type=vmware6StPth | |
+report.vmware6.StPthCsAdAvg.command=--title="VMware6 storagePath.commandsAveraged.average {vmware6StPthName}" \ | |
+--vertical-label="StPthCsAdAvg" \ | |
+DEF:xxx={rrd1}:StPthCsAdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"StPthCsAdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StPthNrRdAdAvg.name=vmware6.storagePath.numberReadAveraged.average | |
+report.vmware6.StPthNrRdAdAvg.columns=StPthNrRdAdAvg | |
+report.vmware6.StPthNrRdAdAvg.propertiesValues=vmware6StPthName | |
+report.vmware6.StPthNrRdAdAvg.type=vmware6StPth | |
+report.vmware6.StPthNrRdAdAvg.command=--title="VMware6 storagePath.numberReadAveraged.average {vmware6StPthName}" \ | |
+--vertical-label="StPthNrRdAdAvg" \ | |
+DEF:xxx={rrd1}:StPthNrRdAdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"StPthNrRdAdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StPthNrWeAdAvg.name=vmware6.storagePath.numberWriteAveraged.average | |
+report.vmware6.StPthNrWeAdAvg.columns=StPthNrWeAdAvg | |
+report.vmware6.StPthNrWeAdAvg.propertiesValues=vmware6StPthName | |
+report.vmware6.StPthNrWeAdAvg.type=vmware6StPth | |
+report.vmware6.StPthNrWeAdAvg.command=--title="VMware6 storagePath.numberWriteAveraged.average {vmware6StPthName}" \ | |
+--vertical-label="StPthNrWeAdAvg" \ | |
+DEF:xxx={rrd1}:StPthNrWeAdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"StPthNrWeAdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StPthRdAvg.name=vmware6.storagePath.read.average | |
+report.vmware6.StPthRdAvg.columns=StPthRdAvg | |
+report.vmware6.StPthRdAvg.propertiesValues=vmware6StPthName | |
+report.vmware6.StPthRdAvg.type=vmware6StPth | |
+report.vmware6.StPthRdAvg.command=--title="VMware6 storagePath.read.average {vmware6StPthName}" \ | |
+--vertical-label="StPthRdAvg" \ | |
+DEF:xxx={rrd1}:StPthRdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"StPthRdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StPthTlRdLyAvg.name=vmware6.storagePath.totalReadLatency.average | |
+report.vmware6.StPthTlRdLyAvg.columns=StPthTlRdLyAvg | |
+report.vmware6.StPthTlRdLyAvg.propertiesValues=vmware6StPthName | |
+report.vmware6.StPthTlRdLyAvg.type=vmware6StPth | |
+report.vmware6.StPthTlRdLyAvg.command=--title="VMware6 storagePath.totalReadLatency.average {vmware6StPthName}" \ | |
+--vertical-label="StPthTlRdLyAvg" \ | |
+DEF:xxx={rrd1}:StPthTlRdLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"StPthTlRdLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StPthTlWeLyAvg.name=vmware6.storagePath.totalWriteLatency.average | |
+report.vmware6.StPthTlWeLyAvg.columns=StPthTlWeLyAvg | |
+report.vmware6.StPthTlWeLyAvg.propertiesValues=vmware6StPthName | |
+report.vmware6.StPthTlWeLyAvg.type=vmware6StPth | |
+report.vmware6.StPthTlWeLyAvg.command=--title="VMware6 storagePath.totalWriteLatency.average {vmware6StPthName}" \ | |
+--vertical-label="StPthTlWeLyAvg" \ | |
+DEF:xxx={rrd1}:StPthTlWeLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"StPthTlWeLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StPthWeAvg.name=vmware6.storagePath.write.average | |
+report.vmware6.StPthWeAvg.columns=StPthWeAvg | |
+report.vmware6.StPthWeAvg.propertiesValues=vmware6StPthName | |
+report.vmware6.StPthWeAvg.type=vmware6StPth | |
+report.vmware6.StPthWeAvg.command=--title="VMware6 storagePath.write.average {vmware6StPthName}" \ | |
+--vertical-label="StPthWeAvg" \ | |
+DEF:xxx={rrd1}:StPthWeAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"StPthWeAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuRdCyAvg.name=vmware6.cpu.reservedCapacity.average | |
+report.vmware6.CpuRdCyAvg.columns=CpuRdCyAvg | |
+report.vmware6.CpuRdCyAvg.type=nodeSnmp | |
+report.vmware6.CpuRdCyAvg.command=--title="VMware6 cpu.reservedCapacity.average" \ | |
+--vertical-label="CpuRdCyAvg" \ | |
+DEF:xxx={rrd1}:CpuRdCyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuRdCyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuTlCyAvg.name=vmware6.cpu.totalCapacity.average | |
+report.vmware6.CpuTlCyAvg.columns=CpuTlCyAvg | |
+report.vmware6.CpuTlCyAvg.type=nodeSnmp | |
+report.vmware6.CpuTlCyAvg.command=--title="VMware6 cpu.totalCapacity.average" \ | |
+--vertical-label="CpuTlCyAvg" \ | |
+DEF:xxx={rrd1}:CpuTlCyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuTlCyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.HbrHbrNetRxAvg.name=vmware6.hbr.hbrNetRx.average | |
+report.vmware6.HbrHbrNetRxAvg.columns=HbrHbrNetRxAvg | |
+report.vmware6.HbrHbrNetRxAvg.type=nodeSnmp | |
+report.vmware6.HbrHbrNetRxAvg.command=--title="VMware6 hbr.hbrNetRx.average" \ | |
+--vertical-label="HbrHbrNetRxAvg" \ | |
+DEF:xxx={rrd1}:HbrHbrNetRxAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"HbrHbrNetRxAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.HbrHbrNetTxAvg.name=vmware6.hbr.hbrNetTx.average | |
+report.vmware6.HbrHbrNetTxAvg.columns=HbrHbrNetTxAvg | |
+report.vmware6.HbrHbrNetTxAvg.type=nodeSnmp | |
+report.vmware6.HbrHbrNetTxAvg.command=--title="VMware6 hbr.hbrNetTx.average" \ | |
+--vertical-label="HbrHbrNetTxAvg" \ | |
+DEF:xxx={rrd1}:HbrHbrNetTxAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"HbrHbrNetTxAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.HbrHbrNumVmsAvg.name=vmware6.hbr.hbrNumVms.average | |
+report.vmware6.HbrHbrNumVmsAvg.columns=HbrHbrNumVmsAvg | |
+report.vmware6.HbrHbrNumVmsAvg.type=nodeSnmp | |
+report.vmware6.HbrHbrNumVmsAvg.command=--title="VMware6 hbr.hbrNumVms.average" \ | |
+--vertical-label="HbrHbrNumVmsAvg" \ | |
+DEF:xxx={rrd1}:HbrHbrNumVmsAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"HbrHbrNumVmsAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemHeapAvg.name=vmware6.mem.heap.average | |
+report.vmware6.MemHeapAvg.columns=MemHeapAvg | |
+report.vmware6.MemHeapAvg.type=nodeSnmp | |
+report.vmware6.MemHeapAvg.command=--title="VMware6 mem.heap.average" \ | |
+--vertical-label="MemHeapAvg" \ | |
+DEF:xxx={rrd1}:MemHeapAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemHeapAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemHeapfreeAvg.name=vmware6.mem.heapfree.average | |
+report.vmware6.MemHeapfreeAvg.columns=MemHeapfreeAvg | |
+report.vmware6.MemHeapfreeAvg.type=nodeSnmp | |
+report.vmware6.MemHeapfreeAvg.command=--title="VMware6 mem.heapfree.average" \ | |
+--vertical-label="MemHeapfreeAvg" \ | |
+DEF:xxx={rrd1}:MemHeapfreeAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemHeapfreeAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemLlSpInAvg.name=vmware6.mem.llSwapIn.average | |
+report.vmware6.MemLlSpInAvg.columns=MemLlSpInAvg | |
+report.vmware6.MemLlSpInAvg.type=nodeSnmp | |
+report.vmware6.MemLlSpInAvg.command=--title="VMware6 mem.llSwapIn.average" \ | |
+--vertical-label="MemLlSpInAvg" \ | |
+DEF:xxx={rrd1}:MemLlSpInAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemLlSpInAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemLlSpOutAvg.name=vmware6.mem.llSwapOut.average | |
+report.vmware6.MemLlSpOutAvg.columns=MemLlSpOutAvg | |
+report.vmware6.MemLlSpOutAvg.type=nodeSnmp | |
+report.vmware6.MemLlSpOutAvg.command=--title="VMware6 mem.llSwapOut.average" \ | |
+--vertical-label="MemLlSpOutAvg" \ | |
+DEF:xxx={rrd1}:MemLlSpOutAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemLlSpOutAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemLowfreeTdAvg.name=vmware6.mem.lowfreethreshold.average | |
+report.vmware6.MemLowfreeTdAvg.columns=MemLowfreeTdAvg | |
+report.vmware6.MemLowfreeTdAvg.type=nodeSnmp | |
+report.vmware6.MemLowfreeTdAvg.command=--title="VMware6 mem.lowfreethreshold.average" \ | |
+--vertical-label="MemLowfreeTdAvg" \ | |
+DEF:xxx={rrd1}:MemLowfreeTdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemLowfreeTdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemRdCyAvg.name=vmware6.mem.reservedCapacity.average | |
+report.vmware6.MemRdCyAvg.columns=MemRdCyAvg | |
+report.vmware6.MemRdCyAvg.type=nodeSnmp | |
+report.vmware6.MemRdCyAvg.command=--title="VMware6 mem.reservedCapacity.average" \ | |
+--vertical-label="MemRdCyAvg" \ | |
+DEF:xxx={rrd1}:MemRdCyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemRdCyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemSharedcommonAvg.name=vmware6.mem.sharedcommon.average | |
+report.vmware6.MemSharedcommonAvg.columns=MemSharedcommonAvg | |
+report.vmware6.MemSharedcommonAvg.type=nodeSnmp | |
+report.vmware6.MemSharedcommonAvg.command=--title="VMware6 mem.sharedcommon.average" \ | |
+--vertical-label="MemSharedcommonAvg" \ | |
+DEF:xxx={rrd1}:MemSharedcommonAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemSharedcommonAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemSpusedAvg.name=vmware6.mem.swapused.average | |
+report.vmware6.MemSpusedAvg.columns=MemSpusedAvg | |
+report.vmware6.MemSpusedAvg.type=nodeSnmp | |
+report.vmware6.MemSpusedAvg.command=--title="VMware6 mem.swapused.average" \ | |
+--vertical-label="MemSpusedAvg" \ | |
+DEF:xxx={rrd1}:MemSpusedAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemSpusedAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemStateLat.name=vmware6.mem.state.latest | |
+report.vmware6.MemStateLat.columns=MemStateLat | |
+report.vmware6.MemStateLat.type=nodeSnmp | |
+report.vmware6.MemStateLat.command=--title="VMware6 mem.state.latest" \ | |
+--vertical-label="MemStateLat" \ | |
+DEF:xxx={rrd1}:MemStateLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemStateLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemSysUsageAvg.name=vmware6.mem.sysUsage.average | |
+report.vmware6.MemSysUsageAvg.columns=MemSysUsageAvg | |
+report.vmware6.MemSysUsageAvg.type=nodeSnmp | |
+report.vmware6.MemSysUsageAvg.command=--title="VMware6 mem.sysUsage.average" \ | |
+--vertical-label="MemSysUsageAvg" \ | |
+DEF:xxx={rrd1}:MemSysUsageAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemSysUsageAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemTlCyAvg.name=vmware6.mem.totalCapacity.average | |
+report.vmware6.MemTlCyAvg.columns=MemTlCyAvg | |
+report.vmware6.MemTlCyAvg.type=nodeSnmp | |
+report.vmware6.MemTlCyAvg.command=--title="VMware6 mem.totalCapacity.average" \ | |
+--vertical-label="MemTlCyAvg" \ | |
+DEF:xxx={rrd1}:MemTlCyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemTlCyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemUdAvg.name=vmware6.mem.unreserved.average | |
+report.vmware6.MemUdAvg.columns=MemUdAvg | |
+report.vmware6.MemUdAvg.type=nodeSnmp | |
+report.vmware6.MemUdAvg.command=--title="VMware6 mem.unreserved.average" \ | |
+--vertical-label="MemUdAvg" \ | |
+DEF:xxx={rrd1}:MemUdAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemUdAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemVmfs.pbc.OdLat.name=vmware6.mem.vmfs.pbc.overhead.latest | |
+report.vmware6.MemVmfs.pbc.OdLat.columns=MemVmfsPbcOdLat | |
+report.vmware6.MemVmfs.pbc.OdLat.type=nodeSnmp | |
+report.vmware6.MemVmfs.pbc.OdLat.command=--title="VMware6 mem.vmfs.pbc.overhead.latest" \ | |
+--vertical-label="MemVmfs.pbc.OdLat" \ | |
+DEF:xxx={rrd1}:MemVmfsPbcOdLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemVmfs.pbc.OdLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemVmfsPbcCpMsRtiLt.name=vmware6.mem.vmfs.pbc.capMissRatio.latest | |
+report.vmware6.MemVmfsPbcCpMsRtiLt.columns=MemVmfsPbcCpMsRtiLt | |
+report.vmware6.MemVmfsPbcCpMsRtiLt.type=nodeSnmp | |
+report.vmware6.MemVmfsPbcCpMsRtiLt.command=--title="VMware6 mem.vmfs.pbc.capMissRatio.latest" \ | |
+--vertical-label="MemVmfsPbcCpMsRtiLt" \ | |
+DEF:xxx={rrd1}:MemVmfsPbcCpMsRtiLt:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemVmfsPbcCpMsRtiLt" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemVmfs.pbc.sizeLat.name=vmware6.mem.vmfs.pbc.size.latest | |
+report.vmware6.MemVmfs.pbc.sizeLat.columns=MemVmfsPbcSizeLat | |
+report.vmware6.MemVmfs.pbc.sizeLat.type=nodeSnmp | |
+report.vmware6.MemVmfs.pbc.sizeLat.command=--title="VMware6 mem.vmfs.pbc.size.latest" \ | |
+--vertical-label="MemVmfs.pbc.sizeLat" \ | |
+DEF:xxx={rrd1}:MemVmfsPbcSizeLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemVmfs.pbc.sizeLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemVmfsPbcSizMaxLat.name=vmware6.mem.vmfs.pbc.sizeMax.latest | |
+report.vmware6.MemVmfsPbcSizMaxLat.columns=MemVmfsPbcSizMaxLat | |
+report.vmware6.MemVmfsPbcSizMaxLat.type=nodeSnmp | |
+report.vmware6.MemVmfsPbcSizMaxLat.command=--title="VMware6 mem.vmfs.pbc.sizeMax.latest" \ | |
+--vertical-label="MemVmfsPbcSizMaxLat" \ | |
+DEF:xxx={rrd1}:MemVmfsPbcSizMaxLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemVmfsPbcSizMaxLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemVmfsPbcWrkSetLat.name=vmware6.mem.vmfs.pbc.workingSet.latest | |
+report.vmware6.MemVmfsPbcWrkSetLat.columns=MemVmfsPbcWrkSetLat | |
+report.vmware6.MemVmfsPbcWrkSetLat.type=nodeSnmp | |
+report.vmware6.MemVmfsPbcWrkSetLat.command=--title="VMware6 mem.vmfs.pbc.workingSet.latest" \ | |
+--vertical-label="MemVmfsPbcWrkSetLat" \ | |
+DEF:xxx={rrd1}:MemVmfsPbcWrkSetLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemVmfsPbcWrkSetLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.MemVmfsPbcWrkStMxLt.name=vmware6.mem.vmfs.pbc.workingSetMax.latest | |
+report.vmware6.MemVmfsPbcWrkStMxLt.columns=MemVmfsPbcWrkStMxLt | |
+report.vmware6.MemVmfsPbcWrkStMxLt.type=nodeSnmp | |
+report.vmware6.MemVmfsPbcWrkStMxLt.command=--title="VMware6 mem.vmfs.pbc.workingSetMax.latest" \ | |
+--vertical-label="MemVmfsPbcWrkStMxLt" \ | |
+DEF:xxx={rrd1}:MemVmfsPbcWrkStMxLt:AVERAGE \ | |
+LINE2:xxx#0000ff:"MemVmfsPbcWrkStMxLt" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.PowerPowerCapAvg.name=vmware6.power.powerCap.average | |
+report.vmware6.PowerPowerCapAvg.columns=PowerPowerCapAvg | |
+report.vmware6.PowerPowerCapAvg.type=nodeSnmp | |
+report.vmware6.PowerPowerCapAvg.command=--title="VMware6 power.powerCap.average" \ | |
+--vertical-label="PowerPowerCapAvg" \ | |
+DEF:xxx={rrd1}:PowerPowerCapAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"PowerPowerCapAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StAdptrMaxTlLyLat.name=vmware6.storageAdapter.maxTotalLatency.latest | |
+report.vmware6.StAdptrMaxTlLyLat.columns=StAdptrMaxTlLyLat | |
+report.vmware6.StAdptrMaxTlLyLat.type=nodeSnmp | |
+report.vmware6.StAdptrMaxTlLyLat.command=--title="VMware6 storageAdapter.maxTotalLatency.latest" \ | |
+--vertical-label="StAdptrMaxTlLyLat" \ | |
+DEF:xxx={rrd1}:StAdptrMaxTlLyLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"StAdptrMaxTlLyLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.StPthMaxTlLyLat.name=vmware6.storagePath.maxTotalLatency.latest | |
+report.vmware6.StPthMaxTlLyLat.columns=StPthMaxTlLyLat | |
+report.vmware6.StPthMaxTlLyLat.type=nodeSnmp | |
+report.vmware6.StPthMaxTlLyLat.command=--title="VMware6 storagePath.maxTotalLatency.latest" \ | |
+--vertical-label="StPthMaxTlLyLat" \ | |
+DEF:xxx={rrd1}:StPthMaxTlLyLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"StPthMaxTlLyLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuCoreUnAvg.name=vmware6.cpu.coreUtilization.average | |
+report.vmware6.CpuCoreUnAvg.columns=CpuCoreUnAvg | |
+report.vmware6.CpuCoreUnAvg.propertiesValues=vmware6CpuName | |
+report.vmware6.CpuCoreUnAvg.type=vmware6Cpu | |
+report.vmware6.CpuCoreUnAvg.command=--title="VMware6 cpu.coreUtilization.average {vmware6CpuName}" \ | |
+--vertical-label="CpuCoreUnAvg" \ | |
+DEF:xxx={rrd1}:CpuCoreUnAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuCoreUnAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.CpuUnAvg.name=vmware6.cpu.utilization.average | |
+report.vmware6.CpuUnAvg.columns=CpuUnAvg | |
+report.vmware6.CpuUnAvg.propertiesValues=vmware6CpuName | |
+report.vmware6.CpuUnAvg.type=vmware6Cpu | |
+report.vmware6.CpuUnAvg.command=--title="VMware6 cpu.utilization.average {vmware6CpuName}" \ | |
+--vertical-label="CpuUnAvg" \ | |
+DEF:xxx={rrd1}:CpuUnAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"CpuUnAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReCpuAcMinLat.name=vmware6.sys.resourceCpuAllocMin.latest | |
+report.vmware6.SysReCpuAcMinLat.columns=SysReCpuAcMinLat | |
+report.vmware6.SysReCpuAcMinLat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReCpuAcMinLat.type=vmware6Sys | |
+report.vmware6.SysReCpuAcMinLat.command=--title="VMware6 sys.resourceCpuAllocMin.latest {vmware6SysName}" \ | |
+--vertical-label="SysReCpuAcMinLat" \ | |
+DEF:xxx={rrd1}:SysReCpuAcMinLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReCpuAcMinLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReCpuAcSsLat.name=vmware6.sys.resourceCpuAllocShares.latest | |
+report.vmware6.SysReCpuAcSsLat.columns=SysReCpuAcSsLat | |
+report.vmware6.SysReCpuAcSsLat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReCpuAcSsLat.type=vmware6Sys | |
+report.vmware6.SysReCpuAcSsLat.command=--title="VMware6 sys.resourceCpuAllocShares.latest {vmware6SysName}" \ | |
+--vertical-label="SysReCpuAcSsLat" \ | |
+DEF:xxx={rrd1}:SysReCpuAcSsLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReCpuAcSsLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReCpuAct1Lat.name=vmware6.sys.resourceCpuAct1.latest | |
+report.vmware6.SysReCpuAct1Lat.columns=SysReCpuAct1Lat | |
+report.vmware6.SysReCpuAct1Lat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReCpuAct1Lat.type=vmware6Sys | |
+report.vmware6.SysReCpuAct1Lat.command=--title="VMware6 sys.resourceCpuAct1.latest {vmware6SysName}" \ | |
+--vertical-label="SysReCpuAct1Lat" \ | |
+DEF:xxx={rrd1}:SysReCpuAct1Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReCpuAct1Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReCpuAct5Lat.name=vmware6.sys.resourceCpuAct5.latest | |
+report.vmware6.SysReCpuAct5Lat.columns=SysReCpuAct5Lat | |
+report.vmware6.SysReCpuAct5Lat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReCpuAct5Lat.type=vmware6Sys | |
+report.vmware6.SysReCpuAct5Lat.command=--title="VMware6 sys.resourceCpuAct5.latest {vmware6SysName}" \ | |
+--vertical-label="SysReCpuAct5Lat" \ | |
+DEF:xxx={rrd1}:SysReCpuAct5Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReCpuAct5Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReCpuMaxLd1Lat.name=vmware6.sys.resourceCpuMaxLimited1.latest | |
+report.vmware6.SysReCpuMaxLd1Lat.columns=SysReCpuMaxLd1Lat | |
+report.vmware6.SysReCpuMaxLd1Lat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReCpuMaxLd1Lat.type=vmware6Sys | |
+report.vmware6.SysReCpuMaxLd1Lat.command=--title="VMware6 sys.resourceCpuMaxLimited1.latest {vmware6SysName}" \ | |
+--vertical-label="SysReCpuMaxLd1Lat" \ | |
+DEF:xxx={rrd1}:SysReCpuMaxLd1Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReCpuMaxLd1Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReCpuMaxLd5Lat.name=vmware6.sys.resourceCpuMaxLimited5.latest | |
+report.vmware6.SysReCpuMaxLd5Lat.columns=SysReCpuMaxLd5Lat | |
+report.vmware6.SysReCpuMaxLd5Lat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReCpuMaxLd5Lat.type=vmware6Sys | |
+report.vmware6.SysReCpuMaxLd5Lat.command=--title="VMware6 sys.resourceCpuMaxLimited5.latest {vmware6SysName}" \ | |
+--vertical-label="SysReCpuMaxLd5Lat" \ | |
+DEF:xxx={rrd1}:SysReCpuMaxLd5Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReCpuMaxLd5Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReCpuRun1Lat.name=vmware6.sys.resourceCpuRun1.latest | |
+report.vmware6.SysReCpuRun1Lat.columns=SysReCpuRun1Lat | |
+report.vmware6.SysReCpuRun1Lat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReCpuRun1Lat.type=vmware6Sys | |
+report.vmware6.SysReCpuRun1Lat.command=--title="VMware6 sys.resourceCpuRun1.latest {vmware6SysName}" \ | |
+--vertical-label="SysReCpuRun1Lat" \ | |
+DEF:xxx={rrd1}:SysReCpuRun1Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReCpuRun1Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReCpuRun5Lat.name=vmware6.sys.resourceCpuRun5.latest | |
+report.vmware6.SysReCpuRun5Lat.columns=SysReCpuRun5Lat | |
+report.vmware6.SysReCpuRun5Lat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReCpuRun5Lat.type=vmware6Sys | |
+report.vmware6.SysReCpuRun5Lat.command=--title="VMware6 sys.resourceCpuRun5.latest {vmware6SysName}" \ | |
+--vertical-label="SysReCpuRun5Lat" \ | |
+DEF:xxx={rrd1}:SysReCpuRun5Lat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReCpuRun5Lat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReCpuUsageAvg.name=vmware6.sys.resourceCpuUsage.average | |
+report.vmware6.SysReCpuUsageAvg.columns=SysReCpuUsageAvg | |
+report.vmware6.SysReCpuUsageAvg.propertiesValues=vmware6SysName | |
+report.vmware6.SysReCpuUsageAvg.type=vmware6Sys | |
+report.vmware6.SysReCpuUsageAvg.command=--title="VMware6 sys.resourceCpuUsage.average {vmware6SysName}" \ | |
+--vertical-label="SysReCpuUsageAvg" \ | |
+DEF:xxx={rrd1}:SysReCpuUsageAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReCpuUsageAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReFdUsageLat.name=vmware6.sys.resourceFdUsage.latest | |
+report.vmware6.SysReFdUsageLat.columns=SysReFdUsageLat | |
+report.vmware6.SysReFdUsageLat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReFdUsageLat.type=vmware6Sys | |
+report.vmware6.SysReFdUsageLat.command=--title="VMware6 sys.resourceFdUsage.latest {vmware6SysName}" \ | |
+--vertical-label="SysReFdUsageLat" \ | |
+DEF:xxx={rrd1}:SysReFdUsageLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReFdUsageLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReMemAcMaxLat.name=vmware6.sys.resourceMemAllocMax.latest | |
+report.vmware6.SysReMemAcMaxLat.columns=SysReMemAcMaxLat | |
+report.vmware6.SysReMemAcMaxLat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReMemAcMaxLat.type=vmware6Sys | |
+report.vmware6.SysReMemAcMaxLat.command=--title="VMware6 sys.resourceMemAllocMax.latest {vmware6SysName}" \ | |
+--vertical-label="SysReMemAcMaxLat" \ | |
+DEF:xxx={rrd1}:SysReMemAcMaxLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReMemAcMaxLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReMemAcMinLat.name=vmware6.sys.resourceMemAllocMin.latest | |
+report.vmware6.SysReMemAcMinLat.columns=SysReMemAcMinLat | |
+report.vmware6.SysReMemAcMinLat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReMemAcMinLat.type=vmware6Sys | |
+report.vmware6.SysReMemAcMinLat.command=--title="VMware6 sys.resourceMemAllocMin.latest {vmware6SysName}" \ | |
+--vertical-label="SysReMemAcMinLat" \ | |
+DEF:xxx={rrd1}:SysReMemAcMinLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReMemAcMinLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReMemAcSsLat.name=vmware6.sys.resourceMemAllocShares.latest | |
+report.vmware6.SysReMemAcSsLat.columns=SysReMemAcSsLat | |
+report.vmware6.SysReMemAcSsLat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReMemAcSsLat.type=vmware6Sys | |
+report.vmware6.SysReMemAcSsLat.command=--title="VMware6 sys.resourceMemAllocShares.latest {vmware6SysName}" \ | |
+--vertical-label="SysReMemAcSsLat" \ | |
+DEF:xxx={rrd1}:SysReMemAcSsLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReMemAcSsLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReMemCdLat.name=vmware6.sys.resourceMemConsumed.latest | |
+report.vmware6.SysReMemCdLat.columns=SysReMemCdLat | |
+report.vmware6.SysReMemCdLat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReMemCdLat.type=vmware6Sys | |
+report.vmware6.SysReMemCdLat.command=--title="VMware6 sys.resourceMemConsumed.latest {vmware6SysName}" \ | |
+--vertical-label="SysReMemCdLat" \ | |
+DEF:xxx={rrd1}:SysReMemCdLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReMemCdLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReMemCowLat.name=vmware6.sys.resourceMemCow.latest | |
+report.vmware6.SysReMemCowLat.columns=SysReMemCowLat | |
+report.vmware6.SysReMemCowLat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReMemCowLat.type=vmware6Sys | |
+report.vmware6.SysReMemCowLat.command=--title="VMware6 sys.resourceMemCow.latest {vmware6SysName}" \ | |
+--vertical-label="SysReMemCowLat" \ | |
+DEF:xxx={rrd1}:SysReMemCowLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReMemCowLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReMemMappedLat.name=vmware6.sys.resourceMemMapped.latest | |
+report.vmware6.SysReMemMappedLat.columns=SysReMemMappedLat | |
+report.vmware6.SysReMemMappedLat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReMemMappedLat.type=vmware6Sys | |
+report.vmware6.SysReMemMappedLat.command=--title="VMware6 sys.resourceMemMapped.latest {vmware6SysName}" \ | |
+--vertical-label="SysReMemMappedLat" \ | |
+DEF:xxx={rrd1}:SysReMemMappedLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReMemMappedLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReMemOdLat.name=vmware6.sys.resourceMemOverhead.latest | |
+report.vmware6.SysReMemOdLat.columns=SysReMemOdLat | |
+report.vmware6.SysReMemOdLat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReMemOdLat.type=vmware6Sys | |
+report.vmware6.SysReMemOdLat.command=--title="VMware6 sys.resourceMemOverhead.latest {vmware6SysName}" \ | |
+--vertical-label="SysReMemOdLat" \ | |
+DEF:xxx={rrd1}:SysReMemOdLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReMemOdLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReMemSharedLat.name=vmware6.sys.resourceMemShared.latest | |
+report.vmware6.SysReMemSharedLat.columns=SysReMemSharedLat | |
+report.vmware6.SysReMemSharedLat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReMemSharedLat.type=vmware6Sys | |
+report.vmware6.SysReMemSharedLat.command=--title="VMware6 sys.resourceMemShared.latest {vmware6SysName}" \ | |
+--vertical-label="SysReMemSharedLat" \ | |
+DEF:xxx={rrd1}:SysReMemSharedLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReMemSharedLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReMemSppedLat.name=vmware6.sys.resourceMemSwapped.latest | |
+report.vmware6.SysReMemSppedLat.columns=SysReMemSppedLat | |
+report.vmware6.SysReMemSppedLat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReMemSppedLat.type=vmware6Sys | |
+report.vmware6.SysReMemSppedLat.command=--title="VMware6 sys.resourceMemSwapped.latest {vmware6SysName}" \ | |
+--vertical-label="SysReMemSppedLat" \ | |
+DEF:xxx={rrd1}:SysReMemSppedLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReMemSppedLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReMemTdLat.name=vmware6.sys.resourceMemTouched.latest | |
+report.vmware6.SysReMemTdLat.columns=SysReMemTdLat | |
+report.vmware6.SysReMemTdLat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReMemTdLat.type=vmware6Sys | |
+report.vmware6.SysReMemTdLat.command=--title="VMware6 sys.resourceMemTouched.latest {vmware6SysName}" \ | |
+--vertical-label="SysReMemTdLat" \ | |
+DEF:xxx={rrd1}:SysReMemTdLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReMemTdLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.SysReMemZeroLat.name=vmware6.sys.resourceMemZero.latest | |
+report.vmware6.SysReMemZeroLat.columns=SysReMemZeroLat | |
+report.vmware6.SysReMemZeroLat.propertiesValues=vmware6SysName | |
+report.vmware6.SysReMemZeroLat.type=vmware6Sys | |
+report.vmware6.SysReMemZeroLat.command=--title="VMware6 sys.resourceMemZero.latest {vmware6SysName}" \ | |
+--vertical-label="SysReMemZeroLat" \ | |
+DEF:xxx={rrd1}:SysReMemZeroLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"SysReMemZeroLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetErrorsRxSum.name=vmware6.net.errorsRx.summation | |
+report.vmware6.NetErrorsRxSum.columns=NetErrorsRxSum | |
+report.vmware6.NetErrorsRxSum.propertiesValues=vmware6NetName | |
+report.vmware6.NetErrorsRxSum.type=vmware6Net | |
+report.vmware6.NetErrorsRxSum.command=--title="VMware6 net.errorsRx.summation {vmware6NetName}" \ | |
+--vertical-label="NetErrorsRxSum" \ | |
+DEF:xxx={rrd1}:NetErrorsRxSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetErrorsRxSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetErrorsTxSum.name=vmware6.net.errorsTx.summation | |
+report.vmware6.NetErrorsTxSum.columns=NetErrorsTxSum | |
+report.vmware6.NetErrorsTxSum.propertiesValues=vmware6NetName | |
+report.vmware6.NetErrorsTxSum.type=vmware6Net | |
+report.vmware6.NetErrorsTxSum.command=--title="VMware6 net.errorsTx.summation {vmware6NetName}" \ | |
+--vertical-label="NetErrorsTxSum" \ | |
+DEF:xxx={rrd1}:NetErrorsTxSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetErrorsTxSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.NetUnknownPsSum.name=vmware6.net.unknownProtos.summation | |
+report.vmware6.NetUnknownPsSum.columns=NetUnknownPsSum | |
+report.vmware6.NetUnknownPsSum.propertiesValues=vmware6NetName | |
+report.vmware6.NetUnknownPsSum.type=vmware6Net | |
+report.vmware6.NetUnknownPsSum.command=--title="VMware6 net.unknownProtos.summation {vmware6NetName}" \ | |
+--vertical-label="NetUnknownPsSum" \ | |
+DEF:xxx={rrd1}:NetUnknownPsSum:AVERAGE \ | |
+LINE2:xxx#0000ff:"NetUnknownPsSum" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskDeLyAvg.name=vmware6.disk.deviceLatency.average | |
+report.vmware6.DiskDeLyAvg.columns=DiskDeLyAvg | |
+report.vmware6.DiskDeLyAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskDeLyAvg.type=vmware6Disk | |
+report.vmware6.DiskDeLyAvg.command=--title="VMware6 disk.deviceLatency.average {vmware6DiskName}" \ | |
+--vertical-label="DiskDeLyAvg" \ | |
+DEF:xxx={rrd1}:DiskDeLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskDeLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskDeRdLyAvg.name=vmware6.disk.deviceReadLatency.average | |
+report.vmware6.DiskDeRdLyAvg.columns=DiskDeRdLyAvg | |
+report.vmware6.DiskDeRdLyAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskDeRdLyAvg.type=vmware6Disk | |
+report.vmware6.DiskDeRdLyAvg.command=--title="VMware6 disk.deviceReadLatency.average {vmware6DiskName}" \ | |
+--vertical-label="DiskDeRdLyAvg" \ | |
+DEF:xxx={rrd1}:DiskDeRdLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskDeRdLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskDeWeLyAvg.name=vmware6.disk.deviceWriteLatency.average | |
+report.vmware6.DiskDeWeLyAvg.columns=DiskDeWeLyAvg | |
+report.vmware6.DiskDeWeLyAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskDeWeLyAvg.type=vmware6Disk | |
+report.vmware6.DiskDeWeLyAvg.command=--title="VMware6 disk.deviceWriteLatency.average {vmware6DiskName}" \ | |
+--vertical-label="DiskDeWeLyAvg" \ | |
+DEF:xxx={rrd1}:DiskDeWeLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskDeWeLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskKlLyAvg.name=vmware6.disk.kernelLatency.average | |
+report.vmware6.DiskKlLyAvg.columns=DiskKlLyAvg | |
+report.vmware6.DiskKlLyAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskKlLyAvg.type=vmware6Disk | |
+report.vmware6.DiskKlLyAvg.command=--title="VMware6 disk.kernelLatency.average {vmware6DiskName}" \ | |
+--vertical-label="DiskKlLyAvg" \ | |
+DEF:xxx={rrd1}:DiskKlLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskKlLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskKlRdLyAvg.name=vmware6.disk.kernelReadLatency.average | |
+report.vmware6.DiskKlRdLyAvg.columns=DiskKlRdLyAvg | |
+report.vmware6.DiskKlRdLyAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskKlRdLyAvg.type=vmware6Disk | |
+report.vmware6.DiskKlRdLyAvg.command=--title="VMware6 disk.kernelReadLatency.average {vmware6DiskName}" \ | |
+--vertical-label="DiskKlRdLyAvg" \ | |
+DEF:xxx={rrd1}:DiskKlRdLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskKlRdLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskKlWeLyAvg.name=vmware6.disk.kernelWriteLatency.average | |
+report.vmware6.DiskKlWeLyAvg.columns=DiskKlWeLyAvg | |
+report.vmware6.DiskKlWeLyAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskKlWeLyAvg.type=vmware6Disk | |
+report.vmware6.DiskKlWeLyAvg.command=--title="VMware6 disk.kernelWriteLatency.average {vmware6DiskName}" \ | |
+--vertical-label="DiskKlWeLyAvg" \ | |
+DEF:xxx={rrd1}:DiskKlWeLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskKlWeLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskMaxQeDhAvg.name=vmware6.disk.maxQueueDepth.average | |
+report.vmware6.DiskMaxQeDhAvg.columns=DiskMaxQeDhAvg | |
+report.vmware6.DiskMaxQeDhAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskMaxQeDhAvg.type=vmware6Disk | |
+report.vmware6.DiskMaxQeDhAvg.command=--title="VMware6 disk.maxQueueDepth.average {vmware6DiskName}" \ | |
+--vertical-label="DiskMaxQeDhAvg" \ | |
+DEF:xxx={rrd1}:DiskMaxQeDhAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskMaxQeDhAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskQeLyAvg.name=vmware6.disk.queueLatency.average | |
+report.vmware6.DiskQeLyAvg.columns=DiskQeLyAvg | |
+report.vmware6.DiskQeLyAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskQeLyAvg.type=vmware6Disk | |
+report.vmware6.DiskQeLyAvg.command=--title="VMware6 disk.queueLatency.average {vmware6DiskName}" \ | |
+--vertical-label="DiskQeLyAvg" \ | |
+DEF:xxx={rrd1}:DiskQeLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskQeLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskQeRdLyAvg.name=vmware6.disk.queueReadLatency.average | |
+report.vmware6.DiskQeRdLyAvg.columns=DiskQeRdLyAvg | |
+report.vmware6.DiskQeRdLyAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskQeRdLyAvg.type=vmware6Disk | |
+report.vmware6.DiskQeRdLyAvg.command=--title="VMware6 disk.queueReadLatency.average {vmware6DiskName}" \ | |
+--vertical-label="DiskQeRdLyAvg" \ | |
+DEF:xxx={rrd1}:DiskQeRdLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskQeRdLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskQeWeLyAvg.name=vmware6.disk.queueWriteLatency.average | |
+report.vmware6.DiskQeWeLyAvg.columns=DiskQeWeLyAvg | |
+report.vmware6.DiskQeWeLyAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskQeWeLyAvg.type=vmware6Disk | |
+report.vmware6.DiskQeWeLyAvg.command=--title="VMware6 disk.queueWriteLatency.average {vmware6DiskName}" \ | |
+--vertical-label="DiskQeWeLyAvg" \ | |
+DEF:xxx={rrd1}:DiskQeWeLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskQeWeLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskTlLyAvg.name=vmware6.disk.totalLatency.average | |
+report.vmware6.DiskTlLyAvg.columns=DiskTlLyAvg | |
+report.vmware6.DiskTlLyAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskTlLyAvg.type=vmware6Disk | |
+report.vmware6.DiskTlLyAvg.command=--title="VMware6 disk.totalLatency.average {vmware6DiskName}" \ | |
+--vertical-label="DiskTlLyAvg" \ | |
+DEF:xxx={rrd1}:DiskTlLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskTlLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskTlRdLyAvg.name=vmware6.disk.totalReadLatency.average | |
+report.vmware6.DiskTlRdLyAvg.columns=DiskTlRdLyAvg | |
+report.vmware6.DiskTlRdLyAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskTlRdLyAvg.type=vmware6Disk | |
+report.vmware6.DiskTlRdLyAvg.command=--title="VMware6 disk.totalReadLatency.average {vmware6DiskName}" \ | |
+--vertical-label="DiskTlRdLyAvg" \ | |
+DEF:xxx={rrd1}:DiskTlRdLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskTlRdLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DiskTlWeLyAvg.name=vmware6.disk.totalWriteLatency.average | |
+report.vmware6.DiskTlWeLyAvg.columns=DiskTlWeLyAvg | |
+report.vmware6.DiskTlWeLyAvg.propertiesValues=vmware6DiskName | |
+report.vmware6.DiskTlWeLyAvg.type=vmware6Disk | |
+report.vmware6.DiskTlWeLyAvg.command=--title="VMware6 disk.totalWriteLatency.average {vmware6DiskName}" \ | |
+--vertical-label="DiskTlWeLyAvg" \ | |
+DEF:xxx={rrd1}:DiskTlWeLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DiskTlWeLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.vflModNumAeVMDKsLat.name=vmware6.vflashModule.numActiveVMDKs.latest | |
+report.vmware6.vflModNumAeVMDKsLat.columns=vflModNumAeVMDKsLat | |
+report.vmware6.vflModNumAeVMDKsLat.propertiesValues=vmware6vflashModuleName | |
+report.vmware6.vflModNumAeVMDKsLat.type=vmware6vflashModule | |
+report.vmware6.vflModNumAeVMDKsLat.command=--title="VMware6 vflashModule.numActiveVMDKs.latest {vmware6vflashModuleName}" \ | |
+--vertical-label="vflModNumAeVMDKsLat" \ | |
+DEF:xxx={rrd1}:vflModNumAeVMDKsLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"vflModNumAeVMDKsLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStDeIopsAvg.name=vmware6.datastore.datastoreIops.average | |
+report.vmware6.DaStDeIopsAvg.columns=DaStDeIopsAvg | |
+report.vmware6.DaStDeIopsAvg.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStDeIopsAvg.type=vmware6DaSt | |
+report.vmware6.DaStDeIopsAvg.command=--title="VMware6 datastore.datastoreIops.average {vmware6DaStName}" \ | |
+--vertical-label="DaStDeIopsAvg" \ | |
+DEF:xxx={rrd1}:DaStDeIopsAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStDeIopsAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStDeMaxQeDhLat.name=vmware6.datastore.datastoreMaxQueueDepth.latest | |
+report.vmware6.DaStDeMaxQeDhLat.columns=DaStDeMaxQeDhLat | |
+report.vmware6.DaStDeMaxQeDhLat.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStDeMaxQeDhLat.type=vmware6DaSt | |
+report.vmware6.DaStDeMaxQeDhLat.command=--title="VMware6 datastore.datastoreMaxQueueDepth.latest {vmware6DaStName}" \ | |
+--vertical-label="DaStDeMaxQeDhLat" \ | |
+DEF:xxx={rrd1}:DaStDeMaxQeDhLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStDeMaxQeDhLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStDeNlRdLyLat.name=vmware6.datastore.datastoreNormalReadLatency.latest | |
+report.vmware6.DaStDeNlRdLyLat.columns=DaStDeNlRdLyLat | |
+report.vmware6.DaStDeNlRdLyLat.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStDeNlRdLyLat.type=vmware6DaSt | |
+report.vmware6.DaStDeNlRdLyLat.command=--title="VMware6 datastore.datastoreNormalReadLatency.latest {vmware6DaStName}" \ | |
+--vertical-label="DaStDeNlRdLyLat" \ | |
+DEF:xxx={rrd1}:DaStDeNlRdLyLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStDeNlRdLyLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStDeNlWeLyLat.name=vmware6.datastore.datastoreNormalWriteLatency.latest | |
+report.vmware6.DaStDeNlWeLyLat.columns=DaStDeNlWeLyLat | |
+report.vmware6.DaStDeNlWeLyLat.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStDeNlWeLyLat.type=vmware6DaSt | |
+report.vmware6.DaStDeNlWeLyLat.command=--title="VMware6 datastore.datastoreNormalWriteLatency.latest {vmware6DaStName}" \ | |
+--vertical-label="DaStDeNlWeLyLat" \ | |
+DEF:xxx={rrd1}:DaStDeNlWeLyLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStDeNlWeLyLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStDeRdBytesLat.name=vmware6.datastore.datastoreReadBytes.latest | |
+report.vmware6.DaStDeRdBytesLat.columns=DaStDeRdBytesLat | |
+report.vmware6.DaStDeRdBytesLat.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStDeRdBytesLat.type=vmware6DaSt | |
+report.vmware6.DaStDeRdBytesLat.command=--title="VMware6 datastore.datastoreReadBytes.latest {vmware6DaStName}" \ | |
+--vertical-label="DaStDeRdBytesLat" \ | |
+DEF:xxx={rrd1}:DaStDeRdBytesLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStDeRdBytesLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStDeRdIopsLat.name=vmware6.datastore.datastoreReadIops.latest | |
+report.vmware6.DaStDeRdIopsLat.columns=DaStDeRdIopsLat | |
+report.vmware6.DaStDeRdIopsLat.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStDeRdIopsLat.type=vmware6DaSt | |
+report.vmware6.DaStDeRdIopsLat.command=--title="VMware6 datastore.datastoreReadIops.latest {vmware6DaStName}" \ | |
+--vertical-label="DaStDeRdIopsLat" \ | |
+DEF:xxx={rrd1}:DaStDeRdIopsLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStDeRdIopsLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStDeRdLdMcLat.name=vmware6.datastore.datastoreReadLoadMetric.latest | |
+report.vmware6.DaStDeRdLdMcLat.columns=DaStDeRdLdMcLat | |
+report.vmware6.DaStDeRdLdMcLat.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStDeRdLdMcLat.type=vmware6DaSt | |
+report.vmware6.DaStDeRdLdMcLat.command=--title="VMware6 datastore.datastoreReadLoadMetric.latest {vmware6DaStName}" \ | |
+--vertical-label="DaStDeRdLdMcLat" \ | |
+DEF:xxx={rrd1}:DaStDeRdLdMcLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStDeRdLdMcLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStDeRdOIOLat.name=vmware6.datastore.datastoreReadOIO.latest | |
+report.vmware6.DaStDeRdOIOLat.columns=DaStDeRdOIOLat | |
+report.vmware6.DaStDeRdOIOLat.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStDeRdOIOLat.type=vmware6DaSt | |
+report.vmware6.DaStDeRdOIOLat.command=--title="VMware6 datastore.datastoreReadOIO.latest {vmware6DaStName}" \ | |
+--vertical-label="DaStDeRdOIOLat" \ | |
+DEF:xxx={rrd1}:DaStDeRdOIOLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStDeRdOIOLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStDeVMOdLyLat.name=vmware6.datastore.datastoreVMObservedLatency.latest | |
+report.vmware6.DaStDeVMOdLyLat.columns=DaStDeVMOdLyLat | |
+report.vmware6.DaStDeVMOdLyLat.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStDeVMOdLyLat.type=vmware6DaSt | |
+report.vmware6.DaStDeVMOdLyLat.command=--title="VMware6 datastore.datastoreVMObservedLatency.latest {vmware6DaStName}" \ | |
+--vertical-label="DaStDeVMOdLyLat" \ | |
+DEF:xxx={rrd1}:DaStDeVMOdLyLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStDeVMOdLyLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStDeWeBytesLat.name=vmware6.datastore.datastoreWriteBytes.latest | |
+report.vmware6.DaStDeWeBytesLat.columns=DaStDeWeBytesLat | |
+report.vmware6.DaStDeWeBytesLat.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStDeWeBytesLat.type=vmware6DaSt | |
+report.vmware6.DaStDeWeBytesLat.command=--title="VMware6 datastore.datastoreWriteBytes.latest {vmware6DaStName}" \ | |
+--vertical-label="DaStDeWeBytesLat" \ | |
+DEF:xxx={rrd1}:DaStDeWeBytesLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStDeWeBytesLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStDeWeIopsLat.name=vmware6.datastore.datastoreWriteIops.latest | |
+report.vmware6.DaStDeWeIopsLat.columns=DaStDeWeIopsLat | |
+report.vmware6.DaStDeWeIopsLat.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStDeWeIopsLat.type=vmware6DaSt | |
+report.vmware6.DaStDeWeIopsLat.command=--title="VMware6 datastore.datastoreWriteIops.latest {vmware6DaStName}" \ | |
+--vertical-label="DaStDeWeIopsLat" \ | |
+DEF:xxx={rrd1}:DaStDeWeIopsLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStDeWeIopsLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStDeWeLdMcLat.name=vmware6.datastore.datastoreWriteLoadMetric.latest | |
+report.vmware6.DaStDeWeLdMcLat.columns=DaStDeWeLdMcLat | |
+report.vmware6.DaStDeWeLdMcLat.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStDeWeLdMcLat.type=vmware6DaSt | |
+report.vmware6.DaStDeWeLdMcLat.command=--title="VMware6 datastore.datastoreWriteLoadMetric.latest {vmware6DaStName}" \ | |
+--vertical-label="DaStDeWeLdMcLat" \ | |
+DEF:xxx={rrd1}:DaStDeWeLdMcLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStDeWeLdMcLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStDeWeOIOLat.name=vmware6.datastore.datastoreWriteOIO.latest | |
+report.vmware6.DaStDeWeOIOLat.columns=DaStDeWeOIOLat | |
+report.vmware6.DaStDeWeOIOLat.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStDeWeOIOLat.type=vmware6DaSt | |
+report.vmware6.DaStDeWeOIOLat.command=--title="VMware6 datastore.datastoreWriteOIO.latest {vmware6DaStName}" \ | |
+--vertical-label="DaStDeWeOIOLat" \ | |
+DEF:xxx={rrd1}:DaStDeWeOIOLat:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStDeWeOIOLat" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStSiocAeTePeAvg.name=vmware6.datastore.siocActiveTimePercentage.average | |
+report.vmware6.DaStSiocAeTePeAvg.columns=DaStSiocAeTePeAvg | |
+report.vmware6.DaStSiocAeTePeAvg.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStSiocAeTePeAvg.type=vmware6DaSt | |
+report.vmware6.DaStSiocAeTePeAvg.command=--title="VMware6 datastore.siocActiveTimePercentage.average {vmware6DaStName}" \ | |
+--vertical-label="DaStSiocAeTePeAvg" \ | |
+DEF:xxx={rrd1}:DaStSiocAeTePeAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStSiocAeTePeAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
+ | |
+report.vmware6.DaStSizeNdDeLyAvg.name=vmware6.datastore.sizeNormalizedDatastoreLatency.average | |
+report.vmware6.DaStSizeNdDeLyAvg.columns=DaStSizeNdDeLyAvg | |
+report.vmware6.DaStSizeNdDeLyAvg.propertiesValues=vmware6DaStName | |
+report.vmware6.DaStSizeNdDeLyAvg.type=vmware6DaSt | |
+report.vmware6.DaStSizeNdDeLyAvg.command=--title="VMware6 datastore.sizeNormalizedDatastoreLatency.average {vmware6DaStName}" \ | |
+--vertical-label="DaStSizeNdDeLyAvg" \ | |
+DEF:xxx={rrd1}:DaStSizeNdDeLyAvg:AVERAGE \ | |
+LINE2:xxx#0000ff:"DaStSizeNdDeLyAvg" \ | |
+GPRINT:xxx:AVERAGE:"Avg \\: %8.2lf %s" \ | |
+GPRINT:xxx:MIN:"Min \\: %8.2lf %s" \ | |
+GPRINT:xxx:MAX:"Max \\: %8.2lf %s\\n" | |
diff --git a/snmp-graph.properties.d/xmp-graph.properties b/snmp-graph.properties.d/xmp-graph.properties | |
index 9e06b76..1be5d8e 100644 | |
--- a/snmp-graph.properties.d/xmp-graph.properties | |
+++ b/snmp-graph.properties.d/xmp-graph.properties | |
@@ -116,7 +116,7 @@ report.xmp.diskstats.command=--title="Disk Reads/Writes" \ | |
LINE2:w#00ff00:"Disk Writes" \ | |
GPRINT:w:AVERAGE:"Avg \\: %10.2lf %s" \ | |
GPRINT:w:MIN:"Min \\: %10.2lf %s" \ | |
- GPRINT:w:MAX:"Max \\: %10.2lf %s\\n" | |
+ GPRINT:w:MAX:"Max \\: %10.2lf %s\\n" | |
report.xmp.diskkb.name=Disk KB I/O | |
report.xmp.diskkb.columns=diskReadKB, diskWriteKB | |
@@ -131,7 +131,7 @@ report.xmp.diskkb.command=--title="Disk KB I/O" \ | |
LINE2:w#00ff00:"Disk Write (KB)" \ | |
GPRINT:w:AVERAGE:"Avg \\: %10.2lf %s" \ | |
GPRINT:w:MIN:"Min \\: %10.2lf %s" \ | |
- GPRINT:w:MAX:"Max \\: %10.2lf %s\\n" | |
+ GPRINT:w:MAX:"Max \\: %10.2lf %s\\n" | |
report.xmp.xmpdsz.name=Process Sizes | |
report.xmp.xmpdsz.suppress=xmp.xmpdrss | |
@@ -202,7 +202,7 @@ report.xmp.cpu.command=--title="Process CPU Time" \ | |
LINE2:cputime#0000ff:"CPU Time" \ | |
GPRINT:cputime:AVERAGE:"Avg \\: %5.2lf %s" \ | |
GPRINT:cputime:MIN:"Min \\: %5.2lf %s" \ | |
- GPRINT:cputime:MAX:"Max \\: %5.2lf %s\\n" | |
+ GPRINT:cputime:MAX:"Max \\: %5.2lf %s\\n" | |
report.xmp.bytes.name=Process I/O Bytes | |
report.xmp.bytes.columns=processReadBytes, processWriteBytes | |
diff --git a/startup.properties b/startup.properties | |
index 5e1ecc2..a0f9eae 100644 | |
--- a/startup.properties | |
+++ b/startup.properties | |
@@ -28,10 +28,11 @@ | |
# | |
# Startup core services like logging | |
# | |
-org/ops4j/pax/url/pax-url-aether/2.1.0/pax-url-aether-2.1.0.jar=5 | |
-org/ops4j/pax/url/pax-url-wrap/2.1.0/pax-url-wrap-2.1.0-uber.jar=5 | |
-org/ops4j/pax/logging/pax-logging-api/1.7.4/pax-logging-api-1.7.4.jar=8 | |
-org/ops4j/pax/logging/pax-logging-service/1.7.4/pax-logging-service-1.7.4.jar=8 | |
-org/apache/felix/org.apache.felix.configadmin/1.8.0/org.apache.felix.configadmin-1.8.0.jar=10 | |
-org/apache/felix/org.apache.felix.fileinstall/3.4.2/org.apache.felix.fileinstall-3.4.2.jar=11 | |
-org/apache/karaf/features/org.apache.karaf.features.core/2.4.0/org.apache.karaf.features.core-2.4.0.jar=15 | |
+org/ops4j/pax/url/pax-url-aether/2.4.1/pax-url-aether-2.4.1.jar=5 | |
+org/ops4j/pax/url/pax-url-wrap/2.4.1/pax-url-wrap-2.4.1-uber.jar=5 | |
+org/ops4j/pax/logging/pax-logging-api/1.8.3/pax-logging-api-1.8.3.jar=8 | |
+org/ops4j/pax/logging/pax-logging-service/1.8.3/pax-logging-service-1.8.3.jar=8 | |
+org/apache/felix/org.apache.felix.configadmin/1.8.4/org.apache.felix.configadmin-1.8.4.jar=10 | |
+org/apache/felix/org.apache.felix.fileinstall/3.5.0/org.apache.felix.fileinstall-3.5.0.jar=11 | |
+org/apache/karaf/features/org.apache.karaf.features.core/2.4.3/org.apache.karaf.features.core-2.4.3.jar=15 | |
+ | |
diff --git a/statsd-configuration.xml b/statsd-configuration.xml | |
index a05a8b8..482aa6f 100644 | |
--- a/statsd-configuration.xml | |
+++ b/statsd-configuration.xml | |
@@ -1,7 +1,7 @@ | |
<?xml version="1.0"?> | |
<statistics-daemon-configuration | |
- xmlns:this="http://www.opennms.org/xsd/config/statsd" | |
- xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" | |
+ xmlns:this="http://www.opennms.org/xsd/config/statsd" | |
+ xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" | |
xsi:schemaLocation="http://www.opennms.org/xsd/config/statsd http://www.opennms.org/xsd/config/statistics-daemon-configuration.xsd "> | |
<!-- | |
@@ -26,7 +26,19 @@ | |
<parameter key="attributeMatch" value="ifInOctets"/> | |
</packageReport> | |
</package> | |
- | |
+ | |
+ <package name="IOWAITReports"> | |
+ <packageReport name="TopN_IOWait" description="Top 20 iowait across all nodes" | |
+ schedule="0 17 10 * * ?" retainInterval="2592000000" | |
+ status="off"> | |
+ <parameter key="count" value="20"/> | |
+ <parameter key="consolidationFunction" value="AVERAGE"/> | |
+ <parameter key="relativeTime" value="YESTERDAY"/> | |
+ <parameter key="resourceTypeMatch" value="nodeSnmp"/> | |
+ <parameter key="attributeMatch" value="CpuRawWait"/> | |
+ </packageReport> | |
+ </package> | |
+ | |
<package name="ResponseTimeReports"> | |
<packageReport name="Top10_Response_Weekly" description="Weekly Top 10 responses across all nodes" | |
schedule="0 0 0 ? * MON" retainInterval="2592000000" | |
@@ -37,7 +49,7 @@ | |
<parameter key="resourceTypeMatch" value="responseTime"/> | |
<parameter key="attributeMatch" value="icmp"/> | |
</packageReport> | |
- | |
+ | |
<packageReport name="Top10_Response_This_Month" description="This Month Top 10 responses across all nodes" | |
schedule="0 0 0 L * ?" retainInterval="2592000000" | |
status="off"> | |
@@ -47,7 +59,7 @@ | |
<parameter key="resourceTypeMatch" value="responseTime"/> | |
<parameter key="attributeMatch" value="icmp"/> | |
</packageReport> | |
- | |
+ | |
<packageReport name="Top10_Response_Last_Month" description="Last Month Top 10 responses across all nodes" | |
schedule="0 0 0 1 * ?" retainInterval="2592000000" | |
status="off"> | |
@@ -57,7 +69,7 @@ | |
<parameter key="resourceTypeMatch" value="responseTime"/> | |
<parameter key="attributeMatch" value="icmp"/> | |
</packageReport> | |
- | |
+ | |
<packageReport name="Top10_Response_This_Year" description="This Year Top 10 responses across all nodes" | |
schedule="0 0 0 1 * ?" retainInterval="2592000000" | |
status="off"> | |
@@ -68,9 +80,10 @@ | |
<parameter key="attributeMatch" value="icmp"/> | |
</packageReport> | |
</package> | |
- | |
- | |
+ | |
+ | |
<report name="TopN_InOctets" class-name="org.opennms.netmgt.dao.support.TopNAttributeStatisticVisitor"/> | |
+ <report name="TopN_IOWait" class-name="org.opennms.netmgt.dao.support.TopNAttributeStatisticVisitor"/> | |
<report name="Top10_Response_Weekly" class-name="org.opennms.netmgt.dao.support.TopNAttributeStatisticVisitor"/> | |
<report name="Top10_Response_This_Month" class-name="org.opennms.netmgt.dao.support.TopNAttributeStatisticVisitor"/> | |
<report name="Top10_Response_Last_Month" class-name="org.opennms.netmgt.dao.support.TopNAttributeStatisticVisitor"/> | |
diff --git a/trend-configuration.xml b/trend-configuration.xml | |
new file mode 100644 | |
index 0000000..b0471df | |
--- /dev/null | |
+++ b/trend-configuration.xml | |
@@ -0,0 +1,233 @@ | |
+<?xml version="1.0"?> | |
+<trend-configuration xmlns="http://xmlns.opennms.org/xsd/config/trend"> | |
+ | |
+ <trend-definition name="nodes"> | |
+ <title>Nodes</title> | |
+ <subtitle>with Outages</subtitle> | |
+ <visible>true</visible> | |
+ <icon>glyphicon-fire</icon> | |
+ <trend-attributes> | |
+ <trend-attribute key="sparkWidth" value="100%"/> | |
+ <trend-attribute key="sparkHeight" value="35"/> | |
+ <trend-attribute key="sparkChartRangeMin" value="0"/> | |
+ <trend-attribute key="sparkLineColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkLineWidth" value="1.5"/> | |
+ <trend-attribute key="sparkFillColor" value="#e9b96e"/> | |
+ <trend-attribute key="sparkSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkMinSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkMaxSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkSpotRadius" value="3"/> | |
+ <trend-attribute key="sparkHighlightSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkHighlightLineColor" value="#8f5902"/> | |
+ </trend-attributes> | |
+ <descriptionLink>outage/list.htm?outtype=current</descriptionLink> | |
+ <description>${intValue[23]} Nodes with Outage(s)</description> | |
+ <query> | |
+ <![CDATA[ | |
+ select ( | |
+ select count(distinct nodeid) from outages o, events e where e.eventid = o.svclosteventid and iflostservice < E and (ifregainedservice is null or ifregainedservice > E) | |
+ ) from ( | |
+ select now() - interval '1 hour' * (O + 1) AS S, now() - interval '1 hour' * O as E from generate_series(0, 23) as O | |
+ ) I order by S; | |
+ ]]> | |
+ </query> | |
+ </trend-definition> | |
+ | |
+ <trend-definition name="severity"> | |
+ <title>Severity</title> | |
+ <subtitle>Distribution</subtitle> | |
+ <visible>true</visible> | |
+ <icon>glyphicon-dashboard</icon> | |
+ <trend-attributes> | |
+ <trend-attribute key="sparkType" value="pie"/> | |
+ <trend-attribute key="sparkHeight" value="35"/> | |
+ <trend-attribute key="sparkSliceColors" value="[#CC0000,#FF3300,#FF9900,#FFCC00,#999900,#336600,#999]"/> | |
+ </trend-attributes> | |
+ <descriptionLink>alarm/list.htm</descriptionLink> | |
+ <description>Go to Alarms Page</description> | |
+ <query> | |
+ <![CDATA[ | |
+ select ( | |
+ select count(*) from alarms where firsteventtime > now() - interval '24 hours' and severity = S | |
+ ) from ( | |
+ select S from generate_series(1,7) as S | |
+ ) I order by S desc; | |
+ ]]> | |
+ </query> | |
+ </trend-definition> | |
+ | |
+ <trend-definition name="alarms-new"> | |
+ <title>Alarms</title> | |
+ <subtitle>Occurence</subtitle> | |
+ <visible>true</visible> | |
+ <icon>glyphicon-bell</icon> | |
+ <trend-attributes> | |
+ <trend-attribute key="sparkWidth" value="100%"/> | |
+ <trend-attribute key="sparkType" value="bar"/> | |
+ <trend-attribute key="sparkBarColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkHeight" value="35"/> | |
+ <trend-attribute key="sparkBarWidth" value="3"/> | |
+ <trend-attribute key="sparkBarSpacing" value="2"/> | |
+ </trend-attributes> | |
+ <descriptionLink>alarm/list.htm</descriptionLink> | |
+ <description>${intValue[23]} New Alarm(s)</description> | |
+ <query> | |
+ <![CDATA[ | |
+ select ( | |
+ select count(*) from alarms where firsteventtime > S and firsteventtime <= E | |
+ ) from ( | |
+ select now() - interval '1 hour' * (O + 1) AS S, now() - interval '1 hour' * O as E from generate_series(0, 23) as O | |
+ ) I order by S; | |
+ ]]> | |
+ </query> | |
+ </trend-definition> | |
+ | |
+ <trend-definition name="alarms-total"> | |
+ <title>Alarms</title> | |
+ <subtitle>Unacknowledged</subtitle> | |
+ <visible>true</visible> | |
+ <icon>glyphicon-bell</icon> | |
+ <trend-attributes> | |
+ <trend-attribute key="sparkWidth" value="100%"/> | |
+ <trend-attribute key="sparkHeight" value="40"/> | |
+ <trend-attribute key="sparkChartRangeMin" value="0"/> | |
+ <trend-attribute key="sparkLineColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkLineWidth" value="1.5"/> | |
+ <trend-attribute key="sparkFillColor" value="#e9b96e"/> | |
+ <trend-attribute key="sparkSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkMinSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkMaxSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkSpotRadius" value="3"/> | |
+ <trend-attribute key="sparkHighlightSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkHighlightLineColor" value="#8f5902"/> | |
+ </trend-attributes> | |
+ <descriptionLink>alarm/list.htm</descriptionLink> | |
+ <description>${intValue[23]} Unacknowledged Alarm(s)</description> | |
+ <query> | |
+ <![CDATA[ | |
+ select ( | |
+ select count(*) from alarms where (alarmacktime is null or alarmacktime > E) and firsteventtime <= E | |
+ ) from ( | |
+ select now() - interval '1 hour' * (O + 1) AS S, now() - interval '1 hour' * O as E from generate_series(0, 23) as O | |
+ ) I order by S; | |
+ ]]> | |
+ </query> | |
+ </trend-definition> | |
+ | |
+ <trend-definition name="outages-new"> | |
+ <title>Outages</title> | |
+ <subtitle>Occurence</subtitle> | |
+ <visible>true</visible> | |
+ <icon>glyphicon-exclamation-sign</icon> | |
+ <trend-attributes> | |
+ <trend-attribute key="sparkWidth" value="100%"/> | |
+ <trend-attribute key="sparkType" value="bar"/> | |
+ <trend-attribute key="sparkBarColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkHeight" value="40"/> | |
+ <trend-attribute key="sparkBarWidth" value="3"/> | |
+ <trend-attribute key="sparkBarSpacing" value="2"/> | |
+ </trend-attributes> | |
+ <descriptionLink>outage/list.htm?outtype=both</descriptionLink> | |
+ <description>${intValue[23]} New Outage(s)</description> | |
+ <query> | |
+ <![CDATA[ | |
+ select ( | |
+ select count(*) from outages where iflostservice > S and iflostservice <= E | |
+ ) from ( | |
+ select now() - interval '1 hour' * (O + 1) AS S, now() - interval '1 hour' * O as E from generate_series(0, 23) as O | |
+ ) I order by S; | |
+ ]]> | |
+ </query> | |
+ </trend-definition> | |
+ | |
+ <trend-definition name="outages-total"> | |
+ <title>Outages</title> | |
+ <subtitle>Current</subtitle> | |
+ <visible>true</visible> | |
+ <icon>glyphicon-exclamation-sign</icon> | |
+ <trend-attributes> | |
+ <trend-attribute key="sparkWidth" value="100%"/> | |
+ <trend-attribute key="sparkHeight" value="40"/> | |
+ <trend-attribute key="sparkChartRangeMin" value="0"/> | |
+ <trend-attribute key="sparkLineColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkLineWidth" value="1.5"/> | |
+ <trend-attribute key="sparkFillColor" value="#e9b96e"/> | |
+ <trend-attribute key="sparkSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkMinSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkMaxSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkSpotRadius" value="3"/> | |
+ <trend-attribute key="sparkHighlightSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkHighlightLineColor" value="#8f5902"/> | |
+ </trend-attributes> | |
+ <descriptionLink>outage/list.htm?outtype=current</descriptionLink> | |
+ <description>${intValue[23]} Current Outage(s)</description> | |
+ <query> | |
+ <![CDATA[ | |
+ select ( | |
+ select count(*) from outages where iflostservice < E and (ifregainedservice is null or ifregainedservice > E) | |
+ ) from ( | |
+ select now() - interval '1 hour' * (O + 1) AS S, now() - interval '1 hour' * O as E from generate_series(0, 23) as O | |
+ ) I order by S; | |
+ ]]> | |
+ </query> | |
+ </trend-definition> | |
+ | |
+ <trend-definition name="bsm-new"> | |
+ <title>Business Services</title> | |
+ <subtitle>Problem Occurence</subtitle> | |
+ <visible>true</visible> | |
+ <icon>glyphicon-tasks</icon> | |
+ <trend-attributes> | |
+ <trend-attribute key="sparkWidth" value="100%"/> | |
+ <trend-attribute key="sparkType" value="bar"/> | |
+ <trend-attribute key="sparkBarColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkHeight" value="40"/> | |
+ <trend-attribute key="sparkBarWidth" value="3"/> | |
+ <trend-attribute key="sparkBarSpacing" value="2"/> | |
+ </trend-attributes> | |
+ <descriptionLink>topology?szl=1amp;layout=Hierarchy+Layout&provider=Business+Services</descriptionLink> | |
+ <description>${intValue[23]} New Problem(s)</description> | |
+ <query> | |
+ <![CDATA[ | |
+ select ( | |
+ select count(*) from alarms where firsteventtime > S and firsteventtime <= E and eventuei='uei.opennms.org/bsm/serviceProblem' | |
+ ) from ( | |
+ select now() - interval '1 hour' * (O + 1) AS S, now() - interval '1 hour' * O as E from generate_series(0, 23) as O | |
+ ) I order by S; | |
+ ]]> | |
+ </query> | |
+ </trend-definition> | |
+ | |
+ <trend-definition name="bsm-total"> | |
+ <title>Business Services</title> | |
+ <subtitle>Current Problems</subtitle> | |
+ <visible>true</visible> | |
+ <icon>glyphicon-tasks</icon> | |
+ <trend-attributes> | |
+ <trend-attribute key="sparkWidth" value="100%"/> | |
+ <trend-attribute key="sparkHeight" value="40"/> | |
+ <trend-attribute key="sparkChartRangeMin" value="0"/> | |
+ <trend-attribute key="sparkLineColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkLineWidth" value="1.5"/> | |
+ <trend-attribute key="sparkFillColor" value="#e9b96e"/> | |
+ <trend-attribute key="sparkSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkMinSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkMaxSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkSpotRadius" value="3"/> | |
+ <trend-attribute key="sparkHighlightSpotColor" value="#8f5902"/> | |
+ <trend-attribute key="sparkHighlightLineColor" value="#8f5902"/> | |
+ </trend-attributes> | |
+ <descriptionLink>topology?szl=1amp;layout=Hierarchy+Layout&provider=Business+Services</descriptionLink> | |
+ <description>${intValue[23]} Current Problem(s)</description> | |
+ <query> | |
+ <![CDATA[ | |
+ select ( | |
+ select count(*) from alarms where (alarmacktime is null or alarmacktime > E) and firsteventtime <= E and eventuei='uei.opennms.org/bsm/serviceProblem' | |
+ ) from ( | |
+ select now() - interval '1 hour' * (O + 1) AS S, now() - interval '1 hour' * O as E from generate_series(0, 23) as O | |
+ ) I order by S; | |
+ ]]> | |
+ </query> | |
+ </trend-definition> | |
+ | |
+</trend-configuration> | |
diff --git a/users.xml b/users.xml | |
index d5ada64..c919ffd 100644 | |
--- a/users.xml | |
+++ b/users.xml | |
@@ -11,6 +11,14 @@ | |
<full-name>Administrator</full-name> | |
<user-comments>Default administrator, do not delete</user-comments> | |
<password>21232F297A57A5A743894A0E4A801FC3</password> | |
+ <role>ROLE_ADMIN</role> | |
+ </user> | |
+ <user> | |
+ <user-id>rtc</user-id> | |
+ <full-name>RTC</full-name> | |
+ <user-comments>RTC user, do not delete</user-comments> | |
+ <password>68154466F81BFB532CD70F8C71426356</password> | |
+ <role>ROLE_RTC</role> | |
</user> | |
</users> | |
</userinfo> | |
diff --git a/vmware-datacollection-config.xml b/vmware-datacollection-config.xml | |
index 0b2cc03..6a93515 100644 | |
--- a/vmware-datacollection-config.xml | |
+++ b/vmware-datacollection-config.xml | |
@@ -744,4 +744,383 @@ | |
</vmware-group> | |
</vmware-groups> | |
</vmware-collection> | |
+ | |
+ <!-- | |
+ Configuration file generated for: | |
+ | |
+ Full name.......: VMware vCenter Server 6.0.0 build-3339084 | |
+ API type........: VirtualCenter | |
+ API version.....: 6.0 | |
+ Product name....: VMware VirtualCenter Server | |
+ Product version.: 6.0 | |
+ OS type.........: linux-x64 | |
+ --> | |
+ | |
+ <vmware-collection name="default-VirtualMachine6"> | |
+ <rrd step="300"> | |
+ <rra>RRA:AVERAGE:0.5:1:2016</rra> | |
+ <rra>RRA:AVERAGE:0.5:12:1488</rra> | |
+ <rra>RRA:AVERAGE:0.5:288:366</rra> | |
+ <rra>RRA:MAX:0.5:288:366</rra> | |
+ <rra>RRA:MIN:0.5:288:366</rra> | |
+ </rrd> | |
+ <vmware-groups> | |
+ <vmware-group name="vmware6VrtDisk" resourceType="vmware6VrtDisk"> | |
+ <attrib name="virtualDisk.largeSeeks.latest" alias="VrtDiskLeSsLat" type="Gauge"/> | |
+ <attrib name="virtualDisk.mediumSeeks.latest" alias="VrtDiskMmSsLat" type="Gauge"/> | |
+ <attrib name="virtualDisk.numberReadAveraged.average" alias="VrtDiskNrRdAdAvg" type="Gauge"/> | |
+ <attrib name="virtualDisk.numberWriteAveraged.average" alias="VrtDiskNrWeAdAvg" type="Gauge"/> | |
+ <attrib name="virtualDisk.read.average" alias="VrtDiskRdAvg" type="Gauge"/> | |
+ <attrib name="virtualDisk.readIOSize.latest" alias="VrtDiskRdIOSizeLat" type="Gauge"/> | |
+ <attrib name="virtualDisk.readLoadMetric.latest" alias="VrtDiskRdLdMcLat" type="Gauge"/> | |
+ <attrib name="virtualDisk.readLatencyUS.latest" alias="VrtDiskRdLyUSLat" type="Gauge"/> | |
+ <attrib name="virtualDisk.readOIO.latest" alias="VrtDiskRdOIOLat" type="Gauge"/> | |
+ <attrib name="virtualDisk.smallSeeks.latest" alias="VrtDiskSlSsLat" type="Gauge"/> | |
+ <attrib name="virtualDisk.totalReadLatency.average" alias="VrtDiskTlRdLyAvg" type="Gauge"/> | |
+ <attrib name="virtualDisk.totalWriteLatency.average" alias="VrtDiskTlWeLyAvg" type="Gauge"/> | |
+ <attrib name="virtualDisk.write.average" alias="VrtDiskWeAvg" type="Gauge"/> | |
+ <attrib name="virtualDisk.writeIOSize.latest" alias="VrtDiskWeIOSizeLat" type="Gauge"/> | |
+ <attrib name="virtualDisk.writeLoadMetric.latest" alias="VrtDiskWeLdMcLat" type="Gauge"/> | |
+ <attrib name="virtualDisk.writeLatencyUS.latest" alias="VrtDiskWeLyUSLat" type="Gauge"/> | |
+ <attrib name="virtualDisk.writeOIO.latest" alias="VrtDiskWeOIOLat" type="Gauge"/> | |
+ </vmware-group> | |
+ <vmware-group name="vmware6Node" resourceType="Node"> | |
+ <attrib name="cpu.demand.average" alias="CpuDemandAvg" type="Gauge"/> | |
+ <attrib name="cpu.demandEntitlementRatio.latest" alias="CpuDmdEntRatioLat" type="Gauge"/> | |
+ <attrib name="cpu.entitlement.latest" alias="CpuEntitlementLat" type="Gauge"/> | |
+ <attrib name="cpu.latency.average" alias="CpuLyAvg" type="Gauge"/> | |
+ <attrib name="cpu.overlap.summation" alias="CpuOverlapSum" type="Gauge"/> | |
+ <attrib name="cpu.readiness.average" alias="CpuRdinessAvg" type="Gauge"/> | |
+ <attrib name="cpu.swapwait.summation" alias="CpuSpwaitSum" type="Gauge"/> | |
+ <attrib name="cpu.usage.average" alias="CpuUsageAvg" type="Gauge"/> | |
+ <attrib name="cpu.usagemhz.average" alias="CpuUsagemhzAvg" type="Gauge"/> | |
+ <attrib name="cpu.used.summation" alias="CpuUsedSum" type="Gauge"/> | |
+ <attrib name="cpu.wait.summation" alias="CpuWaitSum" type="Gauge"/> | |
+ <attrib name="datastore.maxTotalLatency.latest" alias="DaStMaxTlLyLat" type="Gauge"/> | |
+ <attrib name="disk.maxTotalLatency.latest" alias="DiskMaxTlLyLat" type="Gauge"/> | |
+ <attrib name="disk.read.average" alias="DiskRdAvg" type="Gauge"/> | |
+ <attrib name="disk.usage.average" alias="DiskUsageAvg" type="Gauge"/> | |
+ <attrib name="disk.write.average" alias="DiskWeAvg" type="Gauge"/> | |
+ <attrib name="mem.active.average" alias="MemAeAvg" type="Gauge"/> | |
+ <attrib name="mem.activewrite.average" alias="MemAeWeAvg" type="Gauge"/> | |
+ <attrib name="mem.consumed.average" alias="MemCdAvg" type="Gauge"/> | |
+ <attrib name="mem.compressionRate.average" alias="MemCnReAvg" type="Gauge"/> | |
+ <attrib name="mem.compressed.average" alias="MemCompressedAvg" type="Gauge"/> | |
+ <attrib name="mem.decompressionRate.average" alias="MemDnReAvg" type="Gauge"/> | |
+ <attrib name="mem.entitlement.average" alias="MemEntitlementAvg" type="Gauge"/> | |
+ <attrib name="mem.granted.average" alias="MemGrantedAvg" type="Gauge"/> | |
+ <attrib name="mem.llSwapInRate.average" alias="MemLlSpInReAvg" type="Gauge"/> | |
+ <attrib name="mem.llSwapOutRate.average" alias="MemLlSpOutReAvg" type="Gauge"/> | |
+ <attrib name="mem.llSwapUsed.average" alias="MemLlSpUsedAvg" type="Gauge"/> | |
+ <attrib name="mem.latency.average" alias="MemLyAvg" type="Gauge"/> | |
+ <attrib name="mem.overhead.average" alias="MemOdAvg" type="Gauge"/> | |
+ <attrib name="mem.overheadMax.average" alias="MemOdMaxAvg" type="Gauge"/> | |
+ <attrib name="mem.overheadTouched.average" alias="MemOdTdAvg" type="Gauge"/> | |
+ <attrib name="mem.shared.average" alias="MemSharedAvg" type="Gauge"/> | |
+ <attrib name="mem.swaptarget.average" alias="MemSpTtAvg" type="Gauge"/> | |
+ <attrib name="mem.swapin.average" alias="MemSpinAvg" type="Gauge"/> | |
+ <attrib name="mem.swapinRate.average" alias="MemSpinReAvg" type="Gauge"/> | |
+ <attrib name="mem.swapout.average" alias="MemSpoutAvg" type="Gauge"/> | |
+ <attrib name="mem.swapoutRate.average" alias="MemSpoutReAvg" type="Gauge"/> | |
+ <attrib name="mem.swapped.average" alias="MemSppedAvg" type="Gauge"/> | |
+ <attrib name="mem.usage.average" alias="MemUsageAvg" type="Gauge"/> | |
+ <attrib name="mem.vmmemctl.average" alias="MemVmmemctlAvg" type="Gauge"/> | |
+ <attrib name="mem.vmmemctltarget.average" alias="MemVmmemctlTtAvg" type="Gauge"/> | |
+ <attrib name="mem.zero.average" alias="MemZeroAvg" type="Gauge"/> | |
+ <attrib name="mem.zipSaved.latest" alias="MemZipSavedLat" type="Gauge"/> | |
+ <attrib name="mem.zipped.latest" alias="MemZippedLat" type="Gauge"/> | |
+ <attrib name="net.broadcastRx.summation" alias="NetBroadcastRxSum" type="Gauge"/> | |
+ <attrib name="net.broadcastTx.summation" alias="NetBroadcastTxSum" type="Gauge"/> | |
+ <attrib name="net.droppedRx.summation" alias="NetDroppedRxSum" type="Gauge"/> | |
+ <attrib name="net.droppedTx.summation" alias="NetDroppedTxSum" type="Gauge"/> | |
+ <attrib name="net.packetsTx.summation" alias="NetPacketsTxSum" type="Gauge"/> | |
+ <attrib name="net.received.average" alias="NetReceivedAvg" type="Gauge"/> | |
+ <attrib name="net.transmitted.average" alias="NetTransmittedAvg" type="Gauge"/> | |
+ <attrib name="power.energy.summation" alias="PowerEnergySum" type="Gauge"/> | |
+ <attrib name="power.power.average" alias="PowerPowerAvg" type="Gauge"/> | |
+ <attrib name="rescpu.actav15.latest" alias="ResCpuActav15Lat" type="Gauge"/> | |
+ <attrib name="rescpu.actav1.latest" alias="ResCpuActav1Lat" type="Gauge"/> | |
+ <attrib name="rescpu.actav5.latest" alias="ResCpuActav5Lat" type="Gauge"/> | |
+ <attrib name="rescpu.actpk15.latest" alias="ResCpuActpk15Lat" type="Gauge"/> | |
+ <attrib name="rescpu.actpk1.latest" alias="ResCpuActpk1Lat" type="Gauge"/> | |
+ <attrib name="rescpu.actpk5.latest" alias="ResCpuActpk5Lat" type="Gauge"/> | |
+ <attrib name="rescpu.maxLimited15.latest" alias="ResCpuMaxLd15Lat" type="Gauge"/> | |
+ <attrib name="rescpu.maxLimited1.latest" alias="ResCpuMaxLd1Lat" type="Gauge"/> | |
+ <attrib name="rescpu.maxLimited5.latest" alias="ResCpuMaxLd5Lat" type="Gauge"/> | |
+ <attrib name="rescpu.runav15.latest" alias="ResCpuRunav15Lat" type="Gauge"/> | |
+ <attrib name="rescpu.runav1.latest" alias="ResCpuRunav1Lat" type="Gauge"/> | |
+ <attrib name="rescpu.runav5.latest" alias="ResCpuRunav5Lat" type="Gauge"/> | |
+ <attrib name="rescpu.runpk15.latest" alias="ResCpuRunpk15Lat" type="Gauge"/> | |
+ <attrib name="rescpu.runpk1.latest" alias="ResCpuRunpk1Lat" type="Gauge"/> | |
+ <attrib name="rescpu.runpk5.latest" alias="ResCpuRunpk5Lat" type="Gauge"/> | |
+ <attrib name="rescpu.sampleCount.latest" alias="ResCpuSeCtLat" type="Gauge"/> | |
+ <attrib name="rescpu.samplePeriod.latest" alias="ResCpuSePeriodLat" type="Gauge"/> | |
+ <attrib name="sys.heartbeat.latest" alias="SysHeartbeatLat" type="Gauge"/> | |
+ <attrib name="sys.osUptime.latest" alias="SysOsUpTeLat" type="Gauge"/> | |
+ <attrib name="sys.uptime.latest" alias="SysUpTeLat" type="Gauge"/> | |
+ <attrib name="virtualDisk.read.average" alias="VrtDiskRdAvg" type="Gauge"/> | |
+ <attrib name="virtualDisk.write.average" alias="VrtDiskWeAvg" type="Gauge"/> | |
+ </vmware-group> | |
+ <vmware-group name="vmware6Cpu" resourceType="vmware6Cpu"> | |
+ <attrib name="cpu.costop.summation" alias="CpuCostopSum" type="Gauge"/> | |
+ <attrib name="cpu.idle.summation" alias="CpuIdleSum" type="Gauge"/> | |
+ <attrib name="cpu.maxlimited.summation" alias="CpuMaxLdSum" type="Gauge"/> | |
+ <attrib name="cpu.overlap.summation" alias="CpuOverlapSum" type="Gauge"/> | |
+ <attrib name="cpu.readiness.average" alias="CpuRdinessAvg" type="Gauge"/> | |
+ <attrib name="cpu.ready.summation" alias="CpuRdySum" type="Gauge"/> | |
+ <attrib name="cpu.run.summation" alias="CpuRunSum" type="Gauge"/> | |
+ <attrib name="cpu.swapwait.summation" alias="CpuSpwaitSum" type="Gauge"/> | |
+ <attrib name="cpu.system.summation" alias="CpuSystemSum" type="Gauge"/> | |
+ <attrib name="cpu.usagemhz.average" alias="CpuUsagemhzAvg" type="Gauge"/> | |
+ <attrib name="cpu.used.summation" alias="CpuUsedSum" type="Gauge"/> | |
+ <attrib name="cpu.wait.summation" alias="CpuWaitSum" type="Gauge"/> | |
+ </vmware-group> | |
+ <vmware-group name="vmware6Net" resourceType="vmware6Net"> | |
+ <attrib name="net.broadcastRx.summation" alias="NetBroadcastRxSum" type="Gauge"/> | |
+ <attrib name="net.broadcastTx.summation" alias="NetBroadcastTxSum" type="Gauge"/> | |
+ <attrib name="net.bytesRx.average" alias="NetBytesRxAvg" type="Gauge"/> | |
+ <attrib name="net.bytesTx.average" alias="NetBytesTxAvg" type="Gauge"/> | |
+ <attrib name="net.droppedRx.summation" alias="NetDroppedRxSum" type="Gauge"/> | |
+ <attrib name="net.droppedTx.summation" alias="NetDroppedTxSum" type="Gauge"/> | |
+ <attrib name="net.multicastRx.summation" alias="NetMulticastRxSum" type="Gauge"/> | |
+ <attrib name="net.multicastTx.summation" alias="NetMulticastTxSum" type="Gauge"/> | |
+ <attrib name="net.packetsRx.summation" alias="NetPacketsRxSum" type="Gauge"/> | |
+ <attrib name="net.packetsTx.summation" alias="NetPacketsTxSum" type="Gauge"/> | |
+ <attrib name="net.received.average" alias="NetReceivedAvg" type="Gauge"/> | |
+ <attrib name="net.transmitted.average" alias="NetTransmittedAvg" type="Gauge"/> | |
+ <attrib name="net.usage.average" alias="NetUsageAvg" type="Gauge"/> | |
+ </vmware-group> | |
+ <vmware-group name="vmware6Disk" resourceType="vmware6Disk"> | |
+ <attrib name="disk.busResets.summation" alias="DiskBusResetsSum" type="Gauge"/> | |
+ <attrib name="disk.commandsAveraged.average" alias="DiskCsAdAvg" type="Gauge"/> | |
+ <attrib name="disk.commandsAborted.summation" alias="DiskCsAdSum" type="Gauge"/> | |
+ <attrib name="disk.commands.summation" alias="DiskCsSum" type="Gauge"/> | |
+ <attrib name="disk.numberReadAveraged.average" alias="DiskNrRdAdAvg" type="Gauge"/> | |
+ <attrib name="disk.numberRead.summation" alias="DiskNrRdSum" type="Gauge"/> | |
+ <attrib name="disk.numberWriteAveraged.average" alias="DiskNrWeAdAvg" type="Gauge"/> | |
+ <attrib name="disk.numberWrite.summation" alias="DiskNrWeSum" type="Gauge"/> | |
+ <attrib name="disk.read.average" alias="DiskRdAvg" type="Gauge"/> | |
+ <attrib name="disk.write.average" alias="DiskWeAvg" type="Gauge"/> | |
+ </vmware-group> | |
+ <vmware-group name="vmware6DaSt" resourceType="vmware6DaSt"> | |
+ <attrib name="datastore.numberReadAveraged.average" alias="DaStNrRdAdAvg" type="Gauge"/> | |
+ <attrib name="datastore.numberWriteAveraged.average" alias="DaStNrWeAdAvg" type="Gauge"/> | |
+ <attrib name="datastore.read.average" alias="DaStRdAvg" type="Gauge"/> | |
+ <attrib name="datastore.totalReadLatency.average" alias="DaStTlRdLyAvg" type="Gauge"/> | |
+ <attrib name="datastore.totalWriteLatency.average" alias="DaStTlWeLyAvg" type="Gauge"/> | |
+ <attrib name="datastore.write.average" alias="DaStWeAvg" type="Gauge"/> | |
+ </vmware-group> | |
+ </vmware-groups> | |
+ </vmware-collection> | |
+ <vmware-collection name="default-HostSystem6"> | |
+ <rrd step="300"> | |
+ <rra>RRA:AVERAGE:0.5:1:2016</rra> | |
+ <rra>RRA:AVERAGE:0.5:12:1488</rra> | |
+ <rra>RRA:AVERAGE:0.5:288:366</rra> | |
+ <rra>RRA:MAX:0.5:288:366</rra> | |
+ <rra>RRA:MIN:0.5:288:366</rra> | |
+ </rrd> | |
+ <vmware-groups> | |
+ <vmware-group name="vmware6StAdptr" resourceType="vmware6StAdptr"> | |
+ <attrib name="storageAdapter.commandsAveraged.average" alias="StAdptrCsAdAvg" type="Gauge"/> | |
+ <attrib name="storageAdapter.numberReadAveraged.average" alias="StAdptrNrRdAdAvg" type="Gauge"/> | |
+ <attrib name="storageAdapter.numberWriteAveraged.average" alias="StAdptrNrWeAdAvg" type="Gauge"/> | |
+ <attrib name="storageAdapter.read.average" alias="StAdptrRdAvg" type="Gauge"/> | |
+ <attrib name="storageAdapter.totalReadLatency.average" alias="StAdptrTlRdLyAvg" type="Gauge"/> | |
+ <attrib name="storageAdapter.totalWriteLatency.average" alias="StAdptrTlWeLyAvg" type="Gauge"/> | |
+ <attrib name="storageAdapter.write.average" alias="StAdptrWeAvg" type="Gauge"/> | |
+ </vmware-group> | |
+ <vmware-group name="vmware6StPth" resourceType="vmware6StPth"> | |
+ <attrib name="storagePath.commandsAveraged.average" alias="StPthCsAdAvg" type="Gauge"/> | |
+ <attrib name="storagePath.numberReadAveraged.average" alias="StPthNrRdAdAvg" type="Gauge"/> | |
+ <attrib name="storagePath.numberWriteAveraged.average" alias="StPthNrWeAdAvg" type="Gauge"/> | |
+ <attrib name="storagePath.read.average" alias="StPthRdAvg" type="Gauge"/> | |
+ <attrib name="storagePath.totalReadLatency.average" alias="StPthTlRdLyAvg" type="Gauge"/> | |
+ <attrib name="storagePath.totalWriteLatency.average" alias="StPthTlWeLyAvg" type="Gauge"/> | |
+ <attrib name="storagePath.write.average" alias="StPthWeAvg" type="Gauge"/> | |
+ </vmware-group> | |
+ <vmware-group name="vmware6Node" resourceType="Node"> | |
+ <attrib name="cpu.costop.summation" alias="CpuCostopSum" type="Gauge"/> | |
+ <attrib name="cpu.demand.average" alias="CpuDemandAvg" type="Gauge"/> | |
+ <attrib name="cpu.latency.average" alias="CpuLyAvg" type="Gauge"/> | |
+ <attrib name="cpu.reservedCapacity.average" alias="CpuRdCyAvg" type="Gauge"/> | |
+ <attrib name="cpu.readiness.average" alias="CpuRdinessAvg" type="Gauge"/> | |
+ <attrib name="cpu.ready.summation" alias="CpuRdySum" type="Gauge"/> | |
+ <attrib name="cpu.swapwait.summation" alias="CpuSpwaitSum" type="Gauge"/> | |
+ <attrib name="cpu.totalCapacity.average" alias="CpuTlCyAvg" type="Gauge"/> | |
+ <attrib name="cpu.usagemhz.average" alias="CpuUsagemhzAvg" type="Gauge"/> | |
+ <attrib name="cpu.wait.summation" alias="CpuWaitSum" type="Gauge"/> | |
+ <attrib name="datastore.maxTotalLatency.latest" alias="DaStMaxTlLyLat" type="Gauge"/> | |
+ <attrib name="disk.maxTotalLatency.latest" alias="DiskMaxTlLyLat" type="Gauge"/> | |
+ <attrib name="disk.usage.average" alias="DiskUsageAvg" type="Gauge"/> | |
+ <attrib name="hbr.hbrNetRx.average" alias="HbrHbrNetRxAvg" type="Gauge"/> | |
+ <attrib name="hbr.hbrNetTx.average" alias="HbrHbrNetTxAvg" type="Gauge"/> | |
+ <attrib name="hbr.hbrNumVms.average" alias="HbrHbrNumVmsAvg" type="Gauge"/> | |
+ <attrib name="mem.active.average" alias="MemAeAvg" type="Gauge"/> | |
+ <attrib name="mem.activewrite.average" alias="MemAeWeAvg" type="Gauge"/> | |
+ <attrib name="mem.consumed.average" alias="MemCdAvg" type="Gauge"/> | |
+ <attrib name="mem.compressionRate.average" alias="MemCnReAvg" type="Gauge"/> | |
+ <attrib name="mem.compressed.average" alias="MemCompressedAvg" type="Gauge"/> | |
+ <attrib name="mem.decompressionRate.average" alias="MemDnReAvg" type="Gauge"/> | |
+ <attrib name="mem.granted.average" alias="MemGrantedAvg" type="Gauge"/> | |
+ <attrib name="mem.heap.average" alias="MemHeapAvg" type="Gauge"/> | |
+ <attrib name="mem.heapfree.average" alias="MemHeapfreeAvg" type="Gauge"/> | |
+ <attrib name="mem.llSwapIn.average" alias="MemLlSpInAvg" type="Gauge"/> | |
+ <attrib name="mem.llSwapInRate.average" alias="MemLlSpInReAvg" type="Gauge"/> | |
+ <attrib name="mem.llSwapOut.average" alias="MemLlSpOutAvg" type="Gauge"/> | |
+ <attrib name="mem.llSwapOutRate.average" alias="MemLlSpOutReAvg" type="Gauge"/> | |
+ <attrib name="mem.llSwapUsed.average" alias="MemLlSpUsedAvg" type="Gauge"/> | |
+ <attrib name="mem.lowfreethreshold.average" alias="MemLowfreeTdAvg" type="Gauge"/> | |
+ <attrib name="mem.latency.average" alias="MemLyAvg" type="Gauge"/> | |
+ <attrib name="mem.overhead.average" alias="MemOdAvg" type="Gauge"/> | |
+ <attrib name="mem.reservedCapacity.average" alias="MemRdCyAvg" type="Gauge"/> | |
+ <attrib name="mem.shared.average" alias="MemSharedAvg" type="Gauge"/> | |
+ <attrib name="mem.sharedcommon.average" alias="MemSharedcommonAvg" type="Gauge"/> | |
+ <attrib name="mem.swapin.average" alias="MemSpinAvg" type="Gauge"/> | |
+ <attrib name="mem.swapinRate.average" alias="MemSpinReAvg" type="Gauge"/> | |
+ <attrib name="mem.swapout.average" alias="MemSpoutAvg" type="Gauge"/> | |
+ <attrib name="mem.swapoutRate.average" alias="MemSpoutReAvg" type="Gauge"/> | |
+ <attrib name="mem.swapused.average" alias="MemSpusedAvg" type="Gauge"/> | |
+ <attrib name="mem.state.latest" alias="MemStateLat" type="Gauge"/> | |
+ <attrib name="mem.sysUsage.average" alias="MemSysUsageAvg" type="Gauge"/> | |
+ <attrib name="mem.totalCapacity.average" alias="MemTlCyAvg" type="Gauge"/> | |
+ <attrib name="mem.unreserved.average" alias="MemUdAvg" type="Gauge"/> | |
+ <attrib name="mem.usage.average" alias="MemUsageAvg" type="Gauge"/> | |
+ <attrib name="mem.vmfs.pbc.overhead.latest" alias="MemVmfsPbcOdLat" type="Gauge"/> | |
+ <attrib name="mem.vmfs.pbc.capMissRatio.latest" alias="MemVmfsPbcCpMsRtiLt" type="Gauge"/> | |
+ <attrib name="mem.vmfs.pbc.size.latest" alias="MemVmfsPbcSizeLat" type="Gauge"/> | |
+ <attrib name="mem.vmfs.pbc.sizeMax.latest" alias="MemVmfsPbcSizMaxLat" type="Gauge"/> | |
+ <attrib name="mem.vmfs.pbc.workingSet.latest" alias="MemVmfsPbcWrkSetLat" type="Gauge"/> | |
+ <attrib name="mem.vmfs.pbc.workingSetMax.latest" alias="MemVmfsPbcWrkStMxLt" type="Gauge"/> | |
+ <attrib name="mem.vmmemctl.average" alias="MemVmmemctlAvg" type="Gauge"/> | |
+ <attrib name="mem.zero.average" alias="MemZeroAvg" type="Gauge"/> | |
+ <attrib name="net.broadcastTx.summation" alias="NetBroadcastTxSum" type="Gauge"/> | |
+ <attrib name="net.droppedRx.summation" alias="NetDroppedRxSum" type="Gauge"/> | |
+ <attrib name="net.multicastRx.summation" alias="NetMulticastRxSum" type="Gauge"/> | |
+ <attrib name="net.received.average" alias="NetReceivedAvg" type="Gauge"/> | |
+ <attrib name="power.energy.summation" alias="PowerEnergySum" type="Gauge"/> | |
+ <attrib name="power.power.average" alias="PowerPowerAvg" type="Gauge"/> | |
+ <attrib name="power.powerCap.average" alias="PowerPowerCapAvg" type="Gauge"/> | |
+ <attrib name="rescpu.actav15.latest" alias="ResCpuActav15Lat" type="Gauge"/> | |
+ <attrib name="rescpu.actav1.latest" alias="ResCpuActav1Lat" type="Gauge"/> | |
+ <attrib name="rescpu.actav5.latest" alias="ResCpuActav5Lat" type="Gauge"/> | |
+ <attrib name="rescpu.actpk15.latest" alias="ResCpuActpk15Lat" type="Gauge"/> | |
+ <attrib name="rescpu.actpk1.latest" alias="ResCpuActpk1Lat" type="Gauge"/> | |
+ <attrib name="rescpu.actpk5.latest" alias="ResCpuActpk5Lat" type="Gauge"/> | |
+ <attrib name="rescpu.maxLimited15.latest" alias="ResCpuMaxLd15Lat" type="Gauge"/> | |
+ <attrib name="rescpu.maxLimited1.latest" alias="ResCpuMaxLd1Lat" type="Gauge"/> | |
+ <attrib name="rescpu.maxLimited5.latest" alias="ResCpuMaxLd5Lat" type="Gauge"/> | |
+ <attrib name="rescpu.runav15.latest" alias="ResCpuRunav15Lat" type="Gauge"/> | |
+ <attrib name="rescpu.runav1.latest" alias="ResCpuRunav1Lat" type="Gauge"/> | |
+ <attrib name="rescpu.runav5.latest" alias="ResCpuRunav5Lat" type="Gauge"/> | |
+ <attrib name="rescpu.runpk15.latest" alias="ResCpuRunpk15Lat" type="Gauge"/> | |
+ <attrib name="rescpu.runpk1.latest" alias="ResCpuRunpk1Lat" type="Gauge"/> | |
+ <attrib name="rescpu.runpk5.latest" alias="ResCpuRunpk5Lat" type="Gauge"/> | |
+ <attrib name="rescpu.sampleCount.latest" alias="ResCpuSeCtLat" type="Gauge"/> | |
+ <attrib name="rescpu.samplePeriod.latest" alias="ResCpuSePeriodLat" type="Gauge"/> | |
+ <attrib name="storageAdapter.maxTotalLatency.latest" alias="StAdptrMaxTlLyLat" type="Gauge"/> | |
+ <attrib name="storagePath.maxTotalLatency.latest" alias="StPthMaxTlLyLat" type="Gauge"/> | |
+ <attrib name="sys.uptime.latest" alias="SysUpTeLat" type="Gauge"/> | |
+ </vmware-group> | |
+ <vmware-group name="vmware6Cpu" resourceType="vmware6Cpu"> | |
+ <attrib name="cpu.coreUtilization.average" alias="CpuCoreUnAvg" type="Gauge"/> | |
+ <attrib name="cpu.idle.summation" alias="CpuIdleSum" type="Gauge"/> | |
+ <attrib name="cpu.utilization.average" alias="CpuUnAvg" type="Gauge"/> | |
+ <attrib name="cpu.usage.average" alias="CpuUsageAvg" type="Gauge"/> | |
+ <attrib name="cpu.used.summation" alias="CpuUsedSum" type="Gauge"/> | |
+ </vmware-group> | |
+ <vmware-group name="vmware6Sys" resourceType="vmware6Sys"> | |
+ <attrib name="sys.resourceCpuAllocMin.latest" alias="SysReCpuAcMinLat" type="Gauge"/> | |
+ <attrib name="sys.resourceCpuAllocShares.latest" alias="SysReCpuAcSsLat" type="Gauge"/> | |
+ <attrib name="sys.resourceCpuAct1.latest" alias="SysReCpuAct1Lat" type="Gauge"/> | |
+ <attrib name="sys.resourceCpuAct5.latest" alias="SysReCpuAct5Lat" type="Gauge"/> | |
+ <attrib name="sys.resourceCpuMaxLimited1.latest" alias="SysReCpuMaxLd1Lat" type="Gauge"/> | |
+ <attrib name="sys.resourceCpuMaxLimited5.latest" alias="SysReCpuMaxLd5Lat" type="Gauge"/> | |
+ <attrib name="sys.resourceCpuRun1.latest" alias="SysReCpuRun1Lat" type="Gauge"/> | |
+ <attrib name="sys.resourceCpuRun5.latest" alias="SysReCpuRun5Lat" type="Gauge"/> | |
+ <attrib name="sys.resourceCpuUsage.average" alias="SysReCpuUsageAvg" type="Gauge"/> | |
+ <attrib name="sys.resourceFdUsage.latest" alias="SysReFdUsageLat" type="Gauge"/> | |
+ <attrib name="sys.resourceMemAllocMax.latest" alias="SysReMemAcMaxLat" type="Gauge"/> | |
+ <attrib name="sys.resourceMemAllocMin.latest" alias="SysReMemAcMinLat" type="Gauge"/> | |
+ <attrib name="sys.resourceMemAllocShares.latest" alias="SysReMemAcSsLat" type="Gauge"/> | |
+ <attrib name="sys.resourceMemConsumed.latest" alias="SysReMemCdLat" type="Gauge"/> | |
+ <attrib name="sys.resourceMemCow.latest" alias="SysReMemCowLat" type="Gauge"/> | |
+ <attrib name="sys.resourceMemMapped.latest" alias="SysReMemMappedLat" type="Gauge"/> | |
+ <attrib name="sys.resourceMemOverhead.latest" alias="SysReMemOdLat" type="Gauge"/> | |
+ <attrib name="sys.resourceMemShared.latest" alias="SysReMemSharedLat" type="Gauge"/> | |
+ <attrib name="sys.resourceMemSwapped.latest" alias="SysReMemSppedLat" type="Gauge"/> | |
+ <attrib name="sys.resourceMemTouched.latest" alias="SysReMemTdLat" type="Gauge"/> | |
+ <attrib name="sys.resourceMemZero.latest" alias="SysReMemZeroLat" type="Gauge"/> | |
+ </vmware-group> | |
+ <vmware-group name="vmware6Net" resourceType="vmware6Net"> | |
+ <attrib name="net.broadcastRx.summation" alias="NetBroadcastRxSum" type="Gauge"/> | |
+ <attrib name="net.broadcastTx.summation" alias="NetBroadcastTxSum" type="Gauge"/> | |
+ <attrib name="net.bytesRx.average" alias="NetBytesRxAvg" type="Gauge"/> | |
+ <attrib name="net.bytesTx.average" alias="NetBytesTxAvg" type="Gauge"/> | |
+ <attrib name="net.droppedRx.summation" alias="NetDroppedRxSum" type="Gauge"/> | |
+ <attrib name="net.droppedTx.summation" alias="NetDroppedTxSum" type="Gauge"/> | |
+ <attrib name="net.errorsRx.summation" alias="NetErrorsRxSum" type="Gauge"/> | |
+ <attrib name="net.errorsTx.summation" alias="NetErrorsTxSum" type="Gauge"/> | |
+ <attrib name="net.multicastRx.summation" alias="NetMulticastRxSum" type="Gauge"/> | |
+ <attrib name="net.multicastTx.summation" alias="NetMulticastTxSum" type="Gauge"/> | |
+ <attrib name="net.packetsRx.summation" alias="NetPacketsRxSum" type="Gauge"/> | |
+ <attrib name="net.packetsTx.summation" alias="NetPacketsTxSum" type="Gauge"/> | |
+ <attrib name="net.received.average" alias="NetReceivedAvg" type="Gauge"/> | |
+ <attrib name="net.transmitted.average" alias="NetTransmittedAvg" type="Gauge"/> | |
+ <attrib name="net.unknownProtos.summation" alias="NetUnknownPsSum" type="Gauge"/> | |
+ <attrib name="net.usage.average" alias="NetUsageAvg" type="Gauge"/> | |
+ </vmware-group> | |
+ <vmware-group name="vmware6Disk" resourceType="vmware6Disk"> | |
+ <attrib name="disk.busResets.summation" alias="DiskBusResetsSum" type="Gauge"/> | |
+ <attrib name="disk.commandsAveraged.average" alias="DiskCsAdAvg" type="Gauge"/> | |
+ <attrib name="disk.commandsAborted.summation" alias="DiskCsAdSum" type="Gauge"/> | |
+ <attrib name="disk.commands.summation" alias="DiskCsSum" type="Gauge"/> | |
+ <attrib name="disk.deviceLatency.average" alias="DiskDeLyAvg" type="Gauge"/> | |
+ <attrib name="disk.deviceReadLatency.average" alias="DiskDeRdLyAvg" type="Gauge"/> | |
+ <attrib name="disk.deviceWriteLatency.average" alias="DiskDeWeLyAvg" type="Gauge"/> | |
+ <attrib name="disk.kernelLatency.average" alias="DiskKlLyAvg" type="Gauge"/> | |
+ <attrib name="disk.kernelReadLatency.average" alias="DiskKlRdLyAvg" type="Gauge"/> | |
+ <attrib name="disk.kernelWriteLatency.average" alias="DiskKlWeLyAvg" type="Gauge"/> | |
+ <attrib name="disk.maxQueueDepth.average" alias="DiskMaxQeDhAvg" type="Gauge"/> | |
+ <attrib name="disk.numberReadAveraged.average" alias="DiskNrRdAdAvg" type="Gauge"/> | |
+ <attrib name="disk.numberRead.summation" alias="DiskNrRdSum" type="Gauge"/> | |
+ <attrib name="disk.numberWriteAveraged.average" alias="DiskNrWeAdAvg" type="Gauge"/> | |
+ <attrib name="disk.numberWrite.summation" alias="DiskNrWeSum" type="Gauge"/> | |
+ <attrib name="disk.queueLatency.average" alias="DiskQeLyAvg" type="Gauge"/> | |
+ <attrib name="disk.queueReadLatency.average" alias="DiskQeRdLyAvg" type="Gauge"/> | |
+ <attrib name="disk.queueWriteLatency.average" alias="DiskQeWeLyAvg" type="Gauge"/> | |
+ <attrib name="disk.read.average" alias="DiskRdAvg" type="Gauge"/> | |
+ <attrib name="disk.totalLatency.average" alias="DiskTlLyAvg" type="Gauge"/> | |
+ <attrib name="disk.totalReadLatency.average" alias="DiskTlRdLyAvg" type="Gauge"/> | |
+ <attrib name="disk.totalWriteLatency.average" alias="DiskTlWeLyAvg" type="Gauge"/> | |
+ <attrib name="disk.write.average" alias="DiskWeAvg" type="Gauge"/> | |
+ </vmware-group> | |
+ <vmware-group name="vmware6vflashModule" resourceType="vmware6vflashModule"> | |
+ <attrib name="vflashModule.numActiveVMDKs.latest" alias="vflModNumAeVMDKsLat" type="Gauge"/> | |
+ </vmware-group> | |
+ <vmware-group name="vmware6DaSt" resourceType="vmware6DaSt"> | |
+ <attrib name="datastore.datastoreIops.average" alias="DaStDeIopsAvg" type="Gauge"/> | |
+ <attrib name="datastore.datastoreMaxQueueDepth.latest" alias="DaStDeMaxQeDhLat" type="Gauge"/> | |
+ <attrib name="datastore.datastoreNormalReadLatency.latest" alias="DaStDeNlRdLyLat" type="Gauge"/> | |
+ <attrib name="datastore.datastoreNormalWriteLatency.latest" alias="DaStDeNlWeLyLat" type="Gauge"/> | |
+ <attrib name="datastore.datastoreReadBytes.latest" alias="DaStDeRdBytesLat" type="Gauge"/> | |
+ <attrib name="datastore.datastoreReadIops.latest" alias="DaStDeRdIopsLat" type="Gauge"/> | |
+ <attrib name="datastore.datastoreReadLoadMetric.latest" alias="DaStDeRdLdMcLat" type="Gauge"/> | |
+ <attrib name="datastore.datastoreReadOIO.latest" alias="DaStDeRdOIOLat" type="Gauge"/> | |
+ <attrib name="datastore.datastoreVMObservedLatency.latest" alias="DaStDeVMOdLyLat" type="Gauge"/> | |
+ <attrib name="datastore.datastoreWriteBytes.latest" alias="DaStDeWeBytesLat" type="Gauge"/> | |
+ <attrib name="datastore.datastoreWriteIops.latest" alias="DaStDeWeIopsLat" type="Gauge"/> | |
+ <attrib name="datastore.datastoreWriteLoadMetric.latest" alias="DaStDeWeLdMcLat" type="Gauge"/> | |
+ <attrib name="datastore.datastoreWriteOIO.latest" alias="DaStDeWeOIOLat" type="Gauge"/> | |
+ <attrib name="datastore.numberReadAveraged.average" alias="DaStNrRdAdAvg" type="Gauge"/> | |
+ <attrib name="datastore.numberWriteAveraged.average" alias="DaStNrWeAdAvg" type="Gauge"/> | |
+ <attrib name="datastore.read.average" alias="DaStRdAvg" type="Gauge"/> | |
+ <attrib name="datastore.siocActiveTimePercentage.average" alias="DaStSiocAeTePeAvg" type="Gauge"/> | |
+ <attrib name="datastore.sizeNormalizedDatastoreLatency.average" alias="DaStSizeNdDeLyAvg" type="Gauge"/> | |
+ <attrib name="datastore.totalReadLatency.average" alias="DaStTlRdLyAvg" type="Gauge"/> | |
+ <attrib name="datastore.totalWriteLatency.average" alias="DaStTlWeLyAvg" type="Gauge"/> | |
+ <attrib name="datastore.write.average" alias="DaStWeAvg" type="Gauge"/> | |
+ </vmware-group> | |
+ </vmware-groups> | |
+ </vmware-collection> | |
</vmware-datacollection-config> | |
diff --git a/xml-datacollection-config.xml b/xml-datacollection-config.xml | |
new file mode 100644 | |
index 0000000..5780949 | |
--- /dev/null | |
+++ b/xml-datacollection-config.xml | |
@@ -0,0 +1,19 @@ | |
+<xml-datacollection-config rrdRepository="/opt/opennms/share/rrd/snmp/" xmlns="http://xmlns.opennms.org/xsd/config/xml-datacollection"> | |
+ <!-- | |
+ To understand how the XML Collector works, and how to configure it, please check the following link: | |
+ http://www.opennms.org/wiki/XML_Collector | |
+ --> | |
+ | |
+ <xml-collection name="xml-elasticsearch-cluster-stats"> | |
+ <rrd step="300"> | |
+ <rra>RRA:AVERAGE:0.5:1:2016</rra> | |
+ <rra>RRA:AVERAGE:0.5:12:1488</rra> | |
+ <rra>RRA:AVERAGE:0.5:288:366</rra> | |
+ <rra>RRA:MAX:0.5:288:366</rra> | |
+ <rra>RRA:MIN:0.5:288:366</rra> | |
+ </rrd> | |
+ <xml-source url="http://{ipaddr}:9200/_cluster/stats"> | |
+ <import-groups>xml-datacollection/elasticsearch-cluster-stats.xml</import-groups> | |
+ </xml-source> | |
+ </xml-collection> | |
+</xml-datacollection-config> | |
diff --git a/xml-datacollection/elasticsearch-cluster-stats.xml b/xml-datacollection/elasticsearch-cluster-stats.xml | |
new file mode 100644 | |
index 0000000..6659e11 | |
--- /dev/null | |
+++ b/xml-datacollection/elasticsearch-cluster-stats.xml | |
@@ -0,0 +1,156 @@ | |
+<xml-groups> | |
+ <!-- | |
+ Elasticsearch Performance metrics. | |
+ | |
+ The idea to define this on a external file is to share the content between multiple XML sources | |
+ | |
+ IMPORTANT: because of the current limitation for the data source inside RRDs/JRBs of 19 | |
+ characters, the XML objects are using a auto-generated ID to ensure the DS | |
+ length. The real variable name is displayed as a comment after the XML object | |
+ definition for information purposes. For graph templates use should refer to | |
+ the performance metrics using the XML Object Name (auto-generated ID) | |
+ --> | |
+ <xml-group name="cluster" resource-type="node" resource-xpath="/"> | |
+ <xml-object name="clusterName" type="STRING" xpath="cluster_name"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-indices" resource-type="node" resource-xpath="/indices"> | |
+ <xml-object name="indicesCount" type="GAUGE" xpath="count"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-indices-shards-index-shards" resource-type="node" resource-xpath="/indices/shards/index/shards"> | |
+ <xml-object name="shardsMin" type="GAUGE" xpath="min"/> | |
+ <xml-object name="shardsMax" type="GAUGE" xpath="max"/> | |
+ <xml-object name="shardsAvg" type="GAUGE" xpath="avg"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-indices-shards-index-primaries" resource-type="node" resource-xpath="/indices/shards/index/primaries"> | |
+ <xml-object name="primariesMin" type="GAUGE" xpath="min"/> | |
+ <xml-object name="primariesMax" type="GAUGE" xpath="max"/> | |
+ <xml-object name="primariesAvg" type="GAUGE" xpath="avg"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-indices-shards-index-replication" resource-type="node" resource-xpath="/indices/shards/index/replication"> | |
+ <xml-object name="replicationMin" type="GAUGE" xpath="min"/> | |
+ <xml-object name="replicationMax" type="GAUGE" xpath="max"/> | |
+ <xml-object name="replicationAvg" type="GAUGE" xpath="avg"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-indices-docs" resource-type="node" resource-xpath="/indices/docs"> | |
+ <xml-object name="docsCount" type="COUNTER" xpath="count"/> | |
+ <xml-object name="docsDeleted" type="COUNTER" xpath="deleted"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-indices-store" resource-type="node" resource-xpath="/indices/store"> | |
+ <xml-object name="storeSizeBytes" type="GAUGE" xpath="size_in_bytes"/> | |
+ <xml-object name="throttleTimeMillis" type="GAUGE" xpath="throttle_time_in_millis"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-indices-fielddata" resource-type="node" resource-xpath="/indices/fielddata"> | |
+ <xml-object name="memorySizeBytes" type="GAUGE" xpath="memory_size_in_bytes"/> | |
+ <xml-object name="memoryEvictions" type="GAUGE" xpath="evictions"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-indices-filtercache" resource-type="node" resource-xpath="/indices/filter_cache"> | |
+ <xml-object name="fltrCacheBytes" type="GAUGE" xpath="memory_size_in_bytes"/> | |
+ <xml-object name="fltrCacheEvictions" type="GAUGE" xpath="evictions"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-indices-idcache" resource-type="node" resource-xpath="/indices/id_cache"> | |
+ <xml-object name="idCacheBytes" type="GAUGE" xpath="memory_size_in_bytes"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-indices-completion" resource-type="node" resource-xpath="/indices/completion"> | |
+ <xml-object name="completionSizeBytes" type="GAUGE" xpath="memory_size_in_bytes"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-indices-segments" resource-type="node" resource-xpath="/indices/segments"> | |
+ <xml-object name="segCount" type="GAUGE" xpath="count"/> | |
+ <xml-object name="segMemory" type="GAUGE" xpath="memory_in_bytes"/> | |
+ <xml-object name="segWriterMem" type="GAUGE" xpath="index_writer_memory_in_bytes"/> | |
+ <xml-object name="segWriterMemMax" type="GAUGE" xpath="index_writer_max_memory_in_bytes"/> | |
+ <xml-object name="segVerMapMem" type="GAUGE" xpath="version_map_memory_in_bytes"/> | |
+ <xml-object name="segFixBitSetMem" type="GAUGE" xpath="fixed_bit_set_memory_in_bytes"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-indices-percolate" resource-type="node" resource-xpath="/indices/percolate"> | |
+ <xml-object name="percTotal" type="GAUGE" xpath="total"/> | |
+ <xml-object name="percTimeInMillis" type="GAUGE" xpath="time_in_millis"/> | |
+ <xml-object name="percCurrent" type="GAUGE" xpath="current"/> | |
+ <xml-object name="percMemSizeBytes" type="GAUGE" xpath="memory_size_in_bytes"/> | |
+ <xml-object name="percQueries" type="GAUGE" xpath="queries"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-nodes-count" resource-type="node" resource-xpath="/nodes/count"> | |
+ <xml-object name="nodesCntTotal" type="GAUGE" xpath="total"/> | |
+ <xml-object name="nodesCntMstOnly" type="GAUGE" xpath="master_only"/> | |
+ <xml-object name="nodesCntDataOnly" type="GAUGE" xpath="data_only"/> | |
+ <xml-object name="nodesCntMasterData" type="GAUGE" xpath="master_data"/> | |
+ <xml-object name="nodesClient" type="GAUGE" xpath="client"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-nodes-os" resource-type="node" resource-xpath="/nodes/os"> | |
+ <xml-object name="osAvailProc" type="GAUGE" xpath="available_processors"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-nodes-os-mem" resource-type="node" resource-xpath="/nodes/os/mem"> | |
+ <xml-object name="osMemTotalBytes" type="GAUGE" xpath="total_in_bytes"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-nodes-os-cpu" resource-type="node" resource-xpath="/nodes/os/cpu"> | |
+ <xml-object name="cpuVendor" type="STRING" xpath="vendor"/> | |
+ <xml-object name="cpuModel" type="STRING" xpath="model"/> | |
+ <xml-object name="cpuMhz" type="STRING" xpath="mhz"/> | |
+ <xml-object name="cpuCacheSize" type="STRING" xpath="cache_size"/> | |
+ <xml-object name="cpuTotalCores" type="GAUGE" xpath="total_cores"/> | |
+ <xml-object name="cpuTotalSockets" type="GAUGE" xpath="total_sockets"/> | |
+ <xml-object name="cpuCoresPerSocket" type="GAUGE" xpath="cores_per_socket"/> | |
+ <xml-object name="cpuCacheSizeBytes" type="STRING" xpath="cache_size_in_bytes"/> | |
+ <xml-object name="cpuCount" type="GAUGE" xpath="count"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-nodes-process-cpu" resource-type="node" resource-xpath="/nodes/process/cpu"> | |
+ <xml-object name="procCpuPercent" type="GAUGE" xpath="percent"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-nodes-process-open-file-descriptors" resource-type="node" resource-xpath="/nodes/process/open_file_descriptors"> | |
+ <xml-object name="openFileDescMin" type="GAUGE" xpath="min"/> | |
+ <xml-object name="openFileDescMax" type="GAUGE" xpath="max"/> | |
+ <xml-object name="openFileDescAvg" type="GAUGE" xpath="avg"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-nodes-jvm" resource-type="node" resource-xpath="/nodes/jvm"> | |
+ <xml-object name="jvmMaxUptimeMillis" type="GAUGE" xpath="max_uptime_in_millis"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-nodes-jvm-versions" resource-type="node" resource-xpath="/nodes/jvm/versions"> | |
+ <xml-object name="jvmVersion" type="STRING" xpath="version"/> | |
+ <xml-object name="jvmVmName" type="STRING" xpath="vm_name"/> | |
+ <xml-object name="jvmVmVersion" type="STRING" xpath="vm_version"/> | |
+ <xml-object name="jvmVmVendor" type="STRING" xpath="vm_vendor"/> | |
+ <xml-object name="jvmCount" type="GAUGE" xpath="count"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-nodes-jvm-mem" resource-type="node" resource-xpath="/nodes/jvm/mem"> | |
+ <xml-object name="jvmMemHeapBytes" type="GAUGE" xpath="heap_used_in_bytes"/> | |
+ <xml-object name="jvmMemMaxBytes" type="GAUGE" xpath="heap_max_in_bytes"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-nodes-jvm-threads" resource-type="node" resource-xpath="/nodes/jvm"> | |
+ <xml-object name="jvmThreads" type="GAUGE" xpath="threads"/> | |
+ </xml-group> | |
+ | |
+ <xml-group name="cluster-nodes-fs" resource-type="node" resource-xpath="/nodes/fs"> | |
+ <xml-object name="fsTotalBytes" type="GAUGE" xpath="total_in_bytes"/> | |
+ <xml-object name="fsFreeBytes" type="GAUGE" xpath="free_in_bytes"/> | |
+ <xml-object name="fsAvailBytes" type="GAUGE" xpath="available_in_bytes"/> | |
+ <xml-object name="fsDskReads" type="GAUGE" xpath="disk_reads"/> | |
+ <xml-object name="fsDskWrites" type="GAUGE" xpath="disk_writes"/> | |
+ <xml-object name="fsDskIoOp" type="GAUGE" xpath="disk_io_op"/> | |
+ <xml-object name="fsDskReadSizeBytes" type="GAUGE" xpath="disk_read_size_in_bytes"/> | |
+ <xml-object name="fsDskWriteSizeBytes" type="GAUGE" xpath="disk_write_size_in_bytes"/> | |
+ <xml-object name="fsDskIoSizeBytes" type="GAUGE" xpath="disk_io_size_in_bytes"/> | |
+ <xml-object name="fsDskQueue" type="GAUGE" xpath="disk_queue"/> | |
+ <xml-object name="fsDskSvcTime" type="GAUGE" xpath="disk_service_time"/> | |
+ </xml-group> | |
+</xml-groups> |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment