Last active
June 21, 2021 11:16
-
-
Save cluther/4320829 to your computer and use it in GitHub Desktop.
Create Zenoss monitoring templates from YAML
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# | |
# COMMAND Example | |
# | |
/ExampleDeviceClass/ExampleCommandTemplate: | |
datasources: | |
echo: | |
type: COMMAND | |
commandTemplate: 'echo "OK|val1=123 val2=987.6' | |
parser: Nagios | |
severity: info | |
cycletime: 10 | |
datapoints: | |
val1: | |
rrdmin: 0 | |
aliases: | |
value1: "100,/" | |
val2: DERIVE_MIN_0 | |
thresholds: | |
high values: | |
type: MinMaxThreshold | |
dsnames: ["ds1_val1", "ds1_val2"] | |
maxval: "99" | |
graphs: | |
Values: | |
units: number | |
miny: 0 | |
graphpoints: | |
Value 1: | |
dpName: "ds1_val1" | |
format: "%7.2lf%s" | |
Value 2: | |
dpName: "ds1_val2" | |
format: "%7.2lf%s" | |
# | |
# SNMP Example | |
# | |
/ExampleDeviceClass/ExampleSNMPTemplate: | |
datasources: | |
hrMemoryUsed: | |
type: SNMP | |
oid: "1.3.6.1.2.1.25.2.3.1.6" | |
datapoints: | |
hrMemoryUsed: GAUGE_MIN_0 | |
hrProcessorLoad: | |
type: SNMP | |
oid: "1.3.6.1.2.1.25.3.3.1.2.1" | |
datapoints: | |
hrProcessorLoad: GAUGE_MIN_0 | |
thresholds: | |
high load: | |
type: MinMaxThreshold | |
dsnames: ["hrProcessorLoad_hrProcessorLoad"] | |
maxval: "95" | |
graphs: | |
CPU Utilization: | |
units: "percent" | |
miny: 0 | |
maxy: 100 | |
graphpoints: | |
Used: | |
dpName: "hrProcessorLoad_hrProcessorLoad" | |
format: "%4.0lf%%" | |
Memory Utilization: | |
units: "percent" | |
miny: 0 | |
maxy: 100 | |
graphpoints: | |
Used: | |
dpName: "hrMemoryUsed_hrMemoryUsed" | |
format: "%7.2lf%%" | |
rpn: "1024,*,${here/hw/totalMemory},/,100,*" | |
# | |
# JMX Example | |
# | |
/ExampleDeviceClass/ExampleJMXTemplate: | |
datasources: | |
heapMemoryUsage: | |
type: JMX | |
jmxPort: "12345" | |
authenticate: True | |
objectName: "java.lang:type=Memory" | |
attributeName: "HeapMemoryUsage" | |
datapoints: | |
committed: GAUGE_MIN_0 | |
used: GAUGE_MIN_0 | |
nonHeapMemoryUsage: | |
type: JMX | |
jmxPort: "12345" | |
authenticate: True | |
objectName: "java.lang:type=Memory" | |
attributeName: "NonHeapMemoryUsage" | |
datapoints: | |
committed: GAUGE_MIN_0 | |
used: GAUGE_MIN_0 | |
graphs: | |
JVM Memory Usage: | |
units: bytes | |
base: true | |
miny: 0 | |
graphpoints: | |
Heap Committed: | |
dpName: heapMemoryUsage_committed | |
Heap Used: | |
dpName: heapMemoryUsage_used | |
NonHeap Committed: | |
dpName: nonHeapMemoryUsage_committed | |
NonHeap Used: | |
dpName: nonHeapMemoryUsage_used |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
#!/usr/bin/env python | |
""" | |
Build monitoring templates from a YAML input file. | |
Look at monitoring_templates.yaml in the same directory for an example | |
of what the YAML schema should be. | |
WARNING: This will delete and recreate the named monitoring templates. | |
So it can potentially be very destructive. | |
""" | |
import logging | |
LOG = logging.getLogger() | |
import os | |
import re | |
import sys | |
import types | |
def die(msg, *args): | |
LOG.fatal(msg, *args) | |
sys.exit(1) | |
try: | |
import Globals | |
except ImportError: | |
die('Must be run as the zenoss user. Try "su - zenoss" first.') | |
from Products.ZenUtils.Utils import unused | |
unused(Globals) | |
from Products.ZenUtils.ZenScriptBase import ZenScriptBase | |
dmd = ZenScriptBase(connect=True).dmd | |
# ZenScriptBase breaks logging. We'll fix it. | |
LOG.removeHandler(LOG.handlers[0]) | |
logging.basicConfig( | |
format='%(message)s', | |
level=logging.DEBUG if os.environ.get('DEBUG') else logging.INFO) | |
from transaction import commit | |
try: | |
import yaml | |
import yaml.constructor | |
except ImportError: | |
die('PyYAML must be installed. Try "easy_install PyYAML" first.') | |
try: | |
# included in standard lib from Python 2.7 | |
from collections import OrderedDict | |
except ImportError: | |
# try importing the backported drop-in replacement | |
# it's available on PyPI | |
try: | |
from ordereddict import OrderedDict | |
except ImportError: | |
die('ordereddict must be installed. Try "easy_install ordereddict" first.') | |
from Products.ZenModel.ComplexGraphPoint import ComplexGraphPoint | |
from Products.ZenModel.DeviceClass import DeviceClass | |
from Products.ZenModel.GraphDefinition import GraphDefinition | |
from Products.ZenModel.GraphPoint import GraphPoint | |
from Products.ZenModel.DataPointGraphPoint import DataPointGraphPoint | |
from Products.ZenModel.RRDDataPoint import RRDDataPoint | |
from Products.ZenModel.RRDDataSource import RRDDataSource | |
from Products.ZenModel.RRDTemplate import RRDTemplate | |
from Products.ZenModel.ThresholdClass import ThresholdClass | |
class OrderedDictYAMLLoader(yaml.Loader): | |
""" | |
A YAML loader that loads mappings into ordered dictionaries. | |
This is used to allow graph definitions to be sequenced naturally | |
without having to specify a sequence attribute. | |
""" | |
def __init__(self, *args, **kwargs): | |
yaml.Loader.__init__(self, *args, **kwargs) | |
self.add_constructor(u'tag:yaml.org,2002:map', type(self).construct_yaml_map) | |
self.add_constructor(u'tag:yaml.org,2002:omap', type(self).construct_yaml_map) | |
def construct_yaml_map(self, node): | |
data = OrderedDict() | |
yield data | |
value = self.construct_mapping(node) | |
data.update(value) | |
def construct_mapping(self, node, deep=False): | |
if isinstance(node, yaml.MappingNode): | |
self.flatten_mapping(node) | |
else: | |
raise yaml.constructor.ConstructorError( | |
None, None, | |
'expected a mapping node, but found %s' % node.id, | |
node.start_mark) | |
mapping = OrderedDict() | |
for key_node, value_node in node.value: | |
key = self.construct_object(key_node, deep=deep) | |
try: | |
hash(key) | |
except TypeError, exc: | |
raise yaml.constructor.ConstructorError( | |
'while constructing a mapping', | |
node.start_mark, 'found unacceptable key (%s)' % exc, | |
key_node.start_mark) | |
value = self.construct_object(value_node, deep=deep) | |
mapping[key] = value | |
return mapping | |
def main(): | |
data = {} | |
if len(sys.argv) < 2: | |
LOG.info("STDIN: loading YAML") | |
data.update(yaml.load(sys.stdin.read(), OrderedDictYAMLLoader) or {}) | |
else: | |
for filename in sys.argv[1:]: | |
LOG.info("{}: loading YAML".format(filename)) | |
with open(filename, 'r') as yaml_file: | |
data.update(yaml.load(yaml_file, OrderedDictYAMLLoader) or {}) | |
for template_path, template_cfg in data.items(): | |
add_template(template_path, template_cfg) | |
commit() | |
def log_for(obj, msg, level=logging.DEBUG): | |
''' | |
Log msg for obj at level. | |
''' | |
parts = None | |
if isinstance(obj, DeviceClass): | |
parts = ( | |
obj.getOrganizerName(),) | |
elif isinstance(obj, RRDTemplate): | |
parts = ( | |
obj.deviceClass().getOrganizerName(), | |
obj.id) | |
elif isinstance(obj, (RRDDataSource, ThresholdClass, GraphDefinition)): | |
parts = ( | |
obj.rrdTemplate().deviceClass().getOrganizerName(), | |
obj.rrdTemplate().id, | |
obj.id) | |
elif isinstance(obj, RRDDataPoint): | |
parts = ( | |
obj.datasource().rrdTemplate().deviceClass().getOrganizerName(), | |
obj.datasource().rrdTemplate().id, | |
obj.datasource().id, | |
obj.id) | |
elif isinstance(obj, GraphPoint): | |
parts = ( | |
obj.graphDef().rrdTemplate().deviceClass().getOrganizerName(), | |
obj.graphDef().rrdTemplate().id, | |
obj.graphDef().id, | |
obj.id) | |
else: | |
parts = () | |
LOG.log(level, '%s: %s', ' - '.join(parts), msg) | |
def get_severity(value): | |
''' | |
Return numeric severity given a string representation of severity. | |
''' | |
try: | |
severity = int(value) | |
except (TypeError, ValueError): | |
severity = { | |
'crit': 5, 'critical': 5, | |
'err': 4, 'error': 4, | |
'warn': 3, 'warning': 3, | |
'info': 2, 'information': 2, 'informational': 2, | |
'debug': 1, 'debugging': 1, | |
'clear': 0, | |
}.get(value.lower()) | |
if severity is None: | |
die("'%s' is not a valid value for severity.", value) | |
return severity | |
def apply_properties(obj, cfg, ignore=None): | |
if ignore: | |
for propname in ignore: | |
if propname in cfg: | |
del(cfg[propname]) | |
given_properties = set(cfg) | |
valid_properties = set(x['id'] for x in obj._properties) | |
invalid_properties = given_properties - valid_properties | |
if len(invalid_properties) == 1: | |
die("%s is an invalid %s property. Valid properties: %s", | |
list(invalid_properties)[0], | |
obj.__class__.__name__, | |
', '.join(valid_properties)) | |
elif len(invalid_properties) > 1: | |
die("%s are invalid %s properties. Valid properties: %s", | |
', '.join(invalid_properties), | |
obj.__class__.__name__, | |
', '.join(valid_properties)) | |
# Map severity names to values. | |
if 'severity' in cfg: | |
cfg['severity'] = get_severity(cfg['severity']) | |
# TODO: Detect property type and do conversion. | |
# Apply cfg items directly to datasource attributes. | |
for k, v in cfg.items(): | |
if k not in ('type', 'datapoints'): | |
log_for(obj, "{} = {}".format(k, v), level=logging.DEBUG) | |
setattr(obj, k, v) | |
def add_template(path, cfg): | |
if '/' not in path: | |
die("%s is not a path. Include device class and template name", path) | |
path_parts = path.split('/') | |
id_ = path_parts[-1] | |
cfg['deviceClass'] = '/'.join(path_parts[:-1]) | |
device_class = dmd.Devices.createOrganizer(cfg['deviceClass']) | |
existing_template = device_class.rrdTemplates._getOb(id_, None) | |
if existing_template: | |
log_for(existing_template, "replacing template", level=logging.INFO) | |
device_class.rrdTemplates._delObject(id_) | |
device_class.manage_addRRDTemplate(id_) | |
template = device_class.rrdTemplates._getOb(id_) | |
if not existing_template: | |
log_for(template, "adding template", level=logging.INFO) | |
if 'targetPythonClass' in cfg: | |
log_for( | |
template, | |
"targetPythonClass = {}".format(cfg['targetPythonClass'])) | |
template.targetPythonClass = cfg['targetPythonClass'] | |
if 'description' in cfg: | |
log_for( | |
template, | |
"description = {}".format(cfg['description'])) | |
template.description = cfg['description'] | |
if 'thresholds' in cfg: | |
log_for( | |
template, "adding {} thresholds".format(len(cfg['thresholds']))) | |
for threshold_id, threshold_cfg in cfg['thresholds'].items(): | |
add_threshold(template, threshold_id, threshold_cfg) | |
if 'datasources' in cfg: | |
log_for(template, "adding {} datasources".format(len(cfg['datasources']))) | |
for datasource_id, datasource_cfg in cfg['datasources'].items(): | |
add_datasource(template, datasource_id, datasource_cfg) | |
if 'graphs' in cfg: | |
log_for(template, "adding {} graphs".format(len(cfg['graphs']))) | |
for graph_id, graph_cfg in cfg['graphs'].items(): | |
add_graph(template, graph_id, graph_cfg) | |
def add_datasource(template, id_, cfg): | |
datasource_types = dict(template.getDataSourceOptions()) | |
if 'type' not in cfg: | |
die('No type for %s/%s. Valid types: %s', | |
template.id, id_, ', '.join(datasource_types)) | |
type_ = datasource_types.get(cfg['type']) | |
if not type_: | |
die("%s is an invalid datasource type. Valid types: %s", | |
cfg['type'], ', '.join(datasource_types)) | |
# Map severity names to values. | |
if 'severity' in cfg: | |
cfg['severity'] = get_severity(cfg['severity']) | |
datasource = template.manage_addRRDDataSource(id_, type_) | |
log_for(datasource, "adding datasource") | |
if 'datapoints' in cfg: | |
log_for(template, "adding {} datapoints".format(len(cfg['datapoints']))) | |
for datapoint_id, datapoint_cfg in cfg['datapoints'].items(): | |
add_datapoint(datasource, datapoint_id, datapoint_cfg) | |
apply_properties(datasource, cfg, ignore=['type', 'datapoints']) | |
def add_datapoint(datasource, id_, cfg): | |
datapoint = datasource.manage_addRRDDataPoint(id_) | |
log_for(datapoint, "adding datapoint") | |
# Handle cfg shortcuts like DERIVE_MIN_0 and GAUGE_MIN_0_MAX_100. | |
if isinstance(cfg, types.StringTypes): | |
log_for(datapoint, "using shortcut syntax") | |
if 'DERIVE' in cfg.upper(): | |
log_for(datapoint, "rrdtype = {}".format('DERIVE')) | |
datapoint.rrdtype = 'DERIVE' | |
min_match = re.search(r'MIN_(\d+)', cfg, re.IGNORECASE) | |
if min_match: | |
rrdmin = min_match.group(1) | |
log_for(datapoint, "rrdmin = {}".format(rrdmin)) | |
datapoint.rrdmin = rrdmin | |
max_match = re.search(r'MAX_(\d+)', cfg, re.IGNORECASE) | |
if max_match: | |
rrdmax = max_match.group(1) | |
log_for(datapoint, "rrdmax = {}".format(rrdmax)) | |
datapoint.rrdmax = rrdmax | |
return | |
if 'aliases' in cfg: | |
log_for(datapoint, "adding {} aliases".format(len(cfg['aliases']))) | |
for alias_id, formula in cfg['aliases'].items(): | |
datapoint.addAlias(alias_id, formula) | |
alias = datapoint.aliases._getOb(alias_id) | |
log_for(alias, "adding alias".format(alias_id)) | |
log_for(alias, "formula = {}".format(formula)) | |
apply_properties(datapoint, cfg, ignore=['aliases']) | |
def add_threshold(template, id_, cfg): | |
if 'type' not in cfg: | |
# Default to MinMaxThreshold since they're the most common. | |
cfg['type'] = 'MinMaxThreshold' | |
if 'dsnames' not in cfg and 'dsname' not in cfg: | |
die("'%s' threshold has no dsname or dsnames attribute", id_) | |
# Shorthand for thresholds that only have one datapoint. | |
if 'dsname' in cfg: | |
cfg['dsnames'] = cfg['dsname'] | |
del(cfg['dsname']) | |
if isinstance(cfg['dsnames'], types.StringTypes): | |
cfg['dsnames'] = [cfg['dsnames']] | |
# Shorthand for datapoints that have the same name as their datasource. | |
for i, dsname in enumerate(cfg['dsnames']): | |
if '_' not in dsname: | |
cfg['dsnames'][i] = '_'.join((dsname, dsname)) | |
threshold_types = dict((y, x) for x, y in template.getThresholdClasses()) | |
type_ = threshold_types.get(cfg['type']) | |
if not type_: | |
die("'%s' is an invalid threshold type. Valid types: %s", | |
cfg['type'], ', '.join(threshold_types)) | |
threshold = template.manage_addRRDThreshold(id_, cfg['type']) | |
log_for(threshold, "adding threshold") | |
apply_properties(threshold, cfg, ignore=['type']) | |
def add_graph(template, id_, cfg): | |
graph = template.manage_addGraphDefinition(id_) | |
log_for(graph, "adding graph") | |
if 'graphpoints' in cfg: | |
log_for(graph, "adding {} graphpoints".format(len(cfg['graphpoints']))) | |
for graphpoint_id, graphpoint_cfg in cfg['graphpoints'].items(): | |
add_graphpoint(graph, graphpoint_id, graphpoint_cfg) | |
apply_properties(graph, cfg, ignore=['graphpoints']) | |
def add_graphpoint(graph, id_, cfg): | |
graphpoint = graph.createGraphPoint(DataPointGraphPoint, id_) | |
log_for(graphpoint, "adding graphpoint") | |
# Shorthand for datapoints that have the same name as their datasource. | |
if '_' not in cfg.get('dpName', '_'): | |
cfg['dpName'] = '_'.join((cfg['dpName'], cfg['dpName'])) | |
# Validate lineType. | |
if 'lineType' in cfg: | |
valid_linetypes = [x[1] for x in ComplexGraphPoint.lineTypeOptions] | |
if cfg['lineType'].upper() in valid_linetypes: | |
cfg['lineType'] = cfg['lineType'].upper() | |
else: | |
die("'%s' is not a valid graphpoint lineType. Valid lineTypes: %s", | |
cfg['lineType'], ', '.join(valid_linetypes)) | |
# Allow color to be specified by color_index instead of directly. This is | |
# useful when you want to keep the normal progression of colors, but need | |
# to add some DONTDRAW graphpoints for calculations. | |
if 'colorindex' in cfg: | |
try: | |
colorindex = int(cfg['colorindex']) % len(GraphPoint.colors) | |
except (TypeError, ValueError): | |
die("graphpoint colorindex must be numeric.") | |
cfg['color'] = GraphPoint.colors[colorindex].lstrip('#') | |
apply_properties(graphpoint, cfg, ignore=['colorindex', 'graphpoints']) | |
if __name__ == '__main__': | |
main() |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment