2018-09-18 22:43:20 +02:00
|
|
|
#!/usr/bin/env python3
|
2017-12-12 18:48:06 +01:00
|
|
|
"""
|
|
|
|
Script to parse StorCLI's JSON output and expose
|
|
|
|
MegaRAID health as Prometheus metrics.
|
2016-10-04 10:38:00 +02:00
|
|
|
|
2017-12-12 18:48:06 +01:00
|
|
|
Tested against StorCLI 'Ver 1.14.12 Nov 25, 2014'.
|
2016-10-04 10:38:00 +02:00
|
|
|
|
2017-12-12 18:48:06 +01:00
|
|
|
StorCLI reference manual:
|
|
|
|
http://docs.avagotech.com/docs/12352476
|
|
|
|
|
|
|
|
Advanced Software Options (ASO) not exposed as metrics currently.
|
|
|
|
|
|
|
|
JSON key abbreviations used by StorCLI are documented in the standard command
|
|
|
|
output, i.e. when you omit the trailing 'J' from the command.
|
|
|
|
"""
|
|
|
|
|
|
|
|
from __future__ import print_function
|
2016-10-04 10:38:00 +02:00
|
|
|
import argparse
|
|
|
|
import json
|
2017-12-12 18:48:06 +01:00
|
|
|
import os
|
2016-10-04 10:38:00 +02:00
|
|
|
import subprocess
|
2018-09-18 22:43:20 +02:00
|
|
|
import shlex
|
|
|
|
from dateutil.parser import parse
|
|
|
|
import collections
|
|
|
|
from enum import IntEnum
|
2016-10-04 10:38:00 +02:00
|
|
|
|
|
|
|
DESCRIPTION = """Parses StorCLI's JSON output and exposes MegaRAID health as
|
|
|
|
Prometheus metrics."""
|
2018-09-18 22:43:20 +02:00
|
|
|
VERSION = '0.0.2'
|
|
|
|
|
|
|
|
storcli_path = ''
|
|
|
|
metric_prefix = 'megaraid_'
|
|
|
|
metric_list = {}
|
|
|
|
metric_list = collections.defaultdict(list)
|
|
|
|
|
|
|
|
|
|
|
|
class VD_State(IntEnum):
|
|
|
|
Optl = 0 # Optimal
|
|
|
|
Dgrd = 1 # Degraded
|
|
|
|
Pdgd = 2 # Partially Degraded
|
|
|
|
OfLn = 3 # Offline
|
|
|
|
Rec = 4 # Recovery
|
|
|
|
Cac = 5 # CacheCade
|
2016-10-04 10:38:00 +02:00
|
|
|
|
|
|
|
|
|
|
|
def main(args):
|
2017-12-12 18:48:06 +01:00
|
|
|
""" main """
|
2018-09-18 22:43:20 +02:00
|
|
|
global storcli_path
|
|
|
|
storcli_path = args.storcli_path
|
|
|
|
data = json.loads(get_storcli_json('/cALL show all J'))
|
|
|
|
|
|
|
|
# All the information is collected underneath the Controllers key
|
|
|
|
data = data['Controllers']
|
|
|
|
|
|
|
|
# try:
|
|
|
|
# overview = status['Response Data']['System Overview']
|
|
|
|
# except KeyError:
|
|
|
|
# pass
|
|
|
|
|
|
|
|
for controller in data:
|
|
|
|
response = controller['Response Data']
|
|
|
|
if response['Version']['Driver Name'] == 'megaraid_sas':
|
|
|
|
handle_megaraid_controller(response)
|
|
|
|
elif response['Version']['Driver Name'] == 'mpt3sas':
|
|
|
|
handle_sas_controller(response)
|
|
|
|
|
|
|
|
# print_dict_to_exporter({'controller_info': [1]}, controller_info_list)
|
|
|
|
# print_dict_to_exporter({'virtual_disk_info': [1]}, vd_info_list)
|
|
|
|
# print_dict_to_exporter({'physical_disk_info': [1]}, pd_info_list)
|
|
|
|
# print_all_metrics(vd_metric_list)
|
|
|
|
print_all_metrics(metric_list)
|
|
|
|
|
|
|
|
|
|
|
|
def handle_sas_controller(response):
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
def handle_megaraid_controller(response):
|
|
|
|
controller_index = response['Basics']['Controller']
|
|
|
|
baselabel = 'controller="{}"'.format(controller_index)
|
|
|
|
|
|
|
|
controller_info_label = baselabel + ',model="{}",serial="{}",fwversion="{}"'.format(
|
|
|
|
response['Basics']['Model'],
|
|
|
|
response['Basics']['Serial Number'],
|
|
|
|
response['Version']['Firmware Version'],
|
|
|
|
)
|
|
|
|
add_metric('controller_info', controller_info_label, 1)
|
|
|
|
|
|
|
|
add_metric('battery_backup_healthy', baselabel, int(response['Status']['BBU Status'] == 0))
|
|
|
|
add_metric('degraded', baselabel, int(response['Status']['Controller Status'] == 'Degraded'))
|
|
|
|
add_metric('failed', baselabel, int(response['Status']['Controller Status'] == 'Failed'))
|
|
|
|
add_metric('healthy', baselabel, int(response['Status']['Controller Status'] == 'Optimal'))
|
|
|
|
add_metric('drive_groups', baselabel, response['Drive Groups'])
|
|
|
|
add_metric('virtual_drives', baselabel, response['Virtual Drives'])
|
|
|
|
add_metric('physical_drives', baselabel, response['Physical Drives'])
|
|
|
|
add_metric('ports', baselabel, response['HwCfg']['Backend Port Count'])
|
|
|
|
add_metric('scheduled_patrol_read', baselabel,
|
|
|
|
int('hrs' in response['Scheduled Tasks']['Patrol Read Reoccurrence']))
|
|
|
|
|
|
|
|
time_difference_seconds = -1
|
|
|
|
system_time = parse(response['Basics'].get('Current System Date/time'))
|
|
|
|
controller_time = parse(response['Basics'].get('Current Controller Date/Time'))
|
|
|
|
if system_time and controller_time:
|
|
|
|
time_difference_seconds = abs(system_time - controller_time).seconds
|
|
|
|
add_metric('time_difference', baselabel, time_difference_seconds)
|
|
|
|
|
|
|
|
for virtual_drive in response['VD LIST']:
|
|
|
|
vd_position = virtual_drive.get('DG/VD')
|
|
|
|
drive_group, volume_group = -1, -1
|
|
|
|
if vd_position:
|
|
|
|
drive_group = vd_position.split('/')[0]
|
|
|
|
volume_group = vd_position.split('/')[1]
|
|
|
|
vd_baselabel = 'controller="{}",DG="{}",VG="{}"'.format(controller_index, drive_group,
|
|
|
|
volume_group)
|
|
|
|
vd_info_label = vd_baselabel + ',name="{}",cache="{}",type="{}"'.format(
|
|
|
|
virtual_drive.get('Name'), virtual_drive.get('Cache'), virtual_drive.get('TYPE'))
|
|
|
|
add_metric('vd_info', vd_info_label, 1)
|
|
|
|
add_metric('vd_status', vd_baselabel, int(VD_State[virtual_drive.get('State')]))
|
|
|
|
|
|
|
|
if response['Physical Drives'] > 0:
|
|
|
|
data = json.loads(get_storcli_json('/cALL/eALL/sALL show all J'))
|
|
|
|
drive_info = data['Controllers'][controller_index]['Response Data']
|
|
|
|
for physical_drive in response['PD LIST']:
|
|
|
|
enclosure = physical_drive.get('EID:Slt').split(':')[0]
|
|
|
|
slot = physical_drive.get('EID:Slt').split(':')[1]
|
|
|
|
|
|
|
|
pd_baselabel = 'controller="{}",enclosure="{}",slot="{}"'.format(
|
|
|
|
controller_index, enclosure, slot)
|
|
|
|
pd_info_label = pd_baselabel + ',disk_id="{}",interface="{}",media="{}",model="{}"'.format(
|
|
|
|
physical_drive.get('DID'), physical_drive.get('Intf'), physical_drive.get('Med'),
|
|
|
|
physical_drive.get('Model').strip())
|
|
|
|
|
|
|
|
drive_identifier = 'Drive /c' + str(controller_index) + '/e' + str(enclosure) + '/s' + str(
|
|
|
|
slot)
|
|
|
|
try:
|
|
|
|
info = drive_info[drive_identifier + ' - Detailed Information']
|
|
|
|
state = info[drive_identifier + ' State']
|
|
|
|
attributes = info[drive_identifier + ' Device attributes']
|
|
|
|
settings = info[drive_identifier + ' Policies/Settings']
|
|
|
|
|
|
|
|
add_metric('pd_shield_counter', pd_baselabel, state['Shield Counter'])
|
|
|
|
add_metric('pd_media_errors_total', pd_baselabel, state['Media Error Count'])
|
|
|
|
add_metric('pd_other_errors_total', pd_baselabel, state['Other Error Count'])
|
|
|
|
add_metric('pd_predictive_errors_total', pd_baselabel,
|
|
|
|
state['Predictive Failure Count'])
|
|
|
|
add_metric('pd_smart_alerted', pd_baselabel,
|
|
|
|
int(state['S.M.A.R.T alert flagged by drive'] == 'Yes'))
|
|
|
|
add_metric('pd_link_speed_gbps', pd_baselabel, attributes['Link Speed'].split('.')[0])
|
|
|
|
add_metric('pd_device_speed_gbps', pd_baselabel,
|
|
|
|
attributes['Device Speed'].split('.')[0])
|
|
|
|
add_metric('pd_commissioned_spare', pd_baselabel,
|
|
|
|
int(settings['Commissioned Spare'] == 'Yes'))
|
|
|
|
add_metric('pd_emergency_spare', pd_baselabel,
|
|
|
|
int(settings['Emergency Spare'] == 'Yes'))
|
|
|
|
pd_info_label += ',firmware="{}"'.format(attributes['Firmware Revision'])
|
|
|
|
except KeyError:
|
|
|
|
pass
|
|
|
|
add_metric('pd_info', pd_info_label, 1)
|
|
|
|
|
|
|
|
|
|
|
|
def add_metric(name, labels, value):
|
|
|
|
global metric_list
|
|
|
|
metric_list[name].append({
|
|
|
|
'labels': labels,
|
|
|
|
'value': value,
|
|
|
|
})
|
|
|
|
|
|
|
|
|
|
|
|
def print_all_metrics(metrics):
|
|
|
|
for metric, measurements in metrics.items():
|
|
|
|
print('# HELP {}{} MegaRAID {}'.format(metric_prefix, metric, metric.replace('_', ' ')))
|
|
|
|
print('# TYPE {}{} gauge'.format(metric_prefix, metric))
|
|
|
|
for measurement in measurements:
|
|
|
|
print('{}{}{} {}'.format(metric_prefix, metric, '{' + measurement['labels'] + '}',
|
|
|
|
measurement['value']))
|
|
|
|
|
|
|
|
|
|
|
|
def get_storcli_json(storcli_args):
|
2017-12-12 18:48:06 +01:00
|
|
|
"""Get storcli output in JSON format."""
|
2018-09-18 22:43:20 +02:00
|
|
|
# Check if storcli is installed and executable
|
|
|
|
if not (os.path.isfile(storcli_path) and os.access(storcli_path, os.X_OK)):
|
|
|
|
SystemExit(1)
|
|
|
|
storcli_cmd = shlex.split(storcli_path + ' ' + storcli_args)
|
|
|
|
proc = subprocess.Popen(
|
|
|
|
storcli_cmd, shell=False, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
|
|
output_json = proc.communicate()[0]
|
|
|
|
|
|
|
|
return output_json.decode("utf-8")
|
2017-12-12 18:48:06 +01:00
|
|
|
|
2016-10-04 10:38:00 +02:00
|
|
|
|
|
|
|
if __name__ == "__main__":
|
2018-09-18 22:43:20 +02:00
|
|
|
PARSER = argparse.ArgumentParser(
|
|
|
|
description=DESCRIPTION, formatter_class=argparse.ArgumentDefaultsHelpFormatter)
|
|
|
|
PARSER.add_argument(
|
|
|
|
'--storcli_path', default='/opt/MegaRAID/storcli/storcli64', help='path to StorCLi binary')
|
|
|
|
PARSER.add_argument('--version', action='version', version='%(prog)s {}'.format(VERSION))
|
2017-12-12 18:48:06 +01:00
|
|
|
ARGS = PARSER.parse_args()
|
2016-10-04 10:38:00 +02:00
|
|
|
|
2017-12-12 18:48:06 +01:00
|
|
|
main(ARGS)
|