210 lines
6.9 KiB
Python
210 lines
6.9 KiB
Python
#!/usr/bin/env python2
|
|
# ver 1.1 CL2019
|
|
|
|
from flask import Flask
|
|
from flask import Response
|
|
import requests
|
|
import json
|
|
import time
|
|
from gevent.wsgi import WSGIServer
|
|
#
|
|
import logging
|
|
import logging.handlers
|
|
#
|
|
import hx_creds
|
|
|
|
'''
|
|
the hx_creds.py looks like this:
|
|
|
|
hosts=[{'host':'10.1.1.1', 'username':'local/root', 'password':'*******'},{'host':'10.1.1.2', 'username':'local/root', 'password':'****'}]
|
|
|
|
|
|
'''
|
|
#
|
|
server_IP ='10.100.253.13'
|
|
server_port = '8082'
|
|
|
|
# Logging config
|
|
logFile="hx_stats_%s_%s.log"%(server_IP,server_port)
|
|
logCount=4
|
|
logBytes=1048576
|
|
|
|
# suppress the unverified request messages (when using self-signed certificates)
|
|
requests.packages.urllib3.disable_warnings()
|
|
|
|
# cache the credentials, if you keep requesting this you will hit th 256 open session/user limit
|
|
tokens = {}
|
|
|
|
app = Flask('HX Stats')
|
|
|
|
# gets called via the http://your_server_ip:port/metrics
|
|
@app.route('/metrics')
|
|
def get_stats():
|
|
results =''
|
|
for host in hx_creds.hosts:
|
|
logger.info("----------- Processing Host: %s -----------"%host['host'])
|
|
# get auth token (either cached or new one)
|
|
authdata = get_auth(host['host'], host['username'], host['password'])
|
|
if not authdata:
|
|
logger.info("Missing token, skipping host: "+host['host'])
|
|
continue
|
|
url = "https://"+host['host']
|
|
|
|
# uri for throughput data with "last 5 min" filter
|
|
uri_MBps = '/render?target=stats.counters.scvmclient.allhosts.nfsBytesRead.cluster.rate&target=stats.counters.scvmclient.allhosts.nfsBytesWritten.cluster.rate&format=json&from=-5min'
|
|
# Get throughput data
|
|
MBps_data = get_stats(authdata,url+uri_MBps)
|
|
if MBps_data:
|
|
try:
|
|
MBps_Read=round(MBps_data[0]['datapoints'][-2][0],3)
|
|
MBps_Write=round(MBps_data[1]['datapoints'][-2][0],3)
|
|
logger.info("Got MBps info")
|
|
|
|
# build the results
|
|
results += 'MBps_Read{host="%s"} %s\n'%(host['host'],str(MBps_Read))
|
|
results += 'MBps_Write{host="%s"} %s\n'%(host['host'],str(MBps_Write))
|
|
|
|
except Exception as e:
|
|
logger.error(e)
|
|
logger.error("Couldn't parse returned throughput data")
|
|
pass
|
|
|
|
# url to get the IOPS data
|
|
uri_IOPS = '/render?target=stats.counters.scvmclient.allhosts.nfsReads.cluster.rate&target=stats.counters.scvmclient.allhosts.nfsWrites.cluster.rate&format=json&from=-5min'
|
|
# get IOPS data
|
|
IOPS_data = get_stats(authdata,url+uri_IOPS)
|
|
if IOPS_data:
|
|
try:
|
|
IOPS_Read=round(IOPS_data[0]['datapoints'][-2][0],3)
|
|
IOPS_Write=round(IOPS_data[1]['datapoints'][-2][0],3)
|
|
logger.info("Got IOPS info")
|
|
|
|
# build the results
|
|
results += 'IOPS_Read{host="%s"} %s\n'%(host['host'],str(IOPS_Read))
|
|
results += 'IOPS_Write{host="%s"} %s\n'%(host['host'],str(IOPS_Write))
|
|
|
|
except Exception as e:
|
|
logger.error(e)
|
|
logger.error("Couldn't parse returned IOPS data")
|
|
pass
|
|
|
|
# url to get Latency data
|
|
uri_Lat ='/render?target=divideSeries(stats.timers.scvmclient.allhosts.nfsReadLatency.cluster.total%2Cstats.counters.scvmclient.allhosts.nfsReads.cluster.count)&target=divideSeries(stats.timers.scvmclient.allhosts.nfsWriteLatency.cluster.total%2Cstats.counters.scvmclient.allhosts.nfsWrites.cluster.count)&format=json&from=-5min'
|
|
|
|
# get latency data
|
|
Lat_data = get_stats(authdata,url+uri_Lat)
|
|
if Lat_data:
|
|
try:
|
|
Lat_Read=round(Lat_data[0]['datapoints'][-2][0],3)
|
|
Lat_Write=round(Lat_data[1]['datapoints'][-2][0],3)
|
|
logger.info("Got Latency info")
|
|
|
|
# build the results
|
|
results += 'Lat_Read{host="%s"} %s\n'%(host['host'],str(Lat_Read))
|
|
results += 'Lat_Write{host="%s"} %s\n'%(host['host'],str(Lat_Write))
|
|
|
|
except Exception as e:
|
|
logger.error(e)
|
|
logger.error("Couldn't parse returned latency data")
|
|
pass
|
|
#
|
|
# When processing data I'm taking one before last record ([-2]), as sometimes the last record is None
|
|
#
|
|
# return the results to the caller
|
|
logger.info("----------- Finished -----------")
|
|
return Response(results, mimetype='text/plain')
|
|
|
|
#
|
|
# Returns Authentication token
|
|
#
|
|
def get_auth(host, username, password):
|
|
|
|
logger.info("Trying to auth "+host)
|
|
|
|
global tokens
|
|
|
|
headers={'content-type':'application/json'}
|
|
payload = {
|
|
"username": username,
|
|
"password": password,
|
|
"client_id": "HxGuiClient",
|
|
"client_secret": "Sunnyvale",
|
|
"redirect_uri": "http://"+host
|
|
}
|
|
|
|
|
|
if tokens.get(host):
|
|
# looks like we have token cached already
|
|
# let's check if it's valid
|
|
|
|
payload = {
|
|
"access_token": tokens.get(host)['access_token'],
|
|
"scope": "READ",
|
|
"token_type": tokens.get(host)['token_type']
|
|
}
|
|
try:
|
|
#validating token
|
|
url = 'https://%s/aaa/v1/validate'%host
|
|
response = requests.post(url,headers=headers,data=json.dumps(payload),verify=False,timeout=4)
|
|
if response.status_code == 200:
|
|
logger.info("Re-using existing auth token")
|
|
return tokens.get(host)
|
|
logger.error("Failed to validate existing token "+response.content)
|
|
except Exception as e:
|
|
logger.error("Post for token validate failed \n"+str(e))
|
|
|
|
# this happens if no cached token found, or existing token was invalid
|
|
url = 'https://%s/aaa/v1/auth?grant_type=password'%host
|
|
payload = {
|
|
"username": username,
|
|
"password": password,
|
|
"client_id": "HxGuiClient",
|
|
"client_secret": "Sunnyvale",
|
|
"redirect_uri": "http://"+host
|
|
}
|
|
|
|
try:
|
|
response = requests.post(url,headers=headers,data=json.dumps(payload),verify=False,timeout=4)
|
|
if response.status_code == 201:
|
|
if response.json().get('access_token'):
|
|
tokens[host]=response.json()
|
|
logger.info("Got token ok")
|
|
return response.json()
|
|
logger.error("Failed get a token "+response.content)
|
|
return None
|
|
except Exception as e:
|
|
logger.error("Post for token get failed \n"+str(e))
|
|
return None
|
|
|
|
#
|
|
# calls HX API
|
|
#
|
|
def get_stats(authdata, url):
|
|
logger.info("call for get_stats")
|
|
try:
|
|
headers = {'Authorization': authdata['token_type'] + ' ' + authdata['access_token'],'Connection':'close'}
|
|
response = requests.get(url,headers=headers,verify=False,timeout=4)
|
|
if response.status_code == 200:
|
|
logger.info("Got data ok")
|
|
return response.json()
|
|
logger.error("Failed to get data "+response.content)
|
|
return None
|
|
except Exception as e:
|
|
logger.error("Post for data failed \n"+str(e))
|
|
return None
|
|
|
|
if __name__ == '__main__':
|
|
print "Service Started"
|
|
# Enable logging
|
|
logger = logging.getLogger("HX-Stats")
|
|
logger.setLevel(logging.DEBUG)
|
|
handler = logging.handlers.RotatingFileHandler(logFile, maxBytes=logBytes, backupCount=logCount)
|
|
formatter = logging.Formatter('%(asctime)s - %(levelname)s - %(message)s')
|
|
handler.setFormatter(formatter)
|
|
logger.addHandler(handler)
|
|
logger.info("-"*25)
|
|
logger.info("HX Stats script started")
|
|
|
|
http_server = WSGIServer((server_IP, int(server_port)), app, log = logger)
|
|
http_server.serve_forever()
|