2019-06-14 16:15:07 +02:00
|
|
|
#!/usr/bin/env python3
|
|
|
|
import os
|
|
|
|
import sys
|
|
|
|
import stat
|
|
|
|
import time
|
2019-06-17 11:41:14 +02:00
|
|
|
import signal
|
2019-06-14 16:15:07 +02:00
|
|
|
import functools
|
|
|
|
import configparser
|
|
|
|
from pprint import pprint
|
2019-06-17 11:41:14 +02:00
|
|
|
import subprocess
|
|
|
|
import diagnostic_util
|
2019-06-17 09:50:26 +02:00
|
|
|
try:
|
|
|
|
import redis
|
|
|
|
import zmq
|
|
|
|
import json
|
|
|
|
import flask
|
2019-06-17 14:42:41 +02:00
|
|
|
import requests
|
2019-06-17 09:50:26 +02:00
|
|
|
from halo import Halo
|
|
|
|
except ModuleNotFoundError as e:
|
|
|
|
print('Dependency not met. Either not in a virtualenv or dependency not installed.')
|
|
|
|
print(f'- Error: {e}')
|
|
|
|
sys.exit(1)
|
2019-06-14 16:15:07 +02:00
|
|
|
|
|
|
|
'''
|
|
|
|
Steps:
|
2019-06-17 09:50:26 +02:00
|
|
|
- check if dependencies exists
|
2019-06-14 16:15:07 +02:00
|
|
|
- check if virtualenv exists
|
|
|
|
- check if configuration is update-to-date
|
|
|
|
- check file permission
|
|
|
|
- check if redis is running and responding
|
|
|
|
- check if able to connect to zmq
|
|
|
|
- check zmq_dispatcher processing queue
|
|
|
|
- check queue status: being filled up / being filled down
|
|
|
|
- check if subscriber responding
|
|
|
|
- check if dispatcher responding
|
|
|
|
- check if server listening
|
|
|
|
- check log static endpoint
|
|
|
|
- check log dynamic endpoint
|
|
|
|
'''
|
|
|
|
|
2019-06-17 14:42:41 +02:00
|
|
|
HOST = 'http://127.0.0.1'
|
|
|
|
PORT = 8001 # overriden by configuration file
|
2019-06-14 16:15:07 +02:00
|
|
|
configuration_file = {}
|
2019-06-17 11:41:14 +02:00
|
|
|
pgrep_subscriber_output = ''
|
|
|
|
pgrep_dispatcher_output = ''
|
|
|
|
|
|
|
|
|
2019-06-17 14:47:51 +02:00
|
|
|
signal.signal(signal.SIGALRM, diagnostic_util.timeout_handler)
|
2019-06-14 16:15:07 +02:00
|
|
|
|
|
|
|
|
|
|
|
def humanize(name, isResult=False):
|
|
|
|
words = name.split('_')
|
|
|
|
if isResult:
|
|
|
|
words = words[1:]
|
|
|
|
words[0] = words[0][0].upper() + words[0][1:]
|
|
|
|
else:
|
|
|
|
words[0] = words[0][0].upper() + words[0][1:] + 'ing'
|
|
|
|
return ' '.join(words)
|
|
|
|
|
|
|
|
|
|
|
|
def add_spinner(_func=None, name='dots'):
|
|
|
|
def decorator_add_spinner(func):
|
|
|
|
@functools.wraps(func)
|
|
|
|
def wrapper_add_spinner(*args, **kwargs):
|
|
|
|
human_func_name = humanize(func.__name__)
|
|
|
|
human_func_result = humanize(func.__name__, isResult=True)
|
|
|
|
flag_skip = False
|
|
|
|
|
|
|
|
with Halo(text=human_func_name, spinner=name) as spinner:
|
|
|
|
result = func(spinner, *args, **kwargs)
|
|
|
|
if isinstance(result, tuple):
|
|
|
|
status, output = result
|
|
|
|
elif isinstance(result, list):
|
|
|
|
status = result[0]
|
|
|
|
output = result[1]
|
|
|
|
elif isinstance(result, bool):
|
|
|
|
status = result
|
|
|
|
output = None
|
|
|
|
else:
|
|
|
|
status = False
|
|
|
|
flag_skip = True
|
|
|
|
spinner.fail(f'{human_func_name} - Function return unexpected result: {str(result)}')
|
|
|
|
|
|
|
|
if not flag_skip:
|
|
|
|
text = human_func_result
|
|
|
|
if output is not None and len(output) > 0:
|
|
|
|
text += f': {output}'
|
|
|
|
|
|
|
|
if isinstance(status, bool) and status:
|
|
|
|
spinner.succeed(text)
|
|
|
|
elif isinstance(status, bool) and not status:
|
|
|
|
spinner.fail(text)
|
|
|
|
else:
|
2019-06-17 09:50:26 +02:00
|
|
|
if status == 'info':
|
|
|
|
spinner.info(text)
|
|
|
|
else:
|
|
|
|
spinner.warn(text)
|
2019-06-14 16:15:07 +02:00
|
|
|
return status
|
|
|
|
return wrapper_add_spinner
|
|
|
|
|
|
|
|
if _func is None:
|
|
|
|
return decorator_add_spinner
|
|
|
|
else:
|
|
|
|
return decorator_add_spinner(_func)
|
|
|
|
|
|
|
|
|
|
|
|
@add_spinner
|
2019-06-18 12:30:07 +02:00
|
|
|
def check_virtual_environment_and_packages(spinner):
|
2019-06-14 16:15:07 +02:00
|
|
|
result = os.environ.get('VIRTUAL_ENV')
|
|
|
|
if result is None:
|
|
|
|
return (False, 'This diagnostic tool should be started inside a virtual environment.')
|
|
|
|
else:
|
2019-06-18 12:30:07 +02:00
|
|
|
if redis.__version__.startswith('2'):
|
2019-06-20 14:16:25 +02:00
|
|
|
return (False, f'''Redis python client have version {redis.__version__}. Version 3.x required.
|
|
|
|
\t➥ [inside virtualenv] pip3 install -U redis''')
|
2019-06-18 12:30:07 +02:00
|
|
|
else:
|
|
|
|
return (True, '')
|
2019-06-14 16:15:07 +02:00
|
|
|
|
|
|
|
|
|
|
|
@add_spinner
|
|
|
|
def check_configuration(spinner):
|
2019-06-17 14:42:41 +02:00
|
|
|
global configuration_file, port
|
2019-06-14 16:15:07 +02:00
|
|
|
configfile = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'config/config.cfg')
|
|
|
|
cfg = configparser.ConfigParser()
|
|
|
|
cfg.read(configfile)
|
|
|
|
configuration_file = cfg
|
|
|
|
cfg = {s: dict(cfg.items(s)) for s in cfg.sections()}
|
2019-06-20 15:40:14 +02:00
|
|
|
configfile_default = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'config/config.cfg.default')
|
2019-06-14 16:15:07 +02:00
|
|
|
cfg_default = configparser.ConfigParser()
|
2019-06-20 15:40:14 +02:00
|
|
|
cfg_default.read(configfile_default)
|
2019-06-14 16:15:07 +02:00
|
|
|
cfg_default = {s: dict(cfg_default.items(s)) for s in cfg_default.sections()}
|
|
|
|
|
|
|
|
# Check if all fields from config.default exists in config
|
|
|
|
result, faulties = diagnostic_util.dict_compare(cfg_default, cfg)
|
|
|
|
if result:
|
2019-06-17 14:42:41 +02:00
|
|
|
port = configuration_file.get("Server", "port")
|
2019-06-14 16:15:07 +02:00
|
|
|
return (True, '')
|
|
|
|
else:
|
2019-06-20 15:40:14 +02:00
|
|
|
return_text = '''Configuration incomplete.
|
|
|
|
\tUpdate your configuration file `config.cfg`.\n\t➥ Faulty fields:\n'''
|
|
|
|
for field_name in faulties:
|
|
|
|
return_text += f'\t\t- {field_name}\n'
|
|
|
|
return (False, return_text)
|
2019-06-14 16:15:07 +02:00
|
|
|
|
|
|
|
|
|
|
|
@add_spinner(name='dot')
|
|
|
|
def check_file_permission(spinner):
|
|
|
|
max_mind_database_path = configuration_file.get('RedisMap', 'pathmaxminddb')
|
|
|
|
st = os.stat(max_mind_database_path)
|
|
|
|
all_read_perm = bool(st.st_mode & stat.S_IROTH) # FIXME: permission may be changed
|
|
|
|
if all_read_perm:
|
|
|
|
return (True, '')
|
|
|
|
else:
|
|
|
|
return (False, 'Maxmin GeoDB might have incorrect read file permission')
|
|
|
|
|
|
|
|
|
|
|
|
@add_spinner
|
|
|
|
def check_redis(spinner):
|
|
|
|
redis_server = redis.StrictRedis(
|
|
|
|
host=configuration_file.get('RedisGlobal', 'host'),
|
|
|
|
port=configuration_file.getint('RedisGlobal', 'port'),
|
|
|
|
db=configuration_file.getint('RedisLog', 'db'))
|
|
|
|
if redis_server.ping():
|
|
|
|
return (True, '')
|
|
|
|
else:
|
|
|
|
return (False, '''Can\'t reach Redis server.
|
|
|
|
\t➥ Make sure it is running and adapt your configuration accordingly''')
|
|
|
|
|
|
|
|
|
|
|
|
@add_spinner
|
|
|
|
def check_zmq(spinner):
|
2019-06-17 15:53:03 +02:00
|
|
|
timeout = 15
|
2019-06-14 16:15:07 +02:00
|
|
|
context = zmq.Context()
|
2019-06-21 16:19:57 +02:00
|
|
|
misp_instances = json.loads(configuration_file.get('RedisGlobal', 'misp_instances'))
|
2019-06-21 15:55:03 +02:00
|
|
|
instances_status = {}
|
|
|
|
for misp_instance in misp_instances:
|
2019-06-21 16:19:57 +02:00
|
|
|
socket = context.socket(zmq.SUB)
|
2019-06-21 15:55:03 +02:00
|
|
|
socket.connect(misp_instance.get('zmq'))
|
|
|
|
socket.setsockopt_string(zmq.SUBSCRIBE, '')
|
|
|
|
poller = zmq.Poller()
|
|
|
|
|
2019-06-21 16:19:57 +02:00
|
|
|
flag_skip = False
|
2019-06-21 15:55:03 +02:00
|
|
|
start_time = time.time()
|
|
|
|
poller.register(socket, zmq.POLLIN)
|
|
|
|
for t in range(1, timeout+1):
|
|
|
|
socks = dict(poller.poll(timeout=1*1000))
|
|
|
|
if len(socks) > 0:
|
|
|
|
if socket in socks and socks[socket] == zmq.POLLIN:
|
|
|
|
rcv_string = socket.recv()
|
|
|
|
if rcv_string.startswith(b'misp_json'):
|
|
|
|
instances_status[misp_instance.get('name')] = True
|
2019-06-21 16:19:57 +02:00
|
|
|
flag_skip = True
|
|
|
|
break
|
2019-06-21 15:55:03 +02:00
|
|
|
else:
|
2019-06-21 16:19:57 +02:00
|
|
|
spinner.text = f'checking zmq of {misp_instance.get("name")} - elapsed time: {int(time.time() - start_time)}s'
|
|
|
|
if not flag_skip:
|
|
|
|
instances_status[misp_instance.get('name')] = False
|
2019-06-21 15:55:03 +02:00
|
|
|
|
|
|
|
results = [s for n, s in instances_status.items()]
|
|
|
|
if all(results):
|
|
|
|
return (True, '')
|
|
|
|
elif any(results):
|
2019-06-21 16:19:57 +02:00
|
|
|
return_text = 'Connection to ZMQ stream(s) failed.\n'
|
|
|
|
for name, status in instances_status.items():
|
|
|
|
return_text += f'\t➥ {name}: {"success" if status else "failed"}\n'
|
2019-06-21 15:55:03 +02:00
|
|
|
return (True, return_text)
|
2019-06-14 16:15:07 +02:00
|
|
|
else:
|
2019-06-21 16:19:57 +02:00
|
|
|
return (False, '''Can\'t connect to the ZMQ stream(s).
|
2019-06-14 16:15:07 +02:00
|
|
|
\t➥ Make sure the MISP ZMQ is running: `/servers/serverSettings/diagnostics`
|
|
|
|
\t➥ Make sure your network infrastucture allows you to connect to the ZMQ''')
|
|
|
|
|
|
|
|
|
2019-06-17 11:41:14 +02:00
|
|
|
@add_spinner
|
|
|
|
def check_processes_status(spinner):
|
|
|
|
global pgrep_subscriber_output, pgrep_dispatcher_output
|
|
|
|
response = subprocess.check_output(
|
|
|
|
["pgrep", "-laf", "zmq_"],
|
|
|
|
universal_newlines=True
|
|
|
|
)
|
|
|
|
for line in response.splitlines():
|
2019-06-21 16:19:57 +02:00
|
|
|
lines = line.split(' ', maxsplit=1)
|
|
|
|
pid, p_name = lines
|
2019-06-19 11:30:49 +02:00
|
|
|
|
2019-06-17 11:41:14 +02:00
|
|
|
if 'zmq_subscriber.py' in p_name:
|
|
|
|
pgrep_subscriber_output = line
|
|
|
|
elif 'zmq_dispatcher.py' in p_name:
|
|
|
|
pgrep_dispatcher_output = line
|
|
|
|
|
|
|
|
if len(pgrep_subscriber_output) == 0:
|
|
|
|
return (False, 'zmq_subscriber is not running')
|
|
|
|
elif len(pgrep_dispatcher_output) == 0:
|
|
|
|
return (False, 'zmq_dispatcher is not running')
|
|
|
|
else:
|
|
|
|
return (True, 'Both processes are running')
|
|
|
|
|
|
|
|
|
|
|
|
@add_spinner
|
|
|
|
def check_subscriber_status(spinner):
|
|
|
|
global pgrep_subscriber_output
|
|
|
|
pool = redis.ConnectionPool(
|
|
|
|
host=configuration_file.get('RedisGlobal', 'host'),
|
|
|
|
port=configuration_file.getint('RedisGlobal', 'port'),
|
|
|
|
db=configuration_file.getint('RedisLIST', 'db'),
|
|
|
|
decode_responses=True)
|
|
|
|
monitor = diagnostic_util.Monitor(pool)
|
|
|
|
commands = monitor.monitor()
|
|
|
|
|
2019-06-17 15:53:03 +02:00
|
|
|
start_time = time.time()
|
2019-06-17 11:41:14 +02:00
|
|
|
signal.alarm(15)
|
|
|
|
try:
|
|
|
|
for i, c in enumerate(commands):
|
|
|
|
if i == 0: # Skip 'OK'
|
|
|
|
continue
|
|
|
|
split = c.split()
|
|
|
|
try:
|
|
|
|
action = split[3]
|
|
|
|
target = split[4]
|
|
|
|
except IndexError:
|
|
|
|
pass
|
|
|
|
if action == '"LPUSH"' and target == f'\"{configuration_file.get("RedisLIST", "listName")}\"':
|
|
|
|
signal.alarm(0)
|
|
|
|
break
|
2019-06-17 15:53:03 +02:00
|
|
|
else:
|
|
|
|
spinner.text = f'Checking subscriber status - elapsed time: {int(time.time() - start_time)}s'
|
2019-06-17 14:47:51 +02:00
|
|
|
except diagnostic_util.TimeoutException:
|
2019-06-17 11:41:14 +02:00
|
|
|
return_text = f'''zmq_subscriber seems not to be working.
|
|
|
|
\t➥ Consider restarting it: {pgrep_subscriber_output}'''
|
|
|
|
return (False, return_text)
|
|
|
|
return (True, 'subscriber is running and populating the buffer')
|
|
|
|
|
|
|
|
|
2019-06-14 16:15:07 +02:00
|
|
|
@add_spinner
|
|
|
|
def check_buffer_queue(spinner):
|
|
|
|
redis_server = redis.StrictRedis(
|
|
|
|
host=configuration_file.get('RedisGlobal', 'host'),
|
|
|
|
port=configuration_file.getint('RedisGlobal', 'port'),
|
|
|
|
db=configuration_file.getint('RedisLIST', 'db'))
|
2019-06-17 09:50:26 +02:00
|
|
|
warning_threshold = 100
|
2019-06-14 16:15:07 +02:00
|
|
|
elements_in_list = redis_server.llen(configuration_file.get('RedisLIST', 'listName'))
|
2019-06-17 09:50:26 +02:00
|
|
|
return_status = 'warning' if elements_in_list > warning_threshold else ('info' if elements_in_list > 0 else True)
|
|
|
|
return_text = f'Currently {elements_in_list} items in the buffer'
|
|
|
|
return (return_status, return_text)
|
|
|
|
|
|
|
|
|
|
|
|
@add_spinner
|
|
|
|
def check_buffer_change_rate(spinner):
|
|
|
|
redis_server = redis.StrictRedis(
|
|
|
|
host=configuration_file.get('RedisGlobal', 'host'),
|
|
|
|
port=configuration_file.getint('RedisGlobal', 'port'),
|
|
|
|
db=configuration_file.getint('RedisLIST', 'db'))
|
|
|
|
|
|
|
|
time_slept = 0
|
|
|
|
sleep_duration = 0.001
|
|
|
|
sleep_max = 10.0
|
|
|
|
refresh_frequency = 1.0
|
|
|
|
next_refresh = 0
|
|
|
|
change_increase = 0
|
|
|
|
change_decrease = 0
|
|
|
|
elements_in_list_prev = 0
|
|
|
|
elements_in_list = int(redis_server.llen(configuration_file.get('RedisLIST', 'listName')))
|
|
|
|
elements_in_inlist_init = elements_in_list
|
|
|
|
consecutive_no_rate_change = 0
|
|
|
|
while True:
|
|
|
|
elements_in_list_prev = elements_in_list
|
|
|
|
elements_in_list = int(redis_server.llen(configuration_file.get('RedisLIST', 'listName')))
|
|
|
|
change_increase += elements_in_list - elements_in_list_prev if elements_in_list - elements_in_list_prev > 0 else 0
|
|
|
|
change_decrease += elements_in_list_prev - elements_in_list if elements_in_list_prev - elements_in_list > 0 else 0
|
|
|
|
|
|
|
|
if next_refresh < time_slept:
|
|
|
|
next_refresh = time_slept + refresh_frequency
|
|
|
|
change_rate_text = f'↑ {change_increase}/sec\t↓ {change_decrease}/sec'
|
2019-06-17 15:53:03 +02:00
|
|
|
spinner.text = f'Buffer: {elements_in_list}\t{change_rate_text}'
|
2019-06-17 09:50:26 +02:00
|
|
|
|
|
|
|
if consecutive_no_rate_change == 3:
|
|
|
|
time_slept = sleep_max
|
|
|
|
if elements_in_list == 0:
|
|
|
|
consecutive_no_rate_change += 1
|
|
|
|
else:
|
|
|
|
consecutive_no_rate_change = 0
|
|
|
|
change_increase = 0
|
|
|
|
change_decrease = 0
|
|
|
|
|
|
|
|
if time_slept >= sleep_max:
|
2019-06-17 16:16:05 +02:00
|
|
|
return_flag = elements_in_list == 0 or (elements_in_list < elements_in_inlist_init or elements_in_list < 2)
|
2019-06-17 09:50:26 +02:00
|
|
|
return_text = f'Buffer is consumed {"faster" if return_flag else "slower" } than being populated'
|
|
|
|
break
|
|
|
|
|
|
|
|
time.sleep(sleep_duration)
|
|
|
|
time_slept += sleep_duration
|
|
|
|
elements_in_inlist_final = int(redis_server.llen(configuration_file.get('RedisLIST', 'listName')))
|
|
|
|
return (return_flag, return_text)
|
2019-06-14 16:15:07 +02:00
|
|
|
|
|
|
|
|
|
|
|
@add_spinner
|
2019-06-17 09:50:26 +02:00
|
|
|
def check_dispatcher_status(spinner):
|
2019-06-14 16:59:00 +02:00
|
|
|
redis_server = redis.StrictRedis(
|
|
|
|
host=configuration_file.get('RedisGlobal', 'host'),
|
|
|
|
port=configuration_file.getint('RedisGlobal', 'port'),
|
|
|
|
db=configuration_file.getint('RedisLIST', 'db'))
|
2019-06-17 09:50:26 +02:00
|
|
|
content = {'content': time.time()}
|
2019-06-14 16:59:00 +02:00
|
|
|
redis_server.rpush(configuration_file.get('RedisLIST', 'listName'),
|
2019-06-17 09:50:26 +02:00
|
|
|
json.dumps({'zmq_name': 'diagnostic_channel', 'content': 'diagnostic_channel ' + json.dumps(content)})
|
2019-06-14 16:59:00 +02:00
|
|
|
)
|
2019-06-17 09:50:26 +02:00
|
|
|
|
|
|
|
return_flag = False
|
|
|
|
return_text = ''
|
|
|
|
time_slept = 0
|
|
|
|
sleep_duration = 0.2
|
|
|
|
sleep_max = 10.0
|
|
|
|
redis_server.delete('diagnostic_tool_response')
|
2019-06-14 16:59:00 +02:00
|
|
|
while True:
|
|
|
|
reply = redis_server.get('diagnostic_tool_response')
|
|
|
|
elements_in_list = redis_server.llen(configuration_file.get('RedisLIST', 'listName'))
|
|
|
|
if reply is None:
|
2019-06-17 09:50:26 +02:00
|
|
|
if time_slept >= sleep_max:
|
|
|
|
return_flag = False
|
2019-06-17 16:16:05 +02:00
|
|
|
return_text = f'zmq_dispatcher did not respond in the given time ({int(sleep_max)}s)'
|
|
|
|
if len(pgrep_dispatcher_output) > 0:
|
|
|
|
return_text += f'\n\t➥ Consider restarting it: {pgrep_dispatcher_output}'
|
|
|
|
else:
|
|
|
|
return_text += '\n\t➥ Consider starting it'
|
2019-06-17 09:50:26 +02:00
|
|
|
break
|
|
|
|
time.sleep(sleep_duration)
|
2019-06-17 15:09:08 +02:00
|
|
|
spinner.text = f'Dispatcher status: No response yet'
|
2019-06-17 09:50:26 +02:00
|
|
|
time_slept += sleep_duration
|
2019-06-14 16:59:00 +02:00
|
|
|
else:
|
2019-06-17 09:50:26 +02:00
|
|
|
return_flag = True
|
|
|
|
return_text = f'Took {float(reply):.2f}s to complete'
|
|
|
|
break
|
2019-06-14 16:15:07 +02:00
|
|
|
|
2019-06-17 09:50:26 +02:00
|
|
|
return (return_flag, return_text)
|
2019-06-14 16:15:07 +02:00
|
|
|
|
|
|
|
|
|
|
|
@add_spinner
|
|
|
|
def check_server_listening(spinner):
|
2019-06-17 14:42:41 +02:00
|
|
|
url = f'{HOST}:{PORT}/_get_log_head'
|
|
|
|
spinner.text = f'Trying to connect to {url}'
|
2019-06-17 15:09:08 +02:00
|
|
|
try:
|
|
|
|
r = requests.get(url)
|
|
|
|
except requests.exceptions.ConnectionError:
|
|
|
|
return (False, f'Can\'t connect to {url}')
|
2019-06-17 14:42:41 +02:00
|
|
|
return (
|
|
|
|
r.status_code == 200,
|
2019-06-17 15:09:08 +02:00
|
|
|
f'{url} {"not " if r.status_code != 200 else ""}reached. Status code [{r.status_code}]'
|
2019-06-17 14:42:41 +02:00
|
|
|
)
|
2019-06-14 16:15:07 +02:00
|
|
|
|
|
|
|
|
|
|
|
@add_spinner
|
2019-06-17 14:42:41 +02:00
|
|
|
def check_server_dynamic_enpoint(spinner):
|
|
|
|
sleep_max = 15
|
|
|
|
start_time = time.time()
|
|
|
|
url = f'{HOST}:{PORT}/_logs'
|
|
|
|
p = subprocess.Popen(
|
|
|
|
['curl', '-sfN', '--header', 'Accept: text/event-stream', url],
|
|
|
|
stdout=subprocess.PIPE,
|
|
|
|
bufsize=1)
|
|
|
|
signal.alarm(sleep_max)
|
2019-06-17 15:53:03 +02:00
|
|
|
return_flag = False
|
|
|
|
return_text = f'Dynamic endpoint returned data but not in the correct format.'
|
2019-06-17 14:42:41 +02:00
|
|
|
try:
|
|
|
|
for line in iter(p.stdout.readline, b''):
|
|
|
|
if line.startswith(b'data: '):
|
|
|
|
data = line[6:]
|
|
|
|
try:
|
|
|
|
j = json.loads(data)
|
|
|
|
return_flag = True
|
|
|
|
return_text = f'Dynamic endpoint returned data (took {time.time()-start_time:.2f}s)'
|
|
|
|
signal.alarm(0)
|
|
|
|
break
|
|
|
|
except Exception as e:
|
|
|
|
return_flag = False
|
|
|
|
return_text = f'Something went wrong. Output {line}'
|
|
|
|
break
|
2019-06-17 14:47:51 +02:00
|
|
|
except diagnostic_util.TimeoutException:
|
2019-06-17 14:42:41 +02:00
|
|
|
return_text = f'Dynamic endpoint did not returned data in the given time ({int(time.time()-start_time)}sec)'
|
|
|
|
return (return_flag, return_text)
|
2019-06-14 16:15:07 +02:00
|
|
|
|
|
|
|
|
|
|
|
def start_diagnostic():
|
2019-06-18 12:30:07 +02:00
|
|
|
if not (check_virtual_environment_and_packages() and check_configuration()):
|
2019-06-14 16:15:07 +02:00
|
|
|
return
|
|
|
|
check_file_permission()
|
|
|
|
check_redis()
|
2019-06-17 11:41:14 +02:00
|
|
|
check_zmq()
|
|
|
|
check_processes_status()
|
|
|
|
check_subscriber_status()
|
2019-06-17 09:50:26 +02:00
|
|
|
if check_buffer_queue() is not True:
|
|
|
|
check_buffer_change_rate()
|
2019-06-17 16:16:05 +02:00
|
|
|
dispatcher_running = check_dispatcher_status()
|
|
|
|
if check_server_listening() and dispatcher_running:
|
2019-06-17 15:09:08 +02:00
|
|
|
check_server_dynamic_enpoint()
|
2019-06-14 16:15:07 +02:00
|
|
|
|
|
|
|
|
|
|
|
def main():
|
|
|
|
start_diagnostic()
|
|
|
|
|
|
|
|
|
|
|
|
if __name__ == '__main__':
|
|
|
|
main()
|