mirror of https://github.com/MISP/misp-dashboard
271 lines
8.5 KiB
Python
Executable File
271 lines
8.5 KiB
Python
Executable File
#!/usr/bin/env python3.5
|
|
|
|
import time, datetime
|
|
import copy
|
|
from pprint import pprint
|
|
import zmq
|
|
import redis
|
|
import random
|
|
import configparser
|
|
import argparse
|
|
import os
|
|
import sys
|
|
import json
|
|
|
|
import util
|
|
import geo_helper
|
|
import contributor_helper
|
|
import users_helper
|
|
import trendings_helper
|
|
|
|
configfile = os.path.join(os.environ['DASH_CONFIG'], 'config.cfg')
|
|
cfg = configparser.ConfigParser()
|
|
cfg.read(configfile)
|
|
|
|
CHANNEL = cfg.get('RedisLog', 'channel')
|
|
LISTNAME = cfg.get('RedisLIST', 'listName')
|
|
|
|
serv_log = redis.StrictRedis(
|
|
host=cfg.get('RedisGlobal', 'host'),
|
|
port=cfg.getint('RedisGlobal', 'port'),
|
|
db=cfg.getint('RedisLog', 'db'))
|
|
serv_redis_db = redis.StrictRedis(
|
|
host=cfg.get('RedisGlobal', 'host'),
|
|
port=cfg.getint('RedisGlobal', 'port'),
|
|
db=cfg.getint('RedisDB', 'db'))
|
|
serv_list = redis.StrictRedis(
|
|
host=cfg.get('RedisGlobal', 'host'),
|
|
port=cfg.getint('RedisGlobal', 'port'),
|
|
db=cfg.getint('RedisLIST', 'db'))
|
|
|
|
geo_helper = geo_helper.Geo_helper(serv_redis_db, cfg)
|
|
contributor_helper = contributor_helper.Contributor_helper(serv_redis_db, cfg)
|
|
users_helper = users_helper.Users_helper(serv_redis_db, cfg)
|
|
trendings_helper = trendings_helper.Trendings_helper(serv_redis_db, cfg)
|
|
|
|
|
|
def publish_log(zmq_name, name, content, channel=CHANNEL):
|
|
to_send = { 'name': name, 'log': json.dumps(content), 'zmqName': zmq_name }
|
|
serv_log.publish(channel, json.dumps(to_send))
|
|
|
|
def getFields(obj, fields):
|
|
jsonWalker = fields.split('.')
|
|
itemToExplore = obj
|
|
lastName = ""
|
|
try:
|
|
for i in jsonWalker:
|
|
itemToExplore = itemToExplore[i]
|
|
lastName = i
|
|
if type(itemToExplore) is list:
|
|
return { 'name': lastName , 'data': itemToExplore }
|
|
else:
|
|
return itemToExplore
|
|
except KeyError as e:
|
|
return ""
|
|
|
|
##############
|
|
## HANDLERS ##
|
|
##############
|
|
|
|
def handler_log(zmq_name, jsonevent):
|
|
print('sending', 'log')
|
|
return
|
|
|
|
def handler_dispatcher(zmq_name, jsonObj):
|
|
if "Event" in jsonObj:
|
|
handler_event(zmq_name, jsonObj)
|
|
|
|
def handler_keepalive(zmq_name, jsonevent):
|
|
print('sending', 'keepalive')
|
|
to_push = [ jsonevent['uptime'] ]
|
|
publish_log(zmq_name, 'Keepalive', to_push)
|
|
|
|
def handler_user(zmq_name, jsondata):
|
|
print('sending', 'user')
|
|
action = jsondata['action']
|
|
json_user = jsondata['User']
|
|
json_org = jsondata['Organisation']
|
|
org = json_org['name']
|
|
if action == 'login': #only consider user login
|
|
timestamp = int(time.time())
|
|
users_helper.add_user_login(timestamp, org)
|
|
else:
|
|
pass
|
|
|
|
def handler_conversation(zmq_name, jsonevent):
|
|
try: #only consider POST, not THREAD
|
|
jsonpost = jsonevent['Post']
|
|
except KeyError:
|
|
return
|
|
print('sending' ,'Post')
|
|
org = jsonpost['org_name']
|
|
categ = None
|
|
action = 'add'
|
|
eventName = 'no name or id yet...'
|
|
contributor_helper.handleContribution(zmq_name, org,
|
|
'Discussion',
|
|
None,
|
|
action,
|
|
isLabeled=False)
|
|
# add Discussion
|
|
nowSec = int(time.time())
|
|
trendings_helper.addTrendingDisc(eventName, nowSec)
|
|
|
|
def handler_object(zmq_name, jsondata):
|
|
print('obj')
|
|
return
|
|
|
|
def handler_sighting(zmq_name, jsondata):
|
|
print('sending' ,'sighting')
|
|
jsonsight = jsondata['Sighting']
|
|
org = jsonsight['Event']['Orgc']['name']
|
|
categ = jsonsight['Attribute']['category']
|
|
action = jsondata.get('action', None)
|
|
contributor_helper.handleContribution(zmq_name, org, 'Sighting', categ, action, pntMultiplier=2)
|
|
handler_attribute(zmq_name, jsonsight, hasAlreadyBeenContributed=True)
|
|
|
|
timestamp = jsonsight.get('date_sighting', None)
|
|
|
|
if jsonsight['type'] == "0": # sightings
|
|
trendings_helper.addSightings(timestamp)
|
|
elif jsonsight['type'] == "1": # false positive
|
|
trendings_helper.addFalsePositive(timestamp)
|
|
|
|
def handler_event(zmq_name, jsonobj):
|
|
#fields: threat_level_id, id, info
|
|
jsonevent = jsonobj['Event']
|
|
|
|
#Add trending
|
|
eventName = jsonevent['info']
|
|
timestamp = jsonevent['timestamp']
|
|
trendings_helper.addTrendingEvent(eventName, timestamp)
|
|
tags = []
|
|
for tag in jsonobj.get('EventTag', []):
|
|
try:
|
|
tags.append(tag['Tag'])
|
|
except KeyError:
|
|
pass
|
|
trendings_helper.addTrendingTags(tags, timestamp)
|
|
|
|
#redirect to handler_attribute
|
|
if 'Attribute' in jsonevent:
|
|
attributes = jsonevent['Attribute']
|
|
if type(attributes) is list:
|
|
for attr in attributes:
|
|
jsoncopy = copy.deepcopy(jsonobj)
|
|
jsoncopy['Attribute'] = attr
|
|
handler_attribute(zmq_name, jsoncopy)
|
|
else:
|
|
handler_attribute(zmq_name, attributes)
|
|
|
|
action = jsonobj.get('action', None)
|
|
eventLabeled = len(jsonobj.get('EventTag', [])) > 0
|
|
org = jsonobj.get('Orgc', {}).get('name', None)
|
|
|
|
if org is not None:
|
|
contributor_helper.handleContribution(zmq_name, org,
|
|
'Event',
|
|
None,
|
|
action,
|
|
isLabeled=eventLabeled)
|
|
|
|
def handler_attribute(zmq_name, jsonobj, hasAlreadyBeenContributed=False):
|
|
# check if jsonattr is an attribute object
|
|
if 'Attribute' in jsonobj:
|
|
jsonattr = jsonobj['Attribute']
|
|
|
|
#Add trending
|
|
categName = jsonattr['category']
|
|
timestamp = jsonattr.get('timestamp', int(time.time()))
|
|
trendings_helper.addTrendingCateg(categName, timestamp)
|
|
tags = []
|
|
for tag in jsonattr.get('Tag', []):
|
|
try:
|
|
tags.append(tag['Tag'])
|
|
except KeyError:
|
|
pass
|
|
trendings_helper.addTrendingTags(tags, timestamp)
|
|
|
|
to_push = []
|
|
for field in json.loads(cfg.get('Log', 'fieldname_order')):
|
|
if type(field) is list:
|
|
to_join = []
|
|
for subField in field:
|
|
to_join.append(getFields(jsonobj, subField))
|
|
to_add = cfg.get('Log', 'char_separator').join(to_join)
|
|
else:
|
|
to_add = getFields(jsonobj, field)
|
|
to_push.append(to_add)
|
|
|
|
#try to get coord from ip
|
|
if jsonattr['category'] == "Network activity":
|
|
geo_helper.getCoordFromIpAndPublish(jsonattr['value'], jsonattr['category'])
|
|
|
|
#try to get coord from ip
|
|
if jsonattr['type'] == "phone-number":
|
|
geo_helper.getCoordFromPhoneAndPublish(jsonattr['value'], jsonattr['category'])
|
|
|
|
if not hasAlreadyBeenContributed:
|
|
eventLabeled = len(jsonobj.get('EventTag', [])) > 0
|
|
action = jsonobj.get('action', None)
|
|
contributor_helper.handleContribution(zmq_name, jsonobj['Event']['Orgc']['name'],
|
|
'Attribute',
|
|
jsonattr['category'],
|
|
action,
|
|
isLabeled=eventLabeled)
|
|
# Push to log
|
|
publish_log(zmq_name, 'Attribute', to_push)
|
|
|
|
|
|
###############
|
|
## MAIN LOOP ##
|
|
###############
|
|
|
|
def process_log(zmq_name, event):
|
|
topic, eventdata = event.split(' ', maxsplit=1)
|
|
jsonevent = json.loads(eventdata)
|
|
try:
|
|
dico_action[topic](zmq_name, jsonevent)
|
|
except KeyError as e:
|
|
print(e)
|
|
|
|
|
|
def main(sleeptime):
|
|
numMsg = 0
|
|
while True:
|
|
content = serv_list.rpop(LISTNAME)
|
|
if content is None:
|
|
print('Processed', numMsg, 'message(s) since last sleep.')
|
|
numMsg = 0
|
|
time.sleep(sleeptime)
|
|
continue
|
|
content = content.decode('utf8')
|
|
the_json = json.loads(content)
|
|
zmqName = the_json['zmq_name']
|
|
content = the_json['content']
|
|
process_log(zmqName, content)
|
|
numMsg += 1
|
|
|
|
|
|
dico_action = {
|
|
"misp_json": handler_dispatcher,
|
|
"misp_json_event": handler_event,
|
|
"misp_json_self": handler_keepalive,
|
|
"misp_json_attribute": handler_attribute,
|
|
"misp_json_object": handler_object,
|
|
"misp_json_sighting": handler_sighting,
|
|
"misp_json_organisation": handler_log,
|
|
"misp_json_user": handler_user,
|
|
"misp_json_conversation": handler_conversation,
|
|
"misp_json_object_reference": handler_log,
|
|
}
|
|
|
|
|
|
if __name__ == "__main__":
|
|
|
|
parser = argparse.ArgumentParser(description='The ZMQ dispatcher. It pops from the redis buffer then redispatch it to the correct handlers')
|
|
parser.add_argument('-s', '--sleep', required=False, dest='sleeptime', type=int, help='The number of second to wait before checking redis list size', default=5)
|
|
args = parser.parse_args()
|
|
|
|
main(args.sleeptime)
|