mirror of https://github.com/MISP/misp-dashboard
Started integration of Logging
parent
4c04bd02c6
commit
1386e58cf8
|
@ -6,11 +6,14 @@ import os
|
||||||
import configparser
|
import configparser
|
||||||
import json
|
import json
|
||||||
import datetime
|
import datetime
|
||||||
|
import logging
|
||||||
import redis
|
import redis
|
||||||
|
|
||||||
import util
|
import util
|
||||||
import users_helper
|
import users_helper
|
||||||
KEYDAY = "CONTRIB_DAY" # To be used by other module
|
KEYDAY = "CONTRIB_DAY" # To be used by other module
|
||||||
|
logging.basicConfig(filename='logs/logs.log', filemode='w', level=logging.INFO)
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class Contributor_helper:
|
class Contributor_helper:
|
||||||
def __init__(self, serv_redis_db, cfg):
|
def __init__(self, serv_redis_db, cfg):
|
||||||
|
@ -95,10 +98,12 @@ class Contributor_helper:
|
||||||
today_str = util.getDateStrFormat(date)
|
today_str = util.getDateStrFormat(date)
|
||||||
keyname = "{}:{}:{}".format(self.keyCateg, today_str, categ)
|
keyname = "{}:{}:{}".format(self.keyCateg, today_str, categ)
|
||||||
self.serv_redis_db.zincrby(keyname, org, count)
|
self.serv_redis_db.zincrby(keyname, org, count)
|
||||||
|
logger.debug('Added to redis: keyname={}, org={}, count={}'.format(keyname, org, count))
|
||||||
|
|
||||||
def publish_log(self, zmq_name, name, content, channel=""):
|
def publish_log(self, zmq_name, name, content, channel=""):
|
||||||
to_send = { 'name': name, 'log': json.dumps(content), 'zmqName': zmq_name }
|
to_send = { 'name': name, 'log': json.dumps(content), 'zmqName': zmq_name }
|
||||||
self.serv_log.publish(channel, json.dumps(to_send))
|
self.serv_log.publish(channel, json.dumps(to_send))
|
||||||
|
logger.debug('Published: {}'.format(json.dumps(to_send)))
|
||||||
|
|
||||||
''' HANDLER '''
|
''' HANDLER '''
|
||||||
#pntMultiplier if one contribution rewards more than others. (e.g. shighting may gives more points than editing)
|
#pntMultiplier if one contribution rewards more than others. (e.g. shighting may gives more points than editing)
|
||||||
|
@ -111,7 +116,7 @@ class Contributor_helper:
|
||||||
nowSec = int(time.time())
|
nowSec = int(time.time())
|
||||||
pnts_to_add = self.default_pnts_per_contribution
|
pnts_to_add = self.default_pnts_per_contribution
|
||||||
|
|
||||||
# if there is a contribution, there is a login (even if ti comes from the API)
|
# if there is a contribution, there is a login (even if it comes from the API)
|
||||||
self.users_helper.add_user_login(nowSec, org)
|
self.users_helper.add_user_login(nowSec, org)
|
||||||
|
|
||||||
# is a valid contribution
|
# is a valid contribution
|
||||||
|
@ -133,6 +138,7 @@ class Contributor_helper:
|
||||||
|
|
||||||
keyname = "{}:{}".format(self.keyLastContrib, util.getDateStrFormat(now))
|
keyname = "{}:{}".format(self.keyLastContrib, util.getDateStrFormat(now))
|
||||||
self.serv_redis_db.zadd(keyname, nowSec, org)
|
self.serv_redis_db.zadd(keyname, nowSec, org)
|
||||||
|
logger.debug('Added to redis: keyname={}, nowSec={}, org={}'.format(keyname, nowSec, org))
|
||||||
self.serv_redis_db.expire(keyname, util.ONE_DAY*7) #expire after 7 day
|
self.serv_redis_db.expire(keyname, util.ONE_DAY*7) #expire after 7 day
|
||||||
|
|
||||||
awards_given = self.updateOrgContributionRank(org, pnts_to_add, action, contribType, eventTime=datetime.datetime.now(), isLabeled=isLabeled, categ=util.noSpaceLower(categ))
|
awards_given = self.updateOrgContributionRank(org, pnts_to_add, action, contribType, eventTime=datetime.datetime.now(), isLabeled=isLabeled, categ=util.noSpaceLower(categ))
|
||||||
|
@ -141,6 +147,7 @@ class Contributor_helper:
|
||||||
# update awards given
|
# update awards given
|
||||||
keyname = "{}:{}".format(self.keyLastAward, util.getDateStrFormat(now))
|
keyname = "{}:{}".format(self.keyLastAward, util.getDateStrFormat(now))
|
||||||
self.serv_redis_db.zadd(keyname, nowSec, json.dumps({'org': org, 'award': award, 'epoch': nowSec }))
|
self.serv_redis_db.zadd(keyname, nowSec, json.dumps({'org': org, 'award': award, 'epoch': nowSec }))
|
||||||
|
logger.debug('Added to redis: keyname={}, nowSec={}, content={}'.format(keyname, nowSec, json.dumps({'org': org, 'award': award, 'epoch': nowSec })))
|
||||||
self.serv_redis_db.expire(keyname, util.ONE_DAY*7) #expire after 7 day
|
self.serv_redis_db.expire(keyname, util.ONE_DAY*7) #expire after 7 day
|
||||||
# publish
|
# publish
|
||||||
self.publish_log(zmq_name, 'CONTRIBUTION', {'org': org, 'award': award, 'epoch': nowSec }, channel=self.CHANNEL_LASTAWARDS)
|
self.publish_log(zmq_name, 'CONTRIBUTION', {'org': org, 'award': award, 'epoch': nowSec }, channel=self.CHANNEL_LASTAWARDS)
|
||||||
|
@ -214,14 +221,17 @@ class Contributor_helper:
|
||||||
if contribType == 'Attribute':
|
if contribType == 'Attribute':
|
||||||
attributeWeekCount = self.serv_redis_db.incrby(keyname.format(org=orgName, orgCateg='ATTR_WEEK_COUNT'), 1)
|
attributeWeekCount = self.serv_redis_db.incrby(keyname.format(org=orgName, orgCateg='ATTR_WEEK_COUNT'), 1)
|
||||||
self.serv_redis_db.expire(keyname.format(org=orgName, orgCateg='ATTR_WEEK_COUNT'), util.ONE_DAY*7)
|
self.serv_redis_db.expire(keyname.format(org=orgName, orgCateg='ATTR_WEEK_COUNT'), util.ONE_DAY*7)
|
||||||
|
logger.debug('Incrby: keyname={}'.format(keyname.format(org=orgName, orgCateg='ATTR_WEEK_COUNT')))
|
||||||
|
|
||||||
if contribType == 'Proposal':
|
if contribType == 'Proposal':
|
||||||
proposalWeekCount = self.serv_redis_db.incrby(keyname.format(org=orgName, orgCateg='PROP_WEEK_COUNT'), 1)
|
proposalWeekCount = self.serv_redis_db.incrby(keyname.format(org=orgName, orgCateg='PROP_WEEK_COUNT'), 1)
|
||||||
|
logger.debug('Incrby: keyname={}'.format(keyname.format(org=orgName, orgCateg='PROP_WEEK_COUNT')))
|
||||||
self.serv_redis_db.expire(keyname.format(org=orgName, orgCateg='PROP_WEEK_COUNT'), util.ONE_DAY*7)
|
self.serv_redis_db.expire(keyname.format(org=orgName, orgCateg='PROP_WEEK_COUNT'), util.ONE_DAY*7)
|
||||||
addContributionToCateg(datetime.datetime.now(), 'proposal')
|
addContributionToCateg(datetime.datetime.now(), 'proposal')
|
||||||
|
|
||||||
if contribType == 'Sighting':
|
if contribType == 'Sighting':
|
||||||
sightingWeekCount = self.serv_redis_db.incrby(keyname.format(org=orgName, orgCateg='SIGHT_WEEK_COUNT'), 1)
|
sightingWeekCount = self.serv_redis_db.incrby(keyname.format(org=orgName, orgCateg='SIGHT_WEEK_COUNT'), 1)
|
||||||
|
logger.debug('Incrby: keyname={}'.format(keyname.format(org=orgName, orgCateg='SIGHT_WEEK_COUNT')))
|
||||||
self.serv_redis_db.expire(keyname.format(org=orgName, orgCateg='SIGHT_WEEK_COUNT'), util.ONE_DAY*7)
|
self.serv_redis_db.expire(keyname.format(org=orgName, orgCateg='SIGHT_WEEK_COUNT'), util.ONE_DAY*7)
|
||||||
self.addContributionToCateg(datetime.datetime.now(), 'sighting', orgName)
|
self.addContributionToCateg(datetime.datetime.now(), 'sighting', orgName)
|
||||||
|
|
||||||
|
@ -230,9 +240,11 @@ class Contributor_helper:
|
||||||
|
|
||||||
if contribType == 'Event':
|
if contribType == 'Event':
|
||||||
eventWeekCount = self.serv_redis_db.incrby(keyname.format(org=orgName, orgCateg='EVENT_WEEK_COUNT'), 1)
|
eventWeekCount = self.serv_redis_db.incrby(keyname.format(org=orgName, orgCateg='EVENT_WEEK_COUNT'), 1)
|
||||||
|
logger.debug('Incrby: keyname={}'.format(keyname.format(org=orgName, orgCateg='EVENT_WEEK_COUNT')))
|
||||||
self.serv_redis_db.expire(keyname.format(org=orgName, orgCateg='EVENT_WEEK_COUNT'), util.ONE_DAY*7)
|
self.serv_redis_db.expire(keyname.format(org=orgName, orgCateg='EVENT_WEEK_COUNT'), util.ONE_DAY*7)
|
||||||
|
|
||||||
eventMonthCount = self.serv_redis_db.incrby(keyname.format(org=orgName, orgCateg='EVENT_MONTH_COUNT'), 1)
|
eventMonthCount = self.serv_redis_db.incrby(keyname.format(org=orgName, orgCateg='EVENT_MONTH_COUNT'), 1)
|
||||||
|
logger.debug('Incrby: keyname={}'.format(keyname.format(org=orgName, orgCateg='EVENT_MONTH_COUNT')))
|
||||||
self.serv_redis_db.expire(keyname.format(org=orgName, orgCateg='EVENT_MONTH_COUNT'), util.ONE_DAY*7)
|
self.serv_redis_db.expire(keyname.format(org=orgName, orgCateg='EVENT_MONTH_COUNT'), util.ONE_DAY*7)
|
||||||
|
|
||||||
# getRequirement parameters
|
# getRequirement parameters
|
||||||
|
@ -275,6 +287,7 @@ class Contributor_helper:
|
||||||
|
|
||||||
for rankReq, ttl in contrib:
|
for rankReq, ttl in contrib:
|
||||||
self.serv_redis_db.set(keyname.format(org=orgName, orgCateg='CONTRIB_REQ_'+str(rankReq)), 1)
|
self.serv_redis_db.set(keyname.format(org=orgName, orgCateg='CONTRIB_REQ_'+str(rankReq)), 1)
|
||||||
|
logger.debug('Set: keyname={}'.format(keyname.format(org=orgName, orgCateg='CONTRIB_REQ_'+str(rankReq))))
|
||||||
self.serv_redis_db.expire(keyname.format(org=orgName, orgCateg='CONTRIB_REQ_'+str(rankReq)), ttl)
|
self.serv_redis_db.expire(keyname.format(org=orgName, orgCateg='CONTRIB_REQ_'+str(rankReq)), ttl)
|
||||||
|
|
||||||
ContributionStatus = self.getCurrentContributionStatus(orgName)
|
ContributionStatus = self.getCurrentContributionStatus(orgName)
|
||||||
|
@ -322,10 +335,12 @@ class Contributor_helper:
|
||||||
def giveBadgeToOrg(self, org, badgeNum):
|
def giveBadgeToOrg(self, org, badgeNum):
|
||||||
keyname = '{mainKey}:{org}:{orgCateg}'
|
keyname = '{mainKey}:{org}:{orgCateg}'
|
||||||
self.serv_redis_db.set(keyname.format(mainKey=self.keyContribReq, org=org, orgCateg='BADGE_'+str(badgeNum)), 1)
|
self.serv_redis_db.set(keyname.format(mainKey=self.keyContribReq, org=org, orgCateg='BADGE_'+str(badgeNum)), 1)
|
||||||
|
logger.debug('Giving badge {} to org {}'.format(org, badgeNum))
|
||||||
|
|
||||||
def removeBadgeFromOrg(self, org, badgeNum):
|
def removeBadgeFromOrg(self, org, badgeNum):
|
||||||
keyname = '{mainKey}:{org}:{orgCateg}'
|
keyname = '{mainKey}:{org}:{orgCateg}'
|
||||||
self.serv_redis_db.delete(keyname.format(mainKey=self.keyContribReq, org=org, orgCateg='BADGE_'+str(badgeNum)))
|
self.serv_redis_db.delete(keyname.format(mainKey=self.keyContribReq, org=org, orgCateg='BADGE_'+str(badgeNum)))
|
||||||
|
logger.debug('Removing badge {} from org {}'.format(org, badgeNum))
|
||||||
|
|
||||||
''' TROPHIES '''
|
''' TROPHIES '''
|
||||||
def getOrgTrophies(self, org):
|
def getOrgTrophies(self, org):
|
||||||
|
@ -360,7 +375,6 @@ class Contributor_helper:
|
||||||
def posToRankMapping(self, pos, totNum):
|
def posToRankMapping(self, pos, totNum):
|
||||||
mapping = self.trophyMapping
|
mapping = self.trophyMapping
|
||||||
mapping_num = [math.ceil(float(float(totNum*i)/float(100))) for i in mapping]
|
mapping_num = [math.ceil(float(float(totNum*i)/float(100))) for i in mapping]
|
||||||
# print(pos, totNum)
|
|
||||||
if pos == 0: #first
|
if pos == 0: #first
|
||||||
position = 1
|
position = 1
|
||||||
else:
|
else:
|
||||||
|
@ -377,10 +391,12 @@ class Contributor_helper:
|
||||||
def giveTrophyPointsToOrg(self, org, categ, points):
|
def giveTrophyPointsToOrg(self, org, categ, points):
|
||||||
keyname = '{mainKey}:{orgCateg}'
|
keyname = '{mainKey}:{orgCateg}'
|
||||||
self.serv_redis_db.zincrby(keyname.format(mainKey=self.keyTrophy, orgCateg=categ), org, points)
|
self.serv_redis_db.zincrby(keyname.format(mainKey=self.keyTrophy, orgCateg=categ), org, points)
|
||||||
|
logger.debug('Giving {} trophy points to {} in {}'.format(points, org, categ))
|
||||||
|
|
||||||
def removeTrophyPointsFromOrg(self, org, categ, points):
|
def removeTrophyPointsFromOrg(self, org, categ, points):
|
||||||
keyname = '{mainKey}:{orgCateg}'
|
keyname = '{mainKey}:{orgCateg}'
|
||||||
self.serv_redis_db.zincrby(keyname.format(mainKey=self.keyTrophy, orgCateg=categ), org, -points)
|
self.serv_redis_db.zincrby(keyname.format(mainKey=self.keyTrophy, orgCateg=categ), org, -points)
|
||||||
|
logger.debug('Removing {} trophy points from {} in {}'.format(points, org, categ))
|
||||||
|
|
||||||
''' AWARDS HELPER '''
|
''' AWARDS HELPER '''
|
||||||
def getLastAwardsFromRedis(self):
|
def getLastAwardsFromRedis(self):
|
||||||
|
|
|
@ -2,6 +2,7 @@ import math, random
|
||||||
import os
|
import os
|
||||||
import json
|
import json
|
||||||
import datetime, time
|
import datetime, time
|
||||||
|
import logging
|
||||||
import json
|
import json
|
||||||
import redis
|
import redis
|
||||||
from collections import OrderedDict
|
from collections import OrderedDict
|
||||||
|
@ -12,6 +13,9 @@ from phonenumbers import geocoder
|
||||||
|
|
||||||
import util
|
import util
|
||||||
|
|
||||||
|
logging.basicConfig(filename='logs/logs.log', filemode='w', level=logging.INFO)
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class Geo_helper:
|
class Geo_helper:
|
||||||
def __init__(self, serv_redis_db, cfg):
|
def __init__(self, serv_redis_db, cfg):
|
||||||
self.serv_redis_db = serv_redis_db
|
self.serv_redis_db = serv_redis_db
|
||||||
|
@ -102,21 +106,22 @@ class Geo_helper:
|
||||||
"regionCode": rep['full_rep'].country.iso_code,
|
"regionCode": rep['full_rep'].country.iso_code,
|
||||||
}
|
}
|
||||||
self.serv_coord.publish(self.CHANNELDISP, json.dumps(to_send))
|
self.serv_coord.publish(self.CHANNELDISP, json.dumps(to_send))
|
||||||
|
logger.info('Published: {}'.format(json.dumps(to_send)))
|
||||||
except ValueError:
|
except ValueError:
|
||||||
print("can't resolve ip")
|
logger.warning("can't resolve ip")
|
||||||
except geoip2.errors.AddressNotFoundError:
|
except geoip2.errors.AddressNotFoundError:
|
||||||
print("Address not in Database")
|
logger.warning("Address not in Database")
|
||||||
|
|
||||||
def getCoordFromPhoneAndPublish(self, phoneNumber, categ):
|
def getCoordFromPhoneAndPublish(self, phoneNumber, categ):
|
||||||
try:
|
try:
|
||||||
rep = phonenumbers.parse(phoneNumber, None)
|
rep = phonenumbers.parse(phoneNumber, None)
|
||||||
if not (phonenumbers.is_valid_number(rep) or phonenumbers.is_possible_number(rep)):
|
if not (phonenumbers.is_valid_number(rep) or phonenumbers.is_possible_number(rep)):
|
||||||
print("Phone number not valid")
|
logger.warning("Phone number not valid")
|
||||||
return
|
return
|
||||||
country_name = geocoder.country_name_for_number(rep, "en")
|
country_name = geocoder.country_name_for_number(rep, "en")
|
||||||
country_code = self.country_to_iso[country_name]
|
country_code = self.country_to_iso[country_name]
|
||||||
if country_code is None:
|
if country_code is None:
|
||||||
print("Non matching ISO_CODE")
|
logger.warning("Non matching ISO_CODE")
|
||||||
return
|
return
|
||||||
coord = self.country_code_to_coord[country_code.lower()] # countrycode is in upper case
|
coord = self.country_code_to_coord[country_code.lower()] # countrycode is in upper case
|
||||||
coord_dic = {'lat': coord['lat'], 'lon': coord['long']}
|
coord_dic = {'lat': coord['lat'], 'lon': coord['long']}
|
||||||
|
@ -141,8 +146,9 @@ class Geo_helper:
|
||||||
"regionCode": country_code,
|
"regionCode": country_code,
|
||||||
}
|
}
|
||||||
self.serv_coord.publish(self.CHANNELDISP, json.dumps(to_send))
|
self.serv_coord.publish(self.CHANNELDISP, json.dumps(to_send))
|
||||||
|
logger.info('Published: {}'.format(json.dumps(to_send)))
|
||||||
except phonenumbers.NumberParseException:
|
except phonenumbers.NumberParseException:
|
||||||
print("Can't resolve phone number country")
|
logger.warning("Can't resolve phone number country")
|
||||||
|
|
||||||
''' UTIL '''
|
''' UTIL '''
|
||||||
def push_to_redis_geo(self, keyCateg, lon, lat, content):
|
def push_to_redis_geo(self, keyCateg, lon, lat, content):
|
||||||
|
@ -150,12 +156,13 @@ class Geo_helper:
|
||||||
today_str = util.getDateStrFormat(now)
|
today_str = util.getDateStrFormat(now)
|
||||||
keyname = "{}:{}".format(keyCateg, today_str)
|
keyname = "{}:{}".format(keyCateg, today_str)
|
||||||
self.serv_redis_db.geoadd(keyname, lon, lat, content)
|
self.serv_redis_db.geoadd(keyname, lon, lat, content)
|
||||||
|
logger.debug('Added to redis: keyname={}, lon={}, lat={}, content={}'.format(keyname, lon, lat, content))
|
||||||
def push_to_redis_zset(self, keyCateg, toAdd, endSubkey="", count=1):
|
def push_to_redis_zset(self, keyCateg, toAdd, endSubkey="", count=1):
|
||||||
now = datetime.datetime.now()
|
now = datetime.datetime.now()
|
||||||
today_str = util.getDateStrFormat(now)
|
today_str = util.getDateStrFormat(now)
|
||||||
keyname = "{}:{}{}".format(keyCateg, today_str, endSubkey)
|
keyname = "{}:{}{}".format(keyCateg, today_str, endSubkey)
|
||||||
self.serv_redis_db.zincrby(keyname, toAdd, count)
|
self.serv_redis_db.zincrby(keyname, toAdd, count)
|
||||||
|
logger.debug('Added to redis: keyname={}, toAdd={}, count={}'.format(keyname, toAdd, count))
|
||||||
|
|
||||||
def ip_to_coord(self, ip):
|
def ip_to_coord(self, ip):
|
||||||
resp = self.reader.city(ip)
|
resp = self.reader.city(ip)
|
||||||
|
|
|
@ -8,6 +8,7 @@ from time import gmtime as now
|
||||||
from time import sleep, strftime
|
from time import sleep, strftime
|
||||||
import datetime
|
import datetime
|
||||||
import os
|
import os
|
||||||
|
import logging
|
||||||
|
|
||||||
import util
|
import util
|
||||||
import geo_helper
|
import geo_helper
|
||||||
|
@ -19,6 +20,9 @@ configfile = os.path.join(os.environ['DASH_CONFIG'], 'config.cfg')
|
||||||
cfg = configparser.ConfigParser()
|
cfg = configparser.ConfigParser()
|
||||||
cfg.read(configfile)
|
cfg.read(configfile)
|
||||||
|
|
||||||
|
logging.basicConfig(filename='logs/logs.log', filemode='w', level=logging.INFO)
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
app = Flask(__name__)
|
app = Flask(__name__)
|
||||||
|
|
||||||
redis_server_log = redis.StrictRedis(
|
redis_server_log = redis.StrictRedis(
|
||||||
|
@ -99,8 +103,8 @@ class EventMessage():
|
||||||
msg = msg.decode('utf8')
|
msg = msg.decode('utf8')
|
||||||
try:
|
try:
|
||||||
jsonMsg = json.loads(msg)
|
jsonMsg = json.loads(msg)
|
||||||
except json.JSONDecodeError:
|
except json.JSONDecodeError as e:
|
||||||
print('json decode error')
|
logger.error(e)
|
||||||
jsonMsg = { 'name': "undefined" ,'log': json.loads(msg) }
|
jsonMsg = { 'name': "undefined" ,'log': json.loads(msg) }
|
||||||
|
|
||||||
self.feedName = jsonMsg['name']
|
self.feedName = jsonMsg['name']
|
||||||
|
|
|
@ -2,10 +2,14 @@ import math, random
|
||||||
import os
|
import os
|
||||||
import json
|
import json
|
||||||
import datetime, time
|
import datetime, time
|
||||||
|
import logging
|
||||||
from collections import OrderedDict
|
from collections import OrderedDict
|
||||||
|
|
||||||
import util
|
import util
|
||||||
|
|
||||||
|
logging.basicConfig(filename='logs/logs.log', filemode='w', level=logging.INFO)
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class Trendings_helper:
|
class Trendings_helper:
|
||||||
def __init__(self, serv_redis_db, cfg):
|
def __init__(self, serv_redis_db, cfg):
|
||||||
self.serv_redis_db = serv_redis_db
|
self.serv_redis_db = serv_redis_db
|
||||||
|
@ -30,6 +34,7 @@ class Trendings_helper:
|
||||||
else:
|
else:
|
||||||
to_save = data
|
to_save = data
|
||||||
self.serv_redis_db.zincrby(keyname, to_save, 1)
|
self.serv_redis_db.zincrby(keyname, to_save, 1)
|
||||||
|
logger.debug('Added to redis: keyname={}, content={}'.format(keyname, to_save))
|
||||||
|
|
||||||
def addTrendingEvent(self, eventName, timestamp):
|
def addTrendingEvent(self, eventName, timestamp):
|
||||||
self.addGenericTrending(self.keyEvent, eventName, timestamp)
|
self.addGenericTrending(self.keyEvent, eventName, timestamp)
|
||||||
|
@ -53,12 +58,14 @@ class Trendings_helper:
|
||||||
timestampDate_str = util.getDateStrFormat(timestampDate)
|
timestampDate_str = util.getDateStrFormat(timestampDate)
|
||||||
keyname = "{}:{}".format(self.keySigh, timestampDate_str)
|
keyname = "{}:{}".format(self.keySigh, timestampDate_str)
|
||||||
self.serv_redis_db.incrby(keyname, 1)
|
self.serv_redis_db.incrby(keyname, 1)
|
||||||
|
logger.debug('Incrby: keyname={}'.format(keyname))
|
||||||
|
|
||||||
def addFalsePositive(self, timestamp):
|
def addFalsePositive(self, timestamp):
|
||||||
timestampDate = datetime.datetime.fromtimestamp(float(timestamp))
|
timestampDate = datetime.datetime.fromtimestamp(float(timestamp))
|
||||||
timestampDate_str = util.getDateStrFormat(timestampDate)
|
timestampDate_str = util.getDateStrFormat(timestampDate)
|
||||||
keyname = "{}:{}".format(self.keyFalse, timestampDate_str)
|
keyname = "{}:{}".format(self.keyFalse, timestampDate_str)
|
||||||
self.serv_redis_db.incrby(keyname, 1)
|
self.serv_redis_db.incrby(keyname, 1)
|
||||||
|
logger.debug('Incrby: keyname={}'.format(keyname))
|
||||||
|
|
||||||
''' GETTER '''
|
''' GETTER '''
|
||||||
|
|
||||||
|
|
|
@ -2,10 +2,14 @@ import math, random
|
||||||
import os
|
import os
|
||||||
import json
|
import json
|
||||||
import datetime, time
|
import datetime, time
|
||||||
|
import logging
|
||||||
|
|
||||||
import util
|
import util
|
||||||
import contributor_helper
|
import contributor_helper
|
||||||
|
|
||||||
|
logging.basicConfig(filename='logs/logs.log', filemode='w', level=logging.INFO)
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
class Users_helper:
|
class Users_helper:
|
||||||
def __init__(self, serv_redis_db, cfg):
|
def __init__(self, serv_redis_db, cfg):
|
||||||
self.serv_redis_db = serv_redis_db
|
self.serv_redis_db = serv_redis_db
|
||||||
|
@ -21,6 +25,7 @@ class Users_helper:
|
||||||
timestampDate_str = util.getDateHoursStrFormat(timestampDate)
|
timestampDate_str = util.getDateHoursStrFormat(timestampDate)
|
||||||
keyname_timestamp = "{}:{}".format(self.keyTimestampSet, timestampDate_str)
|
keyname_timestamp = "{}:{}".format(self.keyTimestampSet, timestampDate_str)
|
||||||
self.serv_redis_db.sadd(keyname_timestamp, org)
|
self.serv_redis_db.sadd(keyname_timestamp, org)
|
||||||
|
logger.debug('Added to redis: keyname={}, org={}'.format(keyname_timestamp, org))
|
||||||
self.serv_redis_db.expire(keyname_timestamp, 60*60)
|
self.serv_redis_db.expire(keyname_timestamp, 60*60)
|
||||||
|
|
||||||
def hasAlreadyBeenAdded(self, org, timestamp):
|
def hasAlreadyBeenAdded(self, org, timestamp):
|
||||||
|
@ -39,10 +44,12 @@ class Users_helper:
|
||||||
if not self.hasAlreadyBeenAdded(org, timestamp):
|
if not self.hasAlreadyBeenAdded(org, timestamp):
|
||||||
keyname_timestamp = "{}:{}".format(self.keyTimestamp, timestampDate_str)
|
keyname_timestamp = "{}:{}".format(self.keyTimestamp, timestampDate_str)
|
||||||
self.serv_redis_db.sadd(keyname_timestamp, timestamp)
|
self.serv_redis_db.sadd(keyname_timestamp, timestamp)
|
||||||
|
logger.debug('Added to redis: keyname={}, org={}'.format(keyname_timestamp, timestamp))
|
||||||
self.addTemporary(org, timestamp)
|
self.addTemporary(org, timestamp)
|
||||||
|
|
||||||
keyname_org = "{}:{}".format(self.keyOrgLog, timestampDate_str)
|
keyname_org = "{}:{}".format(self.keyOrgLog, timestampDate_str)
|
||||||
self.serv_redis_db.zincrby(keyname_org, org, 1)
|
self.serv_redis_db.zincrby(keyname_org, org, 1)
|
||||||
|
logger.debug('Added to redis: keyname={}, org={}'.format(keyname_org, org))
|
||||||
|
|
||||||
def getUserLogins(self, date):
|
def getUserLogins(self, date):
|
||||||
keyname = "{}:{}".format(self.keyTimestamp, util.getDateStrFormat(date))
|
keyname = "{}:{}".format(self.keyTimestamp, util.getDateStrFormat(date))
|
||||||
|
|
|
@ -2,7 +2,7 @@
|
||||||
|
|
||||||
import time, datetime
|
import time, datetime
|
||||||
import copy
|
import copy
|
||||||
from pprint import pprint
|
import logging
|
||||||
import zmq
|
import zmq
|
||||||
import redis
|
import redis
|
||||||
import random
|
import random
|
||||||
|
@ -22,6 +22,9 @@ configfile = os.path.join(os.environ['DASH_CONFIG'], 'config.cfg')
|
||||||
cfg = configparser.ConfigParser()
|
cfg = configparser.ConfigParser()
|
||||||
cfg.read(configfile)
|
cfg.read(configfile)
|
||||||
|
|
||||||
|
logging.basicConfig(filename='logs/logs.log', filemode='w', level=logging.INFO)
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
CHANNEL = cfg.get('RedisLog', 'channel')
|
CHANNEL = cfg.get('RedisLog', 'channel')
|
||||||
LISTNAME = cfg.get('RedisLIST', 'listName')
|
LISTNAME = cfg.get('RedisLIST', 'listName')
|
||||||
|
|
||||||
|
@ -47,6 +50,7 @@ trendings_helper = trendings_helper.Trendings_helper(serv_redis_db, cfg)
|
||||||
def publish_log(zmq_name, name, content, channel=CHANNEL):
|
def publish_log(zmq_name, name, content, channel=CHANNEL):
|
||||||
to_send = { 'name': name, 'log': json.dumps(content), 'zmqName': zmq_name }
|
to_send = { 'name': name, 'log': json.dumps(content), 'zmqName': zmq_name }
|
||||||
serv_log.publish(channel, json.dumps(to_send))
|
serv_log.publish(channel, json.dumps(to_send))
|
||||||
|
logger.debug('Published: {}'.format(json.dumps(to_send)))
|
||||||
|
|
||||||
def getFields(obj, fields):
|
def getFields(obj, fields):
|
||||||
jsonWalker = fields.split('.')
|
jsonWalker = fields.split('.')
|
||||||
|
@ -68,7 +72,7 @@ def getFields(obj, fields):
|
||||||
##############
|
##############
|
||||||
|
|
||||||
def handler_log(zmq_name, jsonevent):
|
def handler_log(zmq_name, jsonevent):
|
||||||
print('sending', 'log')
|
logger.info('Log not processed')
|
||||||
return
|
return
|
||||||
|
|
||||||
def handler_dispatcher(zmq_name, jsonObj):
|
def handler_dispatcher(zmq_name, jsonObj):
|
||||||
|
@ -76,12 +80,12 @@ def handler_dispatcher(zmq_name, jsonObj):
|
||||||
handler_event(zmq_name, jsonObj)
|
handler_event(zmq_name, jsonObj)
|
||||||
|
|
||||||
def handler_keepalive(zmq_name, jsonevent):
|
def handler_keepalive(zmq_name, jsonevent):
|
||||||
print('sending', 'keepalive')
|
logger.info('Handling keepalive')
|
||||||
to_push = [ jsonevent['uptime'] ]
|
to_push = [ jsonevent['uptime'] ]
|
||||||
publish_log(zmq_name, 'Keepalive', to_push)
|
publish_log(zmq_name, 'Keepalive', to_push)
|
||||||
|
|
||||||
def handler_user(zmq_name, jsondata):
|
def handler_user(zmq_name, jsondata):
|
||||||
print('sending', 'user')
|
logger.info('Handling user')
|
||||||
action = jsondata['action']
|
action = jsondata['action']
|
||||||
json_user = jsondata['User']
|
json_user = jsondata['User']
|
||||||
json_org = jsondata['Organisation']
|
json_org = jsondata['Organisation']
|
||||||
|
@ -93,11 +97,11 @@ def handler_user(zmq_name, jsondata):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
def handler_conversation(zmq_name, jsonevent):
|
def handler_conversation(zmq_name, jsonevent):
|
||||||
|
logger.info('Handling conversation')
|
||||||
try: #only consider POST, not THREAD
|
try: #only consider POST, not THREAD
|
||||||
jsonpost = jsonevent['Post']
|
jsonpost = jsonevent['Post']
|
||||||
except KeyError:
|
except KeyError as e:
|
||||||
return
|
logger.error('Error in handler_conversation: {}'.format(e))
|
||||||
print('sending' ,'Post')
|
|
||||||
org = jsonpost['org_name']
|
org = jsonpost['org_name']
|
||||||
categ = None
|
categ = None
|
||||||
action = 'add'
|
action = 'add'
|
||||||
|
@ -112,11 +116,11 @@ def handler_conversation(zmq_name, jsonevent):
|
||||||
trendings_helper.addTrendingDisc(eventName, nowSec)
|
trendings_helper.addTrendingDisc(eventName, nowSec)
|
||||||
|
|
||||||
def handler_object(zmq_name, jsondata):
|
def handler_object(zmq_name, jsondata):
|
||||||
print('obj')
|
logger.info('Handling object')
|
||||||
return
|
return
|
||||||
|
|
||||||
def handler_sighting(zmq_name, jsondata):
|
def handler_sighting(zmq_name, jsondata):
|
||||||
print('sending' ,'sighting')
|
logger.info('Handling sighting')
|
||||||
jsonsight = jsondata['Sighting']
|
jsonsight = jsondata['Sighting']
|
||||||
org = jsonsight['Event']['Orgc']['name']
|
org = jsonsight['Event']['Orgc']['name']
|
||||||
categ = jsonsight['Attribute']['category']
|
categ = jsonsight['Attribute']['category']
|
||||||
|
@ -132,6 +136,7 @@ def handler_sighting(zmq_name, jsondata):
|
||||||
trendings_helper.addFalsePositive(timestamp)
|
trendings_helper.addFalsePositive(timestamp)
|
||||||
|
|
||||||
def handler_event(zmq_name, jsonobj):
|
def handler_event(zmq_name, jsonobj):
|
||||||
|
logger.info('Handling event')
|
||||||
#fields: threat_level_id, id, info
|
#fields: threat_level_id, id, info
|
||||||
jsonevent = jsonobj['Event']
|
jsonevent = jsonobj['Event']
|
||||||
|
|
||||||
|
@ -170,6 +175,7 @@ def handler_event(zmq_name, jsonobj):
|
||||||
isLabeled=eventLabeled)
|
isLabeled=eventLabeled)
|
||||||
|
|
||||||
def handler_attribute(zmq_name, jsonobj, hasAlreadyBeenContributed=False):
|
def handler_attribute(zmq_name, jsonobj, hasAlreadyBeenContributed=False):
|
||||||
|
logger.info('Handling attribute')
|
||||||
# check if jsonattr is an attribute object
|
# check if jsonattr is an attribute object
|
||||||
if 'Attribute' in jsonobj:
|
if 'Attribute' in jsonobj:
|
||||||
jsonattr = jsonobj['Attribute']
|
jsonattr = jsonobj['Attribute']
|
||||||
|
@ -227,7 +233,7 @@ def process_log(zmq_name, event):
|
||||||
try:
|
try:
|
||||||
dico_action[topic](zmq_name, jsonevent)
|
dico_action[topic](zmq_name, jsonevent)
|
||||||
except KeyError as e:
|
except KeyError as e:
|
||||||
print(e)
|
logger.error(e)
|
||||||
|
|
||||||
|
|
||||||
def main(sleeptime):
|
def main(sleeptime):
|
||||||
|
@ -235,7 +241,7 @@ def main(sleeptime):
|
||||||
while True:
|
while True:
|
||||||
content = serv_list.rpop(LISTNAME)
|
content = serv_list.rpop(LISTNAME)
|
||||||
if content is None:
|
if content is None:
|
||||||
print('Processed', numMsg, 'message(s) since last sleep.')
|
logger.info('Processed {} message(s) since last sleep.'.format(numMsg))
|
||||||
numMsg = 0
|
numMsg = 0
|
||||||
time.sleep(sleeptime)
|
time.sleep(sleeptime)
|
||||||
continue
|
continue
|
||||||
|
|
|
@ -1,8 +1,8 @@
|
||||||
#!/usr/bin/env python3.5
|
#!/usr/bin/env python3.5
|
||||||
|
|
||||||
import time, datetime
|
import time, datetime
|
||||||
from pprint import pprint
|
|
||||||
import zmq
|
import zmq
|
||||||
|
import logging
|
||||||
import redis
|
import redis
|
||||||
import configparser
|
import configparser
|
||||||
import argparse
|
import argparse
|
||||||
|
@ -14,6 +14,9 @@ configfile = os.path.join(os.environ['DASH_CONFIG'], 'config.cfg')
|
||||||
cfg = configparser.ConfigParser()
|
cfg = configparser.ConfigParser()
|
||||||
cfg.read(configfile)
|
cfg.read(configfile)
|
||||||
|
|
||||||
|
logging.basicConfig(filename='logs/logs.log', filemode='w', level=logging.INFO)
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
ZMQ_URL = cfg.get('RedisGlobal', 'zmq_url')
|
ZMQ_URL = cfg.get('RedisGlobal', 'zmq_url')
|
||||||
CHANNEL = cfg.get('RedisLog', 'channel')
|
CHANNEL = cfg.get('RedisLog', 'channel')
|
||||||
LISTNAME = cfg.get('RedisLIST', 'listName')
|
LISTNAME = cfg.get('RedisLIST', 'listName')
|
||||||
|
@ -32,6 +35,7 @@ def put_in_redis_list(zmq_name, content):
|
||||||
content = content.decode('utf8')
|
content = content.decode('utf8')
|
||||||
to_add = {'zmq_name': zmq_name, 'content': content}
|
to_add = {'zmq_name': zmq_name, 'content': content}
|
||||||
serv_list.lpush(LISTNAME, json.dumps(to_add))
|
serv_list.lpush(LISTNAME, json.dumps(to_add))
|
||||||
|
logger.debug('Pushed: {}'.format(json.dumps(to_add)))
|
||||||
|
|
||||||
def main(zmqName):
|
def main(zmqName):
|
||||||
context = zmq.Context()
|
context = zmq.Context()
|
||||||
|
|
Loading…
Reference in New Issue