mirror of https://github.com/MISP/misp-modules
553 lines
32 KiB
Python
Executable File
553 lines
32 KiB
Python
Executable File
'''
|
|
import
|
|
define mandatory
|
|
|
|
'''
|
|
import json
|
|
import base64
|
|
import re
|
|
import zipfile
|
|
import ipaddress
|
|
import io
|
|
import logging
|
|
|
|
misperrors = {'error': 'Error'}
|
|
userConfig = {}
|
|
inputSource = ['file']
|
|
|
|
moduleinfo = {
|
|
'version': '0.10',
|
|
'author': 'Christophe Vandeplas',
|
|
'description': 'Module to import ThreatAnalyzer archive.zip / analysis.json files.',
|
|
'module-type': ['import'],
|
|
'name': 'ThreadAnalyzer Sandbox Import',
|
|
'logo': '',
|
|
'requirements': [],
|
|
'features': 'The module imports MISP Attributes from a ThreatAnalyzer format file. This file can be either ZIP, or JSON format.\nThere is by the way no special feature for users to make the module work.',
|
|
'references': ['https://www.threattrack.com/malware-analysis.aspx'],
|
|
'input': 'ThreatAnalyzer format file',
|
|
'output': 'MISP Event attributes',
|
|
}
|
|
|
|
moduleconfig = []
|
|
log = logging.getLogger('misp-modules')
|
|
|
|
# FIXME - many hardcoded filters should be migrated to import regexes. See also https://github.com/MISP/MISP/issues/2712
|
|
# DISCLAIMER - This module is to be considered as experimental and needs much fine-tuning.
|
|
# more can be done with what's in the ThreatAnalyzer archive.zip
|
|
|
|
|
|
def handler(q=False):
|
|
if q is False:
|
|
return False
|
|
results = []
|
|
zip_starts = 'PK'
|
|
request = json.loads(q)
|
|
data = base64.b64decode(request['data'])
|
|
|
|
if data[:len(zip_starts)].decode() == zip_starts:
|
|
with zipfile.ZipFile(io.BytesIO(data), 'r') as zf:
|
|
# unzipped_files = []
|
|
modified_files_mapping = {}
|
|
# pre-process some of the files in the zip
|
|
for zip_file_name in zf.namelist(): # Get all files in the zip file
|
|
# find the filenames of the modified_files
|
|
if re.match(r"Analysis/proc_\d+/modified_files/mapping\.log", zip_file_name):
|
|
with zf.open(zip_file_name, mode='r', pwd=None) as fp:
|
|
file_data = fp.read()
|
|
for line in file_data.decode("utf-8", 'ignore').split('\n'):
|
|
if not line:
|
|
continue
|
|
if line.count('|') == 3:
|
|
l_fname, l_size, l_md5, l_created = line.split('|')
|
|
if line.count('|') == 4:
|
|
l_fname, l_size, l_md5, l_sha256, l_created = line.split('|')
|
|
l_fname = cleanup_filepath(l_fname)
|
|
if l_fname:
|
|
if l_size == 0:
|
|
results.append({'values': l_fname, 'type': 'filename', 'to_ids': True,
|
|
'categories': ['Artifacts dropped', 'Payload delivery'], 'comment': ''})
|
|
else:
|
|
# file is a non empty sample, upload the sample later
|
|
modified_files_mapping[l_md5] = l_fname
|
|
|
|
# now really process the data
|
|
for zip_file_name in zf.namelist(): # Get all files in the zip file
|
|
# print('Processing file: {}'.format(zip_file_name))
|
|
if re.match(r"Analysis/proc_\d+/modified_files/.+\.", zip_file_name) and "mapping.log" not in zip_file_name:
|
|
sample_md5 = zip_file_name.split('/')[-1].split('.')[0]
|
|
if sample_md5 in modified_files_mapping:
|
|
current_sample_filename = modified_files_mapping[sample_md5]
|
|
# print("{} maps to {}".format(sample_md5, current_sample_filename))
|
|
with zf.open(zip_file_name, mode='r', pwd=None) as fp:
|
|
file_data = fp.read()
|
|
results.append({
|
|
'values': current_sample_filename,
|
|
'data': base64.b64encode(file_data).decode(),
|
|
'type': 'malware-sample', 'categories': ['Artifacts dropped', 'Payload delivery'], 'to_ids': True, 'comment': ''})
|
|
|
|
if 'Analysis/analysis.json' in zip_file_name:
|
|
with zf.open(zip_file_name, mode='r', pwd=None) as fp:
|
|
file_data = fp.read()
|
|
analysis_json = json.loads(file_data.decode('utf-8'))
|
|
results += process_analysis_json(analysis_json)
|
|
try:
|
|
sample_filename = analysis_json.get('analysis').get('@filename')
|
|
if sample_filename:
|
|
with zf.open('sample', mode='r', pwd=None) as fp:
|
|
file_data = fp.read()
|
|
results.append({
|
|
'values': sample_filename,
|
|
'data': base64.b64encode(file_data).decode(),
|
|
'type': 'malware-sample', 'categories': ['Payload delivery', 'Artifacts dropped'], 'to_ids': True, 'comment': ''})
|
|
except Exception:
|
|
# no 'sample' in archive, might be an url analysis, just ignore
|
|
pass
|
|
|
|
else:
|
|
try:
|
|
results = process_analysis_json(json.loads(data.decode('utf-8')))
|
|
except ValueError:
|
|
log.warning('MISP modules {0} failed: uploaded file is not a zip or json file.'.format(request['module']))
|
|
return {'error': 'Uploaded file is not a zip or json file.'}
|
|
pass
|
|
# keep only unique entries based on the value field
|
|
results = list({v['values']: v for v in results}.values())
|
|
r = {'results': results}
|
|
return r
|
|
|
|
|
|
def process_analysis_json(analysis_json):
|
|
if 'analysis' in analysis_json and 'processes' in analysis_json['analysis'] and 'process' in analysis_json['analysis']['processes']:
|
|
# if 'analysis' in analysis_json and '@filename' in analysis_json['analysis']:
|
|
# sample['values'] = analysis_json['analysis']['@filename']
|
|
for process in analysis_json['analysis']['processes']['process']:
|
|
# print_json(process)
|
|
if 'connection_section' in process and 'connection' in process['connection_section']:
|
|
# compensate for absurd behavior of the data format: if one entry = immediately the dict, if multiple entries = list containing dicts
|
|
# this will always create a list, even with only one item
|
|
if isinstance(process['connection_section']['connection'], dict):
|
|
process['connection_section']['connection'] = [process['connection_section']['connection']]
|
|
|
|
# iterate over each entry
|
|
for connection_section_connection in process['connection_section']['connection']:
|
|
# compensate for absurd behavior of the data format: if one entry = immediately the dict, if multiple entries = list containing dicts
|
|
# this will always create a list, even with only one item
|
|
for subsection in ['http_command', 'http_header']:
|
|
if isinstance(connection_section_connection[subsection], dict):
|
|
connection_section_connection[subsection] = [connection_section_connection[subsection]]
|
|
|
|
if 'name_to_ip' in connection_section_connection: # TA 6.1 data format
|
|
connection_section_connection['@remote_ip'] = connection_section_connection['name_to_ip']['@result_addresses']
|
|
connection_section_connection['@remote_hostname'] = connection_section_connection['name_to_ip']['@request_name']
|
|
|
|
connection_section_connection['@remote_ip'] = cleanup_ip(connection_section_connection['@remote_ip'])
|
|
connection_section_connection['@remote_hostname'] = cleanup_hostname(connection_section_connection['@remote_hostname'])
|
|
if connection_section_connection['@remote_ip'] and connection_section_connection['@remote_hostname']:
|
|
val = '{}|{}'.format(connection_section_connection['@remote_hostname'],
|
|
connection_section_connection['@remote_ip'])
|
|
# print("connection_section_connection hostname|ip: {}|{} IDS:yes".format(
|
|
# connection_section_connection['@remote_hostname'],
|
|
# connection_section_connection['@remote_ip'])
|
|
# )
|
|
yield({'values': val, 'type': 'domain|ip', 'categories': ['Network activity'], 'to_ids': True, 'comment': ''})
|
|
elif connection_section_connection['@remote_ip']:
|
|
# print("connection_section_connection ip-dst: {} IDS:yes".format(
|
|
# connection_section_connection['@remote_ip'])
|
|
# )
|
|
yield({'values': connection_section_connection['@remote_ip'], 'type': 'ip-dst', 'to_ids': True, 'comment': ''})
|
|
elif connection_section_connection['@remote_hostname']:
|
|
# print("connection_section_connection hostname: {} IDS:yes".format(
|
|
# connection_section_connection['@remote_hostname'])
|
|
# )
|
|
yield({'values': connection_section_connection['@remote_hostname'], 'type': 'hostname', 'to_ids': True, 'comment': ''})
|
|
if 'http_command' in connection_section_connection:
|
|
for http_command in connection_section_connection['http_command']:
|
|
# print('connection_section_connection HTTP COMMAND: {}\t{}'.format(
|
|
# connection_section_connection['http_command']['@method'], # comment
|
|
# connection_section_connection['http_command']['@url']) # url
|
|
# )
|
|
val = cleanup_url(http_command['@url'])
|
|
if val:
|
|
yield({'values': val, 'type': 'url', 'categories': ['Network activity'], 'to_ids': True, 'comment': http_command['@method']})
|
|
|
|
if 'http_header' in connection_section_connection:
|
|
for http_header in connection_section_connection['http_header']:
|
|
if 'User-Agent:' in http_header['@header']:
|
|
val = http_header['@header'][len('User-Agent: '):]
|
|
yield({'values': val, 'type': 'user-agent', 'categories': ['Network activity'], 'to_ids': False, 'comment': ''})
|
|
elif 'Host:' in http_header['@header']:
|
|
val = http_header['@header'][len('Host: '):]
|
|
if ':' in val:
|
|
try:
|
|
val_port = int(val.split(':')[1])
|
|
except ValueError:
|
|
val_port = False
|
|
val_hostname = cleanup_hostname(val.split(':')[0])
|
|
val_ip = cleanup_ip(val.split(':')[0])
|
|
if val_hostname and val_port:
|
|
val_combined = '{}|{}'.format(val_hostname, val_port)
|
|
# print({'values': val_combined, 'type': 'hostname|port', 'to_ids': True, 'comment': ''})
|
|
yield({'values': val_combined, 'type': 'hostname|port', 'categories': ['Network activity'], 'to_ids': True, 'comment': ''})
|
|
elif val_ip and val_port:
|
|
val_combined = '{}|{}'.format(val_ip, val_port)
|
|
# print({'values': val_combined, 'type': 'ip-dst|port', 'to_ids': True, 'comment': ''})
|
|
yield({'values': val_combined, 'type': 'ip-dst|port', 'to_ids': True, 'comment': ''})
|
|
else:
|
|
continue
|
|
val_hostname = cleanup_hostname(val)
|
|
if val_hostname:
|
|
# print({'values': val_hostname, 'type': 'hostname', 'to_ids': True, 'comment': ''})
|
|
yield({'values': val_hostname, 'type': 'hostname', 'to_ids': True, 'comment': ''})
|
|
else:
|
|
# LATER header not processed
|
|
pass
|
|
if 'filesystem_section' in process and 'create_file' in process['filesystem_section']:
|
|
for filesystem_section_create_file in process['filesystem_section']['create_file']:
|
|
# first skip some items
|
|
if filesystem_section_create_file['@create_disposition'] in {'FILE_OPEN_IF'}:
|
|
continue
|
|
# FIXME - this section is probably not needed considering the 'stored_files stored_created_file' section we process later.
|
|
# print('CREATE FILE: {}\t{}'.format(
|
|
# filesystem_section_create_file['@srcfile'], # filename
|
|
# filesystem_section_create_file['@create_disposition']) # comment - use this to filter out cases
|
|
# )
|
|
|
|
if 'networkoperation_section' in process and 'dns_request_by_addr' in process['networkoperation_section']:
|
|
for networkoperation_section_dns_request_by_addr in process['networkoperation_section']['dns_request_by_addr']:
|
|
# FIXME - it's unclear what this section is for.
|
|
# TODO filter this
|
|
# print('DNS REQUEST: {}\t{}'.format(
|
|
# networkoperation_section_dns_request_by_addr['@request_address'], # ip-dst
|
|
# networkoperation_section_dns_request_by_addr['@result_name']) # hostname
|
|
# ) # => NOT hostname|ip
|
|
pass
|
|
if 'networkoperation_section' in process and 'dns_request_by_name' in process['networkoperation_section']:
|
|
for networkoperation_section_dns_request_by_name in process['networkoperation_section']['dns_request_by_name']:
|
|
networkoperation_section_dns_request_by_name['@request_name'] = cleanup_hostname(networkoperation_section_dns_request_by_name['@request_name'].rstrip('.'))
|
|
networkoperation_section_dns_request_by_name['@result_addresses'] = cleanup_ip(networkoperation_section_dns_request_by_name['@result_addresses'])
|
|
if networkoperation_section_dns_request_by_name['@request_name'] and networkoperation_section_dns_request_by_name['@result_addresses']:
|
|
val = '{}|{}'.format(networkoperation_section_dns_request_by_name['@request_name'],
|
|
networkoperation_section_dns_request_by_name['@result_addresses'])
|
|
# print("networkoperation_section_dns_request_by_name hostname|ip: {}|{} IDS:yes".format(
|
|
# networkoperation_section_dns_request_by_name['@request_name'],
|
|
# networkoperation_section_dns_request_by_name['@result_addresses'])
|
|
# )
|
|
yield({'values': val, 'type': 'domain|ip', 'categories': ['Network activity'], 'to_ids': True, 'comment': ''})
|
|
elif networkoperation_section_dns_request_by_name['@request_name']:
|
|
# print("networkoperation_section_dns_request_by_name hostname: {} IDS:yes".format(
|
|
# networkoperation_section_dns_request_by_name['@request_name'])
|
|
# )
|
|
yield({'values': networkoperation_section_dns_request_by_name['@request_name'], 'type': 'hostname', 'to_ids': True, 'comment': ''})
|
|
elif networkoperation_section_dns_request_by_name['@result_addresses']:
|
|
# this happens when the IP is both in the request_name and result_address.
|
|
# print("networkoperation_section_dns_request_by_name hostname: {} IDS:yes".format(
|
|
# networkoperation_section_dns_request_by_name['@result_addresses'])
|
|
# )
|
|
yield({'values': networkoperation_section_dns_request_by_name['@result_addresses'], 'type': 'ip-dst', 'to_ids': True, 'comment': ''})
|
|
|
|
if 'networkpacket_section' in process and 'connect_to_computer' in process['networkpacket_section']:
|
|
for networkpacket_section_connect_to_computer in process['networkpacket_section']['connect_to_computer']:
|
|
networkpacket_section_connect_to_computer['@remote_hostname'] = cleanup_hostname(networkpacket_section_connect_to_computer['@remote_hostname'])
|
|
networkpacket_section_connect_to_computer['@remote_ip'] = cleanup_ip(networkpacket_section_connect_to_computer['@remote_ip'])
|
|
if networkpacket_section_connect_to_computer['@remote_hostname'] and networkpacket_section_connect_to_computer['@remote_ip']:
|
|
# print("networkpacket_section_connect_to_computer hostname|ip: {}|{} IDS:yes COMMENT:port {}".format(
|
|
# networkpacket_section_connect_to_computer['@remote_hostname'],
|
|
# networkpacket_section_connect_to_computer['@remote_ip'],
|
|
# networkpacket_section_connect_to_computer['@remote_port'])
|
|
# )
|
|
val_combined = "{}|{}".format(networkpacket_section_connect_to_computer['@remote_hostname'], networkpacket_section_connect_to_computer['@remote_ip'])
|
|
yield({'values': val_combined, 'type': 'domain|ip', 'to_ids': True, 'comment': ''})
|
|
elif networkpacket_section_connect_to_computer['@remote_hostname']:
|
|
# print("networkpacket_section_connect_to_computer hostname: {} IDS:yes COMMENT:port {}".format(
|
|
# networkpacket_section_connect_to_computer['@remote_hostname'],
|
|
# networkpacket_section_connect_to_computer['@remote_port'])
|
|
# )
|
|
val_combined = "{}|{}".format(networkpacket_section_connect_to_computer['@remote_hostname'], networkpacket_section_connect_to_computer['@remote_port'])
|
|
yield({'values': val_combined, 'type': 'hostname|port', 'categories': ['Network activity'], 'to_ids': True, 'comment': ''})
|
|
elif networkpacket_section_connect_to_computer['@remote_ip']:
|
|
# print("networkpacket_section_connect_to_computer ip-dst: {} IDS:yes COMMENT:port {}".format(
|
|
# networkpacket_section_connect_to_computer['@remote_ip'],
|
|
# networkpacket_section_connect_to_computer['@remote_port'])
|
|
# )
|
|
val_combined = "{}|{}".format(networkpacket_section_connect_to_computer['@remote_ip'], networkpacket_section_connect_to_computer['@remote_port'])
|
|
yield({'values': val_combined, 'type': 'ip-dst|port', 'to_ids': True, 'comment': ''})
|
|
|
|
if 'registry_section' in process and 'create_key' in process['registry_section']:
|
|
# FIXME this is a complicated section, together with the 'set_value'.
|
|
# it looks like this section is not ONLY about creating registry keys,
|
|
# more about accessing a handle to keys (with specific permissions)
|
|
# maybe we don't want to keep this, in favor of 'set_value'
|
|
for create_key in process['registry_section']['create_key']:
|
|
# print('REG CREATE: {}\t{}'.format(
|
|
# create_key['@desired_access'],
|
|
# create_key['@key_name']))
|
|
pass
|
|
if 'registry_section' in process and 'delete_key' in process['registry_section']:
|
|
# LATER we probably don't want to keep this. Much pollution.
|
|
# Maybe for later once we have filtered out this.
|
|
for delete_key in process['registry_section']['delete_key']:
|
|
# print('REG DELETE: {}'.format(
|
|
# delete_key['@key_name'])
|
|
# )
|
|
pass
|
|
if 'registry_section' in process and 'set_value' in process['registry_section']:
|
|
# FIXME this is a complicated section, together with the 'create_key'.
|
|
for set_value in process['registry_section']['set_value']:
|
|
# '@data_type' == 'REG_BINARY',
|
|
# '@data_type' == 'REG_DWORD',
|
|
# '@data_type' == 'REG_EXPAND_SZ',
|
|
# '@data_type' == 'REG_MULTI_SZ',
|
|
# '@data_type' == 'REG_NONE',
|
|
# '@data_type' == 'REG_QWORD',
|
|
# '@data_type' == 'REG_SZ',
|
|
regkey = cleanup_regkey("{}\\{}".format(set_value['@key_name'], set_value['@value_name']))
|
|
regdata = cleanup_regdata(set_value.get('@data'))
|
|
if not regkey:
|
|
continue
|
|
if set_value['@data_size'] == '0' or not regdata:
|
|
# print('registry_section set_value REG SET: {}\t{}\t{}'.format(
|
|
# set_value['@data_type'],
|
|
# set_value['@key_name'],
|
|
# set_value['@value_name'])
|
|
# )
|
|
yield({'values': regkey, 'type': 'regkey', 'to_ids': True,
|
|
'categories': ['External analysis', 'Persistence mechanism', 'Artifacts dropped'], 'comment': set_value['@data_type']})
|
|
else:
|
|
try:
|
|
# unicode fun...
|
|
# print('registry_section set_value REG SET: {}\t{}\t{}\t{}'.format(
|
|
# set_value['@data_type'],
|
|
# set_value['@key_name'],
|
|
# set_value['@value_name'],
|
|
# set_value['@data'])
|
|
# )
|
|
val = "{}|{}".format(regkey, regdata)
|
|
yield({'values': val, 'type': 'regkey|value', 'to_ids': True,
|
|
'categories': ['External analysis', 'Persistence mechanism', 'Artifacts dropped'], 'comment': set_value['@data_type']})
|
|
except Exception as e:
|
|
print("EXCEPTION registry_section {}".format(e))
|
|
# TODO - maybe we want to handle these later, or not...
|
|
pass
|
|
pass
|
|
|
|
if 'stored_files' in process and 'stored_created_file' in process['stored_files']:
|
|
for stored_created_file in process['stored_files']['stored_created_file']:
|
|
stored_created_file['@filename'] = cleanup_filepath(stored_created_file['@filename'])
|
|
if stored_created_file['@filename']:
|
|
if stored_created_file['@filesize'] != '0':
|
|
val = '{}|{}'.format(stored_created_file['@filename'], stored_created_file['@md5'])
|
|
# print("stored_created_file filename|md5: {}|{} IDS:yes".format(
|
|
# stored_created_file['@filename'], # filename
|
|
# stored_created_file['@md5']) # md5
|
|
# ) # => filename|md5
|
|
yield({'values': val, 'type': 'filename|md5', 'to_ids': True,
|
|
'categories': ['Artifacts dropped', 'Payload delivery'], 'comment': ''})
|
|
|
|
else:
|
|
# print("stored_created_file filename: {} IDS:yes".format(
|
|
# stored_created_file['@filename']) # filename
|
|
# ) # => filename
|
|
yield({'values': stored_created_file['@filename'],
|
|
'type': 'filename', 'to_ids': True,
|
|
'categories': ['Artifacts dropped', 'Payload delivery'], 'comment': ''})
|
|
|
|
if 'stored_files' in process and 'stored_modified_file' in process['stored_files']:
|
|
for stored_modified_file in process['stored_files']['stored_modified_file']:
|
|
stored_modified_file['@filename'] = cleanup_filepath(stored_modified_file['@filename'])
|
|
if stored_modified_file['@filename']:
|
|
if stored_modified_file['@filesize'] != '0':
|
|
val = '{}|{}'.format(stored_modified_file['@filename'], stored_modified_file['@md5'])
|
|
# print("stored_modified_file MODIFY FILE: {}\t{}".format(
|
|
# stored_modified_file['@filename'], # filename
|
|
# stored_modified_file['@md5']) # md5
|
|
# ) # => filename|md5
|
|
yield({'values': val, 'type': 'filename|md5', 'to_ids': True,
|
|
'categories': ['Artifacts dropped', 'Payload delivery'],
|
|
'comment': 'modified'})
|
|
else:
|
|
# print("stored_modified_file MODIFY FILE: {}\t{}".format(
|
|
# stored_modified_file['@filename']) # filename
|
|
# ) # => filename
|
|
yield({'values': stored_modified_file['@filename'], 'type': 'filename', 'to_ids': True,
|
|
'categories': ['Artifacts dropped', 'Payload delivery'],
|
|
'comment': 'modified'})
|
|
|
|
|
|
def add_file(filename, results, hash, index, filedata=None):
|
|
pass
|
|
# results.append({'values': filename, 'data': "{}|{}".format(filename, filedata.decode()), 'type': 'malware-sample',
|
|
# 'categories': ['Artifacts dropped', 'Payload delivery']})
|
|
|
|
|
|
def add_file_zip():
|
|
# if 'malware-sample' in request:
|
|
# sample_filename = request.get("malware-sample").split("|", 1)[0]
|
|
# data = base64.b64decode(data)
|
|
# fl = io.BytesIO(data)
|
|
# zf = zipfile.ZipFile(fl)
|
|
# sample_hashname = zf.namelist()[0]
|
|
# data = zf.read(sample_hashname, b"infected")
|
|
# zf.close()
|
|
pass
|
|
|
|
|
|
def print_json(data):
|
|
print(json.dumps(data, sort_keys=True, indent=4, separators=(',', ': ')))
|
|
|
|
|
|
def list_in_string(lst, data, regex=False):
|
|
for item in lst:
|
|
if regex:
|
|
if re.search(item, data, flags=re.IGNORECASE):
|
|
return True
|
|
else:
|
|
if item in data:
|
|
return True
|
|
|
|
|
|
def cleanup_ip(item):
|
|
# you should exclude private IP ranges via import regexes
|
|
noise_substrings = {
|
|
'224.0.0.',
|
|
'127.0.0.',
|
|
'8.8.8.8',
|
|
'8.8.4.4',
|
|
'0.0.0.0',
|
|
'NONE'
|
|
}
|
|
if list_in_string(noise_substrings, item):
|
|
return None
|
|
try:
|
|
ipaddress.ip_address(item)
|
|
return item
|
|
except ValueError:
|
|
return None
|
|
|
|
|
|
def cleanup_hostname(item):
|
|
noise_substrings = {
|
|
'wpad',
|
|
'teredo.ipv6.microsoft.com',
|
|
'WIN7SP1-x64-UNP'
|
|
}
|
|
# take away common known bad
|
|
if list_in_string(noise_substrings, item):
|
|
return None
|
|
# eliminate IP addresses
|
|
try:
|
|
ipaddress.ip_address(item)
|
|
except ValueError:
|
|
# this is not an IP, so continue
|
|
return item
|
|
return None
|
|
|
|
|
|
def cleanup_url(item):
|
|
if item in ['/']:
|
|
return None
|
|
return item
|
|
|
|
|
|
def cleanup_filepath(item):
|
|
noise_substrings = {
|
|
'\\AppData\\Local\\GDIPFONTCACHEV1.DAT',
|
|
'\\AppData\\Local\\Microsoft\\Internet Explorer\\DOMStore\\',
|
|
'\\AppData\\Local\\Microsoft\\Internet Explorer\\Recovery\\High\\',
|
|
'\\AppData\\Local\\Microsoft\\Windows\\Caches\\',
|
|
'\\AppData\\Local\\Microsoft\\Windows\\Explorer\\thumbcache',
|
|
'\\AppData\\Local\\Microsoft\\Windows\\History\\History.',
|
|
'\\AppData\\Local\\Microsoft\\Windows\\Temporary Internet Files\\Content.',
|
|
'\\AppData\\Local\\Microsoft\\Windows\\WebCache\\',
|
|
'\\AppData\\Local\\Temp\\.*tmp$',
|
|
'\\AppData\\LocalLow\\Microsoft\\CryptnetUrlCache\\',
|
|
'\\AppData\\LocalLow\\Microsoft\\Internet Explorer\\Services\\search_',
|
|
'\\AppData\\Roaming\\Microsoft\\Office\\Recent\\',
|
|
'\\AppData\\Roaming\\Microsoft\\Windows\\Cookies\\',
|
|
'\\AppData\\Roaming\\Microsoft\\Windows\\Recent\\',
|
|
'C:\\ProgramData\\Microsoft\\OfficeSoftwareProtectionPlatform\\Cache\\cache.dat',
|
|
'C:\\Windows\\Prefetch\\',
|
|
|
|
'\\AppData\\Roaming\\Adobe\\Acrobat\\9.0\\SharedDataEvents-journal',
|
|
'\\AppData\\Roaming\\Adobe\\Acrobat\\9.0\\UserCache.bin',
|
|
|
|
'\\AppData\\Roaming\\Macromedia\\Flash Player\\macromedia.com\\support\\flashplayer\\sys\\settings.sol',
|
|
'\\AppData\\Roaming\\Adobe\\Flash Player\\NativeCache\\',
|
|
'C:\\Windows\\AppCompat\\Programs\\',
|
|
'C:\\~' # caused by temp file created by MS Office when opening malicious doc/xls/...
|
|
}
|
|
if list_in_string(noise_substrings, item):
|
|
return None
|
|
return item
|
|
|
|
|
|
def cleanup_regkey(item):
|
|
noise_substrings = {
|
|
r'\\CurrentVersion\\Explorer\\FileExts\\[a-z\.]+\\OpenWith',
|
|
r'\\CurrentVersion\\Explorer\\RecentDocs\\',
|
|
r'\\CurrentVersion\\Explorer\\UserAssist\\',
|
|
r'\\Local Settings\\Software\\Microsoft\\Windows\\Shell\\Bag',
|
|
r'\\Software\\Classes\\CLSID\\',
|
|
r'\\Software\\Classes\\Local Settings\\MuiCache\\',
|
|
r'\\Software\\Microsoft\\Internet Explorer\\Main\\WindowsSearch',
|
|
r'\\Software\\Microsoft\\Office\\[0-9\.]+\\',
|
|
r'\\Software\\Microsoft\\Office\\Common\\Smart Tag\\',
|
|
r'\\Software\\Microsoft\\OfficeSoftwareProtectionPlatform\\',
|
|
r'\\Software\\Microsoft\\Shared Tools\\Panose\\',
|
|
r'\\Software\\Microsoft\\Tracing\\',
|
|
r'\\Software\\Microsoft\\Tracing\\powershell_RASAPI32\\',
|
|
r'\\Software\\Microsoft\\Tracing\\powershell_RASMANCS\\',
|
|
r'\\Software\\Microsoft\\Windows\\CurrentVersion\\Action Center\\',
|
|
r'\\Software\\Microsoft\\Windows\\CurrentVersion\\Explorer\\RunMRU\\',
|
|
r'\\Software\\Microsoft\\Windows\\CurrentVersion\\Installer\\UserData\\',
|
|
r'\\Software\\Microsoft\\Windows\\CurrentVersion\\Internet Settings\\',
|
|
r'\\System\\CurrentControlSet\\Services\\RdyBoost\\',
|
|
r'\\Usage\\SpellingAndGrammarFiles'
|
|
}
|
|
if list_in_string(noise_substrings, item, regex=True):
|
|
return None
|
|
return item
|
|
|
|
|
|
def cleanup_regdata(item):
|
|
if not item:
|
|
return None
|
|
item = item.replace('(UNICODE_0x00000000)', '')
|
|
return item
|
|
|
|
|
|
def get_zipped_contents(filename, data, password=None):
|
|
with zipfile.ZipFile(io.BytesIO(data), 'r') as zf:
|
|
unzipped_files = []
|
|
if password is not None:
|
|
password = str.encode(password) # Byte encoded password required
|
|
for zip_file_name in zf.namelist(): # Get all files in the zip file
|
|
# print(zip_file_name)
|
|
with zf.open(zip_file_name, mode='r', pwd=password) as fp:
|
|
file_data = fp.read()
|
|
unzipped_files.append({'values': zip_file_name,
|
|
'data': file_data,
|
|
'comment': 'Extracted from {0}'.format(filename)})
|
|
# print("{} : {}".format(zip_file_name, len(file_data)))
|
|
return unzipped_files
|
|
|
|
|
|
def introspection():
|
|
modulesetup = {}
|
|
try:
|
|
userConfig
|
|
modulesetup['userConfig'] = userConfig
|
|
except NameError:
|
|
pass
|
|
try:
|
|
inputSource
|
|
modulesetup['inputSource'] = inputSource
|
|
except NameError:
|
|
pass
|
|
return modulesetup
|
|
|
|
|
|
def version():
|
|
moduleinfo['config'] = moduleconfig
|
|
return moduleinfo
|