mirror of https://github.com/CIRCL/AIL-framework
2106 lines
70 KiB
Python
Executable File
2106 lines
70 KiB
Python
Executable File
#!/usr/bin/python3
|
|
|
|
"""
|
|
API Helper
|
|
===================
|
|
|
|
|
|
"""
|
|
import base64
|
|
import gzip
|
|
import hashlib
|
|
import json
|
|
import os
|
|
import pickle
|
|
import re
|
|
import sys
|
|
import time
|
|
import uuid
|
|
|
|
from enum import IntEnum, unique
|
|
from datetime import datetime, timedelta
|
|
from dateutil.relativedelta import relativedelta
|
|
from urllib.parse import urlparse, urljoin
|
|
from bs4 import BeautifulSoup
|
|
|
|
from pylacus import PyLacus
|
|
|
|
from pyfaup.faup import Faup
|
|
|
|
# interact with splash_crawler API
|
|
import requests
|
|
requests.packages.urllib3.disable_warnings(requests.packages.urllib3.exceptions.InsecureRequestWarning)
|
|
|
|
sys.path.append(os.environ['AIL_BIN'])
|
|
##################################
|
|
# Import Project packages
|
|
##################################
|
|
from packages import git_status
|
|
from packages import Date
|
|
from lib.ConfigLoader import ConfigLoader
|
|
from lib.objects.Domains import Domain
|
|
from lib.objects import HHHashs
|
|
from lib.objects.Items import Item
|
|
|
|
config_loader = ConfigLoader()
|
|
r_db = config_loader.get_db_conn("Kvrocks_DB")
|
|
r_crawler = config_loader.get_db_conn("Kvrocks_Crawler")
|
|
r_cache = config_loader.get_redis_conn("Redis_Cache")
|
|
|
|
ITEMS_FOLDER = config_loader.get_config_str("Directories", "pastes")
|
|
HAR_DIR = config_loader.get_files_directory('har')
|
|
activate_crawler = config_loader.get_config_str("Crawler", "activate_crawler")
|
|
config_loader = None
|
|
|
|
faup = Faup()
|
|
|
|
# # # # # # # #
|
|
# #
|
|
# COMMON #
|
|
# #
|
|
# # # # # # # #
|
|
|
|
def gen_uuid():
|
|
return str(uuid.uuid4())
|
|
|
|
def generate_uuid():
|
|
return str(uuid.uuid4()).replace('-', '')
|
|
|
|
# # TODO: remove me ?
|
|
def get_current_date(separator=False):
|
|
if separator:
|
|
return datetime.now().strftime("%Y/%m/%d")
|
|
else:
|
|
return datetime.now().strftime("%Y%m%d")
|
|
|
|
def get_date_crawled_items_source(date):
|
|
return os.path.join('crawled', date)
|
|
|
|
def get_har_dir():
|
|
return HAR_DIR
|
|
|
|
def is_valid_onion_domain(domain):
|
|
if not domain.endswith('.onion'):
|
|
return False
|
|
domain = domain.replace('.onion', '', 1)
|
|
if len(domain) == 16: # v2 address
|
|
r_onion = r'[a-z0-9]{16}'
|
|
if re.match(r_onion, domain):
|
|
return True
|
|
elif len(domain) == 56: # v3 address
|
|
r_onion = r'[a-z0-9]{56}'
|
|
if re.fullmatch(r_onion, domain):
|
|
return True
|
|
return False
|
|
|
|
def is_valid_domain(domain):
|
|
faup.decode(domain)
|
|
url_unpack = faup.get()
|
|
unpack_domain = url_unpack['domain'].lower()
|
|
return domain == unpack_domain
|
|
|
|
def get_faup():
|
|
return faup
|
|
|
|
def unpack_url(url):
|
|
f = get_faup()
|
|
f.decode(url)
|
|
url_decoded = f.get()
|
|
port = url_decoded['port']
|
|
if not port:
|
|
if url_decoded['scheme'] == 'http':
|
|
port = 80
|
|
elif url_decoded['scheme'] == 'https':
|
|
port = 443
|
|
else:
|
|
port = 80
|
|
url_decoded['port'] = port
|
|
# decode URL
|
|
try:
|
|
url = url_decoded['url'].decode()
|
|
except AttributeError:
|
|
url = url_decoded['url']
|
|
# if not url_decoded['scheme']:
|
|
# url = f'http://{url}'
|
|
|
|
# Fix case
|
|
url_decoded['domain'] = url_decoded['domain'].lower()
|
|
url_decoded['url'] = url.replace(url_decoded['host'], url_decoded['host'].lower(), 1)
|
|
return url_decoded
|
|
|
|
# # # # # # # #
|
|
# #
|
|
# FAVICON # TODO REWRITE ME
|
|
# #
|
|
# # # # # # # # TODO CREATE NEW OBJECT
|
|
|
|
def get_favicon_from_html(html, domain, url):
|
|
favicon_urls, favicons = extract_favicon_from_html(html, url)
|
|
# add root favicon
|
|
if not favicon_urls:
|
|
favicon_urls.add(f'{urlparse(url).scheme}://{domain}/favicon.ico')
|
|
print(favicon_urls)
|
|
return favicon_urls
|
|
|
|
def extract_favicon_from_html(html, url):
|
|
favicons = set()
|
|
favicons_urls = set()
|
|
|
|
soup = BeautifulSoup(html, 'html.parser')
|
|
all_icons = set()
|
|
# If there are multiple <link rel="icon">s, the browser uses their media,
|
|
# type, and sizes attributes to select the most appropriate icon.
|
|
# If several icons are equally appropriate, the last one is used.
|
|
# If the most appropriate icon is later found to be inappropriate,
|
|
# for example because it uses an unsupported format,
|
|
# the browser proceeds to the next-most appropriate, and so on.
|
|
# # DEBUG: /!\ firefox load all favicon ???
|
|
|
|
# iOS Safari 'apple-touch-icon'
|
|
# Safari pinned tabs 'mask-icon'
|
|
# Android Chrome 'manifest'
|
|
# Edge and IE 12:
|
|
# - <meta name="msapplication-TileColor" content="#aaaaaa"> <meta name="theme-color" content="#ffffff">
|
|
# - <meta name="msapplication-config" content="/icons/browserconfig.xml">
|
|
|
|
# Root Favicon
|
|
f = get_faup()
|
|
f.decode(url)
|
|
url_decoded = f.get()
|
|
root_domain = f"{url_decoded['scheme']}://{url_decoded['domain']}"
|
|
default_icon = f'{root_domain}/favicon.ico'
|
|
favicons_urls.add(default_icon)
|
|
# print(default_icon)
|
|
|
|
# shortcut
|
|
for shortcut in soup.find_all('link', rel='shortcut icon'):
|
|
all_icons.add(shortcut)
|
|
# icons
|
|
for icon in soup.find_all('link', rel='icon'):
|
|
all_icons.add(icon)
|
|
|
|
for mask_icon in soup.find_all('link', rel='mask-icon'):
|
|
all_icons.add(mask_icon)
|
|
for apple_touche_icon in soup.find_all('link', rel='apple-touch-icon'):
|
|
all_icons.add(apple_touche_icon)
|
|
for msapplication in soup.find_all('meta', attrs={'name': 'msapplication-TileImage'}): # msapplication-TileColor
|
|
all_icons.add(msapplication)
|
|
|
|
# msapplication-TileImage
|
|
|
|
# print(all_icons)
|
|
for tag in all_icons:
|
|
icon_url = tag.get('href')
|
|
if icon_url:
|
|
if icon_url.startswith('data:'):
|
|
data = icon_url.split(',', 1)
|
|
if len(data) > 1:
|
|
data = ''.join(data[1].split())
|
|
favicon = base64.b64decode(data)
|
|
if favicon:
|
|
favicons.add(favicon)
|
|
else:
|
|
favicon_url = urljoin(url, icon_url)
|
|
favicons_urls.add(favicon_url)
|
|
elif tag.get('name') == 'msapplication-TileImage':
|
|
icon_url = tag.get('content')
|
|
if icon_url:
|
|
if icon_url.startswith('data:'):
|
|
data = icon_url.split(',', 1)
|
|
if len(data) > 1:
|
|
data = ''.join(data[1].split())
|
|
favicon = base64.b64decode(data)
|
|
if favicon:
|
|
favicons.add(favicon)
|
|
else:
|
|
favicon_url = urljoin(url, icon_url)
|
|
favicons_urls.add(favicon_url)
|
|
print(favicon_url)
|
|
|
|
# print(favicons_urls)
|
|
return favicons_urls, favicons
|
|
|
|
# mmh3.hash(favicon)
|
|
|
|
# # # - - # # #
|
|
|
|
# # # # # # # #
|
|
# #
|
|
# TITLE #
|
|
# #
|
|
# # # # # # # #
|
|
|
|
def extract_title_from_html(html):
|
|
soup = BeautifulSoup(html, 'html.parser')
|
|
title = soup.title
|
|
if title:
|
|
title = title.string
|
|
if title:
|
|
return str(title)
|
|
return ''
|
|
|
|
def extract_description_from_html(html):
|
|
soup = BeautifulSoup(html, 'html.parser')
|
|
description = soup.find('meta', attrs={'name': 'description'})
|
|
if description:
|
|
return description['content']
|
|
return ''
|
|
|
|
def extract_keywords_from_html(html):
|
|
soup = BeautifulSoup(html, 'html.parser')
|
|
keywords = soup.find('meta', attrs={'name': 'keywords'})
|
|
if keywords:
|
|
return keywords['content']
|
|
return ''
|
|
|
|
def extract_author_from_html(html):
|
|
soup = BeautifulSoup(html, 'html.parser')
|
|
keywords = soup.find('meta', attrs={'name': 'author'})
|
|
if keywords:
|
|
return keywords['content']
|
|
return ''
|
|
|
|
# # # - - # # #
|
|
|
|
|
|
# # # # # # # #
|
|
# #
|
|
# HAR #
|
|
# #
|
|
# # # # # # # #
|
|
|
|
def create_har_id(date, item_id):
|
|
item_id = item_id.split('/')[-1]
|
|
return os.path.join(date, f'{item_id}.json.gz')
|
|
|
|
def save_har(har_id, har_content):
|
|
# create dir
|
|
har_dir = os.path.dirname(os.path.join(get_har_dir(), har_id))
|
|
if not os.path.exists(har_dir):
|
|
os.makedirs(har_dir)
|
|
# save HAR
|
|
filename = os.path.join(get_har_dir(), har_id)
|
|
with gzip.open(filename, 'wb') as f:
|
|
f.write(json.dumps(har_content).encode())
|
|
|
|
def get_all_har_ids():
|
|
har_ids = []
|
|
today_root_dir = os.path.join(HAR_DIR, Date.get_today_date_str(separator=True))
|
|
dirs_year = set()
|
|
for ydir in next(os.walk(HAR_DIR))[1]:
|
|
if len(ydir) == 4:
|
|
try:
|
|
int(ydir)
|
|
dirs_year.add(ydir)
|
|
except (TypeError, ValueError):
|
|
pass
|
|
|
|
if os.path.exists(today_root_dir):
|
|
for file in [f for f in os.listdir(today_root_dir) if os.path.isfile(os.path.join(today_root_dir, f))]:
|
|
har_id = os.path.relpath(os.path.join(today_root_dir, file), HAR_DIR)
|
|
har_ids.append(har_id)
|
|
|
|
for ydir in sorted(dirs_year, reverse=False):
|
|
search_dear = os.path.join(HAR_DIR, ydir)
|
|
for root, dirs, files in os.walk(search_dear):
|
|
for file in files:
|
|
if root != today_root_dir:
|
|
har_id = os.path.relpath(os.path.join(root, file), HAR_DIR)
|
|
har_ids.append(har_id)
|
|
return har_ids
|
|
|
|
def get_month_har_ids(year, month):
|
|
har_ids = []
|
|
month_path = os.path.join(HAR_DIR, year, month)
|
|
for root, dirs, files in os.walk(month_path):
|
|
for file in files:
|
|
har_id = os.path.relpath(os.path.join(root, file), HAR_DIR)
|
|
har_ids.append(har_id)
|
|
return har_ids
|
|
|
|
|
|
def get_har_content(har_id):
|
|
har_path = os.path.join(HAR_DIR, har_id)
|
|
try:
|
|
with gzip.open(har_path) as f:
|
|
try:
|
|
return json.loads(f.read())
|
|
except json.decoder.JSONDecodeError:
|
|
return {}
|
|
except Exception as e:
|
|
print(e) # TODO LOGS
|
|
return {}
|
|
|
|
def extract_cookies_names_from_har(har):
|
|
cookies = set()
|
|
for entrie in har.get('log', {}).get('entries', []):
|
|
for cookie in entrie.get('request', {}).get('cookies', []):
|
|
name = cookie.get('name')
|
|
if name:
|
|
cookies.add(name)
|
|
for cookie in entrie.get('response', {}).get('cookies', []):
|
|
name = cookie.get('name')
|
|
if name:
|
|
cookies.add(name)
|
|
return cookies
|
|
|
|
def _reprocess_all_hars_cookie_name():
|
|
from lib.objects import CookiesNames
|
|
for har_id in get_all_har_ids():
|
|
domain = har_id.split('/')[-1]
|
|
domain = domain[:-44]
|
|
date = har_id.split('/')
|
|
date = f'{date[-4]}{date[-3]}{date[-2]}'
|
|
for cookie_name in extract_cookies_names_from_har(get_har_content(har_id)):
|
|
print(domain, date, cookie_name)
|
|
cookie = CookiesNames.create(cookie_name)
|
|
cookie.add(date, Domain(domain))
|
|
|
|
def extract_etag_from_har(har): # TODO check response url
|
|
etags = set()
|
|
for entrie in har.get('log', {}).get('entries', []):
|
|
for header in entrie.get('response', {}).get('headers', []):
|
|
if header.get('name') == 'etag':
|
|
# print(header)
|
|
etag = header.get('value')
|
|
if etag:
|
|
etags.add(etag)
|
|
return etags
|
|
|
|
def _reprocess_all_hars_etag():
|
|
from lib.objects import Etags
|
|
for har_id in get_all_har_ids():
|
|
domain = har_id.split('/')[-1]
|
|
domain = domain[:-44]
|
|
date = har_id.split('/')
|
|
date = f'{date[-4]}{date[-3]}{date[-2]}'
|
|
for etag_content in extract_etag_from_har(get_har_content(har_id)):
|
|
print(domain, date, etag_content)
|
|
etag = Etags.create(etag_content)
|
|
etag.add(date, Domain(domain))
|
|
|
|
def extract_hhhash_by_id(har_id, domain, date):
|
|
return extract_hhhash(get_har_content(har_id), domain, date)
|
|
|
|
def extract_hhhash(har, domain, date):
|
|
hhhashs = set()
|
|
urls = set()
|
|
for entrie in har.get('log', {}).get('entries', []):
|
|
url = entrie.get('request').get('url')
|
|
if url not in urls:
|
|
# filter redirect
|
|
if entrie.get('response').get('status') == 200: # != 301:
|
|
# print(url, entrie.get('response').get('status'))
|
|
|
|
f = get_faup()
|
|
f.decode(url)
|
|
domain_url = f.get().get('domain')
|
|
if domain_url == domain:
|
|
|
|
headers = entrie.get('response').get('headers')
|
|
|
|
hhhash_header = HHHashs.build_hhhash_headers(headers)
|
|
hhhash = HHHashs.hhhash_headers(hhhash_header)
|
|
|
|
if hhhash not in hhhashs:
|
|
print('', url, hhhash)
|
|
|
|
# -----
|
|
obj = HHHashs.create(hhhash_header, hhhash)
|
|
obj.add(date, Domain(domain))
|
|
|
|
hhhashs.add(hhhash)
|
|
urls.add(url)
|
|
print()
|
|
print()
|
|
print('HHHASH:')
|
|
for hhhash in hhhashs:
|
|
print(hhhash)
|
|
return hhhashs
|
|
|
|
def _reprocess_all_hars_hhhashs():
|
|
for har_id in get_all_har_ids():
|
|
print()
|
|
print(har_id)
|
|
domain = har_id.split('/')[-1]
|
|
domain = domain[:-44]
|
|
date = har_id.split('/')
|
|
date = f'{date[-4]}{date[-3]}{date[-2]}'
|
|
extract_hhhash_by_id(har_id, domain, date)
|
|
|
|
|
|
|
|
def _gzip_har(har_id):
|
|
har_path = os.path.join(HAR_DIR, har_id)
|
|
new_id = f'{har_path}.gz'
|
|
if not har_id.endswith('.gz'):
|
|
if not os.path.exists(new_id):
|
|
with open(har_path, 'rb') as f:
|
|
content = f.read()
|
|
if content:
|
|
with gzip.open(new_id, 'wb') as f:
|
|
r = f.write(content)
|
|
print(r)
|
|
if os.path.exists(new_id) and os.path.exists(har_path):
|
|
os.remove(har_path)
|
|
print('delete:', har_path)
|
|
|
|
def _gzip_all_hars():
|
|
for har_id in get_all_har_ids():
|
|
_gzip_har(har_id)
|
|
|
|
# # # - - # # #
|
|
|
|
################################################################################
|
|
|
|
# # TODO:
|
|
# # TODO: REVIEW ME THEN REMOVE ME
|
|
def create_cookie_crawler(cookie_dict, domain, crawler_type='web'):
|
|
# check cookie domain filed
|
|
if not 'domain' in cookie_dict:
|
|
cookie_dict['domain'] = f'.{domain}'
|
|
|
|
# tor browser: disable secure cookie
|
|
if crawler_type == 'onion':
|
|
cookie_dict['secure'] = False
|
|
|
|
# force cookie domain
|
|
# url = urlparse(browser_cookie['Host raw'])
|
|
# domain = url.netloc.split(':', 1)[0]
|
|
# cookie_dict['domain'] = '.{}'.format(domain)
|
|
|
|
# change expire date
|
|
cookie_dict['expires'] = (datetime.now() + timedelta(days=10)).strftime('%Y-%m-%dT%H:%M:%S') + 'Z'
|
|
return cookie_dict
|
|
|
|
################################################################################
|
|
################################################################################
|
|
################################################################################
|
|
|
|
def get_cookiejars():
|
|
return r_crawler.smembers('cookiejars:all')
|
|
|
|
def get_cookiejars_global():
|
|
cookiejars = r_crawler.smembers('cookiejars:global')
|
|
if not cookiejars:
|
|
cookiejars = []
|
|
return cookiejars
|
|
|
|
def get_cookiejars_user(user_id):
|
|
cookiejars = r_crawler.smembers(f'cookiejars:user:{user_id}')
|
|
if not cookiejars:
|
|
cookiejars = []
|
|
return cookiejars
|
|
|
|
class Cookiejar:
|
|
|
|
def __init__(self, cookiejar_uuid):
|
|
self.uuid = cookiejar_uuid
|
|
|
|
def exists(self):
|
|
return r_crawler.exists(f'cookiejar:meta:{self.uuid}') # or cookiejar:uuid
|
|
|
|
def get_date(self):
|
|
return r_crawler.hget(f'cookiejar:meta:{self.uuid}', 'date')
|
|
|
|
def _set_date(self, date):
|
|
r_crawler.hset(f'cookiejar:meta:{self.uuid}', 'date', date)
|
|
|
|
def get_description(self):
|
|
return r_crawler.hget(f'cookiejar:meta:{self.uuid}', 'description')
|
|
|
|
def set_description(self, description):
|
|
r_crawler.hset(f'cookiejar:meta:{self.uuid}', 'description', description)
|
|
|
|
def get_user(self):
|
|
return r_crawler.hget(f'cookiejar:meta:{self.uuid}', 'user')
|
|
|
|
def _set_user(self, user_id):
|
|
return r_crawler.hset(f'cookiejar:meta:{self.uuid}', 'user', user_id)
|
|
|
|
def get_level(self):
|
|
level = r_crawler.hget(f'cookiejar:meta:{self.uuid}', 'level')
|
|
if level:
|
|
level = 1
|
|
else:
|
|
level = 0
|
|
return level
|
|
|
|
def _set_level(self, level):
|
|
if level:
|
|
level = 1
|
|
else:
|
|
level = 0
|
|
r_crawler.hset(f'cookiejar:meta:{self.uuid}', 'level', level)
|
|
|
|
def is_cookie_in_jar(self, cookie_uuid):
|
|
# kvrocks sismember TEMP fix
|
|
try:
|
|
return r_crawler.sismember(f'cookiejar:cookies:{self.uuid}', cookie_uuid)
|
|
except:
|
|
return False
|
|
|
|
def get_cookies_uuid(self):
|
|
return r_crawler.smembers(f'cookiejar:cookies:{self.uuid}')
|
|
|
|
def get_cookies(self, r_json=False):
|
|
l_cookies = []
|
|
for cookie_uuid in self.get_cookies_uuid():
|
|
cookies = Cookie(cookie_uuid)
|
|
l_cookies.append(cookies.get_meta(r_json=r_json))
|
|
return l_cookies
|
|
|
|
def get_nb_cookies(self):
|
|
return r_crawler.scard(f'cookiejar:cookies:{self.uuid}')
|
|
|
|
def get_meta(self, level=False, nb_cookies=False, cookies=False, r_json=False):
|
|
meta = {'uuid': self.uuid,
|
|
'date': self.get_date(),
|
|
'description': self.get_description(),
|
|
'user': self.get_user()}
|
|
if level:
|
|
meta['level'] = self.get_level()
|
|
if nb_cookies:
|
|
meta['nb_cookies'] = self.get_nb_cookies()
|
|
if cookies:
|
|
meta['cookies'] = self.get_cookies(r_json=r_json)
|
|
return meta
|
|
|
|
def add_cookie(self, name, value, cookie_uuid=None, domain=None, httponly=None, path=None, secure=None, text=None):
|
|
if cookie_uuid:
|
|
cookie = Cookie(cookie_uuid)
|
|
if cookie.exists():
|
|
cookie_uuid = generate_uuid()
|
|
else:
|
|
cookie_uuid = generate_uuid()
|
|
r_crawler.sadd(f'cookiejar:cookies:{self.uuid}', cookie_uuid)
|
|
|
|
cookie = Cookie(cookie_uuid)
|
|
cookie.set_cookiejar(self.uuid)
|
|
|
|
cookie.set_field('name', name)
|
|
cookie.set_field('value', value)
|
|
if domain:
|
|
cookie.set_field('domain', domain)
|
|
if httponly:
|
|
cookie.set_field('httpOnly', str(httponly))
|
|
if path:
|
|
cookie.set_field('path', path)
|
|
if secure:
|
|
cookie.set_field('secure', str(secure))
|
|
if text:
|
|
cookie.set_field('path', text)
|
|
return cookie_uuid
|
|
|
|
def delete_cookie(self, cookie_uuid):
|
|
if self.is_cookie_in_jar(cookie_uuid):
|
|
cookie = Cookie(cookie_uuid)
|
|
cookie.delete()
|
|
|
|
def create(self, user_id, description=None, level=1):
|
|
if self.exists():
|
|
raise Exception('Cookiejar already exists')
|
|
|
|
r_crawler.sadd('cookiejars:all', self.uuid)
|
|
if level == 0:
|
|
r_crawler.sadd(f'cookiejars:user:{user_id}', self.uuid)
|
|
else:
|
|
r_crawler.sadd('cookiejars:global', self.uuid)
|
|
|
|
self._set_user(user_id)
|
|
self._set_date(datetime.now().strftime("%Y%m%d"))
|
|
self._set_level(level)
|
|
if description:
|
|
self.set_description(description)
|
|
|
|
def delete(self):
|
|
for cookie_uuid in self.get_cookies_uuid():
|
|
self.delete_cookie(cookie_uuid)
|
|
r_crawler.srem(f'cookiejars:user:{self.get_user()}', self.uuid)
|
|
r_crawler.srem('cookiejars:global', self.uuid)
|
|
r_crawler.srem('cookiejars:all', self.uuid)
|
|
r_crawler.delete(f'cookiejar:meta:{self.uuid}')
|
|
|
|
|
|
def create_cookiejar(user_id, description=None, level=1, cookiejar_uuid=None):
|
|
if cookiejar_uuid:
|
|
cookiejar = Cookiejar(cookiejar_uuid)
|
|
if cookiejar.exists():
|
|
cookiejar_uuid = generate_uuid()
|
|
else:
|
|
cookiejar_uuid = generate_uuid()
|
|
cookiejar = Cookiejar(cookiejar_uuid)
|
|
cookiejar.create(user_id, description=description, level=level)
|
|
return cookiejar_uuid
|
|
|
|
def get_cookiejars_meta_by_iterator(iter_cookiejar_uuid):
|
|
cookiejars_meta = []
|
|
for cookiejar_uuid in iter_cookiejar_uuid:
|
|
cookiejar = Cookiejar(cookiejar_uuid)
|
|
cookiejars_meta.append(cookiejar.get_meta(nb_cookies=True))
|
|
return cookiejars_meta
|
|
|
|
def get_cookiejars_by_user(user_id):
|
|
cookiejars_global = get_cookiejars_global()
|
|
cookiejars_user = get_cookiejars_user(user_id)
|
|
return [*cookiejars_user, *cookiejars_global]
|
|
|
|
## API ##
|
|
|
|
def api_get_cookiejars_selector(user_id):
|
|
cookiejars = []
|
|
for cookiejar_uuid in get_cookiejars_by_user(user_id):
|
|
cookiejar = Cookiejar(cookiejar_uuid)
|
|
description = cookiejar.get_description()
|
|
if not description:
|
|
description = ''
|
|
cookiejars.append(f'{description} : {cookiejar.uuid}')
|
|
return sorted(cookiejars)
|
|
|
|
def api_verify_cookiejar_acl(cookiejar_uuid, user_id):
|
|
cookiejar = Cookiejar(cookiejar_uuid)
|
|
if not cookiejar.exists():
|
|
return {'error': 'unknown cookiejar uuid', 'cookiejar_uuid': cookiejar_uuid}, 404
|
|
if cookiejar.get_level() == 0: # TODO: check if user is admin
|
|
if cookiejar.get_user() != user_id:
|
|
return {'error': 'The access to this cookiejar is restricted'}, 403
|
|
|
|
def api_edit_cookiejar_description(user_id, cookiejar_uuid, description):
|
|
resp = api_verify_cookiejar_acl(cookiejar_uuid, user_id)
|
|
if resp:
|
|
return resp
|
|
cookiejar = Cookiejar(cookiejar_uuid)
|
|
cookiejar.set_description(description)
|
|
return {'cookiejar_uuid': cookiejar_uuid}, 200
|
|
|
|
def api_delete_cookiejar(user_id, cookiejar_uuid):
|
|
resp = api_verify_cookiejar_acl(cookiejar_uuid, user_id)
|
|
if resp:
|
|
return resp
|
|
cookiejar = Cookiejar(cookiejar_uuid)
|
|
cookiejar.delete()
|
|
return {'cookiejar_uuid': cookiejar_uuid}, 200
|
|
|
|
def api_get_cookiejar(cookiejar_uuid, user_id):
|
|
resp = api_verify_cookiejar_acl(cookiejar_uuid, user_id)
|
|
if resp:
|
|
return resp
|
|
cookiejar = Cookiejar(cookiejar_uuid)
|
|
meta = cookiejar.get_meta(level=True, cookies=True, r_json=True)
|
|
return meta, 200
|
|
|
|
# # # # # # # #
|
|
# #
|
|
# COOKIES #
|
|
# #
|
|
# # # # # # # #
|
|
|
|
# # # #
|
|
# Cookies Fields:
|
|
# - name
|
|
# - value
|
|
# - path (optional)
|
|
# - domain (optional)
|
|
# - secure (optional)
|
|
# - httpOnly (optional)
|
|
# - text (optional)
|
|
# # # #
|
|
|
|
# TODO MISP Import
|
|
|
|
class Cookie:
|
|
|
|
def __init__(self, cookie_uuid):
|
|
self.uuid = cookie_uuid
|
|
|
|
def exists(self):
|
|
return r_crawler.exists(f'cookie:meta:{self.uuid}')
|
|
|
|
def get_cookiejar(self):
|
|
return r_crawler.hget(f'cookie:meta:{self.uuid}', 'cookiejar')
|
|
|
|
def set_cookiejar(self, cookiejar_uuid):
|
|
r_crawler.hset(f'cookie:meta:{self.uuid}', 'cookiejar', cookiejar_uuid)
|
|
|
|
def get_name(self):
|
|
return r_crawler.hget(f'cookie:meta:{self.uuid}', 'name')
|
|
|
|
def get_value(self):
|
|
return r_crawler.hget(f'cookie:meta:{self.uuid}', 'value')
|
|
|
|
def _get_field(self, field):
|
|
return r_crawler.hget(f'cookie:meta:{self.uuid}', field)
|
|
|
|
def set_field(self, field, value):
|
|
return r_crawler.hset(f'cookie:meta:{self.uuid}', field, value)
|
|
|
|
def remove_field(self, field):
|
|
return r_crawler.hdel(f'cookie:meta:{self.uuid}', field)
|
|
|
|
def get_fields(self):
|
|
fields = set(r_crawler.hkeys(f'cookie:meta:{self.uuid}'))
|
|
if 'cookiejar' in fields:
|
|
fields.remove('cookiejar')
|
|
return fields
|
|
|
|
# def get_domain(self):
|
|
# return r_crawler.hget(f'cookie:meta:{self.uuid}', 'domain')
|
|
#
|
|
# def get_path(self):
|
|
# return r_crawler.hget(f'cookie:meta:{self.uuid}', 'path')
|
|
#
|
|
# def get_httpOnly(self):
|
|
# return r_crawler.hget(f'cookie:meta:{self.uuid}', 'httpOnly')
|
|
#
|
|
# def get_secure(self):
|
|
# return r_crawler.hget(f'cookie:meta:{self.uuid}', 'secure')
|
|
|
|
# TODO expire ????
|
|
def get_meta(self, r_json=False):
|
|
meta = {}
|
|
# ['domain', 'path', 'httpOnly', 'secure'] + name + value
|
|
for field in self.get_fields():
|
|
value = self._get_field(field)
|
|
if value:
|
|
meta[field] = value
|
|
if r_json:
|
|
data = json.dumps(meta, indent=4, sort_keys=True)
|
|
meta = {'data': data, 'uuid': self.uuid}
|
|
return meta
|
|
|
|
def edit(self, cookie_dict):
|
|
# remove old keys
|
|
for field in self.get_fields():
|
|
if field not in cookie_dict:
|
|
self.remove_field(field)
|
|
# add new keys
|
|
for field in cookie_dict:
|
|
value = cookie_dict[field]
|
|
if value:
|
|
if field == 'secure' or field == 'httpOnly':
|
|
value = str(value)
|
|
self.set_field(field, value)
|
|
|
|
def delete(self):
|
|
cookiejar_uuid = self.get_cookiejar()
|
|
r_crawler.delete(f'cookie:meta:{self.uuid}')
|
|
r_crawler.srem(f'cookiejar:cookies:{cookiejar_uuid}', self.uuid)
|
|
|
|
## API ##
|
|
|
|
def api_get_cookie(user_id, cookie_uuid):
|
|
cookie = Cookie(cookie_uuid)
|
|
if not cookie.exists():
|
|
return {'error': 'unknown cookie uuid', 'cookie_uuid': cookie_uuid}, 404
|
|
resp = api_verify_cookiejar_acl(cookie.get_cookiejar(), user_id)
|
|
if resp:
|
|
return resp
|
|
return cookie.get_meta()
|
|
|
|
def api_edit_cookie(user_id, cookie_uuid, cookie_dict):
|
|
cookie = Cookie(cookie_uuid)
|
|
if not cookie.exists():
|
|
return {'error': 'unknown cookie uuid', 'cookie_uuid': cookie_uuid}, 404
|
|
resp = api_verify_cookiejar_acl(cookie.get_cookiejar(), user_id)
|
|
if resp:
|
|
return resp
|
|
if 'name' not in cookie_dict or 'value' not in cookie_dict or not cookie_dict['name'] or not cookie_dict['value']:
|
|
return {'error': 'cookie name or value not provided'}, 400
|
|
cookie.edit(cookie_dict)
|
|
return cookie.get_meta(), 200
|
|
|
|
def api_create_cookie(user_id, cookiejar_uuid, cookie_dict):
|
|
resp = api_verify_cookiejar_acl(cookiejar_uuid, user_id)
|
|
if resp:
|
|
return resp
|
|
if 'name' not in cookie_dict or 'value' not in cookie_dict or not cookie_dict['name'] or not cookie_dict['value']:
|
|
return {'error': 'cookie name or value not provided'}, 400
|
|
cookiejar = Cookiejar(cookiejar_uuid)
|
|
name = cookie_dict.get('name')
|
|
value = cookie_dict.get('value')
|
|
domain = cookie_dict.get('domain')
|
|
path = cookie_dict.get('path')
|
|
text = cookie_dict.get('text')
|
|
httponly = bool(cookie_dict.get('httponly'))
|
|
secure = bool(cookie_dict.get('secure'))
|
|
cookiejar.add_cookie(name, value, domain=domain, httponly=httponly, path=path, secure=secure, text=text)
|
|
return resp, 200
|
|
|
|
def api_delete_cookie(user_id, cookie_uuid):
|
|
cookie = Cookie(cookie_uuid)
|
|
if not cookie.exists():
|
|
return {'error': 'unknown cookie uuid', 'cookie_uuid': cookie_uuid}, 404
|
|
cookiejar_uuid = cookie.get_cookiejar()
|
|
resp = api_verify_cookiejar_acl(cookiejar_uuid, user_id)
|
|
if resp:
|
|
return resp
|
|
cookiejar = Cookiejar(cookiejar_uuid)
|
|
if not cookiejar.is_cookie_in_jar(cookie_uuid):
|
|
return {'error': 'Cookie isn\'t in the jar', 'cookiejar_uuid': cookiejar_uuid}, 404
|
|
cookiejar.delete_cookie(cookie_uuid)
|
|
return {'cookiejar_uuid': cookiejar_uuid, 'cookie_uuid': cookie_uuid}, 200
|
|
|
|
# def get_cookie_all_keys_name():
|
|
# return ['name', 'value', 'domain', 'path', 'httpOnly', 'secure']
|
|
|
|
## - - ##
|
|
## Cookies import ## # TODO: add browser type ?
|
|
def import_cookies_from_json(json_cookies, cookiejar_uuid):
|
|
cookiejar = Cookiejar(cookiejar_uuid)
|
|
for cookie in json_cookies:
|
|
try:
|
|
cookie_dict = unpack_imported_json_cookie(cookie)
|
|
name = cookie_dict.get('name')
|
|
value = cookie_dict.get('value')
|
|
domain = cookie_dict.get('domain')
|
|
httponly = cookie_dict.get('httponly')
|
|
path = cookie_dict.get('path')
|
|
secure = cookie_dict.get('secure')
|
|
text = cookie_dict.get('text')
|
|
cookiejar.add_cookie(name, value, domain=domain, httponly=httponly, path=path, secure=secure, text=text)
|
|
except KeyError:
|
|
return {'error': 'Invalid cookie key, please submit a valid JSON', 'cookiejar_uuid': cookiejar_uuid}, 400
|
|
|
|
# # TODO: add text field
|
|
def unpack_imported_json_cookie(json_cookie):
|
|
cookie_dict = {'name': json_cookie['Name raw'], 'value': json_cookie['Content raw']}
|
|
if 'Path raw' in json_cookie:
|
|
cookie_dict['path'] = json_cookie['Path raw']
|
|
if 'HTTP only raw' in json_cookie:
|
|
cookie_dict['httponly'] = json_cookie['HTTP only raw'] == 'true'
|
|
if 'Send for' in json_cookie:
|
|
cookie_dict['secure'] = json_cookie['Send for'] == 'Encrypted connections only'
|
|
if 'Host raw' in json_cookie:
|
|
url = urlparse(json_cookie['Host raw'])
|
|
cookie_dict['domain'] = url.netloc.split(':', 1)[0]
|
|
return cookie_dict
|
|
|
|
## - - ##
|
|
#### COOKIEJAR API ####
|
|
def api_import_cookies_from_json(user_id, cookiejar_uuid, json_cookies_str): # # TODO: add catch
|
|
resp = api_verify_cookiejar_acl(cookiejar_uuid, user_id)
|
|
if resp:
|
|
return resp
|
|
json_cookies = json.loads(json_cookies_str)
|
|
resp = import_cookies_from_json(json_cookies, cookiejar_uuid)
|
|
if resp:
|
|
return resp, 400
|
|
#### ####
|
|
|
|
|
|
# # # # # # # #
|
|
# #
|
|
# CRAWLER # ###################################################################################
|
|
# #
|
|
# # # # # # # #
|
|
|
|
def get_default_user_agent():
|
|
return 'Mozilla/5.0 (Windows NT 10.0; rv:102.0) Gecko/20100101 Firefox/102.0'
|
|
|
|
def get_last_crawled_domains(domain_type):
|
|
return r_crawler.lrange(f'last_{domain_type}', 0, -1)
|
|
|
|
def update_last_crawled_domain(domain_type, domain, epoch):
|
|
# update list, last crawled domains
|
|
r_crawler.lpush(f'last_{domain_type}', f'{domain}:{epoch}')
|
|
r_crawler.ltrim(f'last_{domain_type}', 0, 15)
|
|
|
|
def create_item_metadata(item_id, url, item_father):
|
|
item = Item(item_id)
|
|
item.set_crawled(url, item_father)
|
|
|
|
def get_gzipped_b64_item(item_id, content):
|
|
try:
|
|
gzipencoded = gzip.compress(content.encode())
|
|
gzip64encoded = base64.standard_b64encode(gzipencoded).decode()
|
|
return gzip64encoded
|
|
except:
|
|
print(f'file error: {item_id}')
|
|
return False
|
|
|
|
def get_crawlers_stats_by_day(date, domain_type):
|
|
return {
|
|
'date': date[0:4] + '-' + date[4:6] + '-' + date[6:8],
|
|
'up': r_crawler.scard(f'{domain_type}_up:{date}'),
|
|
'down': r_crawler.scard(f'{domain_type}_down:{date}'),
|
|
}
|
|
|
|
def get_crawlers_stats_by_month(domain_type, date=None):
|
|
stats = []
|
|
for date in Date.get_month_dates(date=date):
|
|
stats.append(get_crawlers_stats_by_day(date, domain_type))
|
|
return stats
|
|
|
|
def get_crawlers_stats_up_down_by_month(domain_type, date=None):
|
|
stats = {'down': 0, 'up': 0}
|
|
for date in Date.get_month_dates(date=date):
|
|
day = get_crawlers_stats_by_day(date, domain_type)
|
|
stats['down'] += day.get('down', 0)
|
|
stats['up'] += day.get('up', 0)
|
|
return stats
|
|
|
|
def get_crawlers_stats(domain_type=None):
|
|
stats = {}
|
|
date = datetime.now().strftime("%Y%m%d")
|
|
if domain_type:
|
|
domain_types = [domain_type]
|
|
else:
|
|
domain_types = get_crawler_all_types()
|
|
for domain_type in domain_types:
|
|
queue = r_crawler.scard(f'crawler:queue:type:{domain_type}')
|
|
up = r_crawler.scard(f'{domain_type}_up:{date}')
|
|
down = r_crawler.scard(f'{domain_type}_down:{date}')
|
|
crawled = up + down
|
|
stats[domain_type] = {'queue': queue, 'up': up, 'down': down, 'crawled': crawled}
|
|
return stats
|
|
|
|
#### Blocklist ####
|
|
|
|
def get_blacklist():
|
|
return r_crawler.smembers('blacklist:domain')
|
|
|
|
def is_blacklisted_domain(domain):
|
|
# kvrocks sismember TEMP fix
|
|
try:
|
|
return r_crawler.sismember('blacklist:domain', domain)
|
|
except:
|
|
return False
|
|
|
|
def blacklist_domain(domain):
|
|
return r_crawler.sadd('blacklist:domain', domain)
|
|
|
|
def unblacklist_domain(domain):
|
|
return r_crawler.srem('blacklist:domain', domain)
|
|
|
|
def load_blacklist():
|
|
try:
|
|
with open(os.path.join(os.environ['AIL_BIN'], 'crawlers/blacklist.txt'), 'r') as f:
|
|
r_crawler.delete('blacklist:domain')
|
|
lines = f.read().splitlines()
|
|
for line in lines:
|
|
blacklist_domain(line)
|
|
# TODO LOG
|
|
except Exception as e:
|
|
print(e)
|
|
|
|
def api_blacklist_domain(data):
|
|
domain = str(data.get('domain', '')).lower()
|
|
if not is_valid_domain(domain):
|
|
return {'error': 'invalid domain'}, 400
|
|
if is_blacklisted_domain(domain):
|
|
return {'error': 'domain already blacklisted'}, 400
|
|
return blacklist_domain(domain), 200
|
|
|
|
def api_unblacklist_domain(data):
|
|
domain = str(data.get('domain', '')).lower()
|
|
if not is_valid_domain(domain):
|
|
return {'error': 'invalid domain'}, 400
|
|
if not is_blacklisted_domain(domain):
|
|
return {'error': 'domain not blacklisted'}, 404
|
|
return unblacklist_domain(domain), 200
|
|
|
|
#### CRAWLER Scheduler ####
|
|
|
|
@unique
|
|
class ScheduleStatus(IntEnum):
|
|
"""The status of the capture"""
|
|
UNKNOWN = -1
|
|
QUEUED = 0
|
|
SCHEDULED = 1
|
|
ONGOING = 2
|
|
|
|
def get_schedulers_uuid():
|
|
return r_crawler.smembers('scheduler:schedules')
|
|
|
|
def get_schedulers_metas():
|
|
schedulers = []
|
|
for schedule_uuid in get_schedulers_uuid():
|
|
schedule = CrawlerSchedule(schedule_uuid)
|
|
schedulers.append(schedule.get_meta_status())
|
|
return schedulers
|
|
|
|
class CrawlerScheduler:
|
|
|
|
def __init__(self):
|
|
self.min_frequency = 60 # TODO ADD IN CONFIG
|
|
|
|
def update_queue(self):
|
|
for schedule_uuid in get_schedulers_uuid():
|
|
schedule = CrawlerSchedule(schedule_uuid)
|
|
# check if already in scheduler queue
|
|
if schedule.is_scheduled():
|
|
continue
|
|
if schedule.is_tasked():
|
|
continue
|
|
|
|
# EXPIRE ????
|
|
|
|
time_next_run = 0.0
|
|
frequency = schedule.get_frequency() # optional or later -> cron
|
|
if frequency == 'hourly':
|
|
time_next_run = (datetime.now() + timedelta(hours=1)).timestamp()
|
|
elif frequency == 'daily':
|
|
time_next_run = (datetime.now() + timedelta(days=1)).timestamp()
|
|
elif frequency == 'weekly':
|
|
time_next_run = (datetime.now() + timedelta(weeks=1)).timestamp()
|
|
elif frequency == 'monthly':
|
|
time_next_run = (datetime.now() + relativedelta(months=1)).timestamp()
|
|
else:
|
|
months, weeks, days, hours, minutes = frequency.split(':')
|
|
if not months:
|
|
months = 0
|
|
if not weeks:
|
|
weeks = 0
|
|
if not days:
|
|
days = 0
|
|
if not hours:
|
|
hours = 0
|
|
if not minutes:
|
|
minutes = 0
|
|
current_time = datetime.now().timestamp()
|
|
time_next_run = (datetime.now() + relativedelta(months=int(months), weeks=int(weeks),
|
|
days=int(days), hours=int(hours),
|
|
minutes=int(minutes))).timestamp()
|
|
# Make sure the next capture is not scheduled for in a too short interval
|
|
interval_next_capture = time_next_run - current_time
|
|
if interval_next_capture < self.min_frequency:
|
|
# self.logger.warning(f'The next capture is scheduled too soon: {interval_next_capture}s. Minimal interval: {self.min_frequency}s.')
|
|
print(f'The next capture is scheduled too soon: {interval_next_capture}s. Minimal interval: {self.min_frequency}s.')
|
|
time_next_run = (datetime.now() + timedelta(seconds=self.min_frequency)).timestamp()
|
|
|
|
schedule.set_next_run(time_next_run)
|
|
print('scheduled:', schedule_uuid)
|
|
|
|
def process_queue(self):
|
|
now = datetime.now().timestamp()
|
|
for raw_schedule in r_crawler.zrangebyscore('scheduler:queue', '-inf', int(now), withscores=True):
|
|
schedule_uuid, next_run = raw_schedule
|
|
schedule = CrawlerSchedule(schedule_uuid)
|
|
if not schedule.exists():
|
|
return None
|
|
meta = schedule.get_meta()
|
|
task_uuid = create_task(meta['url'], depth=meta['depth'], har=meta['har'], screenshot=meta['screenshot'],
|
|
header=meta['header'],
|
|
cookiejar=meta['cookiejar'], proxy=meta['proxy'],
|
|
tags=meta['tags'],
|
|
user_agent=meta['user_agent'], parent='scheduler', priority=40)
|
|
if task_uuid:
|
|
schedule.set_task(task_uuid)
|
|
r_crawler.zrem('scheduler:queue', schedule_uuid)
|
|
|
|
|
|
# TODO Expire -> stuck in crawler queue or reached delta
|
|
class CrawlerSchedule:
|
|
def __init__(self, schedule_uuid):
|
|
self.uuid = schedule_uuid
|
|
|
|
def exists(self):
|
|
return r_crawler.exists(f'schedule:{self.uuid}')
|
|
|
|
def get_frequency(self):
|
|
return r_crawler.hget(f'schedule:{self.uuid}', 'frequency')
|
|
|
|
def get_user(self):
|
|
return r_crawler.hget(f'schedule:{self.uuid}', 'user')
|
|
|
|
def get_date(self):
|
|
return r_crawler.hget(f'schedule:{self.uuid}', 'date')
|
|
|
|
def get_captures(self): # only scheduled capture ????? exclude manual/discovery
|
|
pass
|
|
|
|
def get_status(self):
|
|
if self.is_scheduled():
|
|
return ScheduleStatus.SCHEDULED
|
|
if self.is_tasked():
|
|
if self.is_ongoing():
|
|
return ScheduleStatus.ONGOING
|
|
else:
|
|
return ScheduleStatus.QUEUED
|
|
return ScheduleStatus.UNKNOWN
|
|
|
|
def get_task_uuid(self):
|
|
return r_crawler.hget(f'schedule:{self.uuid}', 'task')
|
|
|
|
def is_tasked(self):
|
|
task_uuid = self.get_task_uuid()
|
|
if task_uuid:
|
|
task = CrawlerTask(task_uuid)
|
|
tasked = task.exists()
|
|
if not tasked:
|
|
r_crawler.hdel(f'schedule:{self.uuid}', 'task')
|
|
return tasked
|
|
return False
|
|
|
|
def get_task(self):
|
|
task_uuid = self.get_task_uuid()
|
|
if task_uuid:
|
|
return CrawlerTask(task_uuid)
|
|
|
|
def set_task(self, task_uuid):
|
|
return r_crawler.hset(f'schedule:{self.uuid}', 'task', task_uuid)
|
|
|
|
def is_ongoing(self):
|
|
task = self.get_task()
|
|
if task:
|
|
return task.is_ongoing()
|
|
return False
|
|
|
|
def get_next_run(self, r_str=False):
|
|
next_run = r_crawler.zscore('scheduler:queue', self.uuid)
|
|
if next_run and r_str:
|
|
next_run = time.strftime('%Y-%m-%d - %H:%M:%S', time.localtime(int(next_run)))
|
|
return next_run
|
|
|
|
def set_next_run(self, time_next_run):
|
|
r_crawler.zadd('scheduler:queue', mapping={self.uuid: time_next_run})
|
|
|
|
def is_scheduled(self):
|
|
return bool(r_crawler.zscore('scheduler:queue', self.uuid))
|
|
|
|
def get_url(self):
|
|
return r_crawler.hget(f'schedule:{self.uuid}', 'url')
|
|
|
|
def get_depth(self):
|
|
return r_crawler.hget(f'schedule:{self.uuid}', 'depth')
|
|
|
|
def get_har(self):
|
|
return r_crawler.hget(f'schedule:{self.uuid}', 'har') == 'True'
|
|
|
|
def get_screenshot(self):
|
|
return r_crawler.hget(f'schedule:{self.uuid}', 'screenshot') == 'True'
|
|
|
|
def get_header(self):
|
|
r_crawler.hget(f'schedule:{self.uuid}', 'header')
|
|
|
|
def get_cookiejar(self):
|
|
return r_crawler.hget(f'schedule:{self.uuid}', 'cookiejar')
|
|
|
|
def get_proxy(self):
|
|
return r_crawler.hget(f'schedule:{self.uuid}', 'proxy')
|
|
|
|
def get_user_agent(self):
|
|
return r_crawler.hget(f'schedule:{self.uuid}', 'user_agent')
|
|
|
|
def _set_field(self, field, value):
|
|
return r_crawler.hset(f'schedule:{self.uuid}', field, value)
|
|
|
|
def get_tags(self):
|
|
return r_crawler.smembers(f'schedule:tags:{self.uuid}')
|
|
|
|
def set_tags(self, tags=[]):
|
|
for tag in tags:
|
|
r_crawler.sadd(f'schedule:tags:{self.uuid}', tag)
|
|
# Tag.create_custom_tag(tag)
|
|
|
|
def get_meta(self, ui=False):
|
|
meta = {
|
|
'uuid': self.uuid,
|
|
'date': self.get_date(),
|
|
'frequency': self.get_frequency(),
|
|
'user': self.get_user(),
|
|
'url': self.get_url(),
|
|
'depth': self.get_depth(),
|
|
'har': self.get_har(),
|
|
'screenshot': self.get_screenshot(),
|
|
'user_agent': self.get_user_agent(),
|
|
'cookiejar': self.get_cookiejar(),
|
|
'header': self.get_header(),
|
|
'proxy': self.get_proxy(),
|
|
'tags': self.get_tags(),
|
|
}
|
|
status = self.get_status()
|
|
if ui:
|
|
status = status.name
|
|
r_str = True
|
|
else:
|
|
r_str = False
|
|
meta['status'] = status
|
|
meta['next_run'] = self.get_next_run(r_str=r_str)
|
|
return meta
|
|
|
|
def get_meta_status(self): # TODO: Description ? Frequency ???
|
|
meta = {'uuid': self.uuid,
|
|
'url': self.get_url(),
|
|
'user': self.get_user(),
|
|
'tags': self.get_tags(),
|
|
'next_run': self.get_next_run(r_str=True)}
|
|
status = self.get_status()
|
|
if isinstance(status, ScheduleStatus):
|
|
status = status.name
|
|
meta['status'] = status
|
|
return meta
|
|
|
|
def create(self, frequency, user, url,
|
|
depth=1, har=True, screenshot=True, header=None, cookiejar=None, proxy=None, user_agent=None, tags=[]):
|
|
|
|
if self.exists():
|
|
raise Exception('Error: Monitor already exists')
|
|
|
|
url_decoded = unpack_url(url)
|
|
url = url_decoded['url']
|
|
|
|
self._set_field('date', datetime.now().strftime("%Y-%m-%d"))
|
|
self._set_field('frequency', frequency)
|
|
self._set_field('user', user)
|
|
self._set_field('url', url)
|
|
self._set_field('depth', int(depth))
|
|
self._set_field('har', str(har))
|
|
self._set_field('screenshot', str(screenshot))
|
|
|
|
if cookiejar:
|
|
self._set_field('cookiejar', cookiejar)
|
|
if header:
|
|
self._set_field('header', header)
|
|
if proxy:
|
|
if proxy == 'web':
|
|
proxy = None
|
|
elif proxy == 'force_tor' or proxy == 'tor' or proxy == 'onion':
|
|
proxy = 'force_tor'
|
|
self._set_field('proxy', proxy)
|
|
if user_agent:
|
|
self._set_field('user_agent', user_agent)
|
|
|
|
if tags:
|
|
self.set_tags(tags)
|
|
|
|
r_crawler.sadd('scheduler:schedules', self.uuid)
|
|
|
|
def delete(self):
|
|
# remove from schedule queue
|
|
r_crawler.zrem('scheduler:queue', self.uuid)
|
|
|
|
# delete task
|
|
task = self.get_task()
|
|
if task:
|
|
task.delete()
|
|
|
|
# delete meta
|
|
r_crawler.delete(f'schedule:{self.uuid}')
|
|
r_crawler.delete(f'schedule:tags:{self.uuid}')
|
|
r_crawler.srem('scheduler:schedules', self.uuid)
|
|
|
|
def create_schedule(frequency, user, url, depth=1, har=True, screenshot=True, header=None, cookiejar=None, proxy=None, user_agent=None, tags=[]):
|
|
schedule_uuid = gen_uuid()
|
|
schedule = CrawlerSchedule(schedule_uuid)
|
|
schedule.create(frequency, user, url, depth=depth, har=har, screenshot=screenshot, header=header, cookiejar=cookiejar, proxy=proxy, user_agent=user_agent, tags=tags)
|
|
return schedule_uuid
|
|
|
|
# TODO sanityze UUID
|
|
def api_delete_schedule(data):
|
|
schedule_uuid = data.get('uuid')
|
|
schedule = CrawlerSchedule(schedule_uuid)
|
|
if not schedule.exists():
|
|
return {'error': 'unknown schedule uuid', 'uuid': schedule}, 404
|
|
return schedule.delete(), 200
|
|
|
|
#### CRAWLER CAPTURE ####
|
|
|
|
def get_nb_crawler_captures():
|
|
return r_cache.zcard('crawler:captures')
|
|
|
|
def get_crawler_captures():
|
|
return r_crawler.zrange('crawler:captures', 0, -1)
|
|
|
|
def reload_crawler_captures():
|
|
r_cache.delete('crawler:captures')
|
|
for capture_uuid in get_crawler_captures():
|
|
capture = CrawlerCapture(capture_uuid)
|
|
capture.update(None)
|
|
|
|
def _clear_captures():
|
|
for capture_uuid in get_crawler_captures():
|
|
capture = CrawlerCapture(capture_uuid)
|
|
task = capture.get_task()
|
|
task.delete()
|
|
capture.delete()
|
|
print(capture_uuid, 'deleted')
|
|
|
|
@unique
|
|
class CaptureStatus(IntEnum):
|
|
"""The status of the capture"""
|
|
UNKNOWN = -1
|
|
QUEUED = 0
|
|
DONE = 1
|
|
ONGOING = 2
|
|
|
|
class CrawlerCapture:
|
|
|
|
def __init__(self, task_uuid):
|
|
self.uuid = task_uuid
|
|
|
|
def exists(self):
|
|
return r_crawler.hexists('crawler:captures:tasks', self.uuid)
|
|
|
|
def get_task_uuid(self):
|
|
return r_crawler.hget('crawler:captures:tasks', self.uuid)
|
|
|
|
def get_task(self):
|
|
task_uuid = self.get_task_uuid()
|
|
if task_uuid:
|
|
return CrawlerTask(task_uuid)
|
|
|
|
def get_start_time(self, r_str=True):
|
|
start_time = self.get_task().get_start_time()
|
|
if r_str:
|
|
return start_time
|
|
elif not start_time:
|
|
return 0
|
|
else:
|
|
start_time = datetime.strptime(start_time, "%Y/%m/%d - %H:%M.%S").timestamp()
|
|
return int(start_time)
|
|
|
|
def get_status(self):
|
|
status = r_cache.hget(f'crawler:capture:{self.uuid}', 'status')
|
|
if not status:
|
|
status = -1
|
|
return status
|
|
|
|
def is_ongoing(self):
|
|
return self.get_status() == CaptureStatus.ONGOING
|
|
|
|
def create(self, task_uuid):
|
|
if self.exists():
|
|
print(f'Capture {self.uuid} already exists') # TODO LOGS
|
|
return None
|
|
launch_time = int(time.time())
|
|
r_crawler.hset(f'crawler:task:{task_uuid}', 'capture', self.uuid)
|
|
r_crawler.hset('crawler:captures:tasks', self.uuid, task_uuid)
|
|
r_crawler.zadd('crawler:captures', {self.uuid: launch_time})
|
|
r_cache.hset(f'crawler:capture:{self.uuid}', 'launch_time', launch_time)
|
|
r_cache.zadd('crawler:captures', {self.uuid: launch_time})
|
|
|
|
def update(self, status):
|
|
# Error or Reload
|
|
if not status:
|
|
r_cache.hset(f'crawler:capture:{self.uuid}', 'status', CaptureStatus.UNKNOWN.value)
|
|
r_cache.zadd('crawler:captures', {self.uuid: 0})
|
|
else:
|
|
last_check = int(time.time())
|
|
r_cache.hset(f'crawler:capture:{self.uuid}', 'status', status)
|
|
r_cache.zadd('crawler:captures', {self.uuid: last_check})
|
|
|
|
# Crawler
|
|
def remove(self):
|
|
r_crawler.zrem('crawler:captures', self.uuid)
|
|
r_cache.delete(f'crawler:capture:{self.uuid}')
|
|
r_crawler.hdel('crawler:captures:tasks', self.uuid)
|
|
|
|
# Manual
|
|
def delete(self):
|
|
# remove Capture from crawler queue
|
|
r_cache.zrem('crawler:captures', self.uuid)
|
|
self.remove()
|
|
|
|
|
|
def create_capture(capture_uuid, task_uuid):
|
|
capture = CrawlerCapture(capture_uuid)
|
|
capture.create(task_uuid)
|
|
|
|
def get_crawler_capture():
|
|
capture = r_cache.zpopmin('crawler:captures')
|
|
if capture:
|
|
capture = CrawlerCapture(capture[0][0])
|
|
else:
|
|
capture = None
|
|
return capture
|
|
|
|
# TODO add capture times
|
|
def get_captures_status():
|
|
status = []
|
|
for capture_uuid in get_crawler_captures():
|
|
capture = CrawlerCapture(capture_uuid)
|
|
task = capture.get_task()
|
|
domain = task.get_domain()
|
|
dom = Domain(domain)
|
|
meta = {
|
|
'uuid': task.uuid,
|
|
'domain': dom.get_id(),
|
|
'type': dom.get_domain_type(),
|
|
'start_time': capture.get_start_time(),
|
|
'status': capture.get_status(),
|
|
}
|
|
capture_status = capture.get_status()
|
|
if capture_status:
|
|
capture_status = CaptureStatus(int(capture_status)).name
|
|
meta['status'] = capture_status
|
|
status.append(meta)
|
|
return status
|
|
|
|
def delete_captures():
|
|
for capture_uuid in get_crawler_captures():
|
|
capture = CrawlerCapture(capture_uuid)
|
|
capture.delete()
|
|
|
|
##-- CRAWLER STATE --##
|
|
|
|
|
|
#### CRAWLER TASK ####
|
|
|
|
class CrawlerTask:
|
|
|
|
def __init__(self, task_uuid):
|
|
self.uuid = task_uuid
|
|
|
|
def exists(self):
|
|
return r_crawler.exists(f'crawler:task:{self.uuid}')
|
|
|
|
def get_url(self):
|
|
return r_crawler.hget(f'crawler:task:{self.uuid}', 'url')
|
|
|
|
def get_domain(self):
|
|
return r_crawler.hget(f'crawler:task:{self.uuid}', 'domain')
|
|
|
|
def get_depth(self):
|
|
depth = r_crawler.hget(f'crawler:task:{self.uuid}', 'depth')
|
|
if not depth:
|
|
depth = 1
|
|
return int(depth)
|
|
|
|
def get_har(self):
|
|
return r_crawler.hget(f'crawler:task:{self.uuid}', 'har') == '1'
|
|
|
|
def get_screenshot(self):
|
|
return r_crawler.hget(f'crawler:task:{self.uuid}', 'screenshot') == '1'
|
|
|
|
def get_queue(self):
|
|
return r_crawler.hget(f'crawler:task:{self.uuid}', 'queue')
|
|
|
|
def get_user_agent(self):
|
|
user_agent = r_crawler.hget(f'crawler:task:{self.uuid}', 'user_agent')
|
|
if not user_agent:
|
|
user_agent = get_default_user_agent()
|
|
return user_agent
|
|
|
|
def get_cookiejar(self):
|
|
return r_crawler.hget(f'crawler:task:{self.uuid}', 'cookiejar')
|
|
|
|
def get_cookies(self):
|
|
cookiejar = self.get_cookiejar()
|
|
if cookiejar:
|
|
cookiejar = Cookiejar(cookiejar)
|
|
return cookiejar.get_cookies()
|
|
else:
|
|
return []
|
|
|
|
def get_header(self):
|
|
return r_crawler.hget(f'crawler:task:{self.uuid}', 'header')
|
|
|
|
def get_proxy(self):
|
|
return r_crawler.hget(f'crawler:task:{self.uuid}', 'proxy')
|
|
|
|
def get_parent(self):
|
|
return r_crawler.hget(f'crawler:task:{self.uuid}', 'parent')
|
|
|
|
def get_hash(self):
|
|
return r_crawler.hget(f'crawler:task:{self.uuid}', 'hash')
|
|
|
|
def get_start_time(self):
|
|
return r_crawler.hget(f'crawler:task:{self.uuid}', 'start_time')
|
|
|
|
# TODO
|
|
def get_status(self):
|
|
return r_crawler.hget(f'crawler:task:{self.uuid}', 'status') #######################################
|
|
|
|
def get_capture(self):
|
|
return r_crawler.hget(f'crawler:task:{self.uuid}', 'capture')
|
|
|
|
def is_ongoing(self):
|
|
capture_uuid = self.get_capture()
|
|
if capture_uuid:
|
|
return CrawlerCapture(capture_uuid).is_ongoing()
|
|
return False
|
|
|
|
def _set_field(self, field, value):
|
|
return r_crawler.hset(f'crawler:task:{self.uuid}', field, value)
|
|
|
|
def get_tags(self):
|
|
return r_crawler.smembers(f'crawler:task:tags:{self.uuid}')
|
|
|
|
def set_tags(self, tags):
|
|
for tag in tags:
|
|
r_crawler.sadd(f'crawler:task:tags:{self.uuid}', tag)
|
|
# Tag.create_custom_tag(tag)
|
|
|
|
def get_meta(self):
|
|
meta = {
|
|
'uuid': self.uuid,
|
|
'url': self.get_url(),
|
|
'domain': self.get_domain(),
|
|
'depth': self.get_depth(),
|
|
'har': self.get_har(),
|
|
'screenshot': self.get_screenshot(),
|
|
'type': self.get_queue(),
|
|
'user_agent': self.get_user_agent(),
|
|
'cookiejar': self.get_cookiejar(),
|
|
'header': self.get_header(),
|
|
'proxy': self.get_proxy(),
|
|
'parent': self.get_parent(),
|
|
'tags': self.get_tags(),
|
|
}
|
|
return meta
|
|
|
|
# TODO STATUS UPDATE
|
|
# TODO SANITIZE PRIORITY
|
|
# PRIORITY: discovery = 0/10, feeder = 10, manual = 50, auto = 40, test = 100
|
|
def create(self, url, depth=1, har=True, screenshot=True, header=None, cookiejar=None, proxy=None,
|
|
user_agent=None, tags=[], parent='manual', priority=0, external=False):
|
|
if self.exists():
|
|
raise Exception('Error: Task already exists')
|
|
|
|
url_decoded = unpack_url(url)
|
|
url = url_decoded['url']
|
|
domain = url_decoded['domain']
|
|
|
|
dom = Domain(domain)
|
|
|
|
# Discovery crawler
|
|
if priority == 0:
|
|
if is_blacklisted_domain(dom.get_id()):
|
|
return None
|
|
if not dom.exists():
|
|
priority = 10
|
|
# Domain Crawled today or UP this month
|
|
if dom.is_down_today() or dom.is_up_this_month():
|
|
return None
|
|
|
|
har = int(har)
|
|
screenshot = int(screenshot)
|
|
|
|
if proxy == 'web':
|
|
proxy = None
|
|
elif proxy == 'force_tor' or proxy == 'tor' or proxy == 'onion':
|
|
proxy = 'force_tor'
|
|
|
|
# TODO SANITIZE COOKIEJAR -> UUID
|
|
|
|
# Check if already in queue
|
|
hash_query = get_task_hash(url, domain, depth, har, screenshot, priority, proxy, cookiejar, user_agent, header, tags)
|
|
if r_crawler.hexists(f'crawler:queue:hash', hash_query):
|
|
self.uuid = r_crawler.hget(f'crawler:queue:hash', hash_query)
|
|
return self.uuid
|
|
|
|
self._set_field('domain', domain)
|
|
self._set_field('url', url)
|
|
self._set_field('depth', int(depth))
|
|
self._set_field('har', har)
|
|
self._set_field('screenshot', screenshot)
|
|
self._set_field('parent', parent)
|
|
|
|
if cookiejar:
|
|
self._set_field('cookiejar', cookiejar)
|
|
if header:
|
|
self._set_field('header', header)
|
|
if proxy:
|
|
self._set_field('proxy', proxy)
|
|
if user_agent:
|
|
self._set_field('user_agent', user_agent)
|
|
|
|
if tags:
|
|
self.set_tags(tags)
|
|
|
|
r_crawler.hset('crawler:queue:hash', hash_query, self.uuid)
|
|
self._set_field('hash', hash_query)
|
|
if not external:
|
|
self.add_to_db_crawler_queue(priority)
|
|
# UI
|
|
domain_type = dom.get_domain_type()
|
|
r_crawler.sadd(f'crawler:queue:type:{domain_type}', self.uuid)
|
|
self._set_field('queue', domain_type)
|
|
return self.uuid
|
|
|
|
def add_to_db_crawler_queue(self, priority):
|
|
r_crawler.zadd('crawler:queue', {self.uuid: priority})
|
|
|
|
def start(self):
|
|
self._set_field('start_time', datetime.now().strftime("%Y/%m/%d - %H:%M.%S"))
|
|
|
|
def reset(self):
|
|
priority = 49
|
|
r_crawler.hdel(f'crawler:task:{self.uuid}', 'start_time')
|
|
self.add_to_db_crawler_queue(priority)
|
|
|
|
# Crawler
|
|
def remove(self): # zrem cache + DB
|
|
capture_uuid = self.get_capture()
|
|
if capture_uuid:
|
|
capture = CrawlerCapture(capture_uuid)
|
|
capture.remove()
|
|
queue_type = self.get_queue()
|
|
if queue_type:
|
|
r_crawler.srem(f'crawler:queue:type:{queue_type}', self.uuid)
|
|
task_hash = self.get_hash()
|
|
if task_hash:
|
|
r_crawler.hdel('crawler:queue:hash', task_hash)
|
|
# meta
|
|
r_crawler.delete(f'crawler:task:{self.uuid}')
|
|
|
|
# Manual
|
|
def delete(self):
|
|
# queue
|
|
r_crawler.zrem('crawler:queue', self.uuid)
|
|
self.remove()
|
|
|
|
|
|
# TODO move to class ???
|
|
def get_task_hash(url, domain, depth, har, screenshot, priority, proxy, cookiejar, user_agent, header, tags):
|
|
to_enqueue = {'domain': domain, 'depth': depth, 'har': har, 'screenshot': screenshot,
|
|
'priority': priority, 'proxy': proxy, 'cookiejar': cookiejar, 'user_agent': user_agent,
|
|
'header': header, 'tags': tags}
|
|
if priority != 0:
|
|
to_enqueue['url'] = url
|
|
return hashlib.sha512(pickle.dumps(to_enqueue)).hexdigest()
|
|
|
|
def add_task_to_lacus_queue():
|
|
task_uuid = r_crawler.zpopmax('crawler:queue')
|
|
if not task_uuid or not task_uuid[0]:
|
|
return None
|
|
task_uuid, priority = task_uuid[0]
|
|
task = CrawlerTask(task_uuid)
|
|
return task, priority
|
|
|
|
# PRIORITY: discovery = 0/10, feeder = 10, manual = 50, auto = 40, test = 100
|
|
def create_task(url, depth=1, har=True, screenshot=True, header=None, cookiejar=None, proxy=None,
|
|
user_agent=None, tags=[], parent='manual', priority=0, task_uuid=None, external=False):
|
|
if task_uuid:
|
|
if CrawlerTask(task_uuid).exists():
|
|
task_uuid = gen_uuid()
|
|
else:
|
|
task_uuid = gen_uuid()
|
|
task = CrawlerTask(task_uuid)
|
|
task_uuid = task.create(url, depth=depth, har=har, screenshot=screenshot, header=header, cookiejar=cookiejar,
|
|
proxy=proxy, user_agent=user_agent, tags=tags, parent=parent, priority=priority,
|
|
external=external)
|
|
return task_uuid
|
|
|
|
|
|
## -- CRAWLER TASK -- ##
|
|
|
|
#### CRAWLER TASK API ####
|
|
|
|
# # TODO: ADD user agent
|
|
# # TODO: sanitize URL
|
|
|
|
def api_parse_task_dict_basic(data, user_id):
|
|
url = data.get('url', None)
|
|
if not url or url == '\n':
|
|
return {'status': 'error', 'reason': 'No url supplied'}, 400
|
|
|
|
screenshot = data.get('screenshot', False)
|
|
if screenshot:
|
|
screenshot = True
|
|
else:
|
|
screenshot = False
|
|
har = data.get('har', False)
|
|
if har:
|
|
har = True
|
|
else:
|
|
har = False
|
|
depth_limit = data.get('depth', 1)
|
|
if depth_limit:
|
|
try:
|
|
depth_limit = int(depth_limit)
|
|
if depth_limit < 0:
|
|
depth_limit = 0
|
|
except ValueError:
|
|
return {'error': 'invalid depth limit'}, 400
|
|
else:
|
|
depth_limit = 0
|
|
|
|
# PROXY
|
|
proxy = data.get('proxy', None)
|
|
if proxy == 'onion' or proxy == 'tor' or proxy == 'force_tor':
|
|
proxy = 'force_tor'
|
|
elif proxy:
|
|
verify = api_verify_proxy(proxy)
|
|
if verify[1] != 200:
|
|
return verify
|
|
|
|
tags = data.get('tags', [])
|
|
|
|
return {'url': url, 'depth_limit': depth_limit, 'har': har, 'screenshot': screenshot, 'proxy': proxy, 'tags': tags}, 200
|
|
|
|
def api_add_crawler_task(data, user_id=None):
|
|
task, resp = api_parse_task_dict_basic(data, user_id)
|
|
if resp != 200:
|
|
return task, resp
|
|
|
|
url = task['url']
|
|
screenshot = task['screenshot']
|
|
har = task['har']
|
|
depth_limit = task['depth_limit']
|
|
proxy = task['proxy']
|
|
tags = task['tags']
|
|
|
|
cookiejar_uuid = data.get('cookiejar', None)
|
|
if cookiejar_uuid:
|
|
cookiejar = Cookiejar(cookiejar_uuid)
|
|
if not cookiejar.exists():
|
|
return {'error': 'unknown cookiejar uuid', 'cookiejar_uuid': cookiejar_uuid}, 404
|
|
level = cookiejar.get_level()
|
|
if level == 0: # # TODO: check if user is admin
|
|
if cookiejar.get_user() != user_id:
|
|
return {'error': 'The access to this cookiejar is restricted'}, 403
|
|
cookiejar_uuid = cookiejar.uuid
|
|
|
|
cookies = data.get('cookies', None)
|
|
if not cookiejar_uuid and cookies:
|
|
# Create new cookiejar
|
|
cookiejar_uuid = create_cookiejar(user_id, "single-shot cookiejar", 1, None)
|
|
cookiejar = Cookiejar(cookiejar_uuid)
|
|
for cookie in cookies:
|
|
try:
|
|
name = cookie.get('name')
|
|
value = cookie.get('value')
|
|
cookiejar.add_cookie(name, value, None, None, None, None, None)
|
|
except KeyError:
|
|
return {'error': 'Invalid cookie key, please submit a valid JSON', 'cookiejar_uuid': cookiejar_uuid}, 400
|
|
|
|
frequency = data.get('frequency', None)
|
|
if frequency:
|
|
if frequency not in ['monthly', 'weekly', 'daily', 'hourly']:
|
|
if not isinstance(frequency, dict):
|
|
return {'error': 'Invalid frequency'}, 400
|
|
else:
|
|
try:
|
|
months = int(frequency.get('months', 0))
|
|
weeks = int(frequency.get('weeks', 0))
|
|
days = int(frequency.get('days', 0))
|
|
hours = int(frequency.get('hours', 0))
|
|
minutes = int(frequency.get('minutes', 0))
|
|
except (TypeError, ValueError):
|
|
return {'error': 'Invalid frequency'}, 400
|
|
if min(months, weeks, days, hours, minutes) < 0:
|
|
return {'error': 'Invalid frequency'}, 400
|
|
if max(months, weeks, days, hours, minutes) <= 0:
|
|
return {'error': 'Invalid frequency'}, 400
|
|
frequency = f'{months}:{weeks}:{days}:{hours}:{minutes}'
|
|
|
|
if frequency:
|
|
# TODO verify user
|
|
task_uuid = create_schedule(frequency, user_id, url, depth=depth_limit, har=har, screenshot=screenshot, header=None,
|
|
cookiejar=cookiejar_uuid, proxy=proxy, user_agent=None, tags=tags)
|
|
else:
|
|
# TODO HEADERS
|
|
# TODO USER AGENT
|
|
task_uuid = create_task(url, depth=depth_limit, har=har, screenshot=screenshot, header=None,
|
|
cookiejar=cookiejar_uuid, proxy=proxy, user_agent=None, tags=tags,
|
|
parent='manual', priority=90)
|
|
|
|
return {'uuid': task_uuid}, 200
|
|
|
|
|
|
#### ####
|
|
|
|
# TODO cookiejar - cookies - frequency
|
|
def api_add_crawler_capture(data, user_id):
|
|
task, resp = api_parse_task_dict_basic(data, user_id)
|
|
if resp != 200:
|
|
return task, resp
|
|
|
|
task_uuid = data.get('task_uuid')
|
|
if not task_uuid:
|
|
return {'error': 'Invalid task_uuid', 'task_uuid': task_uuid}, 400
|
|
capture_uuid = data.get('capture_uuid')
|
|
if not capture_uuid:
|
|
return {'error': 'Invalid capture_uuid', 'capture_uuid': capture_uuid}, 400
|
|
|
|
# parent = data.get('parent')
|
|
|
|
# TODO parent
|
|
task_uuid = create_task(task['url'], depth=task['depth_limit'], har=task['har'], screenshot=task['screenshot'],
|
|
proxy=task['proxy'], tags=task['tags'],
|
|
parent='manual', task_uuid=task_uuid, external=True)
|
|
if not task_uuid:
|
|
return {'error': 'Aborted by Crawler', 'task_uuid': task_uuid, 'capture_uuid': capture_uuid}, 400
|
|
task = CrawlerTask(task_uuid)
|
|
create_capture(capture_uuid, task_uuid)
|
|
task.start()
|
|
return {'uuid': capture_uuid}, 200
|
|
|
|
###################################################################################
|
|
###################################################################################
|
|
###################################################################################
|
|
###################################################################################
|
|
|
|
|
|
#### CRAWLER GLOBAL ####
|
|
|
|
# TODO: # FIXME: config db, dynamic load
|
|
def is_crawler_activated():
|
|
return activate_crawler == 'True'
|
|
|
|
def get_crawler_all_types():
|
|
return ['onion', 'web']
|
|
|
|
##-- CRAWLER GLOBAL --##
|
|
|
|
|
|
#### ####
|
|
|
|
|
|
def is_redirection(domain, last_url):
|
|
url = urlparse(last_url)
|
|
last_domain = url.netloc
|
|
last_domain = last_domain.split('.')
|
|
last_domain = '{}.{}'.format(last_domain[-2], last_domain[-1])
|
|
return domain != last_domain
|
|
|
|
def create_item_id(item_dir, domain):
|
|
# remove /
|
|
domain = domain.replace('/', '_')
|
|
if len(domain) > 215:
|
|
UUID = domain[-215:]+str(uuid.uuid4())
|
|
else:
|
|
UUID = domain+str(uuid.uuid4())
|
|
return os.path.join(item_dir, UUID)
|
|
|
|
# # # # # # # # # # # #
|
|
# #
|
|
# CRAWLER MANAGER # TODO REFACTOR ME
|
|
# #
|
|
# # # # # # # # # # # #
|
|
|
|
#### PROXY ####
|
|
|
|
def api_verify_proxy(proxy_url):
|
|
parsed_proxy = urlparse(proxy_url)
|
|
if parsed_proxy.scheme and parsed_proxy.hostname and parsed_proxy.port:
|
|
if parsed_proxy.scheme in ['http', 'https', 'socks5']:
|
|
if (parsed_proxy.username and parsed_proxy.password) != (
|
|
not parsed_proxy.username and not parsed_proxy.password):
|
|
return proxy_url, 200
|
|
else:
|
|
return {'error': 'You need to enter a username AND a password for your proxy.'}, 400
|
|
else:
|
|
return {'error': 'Proxy scheme not supported: must be http(s) or socks5.'}, 400
|
|
else:
|
|
return {'error': 'Invalid proxy: Check that you entered a scheme, a hostname and a port.'}, 400
|
|
|
|
def get_proxies():
|
|
return r_crawler.smembers('crawler:proxies')
|
|
|
|
class CrawlerProxy:
|
|
def __init__(self, proxy_uuid):
|
|
self.uuid = proxy_uuid
|
|
|
|
def get_description(self):
|
|
return r_crawler.hget(f'crawler:proxy:{self.uuid}', 'description')
|
|
|
|
# Host
|
|
# Port
|
|
# Type -> need test
|
|
def get_url(self):
|
|
return r_crawler.hget(f'crawler:proxy:{self.uuid}', 'url')
|
|
|
|
#### CRAWLER LACUS ####
|
|
|
|
def get_lacus_url():
|
|
return r_db.hget('crawler:lacus', 'url')
|
|
|
|
def get_lacus_api_key():
|
|
return r_db.hget('crawler:lacus', 'key')
|
|
|
|
# TODO Rewrite with new API key
|
|
def get_hidden_lacus_api_key():
|
|
key = get_lacus_api_key()
|
|
if key:
|
|
if len(key) == 41:
|
|
return f'{key[:4]}*********************************{key[-4:]}'
|
|
|
|
# TODO Rewrite with new API key
|
|
def is_valid_api_key(api_key, search=re.compile(r'[^a-zA-Z0-9_-]').search):
|
|
if len(api_key) != 41:
|
|
return False
|
|
return not bool(search(api_key))
|
|
|
|
def save_lacus_url_api(url, api_key):
|
|
r_db.hset('crawler:lacus', 'url', url)
|
|
# r_db.hset('crawler:lacus', 'key', api_key)
|
|
|
|
def is_lacus_connected(delta_check=30):
|
|
last_check = r_cache.hget('crawler:lacus', 'last_check')
|
|
if last_check:
|
|
if int(time.time()) - int(last_check) > delta_check:
|
|
ping_lacus()
|
|
else:
|
|
ping_lacus()
|
|
is_connected = r_cache.hget('crawler:lacus', 'connected')
|
|
return is_connected == 'True'
|
|
|
|
def get_lacus_connection_metadata(force_ping=False):
|
|
dict_manager = {}
|
|
if force_ping:
|
|
dict_manager['status'] = ping_lacus()
|
|
else:
|
|
dict_manager['status'] = is_lacus_connected()
|
|
if not dict_manager['status']:
|
|
dict_manager['status_code'] = r_cache.hget('crawler:lacus', 'status_code')
|
|
dict_manager['error'] = r_cache.hget('crawler:lacus', 'error')
|
|
return dict_manager
|
|
|
|
def get_lacus():
|
|
url = get_lacus_url()
|
|
if url:
|
|
return PyLacus(get_lacus_url())
|
|
|
|
# TODO CATCH EXCEPTIONS
|
|
def ping_lacus():
|
|
# TODO CATCH EXCEPTION
|
|
req_error = None
|
|
lacus = get_lacus()
|
|
if not lacus:
|
|
ping = False
|
|
req_error = {'error': 'Lacus URL undefined', 'status_code': 400}
|
|
else:
|
|
try:
|
|
ping = lacus.is_up
|
|
except:
|
|
req_error = {'error': 'Failed to connect Lacus URL', 'status_code': 400}
|
|
ping = False
|
|
update_lacus_connection_status(ping, req_error=req_error)
|
|
return ping
|
|
|
|
def update_lacus_connection_status(is_connected, req_error=None):
|
|
r_cache.hset('crawler:lacus', 'connected', str(is_connected))
|
|
r_cache.hset('crawler:lacus', 'last_check', int(time.time()))
|
|
if not req_error:
|
|
r_cache.hdel('crawler:lacus', 'error')
|
|
else:
|
|
r_cache.hset('crawler:lacus', 'status_code', req_error['status_code'])
|
|
r_cache.hset('crawler:lacus', 'error', req_error['error'])
|
|
|
|
def api_save_lacus_url_key(data):
|
|
# unpack json
|
|
manager_url = data.get('url', None)
|
|
api_key = data.get('api_key', None)
|
|
if not manager_url: # or not api_key:
|
|
return {'status': 'error', 'reason': 'No url or API key supplied'}, 400
|
|
# check if is valid url
|
|
try:
|
|
result = urlparse(manager_url)
|
|
if not all([result.scheme, result.netloc]):
|
|
return {'status': 'error', 'reason': 'Invalid url'}, 400
|
|
except:
|
|
return {'status': 'error', 'reason': 'Invalid url'}, 400
|
|
|
|
# # check if is valid key CURRENTLY DISABLE
|
|
# if not is_valid_api_key(api_key):
|
|
# return ({'status': 'error', 'reason': 'Invalid API key'}, 400)
|
|
|
|
save_lacus_url_api(manager_url, api_key)
|
|
return {'url': manager_url, 'api_key': get_hidden_lacus_api_key()}, 200
|
|
|
|
def get_crawler_max_captures():
|
|
nb_captures = r_cache.hget('crawler:lacus', 'nb_captures')
|
|
if not nb_captures:
|
|
nb_captures = r_db.hget('crawler:lacus', 'nb_captures')
|
|
if not nb_captures:
|
|
nb_captures = 10
|
|
save_nb_max_captures(nb_captures)
|
|
else:
|
|
r_cache.hset('crawler:lacus', 'nb_captures', int(nb_captures))
|
|
return int(nb_captures)
|
|
|
|
def save_nb_max_captures(nb_captures):
|
|
r_db.hset('crawler:lacus', 'nb_captures', int(nb_captures))
|
|
r_cache.hset('crawler:lacus', 'nb_captures', int(nb_captures))
|
|
|
|
def api_set_crawler_max_captures(data):
|
|
nb_captures = data.get('nb', 10)
|
|
try:
|
|
nb_captures = int(nb_captures)
|
|
if nb_captures < 1:
|
|
nb_captures = 1
|
|
except (TypeError, ValueError):
|
|
return {'error': 'Invalid number of crawlers to launch'}, 400
|
|
save_nb_max_captures(nb_captures)
|
|
return nb_captures, 200
|
|
|
|
## TEST ##
|
|
|
|
def is_test_ail_crawlers_successful():
|
|
return r_db.hget('crawler:tor:test', 'success') == 'True'
|
|
|
|
def get_test_ail_crawlers_message():
|
|
return r_db.hget('crawler:tor:test', 'message')
|
|
|
|
def save_test_ail_crawlers_result(test_success, message):
|
|
r_db.hset('crawler:tor:test', 'success', str(test_success))
|
|
r_db.hset('crawler:tor:test', 'message', message)
|
|
|
|
def test_ail_crawlers():
|
|
# # TODO: test web domain
|
|
if not ping_lacus():
|
|
lacus_url = get_lacus_url()
|
|
error_message = f'Error: Can\'t connect to AIL Lacus, {lacus_url}'
|
|
print(error_message)
|
|
save_test_ail_crawlers_result(False, error_message)
|
|
return False
|
|
|
|
lacus = get_lacus()
|
|
commit_id = git_status.get_last_commit_id_from_local()
|
|
user_agent = f'{commit_id}-AIL LACUS CRAWLER'
|
|
# domain = 'eswpccgr5xyovsahffkehgleqthrasfpfdblwbs4lstd345dwq5qumqd.onion'
|
|
url = 'http://eswpccgr5xyovsahffkehgleqthrasfpfdblwbs4lstd345dwq5qumqd.onion'
|
|
|
|
## LAUNCH CRAWLER, TEST MODE ##
|
|
# set_current_crawler_status(splash_url, 'CRAWLER TEST', started_time=True,
|
|
# crawled_domain='TEST DOMAIN', crawler_type='onion')
|
|
capture_uuid = lacus.enqueue(url=url, depth=0, user_agent=user_agent, proxy='force_tor',
|
|
force=True, general_timeout_in_sec=90)
|
|
status = lacus.get_capture_status(capture_uuid)
|
|
launch_time = int(time.time()) # capture timeout
|
|
while int(time.time()) - launch_time < 90 and status != CaptureStatus.DONE:
|
|
# DEBUG
|
|
print(int(time.time()) - launch_time)
|
|
print(status)
|
|
time.sleep(1)
|
|
status = lacus.get_capture_status(capture_uuid)
|
|
|
|
# TODO CRAWLER STATUS OR QUEUED CAPTURE LIST
|
|
entries = lacus.get_capture(capture_uuid)
|
|
if 'error' in entries:
|
|
save_test_ail_crawlers_result(False, entries['error'])
|
|
return False
|
|
elif 'html' in entries and entries['html']:
|
|
mess = 'It works!'
|
|
if mess in entries['html']:
|
|
save_test_ail_crawlers_result(True, mess)
|
|
return True
|
|
else:
|
|
return False
|
|
elif status == 2:
|
|
save_test_ail_crawlers_result(False, 'Timeout Error')
|
|
else:
|
|
save_test_ail_crawlers_result(False, 'Error')
|
|
return False
|
|
|
|
#### ---- ####
|
|
|
|
# TODO CHECK MIGRATION - Rest API
|
|
|
|
# TODO MIGRATE ME
|
|
# def api_create_crawler_task(user_id, url, screenshot=True, har=True, depth_limit=1, max_pages=100, auto_crawler=False, crawler_delta=3600, crawler_type=None, cookiejar_uuid=None, user_agent=None):
|
|
# # validate url
|
|
# if url is None or url=='' or url=='\n':
|
|
# return ({'error':'invalid depth limit'}, 400)
|
|
|
|
|
|
# TODO MOVE ME IN CRAWLER OR FLASK
|
|
load_blacklist()
|
|
|
|
# if __name__ == '__main__':
|
|
# delete_captures()
|
|
|
|
# item_id = 'crawled/2023/02/20/data.gz'
|
|
# item = Item(item_id)
|
|
# content = item.get_content()
|
|
# temp_url = ''
|
|
# r = extract_favicon_from_html(content, temp_url)
|
|
# print(r)
|
|
# _reprocess_all_hars_cookie_name()
|
|
# _reprocess_all_hars_etag()
|
|
# _gzip_all_hars()
|
|
# _reprocess_all_hars_hhhashs()
|