mirror of https://github.com/CIRCL/AIL-framework
				
				
				
			
		
			
				
	
	
		
			229 lines
		
	
	
		
			7.8 KiB
		
	
	
	
		
			Python
		
	
	
		
			Executable File
		
	
			
		
		
	
	
			229 lines
		
	
	
		
			7.8 KiB
		
	
	
	
		
			Python
		
	
	
		
			Executable File
		
	
| #!/usr/bin/env python3
 | |
| # -*-coding:UTF-8 -*
 | |
| """
 | |
| The ZMQ_Feed_Q Module
 | |
| =====================
 | |
| 
 | |
| This module is consuming the Redis-list created by the ZMQ_Feed_Q Module,
 | |
| And save the item on disk to allow others modules to work on them.
 | |
| 
 | |
| ..todo:: Be able to choose to delete or not the saved item after processing.
 | |
| 
 | |
| ..note:: Module ZMQ_Something_Q and ZMQ_Something are closely bound, always put
 | |
| the same Subscriber name in both of them.
 | |
| 
 | |
| Requirements
 | |
| ------------
 | |
| 
 | |
| *Need running Redis instances.
 | |
| *Need the Mixer or the Importer Module running to be able to work properly.
 | |
| 
 | |
| """
 | |
| 
 | |
| ##################################
 | |
| # Import External packages
 | |
| ##################################
 | |
| import base64
 | |
| import io
 | |
| import gzip
 | |
| import os
 | |
| import sys
 | |
| import time
 | |
| 
 | |
| from hashlib import md5
 | |
| from uuid import uuid4
 | |
| 
 | |
| sys.path.append(os.environ['AIL_BIN'])
 | |
| ##################################
 | |
| # Import Project packages
 | |
| ##################################
 | |
| from modules.abstract_module import AbstractModule
 | |
| from lib.ail_core import get_ail_uuid
 | |
| from lib.ConfigLoader import ConfigLoader
 | |
| from lib.data_retention_engine import update_obj_date
 | |
| from lib.objects.Items import Item
 | |
| 
 | |
| # from lib import Statistics
 | |
| 
 | |
| class Global(AbstractModule):
 | |
|     """
 | |
|     Global module for AIL framework
 | |
|     """
 | |
| 
 | |
|     def __init__(self):
 | |
|         super(Global, self).__init__()
 | |
| 
 | |
|         self.processed_item = 0
 | |
|         self.time_last_stats = time.time()
 | |
| 
 | |
|         config_loader = ConfigLoader()
 | |
| 
 | |
|         # Get and sanitize ITEM DIRECTORY
 | |
|         # # TODO: rename PASTE => ITEM
 | |
|         self.ITEMS_FOLDER = os.path.join(os.environ['AIL_HOME'], config_loader.get_config_str("Directories", "pastes")) + '/'
 | |
|         self.ITEMS_FOLDER = os.path.join(os.path.realpath(self.ITEMS_FOLDER), '')
 | |
| 
 | |
|         # Waiting time in seconds between to message processed
 | |
|         self.pending_seconds = 0.5
 | |
| 
 | |
|         # Send module state to logs
 | |
|         self.logger.info(f"Module {self.module_name} initialized")
 | |
| 
 | |
|     def computeNone(self):
 | |
|         difftime = time.time() - self.time_last_stats
 | |
|         if int(difftime) > 30:
 | |
|             to_print = f'Global; ; ; ;glob Processed {self.processed_item} item(s) in {difftime} s'
 | |
|             print(to_print)
 | |
|             self.redis_logger.debug(to_print)
 | |
| 
 | |
|             self.time_last_stats = time.time()
 | |
|             self.processed_item = 0
 | |
| 
 | |
|     def compute(self, message, r_result=False): # TODO move OBJ ID sanitization to importer
 | |
|         # Recovering the streamed message infos
 | |
|         gzip64encoded = message
 | |
| 
 | |
|         if self.obj.type == 'item':
 | |
|             if gzip64encoded:
 | |
| 
 | |
|                 # Creating the full filepath
 | |
|                 filename = os.path.join(self.ITEMS_FOLDER, self.obj.id)
 | |
|                 filename = os.path.realpath(filename)
 | |
| 
 | |
|                 # Incorrect filename
 | |
|                 if not os.path.commonprefix([filename, self.ITEMS_FOLDER]) == self.ITEMS_FOLDER:
 | |
|                     self.logger.warning(f'Global; Path traversal detected {filename}')
 | |
|                     print(f'Global; Path traversal detected {filename}')
 | |
| 
 | |
|                 else:
 | |
|                     # Decode compressed base64
 | |
|                     decoded = base64.standard_b64decode(gzip64encoded)
 | |
|                     new_file_content = self.gunzip_bytes_obj(filename, decoded)
 | |
| 
 | |
|                     # TODO REWRITE ME
 | |
|                     if new_file_content:
 | |
|                         filename = self.check_filename(filename, new_file_content)
 | |
| 
 | |
|                         if filename:
 | |
|                             # create subdir
 | |
|                             dirname = os.path.dirname(filename)
 | |
|                             if not os.path.exists(dirname):
 | |
|                                 os.makedirs(dirname)
 | |
| 
 | |
|                             with open(filename, 'wb') as f:
 | |
|                                 f.write(decoded)
 | |
| 
 | |
|                             update_obj_date(self.obj.get_date(), 'item')
 | |
| 
 | |
|                             self.add_message_to_queue(obj=self.obj, queue='Item')
 | |
|                             self.processed_item += 1
 | |
| 
 | |
|                             print(self.obj.id)
 | |
|                             if r_result:
 | |
|                                 return self.obj.id
 | |
| 
 | |
|             else:
 | |
|                 self.logger.info(f"Empty Item: {message} not processed")
 | |
|         elif self.obj.type == 'message':
 | |
|             # TODO send to specific object queue => image, ...
 | |
|             self.add_message_to_queue(obj=self.obj, queue='Item')
 | |
|         elif self.obj.type == 'image':
 | |
|             self.add_message_to_queue(obj=self.obj, queue='Image')
 | |
|         else:
 | |
|             self.logger.critical(f"Empty obj: {self.obj} {message} not processed")
 | |
| 
 | |
|     def check_filename(self, filename, new_file_content):
 | |
|         """
 | |
|         Check if file is not a duplicated file
 | |
|         return the filename if new file, else None
 | |
|         """
 | |
| 
 | |
|         # check if file exist
 | |
|         if os.path.isfile(filename):
 | |
|             self.logger.info(f'File already exist {filename}')
 | |
|             print(f'File already exist {filename}')
 | |
| 
 | |
|             # Check that file already exists but content differs
 | |
|             curr_file_content = self.gunzip_file(filename)
 | |
| 
 | |
|             if curr_file_content:
 | |
|                 # Compare file content with message content with MD5 checksums
 | |
|                 curr_file_md5 = md5(curr_file_content).hexdigest()
 | |
|                 new_file_md5 = md5(new_file_content).hexdigest()
 | |
| 
 | |
|                 if new_file_md5 != curr_file_md5:
 | |
|                     # MD5 are not equals, verify filename
 | |
|                     if filename.endswith('.gz'):
 | |
|                         filename = f'{filename[:-3]}_{new_file_md5}.gz'
 | |
|                     else:
 | |
|                         filename = f'{filename}_{new_file_md5}'
 | |
|                     self.logger.debug(f'new file to check: {filename}')
 | |
| 
 | |
|                     if os.path.isfile(filename):
 | |
|                         # Ignore duplicate
 | |
|                         self.logger.debug(f'ignore duplicated file {filename}')
 | |
|                         print(f'ignore duplicated file {filename}')
 | |
|                         filename = None
 | |
| 
 | |
|                 else:
 | |
|                     # Ignore duplicate checksum equals
 | |
|                     self.logger.debug(f'ignore duplicated file {filename}')
 | |
|                     print(f'ignore duplicated file {filename}')
 | |
|                     filename = None
 | |
| 
 | |
|             else:
 | |
|                 # File not unzipped
 | |
|                 filename = None
 | |
| 
 | |
|         return filename
 | |
| 
 | |
|     def gunzip_file(self, filename):
 | |
|         """
 | |
|         Unzip a file
 | |
|         publish stats if failure
 | |
|         """
 | |
|         curr_file_content = None
 | |
| 
 | |
|         try:
 | |
|             with gzip.open(filename, 'rb') as f:
 | |
|                 curr_file_content = f.read()
 | |
|         except EOFError:
 | |
|             self.logger.warning(f'Global; Incomplete file: {filename}')
 | |
|             print(f'Global; Incomplete file: {filename}')
 | |
|             # save daily stats
 | |
|             # self.r_stats.zincrby('module:Global:incomplete_file', 1, datetime.datetime.now().strftime('%Y%m%d'))
 | |
|             # Statistics.
 | |
|         except OSError:
 | |
|             self.logger.warning(f'Global; Not a gzipped file: {filename}')
 | |
|             print(f'Global; Not a gzipped file: {filename}')
 | |
|             # save daily stats
 | |
|             # self.r_stats.zincrby('module:Global:invalid_file', 1, datetime.datetime.now().strftime('%Y%m%d'))
 | |
| 
 | |
|         return curr_file_content
 | |
| 
 | |
|     # # TODO: add stats incomplete_file/Not a gzipped file
 | |
|     def gunzip_bytes_obj(self, filename, bytes_obj):
 | |
|         gunzipped_bytes_obj = None
 | |
|         try:
 | |
|             in_ = io.BytesIO()
 | |
|             in_.write(bytes_obj)
 | |
|             in_.seek(0)
 | |
| 
 | |
|             with gzip.GzipFile(fileobj=in_, mode='rb') as fo:
 | |
|                 gunzipped_bytes_obj = fo.read()
 | |
|         except Exception as e:
 | |
|             self.logger.warning(f'Global; Invalid Gzip file: {filename}, {e}')
 | |
|             print(f'Global; Invalid Gzip file: {filename}, {e}')
 | |
| 
 | |
|         return gunzipped_bytes_obj
 | |
| 
 | |
|     def rreplace(self, s, old, new, occurrence):
 | |
|         li = s.rsplit(old, occurrence)
 | |
|         return new.join(li)
 | |
| 
 | |
| 
 | |
| if __name__ == '__main__':
 | |
| 
 | |
|     module = Global()
 | |
|     module.run()
 |