from pype.api import config, Logger from avalon import io import threading import asyncio import concurrent.futures from concurrent.futures._base import CancelledError from enum import Enum from datetime import datetime from .providers import lib import os log = Logger().get_logger("SyncServer") # test object 5eeb25e411e06a16209ab78e class SyncStatus(Enum): DO_NOTHING = 0 DO_UPLOAD = 1 DO_DOWNLOAD = 2 class SyncServer(): """ WIP Synchronization server that is synching published files from local to any of implemented providers (like GDrive, S3 etc.) Runs in the background and checks all representations, looks for files that are marked to be in different location than 'studio' (temporary), checks if 'created_dt' field is present denoting successful sync with provider destination. ''' - example of synced file test_Cylinder_lookMain_v010.ma to GDrive "files" : [ { "path" : "{root}/Test/Assets/Cylinder/publish/look/lookMain/v010/ test_Cylinder_lookMain_v010.ma", "_id" : ObjectId("5eeb25e411e06a16209ab78f"), "hash" : "test_Cylinder_lookMain_v010,ma|1592468963,24|4822", "size" : NumberLong(4822), "sites" : { "studio" : { "created_dt" : ISODate("2020-05-22T08:05:44.000Z") }, "gdrive" : { "id" : ObjectId("5eeb25e411e06a16209ab78f"), "created_dt" : ISODate("2020-07-16T17:54:35.833Z") } } }, ''' Remote files "id" is real id that could be used in appropriate API. Local files have "id" too, for conformity, contains just file name. It is expected that multiple providers will be implemented in separate classes and registered in 'providers.py'. """ RETRY_CNT = 3 # number of attempts to sync specific file LOCAL_PROVIDER = 'studio' def __init__(self): self.qaction = None self.failed_icon = None self._is_running = False self.presets = None self.lock = threading.Lock() if not io.Session: io.install() io.Session['AVALON_PROJECT'] = 'Test' try: self.presets = config.get_presets()["services"]["sync_server"] except Exception: log.debug(( "There are not set presets for SyncServer." " No credentials provided, no synching possible" ).format(str(self.presets))) self.sync_server_thread = SynchServerThread(self) def get_sync_representations(self): """ Get representations that should be synched, these could be recognised by presence of document in 'files.sites', where key is a provider (GDrive, S3) and value is empty document or document with no value for 'created_dt' field. Currently returning all representations. TODO: filter out representations that shouldnt be synced :return: """ representations = io.find({ "type": "representation" }) return representations def check_status(self, file, provider_name): """ Check synchronization status for single 'file' of single 'representation' by single 'provider'. (Eg. check if 'scene.ma' of lookdev.v10 should be synched to GDrive :param file: of file from representation in Mongo :param provider_name: - gdrive, gdc etc. :return: - one of SyncStatus """ sites = file.get("sites") or {} if isinstance(sites, list): # temporary, old format of 'sites' return SyncStatus.DO_NOTHING provider_rec = sites.get(provider_name) or {} if provider_rec: created_dt = provider_rec.get("created_dt") if not created_dt: tries = self._get_tries_count(file, provider_name) # file will be skipped if unsuccessfully tried over threshold # error metadata needs to be purged manually in DB to reset if tries < self.RETRY_CNT: return SyncStatus.DO_UPLOAD else: local_rec = sites.get(lib.Providers.LOCAL.value) or {} if not local_rec or not local_rec.get("created_dt"): tries = self._get_tries_count(file, self.LOCAL_PROVIDER) # file will be skipped if unsuccessfully tried over # threshold error metadata needs to be purged manually # in DB to reset if tries < self.RETRY_CNT: return SyncStatus.DO_DOWNLOAD return SyncStatus.DO_NOTHING async def upload(self, file, representation, provider_name, tree=None): """ Upload single 'file' of a 'representation' to 'provider'. Source url is taken from 'file' portion, where {root} placeholder is replaced by 'representation.Context.root' Provider could be one of implemented in provider.py. Updates MongoDB, fills in id of file from provider (ie. file_id from GDrive), 'created_dt' - time of upload :param file: of file from representation in Mongo :param representation: of representation :param provider_name: - gdrive, gdc etc. :param tree: - injected memory structure for performance :return: """ # create ids sequentially, upload file in parallel later with self.lock: handler = lib.factory.get_provider(provider_name, tree) remote_file = self._get_remote_file_path(file, handler.get_root_name()) local_root = representation.get("context", {}).get("root") local_file = self._get_local_file_path(file, local_root) target_folder = os.path.dirname(remote_file) folder_id = handler.create_folder(target_folder) if not folder_id: raise NotADirectoryError("Folder {} wasn't created" .format(target_folder)) loop = asyncio.get_running_loop() file_id = await loop.run_in_executor(None, handler.upload_file, local_file, remote_file, True) return file_id async def download(self, file, representation, provider_name, tree=None): """ Downloads file to local folder denoted in representation.Context. :param file: - info about processed file :param representation: - repr that 'file' belongs to :param provider_name: - 'gdrive' etc :param tree: - injected memory structure for performance :return: - 'name' of local file """ with self.lock: handler = lib.factory.get_provider(provider_name, tree) remote_file = self._get_remote_file_path(file, handler.get_root_name()) local_root = representation.get("context", {}).get("root") local_file = self._get_local_file_path(file, local_root) local_folder = os.path.dirname(local_file) os.makedirs(local_folder, exist_ok=True) loop = asyncio.get_running_loop() file_id = await loop.run_in_executor(None, handler.download_file, remote_file, local_file, False) return file_id def update_db(self, new_file_id, file, representation, provider_name, error=None): """ Update 'provider' portion of records in DB with success (file_id) or error (exception) :param new_file_id: :param file: - info about processed file (pulled from DB) :param representation: - parent repr of file (from DB) :param provider_name: - label ('gdrive', 'S3') :param error: - exception message :return: None """ representation_id = representation.get("_id") file_id = file.get("_id") query = { "_id": representation_id, "files._id": file_id } update = {} if new_file_id: update["$set"] = self._get_success_dict(provider_name, new_file_id) # reset previous errors if any update["$unset"] = self._get_error_dict(provider_name, "", "") else: tries = self._get_tries_count(file, provider_name) tries += 1 update["$set"] = self._get_error_dict(provider_name, error, tries) # it actually modifies single _id, but io.update_one not implemented io.update_many( query, update ) status = 'failed' if new_file_id: status = 'succeeded with id {}'.format(new_file_id) source_file = file.get("path", "") log.debug("File {} process {} {}".format(status, source_file, status)) def tray_start(self): self.sync_server_thread.start() def tray_exit(self): self.stop() @property def is_running(self): return self.sync_server_thread.is_running def stop(self): if not self.is_running: return try: log.debug("Stopping synch server server") self.sync_server_thread.is_running = False self.sync_server_thread.stop() except Exception: log.warning( "Error has happened during Killing synchserver server", exc_info=True ) def thread_stopped(self): self._is_running = False def reset_provider_for_file(self, file_id, provider): """ Reset information about synchronization for particular 'file_id' and provider. Useful for testing or forcing file to be reuploaded. :param file_id: file id in representation :param provider: - 'gdrive', 'S3' etc :return: None """ query = { "files._id": file_id } update = { "$unset": {"files.$.sites.{}".format(provider): ""} } # it actually modifies single _id, but io.update_one not implemented io.update_many( query, update ) def _get_success_dict(self, provider, new_file_id): """ Provide success metadata ("id", "created_dt") to be stored in Db. :param provider: used as part of path in DB (files.sites.gdrive) :param new_file_id: id of created file :return: """ val = {"files.$.sites.{}.id".format(provider): new_file_id, "files.$.sites.{}.created_dt".format(provider): datetime.utcnow()} return val def _get_error_dict(self, provider, error="", tries=""): """ Provide error metadata to be stored in Db. Used for set (error and tries provided) or unset mode. :param provider: used as part of path in DB (files.sites.gdrive) :param error: message :param tries: how many times failed :return: """ val = {"files.$.sites.{}.last_failed_dt".format(provider): datetime.utcnow(), "files.$.sites.{}.error".format(provider): error, "files.$.sites.{}.tries".format(provider): tries} return val def _get_tries_count(self, file, provider): """ Get number of failed attempts to synch :param file: - info about specific file :param provider: - gdrive, S3 etc :return: - number of failed attempts """ return file.get("sites", {}).get(provider, {}).get("tries", 0) def _get_local_file_path(self, file, local_root): """ Auxiliary function for replacing rootless path with real path :param file: url to file with {root} :param local_root: value of {root} for local projects :return: - absolute path on local system """ if not local_root: raise ValueError("Unknown local root for file {}") return file.get("path", "").replace('{root}', local_root) def _get_remote_file_path(self, file, root_name): """ Auxiliary function for replacing rootless path with real path :param file: url to file with {root} :param root_name: value of {root} for remote location :return: - absolute path on remote location """ target_root = '/{}'.format(root_name) return file.get("path", "").replace('{root}', target_root) class SynchServerThread(threading.Thread): """ Separate thread running synchronization server with asyncio loop. Stopped when tray is closed. """ def __init__(self, module): super(SynchServerThread, self).__init__() self.module = module self.loop = None self.is_running = False self.executor = concurrent.futures.ThreadPoolExecutor(max_workers=3) def run(self): self.is_running = True try: log.info("Starting synchserver server") self.loop = asyncio.new_event_loop() # create new loop for thread asyncio.set_event_loop(self.loop) self.loop.set_default_executor(self.executor) asyncio.ensure_future(self.check_shutdown(), loop=self.loop) asyncio.ensure_future(self.sync_loop(), loop=self.loop) self.loop.run_forever() except Exception: log.warning( "Synch Server service has failed", exc_info=True ) finally: self.loop.close() # optional async def sync_loop(self): try: while self.is_running: import time from datetime import datetime start_time = time.time() sync_representations = self.module.get_sync_representations() local_label = lib.Providers.LOCAL.value task_files_to_process = [] files_processed_info = [] # process only unique file paths in one batch # multiple representation could have same file path (textures), # upload process can find already uploaded file and reuse same # id processed_file_path = set() for provider in lib.factory.providers.keys(): # first call to get_provider could be expensive, its # building folder tree structure in memory handler = lib.factory.get_provider(provider) tree = handler.get_tree() limit = lib.factory.get_provider_batch_limit(provider) for sync in sync_representations: if limit <= 0: continue files = sync.get("files") or {} if files: for file in files: # skip already processed files file_path = file.get('path', '') if file_path in processed_file_path: continue status = self.module.check_status(file, provider) if status == SyncStatus.DO_UPLOAD: limit -= 1 task = asyncio.create_task( self.module.upload( file, sync, provider, tree)) task_files_to_process.append(task) # store info for exception handling files_processed_info.append((file, sync, provider)) processed_file_path.add(file_path) if status == SyncStatus.DO_DOWNLOAD: limit -= 1 task = asyncio.create_task( self.module.download (file, sync, provider)) task_files_to_process.append(task) files_processed_info.append((file, sync, local_label)) processed_file_path.add(file_path) log.debug("gather tasks len {}". format(len(task_files_to_process))) files_created = await asyncio.gather(*task_files_to_process, return_exceptions=True) for file_id, info in zip(files_created, files_processed_info): file, representation, provider = info error = None if isinstance(file_id, BaseException): error = str(file_id) file_id = None self.module.update_db(file_id, file, representation, provider, error) duration = time.time() - start_time log.debug("One loop took {}".format(duration)) await asyncio.sleep(60) except ConnectionResetError: log.warning("ConnectionResetError in sync loop, trying next loop", exc_info=True) except CancelledError: # just stopping server pass except Exception: self.stop() log.warning("Unhandled exception in sync loop, stopping server", exc_info=True) def stop(self): """Sets is_running flag to false, 'check_shutdown' shuts server down""" self.is_running = False async def check_shutdown(self): """ Future that is running and checks if server should be running periodically. """ while self.is_running: await asyncio.sleep(0.5) tasks = [task for task in asyncio.all_tasks() if task is not asyncio.current_task()] list(map(lambda task: task.cancel(), tasks)) # cancel all the tasks results = await asyncio.gather(*tasks, return_exceptions=True) log.debug(f'Finished awaiting cancelled tasks, results: {results}...') await self.loop.shutdown_asyncgens() # to really make sure everything else has time to stop self.executor.shutdown(wait=True) await asyncio.sleep(0.07) self.loop.stop()