from __future__ import division, print_function, unicode_literals import threading import abc import uuid try: import queue except ImportError: import Queue as queue from datetime import datetime, timedelta from collections import namedtuple from cps import logger log = logger.create() # task 'status' consts STAT_WAITING = 0 STAT_FAIL = 1 STAT_STARTED = 2 STAT_FINISH_SUCCESS = 3 # Only retain this many tasks in dequeued list TASK_CLEANUP_TRIGGER = 20 QueuedTask = namedtuple('QueuedTask', 'num, user, added, task') def _get_main_thread(): for t in threading.enumerate(): if t.__class__.__name__ == '_MainThread': return t raise Exception("main thread not found?!") class ImprovedQueue(queue.Queue): def to_list(self): """ Returns a copy of all items in the queue without removing them. """ with self.mutex: return list(self.queue) #Class for all worker tasks in the background class WorkerThread(threading.Thread): _instance = None @classmethod def getInstance(cls): if cls._instance is None: cls._instance = WorkerThread() return cls._instance def __init__(self): threading.Thread.__init__(self) self.dequeued = list() self.doLock = threading.Lock() self.queue = ImprovedQueue() self.num = 0 self.start() @classmethod def add(cls, user, task): ins = cls.getInstance() ins.num += 1 ins.queue.put(QueuedTask( num=ins.num, user=user, added=datetime.now(), task=task, )) @property def tasks(self): with self.doLock: tasks = self.queue.to_list() + self.dequeued return sorted(tasks, key=lambda x: x.num) def cleanup_tasks(self): with self.doLock: dead = [] alive = [] for x in self.dequeued: (dead if x.task.dead else alive).append(x) # if the ones that we need to keep are within the trigger, do nothing else delta = len(self.dequeued) - len(dead) if delta > TASK_CLEANUP_TRIGGER: ret = alive else: # otherwise, lop off the oldest dead tasks until we hit the target trigger ret = sorted(dead, key=lambda x: x.task.end_time)[-TASK_CLEANUP_TRIGGER:] + alive self.dequeued = sorted(ret, key=lambda x: x.num) # Main thread loop starting the different tasks def run(self): main_thread = _get_main_thread() while main_thread.is_alive(): # this blocks until something is available item = self.queue.get() with self.doLock: # add to list so that in-progress tasks show up self.dequeued.append(item) # once we hit our trigger, start cleaning up dead tasks if len(self.dequeued) > TASK_CLEANUP_TRIGGER: self.cleanup_tasks() # sometimes tasks (like Upload) don't actually have work to do and are created as already finished if item.task.stat is STAT_WAITING: # CalibreTask.start() should wrap all exceptions in it's own error handling item.task.start(self) self.queue.task_done() class CalibreTask: __metaclass__ = abc.ABCMeta def __init__(self, message): self._progress = 0 self.stat = STAT_WAITING self.error = None self.start_time = None self.end_time = None self.message = message self.id = uuid.uuid4() @abc.abstractmethod def run(self, worker_thread): """Provides the caller some human-readable name for this class""" raise NotImplementedError @abc.abstractmethod def name(self): """Provides the caller some human-readable name for this class""" raise NotImplementedError def start(self, *args): self.start_time = datetime.now() self.stat = STAT_STARTED # catch any unhandled exceptions in a task and automatically fail it try: self.run(*args) except Exception as e: self._handleError(str(e)) log.exception(e) self.end_time = datetime.now() @property def stat(self): return self._stat @stat.setter def stat(self, x): self._stat = x @property def progress(self): return self._progress @progress.setter def progress(self, x): if not 0 <= x <= 1: raise ValueError("Task progress should within [0, 1] range") self._progress = x @property def error(self): return self._error @error.setter def error(self, x): self._error = x @property def runtime(self): return (self.end_time or datetime.now()) - self.start_time @property def dead(self): """Determines whether or not this task can be garbage collected We have a separate dictating this because there may be certain tasks that want to override this """ # By default, we're good to clean a task if it's "Done" return self.stat in (STAT_FINISH_SUCCESS, STAT_FAIL) @progress.setter def progress(self, x): # todo: throw error if outside of [0,1] self._progress = x def _handleError(self, error_message): log.error(error_message) self.stat = STAT_FAIL self.progress = 1 self.error = error_message def _handleSuccess(self): self.stat = STAT_FINISH_SUCCESS self.progress = 1