X-Git-Url: https://git.madduck.net/etc/taskwarrior.git/blobdiff_plain/0531935967fd784f762215d35bcae90194e0ff90..f17eb34f55544b849b34e9f73013e0224d01408f:/tasklib/backends.py diff --git a/tasklib/backends.py b/tasklib/backends.py index f817d4f..b15e8bd 100644 --- a/tasklib/backends.py +++ b/tasklib/backends.py @@ -1,22 +1,29 @@ import abc +import copy +import datetime +import json +import logging import os import re +import six import subprocess +from .task import Task, TaskQuerySet, ReadOnlyDictView +from .filters import TaskWarriorFilter +from .serializing import local_zone -VERSION_2_1_0 = six.u('2.1.0') -VERSION_2_2_0 = six.u('2.2.0') -VERSION_2_3_0 = six.u('2.3.0') -VERSION_2_4_0 = six.u('2.4.0') -VERSION_2_4_1 = six.u('2.4.1') -VERSION_2_4_2 = six.u('2.4.2') -VERSION_2_4_3 = six.u('2.4.3') -VERSION_2_4_4 = six.u('2.4.4') -VERSION_2_4_5 = six.u('2.4.5') +DATE_FORMAT_CALC = '%Y-%m-%dT%H:%M:%S' + +logger = logging.getLogger(__name__) class Backend(object): + @abc.abstractproperty + def filter_class(self): + """Returns the TaskFilter class used by this backend""" + pass + @abc.abstractmethod def filter_tasks(self, filter_obj): """Returns a list of Task objects matching the given filter""" @@ -38,17 +45,55 @@ class Backend(object): def stop_task(self, task): pass + @abc.abstractmethod + def complete_task(self, task): + pass + + @abc.abstractmethod + def refresh_task(self, task, after_save=False): + """ + Refreshes the given task. Returns new data dict with serialized + attributes. + """ + pass + + @abc.abstractmethod + def annotate_task(self, task, annotation): + pass + + @abc.abstractmethod + def denotate_task(self, task, annotation): + pass + @abc.abstractmethod def sync(self): """Syncs the backend database with the taskd server""" pass + def convert_datetime_string(self, value): + """ + Converts TW syntax datetime string to a localized datetime + object. This method is not mandatory. + """ + raise NotImplemented + class TaskWarriorException(Exception): pass -class TaskWarrior(object): +class TaskWarrior(Backend): + + VERSION_2_1_0 = six.u('2.1.0') + VERSION_2_2_0 = six.u('2.2.0') + VERSION_2_3_0 = six.u('2.3.0') + VERSION_2_4_0 = six.u('2.4.0') + VERSION_2_4_1 = six.u('2.4.1') + VERSION_2_4_2 = six.u('2.4.2') + VERSION_2_4_3 = six.u('2.4.3') + VERSION_2_4_4 = six.u('2.4.4') + VERSION_2_4_5 = six.u('2.4.5') + def __init__(self, data_location=None, create=True, taskrc_location='~/.taskrc'): self.taskrc_location = os.path.expanduser(taskrc_location) @@ -57,8 +102,9 @@ class TaskWarrior(object): if not os.path.exists(self.taskrc_location): self.taskrc_location = '/' + self._config = None self.version = self._get_version() - self.config = { + self.overrides = { 'confirmation': 'no', 'dependency.confirmation': 'no', # See TW-1483 or taskrc man page 'recurrence.confirmation': 'no', # Necessary for modifying R tasks @@ -68,7 +114,7 @@ class TaskWarrior(object): # 2.4.3 onwards supports 0 as infite bulk, otherwise set just # arbitrary big number which is likely to be large enough - 'bulk': 0 if self.version >= VERSION_2_4_3 else 100000, + 'bulk': 0 if self.version >= self.VERSION_2_4_3 else 100000, } # Set data.location override if passed via kwarg @@ -76,32 +122,124 @@ class TaskWarrior(object): data_location = os.path.expanduser(data_location) if create and not os.path.exists(data_location): os.makedirs(data_location) - self.config['data.location'] = data_location + self.overrides['data.location'] = data_location self.tasks = TaskQuerySet(self) def _get_command_args(self, args, config_override=None): command_args = ['task', 'rc:{0}'.format(self.taskrc_location)] - config = self.config.copy() - config.update(config_override or dict()) - for item in config.items(): + overrides = self.overrides.copy() + overrides.update(config_override or dict()) + for item in overrides.items(): command_args.append('rc.{0}={1}'.format(*item)) command_args.extend(map(six.text_type, args)) return command_args def _get_version(self): p = subprocess.Popen( - ['task', '--version'], - stdout=subprocess.PIPE, - stderr=subprocess.PIPE) + ['task', '--version'], + stdout=subprocess.PIPE, + stderr=subprocess.PIPE) stdout, stderr = [x.decode('utf-8') for x in p.communicate()] return stdout.strip('\n') - def get_config(self): + def _get_modified_task_fields_as_args(self, task): + args = [] + + def add_field(field): + # Add the output of format_field method to args list (defaults to + # field:value) + serialized_value = task._serialize(field, task._data[field]) + + # Empty values should not be enclosed in quotation marks, see + # TW-1510 + if serialized_value is '': + escaped_serialized_value = '' + else: + escaped_serialized_value = six.u("'{0}'").format( + serialized_value) + + format_default = lambda task: six.u("{0}:{1}").format( + field, escaped_serialized_value) + + format_func = getattr(self, 'format_{0}'.format(field), + format_default) + + args.append(format_func(task)) + + # If we're modifying saved task, simply pass on all modified fields + if task.saved: + for field in task._modified_fields: + add_field(field) + # For new tasks, pass all fields that make sense + else: + for field in task._data.keys(): + if field in task.read_only_fields: + continue + add_field(field) + + return args + + def format_depends(self, task): + # We need to generate added and removed dependencies list, + # since Taskwarrior does not accept redefining dependencies. + + # This cannot be part of serialize_depends, since we need + # to keep a list of all depedencies in the _data dictionary, + # not just currently added/removed ones + + old_dependencies = task._original_data.get('depends', set()) + + added = task['depends'] - old_dependencies + removed = old_dependencies - task['depends'] + + # Removed dependencies need to be prefixed with '-' + return 'depends:' + ','.join( + [t['uuid'] for t in added] + + ['-' + t['uuid'] for t in removed] + ) + + def format_description(self, task): + # Task version older than 2.4.0 ignores first word of the + # task description if description: prefix is used + if self.version < self.VERSION_2_4_0: + return task._data['description'] + else: + return six.u("description:'{0}'").format(task._data['description'] or '') + + def convert_datetime_string(self, value): + + if self.version >= self.VERSION_2_4_0: + # For strings, use 'task calc' to evaluate the string to datetime + # available since TW 2.4.0 + args = value.split() + result = self.execute_command(['calc'] + args) + naive = datetime.datetime.strptime(result[0], DATE_FORMAT_CALC) + localized = local_zone.localize(naive) + else: + raise ValueError("Provided value could not be converted to " + "datetime, its type is not supported: {}" + .format(type(value))) + + return localized + + @property + def filter_class(self): + return TaskWarriorFilter + + # Public interface + + @property + def config(self): + # First, check if memoized information is available + if self._config: + return self._config + + # If not, fetch the config using the 'show' command raw_output = self.execute_command( - ['show'], - config_override={'verbose': 'nothing'} - ) + ['show'], + config_override={'verbose': 'nothing'} + ) config = dict() config_regex = re.compile(r'^(?P[^\s]+)\s+(?P[^\s].+$)') @@ -111,7 +249,10 @@ class TaskWarrior(object): if match: config[match.group('key')] = match.group('value').strip() - return config + # Memoize the config dict + self._config = ReadOnlyDictView(config) + + return self._config def execute_command(self, args, config_override=None, allow_failure=True, return_all=False): @@ -141,12 +282,23 @@ class TaskWarrior(object): # of recurrent tasks. # Only necessary for TW up to 2.4.1, fixed in 2.4.2. - if self.version < VERSION_2_4_2: + if self.version < self.VERSION_2_4_2: self.execute_command(['next'], allow_failure=False) + def merge_with(self, path, push=False): + path = path.rstrip('/') + '/' + self.execute_command(['merge', path], config_override={ + 'merge.autopush': 'yes' if push else 'no', + }) + + def undo(self): + self.execute_command(['undo']) + + # Backend interface implementation + def filter_tasks(self, filter_obj): self.enforce_recurrence() - args = ['export', '--'] + filter_obj.get_filter_params() + args = ['export'] + filter_obj.get_filter_params() tasks = [] for line in self.execute_command(args): if line: @@ -159,11 +311,99 @@ class TaskWarrior(object): raise TaskWarriorException('Invalid JSON: %s' % data) return tasks - def merge_with(self, path, push=False): - path = path.rstrip('/') + '/' - self.execute_command(['merge', path], config_override={ - 'merge.autopush': 'yes' if push else 'no', - }) + def save_task(self, task): + """Save a task into TaskWarrior database using add/modify call""" - def undo(self): - self.execute_command(['undo']) + args = [task['uuid'], 'modify'] if task.saved else ['add'] + args.extend(self._get_modified_task_fields_as_args(task)) + output = self.execute_command(args) + + # Parse out the new ID, if the task is being added for the first time + if not task.saved: + id_lines = [l for l in output if l.startswith('Created task ')] + + # Complain loudly if it seems that more tasks were created + # Should not happen + if len(id_lines) != 1 or len(id_lines[0].split(' ')) != 3: + raise TaskWarriorException("Unexpected output when creating " + "task: %s" % '\n'.join(id_lines)) + + # Circumvent the ID storage, since ID is considered read-only + identifier = id_lines[0].split(' ')[2].rstrip('.') + + # Identifier can be either ID or UUID for completed tasks + try: + task._data['id'] = int(identifier) + except ValueError: + task._data['uuid'] = identifier + + # Refreshing is very important here, as not only modification time + # is updated, but arbitrary attribute may have changed due hooks + # altering the data before saving + task.refresh(after_save=True) + + def delete_task(self, task): + self.execute_command([task['uuid'], 'delete']) + + def start_task(self, task): + self.execute_command([task['uuid'], 'start']) + + def stop_task(self, task): + self.execute_command([task['uuid'], 'stop']) + + def complete_task(self, task): + # Older versions of TW do not stop active task at completion + if self.version < self.VERSION_2_4_0 and task.active: + task.stop() + + self.execute_command([task['uuid'], 'done']) + + def annotate_task(self, task, annotation): + args = [task['uuid'], 'annotate', annotation] + self.execute_command(args) + + def denotate_task(self, task, annotation): + args = [task['uuid'], 'denotate', annotation] + self.execute_command(args) + + def refresh_task(self, task, after_save=False): + # We need to use ID as backup for uuid here for the refreshes + # of newly saved tasks. Any other place in the code is fine + # with using UUID only. + args = [task['uuid'] or task['id'], 'export'] + output = self.execute_command(args) + + def valid(output): + return len(output) == 1 and output[0].startswith('{') + + # For older TW versions attempt to uniquely locate the task + # using the data we have if it has been just saved. + # This can happen when adding a completed task on older TW versions. + if (not valid(output) and self.version < self.VERSION_2_4_5 + and after_save): + + # Make a copy, removing ID and UUID. It's most likely invalid + # (ID 0) if it failed to match a unique task. + data = copy.deepcopy(task._data) + data.pop('id', None) + data.pop('uuid', None) + + taskfilter = self.filter_class(self) + for key, value in data.items(): + taskfilter.add_filter_param(key, value) + + output = self.execute_command(['export'] + + taskfilter.get_filter_params()) + + # If more than 1 task has been matched still, raise an exception + if not valid(output): + raise TaskWarriorException( + "Unique identifiers {0} with description: {1} matches " + "multiple tasks: {2}".format( + task['uuid'] or task['id'], task['description'], output) + ) + + return json.loads(output[0]) + + def sync(self): + self.execute_command(['sync'])