X-Git-Url: https://git.madduck.net/etc/taskwarrior.git/blobdiff_plain/b28ec3e9b03f1595eaf4bb82ef1f01c3c4d1c2b6..8724d12824dba6525d3f52d861c8cf3c4583f05e:/tasklib/task.py diff --git a/tasklib/task.py b/tasklib/task.py index 80f08e9..9246e1f 100644 --- a/tasklib/task.py +++ b/tasklib/task.py @@ -1,17 +1,16 @@ import copy import datetime import json +import logging import os import subprocess -import tempfile -import uuid - DATE_FORMAT = '%Y%m%dT%H%M%SZ' - REPR_OUTPUT_SIZE = 10 - PENDING = 'pending' +COMPLETED = 'completed' + +logger = logging.getLogger(__name__) class TaskWarriorException(Exception): @@ -26,23 +25,21 @@ class Task(object): def __init__(self, warrior, data={}): self.warrior = warrior self._data = data - - def __getitem__(self, key): - return self._get_field(key) - - def __setitem__(self, key, val): - self._data[key] = val + self._modified_fields = set() def __unicode__(self): - return self._data.get('description') + return self['description'] - def _get_field(self, key): - hydrate_func = getattr(self, 'deserialize_{0}'.format(key), lambda x:x) + def __getitem__(self, key): + hydrate_func = getattr(self, 'deserialize_{0}'.format(key), + lambda x: x) return hydrate_func(self._data.get(key)) - def _set_field(self, key, value): - dehydrate_func = getattr(self, 'serialize_{0}'.format(key), lambda x:x) + def __setitem__(self, key, value): + dehydrate_func = getattr(self, 'serialize_{0}'.format(key), + lambda x: x) self._data[key] = dehydrate_func(value) + self._modified_fields.add(key) def serialize_due(self, date): return date.strftime(DATE_FORMAT) @@ -65,21 +62,33 @@ class Task(object): ann['entry'], DATE_FORMAT) return ann_list - def regenerate_uuid(self): - self['uuid'] = str(uuid.uuid4()) + def deserialize_tags(self, tags): + if isinstance(tags, basestring): + return tags.split(',') if tags else [] + return tags + + def serialize_tags(self, tags): + return ','.join(tags) if tags else '' def delete(self): - self.warrior.delete_task(self['uuid']) + self.warrior.execute_command([self['id'], 'delete'], config_override={ + 'confirmation': 'no', + }) def done(self): - self.warrior.complete_task(self['uuid']) + self.warrior.execute_command([self['id'], 'done']) - def save(self, delete_first=True): - if self['uuid'] and delete_first: - self.delete() - if not self['uuid'] or delete_first: - self.regenerate_uuid() - self.warrior.import_tasks([self._data]) + def save(self): + args = [self['id'], 'modify'] if self['id'] else ['add'] + args.extend(self._get_modified_fields_as_args()) + self.warrior.execute_command(args) + self._modified_fields.clear() + + def _get_modified_fields_as_args(self): + args = [] + for field in self._modified_fields: + args.append('{}:{}'.format(field, self._data[field])) + return args __repr__ = __unicode__ @@ -100,7 +109,7 @@ class TaskFilter(object): self.filter_params.append('{0}:{1}'.format(key, value)) def get_filter_params(self): - return ['({})'.format(f) for f in self.filter_params if f] + return [f for f in self.filter_params if f] def clone(self): c = self.__class__() @@ -123,8 +132,8 @@ class TaskQuerySet(object): Deep copy of a QuerySet doesn't populate the cache """ obj = self.__class__() - for k,v in self.__dict__.items(): - if k in ('_iter','_result_cache'): + for k, v in self.__dict__.items(): + if k in ('_iter', '_result_cache'): obj.__dict__[k] = None else: obj.__dict__[k] = copy.deepcopy(v, memo) @@ -175,7 +184,7 @@ class TaskQuerySet(object): """ Fetch the tasks which match the current filters. """ - return self.warrior._execute_filter(self.filter_obj) + return self.warrior.filter_tasks(self.filter_obj) def all(self): """ @@ -186,6 +195,9 @@ class TaskQuerySet(object): def pending(self): return self.filter(status=PENDING) + def completed(self): + return self.filter(status=COMPLETED) + def filter(self, *args, **kwargs): """ Returns a new TaskQuerySet with the given filters added. @@ -216,58 +228,58 @@ class TaskQuerySet(object): class TaskWarrior(object): - DEFAULT_FILTERS = { - 'status': 'pending', - } - def __init__(self, data_location='~/.task', create=True): - if not os.path.exists(data_location): + data_location = os.path.expanduser(data_location) + if create and not os.path.exists(data_location): os.makedirs(data_location) self.config = { 'data.location': os.path.expanduser(data_location), } self.tasks = TaskQuerySet(self) - def _get_command_args(self, args): + def _get_command_args(self, args, config_override={}): command_args = ['task', 'rc:/'] - for item in self.config.items(): + config = self.config.copy() + config.update(config_override) + for item in config.items(): command_args.append('rc.{0}={1}'.format(*item)) - command_args.extend(args) + command_args.extend(map(str, args)) return command_args - def _execute_command(self, args): - p = subprocess.Popen(self._get_command_args(args), - stdout=subprocess.PIPE, stderr=subprocess.PIPE) - stdout, stderr = p.communicate() + def execute_command(self, args, config_override={}): + command_args = self._get_command_args( + args, config_override=config_override) + logger.debug(' '.join(command_args)) + p = subprocess.Popen(command_args, stdout=subprocess.PIPE, + stderr=subprocess.PIPE) + stdout, stderr = [x.decode() for x in p.communicate()] if p.returncode: - raise TaskWarriorException(stderr.strip()) + if stderr.strip(): + error_msg = stderr.strip().splitlines()[-1] + else: + error_msg = stdout.strip() + raise TaskWarriorException(error_msg) return stdout.strip().split('\n') - def _execute_filter(self, filter_obj): - args = filter_obj.get_filter_params() + ['export'] + def filter_tasks(self, filter_obj): + args = ['export', '--'] + filter_obj.get_filter_params() tasks = [] - for line in self._execute_command(args): + for line in self.execute_command(args): if line: - tasks.append(Task(self, json.loads(line.strip(',')))) + data = line.strip(',') + try: + tasks.append(Task(self, json.loads(data))) + except ValueError: + raise TaskWarriorException('Invalid JSON: %s' % data) return tasks - def add_task(self, description, project=None): - args = ['add', description] - if project is not None: - args.append('project:{0}'.format(project)) - self._execute_command(args) - - def delete_task(self, task_id): - args = [task_id, 'rc.confirmation:no', 'delete'] - self._execute_command(args) - - def complete_task(self, task_id): - args = [task_id, 'done'] - self._execute_command(args) - - def import_tasks(self, tasks): - fd, path = tempfile.mkstemp() - with open(path, 'w') as f: - f.write(json.dumps(tasks)) - args = ['import', path] - self._execute_command(args) + def merge_with(self, path, push=False): + path = path.rstrip('/') + '/' + self.execute_command(['merge', path], config_override={ + 'merge.autopush': 'yes' if push else 'no', + }) + + def undo(self): + self.execute_command(['undo'], config_override={ + 'confirmation': 'no', + })