import logging
import os
import six
+import sys
import subprocess
DATE_FORMAT = '%Y%m%dT%H%M%SZ'
return ','.join(tags) if tags else ''
def deserialize_tags(self, tags):
- if isinstance(tags, basestring):
+ if isinstance(tags, six.string_types):
return tags.split(',') if tags else []
- return tags
+ return tags or []
def serialize_depends(self, cur_dependencies):
# Return the list of uuids
read_only_fields = []
def _load_data(self, data):
- self._data = data
+ self._data = dict((key, self._deserialize(key, value))
+ for key, value in data.items())
# We need to use a copy for original data, so that changes
- # are not propagated. Shallow copy is alright, since data dict uses only
- # primitive data types
- self._original_data = data.copy()
+ # are not propagated.
+ self._original_data = copy.deepcopy(self._data)
def _update_data(self, data, update_original=False):
"""
updates should already be serialized. If update_original is True, the
original_data dict is updated as well.
"""
-
- self._data.update(data)
+ self._data.update(dict((key, self._deserialize(key, value))
+ for key, value in data.items()))
if update_original:
- self._original_data.update(data)
+ self._original_data = copy.deepcopy(self._data)
+
def __getitem__(self, key):
# This is a workaround to make TaskResource non-iterable
except ValueError:
pass
- return self._deserialize(key, self._data.get(key))
+ if key not in self._data:
+ self._data[key] = self._deserialize(key, None)
+
+ return self._data.get(key)
def __setitem__(self, key, value):
if key in self.read_only_fields:
raise RuntimeError('Field \'%s\' is read-only' % key)
- self._data[key] = self._serialize(key, value)
+ self._data[key] = value
def __str__(self):
s = six.text_type(self.__unicode__())
def __unicode__(self):
return self['description']
+ def __eq__(self, other):
+ # consider 2 annotations equal if they belong to the same task, and
+ # their data dics are the same
+ return self.task == other.task and self._data == other._data
+
__repr__ = __unicode__
"""
pass
+ @classmethod
+ def from_input(cls, input_file=sys.stdin, modify=False):
+ """
+ Creates a Task object, directly from the stdin, by reading one line.
+ If modify=True, two lines are used, first line interpreted as the
+ original state of the Task object, and second line as its new,
+ modified value. This is consistent with the TaskWarrior's hook
+ system.
+
+ Object created by this method should not be saved, deleted
+ or refreshed, as t could create a infinite loop. For this
+ reason, TaskWarrior instance is set to None.
+
+ Input_file argument can be used to specify the input file,
+ but defaults to sys.stdin.
+ """
+
+ # TaskWarrior instance is set to None
+ task = cls(None)
+
+ # Load the data from the input
+ task._load_data(json.loads(input_file.readline().strip()))
+
+ # If this is a on-modify event, we are provided with additional
+ # line of input, which provides updated data
+ if modify:
+ task._update_data(json.loads(input_file.readline().strip()))
+
+ return task
+
def __init__(self, warrior, **kwargs):
self.warrior = warrior
if self._data.get(key) != self._original_data.get(key):
yield key
+ @property
+ def _is_modified(self):
+ return bool(list(self._modified_fields))
+
@property
def completed(self):
return self['status'] == six.text_type('completed')
# to keep a list of all depedencies in the _data dictionary,
# not just currently added/removed ones
- old_dependencies_raw = self._original_data.get('depends','')
- old_dependencies = self.deserialize_depends(old_dependencies_raw)
+ old_dependencies = self._original_data.get('depends', set())
added = self['depends'] - old_dependencies
removed = old_dependencies - self['depends']
self.refresh(only_fields=['status'])
def save(self):
+ if self.saved and not self._is_modified:
+ return
+
args = [self['uuid'], 'modify'] if self.saved else ['add']
args.extend(self._get_modified_fields_as_args())
output = self.warrior.execute_command(args)
# Circumvent the ID storage, since ID is considered read-only
self._data['id'] = int(id_lines[0].split(' ')[2].rstrip('.'))
+ # Refreshing is very important here, as not only modification time
+ # is updated, but arbitrary attribute may have changed due hooks
+ # altering the data before saving
self.refresh()
def add_annotation(self, annotation):
def add_field(field):
# Add the output of format_field method to args list (defaults to
# field:value)
- format_default = lambda k: "{0}:'{1}'".format(k, self._data[k] or '')
+ serialized_value = self._serialize(field, self._data[field]) or ''
+ format_default = lambda: "{0}:{1}".format(
+ field,
+ "'{0}'".format(serialized_value) if serialized_value else ''
+ )
format_func = getattr(self, 'format_{0}'.format(field),
- lambda: format_default(field))
+ format_default)
args.append(format_func())
# If we're modifying saved task, simply pass on all modified fields
else:
self._load_data(new_data)
+ def export_data(self):
+ """
+ Exports current data contained in the Task as JSON
+ """
+
+ # We need to remove spaces for TW-1504, use custom separators
+ return json.dumps(self._data, separators=(',',':'))
class TaskFilter(SerializingObject):
"""
if key == 'uuid':
self.filter_params.insert(0, value)
else:
- self.filter_params.append("{0}:'{1}'".format(key, value))
+ # Surround value with aphostrophes unless it's a empty string
+ value = "'%s'" % value if value else ''
+
+ # We enforce equality match by using 'is' (or 'none') modifier
+ # Without using this syntax, filter fails due to TW-1479
+ modifier = '.is' if value else '.none'
+ key = key + modifier if '.' not in key else key
+
+ self.filter_params.append("{0}:{1}".format(key, value))
def get_filter_params(self):
return [f for f in self.filter_params if f]
self.config = {
'data.location': os.path.expanduser(data_location),
'confirmation': 'no',
+ 'dependency.confirmation': 'no', # See TW-1483 or taskrc man page
}
self.tasks = TaskQuerySet(self)
self.version = self._get_version()