import abc
+import copy
import datetime
import json
import logging
import six
import subprocess
-from .task import Task, TaskQuerySet
+from .task import Task, TaskQuerySet, ReadOnlyDictView
from .filters import TaskWarriorFilter
from .serializing import local_zone
logger = logging.getLogger(__name__)
+
class Backend(object):
@abc.abstractproperty
Converts TW syntax datetime string to a localized datetime
object. This method is not mandatory.
"""
- raise NotImplemented
+ raise NotImplementedError
class TaskWarriorException(Exception):
pass
-class TaskWarrior(object):
+class TaskWarrior(Backend):
VERSION_2_1_0 = six.u('2.1.0')
VERSION_2_2_0 = six.u('2.2.0')
VERSION_2_4_4 = six.u('2.4.4')
VERSION_2_4_5 = six.u('2.4.5')
- def __init__(self, data_location=None, create=True, taskrc_location='~/.taskrc'):
+ def __init__(self, data_location=None, create=True,
+ taskrc_location='~/.taskrc'):
self.taskrc_location = os.path.expanduser(taskrc_location)
# If taskrc does not exist, pass / to use defaults and avoid creating
if not os.path.exists(self.taskrc_location):
self.taskrc_location = '/'
+ self._config = None
self.version = self._get_version()
- self.config = {
+ self.overrides = {
'confirmation': 'no',
'dependency.confirmation': 'no', # See TW-1483 or taskrc man page
'recurrence.confirmation': 'no', # Necessary for modifying R tasks
data_location = os.path.expanduser(data_location)
if create and not os.path.exists(data_location):
os.makedirs(data_location)
- self.config['data.location'] = data_location
+ self.overrides['data.location'] = data_location
self.tasks = TaskQuerySet(self)
def _get_command_args(self, args, config_override=None):
command_args = ['task', 'rc:{0}'.format(self.taskrc_location)]
- config = self.config.copy()
- config.update(config_override or dict())
- for item in config.items():
+ overrides = self.overrides.copy()
+ overrides.update(config_override or dict())
+ for item in overrides.items():
command_args.append('rc.{0}={1}'.format(*item))
- command_args.extend(map(six.text_type, args))
+ command_args.extend([
+ x.decode('utf-8') if isinstance(x, six.binary_type)
+ else six.text_type(x) for x in args
+ ])
return command_args
def _get_version(self):
p = subprocess.Popen(
- ['task', '--version'],
- stdout=subprocess.PIPE,
- stderr=subprocess.PIPE)
+ ['task', '--version'],
+ stdout=subprocess.PIPE,
+ stderr=subprocess.PIPE)
stdout, stderr = [x.decode('utf-8') for x in p.communicate()]
return stdout.strip('\n')
if serialized_value is '':
escaped_serialized_value = ''
else:
- escaped_serialized_value = six.u("'{0}'").format(serialized_value)
+ escaped_serialized_value = six.u("'{0}'").format(
+ serialized_value)
- format_default = lambda task: six.u("{0}:{1}").format(field,
- escaped_serialized_value)
+ format_default = lambda task: six.u("{0}:{1}").format(
+ field, escaped_serialized_value)
format_func = getattr(self, 'format_{0}'.format(field),
format_default)
if task.saved:
for field in task._modified_fields:
add_field(field)
+
# For new tasks, pass all fields that make sense
else:
for field in task._data.keys():
+ # We cannot set stuff that's read only (ID, UUID, ..)
if field in task.read_only_fields:
continue
+ # We do not want to do field deletion for new tasks
+ if task._data[field] is None:
+ continue
+ # Otherwise we're fine
add_field(field)
return args
# Removed dependencies need to be prefixed with '-'
return 'depends:' + ','.join(
- [t['uuid'] for t in added] +
- ['-' + t['uuid'] for t in removed]
- )
+ [t['uuid'] for t in added] +
+ ['-' + t['uuid'] for t in removed]
+ )
def format_description(self, task):
# Task version older than 2.4.0 ignores first word of the
if self.version < self.VERSION_2_4_0:
return task._data['description']
else:
- return six.u("description:'{0}'").format(task._data['description'] or '')
+ return six.u("description:'{0}'").format(
+ task._data['description'] or '',
+ )
def convert_datetime_string(self, value):
naive = datetime.datetime.strptime(result[0], DATE_FORMAT_CALC)
localized = local_zone.localize(naive)
else:
- raise ValueError("Provided value could not be converted to "
- "datetime, its type is not supported: {}"
- .format(type(value)))
+ raise ValueError(
+ 'Provided value could not be converted to '
+ 'datetime, its type is not supported: {}'
+ .format(type(value)),
+ )
return localized
# Public interface
- def get_config(self):
+ @property
+ def config(self):
+ # First, check if memoized information is available
+ if self._config:
+ return self._config
+
+ # If not, fetch the config using the 'show' command
raw_output = self.execute_command(
- ['show'],
- config_override={'verbose': 'nothing'}
- )
+ ['show'],
+ config_override={'verbose': 'nothing'}
+ )
config = dict()
- config_regex = re.compile(r'^(?P<key>[^\s]+)\s+(?P<value>[^\s].+$)')
+ config_regex = re.compile(r'^(?P<key>[^\s]+)\s+(?P<value>[^\s].*$)')
for line in raw_output:
match = config_regex.match(line)
if match:
config[match.group('key')] = match.group('value').strip()
- return config
+ # Memoize the config dict
+ self._config = ReadOnlyDictView(config)
+
+ return self._config
def execute_command(self, args, config_override=None, allow_failure=True,
return_all=False):
command_args = self._get_command_args(
args, config_override=config_override)
- logger.debug(' '.join(command_args))
+ logger.debug(u' '.join(command_args))
+
p = subprocess.Popen(command_args, stdout=subprocess.PIPE,
stderr=subprocess.PIPE)
stdout, stderr = [x.decode('utf-8') for x in p.communicate()]
error_msg = stderr.strip()
else:
error_msg = stdout.strip()
+ error_msg += u'\nCommand used: ' + u' '.join(command_args)
raise TaskWarriorException(error_msg)
# Return all whole triplet only if explicitly asked for
def filter_tasks(self, filter_obj):
self.enforce_recurrence()
- args = ['export', '--'] + filter_obj.get_filter_params()
+ args = ['export'] + filter_obj.get_filter_params()
tasks = []
for line in self.execute_command(args):
if line:
id_lines = [l for l in output if l.startswith('Created task ')]
# Complain loudly if it seems that more tasks were created
- # Should not happen
- if len(id_lines) != 1 or len(id_lines[0].split(' ')) != 3:
- raise TaskWarriorException("Unexpected output when creating "
- "task: %s" % '\n'.join(id_lines))
+ # Should not happen.
+ # Expected output: Created task 1.
+ # Created task 1 (recurrence template).
+ if len(id_lines) != 1 or len(id_lines[0].split(' ')) not in (3, 5):
+ raise TaskWarriorException(
+ 'Unexpected output when creating '
+ 'task: %s' % '\n'.join(id_lines),
+ )
# Circumvent the ID storage, since ID is considered read-only
identifier = id_lines[0].split(' ')[2].rstrip('.')
for key, value in data.items():
taskfilter.add_filter_param(key, value)
- output = self.execute_command(['export', '--'] +
- taskfilter.get_filter_params())
+ output = self.execute_command(['export'] +
+ taskfilter.get_filter_params())
# If more than 1 task has been matched still, raise an exception
if not valid(output):
raise TaskWarriorException(
- "Unique identifiers {0} with description: {1} matches "
- "multiple tasks: {2}".format(
- task['uuid'] or task['id'], task['description'], output)
+ 'Unique identifiers {0} with description: {1} matches '
+ 'multiple tasks: {2}'.format(
+ task['uuid'] or task['id'], task['description'], output)
)
return json.loads(output[0])