Files
deb-python-taskflow/taskflow/engines/action_engine/executor.py
Ivan A. Melnikov 5cbfcba760 Add make_completed_future to async_utils
Instead of keeping this three lines of code inside one of executors,
we move them to special utility for farther re-use.

Change-Id: I8f0ed76737fe29ad9492260cb84b66d09ec07f70
2013-12-25 16:07:59 +02:00

150 lines
4.6 KiB
Python

# -*- coding: utf-8 -*-
# vim: tabstop=4 shiftwidth=4 softtabstop=4
# Copyright (C) 2013 Yahoo! Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import abc
import contextlib
import six
from concurrent import futures
from taskflow.utils import async_utils
from taskflow.utils import misc
from taskflow.utils import threading_utils
@contextlib.contextmanager
def _autobind(task, bind_name, bind_func, **kwargs):
try:
task.bind(bind_name, bind_func, **kwargs)
yield task
finally:
task.unbind(bind_name, bind_func)
def _noop(*args, **kwargs):
pass
def _execute_task(task, arguments, progress_callback):
with _autobind(task, 'update_progress', progress_callback):
try:
result = task.execute(**arguments)
except Exception:
# NOTE(imelnikov): wrap current exception with Failure
# object and return it
result = misc.Failure()
return task, 'executed', result
def _revert_task(task, arguments, result, failures, progress_callback):
kwargs = arguments.copy()
kwargs['result'] = result
kwargs['flow_failures'] = failures
with _autobind(task, 'update_progress', progress_callback):
try:
result = task.revert(**kwargs)
except Exception:
# NOTE(imelnikov): wrap current exception with Failure
# object and return it
result = misc.Failure()
return task, 'reverted', result
@six.add_metaclass(abc.ABCMeta)
class TaskExecutorBase(object):
"""Executes and reverts tasks.
This class takes task and its arguments and executes or reverts it.
It encapsulates knowledge on how task should be executed or reverted:
right now, on separate thread, on another machine, etc.
"""
@abc.abstractmethod
def execute_task(self, task, arguments, progress_callback=_noop):
"""Schedules task execution."""
@abc.abstractmethod
def revert_task(self, task, arguments, result, failures,
progress_callback=_noop):
"""Schedules task reversion"""
@abc.abstractmethod
def wait_for_any(self, fs, timeout=None):
"""Wait for futures returned by this executor to complete"""
def start(self):
"""Prepare to execute tasks"""
pass
def stop(self):
"""Finalize task executor"""
pass
class SerialTaskExecutor(TaskExecutorBase):
"""Execute task one after another."""
def execute_task(self, task, arguments, progress_callback=_noop):
return async_utils.make_completed_future(
_execute_task(task, arguments, progress_callback))
def revert_task(self, task, arguments, result, failures,
progress_callback=_noop):
return async_utils.make_completed_future(
_revert_task(task, arguments, result,
failures, progress_callback))
def wait_for_any(self, fs, timeout=None):
# NOTE(imelnikov): this executor returns only done futures
return fs, []
class ParallelTaskExecutor(TaskExecutorBase):
"""Executes tasks in parallel.
Submits tasks to executor which should provide interface similar
to concurrent.Futures.Executor.
"""
def __init__(self, executor=None):
self._executor = executor
self._own_executor = executor is None
def execute_task(self, task, arguments, progress_callback=_noop):
return self._executor.submit(
_execute_task, task, arguments, progress_callback)
def revert_task(self, task, arguments, result, failures,
progress_callback=_noop):
return self._executor.submit(
_revert_task, task,
arguments, result, failures, progress_callback)
def wait_for_any(self, fs, timeout=None):
return async_utils.wait_for_any(fs, timeout)
def start(self):
if self._own_executor:
thread_count = threading_utils.get_optimal_thread_count()
self._executor = futures.ThreadPoolExecutor(thread_count)
def stop(self):
if self._own_executor:
self._executor.shutdown(wait=True)
self._executor = None