MultiThreaded engine was implemented to execute tasks in parallel: - added parallel action that executes and reverts tasks in parallel; - added thread-safe storage. Change-Id: I4a1f78c95ae5d38660bd32ce21d2b3fb1b2af8ad
166 lines
5.5 KiB
Python
166 lines
5.5 KiB
Python
# -*- coding: utf-8 -*-
|
|
|
|
# vim: tabstop=4 shiftwidth=4 softtabstop=4
|
|
|
|
# Copyright (C) 2013 Yahoo! Inc. All Rights Reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
import logging
|
|
import threading
|
|
import threading2
|
|
import time
|
|
import types
|
|
|
|
|
|
LOG = logging.getLogger(__name__)
|
|
|
|
|
|
def await(check_functor, timeout=None):
|
|
if timeout is not None:
|
|
end_time = time.time() + max(0, timeout)
|
|
else:
|
|
end_time = None
|
|
# Use the same/similar scheme that the python condition class uses.
|
|
delay = 0.0005
|
|
while not check_functor():
|
|
time.sleep(delay)
|
|
if end_time is not None:
|
|
remaining = end_time - time.time()
|
|
if remaining <= 0:
|
|
return False
|
|
delay = min(delay * 2, remaining, 0.05)
|
|
else:
|
|
delay = min(delay * 2, 0.05)
|
|
return True
|
|
|
|
|
|
class MultiLock(object):
|
|
"""A class which can attempt to obtain many locks at once and release
|
|
said locks when exiting.
|
|
|
|
Useful as a context manager around many locks (instead of having to nest
|
|
said individual context managers).
|
|
"""
|
|
|
|
def __init__(self, locks):
|
|
assert len(locks) > 0, "Zero locks requested"
|
|
self._locks = locks
|
|
self._locked = [False] * len(locks)
|
|
|
|
def __enter__(self):
|
|
|
|
def is_locked(lock):
|
|
# NOTE(harlowja): the threading2 lock doesn't seem to have this
|
|
# attribute, so thats why we are checking it existing first.
|
|
if hasattr(lock, 'locked'):
|
|
return lock.locked()
|
|
return False
|
|
|
|
for i in xrange(0, len(self._locked)):
|
|
if self._locked[i] or is_locked(self._locks[i]):
|
|
raise threading.ThreadError("Lock %s not previously released"
|
|
% (i + 1))
|
|
self._locked[i] = False
|
|
for (i, lock) in enumerate(self._locks):
|
|
self._locked[i] = lock.acquire()
|
|
|
|
def __exit__(self, type, value, traceback):
|
|
for (i, locked) in enumerate(self._locked):
|
|
try:
|
|
if locked:
|
|
self._locks[i].release()
|
|
self._locked[i] = False
|
|
except threading.ThreadError:
|
|
LOG.exception("Unable to release lock %s", i + 1)
|
|
|
|
|
|
class CountDownLatch(object):
|
|
"""Similar in concept to the java count down latch."""
|
|
|
|
def __init__(self, count=0):
|
|
self.count = count
|
|
self.lock = threading.Condition()
|
|
|
|
def countDown(self):
|
|
with self.lock:
|
|
self.count -= 1
|
|
if self.count <= 0:
|
|
self.lock.notifyAll()
|
|
|
|
def await(self, timeout=None):
|
|
end_time = None
|
|
if timeout is not None:
|
|
timeout = max(0, timeout)
|
|
end_time = time.time() + timeout
|
|
time_up = False
|
|
with self.lock:
|
|
while True:
|
|
# Stop waiting on these 2 conditions.
|
|
if time_up or self.count <= 0:
|
|
break
|
|
# Was this a spurious wakeup or did we really end??
|
|
self.lock.wait(timeout=timeout)
|
|
if end_time is not None:
|
|
if time.time() >= end_time:
|
|
time_up = True
|
|
else:
|
|
# Reduce the timeout so that we don't wait extra time
|
|
# over what we initially were requested to.
|
|
timeout = end_time - time.time()
|
|
return self.count <= 0
|
|
|
|
|
|
class ThreadGroupExecutor(object):
|
|
"""A simple thread executor that spins up new threads (or greenthreads) for
|
|
each task to be completed (no pool limit is enforced).
|
|
|
|
TODO(harlowja): Likely if we use the more advanced executors that come with
|
|
the concurrent.futures library we can just get rid of this.
|
|
"""
|
|
|
|
def __init__(self, daemonize=True):
|
|
self._threads = []
|
|
self._group = threading2.ThreadGroup()
|
|
self._daemonize = daemonize
|
|
|
|
def submit(self, fn, *args, **kwargs):
|
|
t = threading2.Thread(target=fn, group=self._group,
|
|
args=args, kwargs=kwargs)
|
|
t.daemon = self._daemonize
|
|
self._threads.append(t)
|
|
t.start()
|
|
|
|
def await_termination(self, timeout=None):
|
|
if not self._threads:
|
|
return
|
|
return self._group.join(timeout)
|
|
|
|
|
|
class ThreadSafeMeta(type):
|
|
"""Metaclass that adds locking to all pubic methods of a class"""
|
|
|
|
def __new__(cls, name, bases, attrs):
|
|
from taskflow import decorators
|
|
for attr_name, attr_value in attrs.iteritems():
|
|
if isinstance(attr_value, types.FunctionType):
|
|
if attr_name[0] != '_':
|
|
attrs[attr_name] = decorators.locked(attr_value)
|
|
return super(ThreadSafeMeta, cls).__new__(cls, name, bases, attrs)
|
|
|
|
def __call__(cls, *args, **kwargs):
|
|
instance = super(ThreadSafeMeta, cls).__call__(*args, **kwargs)
|
|
if not hasattr(instance, '_lock'):
|
|
instance._lock = threading.RLock()
|
|
return instance
|