Always retry locking when performing task handoff
There are some Ironic execution workflows where there is not an easy way
to retry, such as when attempting to hand off the processing of an async
task to a conductor. Task handoff can require releasing a lock on the
node, so the next entity processing the task can acquire the lock
itself. However, this is vulnerable to race conditions, as there is no
uniform retry mechanism built in to such handoffs. Consider the
continue_node_deploy/clean logic, which does this:
method = 'continue_node_%s' % operation
# Need to release the lock to let the conductor take it
task.release_resources()
getattr(rpc, method)(task.context, uuid, topic=topic
If another process obtains a lock between the releasing of resources and
the acquiring of the lock during the continue_node_* operation, and
holds the lock longer than the max attempt * interval window (which
defaults to 3 seconds), then the handoff will never complete. Beyond
that, because there is no proper queue for processes waiting on the
lock, there is no fairness, so it's also possible that instead of one
long lock being held, the lock is obtained and held for a short window
several times by other competing processes.
This manifests as nodes occasionally getting stuck in the "DEPLOYING"
state during a deploy. For example, a user may attempt to open or access
the serial console before the deploy is complete--the serial console
process obtains a lock and starves the conductor of the lock, so the
conductor cannot finish the deploy. It's also possible a long heartbeat
or badly-timed sequence of heartbeats could do the same.
To fix this, this commit introduces the concept of a "patient" lock,
which will retry indefinitely until it doesn't encounter the NodeLocked
exception. This overrides any retry behavior.
.. note::
There may be other cases where such a lock is desired.
Story: #2008323
Change-Id: I9937fab18a50111ec56a3fd023cdb9d510a1e990
(cherry picked from commit bfc2ad56d5
)
This commit is contained in:
parent
90da180a17
commit
25a05cf352
|
@ -861,7 +861,7 @@ class ConductorManager(base_manager.BaseConductorManager):
|
|||
|
||||
"""
|
||||
LOG.debug("RPC continue_node_deploy called for node %s.", node_id)
|
||||
with task_manager.acquire(context, node_id, shared=False,
|
||||
with task_manager.acquire(context, node_id, shared=False, patient=True,
|
||||
purpose='continue node deploying') as task:
|
||||
node = task.node
|
||||
|
||||
|
@ -1151,7 +1151,7 @@ class ConductorManager(base_manager.BaseConductorManager):
|
|||
"""
|
||||
LOG.debug("RPC continue_node_clean called for node %s.", node_id)
|
||||
|
||||
with task_manager.acquire(context, node_id, shared=False,
|
||||
with task_manager.acquire(context, node_id, shared=False, patient=True,
|
||||
purpose='continue node cleaning') as task:
|
||||
node = task.node
|
||||
if node.target_provision_state == states.MANAGEABLE:
|
||||
|
|
|
@ -170,7 +170,7 @@ class TaskManager(object):
|
|||
"""
|
||||
|
||||
def __init__(self, context, node_id, shared=False,
|
||||
purpose='unspecified action', retry=True,
|
||||
purpose='unspecified action', retry=True, patient=False,
|
||||
load_driver=True):
|
||||
"""Create a new TaskManager.
|
||||
|
||||
|
@ -185,6 +185,12 @@ class TaskManager(object):
|
|||
lock. Default: False.
|
||||
:param purpose: human-readable purpose to put to debug logs.
|
||||
:param retry: whether to retry locking if it fails. Default: True.
|
||||
:param patient: whether to indefinitely retry locking if it fails.
|
||||
Set this to True if there is an operation that does not
|
||||
have any fallback or built-in retry mechanism, such as
|
||||
finalizing a state transition during deploy/clean.
|
||||
The default retry behavior is to retry a configured
|
||||
number of times and then give up. Default: False.
|
||||
:param load_driver: whether to load the ``driver`` object. Set this to
|
||||
False if loading the driver is undesired or
|
||||
impossible.
|
||||
|
@ -203,6 +209,7 @@ class TaskManager(object):
|
|||
self.node_id = node_id
|
||||
self.shared = shared
|
||||
self._retry = retry
|
||||
self._patient = patient
|
||||
|
||||
self.fsm = states.machine.copy()
|
||||
self._purpose = purpose
|
||||
|
@ -260,7 +267,9 @@ class TaskManager(object):
|
|||
def _lock(self):
|
||||
self._debug_timer.restart()
|
||||
|
||||
if self._retry:
|
||||
if self._patient:
|
||||
attempts = None
|
||||
elif self._retry:
|
||||
attempts = CONF.conductor.node_locked_retry_attempts
|
||||
else:
|
||||
attempts = 1
|
||||
|
|
|
@ -1821,6 +1821,27 @@ class ContinueNodeDeployTestCase(mgr_utils.ServiceSetUpMixin,
|
|||
deployments.do_next_deploy_step,
|
||||
mock.ANY, 1)
|
||||
|
||||
@mock.patch('ironic.conductor.manager.ConductorManager._spawn_worker',
|
||||
autospec=True)
|
||||
def test_continue_node_deploy_locked(self, mock_spawn):
|
||||
"""Test that continuing a deploy via RPC cannot fail due to locks."""
|
||||
max_attempts = 3
|
||||
self.config(node_locked_retry_attempts=max_attempts, group='conductor')
|
||||
prv_state = states.DEPLOYWAIT
|
||||
tgt_prv_state = states.ACTIVE
|
||||
node = obj_utils.create_test_node(self.context, driver='fake-hardware',
|
||||
provision_state=prv_state,
|
||||
target_provision_state=tgt_prv_state,
|
||||
last_error=None,
|
||||
deploy_step=self.deploy_steps[0])
|
||||
self._start_service()
|
||||
with mock.patch.object(objects.Node, 'reserve', autospec=True) as mck:
|
||||
mck.side_effect = (
|
||||
([exception.NodeLocked(node='foo', host='foo')] * max_attempts)
|
||||
+ [node])
|
||||
self.service.continue_node_deploy(self.context, node.uuid)
|
||||
self._stop_service()
|
||||
|
||||
@mock.patch.object(task_manager.TaskManager, 'process_event',
|
||||
autospec=True)
|
||||
@mock.patch('ironic.conductor.manager.ConductorManager._spawn_worker',
|
||||
|
@ -2768,6 +2789,28 @@ class DoNodeCleanTestCase(mgr_utils.ServiceSetUpMixin, db_base.DbTestCase):
|
|||
def test_continue_node_clean_manual_abort_last_clean_step(self):
|
||||
self._continue_node_clean_abort_last_clean_step(manual=True)
|
||||
|
||||
@mock.patch('ironic.conductor.manager.ConductorManager._spawn_worker',
|
||||
autospec=True)
|
||||
def test_continue_node_clean_locked(self, mock_spawn):
|
||||
"""Test that continuing a clean via RPC cannot fail due to locks."""
|
||||
max_attempts = 3
|
||||
self.config(node_locked_retry_attempts=max_attempts, group='conductor')
|
||||
driver_info = {'clean_steps': [self.clean_steps[0]],
|
||||
'clean_step_index': 0}
|
||||
tgt_prov_state = states.AVAILABLE
|
||||
node = obj_utils.create_test_node(
|
||||
self.context, driver='fake-hardware',
|
||||
provision_state=states.CLEANWAIT,
|
||||
target_provision_state=tgt_prov_state, last_error=None,
|
||||
driver_internal_info=driver_info, clean_step=self.clean_steps[0])
|
||||
self._start_service()
|
||||
with mock.patch.object(objects.Node, 'reserve', autospec=True) as mck:
|
||||
mck.side_effect = (
|
||||
([exception.NodeLocked(node='foo', host='foo')] * max_attempts)
|
||||
+ [node])
|
||||
self.service.continue_node_clean(self.context, node.uuid)
|
||||
self._stop_service()
|
||||
|
||||
|
||||
class DoNodeRescueTestCase(mgr_utils.CommonMixIn, mgr_utils.ServiceSetUpMixin,
|
||||
db_base.DbTestCase):
|
||||
|
|
|
@ -222,6 +222,25 @@ class TaskManagerTestCase(db_base.DbTestCase):
|
|||
reserve_mock.assert_called_once_with(self.context, self.host,
|
||||
'fake-node-id')
|
||||
|
||||
def test_excl_lock_exception_patient(
|
||||
self, get_voltgt_mock, get_volconn_mock, get_portgroups_mock,
|
||||
get_ports_mock, build_driver_mock,
|
||||
reserve_mock, release_mock, node_get_mock):
|
||||
retry_attempts = 3
|
||||
self.config(node_locked_retry_attempts=retry_attempts,
|
||||
group='conductor')
|
||||
|
||||
# Fail on the first 3 attempts, succeed on the fourth.
|
||||
reserve_mock.side_effect = (
|
||||
([exception.NodeLocked(node='foo', host='foo')] * 3) + [self.node])
|
||||
|
||||
task_manager.TaskManager(self.context, 'fake-node-id', patient=True)
|
||||
|
||||
expected_calls = [mock.call(self.context, self.host,
|
||||
'fake-node-id')] * 4
|
||||
reserve_mock.assert_has_calls(expected_calls)
|
||||
self.assertEqual(4, reserve_mock.call_count)
|
||||
|
||||
def test_excl_lock_reserve_exception(
|
||||
self, get_voltgt_mock, get_volconn_mock, get_portgroups_mock,
|
||||
get_ports_mock, build_driver_mock,
|
||||
|
|
|
@ -0,0 +1,7 @@
|
|||
---
|
||||
fixes:
|
||||
- |
|
||||
Fixes a bug where a conductor could fail to complete a deployment if there
|
||||
was contention on a shared lock. This would manifest as an instance being
|
||||
stuck in the "deploying" state, though the node had in fact started or even
|
||||
completed its final boot.
|
Loading…
Reference in New Issue