Merge "nova-less-deploy: deploy_instances and undeploy_instances workflows"
This commit is contained in:
commit
3b70004b10
@ -28,3 +28,5 @@ python-keystoneclient>=3.8.0 # Apache-2.0
|
||||
keystoneauth1>=3.4.0 # Apache-2.0
|
||||
tenacity>=4.4.0 # Apache-2.0
|
||||
futures>=3.0.0;python_version=='2.7' or python_version=='2.6' # BSD
|
||||
metalsmith>=0.8.0 # Apache-2.0
|
||||
jsonschema<3.0.0,>=2.6.0 # MIT
|
||||
|
@ -81,6 +81,11 @@ mistral.actions =
|
||||
tripleo.baremetal.validate_nodes = tripleo_common.actions.baremetal:ValidateNodes
|
||||
tripleo.baremetal.get_candidate_nodes = tripleo_common.actions.baremetal:GetCandidateNodes
|
||||
tripleo.baremetal.probe_node = tripleo_common.actions.baremetal:ProbeNode
|
||||
tripleo.baremetal_deploy.check_existing_instances = tripleo_common.actions.baremetal_deploy:CheckExistingInstancesAction
|
||||
tripleo.baremetal_deploy.deploy_node = tripleo_common.actions.baremetal_deploy:DeployNodeAction
|
||||
tripleo.baremetal_deploy.reserve_nodes = tripleo_common.actions.baremetal_deploy:ReserveNodesAction
|
||||
tripleo.baremetal_deploy.undeploy_instance = tripleo_common.actions.baremetal_deploy:UndeployInstanceAction
|
||||
tripleo.baremetal_deploy.wait_for_deploy = tripleo_common.actions.baremetal_deploy:WaitForDeploymentAction
|
||||
tripleo.config.download_config = tripleo_common.actions.config:DownloadConfigAction
|
||||
tripleo.config.get_overcloud_config = tripleo_common.actions.config:GetOvercloudConfig
|
||||
tripleo.container_images.prepare = tripleo_common.actions.container_images:PrepareContainerImageEnv
|
||||
|
343
tripleo_common/actions/baremetal_deploy.py
Normal file
343
tripleo_common/actions/baremetal_deploy.py
Normal file
@ -0,0 +1,343 @@
|
||||
# Copyright 2018 Red Hat, Inc.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import logging
|
||||
|
||||
import jsonschema
|
||||
import metalsmith
|
||||
from metalsmith import sources
|
||||
from mistral_lib import actions
|
||||
import six
|
||||
|
||||
from tripleo_common.actions import base
|
||||
from tripleo_common.utils import keystone
|
||||
|
||||
LOG = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def _provisioner(context):
|
||||
session = keystone.get_session(context)
|
||||
return metalsmith.Provisioner(session=session)
|
||||
|
||||
|
||||
_INSTANCES_INPUT_SCHEMA = {
|
||||
'type': 'array',
|
||||
'items': {
|
||||
'type': 'object',
|
||||
'properties': {
|
||||
'capabilities': {'type': 'object'},
|
||||
'hostname': {'type': 'string',
|
||||
'minLength': 2,
|
||||
'maxLength': 255},
|
||||
'image': {'type': 'string'},
|
||||
'image_checksum': {'type': 'string'},
|
||||
'image_kernel': {'type': 'string'},
|
||||
'image_ramdisk': {'type': 'string'},
|
||||
'name': {'type': 'string'},
|
||||
'nics': {'type': 'array',
|
||||
'items': {'type': 'object',
|
||||
'properties': {
|
||||
'network': {'type': 'string'},
|
||||
'port': {'type': 'string'},
|
||||
'fixed_ip': {'type': 'string'},
|
||||
},
|
||||
'additionalProperties': False}},
|
||||
'profile': {'type': 'string'},
|
||||
'resource_class': {'type': 'string'},
|
||||
'root_size_gb': {'type': 'integer', 'minimum': 4},
|
||||
'swap_size_mb': {'type': 'integer', 'minimum': 64},
|
||||
'traits': {'type': 'array',
|
||||
'items': {'type': 'string'}},
|
||||
},
|
||||
'additionalProperties': False,
|
||||
# Host name is required, but defaults to name in _validate_instances
|
||||
'required': ['hostname'],
|
||||
}
|
||||
}
|
||||
"""JSON schema of the input for these actions."""
|
||||
|
||||
|
||||
class CheckExistingInstancesAction(base.TripleOAction):
|
||||
"""Detect which requested instances have already been provisioned."""
|
||||
|
||||
def __init__(self, instances):
|
||||
super(CheckExistingInstancesAction, self).__init__()
|
||||
self.instances = instances
|
||||
|
||||
def run(self, context):
|
||||
try:
|
||||
_validate_instances(self.instances)
|
||||
except Exception as exc:
|
||||
LOG.error('Failed to validate provided instances. %s', exc)
|
||||
return actions.Result(error=six.text_type(exc))
|
||||
|
||||
provisioner = _provisioner(context)
|
||||
|
||||
not_found = []
|
||||
found = []
|
||||
for request in self.instances:
|
||||
try:
|
||||
instance = provisioner.show_instance(request['hostname'])
|
||||
# TODO(dtantsur): use openstacksdk exceptions when metalsmith
|
||||
# is bumped to 0.9.0.
|
||||
except Exception:
|
||||
not_found.append(request)
|
||||
else:
|
||||
found.append(instance.to_dict())
|
||||
|
||||
if found:
|
||||
LOG.info('Found existing instances: %s',
|
||||
', '.join('%s (on node %s)' % (i['hostname'], i['uuid'])
|
||||
for i in found))
|
||||
if not_found:
|
||||
LOG.info('Instance(s) %s do not exist',
|
||||
', '.join(r['hostname'] for r in not_found))
|
||||
|
||||
return {
|
||||
'not_found': not_found,
|
||||
'instances': found
|
||||
}
|
||||
|
||||
|
||||
class ReserveNodesAction(base.TripleOAction):
|
||||
"""Reserve nodes for requested instances."""
|
||||
|
||||
def __init__(self, instances, default_resource_class='baremetal'):
|
||||
super(ReserveNodesAction, self).__init__()
|
||||
self.instances = instances
|
||||
self.default_resource_class = default_resource_class
|
||||
|
||||
def run(self, context):
|
||||
try:
|
||||
_validate_instances(self.instances)
|
||||
except Exception as exc:
|
||||
LOG.error('Failed to validate provided instances. %s', exc)
|
||||
return actions.Result(error=six.text_type(exc))
|
||||
|
||||
provisioner = _provisioner(context)
|
||||
|
||||
# TODO(dtantsur): looping over instances is not very optimal, change it
|
||||
# to metalsmith plan deployment API when it's available.
|
||||
result = []
|
||||
nodes = []
|
||||
try:
|
||||
for instance in self.instances:
|
||||
LOG.debug('Trying to reserve a node for instance %s', instance)
|
||||
if instance.get('name'):
|
||||
# NOTE(dtantsur): metalsmith accepts list of nodes to pick
|
||||
# from. We implement a simplest case when a user can pick a
|
||||
# node by its name (actually, UUID will also work).
|
||||
candidates = [instance['name']]
|
||||
else:
|
||||
candidates = None
|
||||
|
||||
if instance.get('profile'):
|
||||
# TODO(dtantsur): change to traits?
|
||||
instance.setdefault(
|
||||
'capabilities', {})['profile'] = instance['profile']
|
||||
|
||||
node = provisioner.reserve_node(
|
||||
resource_class=instance.get('resource_class') or
|
||||
self.default_resource_class,
|
||||
capabilities=instance.get('capabilities'),
|
||||
candidates=candidates,
|
||||
traits=instance.get('traits'))
|
||||
LOG.info('Reserved node %s for instance %s', node, instance)
|
||||
nodes.append(node)
|
||||
try:
|
||||
node_id = node.id
|
||||
except AttributeError:
|
||||
# TODO(dtantsur): transition from ironicclient to
|
||||
# openstacksdk, remove when metalsmith is bumped to 0.9.0
|
||||
node_id = node.uuid
|
||||
result.append({'node': node_id, 'instance': instance})
|
||||
except Exception as exc:
|
||||
LOG.exception('Provisioning failed, cleaning up')
|
||||
# Remove all reservations on failure
|
||||
try:
|
||||
_release_nodes(provisioner, nodes)
|
||||
except Exception:
|
||||
LOG.exception('Clean up failed, some nodes may still be '
|
||||
'reserved by failed instances')
|
||||
return actions.Result(
|
||||
error="%s: %s" % (type(exc).__name__, exc)
|
||||
)
|
||||
|
||||
return {'reservations': result}
|
||||
|
||||
|
||||
class DeployNodeAction(base.TripleOAction):
|
||||
"""Provision instance on a previously reserved node."""
|
||||
|
||||
def __init__(self, instance, node, ssh_keys=None,
|
||||
# For compatibility with deployment based on heat+nova
|
||||
ssh_user_name='heat-admin',
|
||||
default_image='overcloud-full',
|
||||
default_network='ctlplane',
|
||||
# 50 is the default for old flavors, subtracting 1G to account
|
||||
# for partitioning and configdrive.
|
||||
default_root_size=49):
|
||||
super(DeployNodeAction, self).__init__()
|
||||
self.instance = instance
|
||||
self.node = node
|
||||
self.config = metalsmith.InstanceConfig(ssh_keys=ssh_keys)
|
||||
self.config.add_user(ssh_user_name, admin=True, sudo=True)
|
||||
self.default_image = default_image
|
||||
self.default_network = default_network
|
||||
self.default_root_size = default_root_size
|
||||
|
||||
def _get_image(self):
|
||||
# TODO(dtantsur): move this logic to metalsmith in 0.9.0
|
||||
image = self.instance.get('image', self.default_image)
|
||||
image_type = _link_type(image)
|
||||
if image_type == 'glance':
|
||||
return sources.GlanceImage(image)
|
||||
else:
|
||||
checksum = self.instance.get('image_checksum')
|
||||
if (checksum and image_type == 'http' and
|
||||
_link_type(checksum) == 'http'):
|
||||
kwargs = {'checksum_url': checksum}
|
||||
else:
|
||||
kwargs = {'checksum': checksum}
|
||||
|
||||
whole_disk_image = not (self.instance.get('image_kernel') or
|
||||
self.instance.get('image_ramdisk'))
|
||||
|
||||
if whole_disk_image:
|
||||
if image_type == 'http':
|
||||
return sources.HttpWholeDiskImage(image, **kwargs)
|
||||
else:
|
||||
return sources.FileWholeDiskImage(image, **kwargs)
|
||||
else:
|
||||
if image_type == 'http':
|
||||
return sources.HttpPartitionImage(
|
||||
image,
|
||||
kernel_url=self.instance.get('image_kernel'),
|
||||
ramdisk_url=self.instance.get('image_ramdisk'),
|
||||
**kwargs)
|
||||
else:
|
||||
return sources.FilePartitionImage(
|
||||
image,
|
||||
kernel_location=self.instance.get('image_kernel'),
|
||||
ramdisk_location=self.instance.get('image_ramdisk'),
|
||||
**kwargs)
|
||||
|
||||
def run(self, context):
|
||||
try:
|
||||
_validate_instances([self.instance])
|
||||
except Exception as exc:
|
||||
LOG.error('Failed to validate the request. %s', exc)
|
||||
return actions.Result(error=six.text_type(exc))
|
||||
|
||||
provisioner = _provisioner(context)
|
||||
|
||||
LOG.debug('Starting provisioning of %s on node %s',
|
||||
self.instance, self.node)
|
||||
try:
|
||||
instance = provisioner.provision_node(
|
||||
self.node,
|
||||
config=self.config,
|
||||
hostname=self.instance['hostname'],
|
||||
image=self._get_image(),
|
||||
nics=self.instance.get('nics',
|
||||
[{'network': self.default_network}]),
|
||||
root_size_gb=self.instance.get('root_size_gb',
|
||||
self.default_root_size),
|
||||
swap_size_mb=self.instance.get('swap_size_mb'),
|
||||
)
|
||||
except Exception as exc:
|
||||
LOG.exception('Provisioning of %s on node %s failed',
|
||||
self.instance, self.node)
|
||||
try:
|
||||
_release_nodes(provisioner, [self.node])
|
||||
except Exception:
|
||||
LOG.exception('Clean up failed, node %s may still be '
|
||||
'reserved by the failed instance', self.node)
|
||||
return actions.Result(
|
||||
error="%s: %s" % (type(exc).__name__, exc)
|
||||
)
|
||||
|
||||
LOG.info('Started provisioning of %s on node %s',
|
||||
self.instance, self.node)
|
||||
return instance.to_dict()
|
||||
|
||||
|
||||
class WaitForDeploymentAction(base.TripleOAction):
|
||||
"""Wait for the instance to be deployed."""
|
||||
|
||||
def __init__(self, instance, timeout=3600):
|
||||
super(WaitForDeploymentAction, self).__init__()
|
||||
self.instance = instance
|
||||
self.timeout = timeout
|
||||
|
||||
def run(self, context):
|
||||
provisioner = _provisioner(context)
|
||||
|
||||
LOG.debug('Waiting for instance %s to provision',
|
||||
self.instance['hostname'])
|
||||
instance = provisioner.wait_for_provisioning([self.instance['uuid']],
|
||||
timeout=self.timeout)[0]
|
||||
LOG.info('Successfully provisioned instance %s',
|
||||
self.instance['hostname'])
|
||||
return instance.to_dict()
|
||||
|
||||
|
||||
class UndeployInstanceAction(base.TripleOAction):
|
||||
"""Undeploy a previously deployed instance."""
|
||||
|
||||
def __init__(self, instance, timeout=1800):
|
||||
super(UndeployInstanceAction, self).__init__()
|
||||
self.instance = instance
|
||||
self.timeout = timeout
|
||||
|
||||
def run(self, context):
|
||||
provisioner = _provisioner(context)
|
||||
|
||||
try:
|
||||
instance = provisioner.show_instance(self.instance)
|
||||
except Exception:
|
||||
LOG.warning('Cannot get instance %s, assuming already deleted',
|
||||
self.instance)
|
||||
return
|
||||
|
||||
LOG.debug('Unprovisioning instance %s', instance.hostname)
|
||||
provisioner.unprovision_node(instance.node, wait=self.timeout)
|
||||
LOG.info('Successfully unprovisioned %s', instance.hostname)
|
||||
|
||||
|
||||
def _validate_instances(instances):
|
||||
for inst in instances:
|
||||
if inst.get('name') and not inst.get('hostname'):
|
||||
inst['hostname'] = inst['name']
|
||||
jsonschema.validate(instances, _INSTANCES_INPUT_SCHEMA)
|
||||
|
||||
|
||||
def _release_nodes(provisioner, nodes):
|
||||
for node in nodes:
|
||||
LOG.debug('Removing reservation from node %s', node)
|
||||
try:
|
||||
provisioner.unprovision_node(node)
|
||||
except Exception:
|
||||
LOG.exception('Unable to release node %s, moving on', node)
|
||||
else:
|
||||
LOG.info('Removed reservation from node %s', node)
|
||||
|
||||
|
||||
def _link_type(image):
|
||||
if image.startswith('http://') or image.startswith('https://'):
|
||||
return 'http'
|
||||
elif image.startswith('file://'):
|
||||
return 'file'
|
||||
else:
|
||||
return 'glance'
|
342
tripleo_common/tests/actions/test_baremetal_deploy.py
Normal file
342
tripleo_common/tests/actions/test_baremetal_deploy.py
Normal file
@ -0,0 +1,342 @@
|
||||
# Copyright 2018 Red Hat, Inc.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from metalsmith import sources
|
||||
import mock
|
||||
|
||||
from tripleo_common.actions import baremetal_deploy
|
||||
from tripleo_common.tests import base
|
||||
|
||||
|
||||
@mock.patch.object(baremetal_deploy, '_provisioner', autospec=True)
|
||||
class TestReserveNodes(base.TestCase):
|
||||
|
||||
def test_success(self, mock_pr):
|
||||
instances = [
|
||||
{'hostname': 'host1', 'profile': 'compute'},
|
||||
{'hostname': 'host2', 'resource_class': 'compute',
|
||||
'capabilities': {'answer': '42'}},
|
||||
{'name': 'control-0', 'traits': ['CUSTOM_GPU']},
|
||||
]
|
||||
action = baremetal_deploy.ReserveNodesAction(instances)
|
||||
result = action.run(mock.Mock())
|
||||
|
||||
self.assertEqual(
|
||||
[{'node': mock_pr.return_value.reserve_node.return_value.id,
|
||||
'instance': req} for req in instances],
|
||||
result['reservations'])
|
||||
mock_pr.return_value.reserve_node.assert_has_calls([
|
||||
mock.call(resource_class='baremetal', traits=None,
|
||||
capabilities={'profile': 'compute'}, candidates=None),
|
||||
mock.call(resource_class='compute', traits=None,
|
||||
capabilities={'answer': '42'}, candidates=None),
|
||||
mock.call(resource_class='baremetal', traits=['CUSTOM_GPU'],
|
||||
capabilities=None, candidates=['control-0']),
|
||||
])
|
||||
self.assertFalse(mock_pr.return_value.unprovision_node.called)
|
||||
|
||||
def test_missing_hostname(self, mock_pr):
|
||||
instances = [
|
||||
{'hostname': 'host1'},
|
||||
{'resource_class': 'compute', 'capabilities': {'answer': '42'}}
|
||||
]
|
||||
action = baremetal_deploy.ReserveNodesAction(instances)
|
||||
result = action.run(mock.Mock())
|
||||
|
||||
self.assertIn("'hostname' is a required property", result.error)
|
||||
self.assertFalse(mock_pr.return_value.reserve_node.called)
|
||||
self.assertFalse(mock_pr.return_value.unprovision_node.called)
|
||||
|
||||
def test_failure(self, mock_pr):
|
||||
instances = [
|
||||
{'hostname': 'host1'},
|
||||
{'hostname': 'host2', 'resource_class': 'compute',
|
||||
'capabilities': {'answer': '42'}},
|
||||
{'hostname': 'host3'},
|
||||
]
|
||||
success_node = mock.Mock(uuid='uuid1')
|
||||
mock_pr.return_value.reserve_node.side_effect = [
|
||||
success_node,
|
||||
RuntimeError("boom"),
|
||||
]
|
||||
action = baremetal_deploy.ReserveNodesAction(instances)
|
||||
result = action.run(mock.Mock())
|
||||
|
||||
self.assertIn('RuntimeError: boom', result.error)
|
||||
mock_pr.return_value.reserve_node.assert_has_calls([
|
||||
mock.call(resource_class='baremetal', capabilities=None,
|
||||
candidates=None, traits=None),
|
||||
mock.call(resource_class='compute', capabilities={'answer': '42'},
|
||||
candidates=None, traits=None)
|
||||
])
|
||||
mock_pr.return_value.unprovision_node.assert_called_once_with(
|
||||
success_node)
|
||||
|
||||
|
||||
@mock.patch.object(baremetal_deploy, '_provisioner', autospec=True)
|
||||
class TestDeployNode(base.TestCase):
|
||||
|
||||
def test_success_defaults(self, mock_pr):
|
||||
action = baremetal_deploy.DeployNodeAction(
|
||||
instance={'hostname': 'host1'},
|
||||
node='1234'
|
||||
)
|
||||
result = action.run(mock.Mock())
|
||||
|
||||
pr = mock_pr.return_value
|
||||
self.assertEqual(
|
||||
pr.provision_node.return_value.to_dict.return_value,
|
||||
result)
|
||||
pr.provision_node.assert_called_once_with(
|
||||
'1234',
|
||||
image=mock.ANY,
|
||||
nics=[{'network': 'ctlplane'}],
|
||||
hostname='host1',
|
||||
root_size_gb=49,
|
||||
swap_size_mb=None,
|
||||
config=mock.ANY,
|
||||
)
|
||||
config = pr.provision_node.call_args[1]['config']
|
||||
self.assertEqual([], config.ssh_keys)
|
||||
self.assertEqual('heat-admin', config.users[0]['name'])
|
||||
source = pr.provision_node.call_args[1]['image']
|
||||
self.assertIsInstance(source, sources.GlanceImage)
|
||||
# TODO(dtantsur): check the image when it's a public field
|
||||
|
||||
def test_success_with_name(self, mock_pr):
|
||||
action = baremetal_deploy.DeployNodeAction(
|
||||
instance={'name': 'host1'},
|
||||
node='1234'
|
||||
)
|
||||
result = action.run(mock.Mock())
|
||||
|
||||
pr = mock_pr.return_value
|
||||
self.assertEqual(
|
||||
pr.provision_node.return_value.to_dict.return_value,
|
||||
result)
|
||||
pr.provision_node.assert_called_once_with(
|
||||
'1234',
|
||||
image=mock.ANY,
|
||||
nics=[{'network': 'ctlplane'}],
|
||||
hostname='host1',
|
||||
root_size_gb=49,
|
||||
swap_size_mb=None,
|
||||
config=mock.ANY,
|
||||
)
|
||||
config = pr.provision_node.call_args[1]['config']
|
||||
self.assertEqual([], config.ssh_keys)
|
||||
self.assertEqual('heat-admin', config.users[0]['name'])
|
||||
|
||||
def test_success(self, mock_pr):
|
||||
pr = mock_pr.return_value
|
||||
action = baremetal_deploy.DeployNodeAction(
|
||||
instance={'hostname': 'host1',
|
||||
'image': 'overcloud-alt',
|
||||
'nics': [{'port': 'abcd'}],
|
||||
'root_size_gb': 100,
|
||||
'swap_size_mb': 4096},
|
||||
node='1234',
|
||||
ssh_keys=['ssh key contents'],
|
||||
ssh_user_name='admin',
|
||||
)
|
||||
result = action.run(mock.Mock())
|
||||
|
||||
self.assertEqual(
|
||||
pr.provision_node.return_value.to_dict.return_value,
|
||||
result)
|
||||
pr.provision_node.assert_called_once_with(
|
||||
'1234',
|
||||
image=mock.ANY,
|
||||
nics=[{'port': 'abcd'}],
|
||||
hostname='host1',
|
||||
root_size_gb=100,
|
||||
swap_size_mb=4096,
|
||||
config=mock.ANY,
|
||||
)
|
||||
config = pr.provision_node.call_args[1]['config']
|
||||
self.assertEqual(['ssh key contents'], config.ssh_keys)
|
||||
self.assertEqual('admin', config.users[0]['name'])
|
||||
source = pr.provision_node.call_args[1]['image']
|
||||
self.assertIsInstance(source, sources.GlanceImage)
|
||||
# TODO(dtantsur): check the image when it's a public field
|
||||
|
||||
# NOTE(dtantsur): limited coverage for source detection since this code is
|
||||
# being moved to metalsmith in 0.9.0.
|
||||
def test_success_http_partition_image(self, mock_pr):
|
||||
action = baremetal_deploy.DeployNodeAction(
|
||||
instance={'hostname': 'host1',
|
||||
'image': 'https://example/image',
|
||||
'image_kernel': 'https://example/kernel',
|
||||
'image_ramdisk': 'https://example/ramdisk',
|
||||
'image_checksum': 'https://example/checksum'},
|
||||
node='1234'
|
||||
)
|
||||
result = action.run(mock.Mock())
|
||||
|
||||
pr = mock_pr.return_value
|
||||
self.assertEqual(
|
||||
pr.provision_node.return_value.to_dict.return_value,
|
||||
result)
|
||||
pr.provision_node.assert_called_once_with(
|
||||
'1234',
|
||||
image=mock.ANY,
|
||||
nics=[{'network': 'ctlplane'}],
|
||||
hostname='host1',
|
||||
root_size_gb=49,
|
||||
swap_size_mb=None,
|
||||
config=mock.ANY,
|
||||
)
|
||||
config = pr.provision_node.call_args[1]['config']
|
||||
self.assertEqual([], config.ssh_keys)
|
||||
self.assertEqual('heat-admin', config.users[0]['name'])
|
||||
source = pr.provision_node.call_args[1]['image']
|
||||
self.assertIsInstance(source, sources.HttpPartitionImage)
|
||||
self.assertEqual('https://example/image', source.url)
|
||||
self.assertEqual('https://example/kernel', source.kernel_url)
|
||||
self.assertEqual('https://example/ramdisk', source.ramdisk_url)
|
||||
self.assertEqual('https://example/checksum', source.checksum_url)
|
||||
|
||||
def test_success_file_partition_image(self, mock_pr):
|
||||
action = baremetal_deploy.DeployNodeAction(
|
||||
instance={'hostname': 'host1',
|
||||
'image': 'file:///var/lib/ironic/image',
|
||||
'image_kernel': 'file:///var/lib/ironic/kernel',
|
||||
'image_ramdisk': 'file:///var/lib/ironic/ramdisk',
|
||||
'image_checksum': 'abcd'},
|
||||
node='1234'
|
||||
)
|
||||
result = action.run(mock.Mock())
|
||||
|
||||
pr = mock_pr.return_value
|
||||
self.assertEqual(
|
||||
pr.provision_node.return_value.to_dict.return_value,
|
||||
result)
|
||||
pr.provision_node.assert_called_once_with(
|
||||
'1234',
|
||||
image=mock.ANY,
|
||||
nics=[{'network': 'ctlplane'}],
|
||||
hostname='host1',
|
||||
root_size_gb=49,
|
||||
swap_size_mb=None,
|
||||
config=mock.ANY,
|
||||
)
|
||||
config = pr.provision_node.call_args[1]['config']
|
||||
self.assertEqual([], config.ssh_keys)
|
||||
self.assertEqual('heat-admin', config.users[0]['name'])
|
||||
source = pr.provision_node.call_args[1]['image']
|
||||
self.assertIsInstance(source, sources.FilePartitionImage)
|
||||
self.assertEqual('file:///var/lib/ironic/image', source.location)
|
||||
self.assertEqual('file:///var/lib/ironic/kernel',
|
||||
source.kernel_location)
|
||||
self.assertEqual('file:///var/lib/ironic/ramdisk',
|
||||
source.ramdisk_location)
|
||||
self.assertEqual('abcd', source.checksum)
|
||||
|
||||
def test_failure(self, mock_pr):
|
||||
pr = mock_pr.return_value
|
||||
action = baremetal_deploy.DeployNodeAction(
|
||||
instance={'hostname': 'host1'},
|
||||
node='1234'
|
||||
)
|
||||
pr.provision_node.side_effect = RuntimeError('boom')
|
||||
result = action.run(mock.Mock())
|
||||
|
||||
self.assertIn('RuntimeError: boom', result.error)
|
||||
pr.provision_node.assert_called_once_with(
|
||||
'1234',
|
||||
image=mock.ANY,
|
||||
nics=[{'network': 'ctlplane'}],
|
||||
hostname='host1',
|
||||
root_size_gb=49,
|
||||
swap_size_mb=None,
|
||||
config=mock.ANY,
|
||||
)
|
||||
pr.unprovision_node.assert_called_once_with('1234')
|
||||
|
||||
|
||||
@mock.patch.object(baremetal_deploy, '_provisioner', autospec=True)
|
||||
class TestCheckExistingInstances(base.TestCase):
|
||||
|
||||
def test_success(self, mock_pr):
|
||||
pr = mock_pr.return_value
|
||||
instances = [
|
||||
{'hostname': 'host1'},
|
||||
{'hostname': 'host2', 'resource_class': 'compute',
|
||||
'capabilities': {'answer': '42'}}
|
||||
]
|
||||
existing = mock.MagicMock()
|
||||
pr.show_instance.side_effect = [
|
||||
RuntimeError('not found'),
|
||||
existing,
|
||||
]
|
||||
action = baremetal_deploy.CheckExistingInstancesAction(instances)
|
||||
result = action.run(mock.Mock())
|
||||
|
||||
self.assertEqual({
|
||||
'instances': [existing.to_dict.return_value],
|
||||
'not_found': [{'hostname': 'host1'}]
|
||||
}, result)
|
||||
pr.show_instance.assert_has_calls([
|
||||
mock.call('host1'), mock.call('host2')
|
||||
])
|
||||
|
||||
def test_missing_hostname(self, mock_pr):
|
||||
instances = [
|
||||
{'hostname': 'host1'},
|
||||
{'resource_class': 'compute', 'capabilities': {'answer': '42'}}
|
||||
]
|
||||
action = baremetal_deploy.CheckExistingInstancesAction(instances)
|
||||
result = action.run(mock.Mock())
|
||||
|
||||
self.assertIn("'hostname' is a required property", result.error)
|
||||
self.assertFalse(mock_pr.return_value.show_instance.called)
|
||||
|
||||
|
||||
@mock.patch.object(baremetal_deploy, '_provisioner', autospec=True)
|
||||
class TestWaitForDeployment(base.TestCase):
|
||||
|
||||
def test_success(self, mock_pr):
|
||||
pr = mock_pr.return_value
|
||||
action = baremetal_deploy.WaitForDeploymentAction(
|
||||
{'hostname': 'compute.cloud', 'uuid': 'uuid1'})
|
||||
result = action.run(mock.Mock())
|
||||
|
||||
pr.wait_for_provisioning.assert_called_once_with(['uuid1'],
|
||||
timeout=3600)
|
||||
inst = pr.wait_for_provisioning.return_value[0]
|
||||
self.assertIs(result, inst.to_dict.return_value)
|
||||
|
||||
|
||||
@mock.patch.object(baremetal_deploy, '_provisioner', autospec=True)
|
||||
class TestUndeployInstance(base.TestCase):
|
||||
|
||||
def test_success(self, mock_pr):
|
||||
pr = mock_pr.return_value
|
||||
action = baremetal_deploy.UndeployInstanceAction('inst1')
|
||||
result = action.run(mock.Mock())
|
||||
self.assertIsNone(result)
|
||||
|
||||
pr.show_instance.assert_called_once_with('inst1')
|
||||
pr.unprovision_node.assert_called_once_with(
|
||||
pr.show_instance.return_value.node, wait=1800)
|
||||
|
||||
def test_not_found(self, mock_pr):
|
||||
pr = mock_pr.return_value
|
||||
pr.show_instance.side_effect = RuntimeError('not found')
|
||||
action = baremetal_deploy.UndeployInstanceAction('inst1')
|
||||
result = action.run(mock.Mock())
|
||||
self.assertIsNone(result)
|
||||
|
||||
pr.show_instance.assert_called_once_with('inst1')
|
||||
self.assertFalse(pr.unprovision_node.called)
|
@ -16,9 +16,10 @@
|
||||
|
||||
import six
|
||||
|
||||
from keystoneauth1.identity.generic import Token as IdentityToken
|
||||
from keystoneauth1 import loading
|
||||
from keystoneauth1 import session as ks_session
|
||||
from keystoneauth1.token_endpoint import Token
|
||||
from keystoneauth1.token_endpoint import Token as SimpleToken
|
||||
from keystoneclient import service_catalog as ks_service_catalog
|
||||
from keystoneclient.v3 import client as ks_client
|
||||
from keystoneclient.v3 import endpoints as ks_endpoints
|
||||
@ -90,10 +91,10 @@ def get_session_and_auth(context, **kwargs):
|
||||
}
|
||||
)
|
||||
|
||||
auth = Token(endpoint=endpoint, token=context.auth_token)
|
||||
auth = SimpleToken(endpoint=endpoint, token=context.auth_token)
|
||||
|
||||
auth_uri = context.auth_uri or CONF.keystone_authtoken.auth_uri
|
||||
ks_auth = Token(
|
||||
ks_auth = SimpleToken(
|
||||
endpoint=auth_uri,
|
||||
token=context.auth_token
|
||||
)
|
||||
@ -108,6 +109,39 @@ def get_session_and_auth(context, **kwargs):
|
||||
}
|
||||
|
||||
|
||||
# NOTE(dtantsur): get_session_and_auth returns a session tied to a specific
|
||||
# service. This function returns a generic session. Eventually we should switch
|
||||
# everything to using it and service-specific Adapter on top.
|
||||
def get_session(context):
|
||||
"""Get a generic session suitable for any service(s).
|
||||
|
||||
:param context: action context
|
||||
:return: keystone `Session`
|
||||
"""
|
||||
try:
|
||||
context = context.security
|
||||
except AttributeError:
|
||||
pass
|
||||
|
||||
auth_uri = context.auth_uri or CONF.keystone_authtoken.auth_uri
|
||||
|
||||
try:
|
||||
# TODO(dtantsur): a better way to detect the project domain?
|
||||
project_domain = context.service_catalog['project']['domain']['name']
|
||||
except KeyError:
|
||||
project_domain = CONF.keystone_authtoken.project_domain_name
|
||||
|
||||
ks_auth = IdentityToken(auth_uri, token=context.auth_token,
|
||||
# NOTE(dtantsur): project scope is required for V3
|
||||
project_name=context.project_name,
|
||||
project_domain_name=project_domain)
|
||||
sess = ks_session.Session(
|
||||
auth=ks_auth,
|
||||
verify=_determine_verify(context)
|
||||
)
|
||||
return sess
|
||||
|
||||
|
||||
def _admin_client(trust_id=None):
|
||||
if CONF.keystone_authtoken.auth_type is None:
|
||||
auth_url = CONF.keystone_authtoken.auth_uri
|
||||
|
190
workbooks/baremetal_deploy.yaml
Normal file
190
workbooks/baremetal_deploy.yaml
Normal file
@ -0,0 +1,190 @@
|
||||
---
|
||||
version: '2.0'
|
||||
name: tripleo.baremetal_deploy.v1
|
||||
description: TripleO Baremetal Deployment Workflows
|
||||
|
||||
workflows:
|
||||
|
||||
_deploy_one:
|
||||
description: Internal workflow to deploy one node
|
||||
|
||||
input:
|
||||
- instance
|
||||
- node
|
||||
- ssh_keys: []
|
||||
- ssh_user_name: heat-admin
|
||||
- timeout: 3600
|
||||
- queue_name: tripleo
|
||||
|
||||
tags:
|
||||
- tripleo-common-managed
|
||||
|
||||
tasks:
|
||||
|
||||
deploy_node:
|
||||
action: tripleo.baremetal_deploy.deploy_node
|
||||
input:
|
||||
instance: <% $.instance %>
|
||||
node: <% $.node %>
|
||||
ssh_keys: <% $.ssh_keys %>
|
||||
ssh_user_name: <% $.ssh_user_name %>
|
||||
publish:
|
||||
instance: <% task().result %>
|
||||
publish-on-error:
|
||||
status: FAILED
|
||||
message: <% task().result %>
|
||||
on-success: wait_for_deploy
|
||||
on-error: send_message
|
||||
|
||||
wait_for_deploy:
|
||||
action: tripleo.baremetal_deploy.wait_for_deploy
|
||||
input:
|
||||
instance: <% $.instance %>
|
||||
timeout: <% $.timeout %>
|
||||
publish:
|
||||
instance: <% task().result %>
|
||||
message: Instance <% task().result.hostname %> deployed successfully
|
||||
publish-on-error:
|
||||
status: FAILED
|
||||
message: <% task().result %>
|
||||
on-complete: send_message
|
||||
|
||||
send_message:
|
||||
workflow: tripleo.messaging.v1.send
|
||||
input:
|
||||
queue_name: <% $.queue_name %>
|
||||
type: <% execution().name %>
|
||||
status: <% $.get('status', 'SUCCESS') %>
|
||||
message: <% $.get('message', '') %>
|
||||
execution: <% execution() %>
|
||||
instance: <% $.instance %>
|
||||
on-success:
|
||||
- fail: <% $.get('status', 'SUCCESS') != 'SUCCESS' %>
|
||||
|
||||
output:
|
||||
instance: <% $.instance %>
|
||||
|
||||
output-on-error:
|
||||
result: <% $.get('message', 'Deployment failed') %>
|
||||
|
||||
|
||||
deploy_instances:
|
||||
description: Deploy instances on bare metal nodes.
|
||||
|
||||
input:
|
||||
- instances
|
||||
- ssh_keys: []
|
||||
- ssh_user_name: heat-admin
|
||||
- timeout: 3600
|
||||
- concurrency: 20
|
||||
- queue_name: tripleo
|
||||
|
||||
tags:
|
||||
- tripleo-common-managed
|
||||
|
||||
tasks:
|
||||
|
||||
find_existing_instances:
|
||||
action: tripleo.baremetal_deploy.check_existing_instances
|
||||
input:
|
||||
instances: <% $.instances %>
|
||||
publish:
|
||||
instances: <% task().result.not_found %>
|
||||
existing_instances: <% task().result.instances %>
|
||||
publish-on-error:
|
||||
status: FAILED
|
||||
message: <% task().result %>
|
||||
on-success: reserve_nodes
|
||||
on-error: send_message
|
||||
|
||||
reserve_nodes:
|
||||
action: tripleo.baremetal_deploy.reserve_nodes
|
||||
input:
|
||||
instances: <% $.instances %>
|
||||
publish:
|
||||
reservations: <% task().result.reservations %>
|
||||
publish-on-error:
|
||||
status: FAILED
|
||||
message: <% task().result %>
|
||||
on-success: deploy_nodes
|
||||
on-error: send_message
|
||||
|
||||
deploy_nodes:
|
||||
with-items: reservation in <% $.reservations %>
|
||||
concurrency: <% $.concurrency %>
|
||||
workflow: _deploy_one
|
||||
input:
|
||||
instance: <% $.reservation.instance %>
|
||||
node: <% $.reservation.node %>
|
||||
ssh_keys: <% $.ssh_keys %>
|
||||
ssh_user_name: <% $.ssh_user_name %>
|
||||
timeout: <% $.timeout %>
|
||||
queue_name: <% $.queue_name %>
|
||||
publish:
|
||||
all_instances: <% task().result.instance + $.existing_instances %>
|
||||
new_instances: <% task().result.instance %>
|
||||
publish-on-error:
|
||||
status: FAILED
|
||||
message: <% task().result %>
|
||||
on-success: publish_result
|
||||
on-error: send_message
|
||||
|
||||
publish_result:
|
||||
publish:
|
||||
ctlplane_ips: <% $.all_instances.toDict($.hostname, $.ip_addresses.ctlplane[0]) %>
|
||||
instances: <% $.all_instances.toDict($.hostname, $) %>
|
||||
on-complete: send_message
|
||||
|
||||
send_message:
|
||||
workflow: tripleo.messaging.v1.send
|
||||
input:
|
||||
queue_name: <% $.queue_name %>
|
||||
type: <% execution().name %>
|
||||
status: <% $.get('status', 'SUCCESS') %>
|
||||
message: <% $.get('message', '') %>
|
||||
execution: <% execution() %>
|
||||
payload:
|
||||
ctlplane_ips: <% $.get('ctlplane_ips', {}) %>
|
||||
instances: <% $.get('instances', {}) %>
|
||||
|
||||
output:
|
||||
ctlplane_ips: <% $.ctlplane_ips %>
|
||||
existing_instances: <% $.existing_instances.toDict($.hostname, $) %>
|
||||
instances: <% $.instances %>
|
||||
new_instances: <% $.new_instances.toDict($.hostname, $) %>
|
||||
|
||||
|
||||
undeploy_instances:
|
||||
description: Undeploy previously deployed instances
|
||||
|
||||
input:
|
||||
- instances
|
||||
- timeout: 3600
|
||||
- concurrency: 20
|
||||
- queue_name: tripleo
|
||||
|
||||
tags:
|
||||
- tripleo-common-managed
|
||||
|
||||
tasks:
|
||||
|
||||
undeploy_instances:
|
||||
with-items: instance in <% $.instances %>
|
||||
concurrency: <% $.concurrency %>
|
||||
action: tripleo.baremetal_deploy.undeploy_instance
|
||||
input:
|
||||
instance: <% $.instance %>
|
||||
timeout: <% $.timeout %>
|
||||
publish-on-error:
|
||||
status: FAILED
|
||||
message: <% task().result %>
|
||||
on-complete: send_message
|
||||
|
||||
send_message:
|
||||
workflow: tripleo.messaging.v1.send
|
||||
input:
|
||||
queue_name: <% $.queue_name %>
|
||||
type: <% execution().name %>
|
||||
status: <% $.get('status', 'SUCCESS') %>
|
||||
message: <% $.get('message', '') %>
|
||||
execution: <% execution() %>
|
Loading…
Reference in New Issue
Block a user