cd940dbcc0
We currently don't use the --timeout option during the node delete execution. This change adds the timeout arg to each playbook execution called. Change-Id: I814e5b14b6f22394417fe8fb808a03cabeb2629e
732 lines
31 KiB
Python
732 lines
31 KiB
Python
# Copyright 2015 Red Hat, Inc.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
#
|
|
|
|
import collections
|
|
import copy
|
|
import datetime
|
|
from io import StringIO
|
|
import ipaddress
|
|
import json
|
|
import logging
|
|
import os
|
|
import sys
|
|
import time
|
|
|
|
from cliff.formatters import table
|
|
from openstack import exceptions as openstack_exc
|
|
from osc_lib import exceptions as oscexc
|
|
from osc_lib.i18n import _
|
|
from osc_lib import utils
|
|
import yaml
|
|
|
|
from tripleoclient import command
|
|
from tripleoclient import constants
|
|
from tripleoclient import exceptions
|
|
from tripleoclient import utils as oooutils
|
|
from tripleoclient.workflows import baremetal
|
|
from tripleoclient.workflows import tripleo_baremetal as tb
|
|
|
|
|
|
class DeleteNode(command.Command):
|
|
"""Delete overcloud nodes."""
|
|
|
|
log = logging.getLogger(__name__ + ".DeleteNode")
|
|
|
|
def get_parser(self, prog_name):
|
|
parser = super(DeleteNode, self).get_parser(prog_name)
|
|
group = parser.add_mutually_exclusive_group(required=True)
|
|
group.add_argument('nodes', metavar='<node>', nargs="*",
|
|
default=[],
|
|
help=_('Node ID(s) to delete (otherwise specified '
|
|
'in the --baremetal-deployment file)'))
|
|
group.add_argument('-b', '--baremetal-deployment',
|
|
metavar='<BAREMETAL DEPLOYMENT FILE>',
|
|
help=_('Configuration file describing the '
|
|
'baremetal deployment'))
|
|
|
|
parser.add_argument('--stack', dest='stack',
|
|
help=_('Name or ID of heat stack to scale '
|
|
'(default=Env: OVERCLOUD_STACK_NAME)'),
|
|
default=utils.env('OVERCLOUD_STACK_NAME',
|
|
default='overcloud'))
|
|
|
|
parser.add_argument(
|
|
'--timeout', metavar='<TIMEOUT>',
|
|
type=int, default=constants.STACK_TIMEOUT, dest='timeout',
|
|
help=_("Timeout in minutes to wait for the nodes to be deleted. "
|
|
"Keep in mind that due to keystone session duration "
|
|
"that timeout has an upper bound of 4 hours ")
|
|
)
|
|
parser.add_argument(
|
|
'--overcloud-ssh-port-timeout',
|
|
help=_('Timeout for the ssh port to become active.'),
|
|
type=int,
|
|
default=constants.ENABLE_SSH_ADMIN_SSH_PORT_TIMEOUT
|
|
)
|
|
parser.add_argument('-y', '--yes',
|
|
help=_('Skip yes/no prompt (assume yes).'),
|
|
default=False,
|
|
action="store_true")
|
|
return parser
|
|
|
|
def _nodes_to_delete(self, parsed_args, roles):
|
|
with oooutils.TempDirs() as tmp:
|
|
unprovision_confirm = os.path.join(
|
|
tmp, 'unprovision_confirm.json')
|
|
|
|
oooutils.run_ansible_playbook(
|
|
playbook='cli-overcloud-node-unprovision.yaml',
|
|
inventory='localhost,',
|
|
workdir=tmp,
|
|
timeout=parsed_args.timeout,
|
|
playbook_dir=constants.ANSIBLE_TRIPLEO_PLAYBOOKS,
|
|
verbosity=oooutils.playbook_verbosity(self=self),
|
|
extra_vars={
|
|
"stack_name": parsed_args.stack,
|
|
"baremetal_deployment": roles,
|
|
"prompt": True,
|
|
"unprovision_confirm": unprovision_confirm,
|
|
}
|
|
)
|
|
with open(unprovision_confirm) as f:
|
|
to_unprovision = json.load(f)
|
|
if isinstance(to_unprovision, dict):
|
|
nodes = to_unprovision.get(
|
|
'instances') + to_unprovision.get('pre_provisioned')
|
|
else:
|
|
nodes = to_unprovision
|
|
if not nodes:
|
|
print('No nodes to unprovision')
|
|
return None, None
|
|
TableArgs = collections.namedtuple(
|
|
'TableArgs', 'print_empty max_width fit_width')
|
|
args = TableArgs(print_empty=True, max_width=-1, fit_width=True)
|
|
nodes_data = [(i.get('hostname', ''),
|
|
i.get('name', ''),
|
|
i.get('id', '')) for i in nodes]
|
|
|
|
node_hostnames = [i['hostname'] for i in nodes if 'hostname' in i]
|
|
|
|
formatter = table.TableFormatter()
|
|
output = StringIO()
|
|
formatter.emit_list(
|
|
column_names=['hostname', 'name', 'id'],
|
|
data=nodes_data,
|
|
stdout=output,
|
|
parsed_args=args
|
|
)
|
|
return output.getvalue(), node_hostnames
|
|
|
|
def _check_skiplist_exists(self, env):
|
|
skiplist = env.get('parameter_defaults',
|
|
{}).get('DeploymentServerBlacklist')
|
|
if skiplist:
|
|
self.log.warning(_('[WARNING] DeploymentServerBlacklist is '
|
|
'ignored when executing scale down actions. If '
|
|
'the node(s) being removed should *NOT* have '
|
|
'any actions executed on them, please shut '
|
|
'them off prior to their removal.'))
|
|
|
|
def _check_timeout(self, start, timeout):
|
|
used = int((time.time() - start) // 60)
|
|
remaining = timeout - used
|
|
if remaining <= 0:
|
|
raise exceptions.DeploymentError(
|
|
'Deployment timed out after %sm' % used
|
|
)
|
|
return remaining
|
|
|
|
def take_action(self, parsed_args):
|
|
self.log.debug("take_action(%s)" % parsed_args)
|
|
|
|
# Start our timer. This will be used to calculate the timeout.
|
|
start = time.time()
|
|
|
|
if parsed_args.baremetal_deployment:
|
|
with open(parsed_args.baremetal_deployment, 'r') as fp:
|
|
roles = yaml.safe_load(fp)
|
|
|
|
nodes_text, nodes = self._nodes_to_delete(parsed_args, roles)
|
|
if nodes_text:
|
|
print(nodes_text)
|
|
else:
|
|
return
|
|
else:
|
|
nodes = parsed_args.nodes
|
|
nodes_text = '\n'.join('- %s' % node for node in nodes)
|
|
if not parsed_args.yes:
|
|
confirm = oooutils.prompt_user_for_confirmation(
|
|
message=_("Are you sure you want to delete these overcloud "
|
|
"nodes [y/N]? "),
|
|
logger=self.log)
|
|
if not confirm:
|
|
raise oscexc.CommandError("Action not confirmed, exiting.")
|
|
|
|
ansible_dir = os.path.join(oooutils.get_default_working_dir(
|
|
parsed_args.stack
|
|
),
|
|
'config-download',
|
|
parsed_args.stack)
|
|
|
|
inventory = os.path.join(ansible_dir,
|
|
'tripleo-ansible-inventory.yaml')
|
|
|
|
ansible_cfg = os.path.join(ansible_dir, 'ansible.cfg')
|
|
key_file = oooutils.get_key(parsed_args.stack)
|
|
|
|
remaining = self._check_timeout(start, parsed_args.timeout)
|
|
|
|
oooutils.run_ansible_playbook(
|
|
playbook='scale_playbook.yaml',
|
|
inventory=inventory,
|
|
workdir=ansible_dir,
|
|
playbook_dir=ansible_dir,
|
|
ansible_cfg=ansible_cfg,
|
|
ssh_user='tripleo-admin',
|
|
limit_hosts=':'.join('%s' % node for node in nodes),
|
|
reproduce_command=True,
|
|
ignore_unreachable=True,
|
|
timeout=remaining,
|
|
extra_env_variables={
|
|
"ANSIBLE_BECOME": True,
|
|
"ANSIBLE_PRIVATE_KEY_FILE": key_file
|
|
}
|
|
)
|
|
|
|
remaining = self._check_timeout(start, parsed_args.timeout)
|
|
|
|
if parsed_args.baremetal_deployment:
|
|
with oooutils.TempDirs() as tmp:
|
|
oooutils.run_ansible_playbook(
|
|
playbook='cli-overcloud-node-unprovision.yaml',
|
|
inventory='localhost,',
|
|
workdir=tmp,
|
|
playbook_dir=constants.ANSIBLE_TRIPLEO_PLAYBOOKS,
|
|
timeout=remaining,
|
|
verbosity=oooutils.playbook_verbosity(self=self),
|
|
extra_vars={
|
|
"stack_name": parsed_args.stack,
|
|
"baremetal_deployment": roles,
|
|
"prompt": False,
|
|
"manage_network_ports": True,
|
|
}
|
|
)
|
|
|
|
|
|
class ProvideNode(command.Command):
|
|
"""Mark nodes as available based on UUIDs or current 'manageable' state."""
|
|
|
|
log = logging.getLogger(__name__ + ".ProvideNode")
|
|
|
|
def get_parser(self, prog_name):
|
|
parser = super(ProvideNode, self).get_parser(prog_name)
|
|
group = parser.add_mutually_exclusive_group(required=True)
|
|
group.add_argument('node_uuids',
|
|
nargs="*",
|
|
metavar="<node_uuid>",
|
|
default=[],
|
|
help=_('Baremetal Node UUIDs for the node(s) to be '
|
|
'provided'))
|
|
group.add_argument("--all-manageable",
|
|
action='store_true',
|
|
help=_("Provide all nodes currently in 'manageable'"
|
|
" state"))
|
|
group.add_argument("--verbosity",
|
|
type=int,
|
|
default=1,
|
|
help=_("Print debug output during execution"))
|
|
return parser
|
|
|
|
def take_action(self, parsed_args):
|
|
self.log.debug("take_action(%s)" % parsed_args)
|
|
|
|
provide = tb.TripleoProvide(verbosity=parsed_args.verbosity)
|
|
|
|
if parsed_args.node_uuids:
|
|
provide.provide(nodes=parsed_args.node_uuids)
|
|
|
|
else:
|
|
provide.provide_manageable_nodes()
|
|
|
|
|
|
class CleanNode(command.Command):
|
|
"""Run node(s) through cleaning."""
|
|
|
|
log = logging.getLogger(__name__ + ".CleanNode")
|
|
|
|
def get_parser(self, prog_name):
|
|
parser = super(CleanNode, self).get_parser(prog_name)
|
|
group = parser.add_mutually_exclusive_group(required=True)
|
|
group.add_argument('node_uuids',
|
|
nargs="*",
|
|
metavar="<node_uuid>",
|
|
default=[],
|
|
help=_('Baremetal Node UUIDs for the node(s) to be '
|
|
'cleaned'))
|
|
group.add_argument("--all-manageable",
|
|
action='store_true',
|
|
help=_("Clean all nodes currently in 'manageable'"
|
|
" state"))
|
|
group.add_argument("--verbosity",
|
|
type=int,
|
|
default=1,
|
|
help=_("Print debug output during execution"))
|
|
parser.add_argument('--provide',
|
|
action='store_true',
|
|
help=_('Provide (make available) the nodes once '
|
|
'cleaned'))
|
|
return parser
|
|
|
|
def take_action(self, parsed_args):
|
|
self.log.debug("take_action(%s)" % parsed_args)
|
|
|
|
nodes = parsed_args.node_uuids
|
|
|
|
clean = tb.TripleoClean(verbosity=parsed_args.verbosity)
|
|
if nodes:
|
|
clean.clean(
|
|
nodes=parsed_args.node_uuids)
|
|
else:
|
|
clean.clean_manageable_nodes()
|
|
|
|
if parsed_args.provide:
|
|
provide = tb.TripleoProvide(verbosity=parsed_args.verbosity)
|
|
if nodes:
|
|
provide.provide(nodes=nodes)
|
|
else:
|
|
provide.provide_manageable_nodes()
|
|
|
|
|
|
class ConfigureNode(command.Command):
|
|
"""Configure Node boot options."""
|
|
|
|
log = logging.getLogger(__name__ + ".ConfigureNode")
|
|
|
|
def get_parser(self, prog_name):
|
|
parser = super(ConfigureNode, self).get_parser(prog_name)
|
|
group = parser.add_mutually_exclusive_group(required=True)
|
|
group.add_argument('node_uuids',
|
|
nargs="*",
|
|
metavar="<node_uuid>",
|
|
default=[],
|
|
help=_('Baremetal Node UUIDs for the node(s) to be '
|
|
'configured'))
|
|
group.add_argument("--all-manageable",
|
|
action='store_true',
|
|
help=_("Configure all nodes currently in "
|
|
"'manageable' state"))
|
|
parser.add_argument(
|
|
'--deploy-kernel',
|
|
default='file://%s/agent.kernel' %
|
|
constants.IRONIC_HTTP_BOOT_BIND_MOUNT,
|
|
help=_('Image with deploy kernel.'))
|
|
parser.add_argument(
|
|
'--deploy-ramdisk',
|
|
default='file://%s/agent.ramdisk' %
|
|
constants.IRONIC_HTTP_BOOT_BIND_MOUNT,
|
|
help=_('Image with deploy ramdisk.'))
|
|
parser.add_argument('--instance-boot-option',
|
|
choices=['local', 'netboot'],
|
|
help=_('Whether to set instances for booting from '
|
|
'local hard drive (local) or network '
|
|
'(netboot).'))
|
|
parser.add_argument('--boot-mode',
|
|
choices=['uefi', 'bios'],
|
|
help=_('Whether to set the boot mode to UEFI '
|
|
'(uefi) or legacy BIOS (bios)'))
|
|
parser.add_argument('--root-device',
|
|
help=_('Define the root device for nodes. '
|
|
'Can be either a list of device names '
|
|
'(without /dev) to choose from or one of '
|
|
'two strategies: largest or smallest. For '
|
|
'it to work this command should be run '
|
|
'after the introspection.'))
|
|
parser.add_argument('--root-device-minimum-size',
|
|
type=int, default=4,
|
|
help=_('Minimum size (in GiB) of the detected '
|
|
'root device. Used with --root-device.'))
|
|
parser.add_argument('--overwrite-root-device-hints',
|
|
action='store_true',
|
|
help=_('Whether to overwrite existing root device '
|
|
'hints when --root-device is used.'))
|
|
parser.add_argument("--verbosity",
|
|
type=int,
|
|
default=1,
|
|
help=_("Print debug output during execution"))
|
|
return parser
|
|
|
|
def take_action(self, parsed_args):
|
|
self.log.debug("take_action(%s)" % parsed_args)
|
|
|
|
conf = tb.TripleoConfigure(
|
|
kernel_name=parsed_args.deploy_kernel,
|
|
ramdisk_name=parsed_args.deploy_ramdisk,
|
|
instance_boot_option=parsed_args.instance_boot_option,
|
|
boot_mode=parsed_args.boot_mode,
|
|
root_device=parsed_args.root_device,
|
|
root_device_minimum_size=parsed_args.root_device_minimum_size,
|
|
overwrite_root_device_hints=(
|
|
parsed_args.overwrite_root_device_hints)
|
|
)
|
|
|
|
if parsed_args.node_uuids:
|
|
conf.configure(
|
|
node_uuids=parsed_args.node_uuids)
|
|
else:
|
|
conf.configure_manageable_nodes()
|
|
|
|
|
|
class DiscoverNode(command.Command):
|
|
"""Discover overcloud nodes by polling their BMCs."""
|
|
|
|
log = logging.getLogger(__name__ + ".DiscoverNode")
|
|
|
|
def get_parser(self, prog_name):
|
|
parser = super(DiscoverNode, self).get_parser(prog_name)
|
|
ip_group = parser.add_mutually_exclusive_group(required=True)
|
|
ip_group.add_argument('--ip', action='append',
|
|
dest='ip_addresses', metavar='<ips>',
|
|
help=_('IP address(es) to probe'))
|
|
ip_group.add_argument('--range', dest='ip_addresses',
|
|
metavar='<range>', help=_('IP range to probe'))
|
|
parser.add_argument('--credentials', metavar='<key:value>',
|
|
action='append', required=True,
|
|
help=_('Key/value pairs of possible credentials'))
|
|
parser.add_argument('--port', action='append', metavar='<ports>',
|
|
type=int, help=_('BMC port(s) to probe'))
|
|
parser.add_argument('--introspect', action='store_true',
|
|
help=_('Introspect the imported nodes'))
|
|
parser.add_argument('--run-validations', action='store_true',
|
|
default=False,
|
|
help=_('Run the pre-deployment validations. These '
|
|
'external validations are from the TripleO '
|
|
'Validations project.'))
|
|
parser.add_argument('--provide', action='store_true',
|
|
help=_('Provide (make available) the nodes'))
|
|
parser.add_argument('--no-deploy-image', action='store_true',
|
|
help=_('Skip setting the deploy kernel and '
|
|
'ramdisk.'))
|
|
parser.add_argument('--instance-boot-option',
|
|
choices=['local', 'netboot'], default='local',
|
|
help=_('Whether to set instances for booting from '
|
|
'local hard drive (local) or network '
|
|
'(netboot).'))
|
|
parser.add_argument('--concurrency', type=int,
|
|
default=20,
|
|
help=_('Maximum number of nodes to introspect at '
|
|
'once.'))
|
|
parser.add_argument('--node-timeout', type=int,
|
|
default=1200,
|
|
help=_('Maximum timeout for node introspection.'))
|
|
parser.add_argument('--max-retries', type=int,
|
|
default=1,
|
|
help=_('Maximum introspection retries.'))
|
|
parser.add_argument('--retry-timeout', type=int,
|
|
default=120,
|
|
help=_('Maximum timeout between introspection'
|
|
'retries'))
|
|
parser.add_argument("--verbosity",
|
|
type=int,
|
|
default=1,
|
|
help=_("Print debug output during execution"))
|
|
return parser
|
|
|
|
# FIXME(tonyb): This is not multi-arch safe :(
|
|
def take_action(self, parsed_args):
|
|
self.log.debug("take_action(%s)" % parsed_args)
|
|
|
|
if parsed_args.no_deploy_image:
|
|
deploy_kernel = None
|
|
deploy_ramdisk = None
|
|
else:
|
|
deploy_kernel = 'file://{}/agent.kernel'.format(
|
|
constants.IRONIC_HTTP_BOOT_BIND_MOUNT
|
|
)
|
|
deploy_ramdisk = 'file://{}/agent.ramdisk'.format(
|
|
constants.IRONIC_HTTP_BOOT_BIND_MOUNT
|
|
)
|
|
|
|
credentials = [list(x.split(':', 1)) for x in parsed_args.credentials]
|
|
kwargs = {}
|
|
# Leave it up to the workflow to figure out the defaults
|
|
if parsed_args.port:
|
|
kwargs['ports'] = parsed_args.port
|
|
|
|
nodes = baremetal.discover_and_enroll(
|
|
self.app.client_manager,
|
|
ip_addresses=parsed_args.ip_addresses,
|
|
credentials=credentials,
|
|
kernel_name=deploy_kernel,
|
|
ramdisk_name=deploy_ramdisk,
|
|
instance_boot_option=parsed_args.instance_boot_option,
|
|
**kwargs
|
|
)
|
|
|
|
nodes_uuids = [node.uuid for node in nodes]
|
|
|
|
if parsed_args.introspect:
|
|
baremetal.introspect(
|
|
self.app.client_manager,
|
|
node_uuids=nodes_uuids,
|
|
run_validations=parsed_args.run_validations,
|
|
concurrency=parsed_args.concurrency,
|
|
node_timeout=parsed_args.node_timeout,
|
|
max_retries=parsed_args.max_retries,
|
|
retry_timeout=parsed_args.retry_timeout,
|
|
)
|
|
|
|
if parsed_args.provide:
|
|
provide = tb.TripleoProvide(verbosity=parsed_args.verbosity)
|
|
provide.provide(nodes=nodes_uuids)
|
|
|
|
|
|
class ExtractProvisionedNode(command.Command):
|
|
|
|
log = logging.getLogger(__name__ + ".ExtractProvisionedNode")
|
|
|
|
def _setup_clients(self):
|
|
self.clients = self.app.client_manager
|
|
self.orchestration_client = self.clients.orchestration
|
|
self.baremetal_client = self.clients.baremetal
|
|
self.network_client = self.clients.network
|
|
|
|
def get_parser(self, prog_name):
|
|
parser = super(ExtractProvisionedNode, self).get_parser(prog_name)
|
|
parser.add_argument('--stack', dest='stack',
|
|
help=_('Name or ID of heat stack '
|
|
'(default=Env: OVERCLOUD_STACK_NAME)'),
|
|
default=utils.env('OVERCLOUD_STACK_NAME',
|
|
default='overcloud'))
|
|
parser.add_argument('-o', '--output',
|
|
metavar='<baremetal_deployment.yaml>',
|
|
help=_('The output file path describing the '
|
|
'baremetal deployment'))
|
|
parser.add_argument('-y', '--yes', default=False, action='store_true',
|
|
help=_('Skip yes/no prompt for existing files '
|
|
'(assume yes).'))
|
|
parser.add_argument('--roles-file', '-r', dest='roles_file',
|
|
required=False,
|
|
help=_('Role data definition file'))
|
|
return parser
|
|
|
|
def _get_subnet_from_net_name_and_ip(self, net_name, ip_addr):
|
|
try:
|
|
network = self.network_client.find_network(net_name)
|
|
except openstack_exc.DuplicateResource:
|
|
raise oscexc.CommandError(
|
|
"Unable to extract role networks. Duplicate network resources "
|
|
"with name %s detected." % net_name)
|
|
|
|
if network is None:
|
|
raise oscexc.CommandError("Unable to extract role networks. "
|
|
"Network %s not found." % net_name)
|
|
|
|
for subnet_id in network.subnet_ids:
|
|
subnet = self.network_client.get_subnet(subnet_id)
|
|
if (ipaddress.ip_address(ip_addr)
|
|
in ipaddress.ip_network(subnet.cidr)):
|
|
subnet_name = subnet.name
|
|
return subnet_name
|
|
|
|
raise oscexc.CommandError("Unable to extract role networks. Could not "
|
|
"find subnet for IP address %(ip)s on "
|
|
"network %(net)s." % {'ip': ip_addr,
|
|
'net': net_name})
|
|
|
|
def take_action(self, parsed_args):
|
|
self.log.debug("take_action(%s)" % parsed_args)
|
|
|
|
self._setup_clients()
|
|
stack = oooutils.get_stack(self.orchestration_client,
|
|
parsed_args.stack)
|
|
tht_j2_sources = oooutils.get_stack_output_item(
|
|
stack, 'TripleoHeatTemplatesJinja2RenderingDataSources') or {}
|
|
|
|
if parsed_args.roles_file:
|
|
roles_file = os.path.abspath(parsed_args.roles_file)
|
|
with open(roles_file, 'r') as fd:
|
|
role_data = yaml.safe_load(fd.read())
|
|
else:
|
|
role_data = tht_j2_sources.get('roles_data')
|
|
if role_data is None:
|
|
raise oscexc.CommandError(
|
|
"Unable to extract. Role data not available in {} stack "
|
|
"output. Please provide the roles data for the deployed "
|
|
"stack by setting the --roles-data argument.".format(
|
|
parsed_args.stack))
|
|
|
|
# Convert role_data to a dict
|
|
role_data = {x['name']: x for x in role_data}
|
|
|
|
host_vars = oooutils.get_stack_output_item(
|
|
stack, 'AnsibleHostVarsMap') or {}
|
|
role_net_ip_map = oooutils.get_stack_output_item(
|
|
stack, 'RoleNetIpMap') or {}
|
|
parameters = stack.to_dict().get('parameters', {})
|
|
parameter_defaults = stack.environment().get('parameter_defaults', {})
|
|
|
|
# list all baremetal nodes and map hostname to node name
|
|
node_details = self.baremetal_client.node.list(detail=True)
|
|
hostname_node_map = {}
|
|
hostname_node_resource = {}
|
|
for node in node_details:
|
|
hostname = node.instance_info.get('display_name')
|
|
if hostname:
|
|
hostname_node_map[hostname] = node.id
|
|
if hostname and node.resource_class:
|
|
hostname_node_resource[hostname] = node.resource_class
|
|
|
|
data = []
|
|
warnings = []
|
|
for role_name, entries in host_vars.items():
|
|
role_count = len(entries)
|
|
|
|
# skip zero count roles
|
|
if not role_count:
|
|
continue
|
|
|
|
if role_name not in role_data:
|
|
raise oscexc.CommandError(
|
|
"Unable to extract. Invalid role file. Role {} is not "
|
|
"defined in roles file {}".format(role_name, roles_file))
|
|
|
|
role = collections.OrderedDict()
|
|
role['name'] = role_name
|
|
role['count'] = role_count
|
|
|
|
hostname_format = parameters.get('%sHostnameFormat' % role_name)
|
|
if hostname_format:
|
|
role['hostname_format'] = hostname_format
|
|
|
|
defaults = role['defaults'] = {}
|
|
|
|
# Add networks to the role default section
|
|
role_networks = defaults['networks'] = []
|
|
for net_name, ips in role_net_ip_map[role_name].items():
|
|
subnet_name = self._get_subnet_from_net_name_and_ip(net_name,
|
|
ips[0])
|
|
if net_name == constants.CTLPLANE_NET_NAME:
|
|
role_networks.append({'network': net_name,
|
|
'vif': True})
|
|
else:
|
|
role_networks.append({'network': net_name,
|
|
'subnet': subnet_name})
|
|
|
|
# Add network config to role defaults section
|
|
net_conf = defaults['network_config'] = {}
|
|
net_conf['template'] = parameters.get(
|
|
role_name + 'NetworkConfigTemplate')
|
|
if net_conf['template'] is None:
|
|
warnings.append(
|
|
'WARNING: No network config found for role {}. Please '
|
|
'edit the file and set the path to the correct network '
|
|
'config template.'.format(role_name))
|
|
|
|
if parameters.get(role_name + 'NetworkDeploymentActions'):
|
|
network_deployment_actions = parameters.get(
|
|
role_name + 'NetworkDeploymentActions')
|
|
else:
|
|
network_deployment_actions = parameters.get(
|
|
'NetworkDeploymentActions', ['CREATE'])
|
|
|
|
net_conf['network_config_update'] = (
|
|
'UPDATE' in network_deployment_actions)
|
|
|
|
# The NetConfigDataLookup parameter is of type: json, but when
|
|
# not set it returns as string '{}'
|
|
ncdl = parameters.get('NetConfigDataLookup')
|
|
if isinstance(ncdl, str):
|
|
ncdl = json.loads(ncdl)
|
|
if ncdl:
|
|
net_conf['net_config_data_lookup'] = ncdl
|
|
|
|
if parameters.get('DnsSearchDomains'):
|
|
net_conf['dns_search_domains'] = parameters.get(
|
|
'DnsSearchDomains')
|
|
|
|
net_conf['physical_bridge_name'] = parameters.get(
|
|
'NeutronPhysicalBridge', 'br-ex')
|
|
net_conf['public_interface_name'] = parameters.get(
|
|
'NeutronPublicInterface', 'nic1')
|
|
|
|
if role_data[role_name].get('default_route_networks'):
|
|
net_conf['default_route_network'] = role_data[role_name].get(
|
|
'default_route_networks')
|
|
if role_data[role_name].get('networks_skip_config'):
|
|
net_conf['networks_skip_config'] = role_data[role_name].get(
|
|
'networks_skip_config')
|
|
|
|
# Add individual instances
|
|
ips_from_pool = parameter_defaults.get(
|
|
'{}IPs'.format(role_name), {})
|
|
instances = role['instances'] = []
|
|
for idx, entry in enumerate(sorted(entries)):
|
|
instance = {'hostname': entry}
|
|
|
|
if entry in hostname_node_map:
|
|
instance['name'] = hostname_node_map[entry]
|
|
|
|
if entry in hostname_node_resource:
|
|
instance['resource_class'] = hostname_node_resource[entry]
|
|
|
|
if ips_from_pool:
|
|
instance['networks'] = copy.deepcopy(role_networks)
|
|
for net in instance['networks']:
|
|
net['fixed_ip'] = (
|
|
role_net_ip_map[role_name][net['network']][idx])
|
|
|
|
instances.append(instance)
|
|
|
|
data.append(role)
|
|
|
|
# Write the file header
|
|
file_data = StringIO()
|
|
file_data.write('# Generated with the following on %s\n#\n' %
|
|
datetime.datetime.now().isoformat())
|
|
file_data.write('# openstack %s\n#\n\n' %
|
|
' '.join(self.app.command_options))
|
|
# Write any warnings in the file header
|
|
for warning in warnings:
|
|
file_data.write('# {}\n'.format(warning))
|
|
if warnings:
|
|
file_data.write(('#\n\n'))
|
|
# Write the data
|
|
if data:
|
|
yaml.dump(data, file_data, RoleDataDumper, width=120,
|
|
default_flow_style=False)
|
|
|
|
if parsed_args.output:
|
|
if (os.path.exists(parsed_args.output)
|
|
and not parsed_args.yes and sys.stdin.isatty()):
|
|
prompt_response = input(
|
|
('Overwrite existing file %s [y/N]?' % parsed_args.output)
|
|
).lower()
|
|
if not prompt_response.startswith('y'):
|
|
raise oscexc.CommandError(
|
|
"Will not overwrite existing file:"
|
|
" %s" % parsed_args.output)
|
|
with open(parsed_args.output, 'w+') as fp:
|
|
fp.write(file_data.getvalue())
|
|
self.app.stdout.write(file_data.getvalue())
|
|
|
|
|
|
class RoleDataDumper(yaml.SafeDumper):
|
|
def represent_ordered_dict(self, data):
|
|
return self.represent_dict(data.items())
|
|
|
|
|
|
RoleDataDumper.add_representer(collections.OrderedDict,
|
|
RoleDataDumper.represent_ordered_dict)
|