# Copyright 2015 Red Hat, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. # import collections import copy import datetime from io import StringIO import ipaddress import json import logging import os import sys import time from cliff.formatters import table from openstack import exceptions as openstack_exc from osc_lib import exceptions as oscexc from osc_lib.i18n import _ from osc_lib import utils import tempfile import yaml from tripleoclient import command from tripleoclient import constants from tripleoclient import exceptions from tripleoclient import utils as oooutils from tripleoclient.workflows import baremetal from tripleoclient.workflows import tripleo_baremetal as tb class DeleteNode(command.Command): """Delete overcloud nodes.""" log = logging.getLogger(__name__ + ".DeleteNode") def get_parser(self, prog_name): parser = super(DeleteNode, self).get_parser(prog_name) group = parser.add_mutually_exclusive_group(required=True) group.add_argument('nodes', metavar='', nargs="*", default=[], help=_('Node ID(s) to delete (otherwise specified ' 'in the --baremetal-deployment file)')) group.add_argument('-b', '--baremetal-deployment', metavar='', help=_('Configuration file describing the ' 'baremetal deployment')) parser.add_argument('--stack', dest='stack', help=_('Name or ID of heat stack to scale ' '(default=Env: OVERCLOUD_STACK_NAME)'), default=utils.env('OVERCLOUD_STACK_NAME', default='overcloud')) parser.add_argument( '--timeout', metavar='', type=int, default=constants.STACK_TIMEOUT, dest='timeout', help=_("Timeout in minutes to wait for the nodes to be deleted. " "Keep in mind that due to keystone session duration " "that timeout has an upper bound of 4 hours ") ) parser.add_argument( '--overcloud-ssh-port-timeout', help=_('Timeout for the ssh port to become active.'), type=int, default=constants.ENABLE_SSH_ADMIN_SSH_PORT_TIMEOUT ) parser.add_argument('-y', '--yes', help=_('Skip yes/no prompt (assume yes).'), default=False, action="store_true") return parser def _nodes_to_delete(self, parsed_args, roles): with oooutils.TempDirs() as tmp: unprovision_confirm = os.path.join( tmp, 'unprovision_confirm.json') oooutils.run_ansible_playbook( playbook='cli-overcloud-node-unprovision.yaml', inventory='localhost,', workdir=tmp, timeout=parsed_args.timeout, playbook_dir=constants.ANSIBLE_TRIPLEO_PLAYBOOKS, verbosity=oooutils.playbook_verbosity(self=self), extra_vars={ "stack_name": parsed_args.stack, "baremetal_deployment": roles, "prompt": True, "unprovision_confirm": unprovision_confirm, } ) with open(unprovision_confirm) as f: to_unprovision = json.load(f) if isinstance(to_unprovision, dict): nodes = to_unprovision.get( 'instances') + to_unprovision.get('pre_provisioned') else: nodes = to_unprovision if not nodes: print('No nodes to unprovision') return None, None TableArgs = collections.namedtuple( 'TableArgs', 'print_empty max_width fit_width') args = TableArgs(print_empty=True, max_width=-1, fit_width=True) nodes_data = [(i.get('hostname', ''), i.get('name', ''), i.get('id', '')) for i in nodes] node_hostnames = [i['hostname'] for i in nodes if 'hostname' in i] formatter = table.TableFormatter() output = StringIO() formatter.emit_list( column_names=['hostname', 'name', 'id'], data=nodes_data, stdout=output, parsed_args=args ) return output.getvalue(), node_hostnames def _check_skiplist_exists(self, env): skiplist = env.get('parameter_defaults', {}).get('DeploymentServerBlacklist') if skiplist: self.log.warning(_('[WARNING] DeploymentServerBlacklist is ' 'ignored when executing scale down actions. If ' 'the node(s) being removed should *NOT* have ' 'any actions executed on them, please shut ' 'them off prior to their removal.')) def _check_timeout(self, start, timeout): used = int((time.time() - start) // 60) remaining = timeout - used if remaining <= 0: raise exceptions.DeploymentError( 'Deployment timed out after %sm' % used ) return remaining def take_action(self, parsed_args): self.log.debug("take_action(%s)" % parsed_args) # Start our timer. This will be used to calculate the timeout. start = time.time() if parsed_args.baremetal_deployment: with open(parsed_args.baremetal_deployment, 'r') as fp: roles = yaml.safe_load(fp) nodes_text, nodes = self._nodes_to_delete(parsed_args, roles) if nodes_text: print(nodes_text) else: return else: nodes = parsed_args.nodes nodes_text = '\n'.join('- %s' % node for node in nodes) if not parsed_args.yes: confirm = oooutils.prompt_user_for_confirmation( message=_("Are you sure you want to delete these overcloud " "nodes [y/N]? "), logger=self.log) if not confirm: raise oscexc.CommandError("Action not confirmed, exiting.") ansible_dir = os.path.join(oooutils.get_default_working_dir( parsed_args.stack ), 'config-download', parsed_args.stack) inventory = os.path.join(ansible_dir, 'tripleo-ansible-inventory.yaml') ansible_cfg = os.path.join(ansible_dir, 'ansible.cfg') key_file = oooutils.get_key(parsed_args.stack) remaining = self._check_timeout(start, parsed_args.timeout) oooutils.run_ansible_playbook( playbook='scale_playbook.yaml', inventory=inventory, workdir=ansible_dir, playbook_dir=ansible_dir, ansible_cfg=ansible_cfg, ssh_user='tripleo-admin', limit_hosts=':'.join('%s' % node for node in nodes), reproduce_command=True, ignore_unreachable=True, timeout=remaining, extra_env_variables={ "ANSIBLE_BECOME": True, "ANSIBLE_PRIVATE_KEY_FILE": key_file } ) remaining = self._check_timeout(start, parsed_args.timeout) if parsed_args.baremetal_deployment: with oooutils.TempDirs() as tmp: oooutils.run_ansible_playbook( playbook='cli-overcloud-node-unprovision.yaml', inventory='localhost,', workdir=tmp, playbook_dir=constants.ANSIBLE_TRIPLEO_PLAYBOOKS, timeout=remaining, verbosity=oooutils.playbook_verbosity(self=self), extra_vars={ "stack_name": parsed_args.stack, "baremetal_deployment": roles, "prompt": False, "manage_network_ports": True, } ) class ProvideNode(command.Command): """Mark nodes as available based on UUIDs or current 'manageable' state.""" log = logging.getLogger(__name__ + ".ProvideNode") def get_parser(self, prog_name): parser = super(ProvideNode, self).get_parser(prog_name) group = parser.add_mutually_exclusive_group(required=True) group.add_argument('node_uuids', nargs="*", metavar="", default=[], help=_('Baremetal Node UUIDs for the node(s) to be ' 'provided')) group.add_argument("--all-manageable", action='store_true', help=_("Provide all nodes currently in 'manageable'" " state")) group.add_argument("--verbosity", type=int, default=1, help=_("Print debug output during execution")) return parser def take_action(self, parsed_args): self.log.debug("take_action(%s)" % parsed_args) provide = tb.TripleoProvide(verbosity=parsed_args.verbosity) if parsed_args.node_uuids: provide.provide(nodes=parsed_args.node_uuids) else: provide.provide_manageable_nodes() class CleanNode(command.Command): """Run node(s) through cleaning.""" log = logging.getLogger(__name__ + ".CleanNode") def get_parser(self, prog_name): parser = super(CleanNode, self).get_parser(prog_name) group = parser.add_mutually_exclusive_group(required=True) group.add_argument('node_uuids', nargs="*", metavar="", default=[], help=_('Baremetal Node UUIDs for the node(s) to be ' 'cleaned')) group.add_argument("--all-manageable", action='store_true', help=_("Clean all nodes currently in 'manageable'" " state")) group.add_argument("--verbosity", type=int, default=1, help=_("Print debug output during execution")) parser.add_argument('--provide', action='store_true', help=_('Provide (make available) the nodes once ' 'cleaned')) return parser def take_action(self, parsed_args): self.log.debug("take_action(%s)" % parsed_args) nodes = parsed_args.node_uuids clean = tb.TripleoClean(verbosity=parsed_args.verbosity) if nodes: clean.clean( nodes=parsed_args.node_uuids) else: clean.clean_manageable_nodes() if parsed_args.provide: provide = tb.TripleoProvide(verbosity=parsed_args.verbosity) if nodes: provide.provide(nodes=nodes) else: provide.provide_manageable_nodes() class ConfigureNode(command.Command): """Configure Node boot options.""" log = logging.getLogger(__name__ + ".ConfigureNode") def get_parser(self, prog_name): parser = super(ConfigureNode, self).get_parser(prog_name) group = parser.add_mutually_exclusive_group(required=True) group.add_argument('node_uuids', nargs="*", metavar="", default=[], help=_('Baremetal Node UUIDs for the node(s) to be ' 'configured')) group.add_argument("--all-manageable", action='store_true', help=_("Configure all nodes currently in " "'manageable' state")) parser.add_argument( '--deploy-kernel', default='file://%s/agent.kernel' % constants.IRONIC_HTTP_BOOT_BIND_MOUNT, help=_('Image with deploy kernel.')) parser.add_argument( '--deploy-ramdisk', default='file://%s/agent.ramdisk' % constants.IRONIC_HTTP_BOOT_BIND_MOUNT, help=_('Image with deploy ramdisk.')) parser.add_argument('--instance-boot-option', choices=['local', 'netboot'], help=_('Whether to set instances for booting from ' 'local hard drive (local) or network ' '(netboot).')) parser.add_argument('--boot-mode', choices=['uefi', 'bios'], help=_('Whether to set the boot mode to UEFI ' '(uefi) or legacy BIOS (bios)')) parser.add_argument('--root-device', help=_('Define the root device for nodes. ' 'Can be either a list of device names ' '(without /dev) to choose from or one of ' 'two strategies: largest or smallest. For ' 'it to work this command should be run ' 'after the introspection.')) parser.add_argument('--root-device-minimum-size', type=int, default=4, help=_('Minimum size (in GiB) of the detected ' 'root device. Used with --root-device.')) parser.add_argument('--overwrite-root-device-hints', action='store_true', help=_('Whether to overwrite existing root device ' 'hints when --root-device is used.')) parser.add_argument("--verbosity", type=int, default=1, help=_("Print debug output during execution")) return parser def take_action(self, parsed_args): self.log.debug("take_action(%s)" % parsed_args) conf = tb.TripleoConfigure( kernel_name=parsed_args.deploy_kernel, ramdisk_name=parsed_args.deploy_ramdisk, instance_boot_option=parsed_args.instance_boot_option, boot_mode=parsed_args.boot_mode, root_device=parsed_args.root_device, root_device_minimum_size=parsed_args.root_device_minimum_size, overwrite_root_device_hints=( parsed_args.overwrite_root_device_hints) ) if parsed_args.node_uuids: conf.configure( node_uuids=parsed_args.node_uuids) else: conf.configure_manageable_nodes() class DiscoverNode(command.Command): """Discover overcloud nodes by polling their BMCs.""" log = logging.getLogger(__name__ + ".DiscoverNode") def get_parser(self, prog_name): parser = super(DiscoverNode, self).get_parser(prog_name) ip_group = parser.add_mutually_exclusive_group(required=True) ip_group.add_argument('--ip', action='append', dest='ip_addresses', metavar='', help=_('IP address(es) to probe')) ip_group.add_argument('--range', dest='ip_addresses', metavar='', help=_('IP range to probe')) parser.add_argument('--credentials', metavar='', action='append', required=True, help=_('Key/value pairs of possible credentials')) parser.add_argument('--port', action='append', metavar='', type=int, help=_('BMC port(s) to probe')) parser.add_argument('--introspect', action='store_true', help=_('Introspect the imported nodes')) parser.add_argument('--run-validations', action='store_true', default=False, help=_('Run the pre-deployment validations. These ' 'external validations are from the TripleO ' 'Validations project.')) parser.add_argument('--provide', action='store_true', help=_('Provide (make available) the nodes')) parser.add_argument('--no-deploy-image', action='store_true', help=_('Skip setting the deploy kernel and ' 'ramdisk.')) parser.add_argument('--instance-boot-option', choices=['local', 'netboot'], default='local', help=_('Whether to set instances for booting from ' 'local hard drive (local) or network ' '(netboot).')) parser.add_argument('--concurrency', type=int, default=20, help=_('Maximum number of nodes to introspect at ' 'once.')) parser.add_argument('--node-timeout', type=int, default=1200, help=_('Maximum timeout for node introspection.')) parser.add_argument('--max-retries', type=int, default=1, help=_('Maximum introspection retries.')) parser.add_argument('--retry-timeout', type=int, default=120, help=_('Maximum timeout between introspection' 'retries')) parser.add_argument("--verbosity", type=int, default=1, help=_("Print debug output during execution")) return parser # FIXME(tonyb): This is not multi-arch safe :( def take_action(self, parsed_args): self.log.debug("take_action(%s)" % parsed_args) if parsed_args.no_deploy_image: deploy_kernel = None deploy_ramdisk = None else: deploy_kernel = 'file://{}/agent.kernel'.format( constants.IRONIC_HTTP_BOOT_BIND_MOUNT ) deploy_ramdisk = 'file://{}/agent.ramdisk'.format( constants.IRONIC_HTTP_BOOT_BIND_MOUNT ) credentials = [list(x.split(':', 1)) for x in parsed_args.credentials] kwargs = {} # Leave it up to the workflow to figure out the defaults if parsed_args.port: kwargs['ports'] = parsed_args.port nodes = baremetal.discover_and_enroll( self.app.client_manager, ip_addresses=parsed_args.ip_addresses, credentials=credentials, kernel_name=deploy_kernel, ramdisk_name=deploy_ramdisk, instance_boot_option=parsed_args.instance_boot_option, **kwargs ) nodes_uuids = [node.uuid for node in nodes] if parsed_args.introspect: baremetal.introspect( self.app.client_manager, node_uuids=nodes_uuids, run_validations=parsed_args.run_validations, concurrency=parsed_args.concurrency, node_timeout=parsed_args.node_timeout, max_retries=parsed_args.max_retries, retry_timeout=parsed_args.retry_timeout, ) if parsed_args.provide: provide = tb.TripleoProvide(verbosity=parsed_args.verbosity) provide.provide(nodes=nodes_uuids) class ExtractProvisionedNode(command.Command): log = logging.getLogger(__name__ + ".ExtractProvisionedNode") def _setup_clients(self): self.clients = self.app.client_manager self.orchestration_client = self.clients.orchestration self.baremetal_client = self.clients.baremetal self.network_client = self.clients.network def get_parser(self, prog_name): parser = super(ExtractProvisionedNode, self).get_parser(prog_name) parser.add_argument('--stack', dest='stack', help=_('Name or ID of heat stack ' '(default=Env: OVERCLOUD_STACK_NAME)'), default=utils.env('OVERCLOUD_STACK_NAME', default='overcloud')) parser.add_argument('-o', '--output', metavar='', help=_('The output file path describing the ' 'baremetal deployment')) parser.add_argument('-y', '--yes', default=False, action='store_true', help=_('Skip yes/no prompt for existing files ' '(assume yes).')) parser.add_argument('--roles-file', '-r', dest='roles_file', required=False, help=_('Role data definition file')) parser.add_argument('--networks-file', '-n', dest='networks_file', required=False, help=_('Network data definition file')) parser.add_argument('--working-dir', action='store', help=_('The working directory for the deployment ' 'where all input, output, and generated ' 'files will be stored.\nDefaults to ' '"$HOME/overcloud-deploy/"')) return parser def _get_subnet_from_net_name_and_ip(self, net_name, ip_addr): try: network = self.network_client.find_network(net_name) except openstack_exc.DuplicateResource: raise oscexc.CommandError( "Unable to extract role networks. Duplicate network resources " "with name %s detected." % net_name) if network is None: raise oscexc.CommandError("Unable to extract role networks. " "Network %s not found." % net_name) for subnet_id in network.subnet_ids: subnet = self.network_client.get_subnet(subnet_id) if (ipaddress.ip_address(ip_addr) in ipaddress.ip_network(subnet.cidr)): subnet_name = subnet.name return subnet_name raise oscexc.CommandError("Unable to extract role networks. Could not " "find subnet for IP address %(ip)s on " "network %(net)s." % {'ip': ip_addr, 'net': net_name}) def _convert_heat_nic_conf_to_j2(self, stack, role_name, network_data, resource_registry, parsed_args): heat_nic_conf = resource_registry.get( 'OS::TripleO::{}::Net::SoftwareConfig'.format(role_name)) if heat_nic_conf is None or heat_nic_conf == 'OS::Heat::None': return None j2_nic_conf_dir = os.path.join(self.working_dir, 'nic-configs') oooutils.makedirs(j2_nic_conf_dir) heat_nic_conf_basename = os.path.basename(heat_nic_conf) tmp_heat_nic_conf_path = os.path.join(j2_nic_conf_dir, heat_nic_conf_basename) heat_nic_conf_content = stack.files().get(heat_nic_conf) j2_nic_conf_basename = (heat_nic_conf_basename.rstrip('.yaml') + '.j2') j2_nic_conf_path = os.path.join(j2_nic_conf_dir, j2_nic_conf_basename) tmp_net_data_fd, tmp_net_data_path = tempfile.mkstemp(suffix=".yaml") try: with open(tmp_net_data_path, 'w') as tmp_net_data: tmp_net_data.write(yaml.safe_dump(network_data)) with open(tmp_heat_nic_conf_path, 'w') as tmp_heat_nic_conf: tmp_heat_nic_conf.write(heat_nic_conf_content) cmd = ['/usr/share/openstack-tripleo-heat-templates/tools/' 'convert_heat_nic_config_to_ansible_j2.py'] if parsed_args.yes: cmd.extend(['--yes']) cmd.extend(['--stack', stack.stack_name, '--networks_file', tmp_net_data_path, tmp_heat_nic_conf_path]) retcode = oooutils.run_command_and_log(self.log, cmd) finally: try: os.remove(tmp_net_data_path) except (IsADirectoryError, FileNotFoundError, PermissionError): pass try: os.remove(tmp_heat_nic_conf_path) except (IsADirectoryError, FileNotFoundError, PermissionError): pass return j2_nic_conf_path if retcode == 0 else None def take_action(self, parsed_args): self.log.debug("take_action(%s)" % parsed_args) if not parsed_args.working_dir: self.working_dir = oooutils.get_default_working_dir( parsed_args.stack) else: self.working_dir = parsed_args.working_dir oooutils.makedirs(self.working_dir) self._setup_clients() stack = oooutils.get_stack(self.orchestration_client, parsed_args.stack) tht_j2_sources = oooutils.get_stack_output_item( stack, 'TripleoHeatTemplatesJinja2RenderingDataSources') or {} if parsed_args.roles_file: roles_file = os.path.abspath(parsed_args.roles_file) with open(roles_file, 'r') as fd: role_data = yaml.safe_load(fd.read()) else: role_data = tht_j2_sources.get('roles_data') if role_data is None: raise oscexc.CommandError( "Unable to extract. Role data not available in {} stack " "output. Please provide the roles data for the deployed " "stack by setting the --roles-data argument.".format( parsed_args.stack)) if parsed_args.networks_file: networks_file = os.path.abspath(parsed_args.networks_file) with open(networks_file, 'r') as fd: network_data = yaml.safe_load(fd.read()) else: network_data = tht_j2_sources.get('networks_data') if network_data is None: raise oscexc.CommandError( "Unable to extract. Network data not available in {} " "stack output. Please provide the networks data for the " "deployed stack by setting the --networks-data argument." .format(parsed_args.stack)) # Convert role_data to a dict role_data = {x['name']: x for x in role_data} host_vars = oooutils.get_stack_output_item( stack, 'AnsibleHostVarsMap') or {} role_net_ip_map = oooutils.get_stack_output_item( stack, 'RoleNetIpMap') or {} parameters = stack.to_dict().get('parameters', {}) parameter_defaults = stack.environment().get('parameter_defaults', {}) resource_registry = stack.environment().get('resource_registry', {}) # list all baremetal nodes and map hostname to node name node_details = self.baremetal_client.node.list(detail=True) hostname_node_map = {} hostname_node_resource = {} for node in node_details: hostname = node.instance_info.get('display_name') if hostname: hostname_node_map[hostname] = node.id if hostname and node.resource_class: hostname_node_resource[hostname] = node.resource_class data = [] warnings = [] network_environment = {'resource_registry': {}, 'parameter_defaults': {}} for role_name, entries in host_vars.items(): role_count = len(entries) net_parameter_defaults = {} net_resource_registry = { "OS::TripleO::{}::Net::SoftwareConfig".format(role_name): "OS::Heat::None"} network_environment['resource_registry'].update( net_resource_registry) # skip zero count roles if not role_count: continue if role_name not in role_data: raise oscexc.CommandError( "Unable to extract. Invalid role file. Role {} is not " "defined in roles file {}".format(role_name, roles_file)) role = collections.OrderedDict() role['name'] = role_name role['count'] = role_count hostname_format = parameters.get('%sHostnameFormat' % role_name) if hostname_format: role['hostname_format'] = hostname_format defaults = role['defaults'] = {} # Add networks to the role default section role_networks = defaults['networks'] = [] for net_name, ips in role_net_ip_map[role_name].items(): subnet_name = self._get_subnet_from_net_name_and_ip(net_name, ips[0]) if net_name == constants.CTLPLANE_NET_NAME: role_networks.append({'network': net_name, 'vif': True}) else: role_networks.append({'network': net_name, 'subnet': subnet_name}) # Add network config to role defaults section net_conf = defaults['network_config'] = {} net_conf['template'] = parameters.get( role_name + 'NetworkConfigTemplate') if net_conf['template'] is None: net_conf['template'] = self._convert_heat_nic_conf_to_j2( stack, role_name, network_data, resource_registry, parsed_args) if net_conf['template'] is None: warnings.append( 'WARNING: No network config found for role {}. ' 'Please edit the file and set the path to the correct ' 'network config template.'.format(role_name)) else: net_parameter_defaults = { "{}NetworkConfigTemplate".format(role_name): "{}".format(net_conf['template'])} warnings.append( 'WARNING: Network config for role {} was ' 'automatically converted from Heat template to ' 'Ansible Jinja2 template. Please review the file: {}' .format(role_name, net_conf['template'])) if parameters.get(role_name + 'NetworkDeploymentActions'): network_deployment_actions = parameters.get( role_name + 'NetworkDeploymentActions') else: network_deployment_actions = parameters.get( 'NetworkDeploymentActions', ['CREATE']) net_conf['network_config_update'] = ( 'UPDATE' in network_deployment_actions) # The NetConfigDataLookup parameter is of type: json, but when # not set it returns as string '{}' ncdl = parameters.get('NetConfigDataLookup') if isinstance(ncdl, str): ncdl = json.loads(ncdl) if ncdl: net_conf['net_config_data_lookup'] = ncdl if parameters.get('DnsSearchDomains'): net_conf['dns_search_domains'] = parameters.get( 'DnsSearchDomains') net_conf['physical_bridge_name'] = parameters.get( 'NeutronPhysicalBridge', 'br-ex') net_conf['public_interface_name'] = parameters.get( 'NeutronPublicInterface', 'nic1') if role_data[role_name].get('default_route_networks'): net_conf['default_route_network'] = role_data[role_name].get( 'default_route_networks') if role_data[role_name].get('networks_skip_config'): net_conf['networks_skip_config'] = role_data[role_name].get( 'networks_skip_config') # Add individual instances ips_from_pool = parameter_defaults.get( '{}IPs'.format(role_name), {}) instances = role['instances'] = [] for idx, entry in enumerate(sorted(entries)): instance = {'hostname': entry} if entry in hostname_node_map: instance['name'] = hostname_node_map[entry] if entry in hostname_node_resource: instance['resource_class'] = hostname_node_resource[entry] if ips_from_pool: instance['networks'] = copy.deepcopy(role_networks) for net in instance['networks']: net['fixed_ip'] = ( role_net_ip_map[role_name][net['network']][idx]) instances.append(instance) if net_parameter_defaults != {}: network_environment['parameter_defaults'].update( net_parameter_defaults) data.append(role) # Write the file header file_data = StringIO() file_data.write('# Generated with the following on %s\n#\n' % datetime.datetime.now().isoformat()) file_data.write('# openstack %s\n#\n\n' % ' '.join(self.app.command_options)) # Write any warnings in the file header for warning in warnings: file_data.write('# {}\n'.format(warning)) if warnings: file_data.write(('#\n\n')) # Write the data if data: yaml.dump(data, file_data, RoleDataDumper, width=120, default_flow_style=False) if len(network_environment['parameter_defaults']) > 0: net_env_file = os.path.join(self.working_dir, "{}-network-environment.yaml".format( parsed_args.stack)) with open(net_env_file, 'w+') as nfp: nfp.write(yaml.dump(network_environment, width=120, default_flow_style=False)) if parsed_args.output: if (os.path.exists(parsed_args.output) and not parsed_args.yes and sys.stdin.isatty()): prompt_response = input( ('Overwrite existing file %s [y/N]?' % parsed_args.output) ).lower() if not prompt_response.startswith('y'): raise oscexc.CommandError( "Will not overwrite existing file:" " %s" % parsed_args.output) with open(parsed_args.output, 'w+') as fp: fp.write(file_data.getvalue()) self.app.stdout.write(file_data.getvalue()) class RoleDataDumper(yaml.SafeDumper): def represent_ordered_dict(self, data): return self.represent_dict(data.items()) RoleDataDumper.add_representer(collections.OrderedDict, RoleDataDumper.represent_ordered_dict)