python-tripleoclient/tripleoclient/v1/baremetal.py

560 lines
21 KiB
Python

# Copyright 2015 Red Hat, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
#
from __future__ import print_function
import argparse
import csv
import json
import logging
import sys
import time
from cliff import command
from cliff import lister
from ironic_inspector_client import client as inspector_client
from openstackclient.common import utils as osc_utils
from os_cloud_config import nodes
from tripleoclient import exceptions
from tripleoclient import utils
def _csv_to_nodes_dict(nodes_csv):
"""Convert CSV to a list of dicts formatted for os_cloud_config
Given a CSV file in the format below, convert it into the
structure expected by os_could_config JSON files.
pm_type, pm_addr, pm_user, pm_password, mac
"""
data = []
for row in csv.reader(nodes_csv):
node = {
"pm_user": row[2],
"pm_addr": row[1],
"pm_password": row[3],
"pm_type": row[0],
"mac": [
row[4]
]
}
data.append(node)
return data
class ValidateInstackEnv(command.Command):
"""Validate `instackenv.json` which is used in `baremetal import`."""
auth_required = False
log = logging.getLogger(__name__ + ".ValidateInstackEnv")
def get_parser(self, prog_name):
parser = super(ValidateInstackEnv, self).get_parser(prog_name)
parser.add_argument(
'-f', '--file', dest='instackenv',
help="Path to the instackenv.json file.",
default='instackenv.json')
return parser
def take_action(self, parsed_args):
self.log.debug("take_action(%s)" % parsed_args)
self.error_count = 0
with open(parsed_args.instackenv, 'r') as net_file:
env_data = json.load(net_file)
maclist = []
baremetal_ips = []
for node in env_data['nodes']:
self.log.info("Checking node %s" % node['pm_addr'])
try:
if len(node['pm_password']) == 0:
self.log.error('ERROR: Password 0 length.')
self.error_count += 1
except Exception as e:
self.log.error('ERROR: Password does not exist: %s', e)
self.error_count += 1
try:
if len(node['pm_user']) == 0:
self.log.error('ERROR: User 0 length.')
self.error_count += 1
except Exception as e:
self.log.error('ERROR: User does not exist: %s', e)
self.error_count += 1
try:
if len(node['mac']) == 0:
self.log.error('ERROR: MAC address 0 length.')
self.error_count += 1
maclist.extend(node['mac'])
except Exception as e:
self.log.error('ERROR: MAC address does not exist: %s', e)
self.error_count += 1
if node['pm_type'] == "pxe_ssh":
self.log.debug("Identified virtual node")
if node['pm_type'] == "pxe_ipmitool":
self.log.debug("Identified baremetal node")
cmd = ('ipmitool -R 1 -I lanplus -H %s -U %s -P %s chassis '
'status' % (node['pm_addr'], node['pm_user'],
node['pm_password']))
self.log.debug("Executing: %s", cmd)
status = utils.run_shell(cmd)
if status != 0:
self.log.error('ERROR: ipmitool failed')
self.error_count += 1
baremetal_ips.append(node['pm_addr'])
if not utils.all_unique(baremetal_ips):
self.log.error('ERROR: Baremetals IPs are not all unique.')
self.error_count += 1
else:
self.log.debug('Baremetal IPs are all unique.')
if not utils.all_unique(maclist):
self.log.error('ERROR: MAC addresses are not all unique.')
self.error_count += 1
else:
self.log.debug('MAC addresses are all unique.')
if self.error_count == 0:
print('SUCCESS: found 0 errors')
else:
print('FAILURE: found %d errors' % self.error_count)
class ImportBaremetal(command.Command):
"""Import baremetal nodes from a JSON or CSV file"""
log = logging.getLogger(__name__ + ".ImportBaremetal")
def get_parser(self, prog_name):
parser = super(ImportBaremetal, self).get_parser(prog_name)
parser.add_argument('-s', '--service-host', dest='service_host',
help='Nova compute service host to register nodes '
'with')
parser.add_argument('--json', dest='json', action='store_true')
parser.add_argument('--csv', dest='csv', action='store_true')
parser.add_argument('file_in', type=argparse.FileType('r'))
return parser
def take_action(self, parsed_args):
self.log.debug("take_action(%s)" % parsed_args)
# We need JSON or CSV to be specified, not both.
if parsed_args.json == parsed_args.csv:
print("ERROR: Either --json or --csv needs to be specified.",
file=sys.stderr)
return
if parsed_args.json is True:
nodes_json = json.load(parsed_args.file_in)
if 'nodes' in nodes_json:
nodes_json = nodes_json['nodes']
else:
nodes_json = _csv_to_nodes_dict(parsed_args.file_in)
nodes.register_all_nodes(
parsed_args.service_host,
nodes_json,
client=self.app.client_manager.tripleoclient.baremetal(),
keystone_client=self.app.client_manager.identity)
class IntrospectionParser(object):
def get_parser(self, prog_name):
parser = super(IntrospectionParser, self).get_parser(prog_name)
parser.add_argument(
'--inspector-url',
default=osc_utils.env('INSPECTOR_URL', default=None),
help='inspector URL, defaults to localhost (env: INSPECTOR_URL).')
return parser
class StartBaremetalIntrospectionBulk(IntrospectionParser, command.Command):
"""Start bulk introspection on all baremetal nodes"""
log = logging.getLogger(__name__ + ".StartBaremetalIntrospectionBulk")
def get_parser(self, prog_name):
parser = super(
StartBaremetalIntrospectionBulk, self).get_parser(prog_name)
return parser
def take_action(self, parsed_args):
self.log.debug("take_action(%s)" % parsed_args)
client = self.app.client_manager.tripleoclient.baremetal()
auth_token = self.app.client_manager.auth_ref.auth_token
node_uuids = []
print("Setting available nodes to manageable...")
self.log.debug("Moving available nodes to manageable state.")
available_nodes = [node for node in client.node.list()
if node.provision_state == "available"]
for uuid in utils.set_nodes_state(client, available_nodes, 'manage',
'manageable'):
self.log.debug("Node {0} has been set to manageable.".format(uuid))
for node in client.node.list():
if node.provision_state != "manageable":
continue
node_uuids.append(node.uuid)
print("Starting introspection of node: {0}".format(node.uuid))
inspector_client.introspect(
node.uuid,
base_url=parsed_args.inspector_url,
auth_token=auth_token)
# NOTE(dtantsur): PXE firmware on virtual machines misbehaves when
# a lot of nodes start DHCPing simultaneously: it ignores NACK from
# DHCP server, tries to get the same address, then times out. Work
# around it by using sleep, anyway introspection takes much longer.
time.sleep(5)
print("Waiting for introspection to finish...")
has_errors = False
for uuid, status in utils.wait_for_node_introspection(
inspector_client, auth_token, parsed_args.inspector_url,
node_uuids):
if status['error'] is None:
print("Introspection for UUID {0} finished successfully."
.format(uuid))
else:
print("Introspection for UUID {0} finished with error: {1}"
.format(uuid, status['error']))
has_errors = True
clients = self.app.client_manager
baremetal_client = clients.tripleoclient.baremetal()
print("Setting manageable nodes to available...")
self.log.debug("Moving manageable nodes to available state.")
available_nodes = [node for node in client.node.list()
if node.provision_state == "manageable"]
for uuid in utils.set_nodes_state(
baremetal_client, baremetal_client.node.list(), 'provide',
'available', skipped_states=("available", "active")):
print("Node {0} has been set to available.".format(uuid))
if has_errors:
print("Introspection completed with errors.")
else:
print("Introspection completed.")
class StatusBaremetalIntrospectionBulk(IntrospectionParser, lister.Lister):
"""Get the status of all baremetal nodes"""
log = logging.getLogger(__name__ + ".StatusBaremetalIntrospectionBulk")
def take_action(self, parsed_args):
self.log.debug("take_action(%s)" % parsed_args)
client = self.app.client_manager.tripleoclient.baremetal()
auth_token = self.app.client_manager.auth_ref.auth_token
statuses = []
for node in client.node.list():
self.log.debug("Getting introspection status of Ironic node {0}"
.format(node.uuid))
statuses.append((node.uuid, inspector_client.get_status(
node.uuid,
base_url=parsed_args.inspector_url,
auth_token=auth_token)))
return (
("Node UUID", "Finished", "Error"),
list((node_uuid, status['finished'], status['error'])
for (node_uuid, status) in statuses)
)
class ConfigureReadyState(IntrospectionParser, command.Command):
"""Configure all baremetal nodes for enrollment"""
log = logging.getLogger(__name__ + ".ConfigureReadyState")
sleep_time = 15
loops = 120
def _configure_bios(self, nodes):
for node in nodes:
print("Configuring BIOS for node {0}".format(node.uuid))
self.bm_client.node.vendor_passthru(
node.uuid, 'configure_bios_settings', http_method='POST')
# NOTE(ifarkas): give the DRAC card some time to process the job
time.sleep(self.sleep_time)
def _configure_root_raid_volumes(self, nodes):
for node in nodes:
print("Configuring root RAID volume for node {0}"
.format(node.uuid))
self.bm_client.node.vendor_passthru(
node.uuid, 'create_raid_configuration',
{'create_root_volume': True, 'create_nonroot_volumes': False},
'POST')
# NOTE(ifarkas): give the DRAC card some time to process the job
time.sleep(self.sleep_time)
def _configure_nonroot_raid_volumes(self, nodes):
for node in nodes:
print("Configuring non-root RAID volume for node {0}"
.format(node.uuid))
self.bm_client.node.vendor_passthru(
node.uuid, 'create_raid_configuration',
{'create_root_volume': False, 'create_nonroot_volumes': True},
'POST')
# NOTE(ifarkas): give the DRAC card some time to process the job
time.sleep(self.sleep_time)
def _wait_for_drac_config_jobs(self, nodes):
for node in nodes:
print("Waiting for DRAC config jobs to finish on node {0}"
.format(node.uuid))
for _ in range(self.loops):
resp = self.bm_client.node.vendor_passthru(
node.uuid, 'list_unfinished_jobs', http_method='GET')
if not resp.unfinished_jobs:
break
time.sleep(self.sleep_time)
else:
msg = ("Timed out waiting for DRAC config jobs on node {0}"
.format(node.uuid))
raise exceptions.Timeout(msg)
def _delete_raid_volumes(self, nodes):
nodes_with_reboot_request = set()
for node in nodes:
print("Deleting RAID volumes on node {0}".format(node.uuid))
resp = self.bm_client.node.vendor_passthru(
node.uuid, 'list_virtual_disks', http_method='GET')
virtual_disks = resp.virtual_disks
changed_raid_controllers = set()
for disk in virtual_disks:
self.bm_client.node.vendor_passthru(
node.uuid, 'delete_virtual_disk',
{'virtual_disk': disk['id']}, 'POST')
changed_raid_controllers.add(disk['controller'])
if changed_raid_controllers:
nodes_with_reboot_request.add(node)
for controller in changed_raid_controllers:
self.bm_client.node.vendor_passthru(
node.uuid, 'apply_pending_raid_config',
{'raid_controller': controller}, 'POST')
# NOTE(ifarkas): give the DRAC card some time to process the job
time.sleep(self.sleep_time)
return nodes_with_reboot_request
def _change_power_state(self, nodes, target_power_state):
for node in nodes:
print("Changing power state on "
"node {0} to {1}".format(node.uuid, target_power_state))
self.bm_client.node.set_power_state(node.uuid, target_power_state)
def _run_introspection(self, nodes):
auth_token = self.app.client_manager.auth_ref.auth_token
node_uuids = []
for node in nodes:
print("Starting introspection on node {0}".format(node.uuid))
inspector_client.introspect(
node.uuid,
base_url=self.inspector_url,
auth_token=auth_token)
node_uuids.append(node.uuid)
print("Waiting for introspection to finish")
for uuid, status in utils.wait_for_node_introspection(
inspector_client, auth_token, self.inspector_url,
node_uuids):
if status['error'] is None:
print("Introspection for node {0} finished successfully."
.format(uuid))
else:
print("Introspection for node {0} finished with error: {1}"
.format(uuid, status['error']))
def get_parser(self, prog_name):
parser = super(ConfigureReadyState, self).get_parser(prog_name)
parser.add_argument('--delete-existing-raid-volumes',
dest='delete_raid_volumes', action='store_true')
return parser
def take_action(self, parsed_args):
self.log.debug("take_action(%s)" % parsed_args)
self.bm_client = (
self.app.client_manager.tripleoclient.baremetal())
self.inspector_url = parsed_args.inspector_url
drac_nodes = [node for node in self.bm_client.node.list(detail=True)
if 'drac' in node.driver]
if parsed_args.delete_raid_volumes:
changed_nodes = self._delete_raid_volumes(drac_nodes)
self._change_power_state(changed_nodes, 'reboot')
self._wait_for_drac_config_jobs(changed_nodes)
self._configure_root_raid_volumes(drac_nodes)
self._configure_bios(drac_nodes)
self._change_power_state(drac_nodes, 'reboot')
self._wait_for_drac_config_jobs(drac_nodes)
self._run_introspection(drac_nodes)
self._configure_nonroot_raid_volumes(drac_nodes)
self._change_power_state(drac_nodes, 'reboot')
self._wait_for_drac_config_jobs(drac_nodes)
self._change_power_state(drac_nodes, 'off')
class ConfigureBaremetalBoot(command.Command):
"""Configure baremetal boot for all nodes"""
log = logging.getLogger(__name__ + ".ConfigureBaremetalBoot")
loops = 12
sleep_time = 10
def get_parser(self, prog_name):
parser = super(ConfigureBaremetalBoot, self).get_parser(prog_name)
parser.add_argument('--deploy-kernel',
default='bm-deploy-kernel',
help='Image with deploy kernel.')
parser.add_argument('--deploy-ramdisk',
default='bm-deploy-ramdisk',
help='Image with deploy ramdisk.')
return parser
def take_action(self, parsed_args):
self.log.debug("take_action(%s)" % parsed_args)
bm_client = self.app.client_manager.tripleoclient.baremetal()
image_client = self.app.client_manager.image
try:
kernel_id = osc_utils.find_resource(
image_client.images, parsed_args.deploy_kernel).id
except AttributeError:
self.log.error("Please make sure that an image named \"%s\" exists"
" in Glance and is the only image with this name."
% parsed_args.deploy_ramdisk)
return
try:
ramdisk_id = osc_utils.find_resource(
image_client.images, parsed_args.deploy_ramdisk).id
except AttributeError:
self.log.error("Please make sure that an image named \"%s\" exists"
" in Glance and is the only image with this name."
% parsed_args.deploy_ramdisk)
return
self.log.debug("Using kernel ID: {0} and ramdisk ID: {1}".format(
kernel_id, ramdisk_id))
for node in bm_client.node.list(maintenance=False):
# NOTE(bnemec): Ironic won't let us update the node while the
# power_state is transitioning.
if node.power_state is None:
self.log.warning('Node %s power state is in transition. '
'Waiting up to %d seconds for it to '
'complete.',
node.uuid,
self.loops * self.sleep_time)
for _ in range(self.loops):
time.sleep(self.sleep_time)
node = bm_client.node.get(node.uuid)
if node.power_state is not None:
break
else:
msg = ('Timed out waiting for node %s power state.' %
node.uuid)
raise exceptions.Timeout(msg)
# Get the full node info
node_detail = bm_client.node.get(node.uuid)
capabilities = node_detail.properties.get('capabilities', None)
# Only update capabilities to add boot_option if it doesn't exist.
if capabilities:
if 'boot_option' not in capabilities:
capabilities = "boot_option:local,%s" % capabilities
else:
capabilities = "boot_option:local"
self.log.debug("Configuring boot for Node {0}".format(
node.uuid))
bm_client.node.update(node.uuid, [
{
'op': 'add',
'path': '/properties/capabilities',
'value': capabilities,
},
{
'op': 'add',
'path': '/driver_info/deploy_ramdisk',
'value': ramdisk_id,
},
{
'op': 'add',
'path': '/driver_info/deploy_kernel',
'value': kernel_id,
},
])
class ShowNodeCapabilities(lister.Lister):
"""List the capabilities for all Nodes"""
log = logging.getLogger(__name__ + ".ShowNodeProfile")
def take_action(self, parsed_args):
bm_client = self.app.client_manager.tripleoclient.baremetal()
rows = []
for node in bm_client.node.list():
node_detail = bm_client.node.get(node.uuid)
capabilities = node_detail.properties.get('capabilities')
rows.append((node.uuid, capabilities))
return (("Node UUID", "Node Capabilities"), rows, )