config/sysinv/sysinv/sysinv/sysinv/api/controllers/v1/host.py

6386 lines
280 KiB
Python

# vim: tabstop=4 shiftwidth=4 softtabstop=4
#
# Copyright 2013 Hewlett-Packard Development Company, L.P.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
#
# Copyright (c) 2013-2019 Wind River Systems, Inc.
#
# SPDX-License-Identifier: Apache-2.0
import ast
import cgi
import copy
import json
import math
import os
import re
import xml.etree.ElementTree as ET
import xml.etree.ElementTree as et
from xml.dom import minidom as dom
from sqlalchemy.orm.exc import NoResultFound
import jsonpatch
import netaddr
import pecan
import six
import psutil
import tsconfig.tsconfig as tsc
import wsme
import wsmeext.pecan as wsme_pecan
from wsme import types as wtypes
from controllerconfig import HOST_XML_ATTRIBUTES
from fm_api import constants as fm_constants
from fm_api import fm_api
from pecan import expose
from pecan import rest
from sysinv import objects
from sysinv.api.controllers.v1 import ethernet_port
from sysinv.api.controllers.v1 import port
from sysinv.api.controllers.v1 import address as address_api
from sysinv.api.controllers.v1 import base
from sysinv.api.controllers.v1 import collection
from sysinv.api.controllers.v1 import cpu as cpu_api
from sysinv.api.controllers.v1 import cpu_utils
from sysinv.api.controllers.v1 import disk
from sysinv.api.controllers.v1 import partition
from sysinv.api.controllers.v1 import ceph_mon
from sysinv.api.controllers.v1 import interface as interface_api
from sysinv.api.controllers.v1 import lvg as lvg_api
from sysinv.api.controllers.v1 import memory
from sysinv.api.controllers.v1 import node as node_api
from sysinv.api.controllers.v1 import profile
from sysinv.api.controllers.v1 import pv as pv_api
from sysinv.api.controllers.v1 import sensor as sensor_api
from sysinv.api.controllers.v1 import sensorgroup
from sysinv.api.controllers.v1 import storage
from sysinv.api.controllers.v1 import label
from sysinv.api.controllers.v1 import link
from sysinv.api.controllers.v1 import lldp_agent
from sysinv.api.controllers.v1 import lldp_neighbour
from sysinv.api.controllers.v1 import mtce_api
from sysinv.api.controllers.v1 import pci_device
from sysinv.api.controllers.v1 import route
from sysinv.api.controllers.v1 import sm_api
from sysinv.api.controllers.v1 import state
from sysinv.api.controllers.v1 import types
from sysinv.api.controllers.v1 import utils
from sysinv.api.controllers.v1 import interface_network
from sysinv.api.controllers.v1 import interface_datanetwork
from sysinv.api.controllers.v1 import vim_api
from sysinv.api.controllers.v1 import patch_api
from sysinv.common import ceph
from sysinv.common import constants
from sysinv.common import exception
from sysinv.common import utils as cutils
from sysinv.helm import common as helm_common
from sysinv.openstack.common import log
from sysinv.openstack.common import uuidutils
from sysinv.openstack.common.gettextutils import _
from sysinv.common.storage_backend_conf import StorageBackendConfig
from sysinv.common import health
LOG = log.getLogger(__name__)
KEYRING_BM_SERVICE = "BM"
ERR_CODE_LOCK_SOLE_SERVICE_PROVIDER = "-1003"
def _get_controller_address(hostname):
return utils.lookup_static_ip_address(hostname,
constants.NETWORK_TYPE_MGMT)
def _get_storage_address(hostname):
return utils.lookup_static_ip_address(hostname,
constants.NETWORK_TYPE_MGMT)
class HostProvisionState(state.State):
@classmethod
def convert_with_links(cls, rpc_ihost, expand=True):
provision_state = HostProvisionState()
provision_state.current = rpc_ihost.provision_state
url_arg = '%s/state/provision' % rpc_ihost.uuid
provision_state.links = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts', url_arg),
link.Link.make_link('bookmark',
pecan.request.host_url,
'ihosts', url_arg,
bookmark=True)
]
if expand:
provision_state.target = rpc_ihost.target_provision_state
# TODO(lucasagomes): get_next_provision_available_states
provision_state.available = []
return provision_state
class HostProvisionStateController(rest.RestController):
# GET ihosts/<uuid>/state/provision
@wsme_pecan.wsexpose(HostProvisionState, six.text_type)
def get(self, ihost_id):
ihost = objects.host.get_by_uuid(pecan.request.context,
ihost_id)
provision_state = HostProvisionState.convert_with_links(ihost)
return provision_state
# PUT ihosts/<uuid>/state/provision
@wsme_pecan.wsexpose(HostProvisionState, six.text_type, six.text_type, status=202)
def put(self, ihost_id, target):
"""Set the provision state of the machine."""
# TODO(lucasagomes): Test if target is a valid state and if it's able
# to transition to the target state from the current one
# TODO(lucasagomes): rpcapi.start_provision_state_change()
raise NotImplementedError()
LOCK_NAME_STATE = 'HostStatesController'
class HostStates(base.APIBase):
"""API representation of the states of a ihost."""
# power = ihostPowerState
# "The current power state of the ihost"
provision = HostProvisionState
"The current provision state of the ihost"
@classmethod
def convert_with_links(cls, rpc_ihost):
states = HostStates()
# states.power = ihostPowerState.convert_with_links(rpc_ihost,
# expand=False)
states.provision = HostProvisionState.convert_with_links(
rpc_ihost,
expand=False)
return states
class HostStatesController(rest.RestController):
_custom_actions = {
'host_cpus_modify': ['PUT'],
'update_install_uuid': ['PUT'],
}
# GET ihosts/<uuid>/state
@wsme_pecan.wsexpose(HostStates, six.text_type)
def get(self, ihost_id):
"""List or update the state of a ihost."""
ihost = objects.host.get_by_uuid(pecan.request.context,
ihost_id)
state = HostStates.convert_with_links(ihost)
return state
def _get_host_cpus_collection(self, host_uuid):
cpus = pecan.request.dbapi.icpu_get_by_ihost(host_uuid)
return cpu_api.CPUCollection.convert_with_links(cpus,
limit=None,
url=None,
expand=None,
sort_key=None,
sort_dir=None)
# PUT ihosts/<uuid>/state/update_install_uuid
@cutils.synchronized(LOCK_NAME_STATE)
@wsme_pecan.wsexpose(HostStates, types.uuid, body=six.text_type)
def update_install_uuid(self, host_uuid, install_uuid):
""" Update install_uuid in /etc/platform/platform.conf
on the specified host.
:param host_uuid: UUID of the host
:param install_uuid: install_uuid.
"""
LOG.info("update_install_uuid host_uuid=%s install_uuid=%s" %
(host_uuid, install_uuid))
pecan.request.rpcapi.update_install_uuid(pecan.request.context,
host_uuid,
install_uuid)
# PUT ihosts/<uuid>/state/host_cpus_modify
@cutils.synchronized(cpu_api.LOCK_NAME)
@wsme_pecan.wsexpose(cpu_api.CPUCollection, types.uuid, body=[six.text_type])
def host_cpus_modify(self, host_uuid, capabilities):
""" Perform bulk host cpus modify.
:param host_uuid: UUID of the host
:param capabilities: dictionary of update cpu function and sockets.
Example:
capabilities=[{'function': 'platform', 'sockets': [{'0': 1}, {'1': 0}]},
{'function': 'vswitch', 'sockets': [{'0': 2}]},
{'function': 'shared', 'sockets': [{'0': 1}, {'1': 1}]}]
"""
def cpu_function_sort_key(capability):
function = capability.get('function', '')
if function.lower() == constants.PLATFORM_FUNCTION.lower():
rank = 0
elif function.lower() == constants.SHARED_FUNCTION.lower():
rank = 1
elif function.lower() == constants.VSWITCH_FUNCTION.lower():
rank = 2
elif function.lower() == constants.APPLICATION_FUNCTION.lower():
rank = 3
else:
rank = 4
return rank
specified_function = None
# patch_obj = jsonpatch.JsonPatch(patch)
# for p in patch_obj:
# if p['path'] == '/capabilities':
# capabilities = p['value']
# break
LOG.info("host_cpus_modify host_uuid=%s capabilities=%s" %
(host_uuid, capabilities))
ihost = pecan.request.dbapi.ihost_get(host_uuid)
cpu_api._check_host(ihost)
ihost.nodes = pecan.request.dbapi.inode_get_by_ihost(ihost.uuid)
num_nodes = len(ihost.nodes)
# Perform allocation in platform, shared, vswitch order
sorted_capabilities = sorted(capabilities, key=cpu_function_sort_key)
for icap in sorted_capabilities:
specified_function = icap.get('function', None)
specified_sockets = icap.get('sockets', None)
if not specified_function or not specified_sockets:
raise wsme.exc.ClientSideError(
_('host %s: cpu function=%s or socket=%s not specified '
'for host %s.') % (host_uuid,
specified_function,
specified_sockets))
capability = {}
for specified_socket in specified_sockets:
socket, value = specified_socket.items()[0]
if int(socket) >= num_nodes:
raise wsme.exc.ClientSideError(
_('There is no Processor (Socket) '
'%s on this host.') % socket)
capability.update({'num_cores_on_processor%s' % socket:
int(value)})
LOG.debug("host_cpus_modify capability=%s" % capability)
# Query the database to get the current set of CPUs and then
# organize the data by socket and function for convenience.
ihost.cpus = pecan.request.dbapi.icpu_get_by_ihost(ihost.uuid)
cpu_utils.restructure_host_cpu_data(ihost)
# Get the CPU counts for each socket and function for this host
cpu_counts = cpu_utils.get_cpu_counts(ihost)
# Update the CPU counts for each socket and function for this host based
# on the incoming requested core counts
if (specified_function.lower() == constants.VSWITCH_FUNCTION.lower()):
cpu_counts = cpu_api._update_vswitch_cpu_counts(ihost, None,
cpu_counts,
capability)
elif (specified_function.lower() == constants.SHARED_FUNCTION.lower()):
cpu_counts = cpu_api._update_shared_cpu_counts(ihost, None,
cpu_counts,
capability)
elif (specified_function.lower() == constants.PLATFORM_FUNCTION.lower()):
cpu_counts = cpu_api._update_platform_cpu_counts(ihost, None,
cpu_counts,
capability)
# Semantic check to ensure the minimum/maximum values are enforced
error_msg = cpu_utils.check_core_allocations(ihost, cpu_counts,
specified_function)
if error_msg:
raise wsme.exc.ClientSideError(_(error_msg))
# Update cpu assignments to new values
cpu_utils.update_core_allocations(ihost, cpu_counts)
for cpu in ihost.cpus:
function = cpu_utils.get_cpu_function(ihost, cpu)
if function == constants.NO_FUNCTION:
raise wsme.exc.ClientSideError(_('Could not determine '
'assigned function for CPU %d' % cpu.cpu))
if (not cpu.allocated_function or
cpu.allocated_function.lower() != function.lower()):
values = {'allocated_function': function}
LOG.info("icpu_update uuid=%s value=%s" %
(cpu.uuid, values))
pecan.request.dbapi.icpu_update(cpu.uuid, values)
# perform inservice apply
pecan.request.rpcapi.update_cpu_config(pecan.request.context,
host_uuid)
return self._get_host_cpus_collection(ihost.uuid)
class Host(base.APIBase):
"""API representation of a host.
This class enforces type checking and value constraints, and converts
between the internal object model and the API representation
of an ihost.
"""
# NOTE: translate 'id' publicly to 'uuid' internally
id = int
uuid = wtypes.text
hostname = wtypes.text
invprovision = wtypes.text
"Represent the current (not transition) provision state of the ihost"
mgmt_mac = wtypes.text
"Represent the provisioned Boot mgmt MAC address of the ihost."
mgmt_ip = wtypes.text
"Represent the provisioned Boot mgmt IP address of the ihost."
bm_ip = wtypes.text
"Discovered board management IP address of the ihost."
bm_type = wtypes.text
"Represent the board management type of the ihost."
bm_username = wtypes.text
"Represent the board management username of the ihost."
bm_password = wtypes.text
"Represent the board management password of the ihost."
personality = wtypes.text
"Represent the personality of the ihost"
subfunctions = wtypes.text
"Represent the subfunctions of the ihost"
subfunction_oper = wtypes.text
"Represent the subfunction operational state of the ihost"
subfunction_avail = wtypes.text
"Represent the subfunction availability status of the ihost"
# target_provision_state = wtypes.text
# "The user modified desired provision state of the ihost."
# NOTE: allow arbitrary dicts for driver_info and extra so that drivers
# and vendors can expand on them without requiring API changes.
# NOTE: translate 'driver_info' internally to 'management_configuration'
serialid = wtypes.text
administrative = wtypes.text
operational = wtypes.text
availability = wtypes.text
# The 'action' field is used for action based administration compared
# to existing state change administration.
# Actions like 'reset','reboot', and 'reinstall' are now supported
# by this new method along with 'lock' and 'unlock'.
action = wtypes.text
ihost_action = wtypes.text
'Represent the current action task in progress'
vim_progress_status = wtypes.text
'Represent the vim progress status'
task = wtypes.text
"Represent the mtce task state"
mtce_info = wtypes.text
"Represent the mtce info"
reserved = wtypes.text
config_status = wtypes.text
"Represent the configuration status of this ihost."
config_applied = wtypes.text
"Represent the configuration already applied to this ihost."
config_target = wtypes.text
"Represent the configuration which needs to be applied to this ihost."
# Host uptime
uptime = int
# NOTE: properties should use a class to enforce required properties
# current list: arch, cpus, disk, ram, image
location = {wtypes.text: utils.ValidTypes(wtypes.text, six.integer_types)}
capabilities = {wtypes.text: utils.ValidTypes(wtypes.text,
six.integer_types)}
# NOTE: translate 'isystem_id' to a link to the isystem resource
# and accept a isystem uuid when creating an ihost.
# (Leaf not ihost)
forisystemid = int
isystem_uuid = types.uuid
"The UUID of the system this host belongs to"
iprofile_uuid = types.uuid
"The UUID of the iprofile to apply to host"
peers = types.MultiType({dict})
"This peers of this host in the cluster"
links = [link.Link]
"A list containing a self link and associated ihost links"
iinterfaces = [link.Link]
"Links to the collection of iinterfaces on this ihost"
ports = [link.Link]
"Links to the collection of Ports on this ihost"
ethernet_ports = [link.Link]
"Links to the collection of EthernetPorts on this ihost"
inodes = [link.Link]
"Links to the collection of inodes on this ihost"
icpus = [link.Link]
"Links to the collection of icpus on this ihost"
imemorys = [link.Link]
"Links to the collection of imemorys on this ihost"
istors = [link.Link]
"Links to the collection of istors on this ihost"
idisks = [link.Link]
"Links to the collection of idisks on this ihost"
partitions = [link.Link]
"Links to the collection of partitions on this ihost"
ceph_mon = [link.Link]
"Links to the collection of ceph monitors on this ihost"
ipvs = [link.Link]
"Links to the collection of ipvs on this ihost"
ilvgs = [link.Link]
"Links to the collection of ilvgs on this ihost"
isensors = [link.Link]
"Links to the collection of isensors on this ihost"
isensorgroups = [link.Link]
"Links to the collection of isensorgruops on this ihost"
pci_devices = [link.Link]
"Links to the collection of pci_devices on this host"
lldp_agents = [link.Link]
"Links to the collection of LldpAgents on this ihost"
lldp_neighbours = [link.Link]
"Links to the collection of LldpNeighbours on this ihost"
labels = [link.Link]
"Links to the collection of labels assigned to this host"
boot_device = wtypes.text
rootfs_device = wtypes.text
install_output = wtypes.text
console = wtypes.text
tboot = wtypes.text
vsc_controllers = wtypes.text
"Represent the VSC controllers used by this ihost."
ttys_dcd = wtypes.text
"Enable or disable serial console carrier detect"
software_load = wtypes.text
"The current load software version"
target_load = wtypes.text
"The target load software version"
install_state = wtypes.text
"Represent the install state"
install_state_info = wtypes.text
"Represent install state extra information if there is any"
iscsi_initiator_name = wtypes.text
"The iscsi initiator name (only used for worker hosts)"
def __init__(self, **kwargs):
self.fields = list(objects.host.fields.keys())
for k in self.fields:
setattr(self, k, kwargs.get(k))
self.fields.append('iprofile_uuid')
setattr(self, 'iprofile_uuid', kwargs.get('iprofile_uuid', None))
self.fields.append('peers')
setattr(self, 'peers', kwargs.get('peers', None))
@classmethod
def convert_with_links(cls, rpc_ihost, expand=True):
minimum_fields = ['id', 'uuid', 'hostname',
'personality', 'subfunctions',
'subfunction_oper', 'subfunction_avail',
'administrative', 'operational', 'availability',
'invprovision',
'task', 'mtce_info', 'action', 'uptime', 'reserved',
'ihost_action', 'vim_progress_status',
'mgmt_mac', 'mgmt_ip', 'location',
'bm_ip', 'bm_type', 'bm_username',
'isystem_uuid', 'capabilities', 'serialid',
'config_status', 'config_applied', 'config_target',
'created_at', 'updated_at', 'boot_device',
'rootfs_device', 'install_output', 'console',
'tboot', 'vsc_controllers', 'ttys_dcd',
'software_load', 'target_load', 'peers', 'peer_id',
'install_state', 'install_state_info',
'iscsi_initiator_name']
fields = minimum_fields if not expand else None
uhost = Host.from_rpc_object(rpc_ihost, fields)
uhost.links = [link.Link.make_link('self', pecan.request.host_url,
'ihosts', uhost.uuid),
link.Link.make_link('bookmark',
pecan.request.host_url,
'ihosts', uhost.uuid,
bookmark=True)
]
if expand:
uhost.iinterfaces = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/iinterfaces"),
link.Link.make_link(
'bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/iinterfaces",
bookmark=True)
]
uhost.ports = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/ports"),
link.Link.make_link(
'bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/ports",
bookmark=True)
]
uhost.ethernet_ports = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/ethernet_ports"),
link.Link.make_link(
'bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/ethernet_ports",
bookmark=True)
]
uhost.inodes = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/inodes"),
link.Link.make_link(
'bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/inodes",
bookmark=True)
]
uhost.icpus = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/icpus"),
link.Link.make_link(
'bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/icpus",
bookmark=True)
]
uhost.imemorys = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/imemorys"),
link.Link.make_link(
'bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/imemorys",
bookmark=True)
]
uhost.istors = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/istors"),
link.Link.make_link(
'bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/istors",
bookmark=True)
]
uhost.idisks = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/idisks"),
link.Link.make_link(
'bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/idisks",
bookmark=True)
]
uhost.partitions = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/partitions"),
link.Link.make_link(
'bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/partitions",
bookmark=True)
]
uhost.ceph_mon = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/ceph_mon"),
link.Link.make_link(
'bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/ceph_mon",
bookmark=True)
]
uhost.ipvs = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/ipvs"),
link.Link.make_link(
'bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/ipvs",
bookmark=True)
]
uhost.ilvgs = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/ilvgs"),
link.Link.make_link(
'bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/ilvgs",
bookmark=True)
]
uhost.isensors = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/isensors"),
link.Link.make_link('bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/isensors",
bookmark=True)
]
uhost.isensorgroups = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/isensorgroups"),
link.Link.make_link('bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/isensorgroups",
bookmark=True)
]
uhost.pci_devices = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/pci_devices"),
link.Link.make_link('bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/pci_devices",
bookmark=True)
]
uhost.lldp_agents = [
link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/lldp_agents"),
link.Link.make_link('bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/lldp_agents",
bookmark=True)
]
uhost.lldp_neighbours = [
link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/lldp_neighbors"),
link.Link.make_link('bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/lldp_neighbors",
bookmark=True)
]
uhost.labels = [link.Link.make_link('self',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/labels"),
link.Link.make_link('bookmark',
pecan.request.host_url,
'ihosts',
uhost.uuid + "/labels",
bookmark=True)
]
# Don't expose the vsc_controllers field if we are not configured with
# the nuage_vrs vswitch or we are not a worker node.
vswitch_type = utils.get_vswitch_type()
if (vswitch_type != constants.VSWITCH_TYPE_NUAGE_VRS or
uhost.personality != constants.WORKER):
uhost.vsc_controllers = wtypes.Unset
uhost.peers = None
if uhost.peer_id:
ipeers = pecan.request.dbapi.peer_get(uhost.peer_id)
uhost.peers = {'name': ipeers.name, 'hosts': ipeers.hosts}
return uhost
class HostCollection(collection.Collection):
"""API representation of a collection of ihosts."""
ihosts = [Host]
"A list containing ihosts objects"
def __init__(self, **kwargs):
self._type = 'ihosts'
@classmethod
def convert_with_links(cls, ihosts, limit, url=None,
expand=False, **kwargs):
collection = HostCollection()
collection.ihosts = [
Host.convert_with_links(n, expand) for n in ihosts]
collection.next = collection.get_next(limit, url=url, **kwargs)
return collection
class HostUpdate(object):
"""Host update helper class.
"""
CONTINUE = "continue"
EXIT_RETURN_HOST = "exit_return_host"
EXIT_UPDATE_PREVAL = "exit_update_preval"
FAILED = "failed"
PASSED = "passed"
# Allow mtce to do the SWACT and FORCE_SWACT?
ACTIONS_TO_TASK_DISPLAY_CHOICES = (
(None, _("")),
("", _("")),
(constants.UNLOCK_ACTION, _("Unlocking")),
(constants.FORCE_UNLOCK_ACTION, _("Force Unlocking")),
(constants.LOCK_ACTION, _("Locking")),
(constants.FORCE_LOCK_ACTION, _("Force Locking")),
(constants.RESET_ACTION, _("Resetting")),
(constants.REBOOT_ACTION, _("Rebooting")),
(constants.REINSTALL_ACTION, _("Reinstalling")),
(constants.POWERON_ACTION, _("Powering-on")),
(constants.POWEROFF_ACTION, _("Powering-off")),
(constants.SWACT_ACTION, _("Swacting")),
(constants.FORCE_SWACT_ACTION, _("Force-Swacting")),
)
def __init__(self, ihost_orig, ihost_patch, delta):
self.ihost_orig = dict(ihost_orig)
self.ihost_patch = dict(ihost_patch)
self._delta = list(delta)
self._iprofile_uuid = None
self._ihost_val_prenotify = {}
self._ihost_val = {}
self._configure_required = False
self._notify_vim = False
self._notify_mtce = False
self._notify_availability = None
self._notify_vim_add_host = False
self._notify_action_lock = False
self._notify_action_lock_force = False
self._skip_notify_mtce = False
self._bm_type_changed_to_none = False
self._nextstep = self.CONTINUE
self._action = None
self.displayid = ihost_patch.get('hostname')
if not self.displayid:
self.displayid = ihost_patch.get('uuid')
LOG.debug("ihost_orig=%s, ihost_patch=%s, delta=%s" %
(self.ihost_orig, self.ihost_patch, self.delta))
@property
def action(self):
return self._action
@action.setter
def action(self, val):
self._action = val
@property
def delta(self):
return self._delta
@property
def nextstep(self):
return self._nextstep
@nextstep.setter
def nextstep(self, val):
self._nextstep = val
@property
def iprofile_uuid(self):
return self._iprofile_uuid
@iprofile_uuid.setter
def iprofile_uuid(self, val):
self._iprofile_uuid = val
@property
def configure_required(self):
return self._configure_required
@configure_required.setter
def configure_required(self, val):
self._configure_required = val
@property
def bm_type_changed_to_none(self):
return self._bm_type_changed_to_none
@bm_type_changed_to_none.setter
def bm_type_changed_to_none(self, val):
self._bm_type_changed_to_none = val
@property
def notify_vim_add_host(self):
return self._notify_vim_add_host
@notify_vim_add_host.setter
def notify_vim_add_host(self, val):
self._notify_vim_add_host = val
@property
def skip_notify_mtce(self):
return self._skip_notify_mtce
@skip_notify_mtce.setter
def skip_notify_mtce(self, val):
self._skip_notify_mtce = val
@property
def notify_action_lock(self):
return self._notify_action_lock
@notify_action_lock.setter
def notify_action_lock(self, val):
self._notify_action_lock = val
@property
def notify_action_lock_force(self):
return self._notify_action_lock_force
@notify_action_lock_force.setter
def notify_action_lock_force(self, val):
self._notify_action_lock_force = val
@property
def ihost_val_prenotify(self):
return self._ihost_val_prenotify
def ihost_val_prenotify_update(self, val):
self._ihost_val_prenotify.update(val)
@property
def ihost_val(self):
return self._ihost_val
def ihost_val_update(self, val):
self._ihost_val.update(val)
@property
def notify_vim(self):
return self._notify_vim
@notify_vim.setter
def notify_vim(self, val):
self._notify_vim = val
@property
def notify_mtce(self):
return self._notify_mtce
@notify_mtce.setter
def notify_mtce(self, val):
self._notify_mtce = val
@property
def notify_availability(self):
return self._notify_availability
@notify_availability.setter
def notify_availability(self, val):
self._notify_availability = val
def get_task_from_action(self, action):
"""Lookup the task value in the action to task dictionary."""
display_choices = self.ACTIONS_TO_TASK_DISPLAY_CHOICES
display_value = [display for (value, display) in display_choices
if value and value.lower() == (action or '').lower()]
if display_value:
return display_value[0]
return None
LOCK_NAME = 'HostController'
LOCK_NAME_SYS = 'HostControllerSys'
class HostController(rest.RestController):
"""REST controller for ihosts."""
state = HostStatesController()
"Expose the state controller action as a sub-element of ihosts"
iinterfaces = interface_api.InterfaceController(
from_ihosts=True)
"Expose iinterfaces as a sub-element of ihosts"
ports = port.PortController(
from_ihosts=True)
"Expose ports as a sub-element of ihosts"
ethernet_ports = ethernet_port.EthernetPortController(
from_ihosts=True)
"Expose ethernet_ports as a sub-element of ihosts"
inodes = node_api.NodeController(from_ihosts=True)
"Expose inodes as a sub-element of ihosts"
icpus = cpu_api.CPUController(from_ihosts=True)
"Expose icpus as a sub-element of ihosts"
imemorys = memory.MemoryController(from_ihosts=True)
"Expose imemorys as a sub-element of ihosts"
istors = storage.StorageController(from_ihosts=True)
"Expose istors as a sub-element of ihosts"
idisks = disk.DiskController(from_ihosts=True)
"Expose idisks as a sub-element of ihosts"
partitions = partition.PartitionController(from_ihosts=True)
"Expose partitions as a sub-element of ihosts"
ceph_mon = ceph_mon.CephMonController(from_ihosts=True)
"Expose ceph_mon as a sub-element of ihosts"
ipvs = pv_api.PVController(from_ihosts=True)
"Expose ipvs as a sub-element of ihosts"
ilvgs = lvg_api.LVGController(from_ihosts=True)
"Expose ilvgs as a sub-element of ihosts"
addresses = address_api.AddressController(parent="ihosts")
"Expose addresses as a sub-element of ihosts"
routes = route.RouteController(parent="ihosts")
"Expose routes as a sub-element of ihosts"
isensors = sensor_api.SensorController(from_ihosts=True)
"Expose isensors as a sub-element of ihosts"
isensorgroups = sensorgroup.SensorGroupController(from_ihosts=True)
"Expose isensorgroups as a sub-element of ihosts"
pci_devices = pci_device.PCIDeviceController(from_ihosts=True)
"Expose pci_devices as a sub-element of ihosts"
lldp_agents = lldp_agent.LLDPAgentController(
from_ihosts=True)
"Expose lldp_agents as a sub-element of ihosts"
lldp_neighbours = lldp_neighbour.LLDPNeighbourController(
from_ihosts=True)
"Expose lldp_neighbours as a sub-element of ihosts"
labels = label.LabelController(from_ihosts=True)
"Expose labels as a sub-element of ihosts"
interface_networks = interface_network.InterfaceNetworkController(
parent="ihosts")
"Expose interface_networks as a sub-element of ihosts"
interface_datanetworks = interface_datanetwork.InterfaceDataNetworkController(
parent="ihosts")
"Expose interface_datanetworks as a sub-element of ihosts"
_custom_actions = {
'detail': ['GET'],
'bulk_add': ['POST'],
'bulk_export': ['GET'],
'upgrade': ['POST'],
'downgrade': ['POST'],
'install_progress': ['POST'],
}
def __init__(self, from_isystem=False):
self._from_isystem = from_isystem
self._mtc_address = constants.LOCALHOST_HOSTNAME
self._mtc_port = 2112
self._ceph = ceph.CephApiOperator()
self._api_token = None
# self._name = 'api-host'
def _ihosts_get(self, isystem_id, marker, limit, personality,
sort_key, sort_dir):
if self._from_isystem and not isystem_id: # TODO: check uuid
raise exception.InvalidParameterValue(_(
"System id not specified."))
limit = utils.validate_limit(limit)
sort_dir = utils.validate_sort_dir(sort_dir)
marker_obj = None
if marker:
marker_obj = objects.host.get_by_uuid(pecan.request.context,
marker)
if isystem_id:
ihosts = pecan.request.dbapi.ihost_get_by_isystem(
isystem_id, limit,
marker_obj,
sort_key=sort_key,
sort_dir=sort_dir)
else:
if personality:
ihosts = pecan.request.dbapi.ihost_get_by_personality(
personality, limit, marker_obj,
sort_key=sort_key,
sort_dir=sort_dir)
else:
ihosts = pecan.request.dbapi.ihost_get_list(
limit, marker_obj,
sort_key=sort_key,
sort_dir=sort_dir)
for h in ihosts:
self._update_controller_personality(h)
return ihosts
@staticmethod
def _update_subfunctions(ihost):
subfunctions = ihost.get('subfunctions') or ""
personality = ihost.get('personality') or ""
# handle race condition with subfunctions being updated late.
if not subfunctions:
LOG.info("update_subfunctions: subfunctions not set. personality=%s" %
personality)
if personality == constants.CONTROLLER:
subfunctions = ','.join(tsc.subfunctions)
else:
subfunctions = personality
ihost['subfunctions'] = subfunctions
subfunctions_set = set(subfunctions.split(','))
if personality not in subfunctions_set:
# Automatically add it
subfunctions_list = list(subfunctions_set)
subfunctions_list.insert(0, personality)
subfunctions = ','.join(subfunctions_list)
LOG.info("%s personality=%s update subfunctions=%s" %
(ihost.get('hostname'), personality, subfunctions))
LOG.debug("update_subfunctions: personality=%s subfunctions=%s" %
(personality, subfunctions))
return subfunctions
@staticmethod
def _update_controller_personality(host):
if host['personality'] == constants.CONTROLLER:
if utils.is_host_active_controller(host):
activity = 'Controller-Active'
else:
activity = 'Controller-Standby'
host['capabilities'].update({'Personality': activity})
@wsme_pecan.wsexpose(HostCollection, six.text_type, six.text_type, int, six.text_type,
six.text_type, six.text_type)
def get_all(self, isystem_id=None, marker=None, limit=None,
personality=None,
sort_key='id', sort_dir='asc'):
"""Retrieve a list of ihosts."""
ihosts = self._ihosts_get(
isystem_id, marker, limit, personality, sort_key, sort_dir)
return HostCollection.convert_with_links(ihosts, limit,
sort_key=sort_key,
sort_dir=sort_dir)
@wsme_pecan.wsexpose(six.text_type, six.text_type, body=six.text_type)
def install_progress(self, uuid, install_state,
install_state_info=None):
""" Update the install status for the given host."""
LOG.debug("Update host uuid %s with install_state=%s "
"and install_state_info=%s" %
(uuid, install_state, install_state_info))
if install_state == constants.INSTALL_STATE_INSTALLED:
# After an install a node will reboot right away. Change the state
# to refect this.
install_state = constants.INSTALL_STATE_BOOTING
host = objects.host.get_by_uuid(pecan.request.context, uuid)
pecan.request.dbapi.ihost_update(host['uuid'],
{'install_state': install_state,
'install_state_info':
install_state_info})
@wsme_pecan.wsexpose(HostCollection, six.text_type, six.text_type, int, six.text_type,
six.text_type, six.text_type)
def detail(self, isystem_id=None, marker=None, limit=None,
personality=None,
sort_key='id', sort_dir='asc'):
"""Retrieve a list of ihosts with detail."""
# /detail should only work against collections
parent = pecan.request.path.split('/')[:-1][-1]
if parent != "ihosts":
raise exception.HTTPNotFound
ihosts = self._ihosts_get(
isystem_id, marker, limit, personality, sort_key, sort_dir)
resource_url = '/'.join(['ihosts', 'detail'])
return HostCollection.convert_with_links(ihosts, limit,
url=resource_url,
expand=True,
sort_key=sort_key,
sort_dir=sort_dir)
@wsme_pecan.wsexpose(Host, six.text_type)
def get_one(self, uuid):
"""Retrieve information about the given ihost."""
if self._from_isystem:
raise exception.OperationNotPermitted
rpc_ihost = objects.host.get_by_uuid(pecan.request.context,
uuid)
self._update_controller_personality(rpc_ihost)
return Host.convert_with_links(rpc_ihost)
def _block_add_host_semantic_checks(self, ihost_dict):
if not self._no_controllers_exist() and \
ihost_dict.get('personality') is None:
# Semantic Check: Prevent adding any new host(s) until there is
# an unlocked-enabled controller to manage them.
controller_list = pecan.request.dbapi.ihost_get_by_personality(
personality=constants.CONTROLLER)
have_unlocked_enabled_controller = False
for c in controller_list:
if (c['administrative'] == constants.ADMIN_UNLOCKED and
c['operational'] == constants.OPERATIONAL_ENABLED):
have_unlocked_enabled_controller = True
break
if not have_unlocked_enabled_controller:
raise wsme.exc.ClientSideError(_(
"Provisioning request for new host '%s' is not permitted "
"while there is no unlocked-enabled controller. Unlock "
"controller-0, wait for it to enable and then retry.") %
ihost_dict.get('mgmt_mac'))
def _new_host_semantic_checks(self, ihost_dict):
if not self._no_controllers_exist():
self._block_add_host_semantic_checks(ihost_dict)
mgmt_network = pecan.request.dbapi.network_get_by_type(
constants.NETWORK_TYPE_MGMT)
if mgmt_network.dynamic and ihost_dict.get('mgmt_ip'):
raise wsme.exc.ClientSideError(_(
"Host-add Rejected: Cannot specify a mgmt_ip when dynamic "
"address allocation is configured"))
elif (not mgmt_network.dynamic and
not ihost_dict.get('mgmt_ip') and
ihost_dict.get('personality') not in
[constants.STORAGE, constants.CONTROLLER]):
raise wsme.exc.ClientSideError(_(
"Host-add Rejected: Cannot add a worker host without "
"specifying a mgmt_ip when static address allocation is "
"configured."))
# Check whether vsc_controllers is set and perform semantic
# checking if necessary.
if ihost_dict['vsc_controllers']:
self._semantic_check_vsc_controllers(
ihost_dict, ihost_dict['vsc_controllers'])
# Check whether the system mode is simplex
if utils.get_system_mode() == constants.SYSTEM_MODE_SIMPLEX:
raise wsme.exc.ClientSideError(_(
"Host-add Rejected: Adding a host on a simplex system "
"is not allowed."))
personality = ihost_dict['personality']
if not ihost_dict['hostname']:
if personality not in (constants.CONTROLLER, constants.STORAGE):
raise wsme.exc.ClientSideError(_(
"Host-add Rejected. Must provide a hostname for a node of "
"personality %s") % personality)
else:
self._validate_hostname(ihost_dict['hostname'], personality)
HostController._personality_license_check(personality)
def _do_post(self, ihost_dict):
"""Create a new ihost based off a dictionary of attributes """
log_start = cutils.timestamped("ihost_post_start")
LOG.info("SYS_I host %s %s add" % (ihost_dict['hostname'],
log_start))
power_on = ihost_dict.get('power_on', None)
ihost_obj = None
# Semantic checks for adding a new node
if self._from_isystem:
raise exception.OperationNotPermitted
self._new_host_semantic_checks(ihost_dict)
current_ihosts = pecan.request.dbapi.ihost_get_list()
hostnames = [h['hostname'] for h in current_ihosts]
# Check for missing/invalid hostname
# ips/hostnames are automatic for controller & storage nodes
if ihost_dict['personality'] not in (constants.CONTROLLER,
constants.STORAGE):
if ihost_dict['hostname'] in hostnames:
raise wsme.exc.ClientSideError(
_("Host-add Rejected: Hostname already exists"))
if ihost_dict.get('mgmt_ip') and ihost_dict['mgmt_ip'] in \
[h['mgmt_ip'] for h in current_ihosts]:
raise wsme.exc.ClientSideError(
_("Host-add Rejected: Host with mgmt_ip %s already "
"exists") % ihost_dict['mgmt_ip'])
try:
ihost_obj = pecan.request.dbapi.ihost_get_by_mgmt_mac(
ihost_dict['mgmt_mac'])
# A host with this MAC already exists. We will allow it to be
# added if the hostname and personality have not been set.
if ihost_obj['hostname'] or ihost_obj['personality']:
raise wsme.exc.ClientSideError(
_("Host-add Rejected: Host with mgmt_mac %s already "
"exists") % ihost_dict['mgmt_mac'])
# Check DNSMASQ for ip/mac already existing
# -> node in use by someone else or has already been booted
elif (not ihost_obj and self._dnsmasq_mac_exists(
ihost_dict['mgmt_mac'])):
raise wsme.exc.ClientSideError(
_("Host-add Rejected: mgmt_mac %s has already been "
"active") % ihost_dict['mgmt_mac'])
# Use the uuid from the existing host
ihost_dict['uuid'] = ihost_obj['uuid']
except exception.NodeNotFound:
# This is a new host
pass
if not ihost_dict.get('uuid'):
ihost_dict['uuid'] = uuidutils.generate_uuid()
ihost_dict['mgmt_mac'] = cutils.validate_and_normalize_mac(
ihost_dict['mgmt_mac'])
# BM handling
defaults = objects.host.get_defaults()
ihost_orig = copy.deepcopy(ihost_dict)
subfunctions = self._update_subfunctions(ihost_dict)
ihost_dict['subfunctions'] = subfunctions
changed_paths = []
delta = set()
for key in defaults:
# Internal values that aren't being modified
if key in ['id', 'updated_at', 'created_at']:
continue
# Update only the new fields
if key in ihost_dict and ihost_dict[key] != defaults[key]:
delta.add(key)
ihost_orig[key] = defaults[key]
bm_list = ['bm_type', 'bm_ip',
'bm_username', 'bm_password']
for bmi in bm_list:
if bmi in ihost_dict:
delta.add(bmi)
changed_paths.append({'path': '/' + str(bmi),
'value': ihost_dict[bmi],
'op': 'replace'})
self._bm_semantic_check_and_update(ihost_orig, ihost_dict,
delta, changed_paths,
current_ihosts)
if ('capabilities' not in ihost_dict or not ihost_dict['capabilities']):
ihost_dict['capabilities'] = {}
# If this is the first controller being set up,
# configure and return
if ihost_dict['personality'] == constants.CONTROLLER:
if self._no_controllers_exist():
pecan.request.rpcapi.create_controller_filesystems(
pecan.request.context, ihost_dict['rootfs_device'])
controller_ihost = pecan.request.rpcapi.create_ihost(
pecan.request.context, ihost_dict)
if 'recordtype' in ihost_dict and \
ihost_dict['recordtype'] != "profile":
pecan.request.rpcapi.configure_ihost(
pecan.request.context,
controller_ihost)
return Host.convert_with_links(controller_ihost)
if ihost_dict['personality'] in (constants.CONTROLLER, constants.STORAGE):
self._controller_storage_node_setup(ihost_dict)
# Validate that management name and IP do not already exist
# If one exists, other value must match in addresses table
mgmt_address_name = cutils.format_address_name(
ihost_dict['hostname'], constants.NETWORK_TYPE_MGMT)
self._validate_address_not_allocated(mgmt_address_name,
ihost_dict.get('mgmt_ip'))
if ihost_dict.get('mgmt_ip'):
self._validate_ip_in_mgmt_network(ihost_dict['mgmt_ip'])
else:
del ihost_dict['mgmt_ip']
# Set host to reinstalling
ihost_dict.update({constants.HOST_ACTION_STATE:
constants.HAS_REINSTALLING})
# Creation/Configuration
if ihost_obj:
# The host exists - do an update.
defaults = objects.host.get_defaults()
for key in defaults:
# Internal values that shouldn't be updated
if key in ['id', 'updated_at', 'created_at', 'uuid']:
continue
# Update only the fields that are not empty and have changed
if (key in ihost_dict and ihost_dict[key] and
(ihost_obj[key] != ihost_dict[key])):
ihost_obj[key] = ihost_dict[key]
ihost_obj = pecan.request.rpcapi.update_ihost(pecan.request.context,
ihost_obj)
else:
# The host doesn't exist - do an add.
LOG.info("create_ihost=%s" % ihost_dict.get('hostname'))
ihost_obj = pecan.request.rpcapi.create_ihost(pecan.request.context,
ihost_dict)
ihost_obj = objects.host.get_by_uuid(pecan.request.context,
ihost_obj.uuid)
pecan.request.dbapi.network_get_by_type(constants.NETWORK_TYPE_MGMT)
# Configure the new ihost
ihost_ret = pecan.request.rpcapi.configure_ihost(pecan.request.context,
ihost_obj)
# Notify maintenance about updated mgmt_ip
ihost_obj['mgmt_ip'] = ihost_ret.mgmt_ip
# Add ihost to mtc
new_ihost_mtc = ihost_obj.as_dict()
new_ihost_mtc.update({'operation': 'add'})
new_ihost_mtc = cutils.removekeys_nonmtce(new_ihost_mtc)
mtc_response = mtce_api.host_add(
self._api_token, self._mtc_address, self._mtc_port, new_ihost_mtc,
constants.MTC_ADD_TIMEOUT_IN_SECS)
if mtc_response is None:
mtc_response = {'status': 'fail',
'reason': 'no response',
'action': 'retry'}
if mtc_response['status'] != 'pass':
# Report mtc error
raise wsme.exc.ClientSideError(_("Maintenance has returned with "
"a status of %s, reason: %s, recommended action: %s") % (
mtc_response.get('status'),
mtc_response.get('reason'),
mtc_response.get('action')))
# once the ihost is added to mtc, attempt to power it on
if power_on is not None and ihost_obj['bm_type'] is not None:
new_ihost_mtc.update({'action': constants.POWERON_ACTION})
mtc_response = {'status': None}
mtc_response = mtce_api.host_modify(
self._api_token, self._mtc_address, self._mtc_port, new_ihost_mtc,
constants.MTC_ADD_TIMEOUT_IN_SECS)
if mtc_response is None:
mtc_response = {'status': 'fail',
'reason': 'no response',
'action': 'retry'}
if mtc_response['status'] != 'pass':
# Report mtc error
raise wsme.exc.ClientSideError(_("Maintenance has returned with "
"a status of %s, reason: %s, recommended action: %s") % (
mtc_response.get('status'),
mtc_response.get('reason'),
mtc_response.get('action')))
# Notify the VIM that the host has been added - must be done after
# the host has been added to mtc and saved to the DB.
LOG.info("VIM notify add host add %s subfunctions=%s" % (
ihost_obj['hostname'], subfunctions))
try:
vim_api.vim_host_add(
self._api_token,
ihost_obj['uuid'],
ihost_obj['hostname'],
subfunctions,
ihost_obj['administrative'],
ihost_obj['operational'],
ihost_obj['availability'],
ihost_obj['subfunction_oper'],
ihost_obj['subfunction_avail'],
constants.VIM_DEFAULT_TIMEOUT_IN_SECS)
except Exception as e:
LOG.warn(_("No response from vim_api %s e=%s" %
(ihost_obj['hostname'], e)))
self._api_token = None
pass # VIM audit will pickup
log_end = cutils.timestamped("ihost_post_end")
LOG.info("SYS_I host %s %s" % (ihost_obj.hostname, log_end))
return Host.convert_with_links(ihost_obj)
@cutils.synchronized(LOCK_NAME)
@expose('json')
def bulk_add(self):
pending_creation = []
success_str = ""
error_str = ""
if utils.get_system_mode() == constants.SYSTEM_MODE_SIMPLEX:
return dict(
success="",
error="Bulk add on a simplex system is not allowed."
)
# Semantic Check: Prevent bulk add until there is an unlocked
# and enabled controller to manage them.
controller_list = pecan.request.dbapi.ihost_get_by_personality(
personality=constants.CONTROLLER)
have_unlocked_enabled_controller = False
for c in controller_list:
if (c['administrative'] == constants.ADMIN_UNLOCKED and
c['operational'] == constants.OPERATIONAL_ENABLED):
have_unlocked_enabled_controller = True
break
if not have_unlocked_enabled_controller:
return dict(
success="",
error="Bulk_add requires enabled controller. Please "
"unlock controller-0, wait for it to enable and then retry."
)
LOG.info("Starting ihost bulk_add operation")
assert isinstance(pecan.request.POST['file'], cgi.FieldStorage)
fileitem = pecan.request.POST['file']
if not fileitem.filename:
return dict(success="", error="Error: No file uploaded")
try:
contents = fileitem.file.read()
# Generate an array of hosts' attributes to be used in creation
root = ET.fromstring(contents)
except Exception:
return dict(
success="",
error="No hosts have been added, invalid XML document"
)
for idx, xmlhost in enumerate(root.findall('host')):
new_ihost = {}
for attr in HOST_XML_ATTRIBUTES:
elem = xmlhost.find(attr)
if elem is not None:
# If the element is found, set the attribute.
# If the text field is empty, set it to the empty string.
new_ihost[attr] = elem.text or ""
else:
# If the element is not found, set the attribute to None.
new_ihost[attr] = None
# This is the expected format of the location field
if new_ihost['location'] is not None:
new_ihost['location'] = {"locn": new_ihost['location']}
# Semantic checks
try:
LOG.debug(new_ihost)
self._new_host_semantic_checks(new_ihost)
except Exception as ex:
culprit = new_ihost.get('hostname') or "with index " + str(idx)
return dict(
success="",
error=" No hosts have been added, error parsing host %s: "
"%s" % (culprit, ex)
)
pending_creation.append(new_ihost)
# Find local network adapter MACs
my_macs = list()
for liSnics in psutil.net_if_addrs().values():
for snic in liSnics:
if snic.family == psutil.AF_LINK:
my_macs.append(snic.address)
# Perform the actual creations
for new_host in pending_creation:
try:
# Configuring for the setup controller, only uses BMC fields
if new_host['mgmt_mac'].lower() in my_macs:
changed_paths = list()
bm_list = ['bm_type', 'bm_ip',
'bm_username', 'bm_password']
for bmi in bm_list:
if bmi in new_host:
changed_paths.append({
'path': '/' + str(bmi),
'value': new_host[bmi],
'op': 'replace'
})
ihost_obj = [ihost
for ihost in pecan.request.dbapi.ihost_get_list()
if ihost['mgmt_mac'] in my_macs]
if len(ihost_obj) != 1:
raise Exception("Unexpected: no/more_than_one host(s) contain(s) a management mac address from local network adapters")
self._patch(ihost_obj[0]['uuid'],
changed_paths, None)
else:
self._do_post(new_host)
if new_host['power_on'] is not None and new_host['bm_type'] is None:
success_str = "%s\n %s Warning: Ignoring <power_on> due to insufficient board management (bm) data." % (success_str, new_host['hostname'])
else:
success_str = "%s\n %s" % (success_str, new_host['hostname'])
except Exception as ex:
LOG.exception(ex)
error_str += " " + (new_host.get('hostname') or
new_host.get('personality')) + \
": " + str(ex) + "\n"
return dict(
success=success_str,
error=error_str
)
@expose('json')
def bulk_export(self):
def host_personality_name_sort_key(host):
if host.personality == constants.CONTROLLER:
rank = 0
elif host.personality == constants.STORAGE:
rank = 1
elif host.personality == constants.WORKER:
rank = 2
else:
rank = 3
return rank, host.hostname
xml_host_node = et.Element('hosts', {'version': cutils.get_sw_version()})
mgmt_network = pecan.request.dbapi.network_get_by_type(
constants.NETWORK_TYPE_MGMT)
host_list = pecan.request.dbapi.ihost_get_list()
sorted_hosts = sorted(host_list, key=host_personality_name_sort_key)
for host in sorted_hosts:
_create_node(host, xml_host_node, host.personality,
mgmt_network.dynamic)
xml_text = dom.parseString(et.tostring(xml_host_node)).toprettyxml()
result = {'content': xml_text}
return result
@cutils.synchronized(LOCK_NAME)
@wsme_pecan.wsexpose(Host, body=Host)
def post(self, host):
"""Create a new ihost."""
ihost_dict = host.as_dict()
# bm_password is not a part of ihost, so retrieve it from the body
body = json.loads(pecan.request.body)
if 'bm_password' in body:
ihost_dict['bm_password'] = body['bm_password']
else:
ihost_dict['bm_password'] = ''
return self._do_post(ihost_dict)
@wsme_pecan.wsexpose(Host, six.text_type, body=[six.text_type])
def patch(self, uuid, patch):
""" Update an existing ihost.
"""
utils.validate_patch(patch)
profile_uuid = None
optimizable = 0
optimize_list = ['/uptime', '/location', '/serialid', '/task']
for p in patch:
# Check if this patch contains a profile
path = p['path']
if path == '/iprofile_uuid':
profile_uuid = p['value']
patch.remove(p)
if path in optimize_list:
optimizable += 1
if len(patch) == optimizable:
return self._patch(uuid, patch, profile_uuid)
elif (pecan.request.user_agent.startswith('mtce') or
pecan.request.user_agent.startswith('vim')):
return self._patch_sys(uuid, patch, profile_uuid)
else:
return self._patch_gen(uuid, patch, profile_uuid)
@cutils.synchronized(LOCK_NAME_SYS)
def _patch_sys(self, uuid, patch, profile_uuid):
return self._patch(uuid, patch, profile_uuid)
@cutils.synchronized(LOCK_NAME)
def _patch_gen(self, uuid, patch, profile_uuid):
return self._patch(uuid, patch, profile_uuid)
@staticmethod
def _validate_capability_is_not_set(old, new):
is_set, __ = new
return not is_set
@staticmethod
def _validate_capability_is_equal(old, new):
return old == new
def _validate_capabilities(self, old_caps, new_caps):
""" Reject updating read-only host capabilities:
1. stor_function. This field is set to 'monitor' for hosts that are
running ceph monitor process: controller-0, controller-1, storage-0.
2. Personality. This field is "virtual": not saved in the database but
returned via API and displayed via "system host-show".
:param old_caps: current host capabilities
:type old_caps: dict
:param new_caps: updated host capabilies (to be set)
:type new_caps: str
:raises: wsme.exc.ClientSideError when attempting to change read-only
capabilities
"""
if type(new_caps) == str:
try:
new_caps = ast.literal_eval(new_caps)
except SyntaxError:
pass
if type(new_caps) != dict:
raise wsme.exc.ClientSideError(
_("Changing capabilities type is not allowed: "
"old_value={}, new_value={}").format(
old_caps, new_caps))
PROTECTED_CAPABILITIES = [
('Personality',
self._validate_capability_is_not_set),
(constants.IHOST_STOR_FUNCTION,
self._validate_capability_is_equal)]
for capability, validate in PROTECTED_CAPABILITIES:
old_is_set, old_value = (
capability in old_caps, old_caps.get(capability))
new_is_set, new_value = (
capability in new_caps, new_caps.get(capability))
if not validate((old_is_set, old_value),
(new_is_set, new_value)):
if old_is_set:
raise wsme.exc.ClientSideError(
_("Changing capability not allowed: "
"name={}, old_value={}, new_value={}. ").format(
capability, old_value, new_value))
else:
raise wsme.exc.ClientSideError(
_("Setting capability not allowed: "
"name={}, value={}. ").format(
capability, new_value))
def _patch(self, uuid, patch, myprofile_uuid):
log_start = cutils.timestamped("ihost_patch_start")
patch_obj = jsonpatch.JsonPatch(patch)
ihost_obj = objects.host.get_by_uuid(pecan.request.context, uuid)
ihost_dict = ihost_obj.as_dict()
self._block_add_host_semantic_checks(ihost_dict)
# Add transient fields that are not stored in the database
ihost_dict['bm_password'] = None
for p in patch:
if p['value'] != 'storage':
break
try:
patched_ihost = jsonpatch.apply_patch(ihost_dict,
patch_obj)
except jsonpatch.JsonPatchException as e:
LOG.exception(e)
raise wsme.exc.ClientSideError(_("Patching Error: %s") % e)
self._validate_capabilities(
ihost_dict['capabilities'], patched_ihost['capabilities'])
defaults = objects.host.get_defaults()
ihost_dict_orig = dict(ihost_obj.as_dict())
for key in defaults:
# Internal values that shouldn't be part of the patch
if key in ['id', 'updated_at', 'created_at']:
continue
# In case of a remove operation, add the missing fields back
# to the document with their default value
if key in ihost_dict and key not in patched_ihost:
patched_ihost[key] = defaults[key]
# Update only the fields that have changed
if ihost_obj[key] != patched_ihost[key]:
ihost_obj[key] = patched_ihost[key]
delta = ihost_obj.obj_what_changed()
delta_handle = list(delta)
uptime_update = False
if 'uptime' in delta_handle:
# There is a log of uptime updates, so just do a debug log
uptime_update = True
LOG.debug("%s %s patch" % (ihost_obj.hostname,
log_start))
else:
LOG.info("%s %s patch" % (ihost_obj.hostname,
log_start))
hostupdate = HostUpdate(ihost_dict_orig, patched_ihost, delta)
if delta_handle:
self._validate_delta(delta_handle)
if delta_handle == ['uptime']:
LOG.debug("%s 1. delta_handle %s" %
(hostupdate.displayid, delta_handle))
else:
LOG.info("%s 1. delta_handle %s" %
(hostupdate.displayid, delta_handle))
else:
LOG.info("%s ihost_patch_end. No changes from %s." %
(hostupdate.displayid, pecan.request.user_agent))
return Host.convert_with_links(ihost_obj)
myaction = patched_ihost.get('action')
if self.action_check(myaction, hostupdate):
LOG.info("%s post action_check hostupdate "
"action=%s notify_vim=%s notify_mtc=%s "
"skip_notify_mtce=%s" %
(hostupdate.displayid,
hostupdate.action,
hostupdate.notify_vim,
hostupdate.notify_mtce,
hostupdate.skip_notify_mtce))
hostupdate.iprofile_uuid = myprofile_uuid
if self.stage_action(myaction, hostupdate):
LOG.info("%s Action staged: %s" %
(hostupdate.displayid, myaction))
else:
LOG.info("%s ihost_patch_end stage_action rc %s" %
(hostupdate.displayid, hostupdate.nextstep))
if hostupdate.nextstep == hostupdate.EXIT_RETURN_HOST:
return Host.convert_with_links(ihost_obj)
elif hostupdate.nextstep == hostupdate.EXIT_UPDATE_PREVAL:
if hostupdate.ihost_val_prenotify:
# update value in db prior to notifications
LOG.info("update ihost_val_prenotify: %s" %
hostupdate.ihost_val_prenotify)
ihost_obj = pecan.request.dbapi.ihost_update(
ihost_obj['uuid'], hostupdate.ihost_val_prenotify)
return Host.convert_with_links(ihost_obj)
if myaction == constants.SUBFUNCTION_CONFIG_ACTION:
self.perform_action_subfunction_config(ihost_obj)
if myaction in delta_handle:
delta_handle.remove(myaction)
LOG.info("%s post action_stage hostupdate "
"action=%s notify_vim=%s notify_mtc=%s "
"skip_notify_mtce=%s" %
(hostupdate.displayid,
hostupdate.action,
hostupdate.notify_vim,
hostupdate.notify_mtce,
hostupdate.skip_notify_mtce))
self._optimize_delta_handling(delta_handle)
if 'administrative' in delta or \
'operational' in delta:
self.stage_administrative_update(hostupdate)
if delta_handle:
LOG.info("%s 2. delta_handle %s" %
(hostupdate.displayid, delta_handle))
self.check_provisioning(hostupdate, patch)
if (hostupdate.ihost_orig['administrative'] ==
constants.ADMIN_UNLOCKED):
self.check_updates_while_unlocked(hostupdate, delta)
current_ihosts = None
hostupdate.bm_type_changed_to_none = \
self._bm_semantic_check_and_update(hostupdate.ihost_orig,
hostupdate.ihost_patch,
delta, patch_obj,
current_ihosts,
hostupdate)
LOG.info("%s post delta_handle hostupdate "
"action=%s notify_vim=%s notify_mtc=%s "
"skip_notify_mtce=%s" %
(hostupdate.displayid,
hostupdate.action,
hostupdate.notify_vim,
hostupdate.notify_mtce,
hostupdate.skip_notify_mtce))
if hostupdate.bm_type_changed_to_none:
hostupdate.ihost_val_update({'bm_ip': None,
'bm_username': None,
'bm_password': None})
if hostupdate.ihost_val_prenotify:
# update value in db prior to notifications
LOG.info("update ihost_val_prenotify: %s" %
hostupdate.ihost_val_prenotify)
pecan.request.dbapi.ihost_update(ihost_obj['uuid'],
hostupdate.ihost_val_prenotify)
if hostupdate.ihost_val:
# apply the staged updates in preparation for update
LOG.info("%s apply ihost_val %s" %
(hostupdate.displayid, hostupdate.ihost_val))
for k, v in hostupdate.ihost_val.items():
ihost_obj[k] = v
LOG.debug("AFTER Apply ihost_val %s to iHost %s" %
(hostupdate.ihost_val, ihost_obj.as_dict()))
if 'personality' in delta:
self._update_subfunctions(ihost_obj)
if hostupdate.notify_vim:
action = hostupdate.action
LOG.info("Notify VIM host action %s action=%s" % (
ihost_obj['hostname'], action))
try:
vim_api.vim_host_action(
self._api_token,
ihost_obj['uuid'],
ihost_obj['hostname'],
action,
constants.VIM_DEFAULT_TIMEOUT_IN_SECS)
except Exception as e:
LOG.warn(_("No response vim_api %s on action=%s e=%s" %
(ihost_obj['hostname'], action, e)))
self._api_token = None
if action == constants.FORCE_LOCK_ACTION:
pass
else:
# reject continuation if VIM rejects action
raise wsme.exc.ClientSideError(_(
"VIM API Error or Timeout on action = %s "
"Please retry and if problem persists then "
"contact your system administrator.") % action)
if hostupdate.configure_required:
LOG.info("%s Perform configure_ihost." % hostupdate.displayid)
if not ((ihost_obj['hostname']) and (ihost_obj['personality'])):
raise wsme.exc.ClientSideError(
_("Please provision 'hostname' and 'personality'."))
ihost_ret = pecan.request.rpcapi.configure_ihost(
pecan.request.context, ihost_obj)
pecan.request.dbapi.ihost_update(
ihost_obj['uuid'], {'capabilities': ihost_obj['capabilities']})
# Notify maintenance about updated mgmt_ip
ihost_obj['mgmt_ip'] = ihost_ret.mgmt_ip
hostupdate.notify_mtce = True
pecan.request.dbapi.ihost_update(ihost_obj['uuid'],
{'capabilities': ihost_obj['capabilities']})
if constants.TASK_REINSTALLING == ihost_obj.task and \
constants.CONFIG_STATUS_REINSTALL == \
ihost_obj.config_status:
# Clear reinstall flag when reinstall starts
ihost_obj.config_status = None
mtc_response = {'status': None}
nonmtc_change_count = 0
if hostupdate.notify_mtce and not hostupdate.skip_notify_mtce:
nonmtc_change_count = self.check_notify_mtce(myaction, hostupdate)
if nonmtc_change_count > 0:
LOG.info("%s Action %s perform notify_mtce" %
(hostupdate.displayid, myaction))
new_ihost_mtc = ihost_obj.as_dict()
new_ihost_mtc = cutils.removekeys_nonmtce(new_ihost_mtc)
if hostupdate.ihost_orig['invprovision'] == constants.PROVISIONED:
new_ihost_mtc.update({'operation': 'modify'})
else:
new_ihost_mtc.update({'operation': 'add'})
new_ihost_mtc.update({"invprovision": ihost_obj['invprovision']})
if hostupdate.notify_action_lock:
new_ihost_mtc['action'] = constants.LOCK_ACTION
elif hostupdate.notify_action_lock_force:
new_ihost_mtc['action'] = constants.FORCE_LOCK_ACTION
elif myaction == constants.FORCE_UNLOCK_ACTION:
new_ihost_mtc['action'] = constants.UNLOCK_ACTION
if new_ihost_mtc['operation'] == 'add':
mtc_response = mtce_api.host_add(
self._api_token, self._mtc_address, self._mtc_port,
new_ihost_mtc,
constants.MTC_DEFAULT_TIMEOUT_IN_SECS)
elif new_ihost_mtc['operation'] == 'modify':
mtc_response = mtce_api.host_modify(
self._api_token, self._mtc_address, self._mtc_port,
new_ihost_mtc,
constants.MTC_DEFAULT_TIMEOUT_IN_SECS,
3)
else:
LOG.warn("Unsupported Operation: %s" % new_ihost_mtc)
mtc_response = None
if mtc_response is None:
mtc_response = {'status': 'fail',
'reason': 'no response',
'action': 'retry'}
ihost_obj['action'] = constants.NONE_ACTION
hostupdate.ihost_val_update({'action': constants.NONE_ACTION})
if ((mtc_response['status'] == 'pass') or
(nonmtc_change_count == 0) or hostupdate.skip_notify_mtce):
ihost_obj.save()
if hostupdate.ihost_patch['operational'] == \
constants.OPERATIONAL_ENABLED:
self._update_add_ceph_state()
if hostupdate.notify_availability:
if (hostupdate.notify_availability ==
constants.VIM_SERVICES_DISABLED):
imsg_dict = {'availability':
constants.AVAILABILITY_OFFLINE}
else:
imsg_dict = {'availability':
constants.VIM_SERVICES_ENABLED}
if (hostupdate.notify_availability !=
constants.VIM_SERVICES_ENABLED):
LOG.error(_("Unexpected notify_availability = %s" %
hostupdate.notify_availability))
LOG.info(_("%s notify_availability=%s" %
(hostupdate.displayid, hostupdate.notify_availability)))
pecan.request.rpcapi.iplatform_update_by_ihost(
pecan.request.context, ihost_obj['uuid'], imsg_dict)
if hostupdate.bm_type_changed_to_none:
ibm_msg_dict = {}
pecan.request.rpcapi.ibm_deprovision_by_ihost(
pecan.request.context,
ihost_obj['uuid'],
ibm_msg_dict)
# Trigger a system app reapply if the host has been unlocked
if (utils.is_kubernetes_config() and patched_ihost.get('action') in
[constants.UNLOCK_ACTION, constants.FORCE_UNLOCK_ACTION]):
self._reapply_system_app()
elif mtc_response['status'] is None:
raise wsme.exc.ClientSideError(
_("Timeout waiting for maintenance response. "
"Please retry and if problem persists then "
"contact your system administrator."))
else:
if hostupdate.configure_required:
# rollback to unconfigure host as mtce has failed the request
invprovision_state = hostupdate.ihost_orig.get('invprovision') or ""
if invprovision_state != constants.PROVISIONED:
LOG.warn("unconfigure ihost %s provision=%s" %
(ihost_obj.uuid, invprovision_state))
pecan.request.rpcapi.unconfigure_ihost(
pecan.request.context,
ihost_obj)
raise wsme.exc.ClientSideError(_("Operation Rejected: %s. %s.") %
(mtc_response['reason'],
mtc_response['action']))
if hostupdate.notify_vim_add_host:
# Notify the VIM that the host has been added - must be done after
# the host has been added to mtc and saved to the DB.
LOG.info("sysinv notify add host add %s subfunctions=%s" %
(ihost_obj['hostname'], ihost_obj['subfunctions']))
try:
vim_api.vim_host_add(
self._api_token,
ihost_obj['uuid'],
ihost_obj['hostname'],
ihost_obj['subfunctions'],
ihost_obj['administrative'],
ihost_obj['operational'],
ihost_obj['availability'],
ihost_obj['subfunction_oper'],
ihost_obj['subfunction_avail'],
constants.VIM_DEFAULT_TIMEOUT_IN_SECS)
except Exception as e:
LOG.warn(_("No response from vim_api %s e=%s" %
(ihost_obj['hostname'], e)))
self._api_token = None
pass # VIM audit will pickup
# check if ttys_dcd is updated and notify the agent via conductor
# if necessary
if 'ttys_dcd' in hostupdate.delta:
self._handle_ttys_dcd_change(hostupdate.ihost_orig,
hostupdate.ihost_patch['ttys_dcd'])
log_end = cutils.timestamped("ihost_patch_end")
if uptime_update:
LOG.debug("host %s %s patch" % (ihost_obj.hostname,
log_end))
else:
LOG.info("host %s %s patch" % (ihost_obj.hostname,
log_end))
if ('administrative' in hostupdate.delta and
hostupdate.ihost_patch['administrative'] ==
constants.ADMIN_LOCKED):
LOG.info("Update host memory for (%s)" % ihost_obj['hostname'])
pecan.request.rpcapi.update_host_memory(pecan.request.context,
ihost_obj['uuid'])
return Host.convert_with_links(ihost_obj)
def _vim_host_add(self, ihost):
LOG.info("sysinv notify vim add host %s personality=%s" % (
ihost['hostname'], ihost['personality']))
subfunctions = self._update_subfunctions(ihost)
try:
# TODO: if self._api_token is None or \
# self._api_token.is_expired():
# self._api_token = rest_api.get_token()
vim_api.vim_host_add(
self._api_token,
ihost['uuid'],
ihost['hostname'],
subfunctions,
ihost['administrative'],
ihost['operational'],
ihost['availability'],
ihost['subfunction_oper'],
ihost['subfunction_avail'],
constants.VIM_DEFAULT_TIMEOUT_IN_SECS)
except Exception as e:
LOG.warn(_("No response from vim_api %s e=%s" %
(ihost['hostname'], e)))
self._api_token = None
pass # VIM audit will pickup
@staticmethod
def _check_host_delete_during_upgrade():
""" Determine whether host delete is allowed during upgrade
returns: boolean False if not allowed
"""
upgrade = None
try:
upgrade = pecan.request.dbapi.software_upgrade_get_one()
except exception.NotFound:
return True
if upgrade:
loads = pecan.request.dbapi.load_get_list()
to_load = cutils.get_imported_load(loads)
active_controller = utils.HostHelper.get_active_controller()
host_upgrade = objects.host_upgrade.get_by_host_id(
pecan.request.context, active_controller.id)
if ((host_upgrade.target_load != to_load.id) or
(host_upgrade.software_load != to_load.id)):
LOG.info("_check_host_delete_during_upgrade %s sw=%s "
"target=%s load=%s" %
(active_controller.hostname,
host_upgrade.target_load,
host_upgrade.software_load,
to_load.id))
return False
return True
@cutils.synchronized(LOCK_NAME)
@wsme_pecan.wsexpose(None, six.text_type, status_code=204)
def delete(self, ihost_id):
"""Delete an ihost.
"""
if utils.get_system_mode() == constants.SYSTEM_MODE_SIMPLEX:
raise wsme.exc.ClientSideError(_(
"Deleting a host on a simplex system is not allowed."))
ihost = objects.host.get_by_uuid(pecan.request.context,
ihost_id)
# Do not allow profiles to be deleted by system host-delete
if ihost['recordtype'] == "profile":
LOG.error("host %s of recordtype %s cannot be deleted via "
"host-delete command."
% (ihost['uuid'], ihost['recordtype']))
raise exception.HTTPNotFound
if ihost['administrative'] == constants.ADMIN_UNLOCKED:
if ihost.hostname is None:
host = ihost.uuid
else:
host = ihost.hostname
raise exception.HostLocked(action=constants.DELETE_ACTION, host=host)
if not self._check_host_delete_during_upgrade():
raise wsme.exc.ClientSideError(_(
"host-delete rejected: not allowed at this upgrade stage"))
personality = ihost.personality
# allow deletion of unprovisioned locked disabled & offline storage hosts
skip_ceph_checks = (
(not ihost['invprovision'] or
ihost['invprovision'] == constants.UNPROVISIONED) and
ihost['administrative'] == constants.ADMIN_LOCKED and
ihost['operational'] == constants.OPERATIONAL_DISABLED and
ihost['availability'] == constants.AVAILABILITY_OFFLINE)
if (personality is not None and
personality.find(constants.STORAGE_HOSTNAME) != -1 and
not skip_ceph_checks):
num_monitors, required_monitors, quorum_names = \
self._ceph.get_monitors_status(pecan.request.dbapi)
if num_monitors < required_monitors:
raise wsme.exc.ClientSideError(_(
"Only %d storage "
"monitor available. At least %s unlocked and "
"enabled hosts with monitors are required. Please"
" ensure hosts with monitors are unlocked and "
"enabled.") %
(num_monitors, constants.MIN_STOR_MONITORS))
# If it is the last storage node to delete, we need to delete
# ceph osd pools and update additional tier status to "defined"
# if no backend is attached to the tier.
storage_nodes = pecan.request.dbapi.ihost_get_by_personality(
constants.STORAGE)
if len(storage_nodes) == 1:
# TODO(CephPoolsDecouple): rework
# delete osd pools
# It would be nice if we have a ceph API that can delete
# all osd pools at once.
if not utils.is_kubernetes_config():
pools = pecan.request.rpcapi.list_osd_pools(
pecan.request.context)
for ceph_pool in pools:
pecan.request.rpcapi.delete_osd_pool(
pecan.request.context, ceph_pool)
# update tier status
tier_list = pecan.request.dbapi.storage_tier_get_list()
for tier in tier_list:
if (tier.name != constants.SB_TIER_DEFAULT_NAMES[
constants.SB_TIER_TYPE_CEPH] and not tier.forbackendid):
pecan.request.dbapi.storage_tier_update(tier.id,
{'status': constants.SB_TIER_STATUS_DEFINED})
LOG.warn("REST API delete host=%s user_agent=%s" %
(ihost['uuid'], pecan.request.user_agent))
if not pecan.request.user_agent.startswith('vim'):
try:
# TODO: if self._api_token is None or \
# self._api_token.is_expired():
# self._api_token = rest_api.get_token()
vim_api.vim_host_delete(
self._api_token,
ihost.uuid,
ihost.hostname,
constants.VIM_DELETE_TIMEOUT_IN_SECS)
except Exception:
LOG.warn(_("No response from vim_api %s " % ihost['uuid']))
raise wsme.exc.ClientSideError(_("System rejected delete "
"request. Please retry and if problem persists then "
"contact your system administrator."))
if (ihost.hostname and ihost.personality and
ihost.invprovision and
ihost.invprovision == constants.PROVISIONED and
(constants.WORKER in ihost.subfunctions)):
# wait for VIM signal
return
openstack_worker = False
if utils.is_kubernetes_config():
labels = objects.label.get_by_host_id(pecan.request.context, ihost.uuid)
for l in labels:
if (constants.COMPUTE_NODE_LABEL ==
str(l.label_key) + '=' + str(l.label_value)):
openstack_worker = True
break
idict = {'operation': constants.DELETE_ACTION,
'uuid': ihost.uuid,
'invprovision': ihost.invprovision}
mtc_response_dict = mtce_api.host_delete(
self._api_token, self._mtc_address, self._mtc_port,
idict, constants.MTC_DELETE_TIMEOUT_IN_SECS)
LOG.info("Mtce Delete Response: %s", mtc_response_dict)
if mtc_response_dict is None:
mtc_response_dict = {'status': 'fail',
'reason': 'no response',
'action': 'retry'}
# Check mtce response prior to attempting delete
if mtc_response_dict.get('status') != 'pass':
self._vim_host_add(ihost)
if mtc_response_dict.get('reason') != 'no response':
raise wsme.exc.ClientSideError(_("Mtce rejected delete request."
"Please retry and if problem persists then contact your "
"system administrator."))
else:
raise wsme.exc.ClientSideError(_("Timeout waiting for system response. Please wait for a "
"few moments. If the host is not deleted,please retry. If "
"problem persists then contact your system administrator."))
pecan.request.rpcapi.unconfigure_ihost(pecan.request.context,
ihost)
# reset the ceph_mon_dev for the controller node being deleted
if ihost.personality == constants.CONTROLLER:
ceph_mons = pecan.request.dbapi.ceph_mon_get_by_ihost(ihost.uuid)
if ceph_mons and ceph_mons[0].device_path:
pecan.request.dbapi.ceph_mon_update(
ceph_mons[0].uuid, {'device_path': None}
)
# Delete the stor entries associated with this host
istors = pecan.request.dbapi.istor_get_by_ihost(ihost['uuid'])
for stor in istors:
try:
self.istors.delete_stor(stor.uuid)
except Exception as e:
# Do not destroy the ihost if the stor cannot be deleted.
LOG.exception(e)
self._vim_host_add(ihost)
raise wsme.exc.ClientSideError(
_("Failed to delete Storage Volumes associated with this "
"host. Please retry and if problem persists then contact"
" your system administrator."))
# Delete the lvgs entries associated with this host
ilvgs = pecan.request.dbapi.ilvg_get_by_ihost(ihost['uuid'])
for lvg in ilvgs:
try:
self.ilvgs.delete(lvg.uuid)
except Exception as e:
# Do not destroy the ihost if the lvg cannot be deleted.
LOG.exception(e)
self._vim_host_add(ihost)
raise wsme.exc.ClientSideError(
_("Failed to delete Local Volume Group(s). Please retry and "
"if problem persists then contact your system "
"administrator."))
# Delete the pvs entries associated with this host
# Note: pvs should have been deleted via cascade with it's lvg.
# This should be unnecessary
ipvs = pecan.request.dbapi.ipv_get_by_ihost(ihost['uuid'])
for pv in ipvs:
try:
self.ipvs.delete(pv.uuid)
except Exception as e:
# Do not destroy the ihost if the stor cannot be deleted.
self._vim_host_add(ihost)
LOG.exception(e)
raise wsme.exc.ClientSideError(
_("Failed to delete Physical Volume(s). Please retry and if "
"problem persists then contact your system "
"administrator."))
# tell conductor to delete the barbican entry associated with this host (if present)
pecan.request.rpcapi.delete_barbican_secret(pecan.request.context,
ihost.uuid)
# Notify patching to drop the host
if ihost.hostname is not None:
try:
# TODO: if self._api_token is None or \
# self._api_token.is_expired():
# self._api_token = rest_api.get_token()
system = pecan.request.dbapi.isystem_get_one()
patch_api.patch_drop_host(
token=self._api_token,
timeout=constants.PATCH_DEFAULT_TIMEOUT_IN_SECS,
hostname=ihost.hostname,
region_name=system.region_name)
except Exception as e:
LOG.warn(_("No response from drop-host patch api %s e=%s" %
(ihost.hostname, e)))
pass
personality = ihost.personality
if (personality is not None and
personality.find(constants.STORAGE_HOSTNAME) != -1 and
ihost.hostname not in [constants.STORAGE_0_HOSTNAME,
constants.STORAGE_1_HOSTNAME] and
ihost.invprovision in [constants.PROVISIONED,
constants.PROVISIONING]):
self._ceph.host_crush_remove(ihost.hostname)
pecan.request.dbapi.ihost_destroy(ihost_id)
# If the host being removed was an openstack worker node, trigger
# a reapply
if openstack_worker:
self._reapply_system_app()
def _reapply_system_app(self):
try:
db_app = objects.kube_app.get_by_name(
pecan.request.context, constants.HELM_APP_OPENSTACK)
if db_app.status == constants.APP_APPLY_SUCCESS:
LOG.info(
"Reapplying the %s app" % constants.HELM_APP_OPENSTACK)
db_app.status = constants.APP_APPLY_IN_PROGRESS
db_app.progress = None
db_app.save()
pecan.request.rpcapi.perform_app_apply(
pecan.request.context, db_app, False)
else:
LOG.info("%s system app is present but not applied, "
"skipping re-apply" % constants.HELM_APP_OPENSTACK)
except exception.KubeAppNotFound:
LOG.info(
"%s system app not present, skipping re-apply" %
constants.HELM_APP_OPENSTACK)
def _check_upgrade_provision_order(self, personality, hostname):
LOG.info("_check_upgrade_provision_order personality=%s, hostname=%s" %
(personality, hostname))
# If this is a simplex system skip this check; there's no other nodes
simplex = (utils.get_system_mode() == constants.SYSTEM_MODE_SIMPLEX)
if simplex:
return
try:
pecan.request.dbapi.software_upgrade_get_one()
except exception.NotFound:
return
loads = pecan.request.dbapi.load_get_list()
target_load = cutils.get_imported_load(loads)
if personality == constants.STORAGE:
if hostname == constants.STORAGE_0_HOSTNAME:
LOG.warn("Allow storage-0 add during upgrade")
else:
LOG.info("Adding storage, ensure controllers upgraded")
self._check_personality_load(constants.CONTROLLER,
target_load)
@cutils.synchronized(LOCK_NAME)
@wsme_pecan.wsexpose(Host, six.text_type, body=six.text_type)
def upgrade(self, uuid, body):
"""Upgrade the host to the specified load"""
# There must be an upgrade in progress
try:
upgrade = pecan.request.dbapi.software_upgrade_get_one()
except exception.NotFound:
raise wsme.exc.ClientSideError(_(
"host-upgrade rejected: An upgrade is not in progress."))
if upgrade.state in [constants.UPGRADE_ABORTING_ROLLBACK,
constants.UPGRADE_ABORTING]:
raise wsme.exc.ClientSideError(_(
"host-upgrade rejected: Aborting Upgrade."))
# Enforce upgrade order
loads = pecan.request.dbapi.load_get_list()
new_target_load = cutils.get_imported_load(loads)
rpc_ihost = objects.host.get_by_uuid(pecan.request.context, uuid)
simplex = (utils.get_system_mode() == constants.SYSTEM_MODE_SIMPLEX)
# If this is a simplex system skip this check, there's no other nodes
if simplex:
pass
elif rpc_ihost.personality == constants.WORKER:
self._check_personality_load(constants.CONTROLLER, new_target_load)
self._check_personality_load(constants.STORAGE, new_target_load)
elif rpc_ihost.personality == constants.STORAGE:
self._check_personality_load(constants.CONTROLLER, new_target_load)
# Ensure we upgrade storage-0 before other storage nodes
if rpc_ihost.hostname != constants.STORAGE_0_HOSTNAME:
self._check_host_load(constants.STORAGE_0_HOSTNAME,
new_target_load)
elif rpc_ihost.hostname == constants.CONTROLLER_0_HOSTNAME:
self._check_host_load(constants.CONTROLLER_1_HOSTNAME,
new_target_load)
# Check upgrade state
if rpc_ihost.hostname == constants.CONTROLLER_1_HOSTNAME or simplex:
if upgrade.state != constants.UPGRADE_STARTED:
raise wsme.exc.ClientSideError(_(
"host-upgrade rejected: Upgrade not in %s state." %
constants.UPGRADE_STARTED))
elif rpc_ihost.hostname == constants.CONTROLLER_0_HOSTNAME:
if upgrade.state != constants.UPGRADE_UPGRADING_CONTROLLERS:
raise wsme.exc.ClientSideError(_(
"host-upgrade rejected: Upgrade not in %s state." %
constants.UPGRADE_UPGRADING_CONTROLLERS))
elif upgrade.state != constants.UPGRADE_UPGRADING_HOSTS:
raise wsme.exc.ClientSideError(_(
"host-upgrade rejected: Upgrade not in %s state." %
constants.UPGRADE_UPGRADING_HOSTS))
if rpc_ihost.personality == constants.STORAGE:
osd_status = self._ceph.check_osds_down_up(rpc_ihost.hostname, True)
if not osd_status:
raise wsme.exc.ClientSideError(
_("Host %s must be locked and "
"all osds must be down.")
% (rpc_ihost.hostname))
if upgrade.state in [constants.UPGRADE_STARTED]:
LOG.info("host-upgrade check upgrade_refresh %s" %
rpc_ihost.hostname)
force = body.get('force', False) is True
self._semantic_check_upgrade_refresh(upgrade, rpc_ihost, force)
# Update the target load for this host
self._update_load(uuid, body, new_target_load)
if rpc_ihost.hostname == constants.CONTROLLER_1_HOSTNAME:
# When controller-1 is upgraded, we do the data migration
upgrade_update = {'state': constants.UPGRADE_DATA_MIGRATION}
pecan.request.dbapi.software_upgrade_update(upgrade.uuid,
upgrade_update)
# Set upgrade flag so controller-1 will upgrade after install
# This flag is guaranteed to be written on controller-0, since
# controller-1 must be locked to run the host-upgrade command.
open(tsc.CONTROLLER_UPGRADE_FLAG, "w").close()
return Host.convert_with_links(rpc_ihost)
@cutils.synchronized(LOCK_NAME)
@wsme_pecan.wsexpose(Host, six.text_type, body=six.text_type)
def downgrade(self, uuid, body):
"""Downgrade the host to the specified load"""
# There must be an upgrade in progress
try:
upgrade = pecan.request.dbapi.software_upgrade_get_one()
except exception.NotFound:
raise wsme.exc.ClientSideError(_(
"host-downgrade rejected: An upgrade is not in progress."))
loads = pecan.request.dbapi.load_get_list()
new_target_load = cutils.get_active_load(loads)
rpc_ihost = objects.host.get_by_uuid(pecan.request.context, uuid)
disable_storage_monitor = False
simplex = (utils.get_system_mode() == constants.SYSTEM_MODE_SIMPLEX)
# If this is a simplex upgrade just check that we are aborting
if simplex:
if upgrade.state not in [constants.UPGRADE_ABORTING_ROLLBACK,
constants.UPGRADE_ABORTING]:
raise wsme.exc.ClientSideError(
_("host-downgrade rejected: The upgrade must be aborted "
"before downgrading."))
# Check if we're doing a rollback
elif upgrade.state == constants.UPGRADE_ABORTING_ROLLBACK:
if rpc_ihost.hostname == constants.CONTROLLER_0_HOSTNAME:
# Before we downgrade controller-0 during a rollback/reinstall
# we check that all other worker/storage nodes are locked and
# offline. We also disable the storage monitor on controller-1
# and set a flag on controller-1 to indicate we are in a
# rollback. When controller-0 comes up it will check for this
# flag and update its database as necessary.
self._semantic_check_rollback()
if StorageBackendConfig.has_backend_configured(
pecan.request.dbapi, constants.CINDER_BACKEND_CEPH):
disable_storage_monitor = True
open(tsc.UPGRADE_ROLLBACK_FLAG, "w").close()
elif rpc_ihost.hostname == constants.CONTROLLER_1_HOSTNAME:
self._check_host_load(constants.CONTROLLER_0_HOSTNAME,
new_target_load)
else:
raise wsme.exc.ClientSideError(_(
"host-downgrade rejected: Rollback is in progress."))
else:
# Enforce downgrade order
if rpc_ihost.personality == constants.CONTROLLER:
self._check_personality_load(constants.WORKER,
new_target_load)
self._check_personality_load(constants.STORAGE,
new_target_load)
if rpc_ihost.hostname == constants.CONTROLLER_1_HOSTNAME:
self._check_host_load(constants.CONTROLLER_0_HOSTNAME,
new_target_load)
elif rpc_ihost.personality == constants.STORAGE:
self._check_personality_load(constants.WORKER,
new_target_load)
if rpc_ihost.hostname == constants.STORAGE_0_HOSTNAME:
self._check_storage_downgrade(new_target_load)
# else we should be a worker node, no need to check other nodes
# Check upgrade state
if rpc_ihost.hostname in [constants.CONTROLLER_0_HOSTNAME,
constants.CONTROLLER_1_HOSTNAME]:
# The controllers are the last nodes to be downgraded.
# There is no way to continue the upgrade after that,
# so force the user to specifically abort the upgrade
# before doing this.
if upgrade.state != constants.UPGRADE_ABORTING:
raise wsme.exc.ClientSideError(_(
"host-downgrade rejected: Upgrade not in %s state." %
constants.UPGRADE_ABORTING))
if rpc_ihost.hostname == constants.CONTROLLER_1_HOSTNAME:
# Clear upgrade flags so controller-1 will not upgrade
# after install. This flag is guaranteed to be written on
# controller-0, since controller-1 must be locked to run
# the host-downgrade command.
try:
os.remove(tsc.CONTROLLER_UPGRADE_FLAG)
except OSError:
LOG.exception("Failed to remove upgrade flag")
try:
os.remove(tsc.CONTROLLER_UPGRADE_COMPLETE_FLAG)
except OSError:
LOG.exception("Failed to remove upgrade complete flag")
try:
os.remove(tsc.CONTROLLER_UPGRADE_FAIL_FLAG)
except OSError:
LOG.exception("Failed to remove upgrade fail flag")
# Check for new hardware since upgrade-start
force = body.get('force', False) is True
self._semantic_check_downgrade_refresh(upgrade, rpc_ihost, force)
if disable_storage_monitor:
# When we downgrade controller-0 during a rollback we need to
# disable the storage monitor on controller-1. We want to ensure
# that when controller-0 comes up it starts with clean ceph data,
# and does not use any stale data that might be present on
# controller-1.
pecan.request.rpcapi.kill_ceph_storage_monitor(
pecan.request.context)
self._update_load(uuid, body, new_target_load)
return Host.convert_with_links(rpc_ihost)
def _semantic_check_rollback(self):
hosts = pecan.request.dbapi.ihost_get_list()
for host in hosts:
if host.personality not in [constants.WORKER, constants.STORAGE]:
continue
if host.administrative != constants.ADMIN_LOCKED or \
host.availability != constants.AVAILABILITY_OFFLINE:
raise wsme.exc.ClientSideError(
_("All worker and storage hosts must be locked and "
"offline before this operation can proceed"))
def _check_personality_load(self, personality, load):
hosts = pecan.request.dbapi.ihost_get_by_personality(personality)
for host in hosts:
host_upgrade = objects.host_upgrade.get_by_host_id(
pecan.request.context, host.id)
if host_upgrade.target_load != load.id or \
host_upgrade.software_load != load.id:
raise wsme.exc.ClientSideError(
_("All %s hosts must be using load %s before this "
"operation can proceed")
% (personality, load.software_version))
def _check_host_load(self, hostname, load):
host = pecan.request.dbapi.ihost_get_by_hostname(hostname)
host_upgrade = objects.host_upgrade.get_by_host_id(
pecan.request.context, host.id)
if host_upgrade.target_load != load.id or \
host_upgrade.software_load != load.id:
raise wsme.exc.ClientSideError(
_("%s must be using load %s before this operation can proceed")
% (hostname, load.software_version))
def _check_storage_downgrade(self, load):
hosts = pecan.request.dbapi.ihost_get_by_personality(constants.STORAGE)
# Ensure all storage nodes are downgraded before storage-0
for host in hosts:
if host.hostname != constants.STORAGE_0_HOSTNAME:
host_upgrade = objects.host_upgrade.get_by_host_id(
pecan.request.context, host.id)
if host_upgrade.target_load != load.id or \
host_upgrade.software_load != load.id:
raise wsme.exc.ClientSideError(
_("All other %s hosts must be using load %s before "
"this operation can proceed")
% (constants.STORAGE, load.software_version))
def _update_load(self, uuid, body, new_target_load):
force = body.get('force', False) is True
rpc_ihost = objects.host.get_by_uuid(pecan.request.context, uuid)
host_upgrade = objects.host_upgrade.get_by_host_id(
pecan.request.context, rpc_ihost.id)
if host_upgrade.target_load == new_target_load.id:
raise wsme.exc.ClientSideError(
_("%s already targeted to install load %s") %
(rpc_ihost.hostname, new_target_load.software_version))
if rpc_ihost.administrative != constants.ADMIN_LOCKED:
raise wsme.exc.ClientSideError(
_("The host must be locked before performing this operation"))
elif rpc_ihost.invprovision != "provisioned":
raise wsme.exc.ClientSideError(_("The host must be provisioned "
"before performing this operation"))
elif not force and rpc_ihost.availability != "online":
raise wsme.exc.ClientSideError(
_("The host must be online to perform this operation"))
if rpc_ihost.personality == constants.STORAGE:
istors = pecan.request.dbapi.istor_get_by_ihost(rpc_ihost.id)
for stor in istors:
istor_obj = objects.storage.get_by_uuid(pecan.request.context,
stor.uuid)
self._ceph.remove_osd_key(istor_obj['osdid'])
if utils.get_system_mode() != constants.SYSTEM_MODE_SIMPLEX:
pecan.request.rpcapi.upgrade_ihost(pecan.request.context,
rpc_ihost,
new_target_load)
host_upgrade.target_load = new_target_load.id
host_upgrade.save()
# There may be alarms, clear them
entity_instance_id = "%s=%s" % (fm_constants.FM_ENTITY_TYPE_HOST,
rpc_ihost.hostname)
fm_api_obj = fm_api.FaultAPIs()
fm_api_obj.clear_fault(
fm_constants.FM_ALARM_ID_HOST_VERSION_MISMATCH,
entity_instance_id)
if rpc_ihost.availability == "online":
new_ihost_mtc = rpc_ihost.as_dict()
new_ihost_mtc.update({'operation': 'modify'})
new_ihost_mtc.update({'action': constants.REINSTALL_ACTION})
new_ihost_mtc = cutils.removekeys_nonmtce(new_ihost_mtc)
mtc_response = mtce_api.host_modify(
self._api_token, self._mtc_address, self._mtc_port,
new_ihost_mtc, constants.MTC_ADD_TIMEOUT_IN_SECS)
if mtc_response is None:
mtc_response = {'status': 'fail',
'reason': 'no response',
'action': 'retry'}
if mtc_response['status'] != 'pass':
# Report mtc error
raise wsme.exc.ClientSideError(_("Maintenance has returned with "
"a status of %s, reason: %s, recommended action: %s") % (
mtc_response.get('status'),
mtc_response.get('reason'),
mtc_response.get('action')))
@staticmethod
def _validate_ip_in_mgmt_network(ip):
network = pecan.request.dbapi.network_get_by_type(
constants.NETWORK_TYPE_MGMT)
utils.validate_address_within_nework(ip, network)
@staticmethod
def _validate_address_not_allocated(name, ip_address):
"""Validate that address isn't allocated
:param name: Address name to check isn't allocated.
:param ip_address: IP address to check isn't allocated.
"""
if name and ip_address:
try:
address = \
pecan.request.dbapi.address_get_by_address(ip_address)
if address.name != name:
raise exception.AddressAlreadyAllocated(address=ip_address)
except exception.AddressNotFoundByAddress:
pass
try:
address = pecan.request.dbapi.address_get_by_name(name)
if address.address != ip_address:
raise exception.AddressAlreadyAllocated(address=name)
except exception.AddressNotFoundByName:
pass
@staticmethod
def _dnsmasq_mac_exists(mac_addr):
"""Check the dnsmasq.hosts file for an existing mac.
:param mac_addr: mac address to check for.
"""
dnsmasq_hosts_file = tsc.CONFIG_PATH + 'dnsmasq.hosts'
with open(dnsmasq_hosts_file, 'r') as f_in:
for line in f_in:
if mac_addr in line:
return True
return False
@staticmethod
def _no_controllers_exist():
current_ihosts = pecan.request.dbapi.ihost_get_list()
hostnames = [h['hostname'] for h in current_ihosts]
return constants.CONTROLLER_0_HOSTNAME not in hostnames and \
constants.CONTROLLER_1_HOSTNAME not in hostnames
@staticmethod
def _validate_delta(delta):
restricted_updates = ['uuid', 'id', 'created_at', 'updated_at',
'cstatus',
'mgmt_mac', 'mgmt_ip',
'invprovision', 'recordtype',
'ihost_action',
'action_state',
'iconfig_applied',
'iconfig_target']
if not pecan.request.user_agent.startswith('mtce'):
# Allow mtc to modify these through sysinv-api.
mtce_only_updates = ['administrative',
'availability',
'operational',
'subfunction_oper',
'subfunction_avail',
'reserved',
'mtce_info',
'task',
'uptime']
restricted_updates.extend(mtce_only_updates)
if not pecan.request.user_agent.startswith('vim'):
vim_only_updates = ['vim_progress_status']
restricted_updates.extend(vim_only_updates)
intersection = set.intersection(set(delta), set(restricted_updates))
if intersection:
raise wsme.exc.ClientSideError(
_("Change %s contains restricted %s." % (delta, intersection)))
else:
LOG.debug("PASS deltaset=%s restricted_updates %s" %
(delta, intersection))
@staticmethod
def _valid_storage_hostname(hostname):
return bool(re.match('^%s-[0-9]+$' % constants.STORAGE_HOSTNAME,
hostname))
def _validate_hostname(self, hostname, personality):
if personality and personality == constants.WORKER:
# Fix of invalid hostnames
err_tl = 'Name restricted to at most 255 characters.'
err_ic = 'Name may only contain letters, ' \
'numbers, underscores, periods and hyphens.'
myexpression = re.compile("^[\w\.\-]+$")
if not myexpression.match(hostname):
raise wsme.exc.ClientSideError(_(err_ic))
if len(hostname) > 255:
raise wsme.exc.ClientSideError(_(err_tl))
non_worker_hosts = ([constants.CONTROLLER_0_HOSTNAME,
constants.CONTROLLER_1_HOSTNAME])
if (hostname and (hostname in non_worker_hosts) or
hostname.startswith(constants.STORAGE_HOSTNAME)):
raise wsme.exc.ClientSideError(
_("%s Reject attempt to configure "
"invalid hostname for personality %s." %
(hostname, personality)))
else:
if personality and personality == constants.CONTROLLER:
valid_hostnames = [constants.CONTROLLER_0_HOSTNAME,
constants.CONTROLLER_1_HOSTNAME]
if hostname not in valid_hostnames:
raise wsme.exc.ClientSideError(
_("Host with personality=%s can only have a hostname "
"from %s" % (personality, valid_hostnames)))
elif personality and personality == constants.STORAGE:
if not self._valid_storage_hostname(hostname):
raise wsme.exc.ClientSideError(
_("Host with personality=%s can only have a hostname "
"starting with %s-(number)" %
(personality, constants.STORAGE_HOSTNAME)))
else:
raise wsme.exc.ClientSideError(
_("%s: Reject attempt to configure with "
"invalid personality=%s " %
(hostname, personality)))
@staticmethod
def _check_worker(patched_ihost, hostupdate=None):
# Check for valid worker node setup
hostname = patched_ihost.get('hostname') or ""
if not hostname:
raise wsme.exc.ClientSideError(
_("Host %s of personality %s, must be provisioned with a hostname."
% (patched_ihost.get('uuid'), patched_ihost.get('personality'))))
non_worker_hosts = ([constants.CONTROLLER_0_HOSTNAME,
constants.CONTROLLER_1_HOSTNAME])
if (hostname in non_worker_hosts or
hostname.startswith(constants.STORAGE_HOSTNAME)):
raise wsme.exc.ClientSideError(
_("Hostname %s is not allowed for personality 'worker'. "
"Please check hostname and personality." % hostname))
def _controller_storage_node_setup(self, patched_ihost, hostupdate=None):
# Initially set the subfunction of the host to it's personality
if hostupdate:
patched_ihost = hostupdate.ihost_patch
patched_ihost['subfunctions'] = patched_ihost['personality']
if patched_ihost['personality'] == constants.CONTROLLER:
controller_0_exists = False
controller_1_exists = False
current_ihosts = \
pecan.request.dbapi.ihost_get_by_personality(constants.CONTROLLER)
for h in current_ihosts:
if h['hostname'] == constants.CONTROLLER_0_HOSTNAME:
controller_0_exists = True
elif h['hostname'] == constants.CONTROLLER_1_HOSTNAME:
controller_1_exists = True
if controller_0_exists and controller_1_exists:
raise wsme.exc.ClientSideError(
_("Two controller nodes have already been configured. "
"This host can not be configured as a controller."))
# Look up the IP address to use for this controller and set
# the hostname.
if controller_0_exists:
hostname = constants.CONTROLLER_1_HOSTNAME
mgmt_ip = _get_controller_address(hostname)
if hostupdate:
hostupdate.ihost_val_update({'hostname': hostname,
'mgmt_ip': mgmt_ip})
else:
patched_ihost['hostname'] = hostname
patched_ihost['mgmt_ip'] = mgmt_ip
elif controller_1_exists:
hostname = constants.CONTROLLER_0_HOSTNAME
mgmt_ip = _get_controller_address(hostname)
if hostupdate:
hostupdate.ihost_val_update({'hostname': hostname,
'mgmt_ip': mgmt_ip})
else:
patched_ihost['hostname'] = hostname
patched_ihost['mgmt_ip'] = mgmt_ip
else:
raise wsme.exc.ClientSideError(
_("Attempting to provision a controller when none "
"exists. This is impossible."))
# Subfunctions can be set directly via the config file
subfunctions = ','.join(tsc.subfunctions)
if hostupdate:
hostupdate.ihost_val_update({'subfunctions': subfunctions})
else:
patched_ihost['subfunctions'] = subfunctions
elif patched_ihost['personality'] == constants.STORAGE:
# Storage nodes are only allowed if we are configured to use ceph.
if not StorageBackendConfig.has_backend_configured(
pecan.request.dbapi,
constants.SB_TYPE_CEPH
):
raise wsme.exc.ClientSideError(
_("Storage nodes can only be configured if storage "
"cluster is configured for the Ceph backend."))
# Storage nodes are allowed when using the CEPH_STORAGE_MODEL model
stor_model = ceph.get_ceph_storage_model()
if stor_model not in [constants.CEPH_STORAGE_MODEL, constants.CEPH_UNDEFINED_MODEL]:
# Adding storage-0 when storage model is CEPH_UNDEFINED_MODEL will
# set it to CEPH_STORAGE_MODEL.
raise wsme.exc.ClientSideError(
_("Storage nodes can not be configured for "
"the '%s' storage model." % stor_model))
current_storage_ihosts = \
pecan.request.dbapi.ihost_get_by_personality(constants.STORAGE)
current_storage = []
for h in current_storage_ihosts:
if self._valid_storage_hostname(h['hostname']):
current_storage.append(h['hostname'])
max_storage_hostnames = ["storage-%s" % x for x in
range(len(current_storage_ihosts) + 1)]
# Look up IP address to use storage hostname
for h in reversed(max_storage_hostnames):
if h not in current_storage:
hostname = h
mgmt_ip = _get_storage_address(hostname)
LOG.info("Found new hostname=%s mgmt_ip=%s "
"current_storage=%s" %
(hostname, mgmt_ip, current_storage))
break
if patched_ihost['hostname']:
if patched_ihost['hostname'] != hostname:
raise wsme.exc.ClientSideError(
_("Storage name %s not allowed. Expected %s. "
"Storage nodes can be one of: "
"storage-#." %
(patched_ihost['hostname'], hostname)))
if hostupdate:
hostupdate.ihost_val_update({'hostname': hostname,
'mgmt_ip': mgmt_ip})
else:
patched_ihost['hostname'] = hostname
patched_ihost['mgmt_ip'] = mgmt_ip
@staticmethod
def _optimize_delta_handling(delta_handle):
"""Optimize specific patch operations.
Updates delta_handle to identify remaining patch semantics to check.
"""
optimizable = ['location', 'serialid']
if pecan.request.user_agent.startswith('mtce'):
mtc_optimizable = ['operational', 'availability', 'task', 'uptime',
'subfunction_oper', 'subfunction_avail']
optimizable.extend(mtc_optimizable)
for k in optimizable:
if k in delta_handle:
delta_handle.remove(k)
@staticmethod
def _semantic_check_interface_addresses(ihost, interface, min_count=0):
"""
Perform IP address semantic checks on a specific interface.
"""
count = 0
if interface.networktype not in address_api.ALLOWED_NETWORK_TYPES:
return
# Check IPv4 address presence
addresses = pecan.request.dbapi.addresses_get_by_interface(
interface['id'], family=constants.IPV4_FAMILY)
count += len(addresses)
if interface.ipv4_mode == constants.IPV4_STATIC:
if not addresses:
msg = (_("Interface %(ifname)s on host %(host)s is configured "
"for IPv4 static address but has no configured "
"IPv4 address") %
{'host': ihost['hostname'],
'ifname': interface.ifname})
raise wsme.exc.ClientSideError(msg)
# Check IPv6 address presence
addresses = pecan.request.dbapi.addresses_get_by_interface(
interface['id'], family=constants.IPV6_FAMILY)
count += len(addresses)
if interface.ipv6_mode == constants.IPV6_STATIC:
if not addresses:
msg = (_("Interface %(ifname)s on host %(host)s is configured "
"for IPv6 static address but has no configured "
"IPv6 address") %
{'host': ihost['hostname'],
'ifname': interface.ifname})
raise wsme.exc.ClientSideError(msg)
if min_count and (count < min_count):
msg = (_("Expecting at least %(min)s IP address(es) on "
"%(networktype)s interface %(ifname)s; found %(count)s") %
{'min': min_count,
'networktype': interface.networktype,
'ifname': interface.ifname,
'count': count})
raise wsme.exc.ClientSideError(msg)
@staticmethod
def _semantic_check_sriov_interface(host, interface, force_unlock=False):
"""
Perform semantic checks on an SRIOV interface.
"""
if (force_unlock or
interface.ifclass != constants.INTERFACE_CLASS_PCI_SRIOV):
return
if_configured_sriov_numvfs = interface.sriov_numvfs
if not if_configured_sriov_numvfs:
return
ports = pecan.request.dbapi.port_get_by_host_interface(
host['id'], interface.id)
for p in ports:
if (p.sriov_vfs_pci_address and
if_configured_sriov_numvfs ==
len(p.sriov_vfs_pci_address.split(','))):
LOG.info("check sriov_numvfs=%s sriov_vfs_pci_address=%s" %
(if_configured_sriov_numvfs, p.sriov_vfs_pci_address))
break
else:
msg = (_("Expecting number of interface sriov_numvfs=%s. "
"Please wait a few minutes for inventory update and "
"retry host-unlock." %
if_configured_sriov_numvfs))
LOG.info(msg)
pecan.request.rpcapi.update_sriov_config(
pecan.request.context,
host['uuid'])
raise wsme.exc.ClientSideError(msg)
def _semantic_check_unlock_upgrade(self, ihost, force_unlock=False):
"""
Perform semantic checks related to upgrades prior to unlocking host.
"""
if ihost['hostname'] != constants.CONTROLLER_1_HOSTNAME:
return
# Don't allow unlock of controller-1 if it is being upgraded
try:
upgrade = pecan.request.dbapi.software_upgrade_get_one()
except exception.NotFound:
# No upgrade in progress
return
if upgrade.state == constants.UPGRADE_DATA_MIGRATION:
msg = _("Can not unlock %s while migrating data. "
"Wait for data migration to complete." % ihost['hostname'])
raise wsme.exc.ClientSideError(msg)
elif upgrade.state == constants.UPGRADE_DATA_MIGRATION_FAILED:
msg = _("Can not unlock %s because data migration "
"failed. Please abort upgrade and downgrade host." %
ihost['hostname'])
raise wsme.exc.ClientSideError(msg)
# Check for new hardware since upgrade-start
self._semantic_check_upgrade_refresh(upgrade, ihost, force_unlock)
@staticmethod
def _semantic_check_oam_interface(ihost):
"""
Perform semantic checks against oam interface to ensure validity of
the node configuration prior to unlocking it.
"""
interfaces = (
pecan.request.dbapi.iinterface_get_by_ihost(ihost['uuid']))
for interface in interfaces:
if interface.networktype == constants.NETWORK_TYPE_OAM:
break
else:
msg = _("Can not unlock a controller host without an oam "
"interface. "
"Add an oam interface before re-attempting this command.")
raise wsme.exc.ClientSideError(msg)
@staticmethod
def _semantic_check_interface_datanets(interface):
"""
Perform data network semantics on a specific interface to ensure
that any data networks that have special requirements on the
interface have been satisfied.
"""
if interface.ifclass != constants.NETWORK_TYPE_DATA:
return
ifdatanets = \
pecan.request.dbapi.interface_datanetwork_get_by_interface(
interface.uuid)
# Check for VXLAN data networks that require IP addresses
for ifdn in ifdatanets:
if ifdn.datanetwork_network_type != \
constants.DATANETWORK_TYPE_VXLAN:
continue
dn = pecan.request.dbapi.datanetwork_get(ifdn.datanetwork_uuid)
if not dn.multicast_group:
# static range; fallback to generic check
continue
# Check for address family specific ranges
address = netaddr.IPAddress(dn.multicast_group)
if ((address.version == constants.IPV4_FAMILY) and
(interface.ipv4_mode == constants.IPV4_DISABLED or not
interface.ipv4_mode)):
msg = (_("Interface %(ifname)s is associated to VXLAN "
"data network %(name)s which requires an "
"IPv4 address") %
{'ifname': interface.ifname,
'name': ifdn.datanetwork_name})
raise wsme.exc.ClientSideError(msg)
if ((address.version == constants.IPV6_FAMILY) and
(interface.ipv6_mode == constants.IPV6_DISABLED or not
interface.ipv6_mode)):
msg = (_("Interface %(ifname)s is associated to VXLAN "
"data network %(name)s which requires an "
"IPv6 address") %
{'ifname': interface.ifname,
'name': ifdn.datanetwork_name})
raise wsme.exc.ClientSideError(msg)
# Check for at least 1 address if no ranges exist yet
if ((interface.ipv4_mode == constants.IPV4_DISABLED) and
(interface.ipv6_mode == constants.IPV6_DISABLED) or
(not interface.ipv4_mode and not interface.ipv6_mode)):
msg = (_("Interface %(ifname)s is associated to VXLAN "
"data network %(name)s which requires an IP "
"address") %
{'ifname': interface.ifname,
'name': ifdn.datanetwork_name})
raise wsme.exc.ClientSideError(msg)
@staticmethod
def _semantic_check_non_accelerated_interface_support(interface):
"""
Perform semantics checks against interfaces to ensure they are
supported in vswitch as native DPDK interfaces.
"""
ports = pecan.request.dbapi.ethernet_port_get_by_interface(interface.uuid)
for p in ports:
if not p.dpdksupport:
msg = _("%s is non-accelerated interface which "
"is not supported by current vswitch" % p.name)
raise wsme.exc.ClientSideError(msg)
def _semantic_check_data_interfaces(
self, ihost, kubernetes_config, force_unlock=False):
"""
Perform semantic checks against data interfaces to ensure validity of
the node configuration prior to unlocking it.
"""
ihost_iinterfaces = (
pecan.request.dbapi.iinterface_get_by_ihost(ihost['uuid']))
vswitch_type = utils.get_vswitch_type()
data_interface_configured = False
for iif in ihost_iinterfaces:
if ((vswitch_type == constants.VSWITCH_TYPE_OVS_DPDK) and
(iif.ifclass == constants.INTERFACE_CLASS_DATA)):
self._semantic_check_non_accelerated_interface_support(iif)
self._semantic_check_interface_datanets(iif)
self._semantic_check_interface_addresses(ihost, iif)
if not iif.ifclass:
continue
self._semantic_check_sriov_interface(ihost, iif, force_unlock)
if iif.ifclass == constants.NETWORK_TYPE_DATA:
data_interface_configured = True
if not data_interface_configured and not kubernetes_config:
msg = _("Can not unlock a worker host without data interfaces. "
"Add at least one data interface before re-attempting "
"this command.")
raise wsme.exc.ClientSideError(msg)
def _semantic_check_data_addresses(self, ihost):
"""
Perform semantic checks against data addresses to ensure validity of
the node configuration prior to unlocking it. Ensure that there is
only 1 IP address configured when SDN is configured otherwise the SDN
controller will be confused on won't know how to map the VXLAN VTEP
endpoints.
"""
sdn_enabled = utils.get_sdn_enabled()
if not sdn_enabled:
return
address_count = 0
interfaces = pecan.request.dbapi.iinterface_get_by_ihost(ihost['uuid'])
for iface in interfaces:
if iface.networktype != constants.NETWORK_TYPE_DATA:
continue
addresses = (
pecan.request.dbapi.addresses_get_by_interface(iface['uuid']))
address_count += len(addresses)
if address_count > 1:
msg = _("Can not unlock a worker host with multiple data "
"addresses while in SDN mode.")
raise wsme.exc.ClientSideError(msg)
@staticmethod
def _semantic_check_vswitch_type_attributes(ihost):
"""
Perform semantic checks host label openstack-compute-node if ovs or ovs-dpdk
vswitch type is enabled since allocating 2M hugepage is needed
validity of the node configuration prior to unlocking it.
"""
vswitch_type = utils.get_vswitch_type()
if vswitch_type == constants.VSWITCH_TYPE_NONE:
return
# Check whether compute_label has been assigned
if utils.is_openstack_compute(ihost) is not True:
raise wsme.exc.ClientSideError(
_("Can not unlock worker host %s without "
" %s label if config %s. Action: assign "
"%s label for this host prior to unlock."
% (ihost['hostname'], helm_common.LABEL_COMPUTE_LABEL,
vswitch_type, helm_common.LABEL_COMPUTE_LABEL)))
@staticmethod
def _semantic_check_data_vrs_attributes(ihost):
"""
Perform semantic checks against data-vrs specific attributes to ensure
validity of the node configuration prior to unlocking it.
"""
vswitch_type = utils.get_vswitch_type()
if vswitch_type != constants.VSWITCH_TYPE_NUAGE_VRS:
return
# Check whether the vsc_controllers have been configured
if not ihost['vsc_controllers']:
raise wsme.exc.ClientSideError(
_("Can not unlock worker host %s without "
"vsc_controllers. Action: Configure "
"vsc_controllers for this host prior to unlock."
% ihost['hostname']))
def _semantic_check_data_routes(self, ihost):
"""
Perform semantic checks against data routes to ensure that any routes
that are configured are reachable via some local address. This check
is already performed before allowing routes to be added but is repeated
here because local static IP addresses are not preserved when interface
profiles are used. This is a reminder to the operator that even though
a profile was used to setup interfaces and routes they still need to
configure IP addresses. Alternatively, the operation can setup an IP
address pool and link that to the interface and addresses will
automatically be added when the profile is applied.
"""
routes = pecan.request.dbapi.routes_get_by_host(ihost['id'])
for r in routes:
route = r.as_dict()
try:
self.routes._check_reachable_gateway(
route['interface_id'], route)
except exception.RouteGatewayNotReachable:
msg = _("Can not unlock a worker host with routes that are "
"not reachable via a local IP address. Add an IP "
"address in the same subnet as each route gateway "
"address before re-attempting this command.")
raise wsme.exc.ClientSideError(msg)
def _semantic_check_sdn_attributes(self, ihost):
"""
Perform semantic checks when SDN Networking is configured to ensure
that at least one SDN controller is configured. Additionally,
check if the Neutron service parameters for ML2 driver,
and L2 Agent have been configured properly.
"""
sdn_enabled = utils.get_sdn_enabled()
if not sdn_enabled:
return
# check sdn controller list
try:
msg = _("An enabled SDN controller is required when SDN is configured. "
"Please configure an SDN controller before re-attempting this "
"command.")
sdn_controllers = pecan.request.dbapi.sdn_controller_get_list()
if not sdn_controllers or len(sdn_controllers) == 0:
raise wsme.exc.ClientSideError(msg)
for controller in sdn_controllers:
if (controller and (controller.state ==
constants.SDN_CONTROLLER_STATE_ENABLED)):
break
else:
raise wsme.exc.ClientSideError(msg)
except NoResultFound:
raise wsme.exc.ClientSideError(msg)
# check to see if Neutron ML2 service parameters have been configured
neutron_parameters = []
for section in [constants.SERVICE_PARAM_SECTION_NETWORK_ML2,
constants.SERVICE_PARAM_SECTION_NETWORK_ML2_ODL,
constants.SERVICE_PARAM_SECTION_NETWORK_DEFAULT]:
try:
parm_list = pecan.request.dbapi.service_parameter_get_all(
service=constants.SERVICE_TYPE_NETWORK,
section=section)
neutron_parameters = neutron_parameters + parm_list
except NoResultFound:
msg = _("Cannot unock a worker host without %s->%s "
",SDN service parameters being configured. "
"Add appropriate service parameters before "
"re-attempting this command." %
(constants.SERVICE_TYPE_NETWORK, section))
raise wsme.exc.ClientSideError(msg)
# The service parameter general semantic checks for Mandatory
# parameters happen within the service parameter schema. However
# doing it there for Neutron ML2 would require us to seed the
# DB with default values for these service params as well
# as add further complexity to the parameter validation logic.
#
# It is simpler to do that check here prior to Unlock.
sdn_service_parameters = \
constants.SERVICE_PARAM_NETWORK_ML2_COMPULSORY
for sdn_param in sdn_service_parameters:
found = False
for param in neutron_parameters:
# value validation is done in service_parameter
if param['name'] == sdn_param:
found = True
break
if not found:
msg = _("Cannot unlock a worker host without "
"\"%s\" SDN service parameter configured. "
"Add service parameter before re-attempting "
"this command." % sdn_param)
raise wsme.exc.ClientSideError(msg)
@staticmethod
def _auto_adjust_memory_for_node(ihost, node):
"""
Detect whether required reserved memory has changed (eg., due to patch
that changes common/constants.py). If reserved memory is larger than
the current setting, push that change to the database. Automatically
adjust pending 2M and 1G memory based on the delta of required reserve
and previous setting.
"""
# Determine required platform reserved memory for this numa node
low_core = cutils.is_low_core_system(ihost, pecan.request.dbapi)
reserved = cutils. \
get_required_platform_reserved_memory(ihost, node['numa_node'], low_core)
# Determine configured memory for this numa node
mems = pecan.request.dbapi.imemory_get_by_inode(node['id'])
# Make adjustment to 2M and 1G hugepages to accomodate an
# increase in platform reserved memory.
for m in mems:
# ignore updates when no change required
if m.platform_reserved_mib is None or \
m.platform_reserved_mib == reserved:
continue
if m.platform_reserved_mib > reserved:
LOG.info("%s auto_adjust_memory numa_node=%d, "
"keep platform_reserved=%d > required=%d"
% (ihost['hostname'], node['numa_node'],
m.platform_reserved_mib, reserved))
continue
# start with current measured hugepage
if m.vm_hugepages_nr_2M is not None:
n_2M = m.vm_hugepages_nr_2M
else:
n_2M = None
if m.vm_hugepages_nr_1G is not None:
n_1G = m.vm_hugepages_nr_1G
else:
n_1G = None
# adjust current measurements
d_MiB = reserved - m.platform_reserved_mib
d_2M = int(d_MiB / constants.MIB_2M)
d_1G = int((d_MiB + 512) / constants.MIB_1G)
if n_2M is not None and n_2M - d_2M > 0:
d_1G = 0
n_2M -= d_2M
else:
d_2M = 0
if n_1G is not None and n_1G - d_1G > 0:
n_1G -= d_1G
else:
d_1G = 0
# override with pending values
if m.vm_hugepages_nr_2M_pending is not None:
n_2M = m.vm_hugepages_nr_2M_pending
if m.vm_hugepages_nr_1G_pending is not None:
n_1G = m.vm_hugepages_nr_1G_pending
values = {}
values.update({'platform_reserved_mib': reserved})
if n_2M is not None:
values.update({'vm_hugepages_nr_2M_pending': n_2M})
if n_1G is not None:
values.update({'vm_hugepages_nr_1G_pending': n_1G})
LOG.info("%s auto_adjust_memory numa_node=%d, "
"+2M=%d, +1G=%d, values=%s"
% (ihost['hostname'], node['numa_node'],
-d_2M, -d_1G, values))
pecan.request.dbapi.imemory_update(m.uuid, values)
return None
@staticmethod
def _semantic_check_memory_for_node(ihost, node):
"""
Perform memory semantic checks on a specific numa node.
"""
# Determine the allocated memory for this numa node
total_allocated_platform_reserved_mib = 0
mems = pecan.request.dbapi.imemory_get_by_inode(node['id'])
pending_2M_memory = False
pending_1G_memory = False
for m in mems:
memtotal = m.node_memtotal_mib
allocated = m.platform_reserved_mib
if m.hugepages_configured:
if m.vswitch_hugepages_reqd is not None:
allocated += m.vswitch_hugepages_reqd * m.vswitch_hugepages_size_mib
else:
allocated += m.vswitch_hugepages_nr * m.vswitch_hugepages_size_mib
if m.vm_hugepages_nr_2M_pending is not None:
allocated += constants.MIB_2M * m.vm_hugepages_nr_2M_pending
pending_2M_memory = True
if m.vm_hugepages_nr_1G_pending is not None:
allocated += constants.MIB_1G * m.vm_hugepages_nr_1G_pending
pending_1G_memory = True
elif m.vm_hugepages_nr_1G:
allocated += constants.MIB_1G * m.vm_hugepages_nr_1G
LOG.debug("MemTotal=%s allocated=%s" % (memtotal, allocated))
if memtotal < allocated:
msg = (_("Rejected: Total allocated memory exceeds the total memory of "
"%(host)s numa node %(node)s "
) %
{'host': ihost['hostname'],
'node': node['numa_node']})
raise wsme.exc.ClientSideError(msg)
total_allocated_platform_reserved_mib += m.platform_reserved_mib
return (total_allocated_platform_reserved_mib,
pending_2M_memory, pending_1G_memory)
@staticmethod
def _align_pending_memory(ihost, align_2M_memory, align_1G_memory):
"""
Update pending fields as required without clearing other settings.
"""
ihost_inodes = pecan.request.dbapi.inode_get_by_ihost(ihost['uuid'])
for node in ihost_inodes:
mems = pecan.request.dbapi.imemory_get_by_inode(node['id'])
for m in mems:
values = {}
if (m.vm_hugepages_nr_2M_pending is None and
m.vm_hugepages_nr_2M and align_2M_memory):
values.update({'vm_hugepages_nr_2M_pending':
m.vm_hugepages_nr_2M})
if (m.vm_hugepages_nr_1G_pending is None and
m.vm_hugepages_nr_1G and align_1G_memory):
values.update({'vm_hugepages_nr_1G_pending':
m.vm_hugepages_nr_1G})
if values:
LOG.info("%s align_pending_memory uuid=%s" %
(ihost['hostname'], values))
pecan.request.dbapi.imemory_update(m.uuid, values)
@staticmethod
def _update_huge_pages(ihost):
"""
Update the host huge pages.
"""
ihost_inodes = pecan.request.dbapi.inode_get_by_ihost(ihost['uuid'])
labels = pecan.request.dbapi.label_get_by_host(ihost['uuid'])
vswitch_type = utils.get_vswitch_type()
for node in ihost_inodes:
mems = pecan.request.dbapi.imemory_get_by_inode(node['id'])
for m in mems:
if m.hugepages_configured:
value = {}
vs_hugepages_nr = m.vswitch_hugepages_nr
# allocate the default vswitch huge pages if required
if vswitch_type != constants.VSWITCH_TYPE_NONE and \
vs_hugepages_nr == 0:
vs_hugepages_nr = constants.VSWITCH_MEMORY_MB \
// m.vswitch_hugepages_size_mib
value.update({'vswitch_hugepages_nr': vs_hugepages_nr})
vm_hugepages_nr_2M = m.vm_hugepages_nr_2M_pending \
if m.vm_hugepages_nr_2M_pending is not None \
else m.vm_hugepages_nr_2M
vm_hugepages_nr_1G = m.vm_hugepages_nr_1G_pending \
if m.vm_hugepages_nr_1G_pending is not None \
else m.vm_hugepages_nr_1G
# calculate 90% 2M pages if the huge pages have not been
# allocated and the compute label is set
if cutils.has_openstack_compute(labels) and \
vm_hugepages_nr_2M == 0 and \
vm_hugepages_nr_1G == 0 and \
cutils.is_default_huge_pages_required(ihost):
vm_hugepages_nr_2M = m.vm_hugepages_possible_2M * 0.9
value.update({'vm_hugepages_nr_2M': vm_hugepages_nr_2M})
vm_hugepages_4K = \
(m.node_memtotal_mib - m.platform_reserved_mib)
vm_hugepages_4K -= \
(vs_hugepages_nr * m.vswitch_hugepages_size_mib)
vm_hugepages_4K -= \
(constants.MIB_2M * vm_hugepages_nr_2M)
vm_hugepages_4K -= \
(constants.MIB_1G * vm_hugepages_nr_1G)
vm_hugepages_4K = \
(constants.NUM_4K_PER_MiB * vm_hugepages_4K)
# Clip 4K pages
min_4K = 32 * constants.Ki / 4
if vm_hugepages_4K < min_4K:
vm_hugepages_4K = 0
value.update({'vm_hugepages_nr_4K': vm_hugepages_4K})
LOG.info("Updating mem values of host(%s) node(%d): %s" %
(ihost['hostname'], node['id'], str(value)))
pecan.request.dbapi.imemory_update(m.uuid, value)
@staticmethod
def _semantic_mtc_check_action(hostupdate, action):
"""
Perform semantic checks with patch action vs current state
returns: notify_mtc_check_action
"""
notify_mtc_check_action = True
ihost = hostupdate.ihost_orig
patched_ihost = hostupdate.ihost_patch
if action in [constants.VIM_SERVICES_DISABLED,
constants.VIM_SERVICES_DISABLE_FAILED,
constants.VIM_SERVICES_DISABLE_EXTEND,
constants.VIM_SERVICES_ENABLED,
constants.VIM_SERVICES_DELETE_FAILED]:
# These are not mtce actions
return notify_mtc_check_action
LOG.info("%s _semantic_mtc_check_action %s" %
(hostupdate.displayid, action))
# Semantic Check: Auto-Provision: Reset, Reboot or Power-On case
if ((cutils.host_has_function(ihost, constants.WORKER)) and
(ihost['administrative'] == constants.ADMIN_LOCKED) and
((patched_ihost['action'] == constants.RESET_ACTION) or
(patched_ihost['action'] == constants.REBOOT_ACTION) or
(patched_ihost['action'] == constants.POWERON_ACTION) or
(patched_ihost['action'] == constants.POWEROFF_ACTION))):
notify_mtc_check_action = True
return notify_mtc_check_action
def _bm_semantic_check_and_update(self, ohost, phost, delta, patch_obj,
current_ihosts=None, hostupdate=None):
""" Parameters:
ohost: object original host
phost: mutable dictionary patch host
delta: default keys changed
patch_obj: all changed paths
returns bm_type_changed_to_none
"""
# NOTE: since the bm_mac is still in the DB; this is just to disallow user to modify it.
if 'bm_mac' in delta:
raise wsme.exc.ClientSideError(
_("Patching Error: can't replace non-existent object 'bm_mac' "))
bm_type_changed_to_none = False
bm_set = {'bm_type',
'bm_ip',
'bm_username',
'bm_password'}
password_exists = any(p['path'] == '/bm_password' for p in patch_obj)
if not (delta.intersection(bm_set) or password_exists):
return bm_type_changed_to_none
if hostupdate:
hostupdate.notify_mtce = True
patch_bm_password = None
for p in patch_obj:
if p['path'] == '/bm_password':
patch_bm_password = p['value']
password_exists = password_exists and patch_bm_password is not None
bm_type_orig = ohost.get('bm_type') or ""
bm_type_patch = phost.get('bm_type') or ""
if bm_type_patch.lower() == 'none':
bm_type_patch = ''
if (not bm_type_patch) and (bm_type_orig != bm_type_patch):
LOG.info("bm_type None from %s to %s." %
(ohost['bm_type'], phost['bm_type']))
bm_type_changed_to_none = True
if 'bm_ip' in delta:
obm_ip = ohost['bm_ip'] or ""
nbm_ip = phost['bm_ip'] or ""
LOG.info("bm_ip in delta=%s obm_ip=%s nbm_ip=%s" %
(delta, obm_ip, nbm_ip))
if obm_ip != nbm_ip:
if (pecan.request.user_agent.startswith('mtce') and
not bm_type_changed_to_none):
raise wsme.exc.ClientSideError(
_("%s: Rejected: %s Board Management "
"controller IP Address is not user-modifiable." %
(constants.REGION_PRIMARY, phost['hostname'])))
if (phost['bm_ip'] or phost['bm_type'] or phost['bm_username']):
if (not phost['bm_type'] or
(phost['bm_type'] and phost['bm_type'].lower() ==
constants.BM_TYPE_NONE)) and not bm_type_changed_to_none:
raise wsme.exc.ClientSideError(
_("%s: Rejected: Board Management controller Type "
"is not provisioned. Provisionable values: 'bmc'."
% phost['hostname']))
elif not phost['bm_username']:
raise wsme.exc.ClientSideError(
_("%s: Rejected: Board Management controller username "
"is not configured." % phost['hostname']))
# Semantic Check: Validate BM type against supported list
# ilo, quanta is kept for backwards compatability only
valid_bm_type_list = [None, 'None', constants.BM_TYPE_NONE,
constants.BM_TYPE_GENERIC,
'ilo', 'ilo3', 'ilo4', 'quanta']
if not phost['bm_type']:
phost['bm_type'] = None
if not (phost['bm_type'] in valid_bm_type_list):
raise wsme.exc.ClientSideError(
_("%s: Rejected: '%s' is not a supported board management "
"type. Must be one of %s" %
(phost['hostname'],
phost['bm_type'],
valid_bm_type_list)))
bm_type_str = phost['bm_type']
if (phost['bm_type'] and
bm_type_str.lower() != constants.BM_TYPE_NONE):
LOG.info("Updating bm_type from %s to %s" %
(phost['bm_type'], constants.BM_TYPE_GENERIC))
phost['bm_type'] = constants.BM_TYPE_GENERIC
if hostupdate:
hostupdate.ihost_val_update(
{'bm_type': constants.BM_TYPE_GENERIC})
else:
phost['bm_type'] = None
if hostupdate:
hostupdate.ihost_val_update({'bm_type': None})
if (phost['bm_type'] and phost['bm_ip'] and
(ohost['bm_ip'] != phost['bm_ip'])):
if not cutils.is_valid_ip(phost['bm_ip']):
raise wsme.exc.ClientSideError(
_("%s: Rejected: Board Management controller IP Address "
"is not valid." % phost['hostname']))
if current_ihosts and ('bm_ip' in phost):
bm_ips = [h['bm_ip'] for h in current_ihosts]
if phost['bm_ip'] and (phost['bm_ip'] in bm_ips):
raise wsme.exc.ClientSideError(
_("Host-add Rejected: bm_ip %s already exists") % phost['bm_ip'])
# Update barbican with updated board management credentials (if supplied)
if (ohost['bm_username'] and phost['bm_username'] and
(ohost['bm_username'] != phost['bm_username'])):
if not password_exists:
raise wsme.exc.ClientSideError(
_("%s Rejected: username change attempt from %s to %s "
"without corresponding password." %
(phost['hostname'],
ohost['bm_username'],
phost['bm_username'])))
if password_exists and patch_bm_password:
pecan.request.rpcapi.create_barbican_secret(pecan.request.context,
phost['uuid'],
patch_bm_password)
LOG.info("%s bm semantic checks for user_agent %s passed" %
(phost['hostname'], pecan.request.user_agent))
return bm_type_changed_to_none
@staticmethod
def _semantic_check_vsc_controllers(ihost, vsc_controllers):
"""
Perform semantic checking for vsc_controllers attribute.
:param ihost: unpatched ihost dictionary
:param vsc_controllers: attribute supplied in patch
"""
# Don't expose the vsc_controllers field if we are not configured with
# the nuage_vrs vswitch or we are not a worker node.
vswitch_type = utils.get_vswitch_type()
if (vswitch_type != constants.VSWITCH_TYPE_NUAGE_VRS or
ihost['personality'] != constants.WORKER):
raise wsme.exc.ClientSideError(
_("The vsc_controllers property is not applicable to this "
"host."))
# When semantic checking a new host the administrative key will not
# be in the dictionary.
if 'administrative' in ihost and ihost['administrative'] != constants.ADMIN_LOCKED:
raise wsme.exc.ClientSideError(
_("Host must be locked before updating vsc_controllers."))
if vsc_controllers:
vsc_list = vsc_controllers.split(',')
if len(vsc_list) != 2:
raise wsme.exc.ClientSideError(
_("Rejected: two VSC controllers (active and standby) "
"must be specified (comma separated)."))
for vsc_ip_str in vsc_list:
try:
vsc_ip = netaddr.IPAddress(vsc_ip_str)
if vsc_ip.version != 4:
raise wsme.exc.ClientSideError(
_("Invalid vsc_controller IP version - only IPv4 "
"supported"))
except netaddr.AddrFormatError:
raise wsme.exc.ClientSideError(
_("Rejected: invalid VSC controller IP address: %s" %
vsc_ip_str))
@staticmethod
def _semantic_check_cinder_volumes(ihost):
"""
Perform semantic checking for cinder volumes storage
:param ihost_uuid: uuid of host with controller functionality
"""
# deny unlock if cinder-volumes is not configured on a controller host
if StorageBackendConfig.has_backend(pecan.request.dbapi,
constants.CINDER_BACKEND_LVM):
msg = _("Cinder's LVM backend is enabled. "
"A configured cinder-volumes PV is required "
"on host %s prior to unlock.") % ihost['hostname']
host_pvs = pecan.request.dbapi.ipv_get_by_ihost(ihost['uuid'])
for pv in host_pvs:
if pv.lvm_vg_name == constants.LVG_CINDER_VOLUMES:
if pv.pv_state not in [constants.PV_ADD, constants.PROVISIONED]:
raise wsme.exc.ClientSideError(msg)
break
else:
raise wsme.exc.ClientSideError(msg)
@staticmethod
def _semantic_check_filesystem_sizes(ihost):
"""
Perform checks for filesystem consistency across controllers
:param ihost: host information of host with controller functionality
"""
# Unlocking the active controller happens only after running
# config_controller or on a one-node system, so this check isn't
# needed in such scenarios
if (utils.is_host_active_controller(ihost) or
utils.is_host_simplex_controller(ihost)):
return
# The check should only happen the first time the standby controller
# is unlocked, so we check if the node has already been provisioned
# or is in a provisioning state (meaning the unlock is in progress)
# After the second controller is provisioned, the filesystem resize
# consistency checks prevent any inconsistencies between nodes
if (ihost['invprovision'] and
ihost['invprovision'] != constants.UNPROVISIONED):
LOG.info("Controller host %s provisioning or already provisioned. "
"Skipping filesystem checks." % ihost['hostname'])
return
active_controller = utils.HostHelper.get_active_controller()
ihost_ilvgs = pecan.request.dbapi.ilvg_get_by_ihost(active_controller.uuid)
for lvg in ihost_ilvgs:
if lvg.lvm_vg_name == constants.LVG_CGTS_VG:
if (not lvg.lvm_vg_size or not lvg.lvm_vg_total_pe):
# Should not happen for active controller, but we should check
# this anyway.
raise wsme.exc.ClientSideError(
_("Active controller %s volume group not yet inventoried.") %
constants.LVG_CGTS_VG)
lvm_vg_used_pe = int(lvg.lvm_vg_total_pe) - int(lvg.lvm_vg_free_pe)
active_controller_used = (
int(lvg.lvm_vg_size) * lvm_vg_used_pe / int(lvg.lvm_vg_total_pe))
# For the standby controller the PVs are not yet allocated to the volume
# group, so we can't get the size directly from volume-group info
# For the standby controller the allocated space is the sum between:
# - cgts-vg space allocated by kickstarts
# - partition PVs assigned to cgts-vg
# - disk PVs assigned to cgts-vg
standby_controller_allocated_space = 0
standby_pvs = pecan.request.dbapi.ipv_get_by_ihost(ihost['uuid'])
for pv in standby_pvs:
if pv.lvm_vg_name == constants.LVG_CGTS_VG:
if pv.lvm_pv_size:
standby_controller_allocated_space += int(
pv.lvm_pv_size)
elif pv.pv_type == constants.PV_TYPE_PARTITION:
part_info = pecan.request.dbapi.partition_get_by_ipv(pv['uuid'])
standby_controller_allocated_space += int(
part_info[0].size_mib) * (1024**2)
elif pv.pv_type == constants.PV_TYPE_DISK:
disk_info = pecan.request.dbapi.idisk_get_by_ipv(pv['uuid'])
standby_controller_allocated_space += int(
disk_info[0].size_mib) * (1024**2)
LOG.info("Active controller filesystem space used: %s" %
str(active_controller_used))
LOG.info("Standby controller filesystem allocated space: %s" %
str(standby_controller_allocated_space))
if (active_controller_used > standby_controller_allocated_space):
# Round up the needed space from float to integer
needed_space = math.ceil(float(
active_controller_used -
standby_controller_allocated_space) / (1024 ** 3))
msg = _("Standby controller does not have enough space allocated to "
"%(vg_name)s volume-group in order to create all filesystems. "
"Please assign an extra %(needed)d GB to the volume group.") % {
'vg_name': constants.LVG_CGTS_VG, 'needed': needed_space}
raise wsme.exc.ClientSideError(msg)
@staticmethod
def _semantic_check_storage_backend(ihost):
"""
Perform semantic checking for storage backends
:param ihost_uuid: uuid of host with controller functionality
"""
# deny operation if any storage backend is either configuring or in error
backends = pecan.request.dbapi.storage_backend_get_list()
for bk in backends:
if bk['state'] != constants.SB_STATE_CONFIGURED:
# TODO(oponcea): Remove once sm supports in-service configuration
if (bk['backend'] != constants.SB_TYPE_CEPH or
bk['task'] != constants.SB_TASK_RECONFIG_CONTROLLER or
ihost['personality'] != constants.CONTROLLER):
msg = _("%(backend)s is %(notok)s. All storage backends must "
"be %(ok)s before operation "
"is allowed.") % {'backend': bk['backend'].title(),
'notok': bk['state'],
'ok': constants.SB_STATE_CONFIGURED}
raise wsme.exc.ClientSideError(msg)
@staticmethod
def _new_host_hardware_since_upgrade(host, upgrade_created_at):
"""
Determines the new hardware on the host since the upgrade started.
:param host host object
:param upgrade_created_at upgrade start timestamp
returns: new_hw tuple of new hardware on host
"""
new_hw = []
disks = pecan.request.dbapi.idisk_get_by_ihost(host.id)
new_disks = [x.uuid for x in disks
if x.created_at and (x.created_at > upgrade_created_at)]
if new_disks:
new_hw.append(('disks', host.hostname, new_disks))
interfaces = pecan.request.dbapi.iinterface_get_by_ihost(host.id)
new_interfaces = [
x.uuid for x in interfaces
if x.created_at and (x.created_at > upgrade_created_at)]
if new_interfaces:
new_hw.append(('interfaces', host.hostname, new_interfaces))
stors = pecan.request.dbapi.istor_get_by_ihost(host.id)
new_stors = [x.uuid for x in stors
if x.created_at and (x.created_at > upgrade_created_at)]
if new_stors:
new_hw.append(('stors', host.hostname, new_stors))
return new_hw
def _semantic_check_upgrade_refresh(self, upgrade, ihost, force):
"""
Determine whether upgrade should be aborted/refreshed due to
new hardware since upgrade start
"""
if force:
LOG.info("_semantic_check_upgrade_refresh check force")
return
if ihost['hostname'] != constants.CONTROLLER_1_HOSTNAME:
return
if upgrade.state not in [constants.UPGRADE_STARTED,
constants.UPGRADE_DATA_MIGRATION,
constants.UPGRADE_DATA_MIGRATION_COMPLETE,
constants.UPGRADE_UPGRADING_CONTROLLERS]:
LOG.info("_semantic_check_upgrade_refresh allow upgrade state=%s" %
upgrade.state)
return
upgrade_created_at = upgrade.created_at
# check for new host hardware since upgrade started
hosts = pecan.request.dbapi.ihost_get_list()
new_hw = []
for h in hosts:
if not h.personality:
continue
if h.created_at > upgrade_created_at:
new_hw.append(('host', h.hostname, h.uuid))
break
new_hw_h = self._new_host_hardware_since_upgrade(
h, upgrade_created_at)
if new_hw_h:
new_hw.extend(new_hw_h)
if new_hw:
msg = _("New hardware %s detected after upgrade started at %s. "
"Upgrade should be aborted."
% (new_hw, upgrade_created_at))
raise wsme.exc.ClientSideError(msg)
def _semantic_check_downgrade_refresh(self, upgrade, ihost, force):
"""
Determine whether downgrade should be aborted due to
new hardware since upgrade start
"""
if force:
LOG.info("_semantic_check_downgrade_refresh check force")
return
if upgrade.state not in [constants.UPGRADE_ABORTING,
constants.UPGRADE_ABORTING_ROLLBACK]:
LOG.info("_semantic_check_downgrade_refresh allow upgrade state=%s" %
upgrade.state)
return
upgrade_created_at = upgrade.created_at
# check for new host hardware since upgrade started
hosts = pecan.request.dbapi.ihost_get_list()
new_hw = []
for h in hosts:
if not h.personality:
continue
if h.created_at > upgrade_created_at:
new_hw.append(('host', h.hostname, h.uuid))
new_hw_h = self._new_host_hardware_since_upgrade(
h, upgrade_created_at)
if new_hw_h:
new_hw.extend(new_hw_h)
if new_hw:
new_host_hw = [(new_hw_type, name, info) for (new_hw_type, name, info) in new_hw
if name == ihost['hostname']]
if new_host_hw:
msg = _("New host %s detected after upgrade started at %s. "
"Host can not be downgraded."
% (ihost['hostname'], upgrade_created_at))
raise wsme.exc.ClientSideError(msg)
else:
# Acceptable to downgrade this host
msg = _("New host hardware %s detected after upgrade "
"started at %s. "
"Allow downgrade of %s during upgrade abort phase."
% (new_hw, upgrade_created_at, ihost['hostname']))
LOG.info(msg)
return
@staticmethod
def _semantic_check_nova_local_storage(ihost_uuid, personality):
"""
Perform semantic checking for nova local storage
:param ihost_uuid: uuid of host with worker functionality
:param personality: personality of host with worker functionality
"""
# query volume groups
nova_local_storage_lvg = None
ihost_ilvgs = pecan.request.dbapi.ilvg_get_by_ihost(ihost_uuid)
for lvg in ihost_ilvgs:
if lvg.lvm_vg_name == constants.LVG_NOVA_LOCAL:
nova_local_storage_lvg = lvg
break
# Prevent unlock if nova-local volume group has: invalid state
# (e.g., removing), invalid instance_backing, no physical
# volumes allocated.
if nova_local_storage_lvg:
if nova_local_storage_lvg.vg_state == constants.LVG_DEL:
raise wsme.exc.ClientSideError(
_("A host with worker functionality requires a "
"nova-local volume group prior to being enabled. It is "
"currently set to be removed on unlock. Please update "
"the storage settings for the host."))
else:
# Make sure that we have physical volumes allocated to the
# volume group
ihost_ipvs = pecan.request.dbapi.ipv_get_by_ihost(ihost_uuid)
lvg_has_pvs = False
for pv in ihost_ipvs:
if ((pv.lvm_vg_name == nova_local_storage_lvg.lvm_vg_name) and
(pv.pv_state != constants.PV_DEL)):
lvg_has_pvs = True
if not lvg_has_pvs:
raise wsme.exc.ClientSideError(
_("A host with worker functionality requires a "
"nova-local volume group prior to being enabled."
"The nova-local volume group does not contain any "
"physical volumes in the adding or provisioned "
"state."))
lvg_capabilities = nova_local_storage_lvg['capabilities']
instance_backing = lvg_capabilities.get(
constants.LVG_NOVA_PARAM_BACKING)
if instance_backing not in [
constants.LVG_NOVA_BACKING_IMAGE,
constants.LVG_NOVA_BACKING_REMOTE]:
raise wsme.exc.ClientSideError(
_("A host with worker functionality and a "
"nova-local volume group requires that a valid "
"instance backing is configured. "))
else:
# This method is only called with hosts that have a worker
# subfunction and is locked or if subfunction_config action is
# being called. Without a nova-local volume group, prevent
# unlocking.
if personality == constants.CONTROLLER:
host_description = 'controller with worker functionality'
else:
host_description = 'worker'
msg = _('A %s requires a nova-local volume group prior to being '
'enabled. Please update the storage settings for the '
'host.') % host_description
raise wsme.exc.ClientSideError('%s' % msg)
@staticmethod
def _semantic_check_restore_complete(ihost):
"""
During a restore procedure, checks worker nodes can be unlocked
only after running "config_controller --restore-complete"
"""
if os.path.isfile(tsc.RESTORE_SYSTEM_FLAG):
raise wsme.exc.ClientSideError(
_("Cannot unlock host %s. Please restore any volumes "
"and then complete the restore procedure by running "
"'config_controller --restore-complete' first. "
"Please refer to system admin guide for more details.") %
(ihost['hostname']))
def _semantic_check_worker_cpu_assignments(self, host):
"""
Perform semantic checks that cpu assignments are valid. Changes in
vswitch_type may alter vswitch core requirements.
"""
# If vswitch_type is none, enforce 0 vswitch cpu cores
vswitch_type = cutils.get_vswitch_type(pecan.request.dbapi)
host_cpus = pecan.request.dbapi.icpu_get_by_ihost(host['uuid'])
if vswitch_type == constants.VSWITCH_TYPE_NONE:
for cpu in host_cpus:
if cpu.allocated_function == constants.VSWITCH_FUNCTION:
raise wsme.exc.ClientSideError(
_('vSwitch cpus can only be used with a vswitch_type '
'specified.'))
elif vswitch_type == constants.VSWITCH_TYPE_OVS_DPDK:
has_vswitch_cpus = False
for cpu in host_cpus:
if cpu.allocated_function == constants.VSWITCH_FUNCTION:
has_vswitch_cpus = True
break
if not has_vswitch_cpus:
raise wsme.exc.ClientSideError(
_('vSwitch cpus must be enabled on this host for ovs-dpdk '
'to function.'))
@staticmethod
def _handle_ttys_dcd_change(ihost, ttys_dcd):
"""
Handle serial line carrier detection enable or disable request.
:param ihost: unpatched ihost dictionary
:param ttys_dcd: attribute supplied in patch
"""
LOG.info("%s _handle_ttys_dcd_change from %s to %s" %
(ihost['hostname'], ihost['ttys_dcd'], ttys_dcd))
# check if the flag is changed
if ttys_dcd is not None:
if ihost['ttys_dcd'] is None or ihost['ttys_dcd'] != ttys_dcd:
if ((ihost['administrative'] == constants.ADMIN_LOCKED and
ihost['availability'] == constants.AVAILABILITY_ONLINE) or
(ihost['administrative'] == constants.ADMIN_UNLOCKED and
ihost['operational'] == constants.OPERATIONAL_ENABLED)):
LOG.info("Notify conductor ttys_dcd change: (%s) (%s)" %
(ihost['uuid'], ttys_dcd))
pecan.request.rpcapi.configure_ttys_dcd(
pecan.request.context, ihost['uuid'], ttys_dcd)
def action_check(self, action, hostupdate):
"""Performs semantic checks related to action"""
if not action or (action.lower() == constants.NONE_ACTION):
rc = False
return rc
valid_actions = [constants.UNLOCK_ACTION,
constants.FORCE_UNLOCK_ACTION,
constants.LOCK_ACTION,
constants.FORCE_LOCK_ACTION,
constants.SWACT_ACTION,
constants.FORCE_SWACT_ACTION,
constants.RESET_ACTION,
constants.REBOOT_ACTION,
constants.REINSTALL_ACTION,
constants.POWERON_ACTION,
constants.POWEROFF_ACTION,
constants.VIM_SERVICES_ENABLED,
constants.VIM_SERVICES_DISABLED,
constants.VIM_SERVICES_DISABLE_FAILED,
constants.VIM_SERVICES_DISABLE_EXTEND,
constants.VIM_SERVICES_DELETE_FAILED,
constants.APPLY_PROFILE_ACTION,
constants.SUBFUNCTION_CONFIG_ACTION]
if action not in valid_actions:
raise wsme.exc.ClientSideError(
_("'%s' is not a supported maintenance action") % action)
force_unlock = False
if action == constants.FORCE_UNLOCK_ACTION:
# set force_unlock for semantic check and update action
# for compatability with vim and mtce
action = constants.UNLOCK_ACTION
force_unlock = True
hostupdate.action = action
rc = True
if action == constants.UNLOCK_ACTION:
# Set ihost_action in DB as early as possible as we need
# it as a synchronization point for things like lvg/pv
# deletion which is not allowed when ihost is unlokced
# or in the process of unlocking.
rc = self.update_ihost_action(action, hostupdate)
if rc:
pecan.request.dbapi.ihost_update(hostupdate.ihost_orig['uuid'],
hostupdate.ihost_val_prenotify)
try:
self.check_unlock(hostupdate, force_unlock)
except Exception:
LOG.info("host unlock check didn't pass, "
"so set the ihost_action back to None and re-raise the exception")
self.update_ihost_action(None, hostupdate)
pecan.request.dbapi.ihost_update(hostupdate.ihost_orig['uuid'],
hostupdate.ihost_val_prenotify)
raise
elif action == constants.LOCK_ACTION:
if self.check_lock(hostupdate):
rc = self.update_ihost_action(action, hostupdate)
elif action == constants.FORCE_LOCK_ACTION:
if self.check_force_lock(hostupdate):
rc = self.update_ihost_action(action, hostupdate)
elif action == constants.SWACT_ACTION:
self.check_swact(hostupdate)
elif action == constants.FORCE_SWACT_ACTION:
self.check_force_swact(hostupdate)
elif action == constants.REBOOT_ACTION:
self.check_reboot(hostupdate)
elif action == constants.RESET_ACTION:
self.check_reset(hostupdate)
elif action == constants.REINSTALL_ACTION:
self.check_reinstall(hostupdate)
elif action == constants.POWERON_ACTION:
self.check_poweron(hostupdate)
elif action == constants.POWEROFF_ACTION:
self.check_poweroff(hostupdate)
elif action == constants.VIM_SERVICES_ENABLED:
# hostupdate.notify_availability = constants.VIM_SERVICES_ENABLED
# self.update_ihost_action(action, hostupdate)
self.update_vim_progress_status(action, hostupdate)
elif action == constants.VIM_SERVICES_DISABLED:
# self.notify_availability = constants.VIM_SERVICES_DISABLED
self.update_vim_progress_status(action, hostupdate)
# rc = self.update_ihost_action(action, hostupdate)
elif action == constants.VIM_SERVICES_DISABLE_FAILED:
self.update_vim_progress_status(action, hostupdate)
elif action == constants.VIM_SERVICES_DISABLE_EXTEND:
self.update_vim_progress_status(action, hostupdate)
elif action == constants.VIM_SERVICES_DELETE_FAILED:
self.update_vim_progress_status(action, hostupdate)
elif action == constants.APPLY_PROFILE_ACTION:
self._check_apply_profile(hostupdate)
elif action == constants.SUBFUNCTION_CONFIG_ACTION:
self._check_subfunction_config(hostupdate)
self._semantic_check_nova_local_storage(
hostupdate.ihost_patch['uuid'],
hostupdate.ihost_patch['personality'])
else:
raise wsme.exc.ClientSideError(_(
"action_check unrecognized action: %s" % action))
if action in constants.MTCE_ACTIONS:
if self._semantic_mtc_check_action(hostupdate, action):
hostupdate.notify_mtce = True
task_val = hostupdate.get_task_from_action(action)
if task_val:
hostupdate.ihost_val_update({'task': task_val})
elif 'administrative' in hostupdate.delta:
# administrative state changed, update task, ihost_action in case
hostupdate.ihost_val_update({'task': "",
'ihost_action': ""})
LOG.info("%s action=%s ihost_val_prenotify: %s ihost_val: %s" %
(hostupdate.displayid,
hostupdate.action,
hostupdate.ihost_val_prenotify,
hostupdate.ihost_val))
if hostupdate.ihost_val_prenotify:
LOG.info("%s host.update.ihost_val_prenotify %s" %
(hostupdate.displayid, hostupdate.ihost_val_prenotify))
if self.check_notify_vim(action):
hostupdate.notify_vim = True
if self.check_notify_mtce(action, hostupdate) > 0:
hostupdate.notify_mtce = True
LOG.info("%s action_check action=%s, notify_vim=%s "
"notify_mtce=%s rc=%s" %
(hostupdate.displayid,
action,
hostupdate.notify_vim,
hostupdate.notify_mtce,
rc))
return rc
@staticmethod
def check_notify_vim(action):
if action in constants.VIM_ACTIONS:
return True
else:
return False
@staticmethod
def _check_apply_profile(hostupdate):
ihost = hostupdate.ihost_orig
if (ihost['administrative'] == constants.ADMIN_UNLOCKED and
not utils.is_host_simplex_controller(ihost)):
raise wsme.exc.ClientSideError(
_("Can not apply profile to an 'unlocked' host %s; "
"Please 'Lock' first." % hostupdate.displayid))
if utils.get_system_mode() == constants.SYSTEM_MODE_SIMPLEX:
raise wsme.exc.ClientSideError(_(
"Applying a profile on a simplex system is not allowed."))
return True
@staticmethod
def check_notify_mtce(action, hostupdate):
"""Determine whether mtce should be notified of this patch request
returns: Integer (nonmtc_change_count)
"""
nonmtc_change_count = 0
if action in constants.VIM_ACTIONS:
return nonmtc_change_count
elif action in constants.CONFIG_ACTIONS:
return nonmtc_change_count
elif action in constants.VIM_SERVICES_ENABLED:
return nonmtc_change_count
mtc_ignore_list = ['administrative', 'availability', 'operational',
'task', 'config_status', 'uptime', 'capabilities',
'ihost_action',
'subfunction_oper', 'subfunction_avail',
'vim_progress_status'
'location', 'serialid', 'invprovision']
if pecan.request.user_agent.startswith('mtce'):
mtc_ignore_list.append('bm_ip')
nonmtc_change_count = len(set(hostupdate.delta) - set(mtc_ignore_list))
return nonmtc_change_count
@staticmethod
def stage_administrative_update(hostupdate):
# Always configure when the host is unlocked - this will set the
# hostname and allow the node to boot and configure itself.
# NOTE: This is being hit the second time through this function on
# the unlock. The first time through, the "action" is set to unlock
# on the patched_iHost, but the "administrative" is still locked.
# Once maintenance processes the unlock, they do another patch and
# set the "administrative" to unlocked.
if 'administrative' in hostupdate.delta and \
hostupdate.ihost_patch['administrative'] == \
constants.ADMIN_UNLOCKED:
if hostupdate.ihost_orig['invprovision'] == \
constants.UNPROVISIONED or \
hostupdate.ihost_orig['invprovision'] is None:
LOG.info("stage_administrative_update: provisioning")
hostupdate.ihost_val_update({'invprovision':
constants.PROVISIONING})
if 'operational' in hostupdate.delta and \
hostupdate.ihost_patch['operational'] == \
constants.OPERATIONAL_ENABLED:
if hostupdate.ihost_orig['invprovision'] == constants.PROVISIONING:
# first time unlocked successfully
LOG.info("stage_administrative_update: provisioned")
hostupdate.ihost_val_update(
{'invprovision': constants.PROVISIONED}
)
@staticmethod
def _check_provisioned_storage_hosts():
# Get provisioned storage hosts
ihosts = pecan.request.dbapi.ihost_get_by_personality(
constants.STORAGE
)
host_names = []
for ihost in ihosts:
if ihost.invprovision == constants.PROVISIONED:
host_names.append(ihost.hostname)
LOG.info("Provisioned storage node(s) %s" % host_names)
# Get replication
replication, __ = StorageBackendConfig.get_ceph_max_replication(pecan.request.dbapi)
expected_hosts = \
constants.CEPH_REPLICATION_GROUP0_HOSTS[int(replication)]
current_exp_hosts = set(expected_hosts) & set(host_names)
# Check expected versus provisioned
if len(current_exp_hosts) == replication:
return True
else:
return False
def _update_add_ceph_state(self):
api = pecan.request.dbapi
backend = StorageBackendConfig.get_configuring_backend(api)
if backend and backend.backend == constants.SB_TYPE_CEPH:
ihosts = api.ihost_get_by_personality(
constants.CONTROLLER
)
for ihost in ihosts:
if ihost.config_status == constants.CONFIG_STATUS_OUT_OF_DATE:
return
# check if customer needs to install storage nodes
if backend.task == constants.SB_TASK_RECONFIG_CONTROLLER:
stor_model = ceph.get_ceph_storage_model()
if (HostController._check_provisioned_storage_hosts() or
stor_model == constants.CEPH_CONTROLLER_MODEL):
# This means that either:
# 1. Storage nodes are already provisioned (this is not
# the first time Ceph is configured) or
# 2. We are on a standard config and we don't need to
# configure storage nodes at all.
api.storage_backend_update(backend.uuid, {
'state': constants.SB_STATE_CONFIGURED,
'task': None
})
else:
# Storage nodes are not yet provisioned
api.storage_backend_update(backend.uuid, {
'state': constants.SB_STATE_CONFIGURED,
'task': constants.SB_TASK_PROVISION_STORAGE
})
return
backend = StorageBackendConfig.get_configured_backend(
api,
constants.CINDER_BACKEND_CEPH
)
if not backend:
return
if backend.task == constants.SB_TASK_PROVISION_STORAGE:
if HostController._check_provisioned_storage_hosts():
api.storage_backend_update(backend.uuid, {
'task': constants.SB_TASK_RECONFIG_WORKER
})
# update manifest for all online/enabled worker nodes
# live apply new ceph manifest for all worker nodes that
# are online/enabled. The rest will pickup when unlock
LOG.info(
'Apply new Ceph manifest to provisioned worker nodes.'
)
pecan.request.rpcapi.update_ceph_base_config(
pecan.request.context, personalities=[constants.WORKER]
)
# mark all tasks completed after updating the manifests for
# all worker nodes.
api.storage_backend_update(backend.uuid, {'task': None})
elif backend.task == constants.SB_TASK_RESIZE_CEPH_MON_LV:
ihosts = pecan.request.dbapi.ihost_get_list()
personalities = [constants.CONTROLLER, constants.STORAGE]
for ihost in ihosts:
if ihost.config_status == constants.CONFIG_STATUS_OUT_OF_DATE \
and ihost.personality in personalities:
break
else:
# all storage controller nodes are up to date
api.storage_backend_update(backend.uuid, {'task': None})
# workflow of installing object gateway is completed
elif backend.task == constants.SB_TASK_ADD_OBJECT_GATEWAY:
ihosts = api.ihost_get_by_personality(
constants.CONTROLLER
)
for ihost in ihosts:
if ihost.config_status == constants.CONFIG_STATUS_OUT_OF_DATE:
return
api.storage_backend_update(backend.uuid, {
'state': constants.SB_STATE_CONFIGURED,
'task': None
})
elif backend.task == constants.SB_TASK_RESTORE:
ihosts = api.ihost_get_by_personality(
constants.STORAGE
)
if ihosts:
# For storage setup, no change is required.
LOG.info("This is a storage setup. No change.")
storage_enabled = 0
for ihost in ihosts:
if ihost.operational == constants.OPERATIONAL_ENABLED:
storage_enabled = storage_enabled + 1
if storage_enabled and storage_enabled == len(ihosts):
LOG.info("All storage hosts are %s. Restore crushmap..." %
constants.OPERATIONAL_ENABLED)
try:
if not pecan.request.rpcapi.restore_ceph_config(
pecan.request.context, after_storage_enabled=True):
raise Exception("restore_ceph_config returned false")
except Exception as e:
raise wsme.exc.ClientSideError(
_("Restore Ceph config failed: %s" % e))
elif utils.is_aio_system(pecan.request.dbapi):
# For AIO, Ceph restore is done in ceph puppet
if not os.path.isfile(tsc.RESTORE_IN_PROGRESS_FLAG):
api.storage_backend_update(backend.uuid, {'task': None})
else:
# This is ceph restore for 2+2.
# If config_controller restore is still in progress, we wait.
if os.path.isfile(tsc.RESTORE_IN_PROGRESS_FLAG):
LOG.info("Restore flag is still on. Do nothing now. ")
return
active_mons, required_mons, __ = \
self._ceph.get_monitors_status(pecan.request.dbapi)
if required_mons > active_mons:
LOG.info("Not enough monitors yet available to fix crushmap.")
else:
LOG.info("Restore Ceph config ...")
# First restore ceph config
try:
if not pecan.request.rpcapi.restore_ceph_config(
pecan.request.context):
raise Exception("restore_ceph_config returned false")
except Exception as e:
raise wsme.exc.ClientSideError(
_("Restore Ceph config failed: %s" % e))
# Set Ceph backend task to None
api.storage_backend_update(backend.uuid, {'task': None})
# Apply runtime manifests for OSDs on two controller nodes.
c_hosts = api.ihost_get_by_personality(
constants.CONTROLLER
)
runtime_manifests = True
for c_host in c_hosts:
istors = pecan.request.dbapi.istor_get_by_ihost(c_host.uuid)
for stor in istors:
pecan.request.rpcapi.update_ceph_osd_config(
pecan.request.context,
c_host,
stor.uuid,
runtime_manifests)
@staticmethod
def update_ihost_action(action, hostupdate):
if action is None:
preval = {'ihost_action': ''}
elif action == constants.FORCE_LOCK_ACTION:
preval = {'ihost_action': constants.FORCE_LOCK_ACTION}
elif action == constants.LOCK_ACTION:
preval = {'ihost_action': constants.LOCK_ACTION}
elif (action == constants.UNLOCK_ACTION or
action == constants.FORCE_UNLOCK_ACTION):
preval = {'ihost_action': constants.UNLOCK_ACTION}
else:
LOG.error("update_ihost_action unsupported action: %s" % action)
return False
hostupdate.ihost_val_prenotify.update(preval)
hostupdate.ihost_val.update(preval)
task_val = hostupdate.get_task_from_action(action)
if task_val:
hostupdate.ihost_val_update({'task': task_val})
return True
@staticmethod
def update_vim_progress_status(action, hostupdate):
LOG.info("%s Pending update_vim_progress_status %s" %
(hostupdate.displayid, action))
return True
def check_provisioning(self, hostupdate, patch):
# Once the host has been provisioned lock down additional fields
ihost = hostupdate.ihost_patch
delta = hostupdate.delta
provision_state = [constants.PROVISIONED, constants.PROVISIONING]
if hostupdate.ihost_orig['invprovision'] in provision_state:
state_rel_path = ['hostname', 'personality', 'subfunctions']
if any(p in state_rel_path for p in delta):
raise wsme.exc.ClientSideError(
_("The following fields can not be modified because "
"this host %s has been configured: "
"hostname, personality, subfunctions" %
hostupdate.ihost_orig['hostname']))
# Check whether any configurable installation parameters are updated
install_parms = ['boot_device', 'rootfs_device', 'install_output', 'console', 'tboot']
if any(p in install_parms for p in delta):
# Disallow changes if the node is not locked
if ihost['administrative'] != constants.ADMIN_LOCKED:
raise wsme.exc.ClientSideError(
_("Host must be locked before updating "
"installation parameters."))
# An update to PXE boot information is required
hostupdate.configure_required = True
# Check whether vsc_controllers semantic checks are needed
if 'vsc_controllers' in hostupdate.delta:
self._semantic_check_vsc_controllers(
hostupdate.ihost_orig,
hostupdate.ihost_patch['vsc_controllers'])
if 'personality' in delta:
LOG.info("iHost['personality']=%s" %
hostupdate.ihost_orig['personality'])
if hostupdate.ihost_orig['personality']:
raise wsme.exc.ClientSideError(
_("Can not change personality after it has been set. "
"Host %s must be deleted and re-added in order to change "
"the personality." % hostupdate.ihost_orig['hostname']))
if (hostupdate.ihost_patch['personality'] in
(constants.CONTROLLER, constants.STORAGE)):
self._controller_storage_node_setup(hostupdate.ihost_patch,
hostupdate)
# check the subfunctions are updated properly
LOG.info("hostupdate.ihost_patch.subfunctions %s" %
hostupdate.ihost_patch['subfunctions'])
elif hostupdate.ihost_patch['personality'] == constants.WORKER:
self._check_worker(hostupdate.ihost_patch, hostupdate)
else:
LOG.error("Unexpected personality: %s" %
hostupdate.ihost_patch['personality'])
hostname = (hostupdate.ihost_val.get('hostname') or
hostupdate.ihost_patch['hostname'])
# Check host personality provisioning order during upgrades
self._check_upgrade_provision_order(
hostupdate.ihost_patch['personality'],
hostname)
# Always configure when the personality has been set - this will
# set up the PXE boot information so the software can be installed
hostupdate.configure_required = True
# Notify VIM when the personality is set.
hostupdate.notify_vim_add_host = True
if constants.SUBFUNCTIONS in delta:
if hostupdate.ihost_orig[constants.SUBFUNCTIONS]:
raise wsme.exc.ClientSideError(
_("Can not change subfunctions after it has been set. "
"Host %s must be deleted and re-added in order to change "
"the subfunctions." % hostupdate.ihost_orig['hostname']))
if hostupdate.ihost_patch['personality'] == constants.WORKER:
valid_subfunctions = (constants.WORKER,
constants.LOWLATENCY)
elif hostupdate.ihost_patch['personality'] == constants.CONTROLLER:
valid_subfunctions = (constants.CONTROLLER,
constants.WORKER,
constants.LOWLATENCY)
elif hostupdate.ihost_patch['personality'] == constants.STORAGE:
# Comparison is expecting a list
valid_subfunctions = (constants.STORAGE, constants.STORAGE)
subfunctions_set = \
set(hostupdate.ihost_patch[constants.SUBFUNCTIONS].split(','))
if not subfunctions_set.issubset(valid_subfunctions):
raise wsme.exc.ClientSideError(
("%s subfunctions %s contains unsupported values. Allowable: %s." %
(hostupdate.displayid, subfunctions_set, valid_subfunctions)))
if hostupdate.ihost_patch['personality'] == constants.WORKER:
if constants.WORKER not in subfunctions_set:
# Automatically add it
subfunctions_list = list(subfunctions_set)
subfunctions_list.insert(0, constants.WORKER)
subfunctions = ','.join(subfunctions_list)
LOG.info("%s update subfunctions=%s" %
(hostupdate.displayid, subfunctions))
hostupdate.ihost_val_prenotify.update({'subfunctions': subfunctions})
hostupdate.ihost_val.update({'subfunctions': subfunctions})
# The hostname for a controller or storage node cannot be modified
# Disallow hostname changes
if 'hostname' in delta:
if hostupdate.ihost_orig['hostname']:
if (hostupdate.ihost_patch['hostname'] !=
hostupdate.ihost_orig['hostname']):
raise wsme.exc.ClientSideError(
_("The hostname field can not be modified because "
"the hostname %s has already been configured. "
"If changing hostname is required, please delete "
"this host, then readd." %
hostupdate.ihost_orig['hostname']))
# TODO: evaluate for efficiency
for attribute in patch:
# check for duplicate attributes
for attribute2 in patch:
if attribute['path'] == attribute2['path']:
if attribute['value'] != attribute2['value']:
err_dp = 'Illegal duplicate parameters passed.'
raise wsme.exc.ClientSideError(_(err_dp))
if 'personality' in delta or 'hostname' in delta:
personality = hostupdate.ihost_patch.get('personality') or ""
hostname = hostupdate.ihost_patch.get('hostname') or ""
if personality and hostname:
self._validate_hostname(hostname, personality)
if 'personality' in delta:
HostController._personality_license_check(
hostupdate.ihost_patch['personality'])
@staticmethod
def _personality_license_check(personality):
if personality == constants.CONTROLLER:
return
if not personality:
return
if personality == constants.WORKER and utils.is_aio_duplex_system():
if utils.get_worker_count() >= constants.AIO_DUPLEX_MAX_WORKERS:
msg = _("All-in-one Duplex is restricted to "
"%s workers.") % constants.AIO_DUPLEX_MAX_WORKERS
raise wsme.exc.ClientSideError(msg)
else:
return
if (utils.SystemHelper.get_product_build() ==
constants.TIS_AIO_BUILD):
msg = _("Personality [%s] for host is not compatible "
"with installed software. ") % personality
raise wsme.exc.ClientSideError(msg)
@staticmethod
def check_reset(hostupdate):
"""Check semantics on host-reset."""
if utils.get_system_mode() == constants.SYSTEM_MODE_SIMPLEX:
raise wsme.exc.ClientSideError(
_("Can not 'Reset' a simplex system"))
if hostupdate.ihost_orig['administrative'] == constants.ADMIN_UNLOCKED:
raise wsme.exc.ClientSideError(
_("Can not 'Reset' an 'unlocked' host %s; "
"Please 'Lock' first" % hostupdate.displayid))
return True
@staticmethod
def check_poweron(hostupdate):
# Semantic Check: State Dependency: Power-On case
if (hostupdate.ihost_orig['administrative'] ==
constants.ADMIN_UNLOCKED):
raise wsme.exc.ClientSideError(
_("Can not 'Power-On' an already Powered-on "
"and 'unlocked' host %s" % hostupdate.displayid))
if utils.get_system_mode() == constants.SYSTEM_MODE_SIMPLEX:
raise wsme.exc.ClientSideError(
_("Can not 'Power-On' an already Powered-on "
"simplex system"))
@staticmethod
def check_poweroff(hostupdate):
# Semantic Check: State Dependency: Power-Off case
if utils.get_system_mode() == constants.SYSTEM_MODE_SIMPLEX:
raise wsme.exc.ClientSideError(
_("Can not 'Power-Off' a simplex system via "
"system commands"))
if (hostupdate.ihost_orig['administrative'] ==
constants.ADMIN_UNLOCKED):
raise wsme.exc.ClientSideError(
_("Can not 'Power-Off' an 'unlocked' host %s; "
"Please 'Lock' first" % hostupdate.displayid))
@staticmethod
def check_reinstall(hostupdate):
""" Semantic Check: State Dependency: Reinstall case"""
if utils.get_system_mode() == constants.SYSTEM_MODE_SIMPLEX:
raise wsme.exc.ClientSideError(_(
"Reinstalling a simplex system is not allowed."))
ihost = hostupdate.ihost_orig
if ihost['administrative'] == constants.ADMIN_UNLOCKED:
raise wsme.exc.ClientSideError(
_("Can not 'Reinstall' an 'unlocked' host %s; "
"Please 'Lock' first" % hostupdate.displayid))
elif ((ihost['administrative'] == constants.ADMIN_LOCKED) and
(ihost['availability'] != "online")):
raise wsme.exc.ClientSideError(
_("Can not 'Reinstall' %s while it is 'offline'. "
"Please wait for this host's availability state "
"to be 'online' and then re-issue the reinstall "
"command." % hostupdate.displayid))
hostupdate.configure_required = True
def check_unlock(self, hostupdate, force_unlock=False):
"""Check semantics on host-unlock."""
if (hostupdate.action != constants.UNLOCK_ACTION and
hostupdate.action != constants.FORCE_UNLOCK_ACTION):
LOG.error("check_unlock unexpected action: %s" % hostupdate.action)
return False
# Semantic Check: Don't unlock if installation failed
if (hostupdate.ihost_orig['install_state'] ==
constants.INSTALL_STATE_FAILED):
raise wsme.exc.ClientSideError(
_("Cannot unlock host %s due to installation failure" %
hostupdate.displayid))
# Semantic Check: Avoid Unlock of Unlocked Host
if hostupdate.ihost_orig['administrative'] == constants.ADMIN_UNLOCKED:
raise wsme.exc.ClientSideError(
_("Avoiding 'unlock' action on already "
"'unlocked' host %s" % hostupdate.ihost_orig['hostname']))
# Semantic Check: Action Dependency: Power-Off / Unlock case
if (hostupdate.ihost_orig['availability'] ==
constants.POWEROFF_ACTION):
raise wsme.exc.ClientSideError(
_("Can not 'Unlock a Powered-Off' host %s; Power-on, "
"wait for 'online' status and then 'unlock'" %
hostupdate.displayid))
# Semantic Check: Action Dependency: Online / Unlock case
if (not force_unlock and hostupdate.ihost_orig['availability'] !=
constants.AVAILABILITY_ONLINE):
raise wsme.exc.ClientSideError(
_("Host %s is not online. "
"Wait for 'online' availability status and then 'unlock'" %
hostupdate.displayid))
# Semantic Check: Don't unlock when running incorrect software load
host_upgrade = objects.host_upgrade.get_by_host_id(
pecan.request.context, hostupdate.ihost_orig['id'])
if host_upgrade.software_load != host_upgrade.target_load and \
hostupdate.ihost_orig['hostname'] != \
constants.CONTROLLER_1_HOSTNAME:
raise wsme.exc.ClientSideError(
_("Can not Unlock a host running the incorrect "
"software load. Reinstall the host to correct."))
# To unlock, we need the following additional fields
if not (hostupdate.ihost_patch['mgmt_mac'] and
hostupdate.ihost_patch['mgmt_ip'] and
hostupdate.ihost_patch['hostname'] and
hostupdate.ihost_patch['personality'] and
hostupdate.ihost_patch['subfunctions']):
raise wsme.exc.ClientSideError(
_("Can not unlock an unprovisioned host %s. "
"Please perform 'Edit Host' to provision host."
% hostupdate.displayid))
# To unlock, ensure reinstall has completed
action_state = hostupdate.ihost_orig[constants.HOST_ACTION_STATE]
if (action_state and
action_state == constants.HAS_REINSTALLING):
if not force_unlock:
raise wsme.exc.ClientSideError(
_("Can not unlock host %s undergoing reinstall. "
"Please ensure host has completed reinstall prior to unlock."
% hostupdate.displayid))
else:
LOG.warn("Allowing force-unlock of host %s "
"undergoing reinstall." % hostupdate.displayid)
personality = hostupdate.ihost_patch.get('personality')
if personality == constants.CONTROLLER:
self.check_unlock_controller(hostupdate, force_unlock)
if cutils.host_has_function(hostupdate.ihost_patch, constants.WORKER):
self.check_unlock_worker(hostupdate, force_unlock)
elif personality == constants.STORAGE:
self.check_unlock_storage(hostupdate)
self.check_unlock_interfaces(hostupdate)
self.unlock_update_mgmt_interface(hostupdate.ihost_patch)
self.check_unlock_partitions(hostupdate)
self.check_unlock_patching(hostupdate, force_unlock)
hostupdate.configure_required = True
if (os.path.isfile(constants.ANSIBLE_BOOTSTRAP_FLAG) and
hostupdate.ihost_patch['hostname'] == 'controller-0'):
# For the first unlock of the initial controller bootstrapped by
# Ansible, don't notify vim.
hostupdate.notify_vim = False
else:
hostupdate.notify_vim = True
return True
def check_unlock_patching(self, hostupdate, force_unlock):
"""Check whether the host is patch current.
"""
if force_unlock:
return
phosts = []
try:
# Token is optional for admin url
# if (self._api_token is None or self._api_token.is_expired()):
# self._api_token = rest_api.get_token()
system = pecan.request.dbapi.isystem_get_one()
response = patch_api.patch_query_hosts(
token=None,
timeout=constants.PATCH_DEFAULT_TIMEOUT_IN_SECS,
region_name=system.region_name)
phosts = response['data']
except Exception as e:
LOG.warn(_("No response from patch api %s e=%s" %
(hostupdate.displayid, e)))
self._api_token = None
return
for phost in phosts:
if phost.get('hostname') == hostupdate.ihost_patch.get('hostname'):
if not phost.get('patch_current'):
raise wsme.exc.ClientSideError(
_("host-unlock rejected: Not patch current. "
"'sw-patch host-install %s' is required." %
hostupdate.displayid))
def check_lock(self, hostupdate):
"""Check semantics on host-lock."""
LOG.info("%s ihost check_lock" % hostupdate.displayid)
if hostupdate.action != constants.LOCK_ACTION:
LOG.error("%s check_lock unexpected action: %s" %
(hostupdate.displayid, hostupdate.action))
return False
# Semantic Check: Avoid Lock of Locked Host
if hostupdate.ihost_orig['administrative'] == constants.ADMIN_LOCKED:
# TOCHECK: previously resetting vals
raise wsme.exc.ClientSideError(
_("Avoiding %s action on already "
"'locked' host %s" % (hostupdate.ihost_patch['action'],
hostupdate.ihost_orig['hostname'])))
# personality specific lock checks
personality = hostupdate.ihost_patch.get('personality')
if personality == constants.CONTROLLER:
self.check_lock_controller(hostupdate)
elif personality == constants.STORAGE:
self.check_lock_storage(hostupdate)
subfunctions_set = \
set(hostupdate.ihost_patch[constants.SUBFUNCTIONS].split(','))
if (personality == constants.WORKER or
constants.WORKER in subfunctions_set):
self.check_lock_worker(hostupdate)
hostupdate.notify_vim = True
hostupdate.notify_mtce = True
return True
def check_force_lock(self, hostupdate):
# personality specific lock checks
personality = hostupdate.ihost_patch.get('personality')
if personality == constants.CONTROLLER:
self.check_lock_controller(hostupdate, force=True)
elif personality == constants.STORAGE:
self.check_lock_storage(hostupdate, force=True)
return True
def check_lock_controller(self, hostupdate, force=False):
"""Pre lock semantic checks for controller"""
LOG.info("%s ihost check_lock_controller" % hostupdate.displayid)
if utils.get_system_mode() != constants.SYSTEM_MODE_SIMPLEX:
active = utils.is_host_active_controller(hostupdate.ihost_orig)
if active:
raise wsme.exc.ClientSideError(
_("%s : Rejected: Can not lock an active "
"controller.") % hostupdate.ihost_orig['hostname'])
# Reject lock while Ceph OSD storage devices are configuring
if not force:
stors = pecan.request.dbapi.istor_get_by_ihost(
hostupdate.ihost_orig['uuid']
)
for stor in stors:
if stor.state == constants.SB_STATE_CONFIGURING:
raise wsme.exc.ClientSideError(
_("%s : Rejected: Can not lock a controller "
"with storage devices in '%s' state.") %
(hostupdate.ihost_orig['hostname'],
constants.SB_STATE_CONFIGURING))
if StorageBackendConfig.has_backend_configured(
pecan.request.dbapi,
constants.SB_TYPE_CEPH):
query_hosts = None
stor_model = ceph.get_ceph_storage_model()
if stor_model == constants.CEPH_STORAGE_MODEL:
query_hosts = constants.STORAGE
elif stor_model == constants.CEPH_CONTROLLER_MODEL:
query_hosts = constants.CONTROLLER
else:
# If backend type is still undefined it means no storage nodes
# have been configured and no worker monitor has been added,
# so it is safe to not check the quorum.
# Or we are dealing with an AIO-SX.
return
try:
st_nodes = pecan.request.dbapi.ihost_get_by_personality(query_hosts)
except exception.NodeNotFound:
# If we don't have any storage nodes we don't need to
# check for quorum. We'll allow the node to be locked.
# We will always have at least one controller, so for
# controllers that also act as storage nodes this should
# never happen.
return
# TODO(oponcea) remove once SM supports in-service config reload
# Allow locking controllers when all storage nodes are locked.
if stor_model == constants.CEPH_STORAGE_MODEL:
for node in st_nodes:
if (node['administrative'] == constants.ADMIN_UNLOCKED):
break
else:
return
if (hostupdate.ihost_orig['administrative'] ==
constants.ADMIN_UNLOCKED and
hostupdate.ihost_orig['operational'] ==
constants.OPERATIONAL_ENABLED):
# If the node is unlocked and enabled we need to check that we
# have enough storage monitors.
# If we are in an upgrade and aborting/rolling back the upgrade
# we need to skip the storage monitor check for controller-1.
# Before we downgrade controller-0 we shutdown the storage
# nodes and disable the storage monitor on controller-1.
# After controller-0 is downgraded and we go to lock
# controller-1, there will only be one storage monitor running
# (on controller-0) and the ceph api will fail/timeout.
check_storage_monitors = True
try:
upgrade = pecan.request.dbapi.software_upgrade_get_one()
except exception.NotFound:
pass
else:
if upgrade.state == constants.UPGRADE_ABORTING_ROLLBACK \
and hostupdate.ihost_orig['hostname'] == \
constants.CONTROLLER_1_HOSTNAME:
check_storage_monitors = False
if check_storage_monitors:
num_monitors, required_monitors, quorum_names = \
self._ceph.get_monitors_status(pecan.request.dbapi)
if (hostupdate.ihost_orig['hostname'] in quorum_names and
num_monitors - 1 < required_monitors):
raise wsme.exc.ClientSideError(_(
"Only %d storage "
"monitor available. At least %s unlocked and "
"enabled hosts with monitors are required. Please"
" ensure hosts with monitors are unlocked and "
"enabled.") %
(num_monitors, constants.MIN_STOR_MONITORS))
if not force:
# sm-lock-pre-check
node_name = hostupdate.displayid
response = sm_api.lock_pre_check(node_name, timeout=30)
if response:
error_code = response.get('error_code')
if ERR_CODE_LOCK_SOLE_SERVICE_PROVIDER == error_code:
impact_svc_list = response.get('impact_service_list')
svc_list = ','.join(impact_svc_list)
if len(impact_svc_list) > 1:
msg = _("Services {svc_list} are only running on "
"{host}, locking {host} will result "
"service outage. If lock {host} is required, "
"please use \"force lock\" command.").format(
svc_list=svc_list, host=node_name)
else:
msg = _("Service {svc_list} is only running on "
"{host}, locking {host} will result "
"service outage. If lock {host} is required, "
"please use \"force lock\" command.").format(
svc_list=svc_list, host=node_name)
raise wsme.exc.ClientSideError(msg)
elif "0" != error_code:
raise wsme.exc.ClientSideError(
_("%s" % response['error_details']))
def check_unlock_controller(self, hostupdate, force_unlock=False):
"""Pre unlock semantic checks for controller"""
LOG.info("%s ihost check_unlock_controller" % hostupdate.displayid)
self._semantic_check_unlock_upgrade(hostupdate.ihost_orig, force_unlock)
self._semantic_check_oam_interface(hostupdate.ihost_orig)
self._semantic_check_cinder_volumes(hostupdate.ihost_orig)
self._semantic_check_filesystem_sizes(hostupdate.ihost_orig)
self._semantic_check_storage_backend(hostupdate.ihost_orig)
# If HTTPS is enabled then we may be in TPM configuration mode
if utils.get_https_enabled():
self._semantic_check_tpm_config(hostupdate.ihost_orig)
def check_unlock_worker(self, hostupdate, force_unlock=False):
"""Check semantics on host-unlock of a worker."""
LOG.info("%s ihost check_unlock_worker" % hostupdate.displayid)
ihost = hostupdate.ihost_orig
if ihost['invprovision'] is None:
raise wsme.exc.ClientSideError(
_("Can not unlock an unconfigured host %s. Please "
"configure host and wait for Availability State "
"'online' prior to unlock." % hostupdate.displayid))
# Check whether a restore was properly completed
self._semantic_check_restore_complete(ihost)
# Disable certain worker unlock checks in a kubernetes config
kubernetes_config = utils.is_kubernetes_config()
if kubernetes_config:
self._semantic_check_data_interfaces(ihost,
kubernetes_config,
force_unlock)
self._semantic_check_vswitch_type_attributes(ihost)
else:
# sdn configuration check
self._semantic_check_sdn_attributes(ihost)
# check whether data route gateways are reachable
self._semantic_check_data_routes(ihost)
# check whether data interfaces have been configured
self._semantic_check_data_interfaces(ihost,
kubernetes_config,
force_unlock)
self._semantic_check_data_addresses(ihost)
self._semantic_check_data_vrs_attributes(ihost)
# Check if cpu assignments are valid
self._semantic_check_worker_cpu_assignments(ihost)
# check if the platform reserved memory is valid
ihost_inodes = pecan.request.dbapi.inode_get_by_ihost(ihost['uuid'])
mib_reserved = 0
mib_reserved_disk_io = 0
align_2M_memory = False
align_1G_memory = False
vswitch_hp_size = None
for node in ihost_inodes:
# If the reserved memory has changed (eg, due to patch that
# changes common/constants.py), then push updated reserved memory
# to database, and automatically adjust 2M and 1G hugepages based
# on the delta. Patch removal will not result in the auto
# incremented value to be brought back down as there is no record
# of the original setting.
self._auto_adjust_memory_for_node(ihost, node)
mems = pecan.request.dbapi.imemory_get_by_inode(node['id'])
for m in mems:
if not vswitch_hp_size:
vswitch_hp_size = m.vswitch_hugepages_size_mib
else:
if m.vswitch_hugepages_size_mib != vswitch_hp_size:
raise wsme.exc.ClientSideError(_(
"Mismatched vswitch socket memory hugepage size."))
# check whether the pending hugepages changes and the current
# platform reserved memory fit within the total memory available
mib_reserved_node, pending_2M_memory, pending_1G_memory = \
self._semantic_check_memory_for_node(ihost, node)
mib_reserved += mib_reserved_node
if pending_2M_memory:
align_2M_memory = True
LOG.info("pending 2M memory node=%s mib_reserved=%s" %
(node.uuid, mib_reserved))
if pending_1G_memory:
align_1G_memory = True
LOG.info("pending 1G memory node=%s mib_reserved=%s" %
(node.uuid, mib_reserved))
mib_reserved_disk_io += constants.DISK_IO_RESIDENT_SET_SIZE_MIB
if align_2M_memory or align_1G_memory:
self._align_pending_memory(ihost, align_2M_memory, align_1G_memory)
# update ihost huge pages allocation
self._update_huge_pages(ihost)
if cutils.is_virtual() or cutils.is_virtual_worker(ihost):
mib_platform_reserved_no_io = mib_reserved
required_platform = \
constants.PLATFORM_CORE_MEMORY_RESERVED_MIB_VBOX
if cutils.host_has_function(ihost, constants.CONTROLLER):
required_platform += \
constants.COMBINED_NODE_CONTROLLER_MEMORY_RESERVED_MIB_VBOX
else:
# If not a controller, add overhead for metadata and vrouters
required_platform += \
constants.NETWORK_METADATA_OVERHEAD_MIB_VBOX
else:
mib_platform_reserved_no_io = mib_reserved - mib_reserved_disk_io
required_platform = constants.PLATFORM_CORE_MEMORY_RESERVED_MIB
if cutils.host_has_function(ihost, constants.CONTROLLER):
low_core = cutils.is_low_core_system(ihost, pecan.request.dbapi)
if low_core:
required_platform += \
constants.COMBINED_NODE_CONTROLLER_MEMORY_RESERVED_MIB_XEOND
else:
required_platform += \
constants.COMBINED_NODE_CONTROLLER_MEMORY_RESERVED_MIB
else:
# If not a controller, add overhead for metadata and vrouters
required_platform += constants.NETWORK_METADATA_OVERHEAD_MIB
LOG.debug("mib_platform_reserved_no_io %s required_platform %s"
% (mib_platform_reserved_no_io, required_platform))
if mib_platform_reserved_no_io < required_platform:
msg = (_("Insufficient memory reserved for platform on %(host)s. "
"Platform memory must be at least %(required)s MiB "
"summed across all numa nodes."
) %
{'host': ihost['hostname'], 'required': required_platform})
raise wsme.exc.ClientSideError(msg)
shared_services = utils.get_shared_services()
if (shared_services is not None and
constants.SERVICE_TYPE_VOLUME in shared_services):
# do not check storage nodes in secondary region as "volume" is
# shared service provided by the primary region.
pass
elif StorageBackendConfig.has_backend_configured(
pecan.request.dbapi,
constants.SB_TYPE_CEPH):
if utils.is_aio_simplex_system(pecan.request.dbapi):
# Check if host has enough OSDs configured for each tier
tiers = pecan.request.dbapi.storage_tier_get_all()
ceph_tiers = [t for t in tiers if t.type == constants.SB_TIER_TYPE_CEPH]
max_replication, __ = \
StorageBackendConfig.get_ceph_max_replication(pecan.request.dbapi)
for tier in ceph_tiers:
replication = max_replication # In case tier has no storage backend configured
if tier.get('forbackendid'):
bk = pecan.request.dbapi.storage_ceph_get(tier.forbackendid)
replication, __ = \
StorageBackendConfig.get_ceph_pool_replication(pecan.request.dbapi, bk)
stors = pecan.request.dbapi.istor_get_by_tier(tier.id)
if len(stors) < replication:
word = 'is' if replication == 1 else 'are'
msg = ("Can not unlock node until at least %(replication)s osd stor %(word)s "
"configured for tier '%(tier)s'."
% {'replication': str(replication), 'word': word, 'tier': tier['name']})
raise wsme.exc.ClientSideError(msg)
else:
if utils.is_aio_duplex_system(pecan.request.dbapi):
host_stors = pecan.request.dbapi.istor_get_by_ihost(ihost['id'])
if not host_stors:
raise wsme.exc.ClientSideError(
_("Can not unlock node until at least one OSD is configured."))
tiers = pecan.request.dbapi.storage_tier_get_all()
ceph_tiers = [t for t in tiers if t.type == constants.SB_TIER_TYPE_CEPH]
# On a two-node configuration, both nodes should have at least one OSD
# in each tier. Otherwise, the cluster is remains in an error state.
for tier in ceph_tiers:
stors = tier['stors']
host_has_osd_in_tier = False
for stor in stors:
if stor['forihostid'] == ihost['id']:
host_has_osd_in_tier = True
if not host_has_osd_in_tier:
raise wsme.exc.ClientSideError(
"Can not unlock node until every storage tier has at least one OSD "
"configured. Tier \"%s\" has no OSD configured." % tier['name'])
else:
storage_nodes = []
try:
storage_nodes = pecan.request.dbapi.ihost_get_by_personality(
personality=constants.STORAGE)
except Exception:
# We are unlocking worker node when no storage nodes are
# defined. This is ok in CEPH_CONTROLLER_MODEL.
pass
is_storage_host_unlocked = False
if storage_nodes:
for node in storage_nodes:
if (node.administrative == constants.ADMIN_UNLOCKED and
(node.operational ==
constants.OPERATIONAL_ENABLED)):
is_storage_host_unlocked = True
break
stor_model = ceph.get_ceph_storage_model()
if (not is_storage_host_unlocked and
not stor_model == constants.CEPH_CONTROLLER_MODEL):
raise wsme.exc.ClientSideError(
_("Can not unlock a worker node until at "
"least one storage node is unlocked and enabled."))
# Local Storage checks
labels = pecan.request.dbapi.label_get_by_host(ihost['uuid'])
if cutils.has_openstack_compute(labels):
self._semantic_check_nova_local_storage(ihost['uuid'],
ihost['personality'])
@staticmethod
def check_unlock_storage(hostupdate):
"""Storage unlock semantic checks"""
# Semantic Check: Cannot unlock a storage node without
# any Storage Volumes (OSDs) configured
LOG.info("%s ihost check_unlock_storage" % hostupdate.displayid)
ihost = hostupdate.ihost_orig
istors = pecan.request.dbapi.istor_get_by_ihost(ihost['uuid'])
if len(istors) == 0:
raise wsme.exc.ClientSideError(
_("Can not unlock a storage node without any Storage Volumes configured"))
ceph_helper = ceph.CephApiOperator()
num_monitors, required_monitors, quorum_names = \
ceph_helper.get_monitors_status(pecan.request.dbapi)
if num_monitors < required_monitors:
raise wsme.exc.ClientSideError(
_("Can not unlock storage node. Only %d storage "
"monitor available. At least %s unlocked and "
"enabled hosts with monitors are required. Please"
" ensure hosts with monitors are unlocked and "
"enabled.") %
(num_monitors, constants.MIN_STOR_MONITORS))
# Check Ceph configuration, if it is wiped out (in the Backup & Restore
# process) then restore the configuration.
try:
if not pecan.request.rpcapi.restore_ceph_config(pecan.request.context):
raise Exception()
except Exception:
raise wsme.exc.ClientSideError(
_("Restore Ceph config failed. Retry unlocking storage node."))
@staticmethod
def check_updates_while_unlocked(hostupdate, delta):
"""Check semantics host-update of an unlocked host."""
ihost = hostupdate.ihost_patch
if ihost['administrative'] == constants.ADMIN_UNLOCKED:
deltaset = set(delta)
restricted_updates = ()
if not pecan.request.user_agent.startswith('mtce'):
# Allow mtc to modify the state throughthe REST API.
# Eventually mtc should switch to using the
# conductor API to modify ihosts because this check will also
# allow users to modify these states (which is bad).
restricted_updates = ('administrative',
'availability',
'operational',
'subfunction_oper',
'subfunction_avail',
'task', 'uptime')
if deltaset.issubset(restricted_updates):
raise wsme.exc.ClientSideError(
("Change set %s contains a subset of restricted %s." %
(deltaset, restricted_updates)))
else:
LOG.debug("PASS deltaset=%s restricted_updates=%s" %
(deltaset, restricted_updates))
if 'administrative' in delta:
# Transition to unlocked
if ihost['ihost_action']:
LOG.info("Host: %s Admin state change to: %s "
"Clearing ihost_action=%s" %
(ihost['uuid'],
ihost['administrative'],
ihost['ihost_action']))
hostupdate.ihost_val_update({'ihost_action': ""})
pass
@staticmethod
def check_force_swact(hostupdate):
"""Pre swact semantic checks for controller"""
# Allow force-swact to continue
return True
@staticmethod
def check_reboot(hostupdate):
"""Pre reboot semantic checks"""
# Semantic Check: State Dependency: Reboot case
if hostupdate.ihost_orig['administrative'] == constants.ADMIN_UNLOCKED:
raise wsme.exc.ClientSideError(
_("Can not 'Reboot' an 'unlocked' host %s; "
"Please 'Lock' first" % hostupdate.displayid))
if utils.get_system_mode() == constants.SYSTEM_MODE_SIMPLEX:
raise wsme.exc.ClientSideError(_(
"Rebooting a simplex system is not allowed."))
return True
@staticmethod
def _semantic_check_tpm_config(ihost):
"""Pre swact/unlock semantic checks for TPM configuration"""
tpmconfig = utils.get_tpm_config()
if tpmconfig:
# retrieve the tpmdevice configuration for this host.
# If this host got Reinstalled or Restored, and it had
# TPM configured on it prior, then we should still find
# a valid tpmdevice entry for this host. Otherwise this
# is a new host or a previous host that was deleted and re-added
tpmdevice = \
pecan.request.dbapi.tpmdevice_get_by_host(ihost['uuid'])
if not tpmdevice or len(tpmdevice) > 1:
raise wsme.exc.ClientSideError(
_("Global TPM configuration found; but "
"no valid TPM Device configuration on host %s." %
ihost['hostname']))
tpmdevice = tpmdevice[0]
if tpmdevice.state == constants.TPMCONFIG_APPLYING:
raise wsme.exc.ClientSideError(
_("TPM configuration in progress on host %s; "
"Please wait for operation to complete "
"before re-attempting." % ihost['hostname']))
elif tpmdevice.state != constants.TPMCONFIG_APPLIED:
# if the TPM certificate for this host is not
# preserved as tpm_data, then disallow unlock/swact
if not tpmdevice.tpm_data:
raise wsme.exc.ClientSideError(
_("TPM configuration not fully applied on host %s; "
"Please run system certificate-install -m tpm_mode "
"before re-attempting." % ihost['hostname']))
def _semantic_check_swact_upgrade(self, from_host, to_host, force_swact=False):
"""
Perform semantic checks related to upgrades prior to swacting host.
"""
# First check if we are in an upgrade
try:
upgrade = pecan.request.dbapi.software_upgrade_get_one()
except exception.NotFound:
# No upgrade in progress so nothing to check
return
# Get the load running on the destination controller
host_upgrade = objects.host_upgrade.get_by_host_id(
pecan.request.context, to_host['id'])
to_host_load_id = host_upgrade.software_load
# Get the load names
from_sw_version = objects.load.get_by_uuid(
pecan.request.context, upgrade.from_load).software_version
to_sw_version = objects.load.get_by_uuid(
pecan.request.context, upgrade.to_load).software_version
to_host_sw_version = objects.load.get_by_uuid(
pecan.request.context, to_host_load_id).software_version
if upgrade.state in [constants.UPGRADE_STARTING,
constants.UPGRADE_STARTED,
constants.UPGRADE_DATA_MIGRATION]:
# Swacting controllers is not supported until database migration is complete
raise wsme.exc.ClientSideError(
_("Swact action not allowed. Upgrade state must be %s") %
(constants.UPGRADE_DATA_MIGRATION_COMPLETE))
if upgrade.state in [constants.UPGRADE_ABORTING,
constants.UPGRADE_ABORTING_ROLLBACK]:
if to_host_load_id == upgrade.to_load:
# Cannot swact to new load if aborting upgrade
raise wsme.exc.ClientSideError(
_("Aborting upgrade: %s must be using load %s before this "
"operation can proceed. Currently using load %s.") %
(to_host['hostname'], from_sw_version, to_host_sw_version))
elif to_host_load_id == upgrade.from_load:
# On CPE loads we must abort before we swact back to the old load
# Any VMs on the active controller will be lost during the swact
if constants.WORKER in to_host.subfunctions:
raise wsme.exc.ClientSideError(
_("Upgrading: %s must be using load %s before this "
"operation can proceed. Currently using load %s.") %
(to_host['hostname'], to_sw_version, to_host_sw_version))
# Check for new hardware since upgrade-start
self._semantic_check_upgrade_refresh(upgrade, to_host, force_swact)
def check_swact(self, hostupdate, force_swact=False):
"""Pre swact semantic checks for controller"""
if hostupdate.ihost_orig['personality'] != constants.CONTROLLER:
raise wsme.exc.ClientSideError(
_("Swact action not allowed for non controller host %s." %
hostupdate.ihost_orig['hostname']))
if hostupdate.ihost_orig['administrative'] == constants.ADMIN_LOCKED:
raise wsme.exc.ClientSideError(
_("Controller is Locked ; No services to Swact"))
if utils.get_system_mode() == constants.SYSTEM_MODE_SIMPLEX:
raise wsme.exc.ClientSideError(_(
"Swact action not allowed for a simplex system."))
# check target controller
ihost_ctrs = pecan.request.dbapi.ihost_get_by_personality(
personality=constants.CONTROLLER)
for ihost_ctr in ihost_ctrs:
if ihost_ctr.hostname != hostupdate.ihost_orig['hostname']:
if (ihost_ctr.operational !=
constants.OPERATIONAL_ENABLED):
raise wsme.exc.ClientSideError(
_("%s is not enabled and has operational "
"state %s."
"Standby controller must be operationally "
"enabled.") %
(ihost_ctr.hostname, ihost_ctr.operational))
if (ihost_ctr.availability ==
constants.AVAILABILITY_DEGRADED):
health_helper = health.Health(pecan.request.dbapi)
degrade_alarms = health_helper.get_alarms_degrade(
pecan.request.context,
alarm_ignore_list=[
fm_constants.FM_ALARM_ID_HA_SERVICE_GROUP_STATE,
fm_constants.FM_ALARM_ID_HA_SERVICE_GROUP_REDUNDANCY,
fm_constants.FM_ALARM_ID_HA_NODE_LICENSE,
fm_constants.FM_ALARM_ID_HA_COMMUNICATION_FAILURE
],
entity_instance_id_filter=ihost_ctr.hostname)
if degrade_alarms:
raise wsme.exc.ClientSideError(
_("%s has degraded availability status. "
"Standby controller must be in available status.") %
(ihost_ctr.hostname))
if constants.WORKER in ihost_ctr.subfunctions:
if (ihost_ctr.subfunction_oper !=
constants.OPERATIONAL_ENABLED):
raise wsme.exc.ClientSideError(
_("%s subfunction is not enabled and has "
"operational state %s."
"Standby controller subfunctions %s "
"must all be operationally enabled.") %
(ihost_ctr.hostname,
ihost_ctr.subfunction_oper,
ihost_ctr.subfunctions))
# deny swact if storage backend not ready
self._semantic_check_storage_backend(ihost_ctr)
if ihost_ctr.config_target:
if ihost_ctr.config_target != ihost_ctr.config_applied:
try:
upgrade = \
pecan.request.dbapi.software_upgrade_get_one()
except exception.NotFound:
upgrade = None
if upgrade and upgrade.state == \
constants.UPGRADE_ABORTING_ROLLBACK:
pass
else:
raise wsme.exc.ClientSideError(
_("%s target Config %s not yet applied."
" Apply target Config via Lock/Unlock prior"
" to Swact") %
(ihost_ctr.hostname, ihost_ctr.config_target))
self._semantic_check_swact_upgrade(hostupdate.ihost_orig,
ihost_ctr,
force_swact)
# If HTTPS is enabled then we may be in TPM mode
if utils.get_https_enabled():
self._semantic_check_tpm_config(ihost_ctr)
# Check: If DRBD is resizing
controller_fs_list = pecan.request.dbapi.controller_fs_get_list()
for controller_fs in controller_fs_list:
if controller_fs['replicated']:
if (controller_fs.get('state') ==
constants.CONTROLLER_FS_RESIZING_IN_PROGRESS):
raise wsme.exc.ClientSideError(
_("drbd '%s' is resizing. Wait for the resizing to "
"complete before issuing Swact") %
(controller_fs['name']))
# Check: Valid Swact action: Pre-Swact Check
response = sm_api.swact_pre_check(hostupdate.ihost_orig['hostname'],
timeout=30)
if response and "0" != response['error_code']:
raise wsme.exc.ClientSideError(
_("%s" % response['error_details']))
def check_lock_storage(self, hostupdate, force=False):
"""Pre lock semantic checks for storage"""
LOG.info("%s ihost check_lock_storage" % hostupdate.displayid)
backend = StorageBackendConfig.get_configured_backend(
pecan.request.dbapi,
constants.CINDER_BACKEND_CEPH
)
if not backend:
raise wsme.exc.ClientSideError(
_("Ceph must be configured as a backend."))
if (backend.task == constants.SB_TASK_RESTORE and force):
LOG.info("%s Allow force-locking as ceph backend is in "
"restore mode" % hostupdate.displayid)
return
ceph_pools_empty = False
if (hostupdate.ihost_orig['administrative'] ==
constants.ADMIN_UNLOCKED and
hostupdate.ihost_orig['operational'] ==
constants.OPERATIONAL_ENABLED):
num_monitors, required_monitors, quorum_names = \
self._ceph.get_monitors_status(pecan.request.dbapi)
if (hostupdate.ihost_orig['hostname'] in quorum_names and
num_monitors - 1 < required_monitors):
raise wsme.exc.ClientSideError(_(
"Only %d storage "
"monitor available. At least %s unlocked and "
"enabled hosts with monitors are required. Please"
" ensure hosts with monitors are unlocked and "
"enabled.") %
(num_monitors, constants.MIN_STOR_MONITORS))
storage_nodes = pecan.request.dbapi.ihost_get_by_personality(
constants.STORAGE)
replication, min_replication = \
StorageBackendConfig.get_ceph_max_replication(pecan.request.dbapi)
available_peer_count = 0
for node in storage_nodes:
if (node['id'] != hostupdate.ihost_orig['id'] and
node['peer_id'] == hostupdate.ihost_orig['peer_id']):
ihost_action_locking = False
ihost_action = node['ihost_action'] or ""
if (ihost_action.startswith(constants.FORCE_LOCK_ACTION) or
ihost_action.startswith(constants.LOCK_ACTION)):
ihost_action_locking = True
if (node['administrative'] == constants.ADMIN_UNLOCKED and
node['operational'] == constants.OPERATIONAL_ENABLED and not
ihost_action_locking):
available_peer_count += 1
if available_peer_count < min_replication:
pools_usage = \
pecan.request.rpcapi.get_ceph_pools_df_stats(pecan.request.context)
if not pools_usage:
raise wsme.exc.ClientSideError(
_("Cannot lock a storage node when ceph pool usage is undetermined."))
ceph_pools_empty = self._ceph.ceph_pools_empty(
pecan.request.dbapi, pools_usage)
if not ceph_pools_empty:
msg = _(
"Cannot lock a storage node when ceph pools are"
" not empty and replication is lost. This may"
" result in data loss. ")
# Ceph pool is not empty and no other enabled storage
# in set, so locking this storage node is not allowed.
raise wsme.exc.ClientSideError(msg)
# Perform checks on storage regardless of operational state
# as a minimum number of monitor is required.
if not force:
# Check if there is upgrade in progress
try:
upgrade = pecan.request.dbapi.software_upgrade_get_one()
if upgrade.state in [constants.UPGRADE_ABORTING_ROLLBACK]:
LOG.info("%s not in a force lock and in an upgrade abort, "
"do not check Ceph status"
% hostupdate.displayid)
return
except exception.NotFound:
pass
if not self._ceph.ceph_status_ok():
LOG.info("%s ceph_status_ok() returned not ok"
% hostupdate.displayid)
host_health = self._ceph.host_osd_status(
hostupdate.ihost_orig['hostname'])
LOG.info("%s check OSD host_health=%s" %
(hostupdate.displayid, host_health))
if (host_health is None or
host_health == constants.CEPH_HEALTH_BLOCK):
LOG.info("%s host_health=%s" %
(hostupdate.displayid, host_health))
if not ceph_pools_empty:
msg = _("Cannot lock a storage node when ceph pools are not empty "
"and replication is lost. This may result in data loss. ")
raise wsme.exc.ClientSideError(msg)
def check_lock_worker(self, hostupdate, force=False):
"""Pre lock semantic checks for worker"""
hostname = hostupdate.ihost_patch.get('hostname')
LOG.info("%s host check_lock_worker" % hostupdate.displayid)
if force:
LOG.info("Forced lock of host: %s" % hostname)
return
system = pecan.request.dbapi.isystem_get_one()
system_mode = system.system_mode
system_type = system.system_type
if system_mode == constants.SYSTEM_MODE_SIMPLEX:
return
# Check upgrade state for controllers with worker subfunction
subfunctions_set = \
set(hostupdate.ihost_patch[constants.SUBFUNCTIONS].split(','))
if (hostupdate.ihost_orig['personality'] == constants.CONTROLLER and
constants.WORKER in subfunctions_set):
upgrade = None
try:
upgrade = pecan.request.dbapi.software_upgrade_get_one()
upgrade_state = upgrade.state
except exception.NotFound:
upgrade_state = None
if upgrade_state in [
constants.UPGRADE_STARTING,
constants.UPGRADE_STARTED,
constants.UPGRADE_DATA_MIGRATION,
constants.UPGRADE_DATA_MIGRATION_COMPLETE,
constants.UPGRADE_DATA_MIGRATION_FAILED]:
if system_type == constants.TIS_AIO_BUILD:
if hostname == constants.CONTROLLER_1_HOSTNAME:
# Allow AIO-DX lock of controller-1
return
raise wsme.exc.ClientSideError(
_("Rejected: Can not lock %s with worker function "
"at this upgrade stage '%s'.") %
(hostupdate.displayid, upgrade_state))
if upgrade_state in [constants.UPGRADE_UPGRADING_CONTROLLERS]:
if system_type == constants.TIS_AIO_BUILD:
# Allow lock for AIO-DX controller-0 after upgrading
# controller-1. Allow lock for AIO-DX controllers.
if hostname == constants.CONTROLLER_0_HOSTNAME:
return
raise wsme.exc.ClientSideError(
_("Rejected: Can not lock %s with worker function "
"at this upgrade stage '%s'.") %
(hostupdate.displayid, upgrade_state))
# Worker node with a Ceph Monitor service? Make sure at least
# two monitors will remain up after lock.
host_id = hostupdate.ihost_orig.get('id')
ceph_mon = pecan.request.dbapi.ceph_mon_get_by_ihost(host_id)
if ceph_mon:
if (hostupdate.ihost_orig['personality'] ==
constants.WORKER and
hostupdate.ihost_orig['administrative'] ==
constants.ADMIN_UNLOCKED and
hostupdate.ihost_orig['operational'] ==
constants.OPERATIONAL_ENABLED):
num_monitors, required_monitors, quorum_names = \
self._ceph.get_monitors_status(pecan.request.dbapi)
if (hostname in quorum_names and
num_monitors - 1 < required_monitors):
raise wsme.exc.ClientSideError(_(
"Only %d Ceph "
"monitors available. At least %s unlocked and "
"enabled hosts with monitors are required. "
"Please ensure hosts with monitors are "
"unlocked and enabled.") %
(num_monitors, constants.MIN_STOR_MONITORS))
def check_unlock_interfaces(self, hostupdate):
"""Semantic check for interfaces on host-unlock."""
ihost = hostupdate.ihost_patch
if ihost['personality'] in [constants.CONTROLLER, constants.WORKER,
constants.STORAGE]:
# Check if there is an management interface on
# controller/worker/storage
ihost_iinterfaces = pecan.request.dbapi.iinterface_get_by_ihost(
ihost['uuid'])
network = pecan.request.dbapi.network_get_by_type(
constants.NETWORK_TYPE_MGMT)
mgmt_interface_configured = False
for iif in ihost_iinterfaces:
if iif.networks and str(network.id) in iif.networks:
mgmt_interface_configured = True
if not mgmt_interface_configured:
msg = _("Cannot unlock host %s "
"without configuring a management interface."
% hostupdate.displayid)
raise wsme.exc.ClientSideError(msg)
if utils.is_kubernetes_config():
# Check if there is a cluster-host interface on
# controller/worker/storage
host_interfaces = pecan.request.dbapi.iinterface_get_by_ihost(
ihost['uuid'])
network = pecan.request.dbapi.network_get_by_type(
constants.NETWORK_TYPE_CLUSTER_HOST)
for iif in host_interfaces:
if iif.networks and str(network.id) in iif.networks:
break
else:
msg = _("Cannot unlock host %s "
"without configuring a cluster-host interface."
% hostupdate.displayid)
raise wsme.exc.ClientSideError(msg)
hostupdate.configure_required = True
def check_unlock_partitions(self, hostupdate):
"""Semantic check for interfaces on host-unlock."""
ihost = hostupdate.ihost_patch
partitions = pecan.request.dbapi.partition_get_by_ihost(ihost['uuid'])
partition_transitory_states = [
constants.PARTITION_CREATE_IN_SVC_STATUS,
constants.PARTITION_DELETING_STATUS,
constants.PARTITION_MODIFYING_STATUS]
for part in partitions:
if part.status in partition_transitory_states:
msg = _("Cannot unlock host %s "
"while partitions on the host are in a "
"creating/deleting/modifying state."
% hostupdate.displayid)
raise wsme.exc.ClientSideError(msg)
@staticmethod
def unlock_update_mgmt_interface(ihost):
# MTU Update: Compute and storage nodes get MTU values for
# management via DHCP. This 'check' updates the 'imtu' value based on
# what will be served via DHCP.
if ihost['personality'] in [constants.WORKER, constants.STORAGE]:
host_list = pecan.request.dbapi.ihost_get_by_personality(
personality=constants.CONTROLLER)
interface_list_active = []
for h in host_list:
if utils.is_host_active_controller(h):
interface_list_active = \
pecan.request.dbapi.iinterface_get_all(h.id)
break
ihost_iinterfaces = \
pecan.request.dbapi.iinterface_get_by_ihost(ihost['uuid'])
# updated management interfaces
idata = {}
for iif in ihost_iinterfaces:
iif_networktype = []
if iif.networktype:
iif_networktype = [network.strip() for network in iif.networktype.split(",")]
if any(network in [constants.NETWORK_TYPE_MGMT] for network in iif_networktype):
for ila in interface_list_active:
ila_networktype = []
if ila.networktype:
ila_networktype = [network.strip() for network in ila.networktype.split(",")]
if any(network in ila_networktype for network in iif_networktype):
idata['imtu'] = ila.imtu
pecan.request.dbapi.iinterface_update(iif.uuid, idata)
break
def stage_action(self, action, hostupdate):
""" Stage the action to be performed.
"""
LOG.info("%s stage_action %s" % (hostupdate.displayid, action))
rc = True
if not action or (action and
action.lower() == constants.NONE_ACTION):
LOG.error("Unrecognized action perform: %s" % action)
return False
if (action == constants.UNLOCK_ACTION or
action == constants.FORCE_UNLOCK_ACTION):
self._handle_unlock_action(hostupdate)
elif action == constants.LOCK_ACTION:
self._handle_lock_action(hostupdate)
elif action == constants.FORCE_LOCK_ACTION:
self._handle_force_lock_action(hostupdate)
elif action == constants.SWACT_ACTION:
self._stage_swact(hostupdate)
elif action == constants.FORCE_SWACT_ACTION:
self._stage_force_swact(hostupdate)
elif action == constants.REBOOT_ACTION:
self._stage_reboot(hostupdate)
elif action == constants.RESET_ACTION:
self._stage_reset(hostupdate)
elif action == constants.REINSTALL_ACTION:
self._stage_reinstall(hostupdate)
elif action == constants.POWERON_ACTION:
self._stage_poweron(hostupdate)
elif action == constants.POWEROFF_ACTION:
self._stage_poweroff(hostupdate)
elif action == constants.VIM_SERVICES_ENABLED:
self._handle_vim_services_enabled(hostupdate)
elif action == constants.VIM_SERVICES_DISABLED:
if not self._handle_vim_services_disabled(hostupdate):
LOG.warn(_("%s Exit _handle_vim_services_disabled" %
hostupdate.ihost_patch['hostname']))
hostupdate.nextstep = hostupdate.EXIT_RETURN_HOST
rc = False
elif action == constants.VIM_SERVICES_DISABLE_FAILED:
if not self._handle_vim_services_disable_failed(hostupdate):
LOG.warn(_("%s Exit _handle_vim_services_disable failed" %
hostupdate.ihost_patch['hostname']))
hostupdate.nextstep = hostupdate.EXIT_RETURN_HOST
rc = False
elif action == constants.VIM_SERVICES_DISABLE_EXTEND:
self._handle_vim_services_disable_extend(hostupdate)
hostupdate.nextstep = hostupdate.EXIT_UPDATE_PREVAL
rc = False
elif action == constants.VIM_SERVICES_DELETE_FAILED:
self._handle_vim_services_delete_failed(hostupdate)
hostupdate.nextstep = hostupdate.EXIT_UPDATE_PREVAL
rc = False
elif action == constants.APPLY_PROFILE_ACTION:
self._stage_apply_profile_action(hostupdate)
elif action == constants.SUBFUNCTION_CONFIG_ACTION:
# Not a mtc action; disable mtc checks and config
self._stage_subfunction_config(hostupdate)
else:
# TODO: raise wsme
LOG.error("%s Unrecognized action perform: %s" %
(hostupdate.displayid, action))
rc = False
if hostupdate.nextstep == hostupdate.EXIT_RETURN_HOST:
LOG.info("%s stage_action aborting request %s %s" %
(hostupdate.displayid,
hostupdate.action,
hostupdate.delta))
return rc
@staticmethod
def _stage_apply_profile_action(hostupdate):
"""Stage apply profile action."""
LOG.info("%s _stage_apply_profile_action uuid=%s profile_uuid=%s" %
(hostupdate.displayid,
hostupdate.ihost_patch['uuid'],
hostupdate.iprofile_uuid))
profile.apply_profile(hostupdate.ihost_patch['uuid'],
hostupdate.iprofile_uuid)
hostupdate.notify_mtce = False
hostupdate.configure_required = False
@staticmethod
def _check_subfunction_config(hostupdate):
"""Check subfunction config."""
LOG.info("%s _check_subfunction_config" % hostupdate.displayid)
patched_ihost = hostupdate.ihost_patch
if patched_ihost['action'] == "subfunction_config":
if not patched_ihost['subfunctions'] or \
patched_ihost['personality'] == patched_ihost['subfunctions']:
raise wsme.exc.ClientSideError(
_("This host is not configured with a subfunction."))
return True
@staticmethod
def _stage_subfunction_config(hostupdate):
"""Stage subfunction config."""
LOG.info("%s _stage_subfunction_config" % hostupdate.displayid)
hostupdate.notify_mtce = False
hostupdate.skip_notify_mtce = True
@staticmethod
def perform_action_subfunction_config(ihost_obj):
"""Perform subfunction config via RPC to conductor."""
LOG.info("%s perform_action_subfunction_config" %
ihost_obj['hostname'])
pecan.request.rpcapi.configure_ihost(pecan.request.context,
ihost_obj,
do_worker_apply=True)
@staticmethod
def _stage_reboot(hostupdate):
"""Stage reboot action."""
LOG.info("%s stage_reboot" % hostupdate.displayid)
hostupdate.notify_mtce = True
def _stage_reinstall(self, hostupdate):
"""Stage reinstall action."""
LOG.info("%s stage_reinstall" % hostupdate.displayid)
# Remove manifests to enable standard install without manifests
# and enable storage allocation change
pecan.request.rpcapi.remove_host_config(
pecan.request.context,
hostupdate.ihost_orig['uuid'])
hostupdate.notify_mtce = True
if hostupdate.ihost_orig['personality'] == constants.STORAGE:
istors = pecan.request.dbapi.istor_get_by_ihost(
hostupdate.ihost_orig['uuid'])
for stor in istors:
istor_obj = objects.storage.get_by_uuid(
pecan.request.context, stor.uuid)
self._ceph.remove_osd_key(istor_obj['osdid'])
hostupdate.ihost_val_update({constants.HOST_ACTION_STATE:
constants.HAS_REINSTALLING})
@staticmethod
def _stage_poweron(hostupdate):
"""Stage poweron action."""
LOG.info("%s stage_poweron" % hostupdate.displayid)
hostupdate.notify_mtce = True
@staticmethod
def _stage_poweroff(hostupdate):
"""Stage poweroff action."""
LOG.info("%s stage_poweroff" % hostupdate.displayid)
hostupdate.notify_mtce = True
@staticmethod
def _stage_swact(hostupdate):
"""Stage swact action."""
LOG.info("%s stage_swact" % hostupdate.displayid)
hostupdate.notify_mtce = True
@staticmethod
def _stage_force_swact(hostupdate):
"""Stage force-swact action."""
LOG.info("%s stage_force_swact" % hostupdate.displayid)
hostupdate.notify_mtce = True
@staticmethod
def _handle_vim_services_enabled(hostupdate):
"""Handle VIM services-enabled signal."""
vim_progress_status = (hostupdate.ihost_orig.get('vim_progress_status') or "")
LOG.info("%s received services-enabled task=%s vim_progress_status=%s"
% (hostupdate.displayid,
hostupdate.ihost_orig['task'],
vim_progress_status))
if (not vim_progress_status or
not vim_progress_status.startswith(constants.VIM_SERVICES_ENABLED)):
hostupdate.notify_availability = constants.VIM_SERVICES_ENABLED
if (not vim_progress_status or
vim_progress_status == constants.VIM_SERVICES_DISABLED):
# otherwise allow the audit to clear the error message
hostupdate.ihost_val_update({'vim_progress_status':
constants.VIM_SERVICES_ENABLED})
hostupdate.skip_notify_mtce = True
@staticmethod
def _handle_vim_services_disabled(hostupdate):
"""Handle VIM services-disabled signal."""
LOG.info("%s _handle_vim_services_disabled'" % hostupdate.displayid)
ihost = hostupdate.ihost_orig
hostupdate.ihost_val_update(
{'vim_progress_status': constants.VIM_SERVICES_DISABLED})
ihost_task_string = ihost['ihost_action'] or ""
if ((ihost_task_string.startswith(constants.LOCK_ACTION) or
ihost_task_string.startswith(constants.FORCE_LOCK_ACTION)) and
ihost['administrative'] != constants.ADMIN_LOCKED):
# passed - skip reset for force-lock
# iHost['ihost_action'] = constants.LOCK_ACTION
hostupdate.notify_availability = constants.VIM_SERVICES_DISABLED
hostupdate.notify_action_lock = True
hostupdate.notify_mtce = True
else:
# return False rather than failing request.
LOG.warn(_("%s Admin action task not Locking or Force Locking "
"upon receipt of 'services-disabled'." %
hostupdate.displayid))
hostupdate.skip_notify_mtce = True
return False
return True
@staticmethod
def _handle_vim_services_disable_extend(hostupdate):
"""Handle VIM services-disable-extend signal."""
ihost_action = hostupdate.ihost_orig['ihost_action'] or ""
result_reason = hostupdate.ihost_patch.get('vim_progress_status') or ""
LOG.info("%s handle_vim_services_disable_extend ihost_action=%s reason=%s" %
(hostupdate.displayid, ihost_action, result_reason))
hostupdate.skip_notify_mtce = True
if ihost_action.startswith(constants.LOCK_ACTION):
val = {'task': constants.LOCKING + '-',
'ihost_action': constants.LOCK_ACTION}
hostupdate.ihost_val_prenotify_update(val)
else:
LOG.warn("%s Skip vim services disable extend ihost action=%s" %
(hostupdate.displayid, ihost_action))
return False
# If the VIM updates vim_progress_status, could post:
# if result_reason:
# hostupdate.ihost_val_prenotify_update({'vim_progress_status':
# result_reason})
# else:
# hostupdate.ihost_val_prenotify_update(
# {'vim_progress_status': constants.VIM_SERVICES_DELETE_FAILED})
LOG.info("services-disable-extend reason=%s" % result_reason)
return True
@staticmethod
def _handle_vim_services_disable_failed(hostupdate):
"""Handle VIM services-disable-failed signal."""
ihost_task_string = hostupdate.ihost_orig['ihost_action'] or ""
LOG.info("%s handle_vim_services_disable_failed ihost_action=%s" %
(hostupdate.displayid, ihost_task_string))
result_reason = hostupdate.ihost_patch.get('vim_progress_status') or ""
if ihost_task_string.startswith(constants.LOCK_ACTION):
hostupdate.skip_notify_mtce = True
val = {'ihost_action': '',
'task': '',
'vim_progress_status': result_reason}
hostupdate.ihost_val_prenotify_update(val)
hostupdate.ihost_val.update(val)
hostupdate.skip_notify_mtce = True
elif ihost_task_string.startswith(constants.FORCE_LOCK_ACTION):
# allow mtce to reset the host
hostupdate.notify_mtce = True
hostupdate.notify_action_lock_force = True
else:
hostupdate.skip_notify_mtce = True
LOG.warn("%s Skipping vim services disable notification task=%s" %
(hostupdate.displayid, ihost_task_string))
return False
if result_reason:
LOG.info("services-disable-failed reason=%s" % result_reason)
hostupdate.ihost_val_update({'vim_progress_status':
result_reason})
else:
hostupdate.ihost_val_update({'vim_progress_status':
constants.VIM_SERVICES_DISABLE_FAILED})
return True
@staticmethod
def _handle_vim_services_delete_failed(hostupdate):
"""Handle VIM services-delete-failed signal."""
ihost_admin = hostupdate.ihost_orig['administrative'] or ""
result_reason = hostupdate.ihost_patch.get('vim_progress_status') or ""
LOG.info("%s handle_vim_services_delete_failed admin=%s reason=%s" %
(hostupdate.displayid, ihost_admin, result_reason))
hostupdate.skip_notify_mtce = True
if ihost_admin.startswith(constants.ADMIN_LOCKED):
val = {'ihost_action': '',
'task': '',
'vim_progress_status': result_reason}
hostupdate.ihost_val_prenotify_update(val)
# hostupdate.ihost_val.update(val)
else:
LOG.warn("%s Skip vim services delete failed notify admin=%s" %
(hostupdate.displayid, ihost_admin))
return False
if result_reason:
hostupdate.ihost_val_prenotify_update({'vim_progress_status':
result_reason})
else:
hostupdate.ihost_val_prenotify_update(
{'vim_progress_status': constants.VIM_SERVICES_DELETE_FAILED})
LOG.info("services-disable-failed reason=%s" % result_reason)
return True
@staticmethod
def _stage_reset(hostupdate):
"""Handle host-reset action."""
LOG.info("%s _stage_reset" % hostupdate.displayid)
hostupdate.notify_mtce = True
def _handle_unlock_action(self, hostupdate):
"""Handle host-unlock action."""
LOG.info("%s _handle_unlock_action" % hostupdate.displayid)
if hostupdate.ihost_patch.get('personality') == constants.STORAGE:
self._handle_unlock_storage_host(hostupdate)
elif hostupdate.ihost_patch.get('personality') == constants.WORKER:
self._handle_unlock_worker_host(hostupdate)
hostupdate.notify_vim_action = False
hostupdate.notify_mtce = True
val = {'ihost_action': constants.UNLOCK_ACTION}
hostupdate.ihost_val_prenotify_update(val)
hostupdate.ihost_val.update(val)
def _handle_unlock_storage_host(self, hostupdate):
self._ceph.update_crushmap(hostupdate)
def _handle_unlock_worker_host(self, hostupdate):
# Update crushmap if we unlocked the worker with a ceph monitor.
monitor_list = pecan.request.dbapi.ceph_mon_get_list()
for mon in monitor_list:
ihost = pecan.request.dbapi.ihost_get(mon['forihostid'])
if ihost.id == hostupdate.ihost_orig['id']:
self._ceph.update_crushmap(hostupdate)
@staticmethod
def _handle_lock_action(hostupdate):
"""Handle host-lock action."""
LOG.info("%s _handle_lock_action" % hostupdate.displayid)
hostupdate.notify_vim_action = True
hostupdate.skip_notify_mtce = True
val = {'ihost_action': constants.LOCK_ACTION}
hostupdate.ihost_val_prenotify_update(val)
hostupdate.ihost_val.update(val)
@staticmethod
def _handle_force_lock_action(hostupdate):
"""Handle host-force-lock action."""
LOG.info("%s _handle_force_lock_action" % hostupdate.displayid)
hostupdate.notify_vim_action = True
hostupdate.skip_notify_mtce = True
val = {'ihost_action': constants.FORCE_LOCK_ACTION}
hostupdate.ihost_val_prenotify_update(val)
hostupdate.ihost_val.update(val)
def _create_node(host, xml_node, personality, is_dynamic_ip):
host_node = et.SubElement(xml_node, 'host')
et.SubElement(host_node, 'personality').text = personality
if personality == constants.WORKER:
et.SubElement(host_node, 'hostname').text = host.hostname
et.SubElement(host_node, 'subfunctions').text = host.subfunctions
et.SubElement(host_node, 'mgmt_mac').text = host.mgmt_mac
if not is_dynamic_ip:
et.SubElement(host_node, 'mgmt_ip').text = host.mgmt_ip
if host.location is not None and 'locn' in host.location:
et.SubElement(host_node, 'location').text = host.location['locn']
pw_on_instruction = _('Uncomment the statement below to power on the host '
'automatically through board management.')
host_node.append(et.Comment(pw_on_instruction))
host_node.append(et.Comment('<power_on />'))
et.SubElement(host_node, 'bm_type').text = host.bm_type
et.SubElement(host_node, 'bm_username').text = host.bm_username
et.SubElement(host_node, 'bm_password').text = ''
et.SubElement(host_node, 'boot_device').text = host.boot_device
et.SubElement(host_node, 'rootfs_device').text = host.rootfs_device
et.SubElement(host_node, 'install_output').text = host.install_output
if host.vsc_controllers is not None:
et.SubElement(host_node, 'vsc_controllers').text = host.vsc_controllers
et.SubElement(host_node, 'console').text = host.console
et.SubElement(host_node, 'tboot').text = host.tboot