e6f4e01ae5
Change-Id: Idd383bdc84fa4ed4f1a43bb11c308f2a296316d7 Closes-Bug: #1621842
763 lines
30 KiB
Python
763 lines
30 KiB
Python
# Copyright 2013 Mirantis, Inc.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
import logging
|
|
import time
|
|
|
|
from devops.helpers.helpers import tcp_ping_
|
|
from devops.helpers.helpers import wait_pass
|
|
from devops.helpers.helpers import wait
|
|
from devops.helpers.metaclasses import SingletonMeta
|
|
from devops.helpers.ntp import sync_time
|
|
from devops.models import Environment
|
|
from keystoneauth1 import exceptions
|
|
from proboscis.asserts import assert_equal
|
|
from proboscis.asserts import assert_true
|
|
import six
|
|
|
|
from core.models.collector_client import CollectorClient
|
|
from core.helpers.log_helpers import logwrap
|
|
from core.helpers.log_helpers import QuietLogger
|
|
|
|
from fuelweb_test.helpers.decorators import revert_info
|
|
from fuelweb_test.helpers.decorators import update_rpm_packages
|
|
from fuelweb_test.helpers.decorators import upload_manifests
|
|
from fuelweb_test.helpers.eb_tables import Ebtables
|
|
from fuelweb_test.helpers.fuel_actions import AdminActions
|
|
from fuelweb_test.helpers.fuel_actions import BaseActions
|
|
from fuelweb_test.helpers.fuel_actions import CobblerActions
|
|
from fuelweb_test.helpers.fuel_actions import NailgunActions
|
|
from fuelweb_test.helpers.fuel_actions import PostgresActions
|
|
from fuelweb_test.helpers.fuel_actions import NessusActions
|
|
from fuelweb_test.helpers.fuel_actions import FuelBootstrapCliActions
|
|
from fuelweb_test.helpers.ssh_manager import SSHManager
|
|
from fuelweb_test.helpers.utils import TimeStat
|
|
from fuelweb_test.helpers.utils import YamlEditor
|
|
from fuelweb_test.helpers import multiple_networks_hacks
|
|
from fuelweb_test.models.fuel_web_client import FuelWebClient
|
|
from fuelweb_test import settings
|
|
from fuelweb_test.settings import iface_alias
|
|
from fuelweb_test import logger
|
|
|
|
|
|
class EnvironmentModel(six.with_metaclass(SingletonMeta, object)):
|
|
"""EnvironmentModel.""" # TODO documentation
|
|
|
|
def __init__(self, config=None):
|
|
if not hasattr(self, "_virt_env"):
|
|
self._virt_env = None
|
|
if not hasattr(self, "_fuel_web"):
|
|
self._fuel_web = None
|
|
self._config = config
|
|
self.ssh_manager = SSHManager()
|
|
self.ssh_manager.initialize(
|
|
self.get_admin_node_ip(),
|
|
login=settings.SSH_CREDENTIALS['login'],
|
|
password=settings.SSH_CREDENTIALS['password']
|
|
)
|
|
self.admin_actions = AdminActions()
|
|
self.base_actions = BaseActions()
|
|
self.cobbler_actions = CobblerActions()
|
|
self.nailgun_actions = NailgunActions()
|
|
self.postgres_actions = PostgresActions()
|
|
self.fuel_bootstrap_actions = FuelBootstrapCliActions()
|
|
|
|
@property
|
|
def fuel_web(self):
|
|
if self._fuel_web is None:
|
|
self._fuel_web = FuelWebClient(self)
|
|
return self._fuel_web
|
|
|
|
def __repr__(self):
|
|
klass, obj_id = type(self), hex(id(self))
|
|
if getattr(self, '_fuel_web'):
|
|
ip = self.fuel_web.admin_node_ip
|
|
else:
|
|
ip = None
|
|
return "[{klass}({obj_id}), ip:{ip}]".format(klass=klass,
|
|
obj_id=obj_id,
|
|
ip=ip)
|
|
|
|
@property
|
|
def admin_node_ip(self):
|
|
return self.fuel_web.admin_node_ip
|
|
|
|
@property
|
|
def collector(self):
|
|
return CollectorClient(settings.ANALYTICS_IP, 'api/v1/json')
|
|
|
|
@logwrap
|
|
def add_syslog_server(self, cluster_id, port=5514):
|
|
self.fuel_web.add_syslog_server(
|
|
cluster_id, self.d_env.router(), port)
|
|
|
|
def bootstrap_nodes(self, devops_nodes, timeout=settings.BOOTSTRAP_TIMEOUT,
|
|
skip_timesync=False):
|
|
"""Lists registered nailgun nodes
|
|
Start vms and wait until they are registered on nailgun.
|
|
:rtype : List of registered nailgun nodes
|
|
"""
|
|
# self.dhcrelay_check()
|
|
|
|
for node in devops_nodes:
|
|
logger.info("Bootstrapping node: {}".format(node.name))
|
|
node.start()
|
|
# TODO(aglarendil): LP#1317213 temporary sleep
|
|
# remove after better fix is applied
|
|
time.sleep(5)
|
|
|
|
with TimeStat("wait_for_nodes_to_start_and_register_in_nailgun"):
|
|
wait(lambda: all(self.nailgun_nodes(devops_nodes)), 15, timeout,
|
|
timeout_msg='Bootstrap timeout for nodes: {}'
|
|
''.format([node.name for node in devops_nodes]))
|
|
|
|
if not skip_timesync:
|
|
self.sync_time()
|
|
return self.nailgun_nodes(devops_nodes)
|
|
|
|
def sync_time(self, nodes_names=None, skip_sync=False):
|
|
if nodes_names is None:
|
|
roles = ['fuel_master', 'fuel_slave']
|
|
nodes_names = [node.name for node in self.d_env.get_nodes()
|
|
if node.role in roles and
|
|
node.driver.node_active(node)]
|
|
logger.info("Please wait while time on nodes: {0} "
|
|
"will be synchronized"
|
|
.format(', '.join(sorted(nodes_names))))
|
|
new_time = sync_time(self.d_env, nodes_names, skip_sync)
|
|
for name in sorted(new_time):
|
|
logger.info("New time on '{0}' = {1}".format(name, new_time[name]))
|
|
|
|
@logwrap
|
|
def get_admin_node_ip(self):
|
|
return str(
|
|
self.d_env.nodes(
|
|
).admin.get_ip_address_by_network_name(
|
|
self.d_env.admin_net))
|
|
|
|
@logwrap
|
|
def get_ebtables(self, cluster_id, devops_nodes):
|
|
return Ebtables(self.get_target_devs(devops_nodes),
|
|
self.fuel_web.client.get_cluster_vlans(cluster_id))
|
|
|
|
def get_keys(self, node, custom=None, build_images=None,
|
|
iso_connect_as='cdrom'):
|
|
params = {
|
|
'device_label': settings.ISO_LABEL,
|
|
'iface': iface_alias('eth0'),
|
|
'ip': node.get_ip_address_by_network_name(
|
|
self.d_env.admin_net),
|
|
'mask': self.d_env.get_network(
|
|
name=self.d_env.admin_net).ip.netmask,
|
|
'gw': self.d_env.router(),
|
|
'hostname': ''.join((settings.FUEL_MASTER_HOSTNAME,
|
|
settings.DNS_SUFFIX)),
|
|
'nat_interface': self.d_env.nat_interface,
|
|
'nameserver': settings.DNS,
|
|
'showmenu': 'yes' if settings.SHOW_FUELMENU else 'no',
|
|
'wait_for_external_config': 'yes',
|
|
'build_images': '1' if build_images else '0',
|
|
'MASTER_NODE_EXTRA_PACKAGES': settings.MASTER_NODE_EXTRA_PACKAGES
|
|
}
|
|
# TODO(akostrikov) add tests for menu items/kernel parameters
|
|
# TODO(akostrikov) refactor it.
|
|
if iso_connect_as == 'usb':
|
|
keys = (
|
|
"<Wait>\n" # USB boot uses boot_menu=yes for master node
|
|
"<F12>\n"
|
|
"2\n"
|
|
)
|
|
else: # cdrom is default
|
|
keys = (
|
|
"<Wait>\n"
|
|
"<Wait>\n"
|
|
"<Wait>\n"
|
|
)
|
|
|
|
keys += (
|
|
"<Esc>\n"
|
|
"<Wait>\n"
|
|
"vmlinuz initrd=initrd.img"
|
|
" inst.ks=cdrom:LABEL=%(device_label)s:/ks.cfg"
|
|
" inst.repo=cdrom:LABEL=%(device_label)s:/"
|
|
" ip=%(ip)s::%(gw)s:%(mask)s:%(hostname)s"
|
|
":%(iface)s:off::: nameserver=%(nameserver)s"
|
|
" showmenu=%(showmenu)s\n"
|
|
" wait_for_external_config=%(wait_for_external_config)s"
|
|
" build_images=%(build_images)s\n"
|
|
" MASTER_NODE_EXTRA_PACKAGES='%(MASTER_NODE_EXTRA_PACKAGES)s'\n"
|
|
" <Enter>\n"
|
|
) % params
|
|
return keys
|
|
|
|
@staticmethod
|
|
def get_target_devs(devops_nodes):
|
|
return [
|
|
interface.target_dev for interface in [
|
|
val for var in map(lambda node: node.interfaces, devops_nodes)
|
|
for val in var]]
|
|
|
|
@property
|
|
def d_env(self):
|
|
if self._virt_env is None:
|
|
if not self._config:
|
|
try:
|
|
return Environment.get(name=settings.ENV_NAME)
|
|
except Exception:
|
|
self._virt_env = Environment.describe_environment(
|
|
boot_from=settings.ADMIN_BOOT_DEVICE)
|
|
self._virt_env.define()
|
|
else:
|
|
try:
|
|
return Environment.get(name=self._config[
|
|
'template']['devops_settings']['env_name'])
|
|
except Exception:
|
|
self._virt_env = Environment.create_environment(
|
|
full_config=self._config)
|
|
self._virt_env.define()
|
|
return self._virt_env
|
|
|
|
def resume_environment(self):
|
|
self.d_env.resume()
|
|
admin = self.d_env.nodes().admin
|
|
|
|
self.ssh_manager.clean_all_connections()
|
|
|
|
try:
|
|
admin.await(self.d_env.admin_net, timeout=30, by_port=8000)
|
|
except Exception as e:
|
|
logger.warning("From first time admin isn't reverted: "
|
|
"{0}".format(e))
|
|
admin.destroy()
|
|
logger.info('Admin node was destroyed. Wait 10 sec.')
|
|
time.sleep(10)
|
|
|
|
admin.start()
|
|
logger.info('Admin node started second time.')
|
|
self.d_env.nodes().admin.await(self.d_env.admin_net)
|
|
self.set_admin_ssh_password()
|
|
self.admin_actions.wait_for_fuel_ready(timeout=600)
|
|
|
|
# set collector address in case of admin node destroy
|
|
if settings.FUEL_STATS_ENABLED:
|
|
self.nailgun_actions.set_collector_address(
|
|
settings.FUEL_STATS_HOST,
|
|
settings.FUEL_STATS_PORT,
|
|
settings.FUEL_STATS_SSL)
|
|
# Restart statsenderd in order to apply new collector address
|
|
self.nailgun_actions.force_fuel_stats_sending()
|
|
self.fuel_web.client.send_fuel_stats(enabled=True)
|
|
logger.info('Enabled sending of statistics to {0}:{1}'.format(
|
|
settings.FUEL_STATS_HOST, settings.FUEL_STATS_PORT
|
|
))
|
|
self.set_admin_ssh_password()
|
|
self.admin_actions.wait_for_fuel_ready()
|
|
|
|
def make_snapshot(self, snapshot_name, description="", is_make=False):
|
|
if settings.MAKE_SNAPSHOT or is_make:
|
|
self.d_env.suspend()
|
|
time.sleep(10)
|
|
|
|
self.d_env.snapshot(snapshot_name, force=True,
|
|
description=description)
|
|
revert_info(snapshot_name, self.get_admin_node_ip(), description)
|
|
|
|
if settings.FUEL_STATS_CHECK:
|
|
self.resume_environment()
|
|
|
|
def nailgun_nodes(self, devops_nodes):
|
|
return [self.fuel_web.get_nailgun_node_by_devops_node(node)
|
|
for node in devops_nodes]
|
|
|
|
def check_slaves_are_ready(self):
|
|
devops_nodes = [node for node in self.d_env.nodes().slaves
|
|
if node.driver.node_active(node)]
|
|
# Bug: 1455753
|
|
time.sleep(30)
|
|
|
|
self.fuel_web.wait_nodes_get_online_state(devops_nodes, timeout=60 * 6)
|
|
return True
|
|
|
|
def revert_snapshot(self, name, skip_timesync=False,
|
|
skip_slaves_check=False):
|
|
if not self.d_env.has_snapshot(name):
|
|
return False
|
|
|
|
logger.info('We have snapshot with such name: {:s}'.format(name))
|
|
|
|
logger.info("Reverting the snapshot '{0}' ....".format(name))
|
|
self.d_env.revert(name)
|
|
|
|
logger.info("Resuming the snapshot '{0}' ....".format(name))
|
|
self.resume_environment()
|
|
|
|
if not skip_timesync:
|
|
self.sync_time()
|
|
else:
|
|
self.sync_time(['admin'])
|
|
try:
|
|
with QuietLogger(upper_log_level=logging.CRITICAL):
|
|
# TODO(astudenov): add timeout_msg
|
|
wait_pass(
|
|
self.fuel_web.client.get_releases,
|
|
expected=(
|
|
exceptions.RetriableConnectionFailure,
|
|
exceptions.UnknownConnectionError),
|
|
timeout=300)
|
|
except exceptions.Unauthorized:
|
|
self.set_admin_keystone_password()
|
|
self.fuel_web.get_nailgun_version()
|
|
|
|
if not skip_slaves_check:
|
|
# TODO(astudenov): add timeout_msg
|
|
wait_pass(lambda: self.check_slaves_are_ready(), timeout=60 * 6)
|
|
return True
|
|
|
|
def set_admin_ssh_password(self):
|
|
new_login = settings.SSH_CREDENTIALS['login']
|
|
new_password = settings.SSH_CREDENTIALS['password']
|
|
try:
|
|
self.ssh_manager.execute_on_remote(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd='date'
|
|
)
|
|
logger.debug('Accessing admin node using SSH: SUCCESS')
|
|
except Exception:
|
|
logger.debug('Accessing admin node using SSH credentials:'
|
|
' FAIL, trying to change password from default')
|
|
self.ssh_manager.initialize(
|
|
admin_ip=self.ssh_manager.admin_ip,
|
|
login='root',
|
|
password='r00tme'
|
|
)
|
|
self.ssh_manager.execute_on_remote(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd='echo -e "{1}\\n{1}" | passwd {0}'.format(new_login,
|
|
new_password)
|
|
)
|
|
self.ssh_manager.initialize(
|
|
admin_ip=self.ssh_manager.admin_ip,
|
|
login=new_login,
|
|
password=new_password
|
|
)
|
|
self.ssh_manager.update_connection(
|
|
ip=self.ssh_manager.admin_ip,
|
|
login=new_login,
|
|
password=new_password
|
|
)
|
|
logger.debug("Admin node password has changed.")
|
|
logger.info("Admin node login name: '{0}' , password: '{1}'".
|
|
format(new_login, new_password))
|
|
|
|
def set_admin_keystone_password(self):
|
|
try:
|
|
self.fuel_web.client.get_releases()
|
|
# TODO(akostrikov) CENTOS7 except exceptions.Unauthorized:
|
|
except:
|
|
self.ssh_manager.execute_on_remote(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd='fuel user --newpass {0} --change-password'.format(
|
|
settings.KEYSTONE_CREDS['password'])
|
|
)
|
|
config_file = self.ssh_manager.execute_on_remote(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd='ls -1 $HOME/.config/fuel/fuel_client.yaml')['stdout_str']
|
|
|
|
with YamlEditor(config_file, ip=self.admin_node_ip) as editor:
|
|
editor.content["OS_USERNAME"] = \
|
|
settings.KEYSTONE_CREDS['username']
|
|
editor.content["OS_PASSWORD"] = \
|
|
settings.KEYSTONE_CREDS['password']
|
|
|
|
with YamlEditor(settings.FUEL_SETTINGS_YAML,
|
|
ip=self.admin_node_ip) as editor:
|
|
editor.content["FUEL_ACCESS"]['user'] = \
|
|
settings.KEYSTONE_CREDS['username']
|
|
editor.content["FUEL_ACCESS"]['password'] = \
|
|
settings.KEYSTONE_CREDS['password']
|
|
|
|
logger.info(
|
|
'New Fuel UI (keystone) username: "{0}", password: "{1}"'
|
|
.format(settings.KEYSTONE_CREDS['username'],
|
|
settings.KEYSTONE_CREDS['password']))
|
|
|
|
def setup_environment(self, custom=settings.CUSTOM_ENV,
|
|
build_images=settings.BUILD_IMAGES,
|
|
iso_connect_as=settings.ADMIN_BOOT_DEVICE,
|
|
security=settings.SECURITY_TEST):
|
|
# Create environment and start the Fuel master node
|
|
admin = self.d_env.nodes().admin
|
|
self.d_env.start([admin])
|
|
|
|
logger.info("Waiting for admin node to start up")
|
|
wait(lambda: admin.driver.node_active(admin), 60,
|
|
timeout_msg='Admin node startup timeout')
|
|
logger.info("Proceed with installation")
|
|
# update network parameters at boot screen
|
|
admin.send_keys(self.get_keys(admin, custom=custom,
|
|
build_images=build_images,
|
|
iso_connect_as=iso_connect_as))
|
|
if settings.SHOW_FUELMENU:
|
|
self.wait_for_fuelmenu()
|
|
else:
|
|
self.wait_for_provisioning()
|
|
|
|
self.set_admin_ssh_password()
|
|
|
|
self.wait_for_external_config()
|
|
if custom:
|
|
self.setup_customisation()
|
|
if security:
|
|
nessus_node = NessusActions(self.d_env)
|
|
nessus_node.add_nessus_node()
|
|
# wait while installation complete
|
|
|
|
self.admin_actions.modify_configs(self.d_env.router())
|
|
self.kill_wait_for_external_config()
|
|
self.wait_bootstrap()
|
|
self.admin_actions.wait_for_fuel_ready()
|
|
|
|
@logwrap
|
|
def enable_force_https(self, admin_node_ip):
|
|
cmd = """
|
|
echo -e '"SSL":\n "force_https": "true"' >> /etc/fuel/astute.yaml
|
|
"""
|
|
self.ssh_manager.execute_on_remote(admin_node_ip, cmd)
|
|
cmd = "find / -name \"nginx_services.pp\""
|
|
puppet_manifest = \
|
|
self.ssh_manager.execute_on_remote(
|
|
admin_node_ip, cmd)['stdout'][0].strip()
|
|
cmd = 'puppet apply {0}'.format(puppet_manifest)
|
|
self.ssh_manager.execute_on_remote(admin_node_ip, cmd)
|
|
cmd = """
|
|
systemctl status nginx.service |
|
|
awk 'match($0, /\s+Active:.*\((\w+)\)/, a) {print a[1]}'
|
|
"""
|
|
wait(lambda: (
|
|
self.ssh_manager.execute_on_remote(
|
|
admin_node_ip, cmd)['stdout'][0] != 'dead'), interval=10,
|
|
timeout=30,
|
|
timeout_msg='Nginx service is dead after trying to enable '
|
|
'it with the command: {}'.format(cmd))
|
|
|
|
# pylint: disable=no-self-use
|
|
@update_rpm_packages
|
|
@upload_manifests
|
|
def setup_customisation(self):
|
|
logger.info('Installing custom packages/manifests '
|
|
'before master node bootstrap...')
|
|
# pylint: enable=no-self-use
|
|
|
|
@logwrap
|
|
def wait_for_provisioning(self,
|
|
timeout=settings.WAIT_FOR_PROVISIONING_TIMEOUT):
|
|
# TODO(astudenov): add timeout_msg
|
|
wait_pass(lambda: tcp_ping_(
|
|
self.d_env.nodes(
|
|
).admin.get_ip_address_by_network_name
|
|
(self.d_env.admin_net), 22), timeout=timeout)
|
|
|
|
@logwrap
|
|
def wait_for_fuelmenu(self,
|
|
timeout=settings.WAIT_FOR_PROVISIONING_TIMEOUT):
|
|
|
|
def check_ssh_connection():
|
|
"""Try to close fuelmenu and check ssh connection"""
|
|
try:
|
|
tcp_ping_(
|
|
self.d_env.nodes(
|
|
).admin.get_ip_address_by_network_name
|
|
(self.d_env.admin_net), 22)
|
|
except Exception:
|
|
# send F8 trying to exit fuelmenu
|
|
self.d_env.nodes().admin.send_keys("<F8>\n")
|
|
return False
|
|
return True
|
|
|
|
wait(check_ssh_connection, interval=30, timeout=timeout,
|
|
timeout_msg="Fuelmenu hasn't appeared during allocated timeout")
|
|
|
|
@logwrap
|
|
def wait_for_external_config(self, timeout=120):
|
|
|
|
wait(lambda: self.ssh_manager.exists_on_remote(
|
|
self.ssh_manager.admin_ip,
|
|
'/var/lock/wait_for_external_config'),
|
|
timeout=600,
|
|
timeout_msg='wait_for_external_config lock file timeout '
|
|
'while bootstrapping the Fuel master node')
|
|
|
|
check_cmd = 'pkill -0 -f wait_for_external_config'
|
|
|
|
wait(
|
|
lambda: self.ssh_manager.execute(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd=check_cmd)['exit_code'] == 0,
|
|
timeout=timeout,
|
|
timeout_msg='wait_for_external_config process timeout '
|
|
'while bootstrapping the Fuel master node')
|
|
|
|
@logwrap
|
|
def kill_wait_for_external_config(self):
|
|
kill_cmd = 'pkill -f "^wait_for_external_config"'
|
|
check_cmd = 'pkill -0 -f "^wait_for_external_config"; [[ $? -eq 1 ]]'
|
|
self.ssh_manager.execute_on_remote(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd=kill_cmd
|
|
)
|
|
self.ssh_manager.execute_on_remote(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd=check_cmd
|
|
)
|
|
|
|
def wait_bootstrap(self):
|
|
logger.info("Waiting while bootstrapping is in progress")
|
|
log_path = "/var/log/puppet/bootstrap_admin_node.log"
|
|
logger.info("Running bootstrap (timeout: {0})".format(
|
|
float(settings.ADMIN_NODE_BOOTSTRAP_TIMEOUT)))
|
|
with TimeStat("admin_node_bootsrap_time", is_uniq=True):
|
|
wait(
|
|
lambda: self.ssh_manager.execute(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd="grep 'Fuel node deployment' '{:s}'".format(log_path)
|
|
)['exit_code'] == 0,
|
|
timeout=(float(settings.ADMIN_NODE_BOOTSTRAP_TIMEOUT)),
|
|
timeout_msg='Fuel master node bootstrap timeout, '
|
|
'please check the log {}'.format(log_path)
|
|
)
|
|
result = self.ssh_manager.execute(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd="grep 'Fuel node deployment "
|
|
"complete' '{:s}'".format(log_path))['exit_code']
|
|
if result != 0:
|
|
raise Exception('Fuel node deployment failed.')
|
|
self.bootstrap_image_check()
|
|
|
|
def dhcrelay_check(self):
|
|
# CentOS 7 is pretty stable with admin iface.
|
|
# TODO(akostrikov) refactor it.
|
|
iface = iface_alias('eth0')
|
|
command = "dhcpcheck discover " \
|
|
"--ifaces {iface} " \
|
|
"--repeat 3 " \
|
|
"--timeout 10".format(iface=iface)
|
|
|
|
out = self.ssh_manager.execute(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd=command
|
|
)['stdout']
|
|
|
|
assert_true(self.get_admin_node_ip() in "".join(out),
|
|
"dhcpcheck doesn't discover master ip")
|
|
|
|
def bootstrap_image_check(self):
|
|
fuel_settings = self.admin_actions.get_fuel_settings()
|
|
if fuel_settings['BOOTSTRAP']['flavor'].lower() != 'ubuntu':
|
|
logger.warning('Default image for bootstrap '
|
|
'is not based on Ubuntu!')
|
|
return
|
|
|
|
bootstrap_images = self.ssh_manager.execute_on_remote(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd='fuel-bootstrap --quiet list'
|
|
)['stdout']
|
|
assert_true(any('active' in line for line in bootstrap_images),
|
|
'Ubuntu bootstrap image wasn\'t built and activated! '
|
|
'See logs in /var/log/fuel-bootstrap-image-build.log '
|
|
'for details.')
|
|
|
|
def admin_install_pkg(self, pkg_name):
|
|
"""Install a package <pkg_name> on the admin node"""
|
|
remote_status = self.ssh_manager.execute(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd="rpm -q {0}'".format(pkg_name)
|
|
)
|
|
if remote_status['exit_code'] == 0:
|
|
logger.info("Package '{0}' already installed.".format(pkg_name))
|
|
else:
|
|
logger.info("Installing package '{0}' ...".format(pkg_name))
|
|
remote_status = self.ssh_manager.execute(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd="yum -y install {0}".format(pkg_name)
|
|
)
|
|
logger.info("Installation of the package '{0}' has been"
|
|
" completed with exit code {1}"
|
|
.format(pkg_name, remote_status['exit_code']))
|
|
return remote_status['exit_code']
|
|
|
|
def admin_run_service(self, service_name):
|
|
"""Start a service <service_name> on the admin node"""
|
|
|
|
self.ssh_manager.execute(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd="service {0} start".format(service_name)
|
|
)
|
|
remote_status = self.ssh_manager.execute(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd="service {0} status".format(service_name)
|
|
)
|
|
if any('running...' in status for status in remote_status['stdout']):
|
|
logger.info("Service '{0}' is running".format(service_name))
|
|
else:
|
|
logger.info("Service '{0}' failed to start"
|
|
" with exit code {1} :\n{2}"
|
|
.format(service_name,
|
|
remote_status['exit_code'],
|
|
remote_status['stdout']))
|
|
|
|
def admin_install_updates(self):
|
|
"""Update packages using yum and install updates via
|
|
update-master-node.sh tool"""
|
|
logger.info('Searching for python-cudet package')
|
|
|
|
search_command = 'yum search python-cudet'
|
|
|
|
search_result = self.ssh_manager.check_call(
|
|
ip=self.ssh_manager.admin_ip,
|
|
command=search_command)
|
|
|
|
assert_true(
|
|
"Warning: No matches found for: " not in search_result.stderr_str,
|
|
"python-cudet wasn't found")
|
|
|
|
install_command = 'yum install -y python-cudet'
|
|
|
|
self.ssh_manager.check_call(
|
|
ip=self.ssh_manager.admin_ip,
|
|
command=install_command,
|
|
verbose=True)
|
|
|
|
logger.info('prepare Fuel node for updating')
|
|
prepare_command = 'update-prepare prepare master'
|
|
|
|
self.ssh_manager.check_call(
|
|
ip=self.ssh_manager.admin_ip,
|
|
command=prepare_command,
|
|
verbose=True)
|
|
|
|
logger.info('update Fuel node')
|
|
update_command = 'update-prepare update master'
|
|
|
|
self.ssh_manager.check_call(
|
|
ip=self.ssh_manager.admin_ip,
|
|
command=update_command,
|
|
verbose=True)
|
|
|
|
logger.info('Update successful')
|
|
|
|
# Modifies a resolv.conf on the Fuel master node and returns
|
|
# its original content.
|
|
# * adds 'nameservers' at start of resolv.conf if merge=True
|
|
# * replaces resolv.conf with 'nameservers' if merge=False
|
|
def modify_resolv_conf(self, nameservers=None, merge=True):
|
|
if nameservers is None:
|
|
nameservers = []
|
|
|
|
resolv_conf = self.ssh_manager.execute(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd='cat /etc/resolv.conf'
|
|
)
|
|
assert_equal(0, resolv_conf['exit_code'],
|
|
'Executing "{0}" on the admin node has failed with: {1}'
|
|
.format('cat /etc/resolv.conf', resolv_conf['stderr']))
|
|
if merge:
|
|
nameservers.extend(resolv_conf['stdout'])
|
|
resolv_keys = ['search', 'domain', 'nameserver']
|
|
resolv_new = "".join('{0}\n'.format(ns) for ns in nameservers
|
|
if any(x in ns for x in resolv_keys))
|
|
logger.debug('echo "{0}" > /etc/resolv.conf'.format(resolv_new))
|
|
echo_cmd = 'echo "{0}" > /etc/resolv.conf'.format(resolv_new)
|
|
echo_result = self.ssh_manager.execute(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd=echo_cmd
|
|
)
|
|
assert_equal(0, echo_result['exit_code'],
|
|
'Executing "{0}" on the admin node has failed with: {1}'
|
|
.format(echo_cmd, echo_result['stderr']))
|
|
return resolv_conf['stdout']
|
|
|
|
@logwrap
|
|
def describe_other_admin_interfaces(self, admin):
|
|
admin_networks = [iface.network.name for iface in admin.interfaces]
|
|
iface_name = None
|
|
for i, network_name in enumerate(admin_networks):
|
|
if 'admin' in network_name and 'admin' != network_name:
|
|
# This will be replaced with actual interface labels
|
|
# form fuel-devops
|
|
iface_name = 'enp0s' + str(i + 3)
|
|
logger.info("Describe Fuel admin node interface {0} for "
|
|
"network {1}".format(iface_name, network_name))
|
|
self.describe_admin_interface(iface_name, network_name)
|
|
|
|
if iface_name:
|
|
return self.ssh_manager.execute(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd="cobbler sync")
|
|
|
|
@logwrap
|
|
def describe_admin_interface(self, admin_if, network_name):
|
|
admin_net_object = self.d_env.get_network(name=network_name)
|
|
admin_network = admin_net_object.ip.network
|
|
admin_netmask = admin_net_object.ip.netmask
|
|
admin_ip = str(self.d_env.nodes(
|
|
).admin.get_ip_address_by_network_name(network_name))
|
|
logger.info(('Parameters for admin interface configuration: '
|
|
'Network - {0}, Netmask - {1}, Interface - {2}, '
|
|
'IP Address - {3}').format(admin_network,
|
|
admin_netmask,
|
|
admin_if,
|
|
admin_ip))
|
|
add_admin_ip = ('DEVICE={0}\\n'
|
|
'ONBOOT=yes\\n'
|
|
'NM_CONTROLLED=no\\n'
|
|
'USERCTL=no\\n'
|
|
'PEERDNS=no\\n'
|
|
'BOOTPROTO=static\\n'
|
|
'IPADDR={1}\\n'
|
|
'NETMASK={2}\\n').format(admin_if,
|
|
admin_ip,
|
|
admin_netmask)
|
|
cmd = ('echo -e "{0}" > /etc/sysconfig/network-scripts/ifcfg-{1};'
|
|
'ifup {1}; ip -o -4 a s {1} | grep -w {2}').format(
|
|
add_admin_ip, admin_if, admin_ip)
|
|
logger.debug('Trying to assign {0} IP to the {1} on master node...'.
|
|
format(admin_ip, admin_if))
|
|
|
|
result = self.ssh_manager.execute(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd=cmd
|
|
)
|
|
assert_equal(result['exit_code'], 0, ('Failed to assign second admin '
|
|
'IP address on master node: {0}').format(result))
|
|
logger.debug('Done: {0}'.format(result['stdout']))
|
|
|
|
# TODO for ssh manager
|
|
multiple_networks_hacks.configure_second_admin_dhcp(
|
|
self.ssh_manager.admin_ip,
|
|
admin_if
|
|
)
|
|
multiple_networks_hacks.configure_second_admin_firewall(
|
|
self.ssh_manager.admin_ip,
|
|
admin_network,
|
|
admin_netmask,
|
|
admin_if,
|
|
self.get_admin_node_ip()
|
|
)
|
|
|
|
@logwrap
|
|
def get_masternode_uuid(self):
|
|
return self.postgres_actions.run_query(
|
|
db='nailgun',
|
|
query="select master_node_uid from master_node_settings limit 1;")
|