Added interface updating after adding compute/controller and before provisioning for test on creation and deployment environment using Fuel CLI. Also running ha and sanity OSTF tests was replaced with assert_os_services_ready and assert_ha_services_ready calls as we should wait for services availability before running OSTF tests. And these functions use running ha and sanity OSTF tests so that we don't have to call it again. Code was partly back-ported from master branch. Change-Id: I7fc831a37acff746578eea79bb7b2445c9d78410 Closes-Bug: #1589942
296 lines
12 KiB
Python
296 lines
12 KiB
Python
# Copyright 2015 Mirantis, Inc.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
from proboscis import test
|
|
from proboscis.asserts import assert_equal
|
|
from proboscis.asserts import assert_false
|
|
from proboscis.asserts import assert_true
|
|
|
|
from devops.error import TimeoutError
|
|
from devops.helpers.helpers import wait
|
|
from fuelweb_test.helpers.checkers import check_cluster_presence
|
|
from fuelweb_test.helpers.checkers import check_cobbler_node_exists
|
|
from fuelweb_test.helpers.decorators import log_snapshot_after_test
|
|
from fuelweb_test.helpers.utils import run_on_remote
|
|
from fuelweb_test.settings import DEPLOYMENT_MODE
|
|
from fuelweb_test.settings import NEUTRON_SEGMENT_TYPE
|
|
from fuelweb_test.settings import OPENSTACK_RELEASE
|
|
from fuelweb_test.tests.base_test_case import SetupEnvironment
|
|
from fuelweb_test.tests.base_test_case import TestBasic
|
|
from fuelweb_test.tests import test_cli_base
|
|
from fuelweb_test import logger
|
|
|
|
|
|
@test(groups=["command_line_minimal"])
|
|
class CommandLineMinimal(TestBasic):
|
|
"""CommandLineMinimal.""" # TODO documentation
|
|
|
|
@test(depends_on=[SetupEnvironment.setup_with_custom_manifests],
|
|
groups=["hiera_deploy"])
|
|
@log_snapshot_after_test
|
|
def hiera_deploy(self):
|
|
"""Deploy cluster with controller node only
|
|
|
|
Scenario:
|
|
1. Start installation of master
|
|
2. Enter "fuelmenu"
|
|
3. Upload custom manifests
|
|
4. Kill "fuelmenu" pid
|
|
5. Deploy hiera manifest
|
|
|
|
Duration 20m
|
|
"""
|
|
self.env.revert_snapshot("empty_custom_manifests")
|
|
|
|
self.env.bootstrap_nodes(
|
|
self.env.d_env.nodes().slaves[:1])
|
|
|
|
cluster_id = self.fuel_web.create_cluster(
|
|
name=self.__class__.__name__,
|
|
mode=DEPLOYMENT_MODE
|
|
)
|
|
self.fuel_web.update_nodes(
|
|
cluster_id,
|
|
{'slave-01': ['controller']}
|
|
)
|
|
with self.env.d_env.get_admin_remote() as remote:
|
|
node_id = self.fuel_web.get_nailgun_node_by_devops_node(
|
|
self.env.d_env.nodes().slaves[0])['id']
|
|
remote.execute('fuel node --node {0} --provision --env {1}'.format
|
|
(node_id, cluster_id))
|
|
self.fuel_web.provisioning_cluster_wait(cluster_id)
|
|
remote.execute('fuel node --node {0} --end hiera --env {1}'.format
|
|
(node_id, cluster_id))
|
|
try:
|
|
wait(lambda: int(
|
|
remote.execute(
|
|
'fuel task | grep deployment | awk \'{print $9}\'')
|
|
['stdout'][0].rstrip()) == 100, timeout=120)
|
|
except TimeoutError:
|
|
raise TimeoutError("hiera manifest was not applied")
|
|
role = remote.execute('ssh -q node-{0} "hiera role"'
|
|
.format(node_id))['stdout'][0].rstrip()
|
|
assert_equal(role, 'primary-controller', "node with deployed hiera "
|
|
"was not found")
|
|
|
|
|
|
@test(groups=["command_line"])
|
|
class CommandLineTest(test_cli_base.CommandLine):
|
|
"""CommandLine.""" # TODO documentation
|
|
|
|
@test(depends_on=[SetupEnvironment.prepare_slaves_3],
|
|
groups=["cli_selected_nodes_deploy"])
|
|
@log_snapshot_after_test
|
|
def cli_selected_nodes_deploy(self):
|
|
"""Create and deploy environment using Fuel CLI
|
|
|
|
Scenario:
|
|
1. Revert snapshot "ready_with_3_slaves"
|
|
2. Create a cluster using Fuel CLI
|
|
3. Provision a controller node using Fuel CLI
|
|
4. Provision two compute+cinder nodes using Fuel CLI
|
|
5. Deploy the controller node using Fuel CLI
|
|
6. Deploy the compute+cinder nodes using Fuel CLI
|
|
7. Run OSTF
|
|
8. Make snapshot "cli_selected_nodes_deploy"
|
|
|
|
Duration 50m
|
|
"""
|
|
self.env.revert_snapshot("ready_with_3_slaves")
|
|
node_ids = [self.fuel_web.get_nailgun_node_by_devops_node(
|
|
self.env.d_env.nodes().slaves[slave_id])['id']
|
|
for slave_id in range(3)]
|
|
release_id = self.fuel_web.get_releases_list_for_os(
|
|
release_name=OPENSTACK_RELEASE)[0]
|
|
|
|
with self.env.d_env.get_admin_remote() as remote:
|
|
|
|
# Create an environment
|
|
if NEUTRON_SEGMENT_TYPE:
|
|
nst = '--nst={0}'.format(NEUTRON_SEGMENT_TYPE)
|
|
else:
|
|
nst = ''
|
|
cmd = ('fuel env create --name={0} --release={1} '
|
|
'{2} --json'.format(self.__class__.__name__,
|
|
release_id, nst))
|
|
env_result = run_on_remote(remote, cmd, jsonify=True)
|
|
cluster_id = env_result['id']
|
|
|
|
# Update network parameters
|
|
self.update_cli_network_configuration(cluster_id, remote)
|
|
|
|
# Update SSL configuration
|
|
self.update_ssl_configuration(cluster_id, remote)
|
|
|
|
# Add and provision a controller node
|
|
logger.info("Add to the cluster and start provisioning "
|
|
"a controller node [{0}]".format(node_ids[0]))
|
|
cmd = ('fuel --env-id={0} node set --node {1} --role=controller'
|
|
.format(cluster_id, node_ids[0]))
|
|
remote.execute(cmd)
|
|
self.update_node_interfaces(node_ids[0])
|
|
cmd = ('fuel --env-id={0} node --provision --node={1} --json'
|
|
.format(cluster_id, node_ids[0]))
|
|
task = run_on_remote(remote, cmd, jsonify=True)
|
|
self.assert_cli_task_success(task, remote, timeout=30 * 60)
|
|
|
|
# Add and provision 2 compute+cinder
|
|
logger.info("Add to the cluster and start provisioning two "
|
|
"compute+cinder nodes [{0},{1}]".format(node_ids[1],
|
|
node_ids[2]))
|
|
cmd = ('fuel --env-id={0} node set --node {1},{2} '
|
|
'--role=compute,cinder'.format(cluster_id,
|
|
node_ids[1], node_ids[2]))
|
|
remote.execute(cmd)
|
|
for node_id in (node_ids[1], node_ids[2]):
|
|
self.update_node_interfaces(node_id)
|
|
cmd = ('fuel --env-id={0} node --provision --node={1},{2} --json'
|
|
.format(cluster_id, node_ids[1], node_ids[2]))
|
|
task = run_on_remote(remote, cmd, jsonify=True)
|
|
self.assert_cli_task_success(task, remote, timeout=10 * 60)
|
|
|
|
# Deploy the controller node
|
|
cmd = ('fuel --env-id={0} node --deploy --node {1} --json'
|
|
.format(cluster_id, node_ids[0]))
|
|
task = run_on_remote(remote, cmd, jsonify=True)
|
|
self.assert_cli_task_success(task, remote, timeout=60 * 60)
|
|
|
|
# Deploy the compute nodes
|
|
cmd = ('fuel --env-id={0} node --deploy --node {1},{2} --json'
|
|
.format(cluster_id, node_ids[1], node_ids[2]))
|
|
task = run_on_remote(remote, cmd, jsonify=True)
|
|
self.assert_cli_task_success(task, remote, timeout=30 * 60)
|
|
|
|
self.fuel_web.assert_os_services_ready(cluster_id)
|
|
self.fuel_web.assert_ha_services_ready(cluster_id)
|
|
self.fuel_web.run_ostf(
|
|
cluster_id=cluster_id,
|
|
test_sets=['smoke'])
|
|
|
|
self.env.make_snapshot("cli_selected_nodes_deploy", is_make=True)
|
|
|
|
@test(depends_on_groups=['cli_selected_nodes_deploy'],
|
|
groups=["cli_node_deletion_check"])
|
|
@log_snapshot_after_test
|
|
def cli_node_deletion_check(self):
|
|
"""Destroy node and remove it from Nailgun using Fuel CLI
|
|
|
|
Scenario:
|
|
1. Revert snapshot 'cli_selected_nodes_deploy'
|
|
2. Check 'slave-03' is present
|
|
3. Destroy 'slave-03'
|
|
4. Wait until 'slave-03' become offline
|
|
5. Delete offline 'slave-03' from db
|
|
6. Check presence of 'slave-03'
|
|
|
|
Duration 30m
|
|
|
|
"""
|
|
self.env.revert_snapshot("cli_selected_nodes_deploy")
|
|
|
|
with self.env.d_env.get_admin_remote() as remote:
|
|
node_id = self.fuel_web.get_nailgun_node_by_devops_node(
|
|
self.env.d_env.nodes().slaves[2])['id']
|
|
|
|
assert_true(check_cobbler_node_exists(remote, node_id),
|
|
"node-{0} is not found".format(node_id))
|
|
self.env.d_env.nodes().slaves[2].destroy()
|
|
try:
|
|
wait(
|
|
lambda: not self.fuel_web.get_nailgun_node_by_devops_node(
|
|
self.env.d_env.nodes().
|
|
slaves[2])['online'], timeout=60 * 6)
|
|
except TimeoutError:
|
|
raise
|
|
with self.env.d_env.get_admin_remote() as remote:
|
|
res = remote.execute('fuel node --node-id {0} --delete-from-db'
|
|
.format(node_id))
|
|
assert_true(
|
|
res['exit_code'] == 0,
|
|
"Offline node-{0} was not"
|
|
"deleted from database".format(node_id))
|
|
|
|
with self.env.d_env.get_admin_remote() as remote:
|
|
try:
|
|
wait(
|
|
lambda: not remote.execute(
|
|
"fuel node | awk '{{print $1}}' | grep -w '{0}'".
|
|
format(node_id))['exit_code'] == 0, timeout=60 * 4)
|
|
except TimeoutError:
|
|
raise TimeoutError(
|
|
"After deletion node-{0} is found in fuel list".
|
|
format(node_id))
|
|
|
|
with self.env.d_env.get_admin_remote() as remote:
|
|
is_cobbler_node_exists = check_cobbler_node_exists(remote, node_id)
|
|
|
|
assert_false(is_cobbler_node_exists,
|
|
"After deletion node-{0} is found in cobbler list".
|
|
format(node_id))
|
|
|
|
with self.env.d_env.get_admin_remote() as remote:
|
|
cluster_id = ''.join(remote.execute(
|
|
"fuel env | tail -n 1 | awk {'print $1'}")['stdout']).rstrip()
|
|
|
|
self.fuel_web.verify_network(cluster_id)
|
|
|
|
self.fuel_web.run_ostf(
|
|
cluster_id=cluster_id, test_sets=['ha', 'smoke', 'sanity'],
|
|
should_fail=1)
|
|
|
|
@test(depends_on_groups=['cli_selected_nodes_deploy'],
|
|
groups=["cli_cluster_deletion"])
|
|
@log_snapshot_after_test
|
|
def cli_cluster_deletion(self):
|
|
"""Delete a cluster using Fuel CLI
|
|
|
|
Scenario:
|
|
1. Revert snapshot 'cli_selected_nodes_deploy'
|
|
2. Delete cluster via cli
|
|
3. Check cluster absence in the list
|
|
|
|
Duration 25m
|
|
|
|
"""
|
|
self.env.revert_snapshot("cli_selected_nodes_deploy")
|
|
|
|
cluster_id = self.fuel_web.get_last_created_cluster()
|
|
|
|
nodes = self.fuel_web.client.list_cluster_nodes(cluster_id)
|
|
online_nodes = [node for node in nodes if node['online']]
|
|
if nodes != online_nodes:
|
|
logger.error(
|
|
'Some slaves do not become online after revert!!'
|
|
' Expected {0} Actual {1}'.format(nodes, online_nodes))
|
|
|
|
self.ssh_manager.execute_on_remote(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd='fuel --env {0} env delete'.format(cluster_id)
|
|
)
|
|
|
|
try:
|
|
wait(lambda:
|
|
self.ssh_manager.execute(
|
|
ip=self.ssh_manager.admin_ip,
|
|
cmd="fuel env | awk '{print $1}' | tail -n 1 | "
|
|
"grep '^.$'"
|
|
)['exit_code'] == 1, timeout=60 * 10)
|
|
except TimeoutError:
|
|
raise TimeoutError(
|
|
"cluster {0} was not deleted".format(cluster_id))
|
|
|
|
assert_false(
|
|
check_cluster_presence(cluster_id, self.env.postgres_actions),
|
|
"cluster {0} is found".format(cluster_id))
|