fuel-qa/fuelweb_test/tests/test_cli.py
Sergey Kulanov e09c2ccad2 [command_line] Ensure operation cluster deleted
To delete operational cluster with fuel-client one must use
--force flag:
     fuel --env ID env delete --force

Change-Id: I76887c24b0f5d8ea8508a59bc726207eb6ba138a
Closes-bug: #1562028
2016-03-25 17:02:11 +02:00

355 lines
16 KiB
Python

# Copyright 2015 Mirantis, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from devops.error import TimeoutError
from devops.helpers.helpers import wait
from proboscis import test
from proboscis.asserts import assert_equal
from proboscis.asserts import assert_false
from proboscis.asserts import assert_true
from fuelweb_test.helpers.checkers import check_cluster_presence
from fuelweb_test.helpers.checkers import check_cobbler_node_exists
from fuelweb_test.helpers.decorators import log_snapshot_after_test
from fuelweb_test.helpers.utils import generate_floating_ranges
from fuelweb_test.settings import DEPLOYMENT_MODE
from fuelweb_test.settings import SSL_CN
from fuelweb_test.settings import PATH_TO_PEM
from fuelweb_test.settings import NEUTRON_SEGMENT_TYPE
from fuelweb_test.settings import OPENSTACK_RELEASE
from fuelweb_test.tests.base_test_case import SetupEnvironment
from fuelweb_test.tests.base_test_case import TestBasic
from fuelweb_test.tests import test_cli_base
from fuelweb_test import logger
@test(groups=["command_line_minimal"])
class CommandLineMinimal(TestBasic):
"""CommandLineMinimal.""" # TODO documentation
@test(depends_on=[SetupEnvironment.setup_with_custom_manifests],
groups=["hiera_deploy"])
@log_snapshot_after_test
def hiera_deploy(self):
"""Deploy cluster with controller node only
Scenario:
1. Start installation of master
2. Enter "fuelmenu"
3. Upload custom manifests
4. Kill "fuelmenu" pid
5. Deploy hiera manifest
Duration 20m
"""
self.env.revert_snapshot("empty_custom_manifests")
self.env.bootstrap_nodes(
self.env.d_env.nodes().slaves[:1])
cluster_id = self.fuel_web.create_cluster(
name=self.__class__.__name__,
mode=DEPLOYMENT_MODE
)
self.fuel_web.update_nodes(
cluster_id,
{'slave-01': ['controller']}
)
with self.env.d_env.get_admin_remote() as remote:
node_id = self.fuel_web.get_nailgun_node_by_devops_node(
self.env.d_env.nodes().slaves[0])['id']
remote.execute('fuel node --node {0} --provision --env {1}'.format
(node_id, cluster_id))
self.fuel_web.provisioning_cluster_wait(cluster_id)
remote.execute('fuel node --node {0} --end hiera --env {1}'.format
(node_id, cluster_id))
try:
wait(lambda: int(
remote.execute(
'fuel task | grep deployment | awk \'{print $9}\'')
['stdout'][0].rstrip()) == 100, timeout=120)
except TimeoutError:
raise TimeoutError("hiera manifest was not applied")
role = remote.execute('ssh -q node-{0} "hiera role"'
.format(node_id))['stdout'][0].rstrip()
assert_equal(role, 'primary-controller', "node with deployed hiera "
"was not found")
@test(groups=["command_line"])
class CommandLineTest(test_cli_base.CommandLine):
"""CommandLine.""" # TODO documentation
@test(depends_on=[SetupEnvironment.prepare_slaves_3],
groups=["cli_selected_nodes_deploy"])
@log_snapshot_after_test
def cli_selected_nodes_deploy(self):
"""Create and deploy environment using Fuel CLI
Scenario:
1. Revert snapshot "ready_with_3_slaves"
2. Create a cluster using Fuel CLI
3. Add floating ranges for public network
4. Provision a controller node using Fuel CLI
5. Provision two compute+cinder nodes using Fuel CLI
6. Deploy the controller node using Fuel CLI
7. Deploy the compute+cinder nodes using Fuel CLI
8. Compare floating ranges
9. Check that all services work by 'https'
10. Check that all services have domain name
11. Check CNs for user owned certificate
12. Check keypairs for user owned ceritificate
13. Run OSTF
14. Make snapshot "cli_selected_nodes_deploy"
Duration 50m
"""
self.env.revert_snapshot("ready_with_3_slaves")
node_ids = [self.fuel_web.get_nailgun_node_by_devops_node(
self.env.d_env.nodes().slaves[
slave_id])['id']for slave_id in range(3)]
release_id = self.fuel_web.get_releases_list_for_os(
release_name=OPENSTACK_RELEASE)[0]
admin_ip = self.ssh_manager.admin_ip
# Create an environment
if NEUTRON_SEGMENT_TYPE:
nst = '--nst={0}'.format(NEUTRON_SEGMENT_TYPE)
else:
nst = ''
cmd = ('fuel env create --name={0} --release={1} {2} --json'.format(
self.__class__.__name__, release_id, nst))
env_result =\
self.ssh_manager.execute_on_remote(admin_ip, cmd,
jsonify=True)['stdout_json']
cluster_id = env_result['id']
# Update network parameters
self.update_cli_network_configuration(cluster_id)
# Change floating ranges
current_floating_range = self.get_floating_ranges(cluster_id)
logger.info("Current floating ranges: {0}".format(
current_floating_range))
first_floating_address = current_floating_range[0][0]
logger.info("First floating address: {0}".format(
first_floating_address))
last_floating_address = current_floating_range[0][1]
logger.info("Last floating address: {0}".format(last_floating_address))
new_floating_range = generate_floating_ranges(first_floating_address,
last_floating_address,
10)
logger.info("New floating range: {0}".format(new_floating_range))
self.change_floating_ranges(cluster_id, new_floating_range)
# Update SSL configuration
self.update_ssl_configuration(cluster_id)
# Add and provision a controller node
logger.info("Add to the cluster \
and start provisioning a controller node [{0}]".format(node_ids[0]))
cmd = ('fuel --env-id={0} node set --node {1}\
--role=controller'.format(cluster_id, node_ids[0]))
self.ssh_manager.execute_on_remote(admin_ip, cmd)
cmd = ('fuel --env-id={0} node --provision --node={1} --json'.format(
cluster_id, node_ids[0]))
task = self.ssh_manager.execute_on_remote(admin_ip,
cmd,
jsonify=True)['stdout_json']
self.assert_cli_task_success(task, timeout=30 * 60)
# Add and provision 2 compute+cinder
logger.info("Add to the cluster and start provisioning two "
"compute+cinder nodes [{0},{1}]".format(node_ids[1],
node_ids[2]))
cmd = ('fuel --env-id={0} node set --node {1},{2} \
--role=compute,cinder'.format(cluster_id, node_ids[1], node_ids[2]))
self.ssh_manager.execute_on_remote(admin_ip, cmd)
cmd = ('fuel --env-id={0} node --provision \
--node={1},{2} --json'.format(cluster_id, node_ids[1], node_ids[2]))
task = self.ssh_manager.execute_on_remote(admin_ip,
cmd,
jsonify=True)['stdout_json']
self.assert_cli_task_success(task, timeout=10 * 60)
# Deploy the controller node
cmd = ('fuel --env-id={0} node --deploy --node {1} --json'.format(
cluster_id, node_ids[0]))
task = self.ssh_manager.execute_on_remote(admin_ip,
cmd,
jsonify=True)['stdout_json']
self.assert_cli_task_success(task, timeout=60 * 60)
# Deploy the compute nodes
cmd = ('fuel --env-id={0} node --deploy --node {1},{2} --json'.format(
cluster_id, node_ids[1], node_ids[2]))
task = self.ssh_manager.execute_on_remote(admin_ip,
cmd,
jsonify=True)['stdout_json']
self.assert_cli_task_success(task, timeout=30 * 60)
# Verify networks
self.fuel_web.verify_network(cluster_id)
controller_nodes = self.fuel_web.get_nailgun_cluster_nodes_by_roles(
cluster_id, ['controller'])
# Get controller ip address
controller_node = controller_nodes[0]['ip']
# Get endpoint list
endpoint_list = self.get_endpoints(controller_node)
logger.info(endpoint_list)
# Check protocol and domain names for endpoints
for endpoint in endpoint_list:
logger.debug(("Endpoint {0} use protocol {1}\
and have domain name {2}".format(endpoint['service_name'],
endpoint['protocol'],
endpoint['domain'])))
assert_equal(endpoint['protocol'], "https",
message=("Endpoint {0} don't use https.".format(
endpoint['service_name'])))
assert_equal(endpoint['domain'], SSL_CN, message=(
"{0} domain name not equal {1}.".format(
endpoint['service_name'], SSL_CN)))
# Check SSL CNs
current_ssl_cn = self.get_current_ssl_cn(controller_node)
logger.info(("CN before cluster deploy {0} \
and after deploy {1}".format(SSL_CN, current_ssl_cn)))
assert_equal(SSL_CN, current_ssl_cn, message="SSL CNs are not equal")
# Check SSL certificate keypairs
with open(PATH_TO_PEM) as pem_file:
old_ssl_keypair = pem_file.read().strip()
current_ssl_keypair = self.get_current_ssl_keypair(controller_node)
logger.info(("SSL keypair before cluster deploy {0} \
and after deploy {1}".format(old_ssl_keypair,
current_ssl_keypair)
))
assert_equal(old_ssl_keypair, current_ssl_keypair,
message="SSL keypairs are not equal")
# Check floating ranges are equal after cluster deploy
actual_floating_ranges = self.hiera_floating_ranges(controller_node)
logger.info("Current floating ranges: {0}".format(
actual_floating_ranges))
assert_equal(actual_floating_ranges, new_floating_range,
message="Floating ranges are not equal")
# Run OSTF
self.fuel_web.run_ostf(cluster_id=cluster_id,
test_sets=['ha', 'smoke', 'sanity'])
self.env.make_snapshot("cli_selected_nodes_deploy", is_make=True)
@test(depends_on_groups=['cli_selected_nodes_deploy'],
groups=["cli_node_deletion_check"])
@log_snapshot_after_test
def cli_node_deletion_check(self):
"""Destroy node and remove it from Nailgun using Fuel CLI
Scenario:
1. Revert snapshot 'cli_selected_nodes_deploy'
2. Check 'slave-03' is present
3. Destroy 'slave-03'
4. Wait until 'slave-03' become offline
5. Delete offline 'slave-03' from db
6. Check presence of 'slave-03'
Duration 30m
"""
self.env.revert_snapshot("cli_selected_nodes_deploy")
node_id = self.fuel_web.get_nailgun_node_by_devops_node(
self.env.d_env.nodes().slaves[2])['id']
assert_true(check_cobbler_node_exists(self.ssh_manager.admin_ip,
node_id),
"node-{0} is not found".format(node_id))
self.env.d_env.nodes().slaves[2].destroy()
try:
wait(
lambda: not self.fuel_web.get_nailgun_node_by_devops_node(
self.env.d_env.nodes().
slaves[2])['online'], timeout=60 * 6)
except TimeoutError:
raise
with self.env.d_env.get_admin_remote() as remote:
res = remote.execute('fuel node --node-id {0} --delete-from-db'
.format(node_id))
assert_true(
res['exit_code'] == 0,
"Offline node-{0} was not"
"deleted from database".format(node_id))
with self.env.d_env.get_admin_remote() as remote:
try:
wait(
lambda: not remote.execute(
"fuel node | awk '{{print $1}}' | grep -w '{0}'".
format(node_id))['exit_code'] == 0, timeout=60 * 4)
except TimeoutError:
raise TimeoutError(
"After deletion node-{0} is found in fuel list".
format(node_id))
is_cobbler_node_exists = check_cobbler_node_exists(
self.ssh_manager.admin_ip, node_id)
assert_false(is_cobbler_node_exists,
"After deletion node-{0} is found in cobbler list".
format(node_id))
with self.env.d_env.get_admin_remote() as remote:
cluster_id = ''.join(remote.execute(
"fuel env | tail -n 1 | awk {'print $1'}")['stdout']).rstrip()
self.fuel_web.verify_network(cluster_id)
self.fuel_web.run_ostf(
cluster_id=cluster_id, test_sets=['ha', 'smoke', 'sanity'],
should_fail=1)
@test(depends_on_groups=['cli_selected_nodes_deploy'],
groups=["cli_cluster_deletion"])
@log_snapshot_after_test
def cli_cluster_deletion(self):
"""Delete a cluster using Fuel CLI
Scenario:
1. Revert snapshot 'cli_selected_nodes_deploy'
2. Delete cluster via cli
3. Check cluster absence in the list
Duration 25m
"""
self.env.revert_snapshot("cli_selected_nodes_deploy")
cluster_id = self.fuel_web.get_last_created_cluster()
nodes = self.fuel_web.client.list_cluster_nodes(cluster_id)
online_nodes = [node for node in nodes if node['online']]
if nodes != online_nodes:
logger.error(
'Some slaves do not become online after revert!!'
' Expected {0} Actual {1}'.format(nodes, online_nodes))
self.ssh_manager.execute_on_remote(
ip=self.ssh_manager.admin_ip,
cmd='fuel --env {0} env delete --force'.format(cluster_id)
)
try:
wait(lambda:
self.ssh_manager.execute(
ip=self.ssh_manager.admin_ip,
cmd="fuel env | awk '{print $1}' | tail -n 1 | "
"grep '^.$'"
)['exit_code'] == 1, timeout=60 * 10)
except TimeoutError:
raise TimeoutError(
"cluster {0} was not deleted".format(cluster_id))
assert_false(
check_cluster_presence(cluster_id, self.env.postgres_actions),
"cluster {0} is found".format(cluster_id))