af71a8bf83
HTTP client was replaced by keystoneauth1 session -> update test code
Related-bug: #1609705
Change-Id: I1b3881e7c75fd0163347d4d6605dd979087393eb
(cherry picked from commit 92a2518
)
563 lines
20 KiB
Python
563 lines
20 KiB
Python
# Copyright 2014 Mirantis, Inc.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
from copy import deepcopy
|
|
|
|
from keystoneauth1.exceptions import HttpError
|
|
from proboscis.asserts import assert_equal
|
|
from proboscis import test
|
|
|
|
from fuelweb_test import logger
|
|
from fuelweb_test.helpers.decorators import log_snapshot_after_test
|
|
from fuelweb_test.settings import DEPLOYMENT_MODE
|
|
from fuelweb_test.settings import NEUTRON_SEGMENT
|
|
from fuelweb_test.tests.base_test_case import SetupEnvironment
|
|
from fuelweb_test.tests.test_bonding_base import BondingTest
|
|
|
|
|
|
@test(groups=["bonding_ha_one_controller", "bonding"])
|
|
class BondingHAOneController(BondingTest):
|
|
"""BondingHAOneController.""" # TODO documentation
|
|
|
|
@test(depends_on=[SetupEnvironment.prepare_slaves_3],
|
|
groups=["deploy_bonding_one_controller_tun"])
|
|
@log_snapshot_after_test
|
|
def deploy_bonding_one_controller_tun(self):
|
|
"""Deploy cluster with active-backup bonding and Neutron VXLAN
|
|
|
|
Scenario:
|
|
1. Create cluster
|
|
2. Add 1 node with controller role
|
|
3. Add 1 node with compute role and 1 node with cinder role
|
|
4. Setup bonding for all interfaces (including admin interface
|
|
bonding)
|
|
5. Run network verification
|
|
6. Deploy the cluster
|
|
7. Run network verification
|
|
8. Run OSTF
|
|
|
|
Duration 30m
|
|
Snapshot deploy_bonding_one_controller_tun
|
|
"""
|
|
|
|
self.env.revert_snapshot("ready_with_3_slaves")
|
|
|
|
segment_type = NEUTRON_SEGMENT['tun']
|
|
|
|
self.show_step(1, initialize=True)
|
|
cluster_id = self.fuel_web.create_cluster(
|
|
name=self.__class__.__name__,
|
|
mode=DEPLOYMENT_MODE,
|
|
settings={
|
|
"net_provider": 'neutron',
|
|
"net_segment_type": segment_type,
|
|
}
|
|
)
|
|
|
|
self.show_step(2)
|
|
self.show_step(3)
|
|
self.fuel_web.update_nodes(
|
|
cluster_id, {
|
|
'slave-01': ['controller'],
|
|
'slave-02': ['compute'],
|
|
'slave-03': ['cinder']
|
|
}
|
|
)
|
|
|
|
self.show_step(4)
|
|
nailgun_nodes = self.fuel_web.client.list_cluster_nodes(cluster_id)
|
|
for node in nailgun_nodes:
|
|
self.fuel_web.update_node_networks(
|
|
node['id'], interfaces_dict=deepcopy(self.INTERFACES),
|
|
raw_data=deepcopy(self.BOND_CONFIG)
|
|
)
|
|
self.show_step(5)
|
|
self.fuel_web.verify_network(cluster_id)
|
|
|
|
self.show_step(6)
|
|
self.fuel_web.deploy_cluster_wait(cluster_id, check_services=False)
|
|
|
|
self.show_step(7)
|
|
self.fuel_web.verify_network(cluster_id)
|
|
|
|
self.show_step(8)
|
|
self.fuel_web.run_ostf(cluster_id=cluster_id)
|
|
|
|
self.env.make_snapshot("deploy_bonding_one_controller_tun")
|
|
|
|
@test(depends_on=[SetupEnvironment.prepare_slaves_3],
|
|
groups=["deploy_bonding_one_controller_vlan"])
|
|
@log_snapshot_after_test
|
|
def deploy_bonding_one_controller_vlan(self):
|
|
"""Deploy cluster with active-backup bonding and Neutron VLAN
|
|
|
|
Scenario:
|
|
1. Create cluster
|
|
2. Add 1 node with controller role
|
|
3. Add 1 node with compute role and 1 node with cinder role
|
|
4. Setup bonding for all interfaces (including admin interface
|
|
bonding)
|
|
5. Run network verification
|
|
6. Deploy the cluster
|
|
7. Run network verification
|
|
8. Run OSTF
|
|
|
|
|
|
Duration 30m
|
|
Snapshot deploy_bonding_one_controller_vlan
|
|
"""
|
|
|
|
self.env.revert_snapshot("ready_with_3_slaves")
|
|
|
|
segment_type = NEUTRON_SEGMENT['vlan']
|
|
|
|
self.show_step(1, initialize=True)
|
|
cluster_id = self.fuel_web.create_cluster(
|
|
name=self.__class__.__name__,
|
|
mode=DEPLOYMENT_MODE,
|
|
settings={
|
|
"net_provider": 'neutron',
|
|
"net_segment_type": segment_type,
|
|
}
|
|
)
|
|
|
|
self.show_step(2)
|
|
self.show_step(3)
|
|
self.fuel_web.update_nodes(
|
|
cluster_id, {
|
|
'slave-01': ['controller'],
|
|
'slave-02': ['compute'],
|
|
'slave-03': ['cinder']
|
|
}
|
|
)
|
|
|
|
self.show_step(4)
|
|
nailgun_nodes = self.fuel_web.client.list_cluster_nodes(cluster_id)
|
|
for node in nailgun_nodes:
|
|
self.fuel_web.update_node_networks(
|
|
node['id'], interfaces_dict=deepcopy(self.INTERFACES),
|
|
raw_data=deepcopy(self.BOND_CONFIG)
|
|
)
|
|
|
|
self.show_step(5)
|
|
self.fuel_web.verify_network(cluster_id)
|
|
|
|
self.show_step(6)
|
|
self.fuel_web.deploy_cluster_wait(cluster_id, check_services=False)
|
|
|
|
self.show_step(7)
|
|
self.fuel_web.verify_network(cluster_id)
|
|
|
|
self.show_step(8)
|
|
self.fuel_web.run_ostf(cluster_id=cluster_id)
|
|
|
|
self.env.make_snapshot("deploy_bonding_one_controller_vlan")
|
|
|
|
@test(depends_on=[SetupEnvironment.prepare_slaves_3],
|
|
groups=["negative_admin_bonding_in_lacp_mode"])
|
|
@log_snapshot_after_test
|
|
def negative_admin_bonding_in_lacp_mode(self):
|
|
"""Verify that lacp mode cannot be enabled for admin bond
|
|
|
|
Scenario:
|
|
1. Create cluster
|
|
2. Add 1 node with controller role
|
|
3. Add 1 node with compute role and 1 node with cinder role
|
|
4. Verify that lacp mode cannot be enabled for admin bond
|
|
|
|
Duration 4m
|
|
Snapshot negative_admin_bonding_in_lacp_mode
|
|
"""
|
|
self.env.revert_snapshot("ready_with_3_slaves")
|
|
|
|
segment_type = NEUTRON_SEGMENT['tun']
|
|
|
|
self.show_step(1, initialize=True)
|
|
cluster_id = self.fuel_web.create_cluster(
|
|
name=self.__class__.__name__,
|
|
mode=DEPLOYMENT_MODE,
|
|
settings={
|
|
"net_provider": 'neutron',
|
|
"net_segment_type": segment_type,
|
|
}
|
|
)
|
|
|
|
self.show_step(2)
|
|
self.show_step(3)
|
|
self.fuel_web.update_nodes(
|
|
cluster_id, {
|
|
'slave-01': ['controller'],
|
|
'slave-02': ['compute'],
|
|
'slave-03': ['cinder']
|
|
}
|
|
)
|
|
|
|
self.show_step(4)
|
|
nailgun_nodes = self.fuel_web.client.list_cluster_nodes(cluster_id)
|
|
invalid_bond_conf = deepcopy(self.BOND_CONFIG)
|
|
invalid_bond_conf[1]['mode'] = '802.3ad'
|
|
invalid_bond_conf[1]['bond_properties']['mode'] = '802.3ad'
|
|
interfaces_dict = deepcopy(self.INTERFACES)
|
|
|
|
exp_code = 400
|
|
try:
|
|
self.fuel_web.update_node_networks(
|
|
nailgun_nodes[0]['id'],
|
|
interfaces_dict=interfaces_dict,
|
|
raw_data=invalid_bond_conf)
|
|
except HttpError as exc:
|
|
if exc.http_status != exp_code:
|
|
logger.error(
|
|
'Raised: {exc!s},\n'
|
|
'Expected: {exp} with code={code}'.format(
|
|
exc=exc,
|
|
exp=HttpError,
|
|
code=exp_code))
|
|
raise
|
|
|
|
logger.info('Test PASS: expected exception raised: '
|
|
'{!s}'.format(exc))
|
|
return
|
|
except BaseException as exc:
|
|
logger.error(
|
|
'Raised: {exc!s},\n'
|
|
'Expected: {exp} with code={code}'.format(
|
|
exc=exc,
|
|
exp=HttpError,
|
|
code=exp_code))
|
|
raise
|
|
raise AssertionError(
|
|
'Not raised any exception, while expected '
|
|
'{exp} with code={code}'.format(
|
|
exp=HttpError,
|
|
code=exp_code))
|
|
|
|
|
|
@test(groups=["bonding_neutron", "bonding_ha", "bonding"])
|
|
class BondingHA(BondingTest):
|
|
"""Tests for HA bonding."""
|
|
|
|
@test(depends_on=[SetupEnvironment.prepare_slaves_5],
|
|
groups=["deploy_bonding_neutron_vlan"])
|
|
@log_snapshot_after_test
|
|
def deploy_bonding_neutron_vlan(self):
|
|
"""Deploy cluster with active-backup bonding and Neutron VLAN
|
|
|
|
Scenario:
|
|
1. Create cluster
|
|
2. Add 3 nodes with controller role
|
|
3. Add 1 node with compute role and 1 node with cinder role
|
|
4. Setup bonding for all interfaces (including admin interface
|
|
bonding)
|
|
5. Run network verification
|
|
6. Deploy the cluster
|
|
7. Run network verification
|
|
8. Run OSTF
|
|
9. Save network configuration from slave nodes
|
|
10. Reboot all environment nodes
|
|
11. Verify that network configuration is the same after reboot
|
|
12. Run network verification
|
|
13. Run OSTF
|
|
|
|
Duration 70m
|
|
Snapshot deploy_bonding_neutron_vlan
|
|
"""
|
|
|
|
self.env.revert_snapshot("ready_with_5_slaves")
|
|
|
|
segment_type = NEUTRON_SEGMENT['vlan']
|
|
|
|
self.show_step(1, initialize=True)
|
|
cluster_id = self.fuel_web.create_cluster(
|
|
name=self.__class__.__name__,
|
|
mode=DEPLOYMENT_MODE,
|
|
settings={
|
|
"net_provider": 'neutron',
|
|
"net_segment_type": segment_type,
|
|
}
|
|
)
|
|
|
|
self.show_step(2)
|
|
self.show_step(3)
|
|
self.fuel_web.update_nodes(
|
|
cluster_id, {
|
|
'slave-01': ['controller'],
|
|
'slave-02': ['controller'],
|
|
'slave-03': ['controller'],
|
|
'slave-04': ['compute'],
|
|
'slave-05': ['cinder']
|
|
}
|
|
)
|
|
|
|
net_params = self.fuel_web.client.get_networks(cluster_id)
|
|
|
|
self.show_step(4)
|
|
nailgun_nodes = self.fuel_web.client.list_cluster_nodes(cluster_id)
|
|
for node in nailgun_nodes:
|
|
self.fuel_web.update_node_networks(
|
|
node['id'], interfaces_dict=deepcopy(self.INTERFACES),
|
|
raw_data=deepcopy(self.BOND_CONFIG)
|
|
)
|
|
|
|
self.show_step(5)
|
|
self.fuel_web.verify_network(cluster_id)
|
|
|
|
self.show_step(6)
|
|
self.fuel_web.deploy_cluster_wait(cluster_id, check_services=False)
|
|
|
|
cluster = self.fuel_web.client.get_cluster(cluster_id)
|
|
assert_equal(str(cluster['net_provider']), 'neutron')
|
|
assert_equal(str(net_params["networking_parameters"]
|
|
['segmentation_type']), segment_type)
|
|
|
|
self.show_step(7)
|
|
self.fuel_web.verify_network(cluster_id)
|
|
|
|
self.show_step(8)
|
|
self.fuel_web.run_ostf(cluster_id=cluster_id)
|
|
|
|
self.show_step(9)
|
|
self.show_step(10)
|
|
self.show_step(11)
|
|
self.check_interfaces_config_after_reboot(cluster_id)
|
|
self.fuel_web.assert_ha_services_ready(cluster_id)
|
|
|
|
self.show_step(12)
|
|
self.fuel_web.verify_network(cluster_id)
|
|
|
|
self.show_step(13)
|
|
self.fuel_web.run_ostf(cluster_id=cluster_id)
|
|
|
|
self.env.make_snapshot("deploy_bonding_neutron_vlan")
|
|
|
|
@test(depends_on=[SetupEnvironment.prepare_slaves_5],
|
|
groups=["deploy_bonding_neutron_tun"])
|
|
@log_snapshot_after_test
|
|
def deploy_bonding_neutron_tun(self):
|
|
"""Deploy cluster with active-backup bonding and Neutron VXLAN
|
|
|
|
Scenario:
|
|
1. Create cluster
|
|
2. Add 3 nodes with controller role
|
|
3. Add 1 node with compute role and 1 node with cinder role
|
|
4. Setup bonding for all interfaces (including admin interface
|
|
bonding)
|
|
5. Run network verification
|
|
6. Deploy the cluster
|
|
7. Run network verification
|
|
8. Run OSTF
|
|
9. Save network configuration from slave nodes
|
|
10. Reboot all environment nodes
|
|
11. Verify that network configuration is the same after reboot
|
|
12. Run network verification
|
|
13. Run OST
|
|
|
|
Duration 70m
|
|
Snapshot deploy_bonding_neutron_tun
|
|
"""
|
|
|
|
self.env.revert_snapshot("ready_with_5_slaves")
|
|
|
|
segment_type = NEUTRON_SEGMENT['tun']
|
|
|
|
self.show_step(1, initialize=True)
|
|
cluster_id = self.fuel_web.create_cluster(
|
|
name=self.__class__.__name__,
|
|
mode=DEPLOYMENT_MODE,
|
|
settings={
|
|
"net_provider": 'neutron',
|
|
"net_segment_type": segment_type,
|
|
}
|
|
)
|
|
|
|
self.show_step(2)
|
|
self.show_step(3)
|
|
self.fuel_web.update_nodes(
|
|
cluster_id, {
|
|
'slave-01': ['controller'],
|
|
'slave-02': ['controller'],
|
|
'slave-03': ['controller'],
|
|
'slave-04': ['compute'],
|
|
'slave-05': ['cinder']
|
|
}
|
|
)
|
|
|
|
net_params = self.fuel_web.client.get_networks(cluster_id)
|
|
|
|
self.show_step(4)
|
|
nailgun_nodes = self.fuel_web.client.list_cluster_nodes(cluster_id)
|
|
for node in nailgun_nodes:
|
|
self.fuel_web.update_node_networks(
|
|
node['id'], interfaces_dict=deepcopy(self.INTERFACES),
|
|
raw_data=deepcopy(self.BOND_CONFIG)
|
|
)
|
|
|
|
self.show_step(5)
|
|
self.fuel_web.verify_network(cluster_id)
|
|
|
|
self.show_step(6)
|
|
self.fuel_web.deploy_cluster_wait(cluster_id, check_services=False)
|
|
|
|
cluster = self.fuel_web.client.get_cluster(cluster_id)
|
|
assert_equal(str(cluster['net_provider']), 'neutron')
|
|
assert_equal(str(net_params["networking_parameters"]
|
|
['segmentation_type']), segment_type)
|
|
|
|
self.show_step(7)
|
|
self.fuel_web.verify_network(cluster_id)
|
|
|
|
self.show_step(8)
|
|
self.fuel_web.run_ostf(cluster_id=cluster_id)
|
|
|
|
self.show_step(9)
|
|
self.show_step(10)
|
|
self.show_step(11)
|
|
self.check_interfaces_config_after_reboot(cluster_id)
|
|
self.fuel_web.assert_ha_services_ready(cluster_id)
|
|
|
|
self.show_step(12)
|
|
self.fuel_web.verify_network(cluster_id)
|
|
|
|
self.show_step(13)
|
|
self.fuel_web.run_ostf(cluster_id=cluster_id)
|
|
|
|
self.env.make_snapshot("deploy_bonding_neutron_tun")
|
|
|
|
@test(depends_on=[SetupEnvironment.prepare_slaves_5],
|
|
groups=["bonding_conf_consistency"])
|
|
@log_snapshot_after_test
|
|
def bonding_conf_consistency(self):
|
|
"""Verify that network configuration with bonds is consistent\
|
|
after deployment failure
|
|
|
|
Scenario:
|
|
1. Create an environment
|
|
2. Add 3 nodes with controller role
|
|
3. Add 1 node with compute role
|
|
4. Setup bonding for all interfaces (including admin interface
|
|
bonding)
|
|
5. Run network verification
|
|
6. Update 'connectivity_tests' puppet manifest to cause the\
|
|
deployment process fail right after 'netconfig' task is finished
|
|
7. Start deployment and wait until it fails
|
|
8. Verify that interfaces are not lost from the configured bonds
|
|
9. Restore the initial version of 'connectivity_tests' manifest
|
|
10. Redeploy the cluster and run basic health checks
|
|
11. Run network verification
|
|
|
|
Duration 120m
|
|
Snapshot bonding_conf_consistency
|
|
"""
|
|
|
|
self.env.revert_snapshot("ready_with_5_slaves")
|
|
|
|
self.show_step(1, initialize=True)
|
|
cluster_id = self.fuel_web.create_cluster(
|
|
name=self.__class__.__name__,
|
|
mode=DEPLOYMENT_MODE,
|
|
settings={
|
|
"net_provider": 'neutron',
|
|
"net_segment_type": NEUTRON_SEGMENT['vlan'],
|
|
}
|
|
)
|
|
|
|
self.show_step(2)
|
|
self.show_step(3)
|
|
self.fuel_web.update_nodes(
|
|
cluster_id, {
|
|
'slave-01': ['controller'],
|
|
'slave-02': ['controller'],
|
|
'slave-03': ['controller'],
|
|
'slave-04': ['compute'],
|
|
}
|
|
)
|
|
|
|
self.show_step(4)
|
|
nailgun_nodes = self.fuel_web.client.list_cluster_nodes(cluster_id)
|
|
for node in nailgun_nodes:
|
|
self.fuel_web.update_node_networks(
|
|
node['id'], interfaces_dict=deepcopy(self.INTERFACES),
|
|
raw_data=deepcopy(self.BOND_CONFIG)
|
|
)
|
|
|
|
self.show_step(5)
|
|
self.fuel_web.verify_network(cluster_id)
|
|
|
|
# Get ID of a (pending) primary controller
|
|
pending_ctrl_ids = [n['id'] for n in nailgun_nodes
|
|
if 'controller' in n['pending_roles']]
|
|
node_id = sorted(pending_ctrl_ids)[0]
|
|
|
|
# Get interfaces data of the primary controller for which deployment
|
|
# will be forced to fail
|
|
ifaces_data = self.fuel_web.client.get_node_interfaces(node_id)
|
|
|
|
self.show_step(6)
|
|
pp_file = ("/etc/puppet/modules/osnailyfacter/modular/netconfig/"
|
|
"connectivity_tests.pp")
|
|
with self.env.d_env.get_admin_remote() as admin_node:
|
|
# Backup the manifest to be updated for the sake of the test
|
|
backup_cmd = "cp {0} {1}".format(pp_file, pp_file + "_bak")
|
|
res = admin_node.execute(backup_cmd)
|
|
assert_equal(0, res['exit_code'],
|
|
"Failed to create a backup copy of {0} puppet "
|
|
"manifest on master node".format(pp_file))
|
|
|
|
fail_cmd = ("echo 'fail(\"Emulate deployment failure after "
|
|
"netconfig!\")' >> {0}".format(pp_file))
|
|
res = admin_node.execute(fail_cmd)
|
|
assert_equal(0, res['exit_code'],
|
|
"Failed to update {0} puppet manifest "
|
|
"on master node".format(pp_file))
|
|
|
|
self.show_step(7)
|
|
task = self.fuel_web.deploy_cluster(cluster_id)
|
|
self.fuel_web.assert_task_failed(task)
|
|
|
|
# Get interfaces data after deployment failure on
|
|
# the primary controller
|
|
ifaces_data_latest = self.fuel_web.client.get_node_interfaces(node_id)
|
|
|
|
self.show_step(8)
|
|
# Bond interfaces are always the last objects in the list being
|
|
# returned by 'get node interfaces' API request.
|
|
# So having 2 bonds on the node under test the last 2 objects
|
|
# in the corresponding list are being examined below
|
|
admin_bond_ifaces = ifaces_data[-1]['slaves']
|
|
admin_bond_ifaces_latest = ifaces_data_latest[-1]['slaves']
|
|
assert_equal(len(admin_bond_ifaces), len(admin_bond_ifaces_latest),
|
|
"Admin interface bond config is inconsistent; "
|
|
"interface(s) have disappeared from the bond")
|
|
others_bond_ifaces = ifaces_data[-2]['slaves']
|
|
others_bond_ifaces_latest = ifaces_data_latest[-2]['slaves']
|
|
assert_equal(len(others_bond_ifaces), len(others_bond_ifaces_latest),
|
|
"Other network interfaces bond config is inconsistent; "
|
|
"interface(s) have disappeared from the bond")
|
|
|
|
self.show_step(9)
|
|
with self.env.d_env.get_admin_remote() as admin_node:
|
|
restore_cmd = "cp {0} {1}".format(pp_file + "_bak", pp_file)
|
|
res = admin_node.execute(restore_cmd)
|
|
assert_equal(0, res['exit_code'],
|
|
"Failed to restore the backup copy of {0} puppet "
|
|
"manifest on master node".format(pp_file))
|
|
|
|
self.show_step(10)
|
|
self.fuel_web.deploy_cluster_wait(cluster_id)
|
|
|
|
self.show_step(11)
|
|
self.fuel_web.verify_network(cluster_id)
|
|
|
|
self.env.make_snapshot("bonding_conf_consistency")
|