Merge "Separate test cases by names in multiple Jenkins jobs"
This commit is contained in:
commit
d8526ae95a
|
@ -16,6 +16,11 @@ Admin Node Tests
|
|||
.. automodule:: fuelweb_test.tests.test_admin_node
|
||||
:members:
|
||||
|
||||
Test Admin Node Backup-Restore
|
||||
------------------------------
|
||||
.. automodule:: fuelweb_test.tests.test_backup_restore
|
||||
:members:
|
||||
|
||||
Test Bonding
|
||||
------------
|
||||
.. automodule:: fuelweb_test.tests.test_bonding
|
||||
|
@ -273,4 +278,4 @@ Test Upgrade
|
|||
Test Upgrade Chains
|
||||
-------------------
|
||||
.. automodule:: fuelweb_test.tests.tests_upgrade.test_upgrade_chains
|
||||
:members:
|
||||
:members:
|
||||
|
|
|
@ -42,6 +42,7 @@ class CloseSSHConnectionsPlugin(Plugin):
|
|||
|
||||
def import_tests():
|
||||
from tests import test_admin_node # noqa
|
||||
from tests import test_backup_restore # noqa
|
||||
from tests import test_ceph # noqa
|
||||
from tests import test_environment_action # noqa
|
||||
from tests import test_neutron # noqa
|
||||
|
|
|
@ -92,34 +92,6 @@ class TestAdminNode(TestBasic):
|
|||
assert_equal(True, len(astute_workers) > 1)
|
||||
|
||||
|
||||
@test(groups=["known_issues"])
|
||||
class TestAdminNodeBackupRestore(TestBasic):
|
||||
@test(depends_on=[SetupEnvironment.setup_master],
|
||||
groups=["backup_restore_master_base"])
|
||||
@log_snapshot_after_test
|
||||
def backup_restore_master_base(self):
|
||||
"""Backup/restore master node
|
||||
|
||||
Scenario:
|
||||
1. Revert snapshot "empty"
|
||||
2. Backup master
|
||||
3. Check backup
|
||||
4. Restore master
|
||||
5. Check restore
|
||||
|
||||
Duration 30m
|
||||
|
||||
"""
|
||||
self.env.revert_snapshot("empty")
|
||||
self.fuel_web.backup_master(self.env.d_env.get_admin_remote())
|
||||
checkers.backup_check(self.env.d_env.get_admin_remote())
|
||||
self.fuel_web.restore_master(self.env.d_env.get_admin_remote())
|
||||
self.fuel_web.restore_check_nailgun_api(
|
||||
self.env.d_env.get_admin_remote())
|
||||
checkers.restore_check_sum(self.env.d_env.get_admin_remote())
|
||||
checkers.iptables_check(self.env.d_env.get_admin_remote())
|
||||
|
||||
|
||||
@test(groups=["logrotate"])
|
||||
class TestLogrotateBase(TestBasic):
|
||||
|
||||
|
|
|
@ -0,0 +1,234 @@
|
|||
# Copyright 2015 Mirantis, Inc.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from proboscis.asserts import assert_equal
|
||||
from proboscis import test
|
||||
|
||||
from fuelweb_test.helpers import checkers
|
||||
from fuelweb_test.helpers.decorators import log_snapshot_after_test
|
||||
from fuelweb_test.helpers import os_actions
|
||||
from fuelweb_test.tests.base_test_case import SetupEnvironment
|
||||
from fuelweb_test.tests.base_test_case import TestBasic
|
||||
from fuelweb_test.tests.test_ha_one_controller_base\
|
||||
import HAOneControllerNeutronBase
|
||||
from fuelweb_test.tests.test_neutron_tun import NeutronTunHaBase
|
||||
|
||||
|
||||
@test(groups=["known_issues"])
|
||||
class TestAdminNodeBackupRestore(TestBasic):
|
||||
|
||||
@test(depends_on=[SetupEnvironment.setup_master],
|
||||
groups=["backup_restore_master_base"])
|
||||
@log_snapshot_after_test
|
||||
def backup_restore_master_base(self):
|
||||
"""Backup/restore master node
|
||||
|
||||
Scenario:
|
||||
1. Revert snapshot "empty"
|
||||
2. Backup master
|
||||
3. Check backup
|
||||
4. Restore master
|
||||
5. Check restore
|
||||
6. Check iptables
|
||||
|
||||
Duration 30m
|
||||
|
||||
"""
|
||||
self.env.revert_snapshot("empty")
|
||||
|
||||
with self.env.d_env.get_admin_remote() as remote:
|
||||
self.fuel_web.backup_master(remote)
|
||||
checkers.backup_check(remote)
|
||||
self.fuel_web.restore_master(remote)
|
||||
self.fuel_web.restore_check_nailgun_api(remote)
|
||||
checkers.restore_check_sum(remote)
|
||||
checkers.iptables_check(remote)
|
||||
|
||||
|
||||
@test(groups=["known_issues"])
|
||||
class BackupRestoreHAOneController(HAOneControllerNeutronBase):
|
||||
"""BackupRestoreHAOneController""" # TODO documentation
|
||||
|
||||
@test(depends_on=[SetupEnvironment.prepare_slaves_3],
|
||||
groups=["deploy_ha_one_controller_backup_restore"])
|
||||
@log_snapshot_after_test
|
||||
def deploy_ha_one_controller_backup_restore(self):
|
||||
"""Deploy cluster in HA mode (one controller) with neutron
|
||||
|
||||
Scenario:
|
||||
1. Create cluster in HA mode
|
||||
2. Add 1 node with controller role
|
||||
3. Add 1 node with compute role
|
||||
4. Deploy the cluster
|
||||
5. Validate cluster was set up correctly, there are no dead
|
||||
services, there are no errors in logs
|
||||
6. Verify networks
|
||||
7. Verify network configuration on controller
|
||||
8. Run OSTF
|
||||
|
||||
Duration 30m
|
||||
Snapshot: deploy_ha_one_controller_backup_restore
|
||||
"""
|
||||
super(self.__class__, self).deploy_ha_one_controller_neutron_base(
|
||||
snapshot_name="deploy_ha_one_controller_backup_restore")
|
||||
|
||||
@test(depends_on=[deploy_ha_one_controller_backup_restore],
|
||||
groups=["ha_one_controller_backup_restore"])
|
||||
@log_snapshot_after_test
|
||||
def ha_one_controller_backup_restore(self):
|
||||
"""Backup/restore master node with one controller in cluster
|
||||
|
||||
Scenario:
|
||||
1. Revert snapshot "deploy_ha_one_controller_backup_restore"
|
||||
2. Backup master
|
||||
3. Check backup
|
||||
4. Run OSTF
|
||||
5. Add 1 node with compute role
|
||||
6. Restore master
|
||||
7. Check restore
|
||||
8. Run OSTF
|
||||
|
||||
Duration 35m
|
||||
|
||||
"""
|
||||
self.env.revert_snapshot("deploy_ha_one_controller_backup_restore")
|
||||
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
|
||||
os_conn = os_actions.OpenStackActions(
|
||||
self.fuel_web.get_public_vip(cluster_id),
|
||||
'neutronOneController', 'neutronOneController',
|
||||
'neutronOneController')
|
||||
self.fuel_web.assert_cluster_ready(os_conn, smiles_count=5,
|
||||
networks_count=2, timeout=300)
|
||||
|
||||
with self.env.d_env.get_admin_remote() as remote:
|
||||
# Execute master node backup
|
||||
self.fuel_web.backup_master(remote)
|
||||
# Check created backup
|
||||
checkers.backup_check(remote)
|
||||
|
||||
self.fuel_web.update_nodes(
|
||||
cluster_id, {'slave-03': ['compute']}, True, False)
|
||||
|
||||
assert_equal(
|
||||
3, len(self.fuel_web.client.list_cluster_nodes(cluster_id)))
|
||||
|
||||
with self.env.d_env.get_admin_remote() as remote:
|
||||
self.fuel_web.restore_master(remote)
|
||||
checkers.restore_check_sum(remote)
|
||||
self.fuel_web.restore_check_nailgun_api(remote)
|
||||
checkers.iptables_check(remote)
|
||||
|
||||
assert_equal(
|
||||
2, len(self.fuel_web.client.list_cluster_nodes(cluster_id)))
|
||||
|
||||
self.fuel_web.update_nodes(
|
||||
cluster_id, {'slave-03': ['compute']}, True, False)
|
||||
self.fuel_web.deploy_cluster_wait(cluster_id)
|
||||
|
||||
self.fuel_web.run_ostf(
|
||||
cluster_id=cluster_id)
|
||||
|
||||
self.env.make_snapshot("ha_one_controller_backup_restore")
|
||||
|
||||
|
||||
@test(groups=["known_issues"])
|
||||
class BackupRestoreHA(NeutronTunHaBase):
|
||||
"""BackupRestoreHAOneController""" # TODO documentation
|
||||
|
||||
@test(depends_on=[SetupEnvironment.prepare_slaves_5],
|
||||
groups=["deploy_neutron_tun_ha_backup_restore"])
|
||||
@log_snapshot_after_test
|
||||
def deploy_neutron_tun_ha_backup_restore(self):
|
||||
"""Deploy cluster in HA mode with Neutron VXLAN
|
||||
|
||||
Scenario:
|
||||
1. Create cluster
|
||||
2. Add 3 nodes with controller role
|
||||
3. Add 2 nodes with compute role
|
||||
4. Deploy the cluster
|
||||
5. Run network verification
|
||||
6. Run OSTF
|
||||
|
||||
Duration 80m
|
||||
Snapshot deploy_neutron_tun_ha_backup_restore
|
||||
"""
|
||||
super(self.__class__, self).deploy_neutron_tun_ha_base(
|
||||
snapshot_name="deploy_neutron_tun_ha_backup_restore")
|
||||
|
||||
@test(depends_on_groups=['deploy_neutron_tun_ha_backup_restore'],
|
||||
groups=["neutron_tun_ha_backup_restore"])
|
||||
@log_snapshot_after_test
|
||||
def neutron_tun_ha_backup_restore(self):
|
||||
"""Backup/restore master node with cluster in ha mode
|
||||
|
||||
Scenario:
|
||||
1. Revert snapshot "deploy_neutron_tun_ha"
|
||||
2. Backup master
|
||||
3. Check backup
|
||||
4. Run OSTF
|
||||
5. Add 1 node with compute role
|
||||
6. Restore master
|
||||
7. Check restore
|
||||
8. Run OSTF
|
||||
|
||||
Duration 50m
|
||||
"""
|
||||
self.env.revert_snapshot("deploy_neutron_tun_ha_backup_restore")
|
||||
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
cluster = self.fuel_web.client.get_cluster(cluster_id)
|
||||
assert_equal(str(cluster['net_provider']), 'neutron')
|
||||
os_conn = os_actions.OpenStackActions(
|
||||
self.fuel_web.get_public_vip(cluster_id))
|
||||
# assert_equal(str(cluster['net_segment_type']), segment_type)
|
||||
self.fuel_web.check_fixed_network_cidr(
|
||||
cluster_id, os_conn)
|
||||
|
||||
with self.env.d_env.get_admin_remote() as remote:
|
||||
self.fuel_web.backup_master(remote)
|
||||
checkers.backup_check(remote)
|
||||
|
||||
self.env.bootstrap_nodes(
|
||||
self.env.d_env.nodes().slaves[5:6])
|
||||
self.fuel_web.update_nodes(
|
||||
cluster_id, {'slave-06': ['compute']}, True, False
|
||||
)
|
||||
|
||||
assert_equal(
|
||||
6, len(self.fuel_web.client.list_cluster_nodes(cluster_id)))
|
||||
|
||||
with self.env.d_env.get_admin_remote() as remote:
|
||||
self.fuel_web.restore_master(remote)
|
||||
checkers.restore_check_sum(remote)
|
||||
|
||||
self.fuel_web.restore_check_nailgun_api(remote)
|
||||
checkers.iptables_check(remote)
|
||||
|
||||
assert_equal(
|
||||
5, len(self.fuel_web.client.list_cluster_nodes(cluster_id)))
|
||||
|
||||
self.env.bootstrap_nodes(
|
||||
self.env.d_env.nodes().slaves[5:6])
|
||||
self.fuel_web.update_nodes(
|
||||
cluster_id, {'slave-06': ['compute']}, True, False
|
||||
)
|
||||
self.fuel_web.deploy_cluster_wait(cluster_id)
|
||||
|
||||
self.fuel_web.run_ostf(
|
||||
cluster_id=cluster_id,
|
||||
test_sets=['ha', 'smoke', 'sanity'])
|
||||
|
||||
self.env.make_snapshot("neutron_tun_ha_backup_restore")
|
|
@ -241,7 +241,7 @@ class CustomHostname(TestBasic):
|
|||
self.env.make_snapshot("set_duplicate_hostname")
|
||||
|
||||
@test(depends_on=[SetupEnvironment.prepare_slaves_5],
|
||||
groups=['known_issues', 'set_custom_hostname_for_provisioned_node'])
|
||||
groups=['set_custom_hostname_for_provisioned_node'])
|
||||
@log_snapshot_after_test
|
||||
def set_custom_hostname_for_provisioned_node(self):
|
||||
"""Verify that it is not allowed to change a hostname of a
|
||||
|
|
|
@ -111,7 +111,8 @@ class HAOneControllerNeutron(HAOneControllerNeutronBase):
|
|||
Duration 30m
|
||||
Snapshot: deploy_ha_one_controller_neutron
|
||||
"""
|
||||
super(self.__class__, self).deploy_ha_one_controller_neutron_base()
|
||||
super(self.__class__, self).deploy_ha_one_controller_neutron_base(
|
||||
snapshot_name="deploy_ha_one_controller_neutron")
|
||||
|
||||
@test(depends_on=[deploy_ha_one_controller_neutron],
|
||||
groups=["ha_one_controller_neutron_node_deletion"])
|
||||
|
@ -926,71 +927,6 @@ class UntaggedNetworksNegative(TestBasic):
|
|||
self.fuel_web.assert_task_failed(task)
|
||||
|
||||
|
||||
@test(groups=["known_issues"])
|
||||
class BackupRestoreHAOneController(TestBasic):
|
||||
"""BackupRestoreHAOneController""" # TODO documentation
|
||||
|
||||
@test(depends_on=[HAOneControllerNeutron.deploy_ha_one_controller_neutron],
|
||||
groups=["ha_one_controller_backup_restore"])
|
||||
@log_snapshot_after_test
|
||||
def ha_one_controller_backup_restore(self):
|
||||
"""Backup/restore master node with one controller in cluster
|
||||
|
||||
Scenario:
|
||||
1. Revert snapshot "deploy_ha_one_controller_neutron"
|
||||
2. Backup master
|
||||
3. Check backup
|
||||
4. Run OSTF
|
||||
5. Add 1 node with compute role
|
||||
6. Restore master
|
||||
7. Check restore
|
||||
8. Run OSTF
|
||||
|
||||
Duration 35m
|
||||
|
||||
"""
|
||||
self.env.revert_snapshot("deploy_ha_one_controller_neutron")
|
||||
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
|
||||
os_conn = os_actions.OpenStackActions(
|
||||
self.fuel_web.get_public_vip(cluster_id),
|
||||
'neutronOneController', 'neutronOneController',
|
||||
'neutronOneController')
|
||||
self.fuel_web.assert_cluster_ready(os_conn, smiles_count=5,
|
||||
networks_count=2, timeout=300)
|
||||
|
||||
# Execute master node backup
|
||||
self.fuel_web.backup_master(self.env.d_env.get_admin_remote())
|
||||
|
||||
# Check created backup
|
||||
checkers.backup_check(self.env.d_env.get_admin_remote())
|
||||
|
||||
self.fuel_web.update_nodes(
|
||||
cluster_id, {'slave-03': ['compute']}, True, False)
|
||||
|
||||
assert_equal(
|
||||
3, len(self.fuel_web.client.list_cluster_nodes(cluster_id)))
|
||||
|
||||
self.fuel_web.restore_master(self.env.d_env.get_admin_remote())
|
||||
checkers.restore_check_sum(self.env.d_env.get_admin_remote())
|
||||
self.fuel_web.restore_check_nailgun_api(
|
||||
self.env.d_env.get_admin_remote())
|
||||
checkers.iptables_check(self.env.d_env.get_admin_remote())
|
||||
|
||||
assert_equal(
|
||||
2, len(self.fuel_web.client.list_cluster_nodes(cluster_id)))
|
||||
|
||||
self.fuel_web.update_nodes(
|
||||
cluster_id, {'slave-03': ['compute']}, True, False)
|
||||
self.fuel_web.deploy_cluster_wait(cluster_id)
|
||||
|
||||
self.fuel_web.run_ostf(
|
||||
cluster_id=cluster_id)
|
||||
|
||||
self.env.make_snapshot("ha_one_controller_backup_restore")
|
||||
|
||||
|
||||
@test(groups=["thread_usb"])
|
||||
class HAOneControllerNeutronUSB(HAOneControllerNeutronBase):
|
||||
"""HAOneControllerNeutronUSB.""" # TODO documentation
|
||||
|
|
|
@ -19,9 +19,12 @@ from fuelweb_test.tests.base_test_case import TestBasic
|
|||
|
||||
|
||||
class HAOneControllerNeutronBase(TestBasic):
|
||||
def deploy_ha_one_controller_neutron_base(self):
|
||||
self.env.revert_snapshot("ready_with_3_slaves")
|
||||
"""HAOneControllerNeutronBase""" # TODO documentation
|
||||
|
||||
def deploy_ha_one_controller_neutron_base(
|
||||
self, snapshot_name):
|
||||
|
||||
self.env.revert_snapshot("ready_with_3_slaves")
|
||||
data = {
|
||||
'tenant': 'neutronOneController',
|
||||
'user': 'neutronOneController',
|
||||
|
@ -55,5 +58,4 @@ class HAOneControllerNeutronBase(TestBasic):
|
|||
self.fuel_web.run_ostf(
|
||||
cluster_id=cluster_id)
|
||||
|
||||
self.env.make_snapshot("deploy_ha_one_controller_neutron",
|
||||
is_make=True)
|
||||
self.env.make_snapshot(snapshot_name, is_make=True)
|
||||
|
|
|
@ -94,28 +94,10 @@ class NeutronTun(TestBasic):
|
|||
self.env.make_snapshot("deploy_neutron_tun")
|
||||
|
||||
|
||||
@test(groups=["neutron", "ha", "ha_neutron_tun"])
|
||||
class NeutronTunHa(TestBasic):
|
||||
"""NeutronTunHa.""" # TODO documentation
|
||||
class NeutronTunHaBase(TestBasic):
|
||||
"""NeutronTunHaBase.""" # TODO documentation
|
||||
|
||||
@test(depends_on=[SetupEnvironment.prepare_slaves_5],
|
||||
groups=["deploy_neutron_tun_ha", "ha_neutron_tun"])
|
||||
@log_snapshot_after_test
|
||||
def deploy_neutron_tun_ha(self):
|
||||
"""Deploy cluster in HA mode with Neutron VXLAN
|
||||
|
||||
Scenario:
|
||||
1. Create cluster
|
||||
2. Add 3 nodes with controller role
|
||||
3. Add 2 nodes with compute role
|
||||
4. Deploy the cluster
|
||||
5. Run network verification
|
||||
6. Run OSTF
|
||||
|
||||
Duration 80m
|
||||
Snapshot deploy_neutron_tun_ha
|
||||
|
||||
"""
|
||||
def deploy_neutron_tun_ha_base(self, snapshot_name):
|
||||
self.env.revert_snapshot("ready_with_5_slaves")
|
||||
|
||||
cluster_id = self.fuel_web.create_cluster(
|
||||
|
@ -165,7 +147,32 @@ class NeutronTunHa(TestBasic):
|
|||
cluster_id=cluster_id,
|
||||
test_sets=['ha', 'smoke', 'sanity'])
|
||||
|
||||
self.env.make_snapshot("deploy_neutron_tun_ha")
|
||||
self.env.make_snapshot(snapshot_name)
|
||||
|
||||
|
||||
@test(groups=["neutron", "ha", "ha_neutron_tun"])
|
||||
class NeutronTunHa(NeutronTunHaBase):
|
||||
"""NeutronTunHa.""" # TODO documentation
|
||||
|
||||
@test(depends_on=[SetupEnvironment.prepare_slaves_5],
|
||||
groups=["deploy_neutron_tun_ha", "ha_neutron_tun"])
|
||||
@log_snapshot_after_test
|
||||
def deploy_neutron_tun_ha(self):
|
||||
"""Deploy cluster in HA mode with Neutron VXLAN
|
||||
|
||||
Scenario:
|
||||
1. Create cluster
|
||||
2. Add 3 nodes with controller role
|
||||
3. Add 2 nodes with compute role
|
||||
4. Deploy the cluster
|
||||
5. Run network verification
|
||||
6. Run OSTF
|
||||
|
||||
Duration 80m
|
||||
Snapshot deploy_neutron_tun_ha
|
||||
"""
|
||||
super(self.__class__, self).deploy_neutron_tun_ha_base(
|
||||
snapshot_name="deploy_neutron_tun_ha")
|
||||
|
||||
|
||||
@test(groups=["ha", "ha_neutron_tun"])
|
||||
|
@ -399,81 +406,3 @@ class TestHaNeutronScalability(TestBasic):
|
|||
test_sets=['sanity', 'smoke', 'ha'], should_fail=1)
|
||||
self.env.sync_time()
|
||||
self.env.make_snapshot("neutron_vlan_ha_scalability")
|
||||
|
||||
|
||||
@test(groups=["known_issues", "ha", "ha_neutron_tun"])
|
||||
class BackupRestoreHa(TestBasic):
|
||||
"""BackupRestoreHa.""" # TODO documentation
|
||||
|
||||
@test(depends_on_groups=['deploy_neutron_tun_ha'],
|
||||
groups=["known_issues", "backup_restore_neutron_tun_ha"])
|
||||
@log_snapshot_after_test
|
||||
def backup_restore_neutron_tun_ha(self):
|
||||
"""Backup/restore master node with cluster in ha mode
|
||||
|
||||
Scenario:
|
||||
1. Revert snapshot "deploy_neutron_tun_ha"
|
||||
2. Backup master
|
||||
3. Check backup
|
||||
4. Run OSTF
|
||||
5. Add 1 node with compute role
|
||||
6. Restore master
|
||||
7. Check restore
|
||||
8. Run OSTF
|
||||
|
||||
Duration 50m
|
||||
|
||||
"""
|
||||
self.env.revert_snapshot("deploy_neutron_tun_ha")
|
||||
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
cluster = self.fuel_web.client.get_cluster(cluster_id)
|
||||
assert_equal(str(cluster['net_provider']), 'neutron')
|
||||
os_conn = os_actions.OpenStackActions(
|
||||
self.fuel_web.get_public_vip(cluster_id))
|
||||
# assert_equal(str(cluster['net_segment_type']), segment_type)
|
||||
self.fuel_web.check_fixed_network_cidr(
|
||||
cluster_id, os_conn)
|
||||
|
||||
with self.env.d_env.get_admin_remote() as remote:
|
||||
self.fuel_web.backup_master(remote)
|
||||
|
||||
with self.env.d_env.get_admin_remote() as remote:
|
||||
checkers.backup_check(remote)
|
||||
|
||||
self.env.bootstrap_nodes(
|
||||
self.env.d_env.nodes().slaves[5:6])
|
||||
self.fuel_web.update_nodes(
|
||||
cluster_id, {'slave-06': ['compute']}, True, False
|
||||
)
|
||||
|
||||
assert_equal(
|
||||
6, len(self.fuel_web.client.list_cluster_nodes(cluster_id)))
|
||||
|
||||
with self.env.d_env.get_admin_remote() as remote:
|
||||
self.fuel_web.restore_master(remote)
|
||||
|
||||
with self.env.d_env.get_admin_remote() as remote:
|
||||
checkers.restore_check_sum(remote)
|
||||
|
||||
with self.env.d_env.get_admin_remote() as remote:
|
||||
self.fuel_web.restore_check_nailgun_api(remote)
|
||||
|
||||
with self.env.d_env.get_admin_remote() as remote:
|
||||
checkers.iptables_check(remote)
|
||||
|
||||
assert_equal(
|
||||
5, len(self.fuel_web.client.list_cluster_nodes(cluster_id)))
|
||||
|
||||
self.env.bootstrap_nodes(
|
||||
self.env.d_env.nodes().slaves[5:6])
|
||||
self.fuel_web.update_nodes(
|
||||
cluster_id, {'slave-06': ['compute']}, True, False
|
||||
)
|
||||
self.fuel_web.deploy_cluster_wait(cluster_id)
|
||||
|
||||
self.fuel_web.run_ostf(
|
||||
cluster_id=cluster_id,
|
||||
test_sets=['ha', 'smoke', 'sanity'])
|
||||
|
||||
self.env.make_snapshot("backup_restore_neutron_tun_ha")
|
||||
|
|
|
@ -70,8 +70,7 @@ class UbuntuBootstrap(TestBasic):
|
|||
cmd = 'fuel-bootstrap-image-set ubuntu'
|
||||
run_on_remote(remote, cmd)
|
||||
|
||||
# Should be removed after Bug#1482242 will be fixed
|
||||
with self.env.d_env.get_admin_remote() as remote:
|
||||
# Should be removed after Bug#1482242 will be fixed
|
||||
cmd = 'dockerctl shell cobbler service dnsmasq restart'
|
||||
run_on_remote(remote, cmd)
|
||||
|
||||
|
|
Loading…
Reference in New Issue