Fix minor codestyle violations
Change-Id: I3598bea411649f181457ecc3889190a902feaa91 Related-Bug: #1482643
This commit is contained in:
parent
7a4d175249
commit
a269b98dec
@ -1007,8 +1007,8 @@ def check_repo_managment(remote):
|
||||
run 'yum -y clean all && yum check-update' or
|
||||
'apt-get clean all && apt-get update' exit code should be 0
|
||||
|
||||
:type devops_node: Node
|
||||
:rtype True or False
|
||||
:type remote: SSHClient
|
||||
:rtype Dict
|
||||
"""
|
||||
if OPENSTACK_RELEASE == OPENSTACK_RELEASE_UBUNTU:
|
||||
cmd = "apt-get clean all && apt-get update > /dev/null"
|
||||
@ -1062,7 +1062,7 @@ def check_haproxy_backend(remote,
|
||||
all nodes. Ignoring has a bigger priority.
|
||||
|
||||
:type remote: SSHClient
|
||||
:type service: List
|
||||
:type services: List
|
||||
:type nodes: List
|
||||
:type ignore_services: List
|
||||
:type ignore_nodes: List
|
||||
|
@ -48,8 +48,8 @@ class RallyEngine(object):
|
||||
return [self.container_repo, tag] in existing_images
|
||||
|
||||
def pull_image(self):
|
||||
#TODO(apanchenko): add possibility to load image from local path or
|
||||
#remote link provided in settings, in order to speed up downloading
|
||||
# TODO(apanchenko): add possibility to load image from local path or
|
||||
# remote link provided in settings, in order to speed up downloading
|
||||
cmd = 'docker pull {0}'.format(self.container_repo)
|
||||
logger.debug('Downloading Rally repository/image from registry...')
|
||||
result = self.admin_remote.execute(cmd)
|
||||
|
@ -963,7 +963,7 @@ class FuelWebClient(object):
|
||||
def get_devops_nodes_by_nailgun_nodes(self, nailgun_nodes):
|
||||
"""Return devops node by nailgun node
|
||||
|
||||
:type nailgun_node: List
|
||||
:type nailgun_nodes: List
|
||||
:rtype: list of Nodes or None
|
||||
"""
|
||||
d_nodes = [self.get_devops_node_by_nailgun_node(n) for n
|
||||
|
@ -489,7 +489,7 @@ class ContrailPlugin(TestBasic):
|
||||
|
||||
self.fuel_web.deploy_cluster_wait(self.cluster_id)
|
||||
|
||||
#TODO
|
||||
# TODO:
|
||||
# Tests using north-south connectivity are expected to fail because
|
||||
# they require additional gateway nodes, and specific contrail
|
||||
# settings. This mark is a workaround until it's verified
|
||||
|
@ -116,12 +116,12 @@ class VipReservation(TestBasic):
|
||||
self.fuel_web.run_ostf(cluster_id=cluster_id)
|
||||
|
||||
remote = self.fuel_web.get_ssh_for_node('slave-01')
|
||||
#get vips from hiera
|
||||
# get vips from hiera
|
||||
reserved_vip_pub = remote.execute(
|
||||
'hiera reserved_vip_pub')['stdout'][1].split('"')[3]
|
||||
reserved_vip_mng = remote.execute(
|
||||
'hiera reserved_vip_mng')['stdout'][1].split('"')[3]
|
||||
#get vips from database
|
||||
# get vips from database
|
||||
reserved_vip_pub_db = self.env.postgres_actions.run_query(
|
||||
db='nailgun', query="select ip_addr from ip_addrs where "
|
||||
"vip_type = '\"'\"'reserved_vip_pub'\"'\"';")
|
||||
|
@ -528,7 +528,7 @@ class FloatingIPs(TestBasic):
|
||||
Snapshot: deploy_floating_ips
|
||||
|
||||
"""
|
||||
#Test should be re-worked for neutron according to LP#1481322
|
||||
# Test should be re-worked for neutron according to LP#1481322
|
||||
self.env.revert_snapshot("ready_with_3_slaves")
|
||||
|
||||
cluster_id = self.fuel_web.create_cluster(
|
||||
|
@ -214,7 +214,7 @@ class PatchingTests(TestBasic):
|
||||
self.fuel_web.deploy_cluster_wait(cluster_id,
|
||||
check_services=False)
|
||||
self.fuel_web.verify_network(cluster_id)
|
||||
#sanity set isn't running due to LP1457515
|
||||
# sanity set isn't running due to LP1457515
|
||||
self.fuel_web.run_ostf(cluster_id=cluster_id,
|
||||
test_sets=['smoke', 'ha'])
|
||||
|
||||
@ -231,7 +231,7 @@ class PatchingTests(TestBasic):
|
||||
wait(
|
||||
lambda: self.fuel_web.is_node_discovered(nodes[0]),
|
||||
timeout=6 * 60)
|
||||
#sanity set isn't running due to LP1457515
|
||||
# sanity set isn't running due to LP1457515
|
||||
self.fuel_web.run_ostf(cluster_id=cluster_id,
|
||||
test_sets=['smoke', 'ha'])
|
||||
|
||||
@ -372,7 +372,7 @@ class PatchingMasterTests(TestBasic):
|
||||
self.fuel_web.deploy_cluster_wait(cluster_id,
|
||||
check_services=False)
|
||||
self.fuel_web.verify_network(cluster_id)
|
||||
#sanity set isn't running due to LP1457515
|
||||
# sanity set isn't running due to LP1457515
|
||||
self.fuel_web.run_ostf(cluster_id=cluster_id,
|
||||
test_sets=['smoke', 'ha'])
|
||||
|
||||
@ -388,7 +388,7 @@ class PatchingMasterTests(TestBasic):
|
||||
wait(
|
||||
lambda: self.fuel_web.is_node_discovered(nodes[0]),
|
||||
timeout=6 * 60)
|
||||
#sanity set isn't running due to LP1457515
|
||||
# sanity set isn't running due to LP1457515
|
||||
self.fuel_web.run_ostf(cluster_id=cluster_id,
|
||||
test_sets=['smoke', 'ha'])
|
||||
|
||||
|
@ -132,7 +132,7 @@ class SeparateDbFailover(TestBasic):
|
||||
"""
|
||||
self.env.revert_snapshot("separate_db_service")
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
#destroy one db node
|
||||
# destroy one db node
|
||||
db_node = self.env.d_env.nodes().slaves[3]
|
||||
db_node.destroy()
|
||||
wait(lambda: not self.fuel_web.get_nailgun_node_by_devops_node(
|
||||
@ -161,7 +161,7 @@ class SeparateDbFailover(TestBasic):
|
||||
"""
|
||||
self.env.revert_snapshot("separate_db_service")
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
#restart one db node
|
||||
# restart one db node
|
||||
db_node = self.env.d_env.nodes().slaves[3]
|
||||
self.fuel_web.warm_restart_nodes([db_node])
|
||||
wait(lambda: self.fuel_web.get_nailgun_node_by_devops_node(
|
||||
@ -190,7 +190,7 @@ class SeparateDbFailover(TestBasic):
|
||||
"""
|
||||
self.env.revert_snapshot("separate_db_service")
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
#shutdown primary controller
|
||||
# shutdown primary controller
|
||||
controller = self.fuel_web.get_nailgun_primary_node(
|
||||
self.env.d_env.nodes().slaves[0])
|
||||
logger.debug(
|
||||
|
@ -138,7 +138,7 @@ class SeparateHorizonFailover(TestBasic):
|
||||
"""
|
||||
self.env.revert_snapshot("separate_horizon_service")
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
#destroy one horizon node
|
||||
# destroy one horizon node
|
||||
horizon_node = self.env.d_env.nodes().slaves[3]
|
||||
horizon_node.destroy()
|
||||
wait(lambda: not self.fuel_web.get_nailgun_node_by_devops_node(
|
||||
@ -164,7 +164,7 @@ class SeparateHorizonFailover(TestBasic):
|
||||
"""
|
||||
self.env.revert_snapshot("separate_horizon_service")
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
#restart one horizon node
|
||||
# restart one horizon node
|
||||
horizon_node = self.env.d_env.nodes().slaves[3]
|
||||
self.fuel_web.warm_restart_nodes([horizon_node])
|
||||
wait(lambda: self.fuel_web.get_nailgun_node_by_devops_node(
|
||||
@ -190,7 +190,7 @@ class SeparateHorizonFailover(TestBasic):
|
||||
"""
|
||||
self.env.revert_snapshot("separate_horizon_service")
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
#shutdown primary controller
|
||||
# shutdown primary controller
|
||||
controller = self.fuel_web.get_nailgun_primary_node(
|
||||
self.env.d_env.nodes().slaves[0])
|
||||
logger.debug(
|
||||
|
@ -142,7 +142,7 @@ class SeparateKeystoneFailover(TestBasic):
|
||||
"""
|
||||
self.env.revert_snapshot("separate_keystone_service")
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
#destroy one keystone node
|
||||
# destroy one keystone node
|
||||
keystone_node = self.env.d_env.nodes().slaves[3]
|
||||
keystone_node.destroy()
|
||||
wait(lambda: not self.fuel_web.get_nailgun_node_by_devops_node(
|
||||
@ -170,7 +170,7 @@ class SeparateKeystoneFailover(TestBasic):
|
||||
"""
|
||||
self.env.revert_snapshot("separate_keystone_service")
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
#restart one keystone node
|
||||
# restart one keystone node
|
||||
keystone_node = self.env.d_env.nodes().slaves[3]
|
||||
self.fuel_web.warm_restart_nodes([keystone_node])
|
||||
wait(lambda: self.fuel_web.get_nailgun_node_by_devops_node(
|
||||
@ -197,7 +197,7 @@ class SeparateKeystoneFailover(TestBasic):
|
||||
"""
|
||||
self.env.revert_snapshot("separate_keystone_service")
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
#shutdown primary controller
|
||||
# shutdown primary controller
|
||||
controller = self.fuel_web.get_nailgun_primary_node(
|
||||
self.env.d_env.nodes().slaves[0])
|
||||
logger.debug(
|
||||
|
@ -157,7 +157,7 @@ class SeparateAllFailover(TestBasic):
|
||||
"""
|
||||
self.env.revert_snapshot("separate_all_service")
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
#destroy node with rabbit master
|
||||
# destroy node with rabbit master
|
||||
all_node = self.fuel_web.get_rabbit_master_node(
|
||||
self.env.d_env.nodes().slaves[3].name)
|
||||
all_node.destroy()
|
||||
@ -186,7 +186,7 @@ class SeparateAllFailover(TestBasic):
|
||||
"""
|
||||
self.env.revert_snapshot("separate_all_service")
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
#shutdown primary controller
|
||||
# shutdown primary controller
|
||||
controller = self.fuel_web.get_nailgun_primary_node(
|
||||
self.env.d_env.nodes().slaves[0])
|
||||
logger.debug(
|
||||
|
@ -133,7 +133,7 @@ class SeparateRabbitFailover(TestBasic):
|
||||
"""
|
||||
self.env.revert_snapshot("separate_rabbit_service")
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
#destroy master rabbit node
|
||||
# destroy master rabbit node
|
||||
rabbit_node = self.fuel_web.get_rabbit_master_node(
|
||||
self.env.d_env.nodes().slaves[3].name)
|
||||
rabbit_node.destroy()
|
||||
@ -161,7 +161,7 @@ class SeparateRabbitFailover(TestBasic):
|
||||
"""
|
||||
self.env.revert_snapshot("separate_rabbit_service")
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
#restart rabbit master node
|
||||
# restart rabbit master node
|
||||
rabbit_node = self.fuel_web.get_rabbit_master_node(
|
||||
self.env.d_env.nodes().slaves[3].name)
|
||||
self.fuel_web.warm_restart_nodes([rabbit_node])
|
||||
@ -190,7 +190,7 @@ class SeparateRabbitFailover(TestBasic):
|
||||
"""
|
||||
self.env.revert_snapshot("separate_rabbit_service")
|
||||
cluster_id = self.fuel_web.get_last_created_cluster()
|
||||
#shutdown primary controller
|
||||
# shutdown primary controller
|
||||
controller = self.fuel_web.get_nailgun_primary_node(
|
||||
self.env.d_env.nodes().slaves[0])
|
||||
logger.debug(
|
||||
|
@ -524,7 +524,7 @@ class TestHaNovaFailover(TestHaFailoverBase):
|
||||
Duration
|
||||
|
||||
"""
|
||||
# TODO enable tests when fencing will be implements
|
||||
# TODO: enable tests when fencing will be implements
|
||||
super(self.__class__, self).ha_controller_loss_packages()
|
||||
|
||||
@test(depends_on_groups=['prepare_ha_nova'],
|
||||
|
@ -1001,7 +1001,7 @@ class TestHaFailoverBase(TestBasic):
|
||||
# turn off rabbit master
|
||||
master_rabbit.suspend(False)
|
||||
|
||||
# Wait until Nailgun marked suspended controller as offline
|
||||
# Wait until Nailgun marked suspended controller as offline
|
||||
try:
|
||||
wait(lambda: not self.fuel_web.get_nailgun_node_by_devops_node(
|
||||
master_rabbit)['online'], timeout=60 * 5)
|
||||
|
@ -202,7 +202,7 @@ class UpgradeFuelMaster(base_test_data.TestBasic):
|
||||
8. Run OSTF
|
||||
|
||||
"""
|
||||
#(ddmitriev)TODO: change the snapshot name to actual when reverting 7.0
|
||||
# TODO(ddmitriev): change snapshot name to actual when reverting 7.0
|
||||
if not self.env.d_env.has_snapshot('deploy_neutron_gre_ha'):
|
||||
raise SkipTest()
|
||||
|
||||
@ -291,7 +291,7 @@ class UpgradeFuelMaster(base_test_data.TestBasic):
|
||||
9. Run OSTF
|
||||
|
||||
"""
|
||||
#(ddmitriev)TODO: change the snapshot name to actual when reverting 7.0
|
||||
# TODO(ddmitriev): change snapshot name to actual when reverting 7.0
|
||||
if not self.env.d_env.has_snapshot('deploy_neutron_gre_ha'):
|
||||
raise SkipTest()
|
||||
|
||||
@ -501,7 +501,7 @@ class UpgradeFuelMaster(base_test_data.TestBasic):
|
||||
8. Run OSTF for new cluster
|
||||
9. Run network verification
|
||||
"""
|
||||
#(ddmitriev)TODO: change the snapshot name to actual when reverting 7.0
|
||||
# TODO(ddmitriev): change snapshot name to actual when reverting 7.0
|
||||
if not self.env.d_env.has_snapshot('deploy_neutron_gre'):
|
||||
raise SkipTest()
|
||||
|
||||
@ -611,7 +611,7 @@ class RollbackFuelMaster(base_test_data.TestBasic):
|
||||
9. Run OSTF
|
||||
|
||||
"""
|
||||
#(ddmitriev)TODO: change the snapshot name to actual when reverting 7.0
|
||||
# TODO(ddmitriev): change snapshot name to actual when reverting 7.0
|
||||
if not self.env.d_env.has_snapshot('deploy_neutron_gre_ha'):
|
||||
raise SkipTest()
|
||||
|
||||
@ -754,7 +754,7 @@ class RollbackFuelMaster(base_test_data.TestBasic):
|
||||
9. Run OSTF
|
||||
|
||||
"""
|
||||
#(ddmitriev)TODO: change the snapshot name to actual when reverting 7.0
|
||||
# TODO(ddmitriev): change snapshot name to actual when reverting 7.0
|
||||
if not self.env.d_env.has_snapshot('deploy_neutron_gre'):
|
||||
raise SkipTest()
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user