Merge "Replacing all hard coded cluster status using cluster_utils"

This commit is contained in:
Jenkins 2015-12-15 07:31:26 +00:00 committed by Gerrit Code Review
commit 600032aca6
4 changed files with 23 additions and 15 deletions

View File

@ -25,6 +25,10 @@ from tempest.scenario import manager
TEMPEST_CONF = config.CONF
# cluster status
CLUSTER_STATUS_ACTIVE = "Active"
CLUSTER_STATUS_ERROR = "Error"
LOG = logging.getLogger(__name__)
@ -218,15 +222,16 @@ class BaseDataProcessingTest(manager.ScenarioTest):
s_time = timeutils.utcnow()
while timeutils.delta_seconds(s_time, timeutils.utcnow()) < timeout:
cluster = self.client.clusters.get(cluster_id)
if cluster.status == 'Active':
if cluster.status == CLUSTER_STATUS_ACTIVE:
return
if cluster.status == 'Error':
if cluster.status == CLUSTER_STATUS_ERROR:
raise exceptions.BuildErrorException(
'Cluster failed to build and is in "Error" status.')
'Cluster failed to build and is in %s status.' %
CLUSTER_STATUS_ERROR)
time.sleep(TEMPEST_CONF.data_processing.request_timeout)
raise exceptions.TimeoutException(
'Cluster failed to get to "Active status within %d seconds.'
% timeout)
'Cluster failed to get to %s status within %d seconds.'
% (CLUSTER_STATUS_ACTIVE, timeout))
def create_job_execution(self, **kwargs):

View File

@ -131,7 +131,7 @@ class TestPeriodicBack(base.SaharaWithDbTestCase):
timeutils.set_time_override(datetime.datetime(2005, 2, 1, second=0))
self._make_cluster('1', status='Pending')
self._make_cluster('1', c_u.CLUSTER_STATUS_SPAWNING)
timeutils.set_time_override(datetime.datetime(
2005, 2, 1, minute=59, second=50))
@ -147,7 +147,7 @@ class TestPeriodicBack(base.SaharaWithDbTestCase):
self.override_config('cleanup_time_for_incomplete_clusters', 1)
timeutils.set_time_override(datetime.datetime(2005, 2, 1, second=0))
self._make_cluster('1', status='Pending')
self._make_cluster('1', c_u.CLUSTER_STATUS_SPAWNING)
timeutils.set_time_override(datetime.datetime(
2005, 2, 1, hour=1, second=10))

View File

@ -68,14 +68,16 @@ class TestScalingValidation(u.ValidationTestCase):
ops.get_engine_type_and_version.return_value = "direct.1.1"
ng1 = tu.make_ng_dict('ng', '42', ['namenode'], 1)
cluster = tu.create_cluster("cluster1", "tenant1", "fake", "0.1",
[ng1], status='Validating', id='12321')
[ng1],
status=c_u.CLUSTER_STATUS_VALIDATING,
id='12321')
self._assert_check_scaling(
data={}, cluster=cluster,
expected_message="Cluster cannot be scaled "
"not in 'Active' "
"status. Cluster status: "
"Validating")
"not in '" + c_u.CLUSTER_STATUS_ACTIVE +
"' status. Cluster status: " +
c_u.CLUSTER_STATUS_VALIDATING)
cluster = tu.create_cluster("cluster1", "tenant1", "fake", "0.1",
[ng1], status=c_u.CLUSTER_STATUS_ACTIVE,

View File

@ -35,11 +35,12 @@ class UtilsClusterTest(base.SaharaWithDbTestCase):
def test_change_cluster_status(self):
cluster = self._make_sample()
cluster = cluster_utils.change_cluster_status(
cluster, "Deleting", "desc")
self.assertEqual("Deleting", cluster.status)
cluster, cluster_utils.CLUSTER_STATUS_DELETING, "desc")
self.assertEqual(cluster_utils.CLUSTER_STATUS_DELETING, cluster.status)
self.assertEqual("desc", cluster.status_description)
cluster_utils.change_cluster_status(cluster, "Spawning")
self.assertEqual("Deleting", cluster.status)
cluster_utils.change_cluster_status(
cluster, cluster_utils.CLUSTER_STATUS_SPAWNING)
self.assertEqual(cluster_utils.CLUSTER_STATUS_DELETING, cluster.status)
def test_change_status_description(self):
ctx = context.ctx()