From 4ff5776653c330d0413616590371a39397adf01c Mon Sep 17 00:00:00 2001 From: zhangzhang <zhangzh.fnst@cn.fujitsu.com> Date: Mon, 14 Nov 2016 20:46:09 -0500 Subject: [PATCH] update wait_for to wait_for_status update all atomic action Change-Id: I8dbac12470f81c6ddf30c45e31ead8798d5f7f84 --- .../openstack/scenarios/ceilometer/utils.py | 10 +++--- .../openstack/scenarios/cinder/utils.py | 16 +++++----- .../plugins/openstack/scenarios/ec2/utils.py | 2 +- .../plugins/openstack/scenarios/heat/utils.py | 14 ++++----- .../openstack/scenarios/manila/utils.py | 2 +- .../openstack/scenarios/murano/utils.py | 2 +- .../openstack/scenarios/sahara/utils.py | 2 +- .../scenarios/ceilometer/test_utils.py | 5 +-- .../openstack/scenarios/cinder/test_utils.py | 31 +++++++++++-------- .../openstack/scenarios/ec2/test_utils.py | 2 +- .../openstack/scenarios/heat/test_utils.py | 17 +++++----- .../openstack/scenarios/manila/test_utils.py | 2 +- .../openstack/scenarios/murano/test_utils.py | 2 +- 13 files changed, 57 insertions(+), 50 deletions(-) diff --git a/rally/plugins/openstack/scenarios/ceilometer/utils.py b/rally/plugins/openstack/scenarios/ceilometer/utils.py index 0ab0e477..ff2ea7ee 100644 --- a/rally/plugins/openstack/scenarios/ceilometer/utils.py +++ b/rally/plugins/openstack/scenarios/ceilometer/utils.py @@ -259,11 +259,11 @@ class CeilometerScenario(scenario.OpenStackScenario): :returns: alarm in the set state """ self.clients("ceilometer").alarms.set_state(alarm.alarm_id, state) - return bench_utils.wait_for(alarm, - ready_statuses=[state], - update_resource=bench_utils - .get_from_manager(), - timeout=timeout, check_interval=1) + return bench_utils.wait_for_status(alarm, + ready_statuses=[state], + update_resource=bench_utils + .get_from_manager(), + timeout=timeout, check_interval=1) @atomic.action_timer("ceilometer.list_events") def _list_events(self): diff --git a/rally/plugins/openstack/scenarios/cinder/utils.py b/rally/plugins/openstack/scenarios/cinder/utils.py index 0e43e954..662f23f4 100644 --- a/rally/plugins/openstack/scenarios/cinder/utils.py +++ b/rally/plugins/openstack/scenarios/cinder/utils.py @@ -163,7 +163,7 @@ class CinderScenario(scenario.OpenStackScenario): # check whether the volume is ready => less API calls. self.sleep_between(CONF.openstack.cinder_volume_create_prepoll_delay) - volume = bench_utils.wait_for( + volume = bench_utils.wait_for_status( volume, ready_statuses=["available"], update_resource=bench_utils.get_from_manager(), @@ -234,7 +234,7 @@ class CinderScenario(scenario.OpenStackScenario): new_size = random.randint(new_size["min"], new_size["max"]) volume.extend(volume, new_size) - volume = bench_utils.wait_for( + volume = bench_utils.wait_for_status( volume, ready_statuses=["available"], update_resource=bench_utils.get_from_manager(), @@ -262,7 +262,7 @@ class CinderScenario(scenario.OpenStackScenario): container_format, disk_format) # NOTE (e0ne): upload_to_image changes volume status to uploading so # we need to wait until it will be available. - volume = bench_utils.wait_for( + volume = bench_utils.wait_for_status( volume, ready_statuses=["available"], update_resource=bench_utils.get_from_manager(), @@ -272,7 +272,7 @@ class CinderScenario(scenario.OpenStackScenario): image_id = img["os-volume_upload_image"]["image_id"] image = self.clients("glance").images.get(image_id) wrapper = glance_wrapper.wrap(self._clients.glance, self) - image = bench_utils.wait_for( + image = bench_utils.wait_for_status( image, ready_statuses=["active"], update_resource=wrapper.get_image, @@ -301,7 +301,7 @@ class CinderScenario(scenario.OpenStackScenario): snapshot = client.create_snapshot(volume_id, **kwargs) self.sleep_between(CONF.openstack.cinder_volume_create_prepoll_delay) - snapshot = bench_utils.wait_for( + snapshot = bench_utils.wait_for_status( snapshot, ready_statuses=["available"], update_resource=bench_utils.get_from_manager(), @@ -336,7 +336,7 @@ class CinderScenario(scenario.OpenStackScenario): :param kwargs: Other optional parameters """ backup = self.clients("cinder").backups.create(volume_id, **kwargs) - return bench_utils.wait_for( + return bench_utils.wait_for_status( backup, ready_statuses=["available"], update_resource=bench_utils.get_from_manager(), @@ -372,14 +372,14 @@ class CinderScenario(scenario.OpenStackScenario): restore = self.clients("cinder").restores.restore(backup_id, volume_id) restored_volume = self.clients("cinder").volumes.get(restore.volume_id) backup_for_restore = self.clients("cinder").backups.get(backup_id) - bench_utils.wait_for( + bench_utils.wait_for_status( backup_for_restore, ready_statuses=["available"], update_resource=bench_utils.get_from_manager(), timeout=CONF.openstack.cinder_backup_restore_timeout, check_interval=CONF.openstack.cinder_backup_restore_poll_interval ) - return bench_utils.wait_for( + return bench_utils.wait_for_status( restored_volume, ready_statuses=["available"], update_resource=bench_utils.get_from_manager(), diff --git a/rally/plugins/openstack/scenarios/ec2/utils.py b/rally/plugins/openstack/scenarios/ec2/utils.py index c6c4b721..da661b6c 100644 --- a/rally/plugins/openstack/scenarios/ec2/utils.py +++ b/rally/plugins/openstack/scenarios/ec2/utils.py @@ -55,7 +55,7 @@ class EC2Scenario(scenario.OpenStackScenario): servers = [instance for instance in reservation.instances] self.sleep_between(CONF.openstack.ec2_server_boot_prepoll_delay) - servers = [utils.wait_for( + servers = [utils.wait_for_status( server, ready_statuses=["RUNNING"], update_resource=self._update_resource, diff --git a/rally/plugins/openstack/scenarios/heat/utils.py b/rally/plugins/openstack/scenarios/heat/utils.py index 0534e308..7d6184c0 100644 --- a/rally/plugins/openstack/scenarios/heat/utils.py +++ b/rally/plugins/openstack/scenarios/heat/utils.py @@ -67,7 +67,7 @@ class HeatScenario(scenario.OpenStackScenario): self.sleep_between(CONF.openstack.heat_stack_create_prepoll_delay) - stack = utils.wait_for( + stack = utils.wait_for_status( stack, ready_statuses=["CREATE_COMPLETE"], failure_statuses=["CREATE_FAILED", "ERROR"], @@ -103,7 +103,7 @@ class HeatScenario(scenario.OpenStackScenario): self.sleep_between(CONF.openstack.heat_stack_update_prepoll_delay) - stack = utils.wait_for( + stack = utils.wait_for_status( stack, ready_statuses=["UPDATE_COMPLETE"], failure_statuses=["UPDATE_FAILED", "ERROR"], @@ -121,7 +121,7 @@ class HeatScenario(scenario.OpenStackScenario): :param stack: stack that needs to be checked """ self.clients("heat").actions.check(stack.id) - utils.wait_for( + utils.wait_for_status( stack, ready_statuses=["CHECK_COMPLETE"], failure_statuses=["CHECK_FAILED", "ERROR"], @@ -155,7 +155,7 @@ class HeatScenario(scenario.OpenStackScenario): """ self.clients("heat").actions.suspend(stack.id) - utils.wait_for( + utils.wait_for_status( stack, ready_statuses=["SUSPEND_COMPLETE"], failure_statuses=["SUSPEND_FAILED", "ERROR"], @@ -171,7 +171,7 @@ class HeatScenario(scenario.OpenStackScenario): """ self.clients("heat").actions.resume(stack.id) - utils.wait_for( + utils.wait_for_status( stack, ready_statuses=["RESUME_COMPLETE"], failure_statuses=["RESUME_FAILED", "ERROR"], @@ -188,7 +188,7 @@ class HeatScenario(scenario.OpenStackScenario): """ snapshot = self.clients("heat").stacks.snapshot( stack.id) - utils.wait_for( + utils.wait_for_status( stack, ready_statuses=["SNAPSHOT_COMPLETE"], failure_statuses=["SNAPSHOT_FAILED", "ERROR"], @@ -205,7 +205,7 @@ class HeatScenario(scenario.OpenStackScenario): :param snapshot_id: id of given snapshot """ self.clients("heat").stacks.restore(stack.id, snapshot_id) - utils.wait_for( + utils.wait_for_status( stack, ready_statuses=["RESTORE_COMPLETE"], failure_statuses=["RESTORE_FAILED", "ERROR"], diff --git a/rally/plugins/openstack/scenarios/manila/utils.py b/rally/plugins/openstack/scenarios/manila/utils.py index 7ad4cf32..555795f1 100644 --- a/rally/plugins/openstack/scenarios/manila/utils.py +++ b/rally/plugins/openstack/scenarios/manila/utils.py @@ -61,7 +61,7 @@ class ManilaScenario(scenario.OpenStackScenario): share_proto, size, **kwargs) self.sleep_between(CONF.openstack.manila_share_create_prepoll_delay) - share = utils.wait_for( + share = utils.wait_for_status( share, ready_statuses=["available"], update_resource=utils.get_from_manager(), diff --git a/rally/plugins/openstack/scenarios/murano/utils.py b/rally/plugins/openstack/scenarios/murano/utils.py index c9cf79e6..49d7c01f 100644 --- a/rally/plugins/openstack/scenarios/murano/utils.py +++ b/rally/plugins/openstack/scenarios/murano/utils.py @@ -102,7 +102,7 @@ class MuranoScenario(scenario.OpenStackScenario): session.id) config = CONF.openstack - utils.wait_for( + utils.wait_for_status( environment, ready_statuses=["READY"], update_resource=utils.get_from_manager(["DEPLOY FAILURE"]), diff --git a/rally/plugins/openstack/scenarios/sahara/utils.py b/rally/plugins/openstack/scenarios/sahara/utils.py index 3a1ccd8e..7ad8d5ca 100644 --- a/rally/plugins/openstack/scenarios/sahara/utils.py +++ b/rally/plugins/openstack/scenarios/sahara/utils.py @@ -107,7 +107,7 @@ class SaharaScenario(scenario.OpenStackScenario): self.clients("sahara").node_group_templates.delete(node_group.id) def _wait_active(self, cluster_object): - utils.wait_for( + utils.wait_for_status( resource=cluster_object, ready_statuses=["active"], failure_statuses=["error"], update_resource=self._update_cluster, timeout=CONF.openstack.sahara_cluster_create_timeout, diff --git a/tests/unit/plugins/openstack/scenarios/ceilometer/test_utils.py b/tests/unit/plugins/openstack/scenarios/ceilometer/test_utils.py index 273a6059..bf1e76fe 100644 --- a/tests/unit/plugins/openstack/scenarios/ceilometer/test_utils.py +++ b/tests/unit/plugins/openstack/scenarios/ceilometer/test_utils.py @@ -175,13 +175,14 @@ class CeilometerScenarioTestCase(test.ScenarioTestCase): alarm = mock.Mock() self.clients("ceilometer").alarms.create.return_value = alarm return_alarm = self.scenario._set_alarm_state(alarm, "ok", 100) - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( alarm, ready_statuses=["ok"], update_resource=self.mock_get_from_manager.mock.return_value, timeout=100, check_interval=1) self.mock_get_from_manager.mock.assert_called_once_with() - self.assertEqual(self.mock_wait_for.mock.return_value, return_alarm) + self.assertEqual(self.mock_wait_for_status.mock.return_value, + return_alarm) self._test_atomic_action_timer(self.scenario.atomic_actions(), "ceilometer.set_alarm_state") diff --git a/tests/unit/plugins/openstack/scenarios/cinder/test_utils.py b/tests/unit/plugins/openstack/scenarios/cinder/test_utils.py index df5d739a..576f584f 100644 --- a/tests/unit/plugins/openstack/scenarios/cinder/test_utils.py +++ b/tests/unit/plugins/openstack/scenarios/cinder/test_utils.py @@ -148,7 +148,7 @@ class CinderScenarioTestCase(test.ScenarioTestCase): def test__create_volume(self): return_volume = self.scenario._create_volume(1) - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( self.mock_wrap.return_value.create_volume.return_value, ready_statuses=["available"], update_resource=self.mock_get_from_manager.mock.return_value, @@ -156,7 +156,8 @@ class CinderScenarioTestCase(test.ScenarioTestCase): check_interval=CONF.openstack.cinder_volume_create_poll_interval ) self.mock_get_from_manager.mock.assert_called_once_with() - self.assertEqual(self.mock_wait_for.mock.return_value, return_volume) + self.assertEqual(self.mock_wait_for_status.mock.return_value, + return_volume) self._test_atomic_action_timer(self.scenario.atomic_actions(), "cinder.create_volume") @@ -171,7 +172,7 @@ class CinderScenarioTestCase(test.ScenarioTestCase): self.mock_wrap.return_value.create_volume.assert_called_once_with( 3, display_name="TestVolume") - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( self.mock_wrap.return_value.create_volume.return_value, ready_statuses=["available"], update_resource=self.mock_get_from_manager.mock.return_value, @@ -179,7 +180,8 @@ class CinderScenarioTestCase(test.ScenarioTestCase): check_interval=CONF.openstack.cinder_volume_create_poll_interval ) self.mock_get_from_manager.mock.assert_called_once_with() - self.assertEqual(self.mock_wait_for.mock.return_value, return_volume) + self.assertEqual(self.mock_wait_for_status.mock.return_value, + return_volume) self._test_atomic_action_timer(self.scenario.atomic_actions(), "cinder.create_volume") @@ -230,7 +232,7 @@ class CinderScenarioTestCase(test.ScenarioTestCase): self.scenario._extend_volume(volume, new_size={"min": 1, "max": 5}) volume.extend.assert_called_once_with(volume, 3) - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( volume, ready_statuses=["available"], update_resource=self.mock_get_from_manager.mock.return_value, @@ -245,7 +247,7 @@ class CinderScenarioTestCase(test.ScenarioTestCase): volume = mock.Mock() self.clients("cinder").volumes.extend.return_value = volume self.scenario._extend_volume(volume, 2) - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( volume, ready_statuses=["available"], update_resource=self.mock_get_from_manager.mock.return_value, @@ -270,7 +272,7 @@ class CinderScenarioTestCase(test.ScenarioTestCase): volume.upload_to_image.assert_called_once_with(False, "test_vol", "container", "disk") - self.mock_wait_for.mock.assert_has_calls([ + self.mock_wait_for_status.mock.assert_has_calls([ mock.call( volume, ready_statuses=["available"], @@ -292,7 +294,7 @@ class CinderScenarioTestCase(test.ScenarioTestCase): def test__create_snapshot(self): return_snapshot = self.scenario._create_snapshot("uuid", False) - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( self.mock_wrap.return_value.create_snapshot.return_value, ready_statuses=["available"], update_resource=self.mock_get_from_manager.mock.return_value, @@ -300,7 +302,8 @@ class CinderScenarioTestCase(test.ScenarioTestCase): check_interval=cfg.CONF.openstack .cinder_volume_create_poll_interval) self.mock_get_from_manager.mock.assert_called_once_with() - self.assertEqual(self.mock_wait_for.mock.return_value, return_snapshot) + self.assertEqual(self.mock_wait_for_status.mock.return_value, + return_snapshot) self._test_atomic_action_timer(self.scenario.atomic_actions(), "cinder.create_snapshot") @@ -323,7 +326,7 @@ class CinderScenarioTestCase(test.ScenarioTestCase): def test__create_backup(self): return_backup = self.scenario._create_backup("uuid") - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( self.clients("cinder").backups.create.return_value, ready_statuses=["available"], update_resource=self.mock_get_from_manager.mock.return_value, @@ -331,7 +334,8 @@ class CinderScenarioTestCase(test.ScenarioTestCase): check_interval=cfg.CONF.openstack .cinder_volume_create_poll_interval) self.mock_get_from_manager.mock.assert_called_once_with() - self.assertEqual(self.mock_wait_for.mock.return_value, return_backup) + self.assertEqual(self.mock_wait_for_status.mock.return_value, + return_backup) self._test_atomic_action_timer(self.scenario.atomic_actions(), "cinder.create_backup") @@ -362,7 +366,7 @@ class CinderScenarioTestCase(test.ScenarioTestCase): self.clients("cinder").volumes.get.return_value = restore return_restore = self.scenario._restore_backup(backup.id, None) - self.mock_wait_for.mock.assert_has_calls([ + self.mock_wait_for_status.mock.assert_has_calls([ mock.call( backup, ready_statuses=["available"], @@ -379,7 +383,8 @@ class CinderScenarioTestCase(test.ScenarioTestCase): self.mock_get_from_manager.mock.assert_has_calls([mock.call(), mock.call()]) - self.assertEqual(self.mock_wait_for.mock.return_value, return_restore) + self.assertEqual(self.mock_wait_for_status.mock.return_value, + return_restore) self._test_atomic_action_timer(self.scenario.atomic_actions(), "cinder.restore_backup") diff --git a/tests/unit/plugins/openstack/scenarios/ec2/test_utils.py b/tests/unit/plugins/openstack/scenarios/ec2/test_utils.py index ea347728..e988f9bb 100644 --- a/tests/unit/plugins/openstack/scenarios/ec2/test_utils.py +++ b/tests/unit/plugins/openstack/scenarios/ec2/test_utils.py @@ -66,6 +66,6 @@ class EC2ScenarioTestCase(test.ScenarioTestCase): timeout=CONF.openstack.ec2_server_boot_timeout ) ] - self.mock_wait_for.mock.assert_has_calls(expected) + self.mock_wait_for_status.mock.assert_has_calls(expected) self._test_atomic_action_timer(ec2_scenario.atomic_actions(), "ec2.boot_servers") diff --git a/tests/unit/plugins/openstack/scenarios/heat/test_utils.py b/tests/unit/plugins/openstack/scenarios/heat/test_utils.py index 738e11b2..0bc7e30f 100644 --- a/tests/unit/plugins/openstack/scenarios/heat/test_utils.py +++ b/tests/unit/plugins/openstack/scenarios/heat/test_utils.py @@ -58,7 +58,7 @@ class HeatScenarioTestCase(test.ScenarioTestCase): self.assertIn(self.default_template, kwargs.values()) self.assertIn(self.dummy_files, kwargs.values()) self.assertIn(self.dummy_environment, kwargs.values()) - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( self.stack, update_resource=self.mock_get_from_manager.mock.return_value, ready_statuses=["CREATE_COMPLETE"], @@ -66,7 +66,8 @@ class HeatScenarioTestCase(test.ScenarioTestCase): check_interval=CONF.openstack.heat_stack_create_poll_interval, timeout=CONF.openstack.heat_stack_create_timeout) self.mock_get_from_manager.mock.assert_called_once_with() - self.assertEqual(self.mock_wait_for.mock.return_value, return_stack) + self.assertEqual(self.mock_wait_for_status.mock.return_value, + return_stack) self._test_atomic_action_timer(self.scenario.atomic_actions(), "heat.create_stack") @@ -82,7 +83,7 @@ class HeatScenarioTestCase(test.ScenarioTestCase): self.assertIn(self.dummy_files, kwargs.values()) self.assertIn(self.dummy_environment, kwargs.values()) self.assertIn(self.stack.id, args) - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( self.stack, update_resource=self.mock_get_from_manager.mock.return_value, ready_statuses=["UPDATE_COMPLETE"], @@ -98,7 +99,7 @@ class HeatScenarioTestCase(test.ScenarioTestCase): scenario._check_stack(self.stack) self.clients("heat").actions.check.assert_called_once_with( self.stack.id) - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( self.stack, update_resource=self.mock_get_from_manager.mock.return_value, ready_statuses=["CHECK_COMPLETE"], @@ -129,7 +130,7 @@ class HeatScenarioTestCase(test.ScenarioTestCase): scenario._suspend_stack(self.stack) self.clients("heat").actions.suspend.assert_called_once_with( self.stack.id) - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( self.stack, update_resource=self.mock_get_from_manager.mock.return_value, ready_statuses=["SUSPEND_COMPLETE"], @@ -145,7 +146,7 @@ class HeatScenarioTestCase(test.ScenarioTestCase): scenario._resume_stack(self.stack) self.clients("heat").actions.resume.assert_called_once_with( self.stack.id) - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( self.stack, update_resource=self.mock_get_from_manager.mock.return_value, ready_statuses=["RESUME_COMPLETE"], @@ -161,7 +162,7 @@ class HeatScenarioTestCase(test.ScenarioTestCase): scenario._snapshot_stack(self.stack) self.clients("heat").stacks.snapshot.assert_called_once_with( self.stack.id) - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( self.stack, update_resource=self.mock_get_from_manager.mock.return_value, ready_statuses=["SNAPSHOT_COMPLETE"], @@ -177,7 +178,7 @@ class HeatScenarioTestCase(test.ScenarioTestCase): scenario._restore_stack(self.stack, "dummy_id") self.clients("heat").stacks.restore.assert_called_once_with( self.stack.id, "dummy_id") - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( self.stack, update_resource=self.mock_get_from_manager.mock.return_value, ready_statuses=["RESTORE_COMPLETE"], diff --git a/tests/unit/plugins/openstack/scenarios/manila/test_utils.py b/tests/unit/plugins/openstack/scenarios/manila/test_utils.py index d4d48c1f..573f5c3b 100644 --- a/tests/unit/plugins/openstack/scenarios/manila/test_utils.py +++ b/tests/unit/plugins/openstack/scenarios/manila/test_utils.py @@ -53,7 +53,7 @@ class ManilaScenarioTestCase(test.ScenarioTestCase): share_network=self.scenario.context["tenant"][ consts.SHARE_NETWORKS_CONTEXT_NAME]["share_networks"][0]["id"]) - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( fake_share, ready_statuses=["available"], update_resource=self.mock_get_from_manager.mock.return_value, diff --git a/tests/unit/plugins/openstack/scenarios/murano/test_utils.py b/tests/unit/plugins/openstack/scenarios/murano/test_utils.py index ff2abc53..6917ed1e 100644 --- a/tests/unit/plugins/openstack/scenarios/murano/test_utils.py +++ b/tests/unit/plugins/openstack/scenarios/murano/test_utils.py @@ -89,7 +89,7 @@ class MuranoScenarioTestCase(test.ScenarioTestCase): ) config = CONF.openstack - self.mock_wait_for.mock.assert_called_once_with( + self.mock_wait_for_status.mock.assert_called_once_with( environment, update_resource=self.mock_get_from_manager.mock.return_value, ready_statuses=["READY"],